I proposed a similar voice transcription suggestion (and trigger options) to the Wyze AI team when they solicited suggestions for the AI here:
So yes, I have similar thoughts as you as how that could be cool.
As for visual cues, Wyze had some interesting announcements here:
But the information in that video has basically been the extent of what they’ve said to the public so far.
I don’t know enough about what it would take to distinguish between claps, but I can say that the clapper I have has never activated from anything on accident that wasn’t a deliberate clap, so I would assume an intelligent AI should also be able to be programmed to tell the difference between any of those things with sufficient samples of correct and incorrect detections with sensitivity settins, at least to be as reliable as clappers are, and mine hasn’t falsely triggered in memory. They could launch a pilot or beta test for it and just have users submit events with the various inputs and the AI would do all the work just like it does for meowing and barking and crying detection, etc.
It would be cool to have some more AI triggers though. Currently we basically triggers for DETECTS:
- Sound
- Motion
- Smoke Alarm
- CO Alarm
- Pet
- Vehicle
- Package
- Person
I’d love to see more options we can set as triggers.