Voice assistants are money losing products. If they can do something like processing the wakewords on the device before chosing to send to a server they will. These companies are far too stingy to continuously stream audio to their servers
Someone can correct me if I’m wrong but home assistant is currently struggling with this and is processing everything on your local box because it can’t do wakewords on the device.
I think they’re choosing to do it that way. Raspberry pi’s easily have that capability to do the wake word recognition on device (i think they are also working on that). Esp’s on the other hand, can only stream audio to the server and not much more. Since esp’s are far cheaper than installing a raspberry in each room, they are focusing to do wake word detection on the server not on device.
Exactly. If it is practical and money can be made doing it, then continuous, ambient sound parsing will be the norm. Currently it seems like it’s not a valuable business. When it is valuable to them, they will add a checkbox somewhere in your account to disable it, and most people will not be bothered enough to look for it.
My experiences are much MUCH different. The amount of compute waste is through the roof, and we shrug at +$50k/m provisioning. You don’t even need approvals for that, and you can leave it idle and you MIGHT get a ping from gloudgov after a few months.
Voice assistants are money losing products. If they can do something like processing the wakewords on the device before chosing to send to a server they will. These companies are far too stingy to continuously stream audio to their servers
I think this should be fairly easy to test yourself. Just disconnect from the WAN, say the wake word, and see if the device responds.
He means internet, people. He means disconnect from the internet
Someone can correct me if I’m wrong but home assistant is currently struggling with this and is processing everything on your local box because it can’t do wakewords on the device.
I think they’re choosing to do it that way. Raspberry pi’s easily have that capability to do the wake word recognition on device (i think they are also working on that). Esp’s on the other hand, can only stream audio to the server and not much more. Since esp’s are far cheaper than installing a raspberry in each room, they are focusing to do wake word detection on the server not on device.
Yeah what possible use could this company, whose business model relies on surveillance, have for surveiling you
Exactly. If it is practical and money can be made doing it, then continuous, ambient sound parsing will be the norm. Currently it seems like it’s not a valuable business. When it is valuable to them, they will add a checkbox somewhere in your account to disable it, and most people will not be bothered enough to look for it.
Are they though?
My experiences are much MUCH different. The amount of compute waste is through the roof, and we shrug at +$50k/m provisioning. You don’t even need approvals for that, and you can leave it idle and you MIGHT get a ping from gloudgov after a few months.