Publicly, that is. They have no doubt been doing it in secret since they launched it.
Off-device processing has been the default from day one. The only thing changing is the removal for local processing on certain devices, likely because the new backing AI model will no longer be able to run on that hardware.
Today: “…they will be deleted after Alexa processes your requests.”
Some point in the not-so-distant future: “We are reaching out to let you know that your voice recordings will no longer be deleted. As we continue to expand Alexa’s capabilities, we have decided to no longer support this feature.”
“We lied and paid a $3M fine.”
They could also transcribe the recording and only save that. I mean they absolutely will and surely already did do that.
If you do not want to set your voice recordings setting to ‘Don’t save recordings,’ please follow these steps before March 28th:
Am I the only one curious to know what these steps are? The image cuts off the rest of the email.
If anyone else is wondering, I’ve not found a verbatim quote of the steps but I did see an article that mentioned the consequences. It seems like you will be able to turn this off but it will disable Voice ID:
anyone with their Echo device set to “Don’t save recordings” will see their already-purchased devices’ Voice ID feature bricked. Voice ID enables Alexa to do things like share user-specified calendar events, reminders, music, and more. Previously, Amazon has said that “if you choose not to save any voice recordings, Voice ID may not work.” As of March 28, broken Voice ID is a guarantee for people who don’t let Amazon store their voice recordings.
The old “privacy focused” setting made speech processing local. The new “privacy focused setting” means that processing will happen on a remote server, but Amazon won’t store the audio after it’s been processed. Amazon could still fingerprint voices with the new setting, to know if it was you or your parents/parter/kid/roommate/whomever and give a person specific response, but for now at least they appear to not be doing so.
This all seems like it’s missing the point to me. If you own one of these devices you’re giving up privacy for convenience. With the old privacy setting you were still sending your processed speech to a server nearly every time you interacted with one of those devices because they can’t always react/provide a response on their own. Other than trying to avoid voice fingerprinting, it doesn’t seem like the old setting would gain you much privacy. They still know the device associated to the interaction, know where the device is located, which accounts it’s associated with, what the interaction was, etc. They can then fuse this information with tons of other data collected from different devices, like a phone or computer. They don’t need your unprocessed speech to know way too much about you.
How the fuck does anyone even buy one of these
The same people who buy mobile phones; despite those being bugs/spy-devices.
More directly comparable, is the Ring cameras inside the house.
Easy fix: don’t buy this garbage to begin with. It’s terrible for the environment, terrible for your privacy, of dubious value to begin with.
If every man is an onion, one of my deeper layers is crumudgeon. So take that into account when I say fuck all portable speakers. I’m so tired of hearing everyone’s shitty noise. Just fucking everywhere. It takes one person feeling entitled to blast the shittiest music available to ruin everyone in a 500yd radius’s day. If this is you, I hope you stub your toe on every coffee table, hit your head on every door jam, miss every bus.
So… if you own an inexpensive Alexa device, it just doesn’t have the horsepower to process your requests on-device. Your basic $35 device is just a microphone and a wifi streamer (ok, it also handles buttons and fun LED light effects). The Alexa device SDK can run on a $5 ESP-32. That’s how little it needs to work on-site.
Everything you say is getting sent to the cloud where it is NLP processed, parsed, then turned into command intents and matched against the devices and services you’ve installed. It does a match against the phrase ‘slots’ and returns results which are then turned into voice and played back on the speaker.
With the new LLM-based Alexa+ services, it’s all on the cloud. Very little of the processing can happen on-device. If you want to use the service, don’t be surprised the voice commands end up on the cloud. In most cases, it already was.
If you don’t like it, look into Home Assistant. But last I checked, to keep everything local and not too laggy, you’ll need a super beefy (expensive) local home server. Otherwise, it’s shipping your audio bits out to the cloud as well. There’s no free lunch.
everything you say to your echo/alexa has always been sent to amazon.
theres literally been leaks proving it.
It’s always been this way for the cheap speakers. They’ve no processing power on-board and need the cloud just to tell you the time.
Now they can hear me scream “shut the fuck up Alexa!!!” every time she says “…by the way…” when I just want to know what time it is.
Say this: “Alexa, disable by the way”
“Alexa, from now on, call me ‘Big Dick Daddy from Cincinnati’.”
If anyone remembers the Mycroft Mark II Voice Assistant Kickstarter and was disappointed when development challenges and patent trolls caused the company’s untimely demise, know that hope is not lost for a FOSS/OSHW voice assistant insulated from Big Tech…
FAQ: OVOS, Neon, and the Future of the Mycroft Voice Assistant
Disclaimer: I do not represent any of these organizations in any way; I just believe in their mission and wish them all the success in getting there by spreading the word.
Alexa! Install Linux!
I didn’t even know this was a feature. My understanding has always been that Echo devices work as follows.
- Store a constant small buffer of the past few seconds of audio
- Locally listen for the wake word (typically “Alexa”) using onboard hardware. (This is why you cannot use arbitrary wake words.)
- Upon hearing the wake word, send the buffer from step one along with any fresh audio to the cloud to process what was said.
- Act on what was said. (Turn lights on or off, play Spotify, etc.)
Unless they made some that were able to do step 3 locally entirely I don’t see this as a big deal. They still have to do step 4 remotely.
Also, while they may be “always recording” they don’t transmit everything. It’s only so if you say “Alexaturnthelightsoff” really fast it has a better chance of getting the full sentence.
I’m not trying to defend Amazon, and I don’t necessarily think this is great news or anything, but it doesn’t seem like too too big of a deal unless they made a lot of devices that could parse all speech locally and I didn’t know.
It was a non advertised feature only available in the US and in English only
If you traveled back in time and told J. Edgar Hoover that in the future, the American public voluntarily wire-tapped themselves, he would cream his frilly pink panties.
Great, they made it public.