This past week, some information came out that people contracted to work on Siri's quality control regularly hear up to 30 seconds of our personal conversations that Siri was not supposed to be listening to. This information, according to an unnamed source that spoke to The Guardian, some of that private audio includes conversations with doctors, drug deals, and even people engaged in sexual activity.
There's a lot to unpack here, and iMore's Apple analyst Rene Ritchie has already done some research and spoke to experts in the field of virtual assistant learning and has a deep-dive analysis of the situation right now.
Apple did respond to the issue with an explanation about its process and regulation regarding Siri training, which is also included in its whitepaper and user agreement.
While it's never been a secret that Apple has a department for quality assurance similar to Amazon and Google for Siri, which includes humans hearing accidental recordings of conversations, this reminder (not revelation) strikes me harder. I expect Amazon and Google to be less concerned with my privacy. I know that my browsing and online purchasing activities are being shared with marketers. Apple, on the other hand, stands on a very large soapbox of privacy. So, even though humans listening to recordings of our Siri queries is necessary (at least for now) to building a better experience, I wish Apple would put this information at the forefront of its devices' interfaces.
That friendly little handshake logo you see when you first install a new operating system that links to all the information about Apple's privacy policies? Not clear enough for me. The fact that you can disable Siri altogether and even opt-out of sharing your iCloud activities with Apple's analytics for quality assurance purposes? Not explanatory enough.
It's not that Apple, Google, Amazon, and other companies that are working on virtual assistant technology are being nefarious. They're providing us with details on what happens when we purposefully or accidentally trigger our assistants. It's that we haven't asked these companies to be clearer about explaining it to us. We accept the Terms and Conditions and start asking questions about the weather or our schedule.
It's only now that we, as the consumers of these wonderful virtual assistants, are starting to ask more important questions about who's listening, why they're listening, and how we can opt-out of being listened to.
This is a whole new world, technologically speaking, and we're only starting to grasp the reality of how much privacy we give away in exchange for convenience. Even when we trust a company to protect our privacy, as I do with Apple, we should consider what's being done behind the scenes and ask those holding the keys to help us understand it better.
It is, after all, our privacy, not theirs.
Lory is a renaissance woman, writing news, reviews, and how-to guides for iMore. She also fancies herself a bit of a rock star in her town and spends too much time reading comic books. If she's not typing away at her keyboard, you can probably find her at Disneyland or watching Star Wars (or both).
“It's not that Apple, Google, Amazon, and other companies that are working on virtual assistant technology are being nefarious. https://www.nytimes.com/2019/07/23/health/data-privacy-protection.html
Sorry for the typo. I meant to say this. “It's not that Apple, Google, Amazon, and other companies that are working on virtual assistant technology are being nefarious.” Respectfully, you just don’t know that. With Google nearly deploying a censored search for China and Amazon seeking military contracts we don’t know what intent of these companies with this data. As for Apple for them not being transparent that a actually person can listen to your Siri inquiries is at minimum problematic. While Apple has done a lot in the privacy space, they also done a lot to make appear the data doesn’t leave device, that everything happens on the device. Also anonymized data isn’t truly anonymous. Researchers have broke it: https://www.nytimes.com/2019/07/23/health/data-privacy-protection.htm “We accept the Terms and Conditions and start asking questions about the weather or our schedule.” Here’s another issue, terms and conditions can be pages long, I’ve seen some as long as 60 page, written in legalese often with links to other as long privacy documents. It’s a data dump that in my opinion used to obfuscate how our data is being used. Simple we need GDPR for the US. A set of laws to protect our privacy as oppose to hoping for the good will of a company.
Get the best of iMore in in your inbox, every day!
Thank you for signing up to iMore. You will receive a verification email shortly.
There was a problem. Please refresh the page and try again.