Source: iMore
What you need to know
- Jane Horvath, Apple's chief privacy officer spoke at CES yesterday.
- Horvath said Apple uses technology to screen for specific photos.
- Images found are "reported."
This story has been updated to reflect a change made to the original story from The Telegraph. From the source article: "This story originally said Apple screens photos when they are uploaded to iCloud, Apple's cloud storage service. Ms Horvath and Apple's disclaimer did not mention iCloud, and the company has not specified how it screens material, saying this information could help criminals."
Jane Horvath, Apple's Senior Director of Global Privacy, spoke during a CES panel yesterday and confirmed that Apple scans photos to ensure that they don't contain anything illegal.
The Telegraph reports:
Apple scans photos to check for child sexual abuse images, an executive has said, as tech companies come under pressure to do more to tackle the crime. Jane Horvath, Apple's chief privacy officer, said at a tech conference that the company uses screening technology to look for the illegal images. The company says it disables accounts if Apple finds evidence of child exploitation material, although it does not specify how it discovers it.
Apple updated its privacy policy last year, but it isn't clear exactly when Apple started scanning images to ensure they are above board. It does have a web page specifically dedicated to child safety, too.
Apple is dedicated to protecting children throughout our ecosystem wherever our products are used, and we continue to support innovation in this space. We have developed robust protections at all levels of our software platform and throughout our supply chain. As part of this commitment, Apple uses image matching technology to help find and report child exploitation. Much like spam filters in email, our systems use electronic signatures to find suspected child exploitation. We validate each match with individual review. Accounts with child exploitation content violate our terms and conditions of service, and any accounts we find with this material will be disabled.
While there will surely be people who have an issue with this, Apple isn't the first company to scan images in this way. Many companies use software called PhotoDNA – a solution that was specifically designed to help prevent child exploitation.
"By working collaboratively with industry and sharing PhotoDNA technology, we're continuing the fight to help protect children." – Courtney Gregoire, Assistant General Counsel, Microsoft Digital Crimes Unit
Now that's surely something we can all agree with.
We may earn a commission for purchases using our links. Learn more.

iOS gaming recap: Is that Pokémon Go meets... basketball?
Ever wanted to play basketball but as Pokémon Go? What about hopping into some more dino action now that the Jurassic World series has seemingly wrapped up? Check out this week in iOS gaming.

Huge Apple TV promotion now live ahead of major upgrade
You can now get a $50 gift card when you buy either of Apple's TV models. The company is likely cleaning house ahead of a rumored big upgrade.

New M2 13-inch MacBook Pro teardown shows it's just the old one in disguise
Apple's new M2 13-inch MacBook Pro is now on sale and people are doing what they tend to do with new things — they're taking the machine apart. The latest to take a screwdriver to Apple's new notebook is iFixit, and the result is further confirmation that this is essentially the M1 model but with some small tweaks and a new CPU.

Deck your dorm with the coolest stuff
Dorm rooms can be plain, so you're going to want to deck it out with as much cool stuff as you possibly can - here are our picks for the essentials!