Apples says its child abuse prevention system will only scan already flagged photos

Tips & Techniques

Apple has faced harsh criticism after announcing that it will scan iCloud and iPhones for photos of child sexual abuse. The company has now responded to it and clarified that it will only scan the photos that have already been flagged internationally.

One of the features allows Apple to scan stored iCloud images in search of known CSAM (Child Sexual Abuse Material) images. If any of them are found, Apple will report it to the National Center for Missing and Exploited Children (NCMEC). According to Reuters, Apple said on Friday that it will “hunt only for pictures that have been flagged by clearinghouses in multiple countries.”

This feature uses an automatic scanning system to go through your photos. After they reach a certain threshold, Apple gets an alert and a human is allowed to review the content on your iCloud. Apple initially refused to say what the threshold is, but has now confirmed that it’s 30 images. The company added that this number would eventually be reduced in the future as the system improves. Also, the “surveillance,” as this feature was dubbed by some, will only be focused on the US for now.

When asked whether criticism influenced the decisions, Apple declined to respond. The company said that “the project was still in development and changes were to be expected.”

[via Engadget]

Products You May Like

Articles You May Like

What is Rembrandt Lighting and How to Use It for Portrait Photos
Green Screens Aren’t Bad, Most Just Don’t Know How to Use Them
Rode introduces its newest entry-level on-camera Rode VideoMicro II microphone
What is a Mood Board and How Do You Make One?
Photographer Blames Balenciaga for Photos of Kids with BDSM Props

Leave a Reply

Your email address will not be published. Required fields are marked *