Apple has announced its latest plan to protect young children from sexual predators.
On Thursday, the tech giant confirmed it would begin using new software that will detect and report child sexual abuse material (CSAM) on U.S. iPhones. Apple will utilize a tool known as “NeuralHash,” which can help determine whether a user is trying to store known CSAM on iCloud.
“This matching process is powered by a cryptographic technology called private set intersection, which determines if there is a match without revealing the result,” Apple wrote in the announcement. “The device creates a cryptographic safety voucher that encodes the match result along with additional encrypted data about the image. This voucher is uploaded to iCloud Photos along with the image.”
Once the automated system finds a match, a human will review the image in a question and assess whether it is illegal. If the reviewer concludes the content qualifies as child pornography, the user’s account will be deactivated and the material will be reported to the National Center for Missing and Exploited Children (NCMEC).
While some have applauded Apple’s efforts to beef up their child protection policies, some technology experts have raised concerns the tool will lead to abuse of privacy.
“Regardless of what Apple’s long term plans are, they’ve sent a very clear signal. In their (very influential) opinion, it is safe to build systems that scan users’ phones for prohibited content,” Matthew Green, a security researcher at Johns Hopkins University, said to the Associated Press. “Whether they turn out to be right or wrong on that point hardly matters. This will break the dam — governments will demand it from everyone.”
NeuralHash will be introduced as part of the iOS 15 software update, which is expected to roll out within the next month or two.
Sign up for the Complex Newsletter for breaking news, events, and unique stories.
Follow Complex on: Facebook, Twitter, Instagram, YouTube, Snapchat, TikTok