Top

Apple will scan iPhones, iPads for images of child sex abuse

Apple also said it can\'t actually see user photos or the results of such scans unless there\'s a hit

Washington: Tech giant Apple is adding a series of new child-safety features to its next big operating system updates for iPhones and iPads.

As per Variety, as a part of iOS 15 and iPadOS 15 updates later this year, the company will implement a feature to detect photos stored in iCloud Photos that depict sexually explicit activities involving children.

"This will enable Apple to report these instances to the National Center for Missing and Exploited Children (NCMEC)," the company said in a notice on its website. NCMEC acts as a reporting centre for child sexual abuse material (CSAM) and works in collaboration with law enforcement agencies across the US.

According to Apple, its method of detecting known CSAM is "designed with user privacy in mind." The company said it is not directly accessing customers' photos but instead is using a device-local, hash-based matching system to detect child abuse images.

Apple also said it can't actually see user photos or the results of such scans unless there's a hit.

If there's a match between a user's photos and the CSAM database, Apple then manually reviews each report to confirm the presence of sexually explicit images of children, then will disable the user's account and send a report to NCMEC.

If a user feels their account has been mistakenly flagged, according to Apple, "they can file an appeal to have their account reinstated."

The system provides a high level of accuracy that ensures less than a one-in-1-trillion chance per year of incorrectly flagging a given account, according to Apple.

In addition, with Apple's iOS 15 update, the iPhone's Messages app will add new tools to warn children and their parents if they are receiving or sending sexually explicit photos.

"When receiving this type of content, the photo will be blurred and the child will be warned, presented with helpful resources, and reassured it is okay if they do not want to view this photo," Apple said.

The tech giant added, "As an additional precaution, the child can also be told that, to make sure they are safe, their parents will get a message if they do view it. Similar protections are available if a child attempts to send sexually explicit photos."

Apple's iOS 15 also will provide updates to Siri and Search to "provide parents and children expanded information and help if they encounter unsafe situations."

Siri and Search will intervene when users try to search for child sexual abuse material, displaying prompts that will "explain to users that interest in this topic is harmful and problematic, and provide resources from partners to get help with this issue."

The iOS 15 update is slated to be available in the fall of 2021, available for iPhone 6s and later models.

Next Story