Skip to content

Apple accused of using privacy to excuse ignoring child abuse material on iCloud

A proposed class action suit claims that Apple is hiding behind claims of privacy in order to avoid stopping the storage of child sexual abuse material on iCloud, and alleged grooming over iMessage.

Smartphone displaying a sensitive content warning, laid on a white keyboard.
Apple cancelled its major CSAM proposals but introduced features such as automatic blocking of nudity sent to children

Following a UK organization’s claims that Apple is vastly underreporting incidents of child sexual abuse material (CSAM), a new proposed class action suit says the company is “privacy-washing” its responsibilities.

A new filing with the US District Court for the Northern District of California, has been brought on behalf of an unnamed 9-year-old plaintiff. Listed only as Jane Doe in the complaint, the filing says that she was coerced into making and uploading CSAM on her iPad.

Continue Reading on AppleInsider | Discuss on our Forums