Apple pushes back against child abuse scanning concerns in new FAQ

Illustration by Alex Castro / The Verge

In a new FAQ, Apple has attempted to assuage concerns that its new anti-child abuse measures could be turned into surveillance tools by authoritarian governments. “Let us be clear, this technology is limited to detecting CSAM [child sexual abuse material] stored in iCloud and we will not accede to any government’s request to expand it,” the company writes.

Apple’s new tools, announced last Thursday, include two features designed to protect children. One, called “communication safety,” uses on-device machine learning to identify and blur sexually explicit images received by children in the Messages app, and can notify a parent if a child age 12 and younger decides to view or send such an image. The second is designed to detect known CSAM...

Continue reading…



from The Verge - All Posts https://ift.tt/37tqpzn

SUBSCRIBE TO OUR NEWSLETTER

Related Posts :

“Work hard in silence, let your success be your noise"

0 Response to "Apple pushes back against child abuse scanning concerns in new FAQ"

Post a Comment

ad

Search Your Job