Fri, 15 November , 2024 Home About Us Advertisement Contact Us
Breaking News

‘Child Sexual Abuse Is Stored On iCloud. Apple Allows It’: Anti-Apple Posters Emerge In Front Of Apple Park

Apple Inc.’s (NASDAQ:AAPL)decision to abandon plans to scan iPhones for child sexual abuse material (CSAM) has invited the wrath of protestors, who have now set up banners in front of Apple Park to coincide with the iPhone 15 launch.

What Happened: Protestors have set up anti-Apple banners in front of Apple Park, condemning Apple’s decision to abandon its iPhone CSAM detection plans, implying that the company is somehow enabling it.

“Child sexual abuse is stored on iCloud. Apple allows it,” says the poster put up by Heat Initiative, a collective of child safety experts and advocates. The poster was spotted and shared by an Instagram user.

The Heat Initiative’s website underlines its efforts to bring Apple to make good on its commitment to protect children. Back in 2021, Apple announced plans to scan iPhones and iCloud for CSAM – while child safety experts praised this, it angered privacy advocates.

The Electronic Frontier Foundation, a digital rights group, called it opening a “backdoor” to the private lives of users.

After facing immense backlash, Apple quietly pulled references to CSAM detection on its website and subsequently called it off entirely in 2022.

Why It Matters: CSAM is a huge problem. A case in point is the iCloud account of a 32-year-old man, who was found to be in possession of nearly 2,400 child abuse images in 2019, according to the Heat Initiative.

“We are calling on Apple to detect, report, and remove child sexual abuse images and videos from iCloud,” the collective says.

Here’s What Apple Has To Say

On the other hand, Apple says its initial CSAM detection plans would have violated user privacy. This realization came about after immense backlash from cybersecurity and privacy experts.

The company said this would not only snowball into a privacy issue but also create security risks and new vectors for malicious parties to attack.

Instead, Apple has come up with a different solution with on-device CSAM detection within apps themselves – for example, apps like Messages, FaceTime, AirDrop, and more have on-device nudity detection systems.

Apple has also launched an application programming interface (API) for this so third-party apps can implement it without adversely impacting user privacy – this also aligns with one of the demands of Heat Initiative, but its protests suggest that it is not completely satisfied with the steps Apple has taken so far.

Comments

comments