'Child Sexual Abuse Is Stored On iCloud. Apple Allows It': Anti-Apple Posters Emerge In Front Of Apple Park

Comments
Loading...
Zinger Key Points
  • Apple’s decision to abandon iPhone CSAM detection has angered child safety experts.
  • One such collective, Heat Initiative, has set up protest banners in front of Apple Park.
  • On the other hand, Apple said it had to drop its CSAM detection plans to protest user privacy and close vectors for malicious attacks.
  • Discover Fast-Growing Stocks Every Month

Apple Inc.'s AAPL decision to abandon plans to scan iPhones for child sexual abuse material (CSAM) has invited the wrath of protestors, who have now set up banners in front of Apple Park to coincide with the iPhone 15 launch.

What Happened: Protestors have set up anti-Apple banners in front of Apple Park, condemning Apple's decision to abandon its iPhone CSAM detection plans, implying that the company is somehow enabling it.

See Also: Want To Upgrade To The iPhone 15 or 15 Pro From iPhone 14? Here Are The Trade-In Prices Apple Is Offering

"Child sexual abuse is stored on iCloud. Apple allows it," says the poster put up by Heat Initiative, a collective of child safety experts and advocates. Danish Khan (@dankh4n) spotted and shared the poster on Instagram.

Image Credits – @dankh4n on Instagram.

The Heat Initiative's website underlines its efforts to bring Apple to make good on its commitment to protect children. Back in 2021, Apple announced plans to scan iPhones and iCloud for CSAM – while child safety experts praised this, it angered privacy advocates.

The Electronic Frontier Foundation, a digital rights group, called it opening a “backdoor” to the private lives of users.

After facing immense backlash, Apple quietly pulled references to CSAM detection on its website and subsequently called it off entirely in 2022.

Why It Matters: CSAM is a huge problem. A case in point is the iCloud account of a 32-year-old man, who was found to be in possession of nearly 2,400 child abuse images in 2019, according to the Heat Initiative.

"We are calling on Apple to detect, report, and remove child sexual abuse images and videos from iCloud," the collective says.

Here's What Apple Has To Say

On the other hand, Apple says its initial CSAM detection plans would have violated user privacy. This realization came about after immense backlash from cybersecurity and privacy experts.

The company said this would not only snowball into a privacy issue but also create security risks and new vectors for malicious parties to attack.

Instead, Apple has come up with a different solution with on-device CSAM detection within apps themselves – for example, apps like Messages, FaceTime, AirDrop, and more have on-device nudity detection systems.

Apple has also launched an application programming interface (API) for this so third-party apps can implement it without adversely impacting user privacy – this also aligns with one of the demands of Heat Initiative, but its protests suggest that it is not completely satisfied with the steps Apple has taken so far.

Check out more of Benzinga’s Consumer Tech coverage by following this link.

Read Next: How To Preorder Apple's iPhone 15: Price, Where To Order And Everything You Need To Know

Overview Rating:
Good
62.5%
Technicals Analysis
66
0100
Financials Analysis
60
0100
Overview
Market News and Data brought to you by Benzinga APIs

Posted In:
Benzinga simplifies the market for smarter investing

Trade confidently with insights and alerts from analyst ratings, free reports and breaking news that affects the stocks you care about.

Join Now: Free!