Comment: Apple’s child protection measures get mixed reactions from experts
Ben Lovejoy
- Aug. 6th 2021 6:12 am PT
@benlovejoy
0
The announcement yesterday of
Apple’s child protection measures
confirmed
an earlier report
that the company would begin scanning for child abuse photos on iPhones. The news has seen mixed reactions from experts in both cybersecurity and child safety.
Four concerns had already been raised
before the details were known, and Apple’s announcement addressed two of them …
CSAM scanning concerns
The original concerns included the fact that digital signatures for child sexual abuse materials (CSAM) are deliberately fuzzy, to allow for things like crops and other image adjustments. That creates a risk of false positives, either by chance (concern one) or malicious action (concern two).
Apple addressed these by announcing that action would not be triggered by a single matching image. Those who collect such material tend to have multiple images, so Apple said a certain threshold would be required before a report was generated. The company didn’t reveal what the threshold is, but did say that it reduced the chances of a false positive to less than one in a trillion. Personally, that completely satisfies me.
However, two further risks remain.
The Electronic Frontier Foundation (EFF)
highlighted the misuse risk
, pointing out that there is no way for either Apple or users to audit the digital fingerprints. A government can
tell
Apple that it only contains CSAM hashes, but there is no way for the company to verify that.
Right now, the process is that Apple will manually review flagged images, and only if the review confirms abusive material will the company pass the details to law enforcement. But again, there is no guarantee that the company will be allowed to continue following this process.
Cryptography academic Matthew Green
reiterated this point
after his pre-announcement tweets.
The EFF says
this is more than a theoretical risk:
In Hong Kong, for example,
criticism of the Chinese government
is classified on the same level as terrorism, and is punishable by life imprisonment.
iMessage scanning concerns
Concerns have also been raised about the AI-based scanning iPhones will conduct on photos in iMessage. This scanning doesn’t rely on digital signatures, but instead tries to identify nude photos based on machine-learning.
Again, Apple has protections built in. It’s only for suspected nude photos. It only affects child accounts as part of family groups. The child is warned that an incoming message might be inappropriate, and then chooses whether or not to view it. No external report is generated, only a parent notified if appropriate.
But again, the slippery slope argument is being raised. These are all controls that apply right now, but
the EFF asks
what if a repressive government forces Apple to change the rules?
The organization also argues that false matches are a definite risk here.
Again, that’s not an issue with Apple’s current implementation due to the safeguards included, but creating technology that can scan the contents of private messages has huge potential for future abuse.
The EFF also highlights an issue raised by some child-protection experts: that a parent or legal guardian isn’t always a safe person with whom to share a child’s private messages.
Some of the discussion highlights that tricky tightrope Apple is trying to walk. For example, one protection is that parents are not automatically alerted: The child is warned first, and then given the choice of whether or not to view or send the image. If they choose not to, no alert is generated.
David Thiel
was one of many to point out the obvious flaw there:
Apple’s child protection measures can’t please everyone
Everyone supports Apple’s intentions here, and personally I’m entirely satisfied by the threshold protection against false positives. Apple’s other safeguards are also thoughtful, and ought to be effective. The company is to be applauded for trying to address a serious issue in a careful manner.
At the same time, the slippery slope risks are very real. It is extremely common for a government – even a relatively benign one – to indulge in mission-creep. It first introduces a law that nobody could reasonably oppose, then later widens its scope, sometimes salami-style, one slice at a time. This is especially dangerous in authoritarian regimes.
Conversely, you could argue that by making this system public, Apple just tipped its hand. Now anyone with CSAM on their iPhone knows they should switch off iCloud, and abusers know if they want to send nudes to children, they shouldn’t use iMessage. So you could argue that Apple shouldn’t be doing this at all, or you could argue that it should have done it without telling anyone.
The reality is that there’s no perfect solution here, and every stance Apple could take has both benefits and risks.
Yesterday, before the full details were known, the vast majority of you
opposed the move
. Where do you stand now that the details – and the safeguards – are known? Please again take our poll, and share your thoughts in the comments.
Take Our Poll
Photo:
David Watkis
/
Unsplash
FTC: We use income earning auto affiliate links.
More.
Check out 9to5Mac on YouTube for more Apple news:
You’re reading 9to5Mac — experts who break news about Apple and its surrounding ecosystem, day after day. Be sure to check out
our homepage
for all the latest news, and follow 9to5Mac on
,
, and
to stay in the loop. Don’t know where to start? Check out our
exclusive stories
,
reviews
,
how-tos
, and
subscribe to our YouTube channel
Guides
AAPL Company
Breaking news from Cupertino. We’ll give you the latest from Apple headquarters and decipher fact from fiction from the rumor mill.
Privacy
Privacy is a growing concern in today's world. Follow along with all our coverage related to privacy, security, and more in our guide.
Opinion
About the Author
Ben Lovejoy
@benlovejoy
Ben Lovejoy is a British technology writer and EU Editor for 9to5Mac. He’s known for his op-eds and diary pieces, exploring his experience of Apple products over time, for a more rounded review. He also writes fiction, with two technothriller novels, a couple of SF shorts and a rom-com!
Ben Lovejoy's favorite gear
NordVPN
Apple Watch Series 3