Chat with us, powered by LiveChat

Apple Sparks Privacy Debate over Scanning for Illegal Images

Halt Production
Print Friendly, PDF & Email

EDITOR NOTE: We all want to see child abuse and exploitation stop, but the way Apple is going about it is incredibly scary and a very slippery slope. The tech giant recently announced it will start scanning iPhone users’ devices without permission to look for illegal images of children. It’s a noble cause, no doubt, but this level of corporate surveillance should strike fear in the hearts of every American. If this isn’t a huge reason to purchase private, non-CUSIP gold and silver, I’m not sure what is.

Apple announced Thursday is it planning to scan all iPhones in the United States for child abuse imagery, raising alarm among security experts who said the plan could allow the firm to surveil tens of millions of personal devices for unrelated reasons.

In a blog post, the company confirmed reports saying that new scanning technology is part of a suite of child protection programs that would “evolve and expand.” It will be rolled out as part of iOS 15, which is scheduled for release sometime in August.

Apple, which has often touted itself as a company that promises to safeguard users’ right to privacy, appeared to try and preempt privacy concerns by saying that the software will enhance those protections by avoiding the need to carry out widespread image scanning on its cloud servers.

“This innovative new technology allows Apple to provide valuable and actionable information to [the National Center for Missing and Exploited Children] and law enforcement regarding the proliferation of known CSAM,” said the company, referring to an acronym for child sexual abuse material. “And it does so while providing significant privacy benefits over existing techniques since Apple only learns about users’ photos if they have a collection of known CSAM in their iCloud Photos account. Even in these cases, Apple only learns about images that match known CSAM.”

The Cupertino-based tech giant said the system will utilize breakthrough cryptography technology and artificial intelligence to find abuse material when it is stored in iCloud Photos, said the firm in its blog post. The images will be matched to a known database of illegal images, the firm said, adding that if a certain number of those images are uploaded to iCloud Photos, the company will review them.

Those images—if they’re deemed illegal—will be reported to the National Center for Missing and Exploited Children. The software won’t be applied to videos, Apple added.

“Apple’s expanded protection for children is a game-changer,” John Clark, the president and CEO of the National Center for Missing and Exploited Children, said in a statement on Thursday about the initiative. “The reality is that privacy and child protection can coexist.“

But some security experts and researchers, who stressed they support efforts to combat child abuse, said the program could present significant privacy concerns.

Ross Anderson, professor of security engineering at the University of Cambridge, described Apple’s proposed system as “an absolutely appalling idea,” according to the Financial Times. “It is going to lead to distributed bulk surveillance of … our phones and laptops,” he remarked.

When news of the proposal broke on Wednesday evening, John Hopkins University professor and cryptographer Matthew Green echoed those concerns.

“This sort of tool can be a boon for finding child pornography in people’s phones,” Green wrote on Twitter. “But imagine what it could do in the hands of an authoritarian government?”

Green said that “if you believe Apple won’t allow these tools to be misused [crossed fingers emoji] there’s still a lot to be concerned about,” noting that such “systems rely on a database of ‘problematic media hashes’ that you, as a consumer, can’t review.”

The expert told The Associated Press that he’s concerned Apple could be pressured by other, more authoritarian governments to scan for other types of information.

Microsoft created PhotoDNA to assist companies in identifying child sexual abuse images on the internet, while Facebook and Google have implemented systems to flag and review possibly illegal content.

Original post from  The Epoch Times 

Bank Failure Scenario Cover Small Not Tilted



  • This field is for validation purposes and should be left unchanged.

All articles are provided as a third party analysis and do not necessarily reflect the explicit views of GSI Exchange and should not be construed as financial advice.

Precious Metals and Currency Data Powered by nFusion Solutions