Facebook investors will urge firm to drop end-to-end encryption plans
Shareholders say plans could hinder efforts to detect and stop child exploitation cases
Facebook’s shareholders will urge the company to ditch plans to implement end-to-end encryption across its messaging systems, saying that it could hinder efforts to detect and stop child exploitation cases.
The investors are set to discuss and vote on Facebook’s management plans at the company’s annual shareholders meeting today, which is to be held virtually due to COVID-19 safety measures.
The meeting will allow them to seek a board review of the company’s plans to use end-to-end encryption and its potential impact on child abuse victims.
Facebook CEO Mark Zuckerberg previously stated that “people's private communications should be secure” and that the company’s implementation of end-to-end encryption will prevent “anyone -- including us -- from seeing what people share on our services”.
However, Michael Passoff, CEO of shareholder advocacy service Proxy Impact, said that “shareholders are legitimately concerned that Facebook's role as a facilitator of child abuse and exploitation will spiral even further out of control if it adopts end-to-end encryption without first stopping predators who prey on children”.
“Not only is it the right thing to do, but it is in the best interests of the company which may otherwise face legislative, regulatory, legal, advertising and consumer backlashes,” he added.
Zuckerberg’s plans to implement end-to-end encryption, announced last year, have previously raised concerns about children’s safety.
“When we were deciding whether to go to end-to-end encryption across the different apps, this was one of the things that just weighed the most heavily on me,” Zuckerberg said at Facebook’s internal Q&A session in October 2019.
However, he expressed that he was "optimistic" that criminals could still be identified by other means, such as their patterns of activity.
According to The New York Times, Facebook Messenger alone was responsible for “nearly 12 million of the 18.4 million worldwide reports of” child sexual abuse material (CSAM) in 2018.
Accelerating AI modernisation with data infrastructure
Generate business value from your AI initiativesFree Download
Recommendations for managing AI risks
Integrate your external AI tool findings into your broader security programsFree Download
Modernise your legacy databases in the cloud
An introduction to cloud databasesFree Download
Powering through to innovation
IT agility drive digital transformationFree Download