Facebook investors will urge firm to drop end-to-end encryption plans

Shareholders say plans could hinder efforts to detect and stop child exploitation cases

Facebook website on a computer screen

Facebook’s shareholders will urge the company to ditch plans to implement end-to-end encryption across its messaging systems, saying that it could hinder efforts to detect and stop child exploitation cases.

The investors are set to discuss and vote on Facebook’s management plans at the company’s annual shareholders meeting today, which is to be held virtually due to COVID-19 safety measures.

The meeting will allow them to seek a board review of the company’s plans to use end-to-end encryption and its potential impact on child abuse victims.

Facebook CEO Mark Zuckerberg previously stated that “people's private communications should be secure” and that the company’s implementation of end-to-end encryption will prevent “anyone -- including us -- from seeing what people share on our services”.

However, Michael Passoff, CEO of shareholder advocacy service Proxy Impact, said that  “shareholders are legitimately concerned that Facebook's role as a facilitator of child abuse and exploitation will spiral even further out of control if it adopts end-to-end encryption without first stopping predators who prey on children”.

“Not only is it the right thing to do, but it is in the best interests of the company which may otherwise face legislative, regulatory, legal, advertising and consumer backlashes,” he added.

Zuckerberg’s plans to implement end-to-end encryption, announced last year, have previously raised concerns about children’s safety.

“When we were deciding whether to go to end-to-end encryption across the different apps, this was one of the things that just weighed the most heavily on me,” Zuckerberg said at Facebook’s internal Q&A session in October 2019.

However, he expressed that he was "optimistic" that criminals could still be identified by other means, such as their patterns of activity.

According to The New York Times, Facebook Messenger alone was responsible for “nearly 12 million of the 18.4 million worldwide reports of” child sexual abuse material (CSAM) in 2018.

Featured Resources

Unlocking collaboration: Making software work better together

How to improve collaboration and agility with the right tech

Download now

Four steps to field service excellence

How to thrive in the experience economy

Download now

Six things a developer should know about Postgres

Why enterprises are choosing PostgreSQL

Download now

The path to CX excellence for B2B services

The four stages to thrive in the experience economy

Download now

Most Popular

Microsoft is submerging servers in boiling liquid to prevent Teams outages
data centres

Microsoft is submerging servers in boiling liquid to prevent Teams outages

7 Apr 2021
Hackers are using fake messages to break into WhatsApp accounts
instant messaging (IM)

Hackers are using fake messages to break into WhatsApp accounts

8 Apr 2021
How to find RAM speed, size and type
Laptops

How to find RAM speed, size and type

8 Apr 2021