Facebook hit with $150 billion lawsuit over Myanmar hate speech
The tech giant has been accused of failing to prevent disinformation from thriving on its platform
Legal complaints filed in the UK and US on behalf of the Rohingya refugees claim that the tech giant failed to prevent anti-Rohingya hate speech and disinformation from thriving on its platform.
This included thousands of posts describing the community as animals and foreign invaders, falsely accusing them of crimes, and calling for them to be killed.
Facebook’s algorithm had allegedly amplified such hateful posts on users’ news feeds, with the company failing to hire enough Burmese-speaking content moderators despite record-breaking profits that year.
The spread of anti-Rohingya propaganda ultimately resulted in real-life violence that cost the lives of 24,000 people and displaced up to a million, forcing them into “abject poverty”, according to the class-action complaint filed in California by law firms Edelson and Fields.
The US lawsuit is seeking damages “in excess of $150 billion”. The legal notice to Meta’s London offices has not been made publicly available.
The state of brand protection 2021
A new front opens up in the war for brand safetyFree download
The lawsuit references claims made by a former Facebook employee, who said that the company’s executives “were fully aware that posts ordering hits by the Myanmar government on the minority Muslim Rohingya were spreading wildly on Facebook”, and that “the issue of the Rohingya being targeted on Facebook was well known inside the company for years”.
The claims echo testimonies made by another former-employee-turned-whistleblower, Frances Haugan, who in October told members of the US Congress that Facebook was “literally fanning” ethnic violence in developing countries.
Weeks later, Haugen told UK’s MPs that, due to shortages of moderators, Facebook had been unable to police harmful content in multiple languages around the world, leading to civil unrest in Myanmar in 2017 as well as Ethiopia in 2021.
However, the issue also impacts the UK, she added, due to the fact that Facebook's AI is unable to detect online abuse in British English.
Despite the widely-reported anti-Rohingya violence in Myanmar, the tech giant also failed to prevent the spread of anti-Muslim hate speech on its platform in the Assam region of northeast India, according to the lawsuit.
Meta didn’t respond to IT Pro’s request for comment, yet had previously admitted to being “too slow to prevent misinformation and hate” in Myanmar.
Accelerating AI modernisation with data infrastructure
Generate business value from your AI initiativesFree Download
Recommendations for managing AI risks
Integrate your external AI tool findings into your broader security programsFree Download
Modernise your legacy databases in the cloud
An introduction to cloud databasesFree Download
Powering through to innovation
IT agility drive digital transformationFree Download