Facebook claims AI managed to reduce hate speech by 50%
The social media platform has hit back at claims the tech it uses to fight hate speech is inadequate
On Sunday, the Wall Street Journal (WSJ) published a report based on internal documents and employee testimonials which show that the social media platform manages to remove only “a low-single-digit percentage” of posts that violate its rules of conduct.
The AI used to identify harmful content has problems detecting first-person shooting videos and racist rants, according to the report, as well as telling the difference between cockfighting and car crashes. However, it is more affordable to use than human reviewers, which in 2019 were costing the company “$2 million a week, or $104 million a year”, according to the WSJ.
Facebook’s VP of Integrity, Guy Rosen, issued a response to the article hours after it was published, stating that the hate speech has been reduced by almost 50% during the last three quarters.
According to the company, “prevalence is the most important metric to use because it shows how much hate speech is actually seen on Facebook”.
Build trustworthy AI with MLOps
AI performance, operations, and ethicsFree download
“Recent reporting suggests that our approach to addressing hate speech is much narrower than it actually is, ignoring the fact that hate speech prevalence has dropped to 0.05%, or 5 views per every 10,000 on Facebook,” said Rosen.
When it’s uncertain whether a post violates Facebook’s terms, its visibility is reduced by limiting its distribution and not being recommended to users. This is done in order to protect those who post “content that looks like hate speech but isn’t”, such as “describing experiences with hate speech or condemning it”. The company also stated that 97% of removed content is identified by its algorithm, up from 23.6% in 2016.
Facebook didn’t address the WSJ’s claims that the decision to use AI to monitor hate speech was motivated by costs.
These latest allegations come amid a difficult month for the social media platform, which was recently accused by former product manager turned whistleblower Frances Haugen of repeatedly prioritising profits over user safety. On 4 October, Facebook, as well as its subsidiaries WhatsApp and Instagram, also suffered a six-hour outage.
Four strategies for building a hybrid workplace that works
All indications are that the future of work is hybrid, if it's not here alreadyFree webinar
The digital marketer’s guide to contextual insights and trends
How to use contextual intelligence to uncover new insights and inform strategiesFree Download
Ransomware and Microsoft 365 for business
What you need to know about reducing ransomware riskFree Download
Building a modern strategy for analytics and machine learning success
Turning into business valueFree Download