Facebook updates hate speech detection ahead of Myanmar election
The social media firm will limit message forwarding, and use its improved AI to remove hate speech
Facebook has outlined plans to improve its technology that detects hate speech on its platform ahead of crucial November elections in Myanmar.
The company has expanded its misinformation policy so that it will now remove fake news that could lead to voter suppression or damage the integrity of the election. Facebook will also use its AI to remove content such as hate speech that could lead to offline harm as well as attempts to suppress the vote, according to a blog post.
The firm claims to have invested significantly in proactive detection technology to help catch such content more quickly than before, using AI to identify hate speech in 45 languages, including Burmese.
"In the second quarter of 2020, we took action against 280,000 pieces of content in Myanmar for violations of our Community Standards prohibiting hate speech, of which we detected 97.8% proactively before it was reported to us," Facebook’s director of public policy for Southeast Asia emerging markets, Rafael Frankel said.
"This is up significantly from Q1 2020, when we took action against 51,000 pieces of content for hate speech violations, detecting 83% proactively," he added.
Political ads in Myanmar will also be labelled with a tag showing which individual or organisation has paid for the content, alongside verified badges for the pages of official political organisations. This is in addition to limits on message forwarding to five people, as well as the launch of a third-party fact-checking programme.
The social media platform has previously been criticised for allowing hate speech and prejudice to spread on its platform in the region, and for inadequate processes that "enable genocide", according to New Zealand’s privacy commissioner.
The reliance on AI to detect harmful content, such as hate speech, was also slammed by the global nonprofit advocacy organisation Avaaz. The group published a study in October 2019 suggesting an overreliance on this technology, which is under-developed, was leaving minorities vulnerable.
Attention will also be thrust on Facebook’s role in the upcoming US presidential election, and the nature that false and harmful content spreads on its platforms. The firm’s reliance on technology was also recently criticised for wrongly blocking a pro-Joe Biden ad, telling a Democratical group that the clip violated its policy against "sensational" content.
The decision was later reversed, according to Reuters, with Facebook citing an "enforcement error" for the rejection of only certain versions of the advert, while others were allowed to run.
Five lessons learned from the pivot to a distributed workforce
Delivering continuity and scale with a remote work strategyDownload now
Connected experiences in a digital transformation
Enable businesses to meet the demands of the futureDownload now
Simplify to secure
Reduce complexity by integrating your security ecosystemDownload now
Enhance the safety and security of your people, assets and operations
Enable a true vision of security with an engineered solution based on hyperconverged and storage platformsDownload now