Coronavirus forces social media to rely heavily on AI moderation

The pandemic has put social media’s automated takedown software to the test - with some room for error

Facebook, Twitter, and YouTube warned that more videos and other content could be mistakenly classified as policy violations and removed, due to the companies relying on the judgement of artificial intelligence (AI) during the coronavirus pandemic.

AI has been left to do its content-policing job virtually unattended as tech giants empty their offices and ask staff to work from home to protect them from the virus and curb the spread of the pandemic.

But this unprecedented situation has put social media’s automated takedown software to the test, with some room for error.

“We've invested significantly in automated systems for content review but they are not always as accurate or granular in their analysis of content as human reviewers,” Google announced on their blog. “These systems are configured deliberately to identify content that may violate our policies. So on YouTube there may be an increase in content classified for removal during this time—including some content that does not violate our policies.”

Twitter has also announced that it would also be increasing its use of machine learning and automation. In a blog post, legal, policy and trust & safety lead Vijaya Gadde and VP of sales Matt Derella warned: “While we work to ensure our systems are consistent, they can sometimes lack the context that our teams bring, and this may result in us making mistakes”.

The company also assured users that it would not be permanently suspending any Twitter accounts based solely on the judgement of its automated enforcement systems.

Facebook announced that it would ask its content review contract workers to work from home, yet warned that the staff could not perform some duties remotely “due to safety, privacy and legal reasons”. Nevertheless, it assured that AI would help in tackling content-reviewing workload.

“We believe the investments we’ve made over the past three years have prepared us for this situation,” assured Kang-Xing Jin, Facebook’s head of health. “With fewer people available for human review we’ll continue to prioritize imminent harm and increase our reliance on proactive detection in other areas to remove violating content. We don’t expect this to impact people using our platform in any noticeable way.”

Despite assurances that the situation would not heavily affect the experiences of Facebook users, Jin warned that “there may be some limitations to this approach and we may see some longer response times and make more mistakes as a result”.

Featured Resources

2021 Thales access management index: Global edition

The challenges of trusted access in a cloud-first world

Free download

Transforming higher education for the digital era

The future is yours

Free download

Building a cloud-native, hybrid-multi cloud infrastructure

Get ready for hybrid-multi cloud databases, AI, and machine learning workloads

Free download

The next biggest shopping destination is the cloud

Know why retail businesses must move to the cloud

Free Download

Most Popular

Best Linux distros 2021
operating systems

Best Linux distros 2021

11 Oct 2021
Apple MacBook Pro 15in vs Dell XPS 15: Clash of the titans

Apple MacBook Pro 15in vs Dell XPS 15: Clash of the titans

11 Oct 2021
Supply chain breaches impacted 97% of firms in the past year
supply chain management (SCM)

Supply chain breaches impacted 97% of firms in the past year

12 Oct 2021