EU to propose GDPR-like fines for AI abuses

AI will be prohibited from use in mass surveillance or for ranking social behaviour

The European Union (EU) is set to propose a set of enforceable rules that will restrict the use of artificial intelligence (AI) systems against the threat of hefty GDPR-like fines for flagrant violations.

Under the proposals drafted by the European Commission (EC), organisations operating in the EU will not be allowed to use AI for mass surveillance or for ranking social behaviour, according to Bloomberg. Systems deployed to manipulate human behaviour, exploit information about individuals or groups would also be banned in the EU.

Under the rules, authorisation would be required to use biometric identification systems in the public domain, while high-risk AI applications would need to undergo a thorough inspection before they’re deployed. The high-risk category would include applications that use facial recognition, are involved with physical safety or healthcare, or those used in transport or energy.

Member states, in these cases, would need to appoint assessment bodies to examine whether these systems are trained on unbiased data sets and have sufficient human oversight. They’ll inspect and eventually certify the systems that pass the criteria.

While some companies will be allowed to assess themselves, others will need to be vetted by a third party, which will issue compliance certificates valid for up to five years.

Failure to comply with the terms set out in the proposals will result in a range of punishments including financial penalties up to a maximum of 4% of global revenue, which is the same maximum penalty for violating GDPR.

There are several exemptions in place, however, including the use of AI for safeguarding public security as well as AI systems used exclusively for military purposes.

The EU has long been keen to devise a set of enforceable rules that would govern the use of AI systems by organisations operating in its territories, amid the growing concerns about the consequences of unregulated AI deployments. Google has also been among a sea of voices calling out for some kind of framework governing the use of AI, while the Information Commissioner's Office (ICO) is currently consulting with experts on AI regulation.

In March last year, the EC launched the first phase of this process, setting out in its AI white paper the goals for a regulated industry that taps into its digital single market.

The EU also hopes these rules will serve as a model for other nations and territories to replicate, as has become the case with GDPR since its introduction in 2018 with legislation such as the California Consumer Privacy Act (CCPA).

Featured Resources

B2B under quarantine

Key B2C e-commerce features B2B need to adopt to survive

Download now

The top three IT pains of the new reality and how to solve them

Driving more resiliency with unified operations and service management

Download now

The five essentials from your endpoint security partner

Empower your MSP business to operate efficiently

Download now

How fashion retailers are redesigning their digital future

Fashion retail guide

Download now

Most Popular

The benefits of workload optimisation

The benefits of workload optimisation

16 Jul 2021
Samsung Galaxy S21 5G review: A rose-tinted experience
Mobile Phones

Samsung Galaxy S21 5G review: A rose-tinted experience

14 Jul 2021
RMIT to be first Australian university to implement AWS supercomputing facility
high-performance computing (HPC)

RMIT to be first Australian university to implement AWS supercomputing facility

28 Jul 2021