Artificial intelligence – EU Commission publishes proposed regulations

This website will offer limited functionality in this browser. We only support the recent versions of major browsers like Chrome, Firefox, Safari, and Edge.
The EU Commission has published proposals for the regulation of artificial intelligence (AI). There's a lot to digest but we set out here some of the key points contained in the proposed regulations and their accompanying documents, including why this is of relevance in the UK.
Objectives
The proposed objectives of the regulations are to:
However, the regulations should be seen as part of a wider move by the EU along with its Co-ordinated Plan with Member States to strengthen AI uptake, investment and innovation.
The regulations have been drafted with the above in mind.
How does the proposed regulation define AI?
Defining AI is tricky and there are various across industry and regulators. When it comes to the EU's regulations, the EU recognises that the definition needs to be specific enough to provide certainty but flexible enough to accommodate technological developments. As a result, the definition of AI is:
"software that is developed with one or more [specified] techniques and approaches [e.g. machine learning] and can, for a given set of human-defined objectives, generate outputs such as content, predictions, recommendations, or decisions influencing the environments they interact with"
This is different to the definition of AI adopted by the EU Parliament in October 2020 for its framework for ethical AI (which we wrote about here). Whether anything turns on the difference is not immediately apparent, but does denote the difficulties of pinning down a definition.
To whom and what would the proposed regulations apply?
The regulations would apply to:
The regulations would not apply to:
What do the regulations propose?
The regulations affect different types and uses of AI based on a risk-based approach. There's a lot of detail in the proposed regulations, but in summary (in the words of the EU press release):
Unacceptable risk: AI systems considered a clear threat to the safety, livelihoods and rights of people will be banned. This includes AI systems or applications that manipulate human behaviour to circumvent users' free will (e.g. toys using voice assistance encouraging dangerous behaviour of minors) and systems that allow ‘social scoring' by governments.
High-risk: AI systems identified as high-risk include AI technology used in:
High-risk AI systems will be subject to strict obligations before they can be put on the market:
Limited risk, i.e. AI systems with specific transparency obligations: When using AI systems such as chatbots, users should be aware that they are interacting with a machine so they can take an informed decision to continue or step back.
Minimal risk: The legal proposal allows the free use of applications such as AI-enabled video games or spam filters. The vast majority of AI systems fall into this category. The draft Regulation does not intervene here, as these AI systems represent only minimal or no risk for citizens' rights or safety.
The regulations also propose potentially significant fines for non-compliance.
Breach | Potential penalties |
Prohibition on specific AI systems or data governance requirements for high-risk systems | Up to €30m or, for companies, up to 6% of worldwide annual turnover for the preceding FY, whichever is higher |
Other regulations | Up to €20m or, for companies, up to 4% of worldwide annual turnover for the preceding FY, whichever is higher |
Supply of incorrect, incomplete or misleading information to notified bodies or national competent authorities in reply to a request | Up to €10m or, for companies, up to 2% of worldwide annual turnover for the preceding FY, whichever is higher |
What relevance does this have in the UK after Brexit?
As a result of Brexit, the regulations would not apply directly in the UK. However, they will have an impact. Two reasons include:
Comment
Much of what the regulations propose may already be happening in practice in some form. For example, requirements for high-risk AI specify include detail about the required risk management systems, data governance, technical documentation, and record keeping. However, some of the regulations may require AI developers and users to rethink what they are doing. For example, "high-risk AI systems shall be designed and developed in such a way to ensure that their operation is sufficiently transparent to enable users to interpret the system’s output and use it appropriately."
In any event, should the regulations ultimately proceed and have effect, users and providers or AI systems will want to ensure compliance given the significant penalties for non-compliance. Given the detail of the proposed regulations, and often complexity of AI systems, thought on compliance is likely to be required sooner rather than later.
Europe fit for the Digital Age: Commission proposes new rules and actions for excellence and trust in Artificial Intelligence
https://ec.europa.eu/commission/presscorner/detail/en/ip_21_1682