blog Article

EU AI Legislation: What To Keep in Mind

Author: Joren Verspeurt ,

In our mission to achieve impactful AI, we must keep many things in mind, including the rule of law. This blog post highlights the EU AI Act, a law proposed by the European Commission that would significantly impact how businesses in Europe and worldwide use AI. While this regulation may not take effect immediately, companies should keep it in mind as they build or update AI solutions in 2022.

1-Feb-15-2022-12-55-50-15-PMThe New Regulatory Framework of the EU AI Act

The AI Act creates a regulatory framework for AI, categorizing AI systems based on risk. Each risk category has different requirements companies must follow.

Business blog visual_Radix_Based on Risk

1. Unacceptable Risks

The legislation prohibits the use of AI systems that compromise the safety, rights, or livelihoods of citizens. For example, social scoring (like the system currently used in China) and some forms of real-time facial recognition are deemed unacceptable under the AI Act. If your AI system is categorized as unacceptable, it becomes banned in the EU once the AI Act becomes official law.

2. High-Risk Systems

Risk classification is based on the intended purpose of the AI system and the potential harm it could do to the public. Here are some examples of high-risk AI systems:

  • Consumer credit evaluation – Automated systems for obtaining loans and lines of credit which could bar people from getting fair credit services.
  • Employee recruitment - Automated job application and sorting software that could discriminate applicants.
  • Transportation – Autonomous vehicles that could compromise the safety of citizens.
  • Healthcare – Surgery with AI-assisted software or robots that could lead to harmful outcomes.

There is a lot of debate about what constitutes a high-risk AI system, so this category will likely evolve. Under the proposed AI Act however, companies must complete several steps before deploying a high-risk AI system to the EU market:

  • Complete a conformity assessment to confirm the AI system complies with AI Act requirements.
  • Register the AI system in a public database of high-risk AI systems (to be set up by the European Commission).
  • Sign a declaration of conformity and make sure the AI system carries the CE mark.
  • Monitor the AI system to detect and address any new or additional risks that arise after entering the EU market.

Companies that deploy high-risk AI systems but fail to meet AI Act requirements could face hefty fines.

3. Limited Risks

This category covers AI systems where the harmful effects are limited, but of which users should know about their artificial nature, such as chatbots, virtual assistants, and emotion recognition systems. AI systems in this category have specific transparency requirements they must meet. For example, an AI chatbot should have a mechanism to alert users that they are interacting with a machine instead of a human.

4. Minimal Risks

This category includes AI systems that pose little or no risk to the public, such as email spam filters, customer segmentation systems, and inventory management systems. AI systems in this category have no requirements under the AI Act.

What Could Future EU AI Legislation Regulate?

Besides the new AI Act, the EU is working on a whole slew of related regulations, such as
the Data Governance Act, and the ePrivacy regulation (which would replace the ePrivacy directive as it currently stands).

The vision paper released in 2020 also suggests that other AI-specific legislation may be on its way. We're excited to see how the regulatory landscape evolves over time and will be ahead of the curve when it comes to new regulatory frameworks and compliance.

These future developments could include legal domains where legislation has never applied to AI directly, like liability.

Companies Should Prepare for Future EU AI Legislation Now

In our opinion, companies must understand current legislation impacting AI and prepare for future AI regulatory updates now. They should also proactively build ethical AI solutions since AI legislation will eventually mandate that.

2-Feb-15-2022-12-55-50-23-PMWe will help you build innovative and ethical AI solutions that comply with current and future laws. Driving fairness in AI for example requires making difficult choices, and we can guide you to make the right ones. We’re developing a new version of AutoML that provides fast and accurate predictions as well as explanations for each prediction. And we can work with you on setting up a framework for effective human oversight so that you can ensure accountability when using AI.

Ready to create impact with AI in an ethical and fair way?

Book a call with our AI specialist below to get started.

 
Joren Verspeurt
About The Author

Joren Verspeurt

Joren is a Machine Learning Engineer and Security Officer at Radix. He mixes his experience as a Data Scientist and Software Engineer for companies working in domains as varied as video streaming, life sciences, and cloud telephony. What drives him is finding ways to improve people's interactions with technology by using Al and Machine Learning techniques.

About The Author