The legislation is designed to address the varying levels of risk posed by AI systems, with stricter rules for higher-risk applications. This act is the first global regulation of its kind and is expected to set a precedent for AI governance worldwide.
The AI Act intends to promote the development and adoption of safe, reliable AI systems within the EU’s single market, applicable to both private and public sectors. It aims to balance the advancement of AI technology with the protection of fundamental rights of EU citizens. The regulation covers areas within EU law, exempting systems used solely for military, defence, and research purposes.
Some of the most important provisions of the AI act include:
The enforcement of the AI Act will be managed by an AI Office within the European Commission to oversee rule implementation, as well as a scientific panel of independent experts to support enforcement. An AI Board composed of member state representatives will assist in consistent application of the Act, while an advisory forum to provide technical expertise.
Fines for non-compliance are based on a percentage of the offending company’s global annual turnover or a set amount, whichever is higher. Small and medium-sized enterprises (SMEs) and start-ups face proportionate fines. Before deploying high-risk AI systems, public service entities must assess their impact on fundamental rights. The regulation mandates increased transparency in developing and using high-risk AI systems, requiring registration in the EU database for such systems and disclosure when using emotion recognition technology.
The AI Act encourages innovation through regulatory sandboxes, allowing controlled testing of AI systems in real-world conditions. This framework is designed to facilitate evidence-based regulatory learning and support the development of new AI technologies.
Every day we send out a free e-mail with the most important headlines of the last 24 hours.
Subscribe now