Policy monitor

Council of Europe Framework Convention on Artificial Intelligence and Human Rights, Democracy and the Rule of Law

On 17 May 2024, the Council of Europe adopted the Framework Convention on Artificial Intelligence, the world’s first international legally binding treaty focused on regulating AI systems in alignment with human rights, democracy, and the rule of law. Developed through international collaboration, the Convention offers a comprehensive legal framework for activities within the lifecycle of artificial intelligence undertaken by public authorities and private actors, promoting innovation while addressing risks. The convention was since signed by the EU Commission. The convention was signed by the EU Commission on september 5th 2024.

What: legislation - treaty

Impact score: 1

For who: citizens, AI providers, policy makers

URL: https://rm.coe.int/1680afae3c and https://rm.coe.int/1680afae67

On 17 May 2024, the Council of Europe adopted the Framework Convention on artificial intelligence and human rights, democracy, and the rule of law. This marks the world’s first international legally binding treaty focused on regulating AI systems with respect to human rights, democracy, and the rule of law. It encourages innovation and advancement in AI while addressing the potential risks to human rights, democracy, and the rule of law. To ensure its relevance over time, the framework is designed to be technology-neutral.

Background

The negotiations began in September 2022, under the guidance of the Committee on AI (CAI), a body established by the Council of Europe. The European Commission, negotiating on behalf of the EU, the 46 Council of Europe Member States, and 11 non-member states (the USA, Canada, Mexico, Japan, Israel, Australia, Argentina, Peru, Uruguay, Costa Rica and the Holy See) participated in these negotiations. Input from 68 international stakeholders, including some civil society groups, academia, industry, and other global organizations, helped ensure a comprehensive and inclusive negotiation process. This broad involvement highlights the global nature of the challenges and the opportunities posed by AI systems.

Key highlights

The Convention offers some cornerstones for a legal framework that governs AI systems throughout their lifecycle to ensure compatibility with human rights, democracy, and the rule of law. It covers AI systems used both in public and private sectors. Recognizing the global differences in legal systems, the Convention offers parties two different ways of complying with its principles and obligations towards privacte actors. Parties can either opt to be directly obliged by the relevant convention provisions or, as an alternative, implement other measures to comply with the Convention's provisions.

The Convention includes a number of key concepts, such as:

- Human-centric AI: In particular, the Convention underscores the importance of aligning AI systems with human rights, democracy, and the rule of law.

- Risk-based approach: To manage AI risks effectively, the convention introduces transparency and oversight requirements tailored to specific risks and contexts. The parties are required to provide measures to identify and prevent possible risks, including the potential need for a ban or other measures on AI systems that could pose incompatibilities with human rights.

- Accountability and remedies: The parties should ensure accountability and responsibility for possible negative effects stemming from AI use. In addition, victims of human rights violations due to AI must have access to effective legal remedies and procedural safeguards.

- Regulatory sandboxes for safe innovation: Parties should enable developing, experimenting, and testing AI systems in controlled environments under the supervision of their competent authorities.

- Oversight mechanisms: The Convention requires the establishment of oversight by independent bodies to ensure compliance with the Convention.

- Implementation and follow-up: A follow-up mechanism is included in order to ensure the effective enforcement and implementation of the Convention’s standards.

While there are certain exemptions for national security and research, these activities are still expected to respect international law and democratic principles. However, matters of national defense are outside the scope of the Convention.

Compatibility with the EU AI Act

The European Commission played a significant role in shaping the Convention, resulting in the consistency of the convention with the EU AI Act. Both the Convention and the EU AI Act share key principles, including a risk-based approach, transparency and oversight requirements, measures ensuring accountability and responsibility, measures fostering equality and non-discrimination, and the provision of effective remedies. This consistency ensures that European Union member states can smoothly integrate the Convention’s standards. As the Convention largely overlaps with several principles included in the EU AI Act, it promotes a more harmonized approach to the use of AI. The Convention also reinforces the significance of the OECD’s proposed definition of AI systems. The alignment on definitions is notable, as it aids to establish common ground, facilitating interoperability across different AI governance frameworks globally.

Looking ahead: the road to effective AI governance

The Convention has already been ratified by Andorra, Georgia, Iceland, Norway, the Republic of Moldova, San Marino, the United Kingdom, as well as Israel, the United States, and the European Union. This means that many of the leading AI-producing countries, such as the United States, the United Kingdom and Israel, are now aligning with principles similar to those in the EU AI Act. The Framework Convention is an important step forward in the global dialogue on AI governance, as well as a notable example of consensus by diverse stakeholders on a rapidly evolving technological issue. Ultimately, the success of the Framework Convention in meeting its objectives depends on the commitment of the parties to uphold and implement its principles in their own legal systems. As AI technology continues to evolve at a rapid pace, harmonizing global strategies to AI governance will be essential for maximizing its potential benefits while mitigating its risks.