Intergovernmental

How countries are implementing the OECD Principles for Trustworthy AI

Since signing on to the OECD AI Principles in 2019, countries have been using them as guidance to craft policies to tackle AI risks and capitalise on opportunities. The Principles are a global reference and the first intergovernmental standard for trustworthy AI, reflecting the fundamental challenges and opportunities for economies and societies.

The OECD report The state of implementation of the OECD AI Principles four years on, takes stock of national initiatives to implement the OECD AI Principles worldwide. It overviews national AI strategies, monitoring bodies, expert advisory groups, and monitoring and evaluation frameworks. It reviews regulatory approaches to date and provides examples of initiatives related to each AI principle so policymakers can learn from one another.

Here is a quick look at the report’s most important findings.

Countries worldwide are adopting national AI strategies

In 2017, only a few countries had national AI strategies. Today, the OECD.AI Policy Observatory contains over 50 national strategic and government-wide initiatives for trustworthy AI (Figure 1). All OECD members and partner countries in the Arab region, Africa, and South America are promoting the OECD AI Principles. By May 2023, governments reported more than 930 policy initiatives across 71 jurisdictions in the OECD.AI national policy database​.  

A timeline of AI strategies
Figure 1. A timeline of national AI strategies

Governments are using different governance models to manage national AI policies. Some countries have created governmental coordination bodies for AI. The United Kingdom founded the Government Office for AI, a unit within the Department for Science, Innovation and Technology, and the United States established the Artificial Intelligence Initiative Office (NAIIO), located within the White House Office of Science and Technology Policy.

Other countries tasked existing ministries and established AI inter-ministerial and multi-stakeholder committees to oversee the development and implementation of AI strategies. Examples of this approach include the Governance Committee of the Brazilian AI and Egypt’s National Council for AI (NCAI).

Outside of government, countries are establishing multi-stakeholder groups of AI experts to advise and report on the current and future opportunities, risks and challenges for the public use of AI. France’s National Consultative Committee on Digital Ethics and AI (FNCDE) and Canada’s Advisory Council on AI are pertinent examples.

Countries like Germany and Canada established AI observatories to monitor and evaluate their AI strategy rollouts. While still scarce, these monitoring and evaluation mechanisms should expand across other countries as national AI strategies move into later stages of implementation.

Follow us on LinkedIn

Ethical frameworks, hard and soft laws are among the regulatory frameworks countries use to ensure trustworthy AI systems

Countries are exploring different regulatory approaches to encourage trustworthy AI and mitigate the risks of AI systems. Existing legislation in other areas already regulates AI systems, and countries have started codifying OECD AI Principles and national AI strategies into AI-specific regulatory frameworks.

Several countries have national ethical frameworks and AI principles that align with the OECD AI Principles. In May 2023, 17 of these were included in the OECD database. Japan, Korea, and India provide guidance for developers and operators on how to implement the principles. Colombia has set up an online platform to monitor its framework’s implementation.

Second, many countries are considering AI-specific regulatory approaches (Figure 2). Others, such as Canada and the European Union, have proposed to regulate AI systems across domains and applications with comprehensive frameworks applicable to all sectors. The Committee on Artificial Intelligence of the Council of Europe is drafting a horizontal international legal instrument on AI.

The United States, the United Kingdom, Israel and China have opted for a sectoral approach by establishing cross-sectoral, non-binding principles. This leaves regulators to implement and enforce regulations in their respective sectors.

More broadly, countries are supporting international standardisation efforts. These include ISO’s ISO/IEC 23053 (2022), establishing its Framework for AI Systems Using Machine Learning and ISO 31000 for risk management.

Some countries promote controlled environments for experimentation through regulatory sandboxes, i.e., spaces where authorities engage firms to test innovative products and services that challenge existing legal frameworks. For example, in 2022, Spain created an AI regulatory sandbox as the first pilot programme to test the future proposed EU AI Act. The United Kingdom and Norway also launched regulatory sandbox initiatives that include AI.

Figure 2 Examples of existing and emerging AI-specific regulatory approaches
Figure 2. Examples of existing and emerging AI-specific regulatory approaches

Implementing values-based OECD AI Principles

Policy actions that put the OECD AI values-based principles into concrete, operational initiatives are the most relevant development in the last two years. The OECD AI Principles identify five complementary values-based principles for the responsible stewardship of trustworthy AI: (1.1) inclusive growth, sustainable development and well-being, (1.2) human-centred values and fairness, (1.3) transparency and explainability, (1.4) robustness, security and safety, and (1.5) accountability.

These core values guide national and international principles and policies, including emerging AI-specific legislation.

Inclusive growth, sustainable development, and well-being (Principle 1.1)

Governments are pursuing different approaches to achieve inclusive growth, sustainable development, and well-being. Most national AI strategies and AI ethics frameworks or guidelines for implementing AI refer to this principle.

At the policy level, countries have launched initiatives to ensure that stakeholders, including vulnerable groups, are involved in policy design and benefit from AI systems, including through initiatives to augment human capabilities and enhance creativity. Governments are also funding projects that use AI to address environmental challenges.

Figure 3. Inclusive growth, sustainable development, and well-being select national policies
CanadaQuebec AI Forum
ChileParticipation Process on AI
ColombiaCoordination Bodies for AI Policy Implementation
European CommissionDestination Earth (DestinE)
FranceIA Booster
GermanyAI Lighthouses for the Environment, Climate, Nature, and Resources Initiative
GermanyAUTONOM – Performing Arts and AI programme
TürkiyeBreast Cancer Detection with AI project
United KingdomWomen in Data Science and AI initiative
United StatesNational Institutes of Health’s AIM-AHEAD programme

Human-centred values and fairness (Principle 1.2)

To implement principle 1.2, many governments have issued primarily non-binding guidelines and initiatives to reduce AI biases and promote human rights and human-centred values. By contrast, only some have developed Human Rights Impact Assessments (HRIA) and quality seals.

Figure 4. Human-centred values and fairness, select national policies
FranceAddressing Human Rights Concerns Arising From Facial Recognition Technology
GermanyAI Seal of Quality
KoreaAI Personal Information Protection Self-Checklist
MexicoRecommendations for the Processing of Personal Data derived from the use of AI
The NetherlandsFundamental Right and Algorithmic Impact Assessment
TürkiyeTrustworthy AI Stamp
United KingdomReview into Bias in Algorithmic Decision-Making
United StatesArtificial Intelligence and Algorithmic Fairness Initiative

Transparency and explainability (Principle 1.3)

Governments are taking various approaches to ensure AI transparency, ranging from guidelines for implementing AI to oversight bodies. Regulatory bodies have recognised the importance of AI transparency and explainability. Transparency provisions are already established in existing legislation in areas like data protection, privacy and consumer protection legislation. They are included in proposed AI-specific regulations with specific provisions for the workplace. In the public sector, governments are enhancing transparency around using AI for public services through initiatives such as AI registers.

Figure 5 Transparency and explainability Select national policies
Figure 5. Transparency and explainability, select national policies
CanadaDirective on Automated Decision-Making
CanadaProposed Artificial Intelligence and Data Act (AIDA)
European UnionGeneral Data Protection Regulation (GDPR)
European UnionProposed AI Act
Finland and The NetherlandsOpen AI Register of AI systems’ use in the public sector
SpainProvision to Workers’ Statute Law Detailing Employers’ Algorithmic Transparency Requirements

Robustness, security, and safety (Principle 1.4)

Countries are drawing on guidelines, ethics frameworks, impact assessments, new legislation, amendments to existing legislation and other instruments to implement Principle 1.4. Here, the focus is primarily on traffic laws and regulations intended to prevent unreasonable safety risks and how these apply to AI systems. In addition, countries develop risk management approaches and mechanisms to maintain records of data characteristics.

Figure 6 Robustness security and safety Select national policies
Figure 6. Robustness security and safety, select national policies
AustriaAutomated Driving Regulation
LithuaniaLaw on Road Traffic Safety
MexicoPrinciples and Impact Analysis Guide for the Development and Use of Systems Based on Artificial Intelligence in the Federal Public Administration
TürkiyeNational Data Dicitionary
United StatesNational Security Presidential Memorandum on Protecting the United States’ Advantage in Artificial Intelligence and Related Critical Technologies
UruguayAlgorithmic Impact Study

Accountability (Principle 1.5)

Countries have developed codes of ethical conduct for using or implementing AI in several sectors (public administration, health care, autonomous driving). It is important to note that the proposed AI-specific regulation requires documentation that explains the AI system’s proper functioning throughout its lifecycle. Lastly, countries have established independent oversight bodies to audit algorithm use.

Figure 7 Accountability Select national policies
Figure 7. Accountability, select national policies
CanadaProposed Artificial Intelligence and Data Act (AIDA)
European UnionProposed AI Act
PortugalGuide to Artificial Intelligence in the Public Administration
SingaporeAI in Healthcare Guidelines
The NetherlandsAlgorithm Supervision Body

Implementing the five recommendations to governments

Investing in AI Research & Development (Principle 2.1)

Many countries recognise the importance of policies to support AI R&D and have put forward initiatives to ramp up these efforts. Most national AI strategies focus on AI R&D as one of the key areas for action. Countries have dedicated AI R&D funding and support it through different instruments. Main trends include launching AI R&D-focused policies, plans, and programmes, supporting the creation of national AI research institutes and centres, and consolidating AI research networks and collaborative platforms.

Figure 8 Investing in AI R&D Select national policies
Figure 8. Investing in AI R&D, select national policies
AustraliaAI Solutions to Build a Stronger Australia
CanadaNational AI Institutes (Amii, Vector Institute, Mila)
ChileNational AI Research Centre
European UnionHorizon Europe 2021-2027
FranceNational AI Research Programme
JapanAI R&D Network
United KingdomAI Sector Deal
United StatesAI Researchers Portal

Fostering a digital ecosystem for AI (Principle 2.2)

Embracing AI-enabled transformation depends on the availability of data, infrastructure, and software to train and use AI models at scale. Thus, countries must ensure sufficient AI compute capacity to meet their needs and capture AI’s full economic potential. Fostering a digital ecosystem for AI represents a crucial component of countries’ efforts to advance AI adoption. Efforts include linking AI policies to data access and sharing policies. This strengthens efforts to increase computing capacity, access to infrastructure and investment in NLP technologies.

Figure 9 Fostering a digital ecosystem for AI Select national policies
Figure 9. Fostering a digital ecosystem for AI, select national policies
DenmarkGigaword Project
European CommissionEuropean High Performance Computing Joint Undertaking
FranceHealth Data Hub
NorwayResource Centre for Sharing of Data
Several European CountriesCommon Language Resources and Technology Infrastructure (CLARIN Research Infrastructure Consortium)
United StatesNational AI Research Resource Task Force

 

Fostering an enabling policy environment for AI (Principle 2.3)

Countries encourage trustworthy AI use and innovation by reviewing and adapting existing policies and legislations and adopting AI-specific regulatory frameworks. They also support an agile transition from R&D to commercialisation or deployment of AI by providing controlled environments for experimentation and testing of AI systems.

Figure 10 Fostering an enabling policy environment for AI Select national policies
Figure 10. Fostering an enabling policy environment for AI, select national policies
European UnionProposed AI Act
JapanSocial Principles of Human-Centric AI
KoreaHuman-Centred National Guidelines for AI Ethics
NorwayData Protection Authority Regulatory Sandbox
SingaporeFinTech Regulatory Sandbox
United StatesAI Bill of Rights

AI skills, jobs and labour market transformation (Principle 2.4)

Governments have put initiatives in place for the labour market’s transition and to provide the workforce with the necessary skills for AI through formal education programmes, vocational training, and lifelong learning initiatives. They also launched initiatives to attract and retain AI talent and monitor AI’s impact in the labour market.

Figure 11. AI skills jobs and labour market transformation, select national policies
AustraliaNext-Generation AI Graduates Programme
IndiaAWS Young Builders Challenge
JapanPractical Guidebook on Providing Data for Employee Development in AI and Data Science
KenyaDigital Literacy Programme
SingaporeAI for Industry
SingaporeGuide to Job Redesign in the Age of AI
United StatesAI Training for the Acquisition Workforce Bill
United StatesAmerican Workforce Policy Advisory Board

International and multi-stakeholder cooperation on AI (Principle 2.5)

Countries are engaging international cooperation to promote the beneficial use of AI and address its challenges. This is happening through several types of initiatives, including: i) international research collaborations on AI, ii) trade agreements including language on AI, and iii) cooperation for AI capacity building in developing countries.

Figure 12 International and multi-stakeholder cooperation on AI Select national
Figure 12. International and multi-stakeholder cooperation on AI, select national policies
Australia and SingaporeSingapore-Australia Digital Economy Agreement (SADEA)
CanadaMontreal Declaration for Responsible AI
Chile, New Zealand and SingaporeDigital Economy Partnership Agreement (DEPA)
European UnionHorizon Europe
GermanyArtificial Intelligence for All – FAIR Forward
KoreaAI Ethics Policy Forum
Several Like-Minded CountriesGlobal Partnership on AI (GPAI)
World BankHarnessing AI for Development

At the G7 Summit meeting in May 2023, hosted in Hiroshima, leaders agreed to task their digital and tech ministers to establish the “Hiroshima AI process”, where G7 members continue to discuss generative AI inclusively, in cooperation with the OECD and GPAI by the end of 2023 (Ministry of Foreign Affairs of Japan, 2023). On 7 September 2023, G7 Digital and Tech Ministers issued a statement (MIC, 2023) endorsing: 1) a report by the OECD summarising a stocktaking of priority risks (OECD, 2023), challenges and opportunities of generative AI based on priorities highlighted in the G7 Leaders’ Statement, (2) work towards international guiding principles applicable for all AI actors, (3) developing a code of conduct for organisations developing advanced AI systems, to be presented to the G7 Leaders and (4) project-based cooperation in support of the development of responsible AI tools and best practices



Disclaimer: The opinions expressed and arguments employed herein are solely those of the authors and do not necessarily reflect the official views of the OECD or its member countries. The Organisation cannot be held responsible for possible violations of copyright resulting from the posting of any written material on this website/blog.