• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

AI Certification: Advancing Ethical Practice by Reducing Information Asymmetries

August 18, 2021

🔬 Research summary by Jonas Schuett, Policy Research Intern at DeepMind | Research Fellow at the Legal Priorities Project | PhD Candidate in Law at Goethe University Frankfurt

[Original paper by Peter Cihon, Moritz J. Kleinaltenkamp, Jonas Schuett, Seth D. Baum]


Overview: How can we incentivize the adoption of AI ethics principles? This paper explores the role of certification. Based on a review of the management literature on certification, it shows how AI certification can reduce information asymmetries and incentivize change. It also surveys the current landscape of AI certification schemes and briefly discusses implications for the future of AI research and development.


Introduction

Certification is widely used to convey that an entity has met some sort of performance standard. It includes everything from the certificate that people receive for completing a university degree to certificates for energy efficiency in consumer appliances and quality management in organizations. As AI technology becomes increasingly impactful across society, there can be a role for certification to improve AI governance. This paper presents an overview of AI certification, applying insights from prior research and experience with certification in other domains to the relatively new domain of AI certification.

Key Insights

Certification can reduce information asymmetries

A primary role of certification is to reduce information asymmetries. Information asymmetries are acute in AI systems because the systems are often complex and opaque and users typically lack the data and expertise necessary to understand them. For example, it is difficult or impossible to evaluate from the outside how biased or explainable a model is, or whether it was developed according to certain ethics principles.

Certification can incentivize change

In reducing the asymmetry of information between insiders and outsiders, certification can further serve to incentivize good behavior by the insiders. For example, corporations may be more motivated to achieve ethics standards if they can use certification to demonstrate their achievements to customers who value these achievements.

The current landscape of AI certification

The paper surveys the landscape of AI certification from 2020, identifying seven active and proposed programs:

  • the European Commission White Paper on Artificial Intelligence (this is outdated, see the proposed Artificial Intelligence Act),
  • the IEEE Ethics Certification Program for Autonomous and Intelligence Systems,
  • the Malta AI Innovative Technology Arrangement,
  • the Turing Certification proposed by Australia’s Chief Scientist,
  • the Queen’s University executive education program Principles of AI Implementation,
  • the Finland civics course Elements of AI, and
  • a Danish program in development for labeling IT-security and responsible use of data.

These programs demonstrate the variety of forms AI certification can take, including both public and private, certifying both individuals and groups, and covering a range of AI-related activities.

The value of certification for future AI research and development

Finally, the paper addresses the potential value of certification for future AI technology. Some aspects of certification will likely remain relevant even as the technology changes. For example, the various roles of corporations, their employees and management, governments, and other actors tend to stay the same. Likewise, certification programs can remain relevant over time by emphasizing human and institutional factors. Programs can also build in mechanisms to update their certification criteria as AI technology changes. Looking further into the future, certification may play a constructive role in governance of the processes that lead to the development of advanced systems. Certification could be especially valuable for building trust among rival AI development groups and ensuring that advanced AI systems are built to high standards of safety and ethics.

Between the lines

In summary, certification can be a valuable tool for AI governance. It is not a panacea for ensuring ethical AI, but it can help especially for reducing information asymmetries and incentivizing ethical AI development and use. The paper presents the first-ever research study of AI certification and therefore serves to establish essential fundamentals of the topic, including key terms and concepts.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

This image is a collage with a colourful Japanese vintage landscape showing a mountain, hills, flowers and other plants and a small stream. There are 3 large black data servers placed in the bottom half of the image, with a cloud of black smoke emitting from them, partly obscuring the scenery.

Tech Futures: Crafting Participatory Tech Futures

A network diagram with lots of little emojis, organised in clusters.

Tech Futures: AI For and Against Knowledge

A brightly coloured illustration which can be viewed in any direction. It has many elements to it working together: men in suits around a table, someone in a data centre, big hands controlling the scenes and holding a phone, people in a production line. Motifs such as network diagrams and melting emojis are placed throughout the busy vignettes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part II

A rock embedded with intricate circuit board patterns, held delicately by pale hands drawn in a ghostly style. The contrast between the rough, metallic mineral and the sleek, artificial circuit board illustrates the relationship between raw natural resources and modern technological development. The hands evoke human involvement in the extraction and manufacturing processes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part I

Close-up of a cat sleeping on a computer keyboard

Tech Futures: The threat of AI-generated code to the world’s digital infrastructure

related posts

  • Jack Clark Presenting the 2022 AI Index Report

    Jack Clark Presenting the 2022 AI Index Report

  • Achieving a ‘Good AI Society’: Comparing the Aims and Progress of the EU and the US

    Achieving a ‘Good AI Society’: Comparing the Aims and Progress of the EU and the US

  • Are we ready for a multispecies Westworld?

    Are we ready for a multispecies Westworld?

  • Responsible Generative AI: A Reference Architecture for Designing Foundation Model-based Agents

    Responsible Generative AI: A Reference Architecture for Designing Foundation Model-based Agents

  • Enhancing Trust in AI Through Industry Self-Governance

    Enhancing Trust in AI Through Industry Self-Governance

  • The Evolution of the Draft European Union AI Act after the European Parliament’s Amendments

    The Evolution of the Draft European Union AI Act after the European Parliament’s Amendments

  • Explaining the Principles to Practices Gap in AI

    Explaining the Principles to Practices Gap in AI

  • The State of AI Ethics Report (June 2020)

    The State of AI Ethics Report (June 2020)

  • Designing for Meaningful Human Control in Military Human-Machine Teams

    Designing for Meaningful Human Control in Military Human-Machine Teams

  • Learning to Prompt in the Classroom to Understand AI Limits: A pilot study

    Learning to Prompt in the Classroom to Understand AI Limits: A pilot study

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.