• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

AI Policy Corner: Frontier AI Safety Commitments, AI Seoul Summit 2024

April 28, 2025

✍️ By Alexander Wilhelm.

Alexander is a PhD Student in Political Science and a Graduate Affiliate at the Governance and Responsible AI Lab (GRAIL), Purdue University.


📌 Editor’s Note: This article is part of our AI Policy Corner series, a collaboration between the Montreal AI Ethics Institute (MAIEI) and the Governance and Responsible AI Lab (GRAIL) at Purdue University. The series provides concise insights into critical AI policy developments from the local to international levels, helping our readers stay informed about the evolving landscape of AI governance.


Frontier AI Safety Commitments, AI Seoul Summit 2024

Discussions between governments, civil society, and companies on the ‘safe’ development of AI have advanced through collaborations such as the AI Safety Summit 2023 held in the UK and the AI Seoul Summit 2024. Led by the United Kingdom and the Republic of South Korea, the Seoul Summit resulted in a framework of commitments, known as the Frontier AI Safety Commitments, which 20 organizations, including Anthropic, Microsoft, NVIDIA, and OpenAI, have agreed to. These commitments required signatories to publish “a safety framework focused on severe risks” at the AI Summit in France in February 2025 (See The AI Ethics Brief #158 for more on the Paris AI Action Summit). However, rhetoric at the Paris Summit emphasized the benefits of AI rather than its potential harms and risks, raising questions about the future of the three goals outlined in the Frontier AI Safety Commitments.

Three outcomes of the Frontier AI Safety Commitments

Outcome 1: Organisations effectively identify, assess and manage risks when developing and deploying their frontier AI models and systems.

  • Signatories to the Commitments agree to identify risks relevant to their frontier models, including risks detected by external entities and governments. Frontier models are defined within the Commitments as “highly capable general-purpose AI models or systems that can perform a wide variety of tasks and match or exceed the capabilities present in the most advanced models.” Multiple stakeholders are expected to collaboratively identify unacceptable levels of risk within frontier models, with justifications for the boundaries once they are set. Risk mitigation should then be planned to maintain the acceptable levels, with a commitment not to develop models that fail to meet these standards. 

Outcome 2: Organisations are accountable for safely developing and deploying their frontier AI models and systems.

  • Groups that voluntarily pledge to join the Frontier AI Safety Commitments must update their policies on an ongoing basis, extending the viability of the agreement as these technologies evolve.

Outcome 3: Organisations’ approaches to frontier AI safety are appropriately transparent to external actors, including governments.

  • Signatories are expected to provide transparency to the public except when “doing so would increase risk or divulge sensitive commercial information to a degree disproportionate to the societal benefit.” The caveat, however, is that more details should be provided to “trusted actors,” such as a home government. Finally, external actors should be engaged in the assessment of risk, the organization’s internal plans to safely develop frontier AI models, and their follow through in implementing these plans.

Recent Developments in Frontier AI Governance

The Frontier AI Safety Commitments provide a framework to mitigate risks to safety, security, and transparency, while discussing governance strategies such as disclosure, evaluation, and performance requirements. While not all AI development organizations have signed the Frontier AI Safety Commitments, consensus on frontier AI standards is developing, as the Frontier AI Safety Commitments are reflected in China’s AI Safety Commitments. 

Nonetheless, some experts remain concerned about the voluntary nature of these commitments. The Paris AI Summit’s focus on the promise and opportunity of AI instead of the risks latent in frontier models led to disappointment for some civil society groups. The voluntary commitments remain for the 20 signatories to the Frontier AI Safety Commitments, but the future of such standards is an open question as the focus of AI Summits shifts.

Further Reading

  1. The AI Seoul Summit 2024
  2. Tech Giants Pledge AI Safety Commitments — Including a ‘Kill Switch’ if They Can’t Mitigate Risks
  3. The Bletchley Park Process Could be a Building Block for Global Cooperation on AI Safety

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

A network diagram with lots of little emojis, organised in clusters.

Tech Futures: AI For and Against Knowledge

A brightly coloured illustration which can be viewed in any direction. It has many elements to it working together: men in suits around a table, someone in a data centre, big hands controlling the scenes and holding a phone, people in a production line. Motifs such as network diagrams and melting emojis are placed throughout the busy vignettes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part II

A rock embedded with intricate circuit board patterns, held delicately by pale hands drawn in a ghostly style. The contrast between the rough, metallic mineral and the sleek, artificial circuit board illustrates the relationship between raw natural resources and modern technological development. The hands evoke human involvement in the extraction and manufacturing processes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part I

Close-up of a cat sleeping on a computer keyboard

Tech Futures: The threat of AI-generated code to the world’s digital infrastructure

The undying sun hangs in the sky, as people gather around signal towers, working through their digital devices.

Dreams and Realities in Modi’s AI Impact Summit

related posts

  • AI Policy Corner: Discussing the White House’s 2025 AI Action Plan

    AI Policy Corner: Discussing the White House’s 2025 AI Action Plan

  • AI Policy Corner: AI Governance in East Asia: Comparing the AI Acts of South Korea and Japan

    AI Policy Corner: AI Governance in East Asia: Comparing the AI Acts of South Korea and Japan

  • Maintaining fairness across distribution shift: do we have viable solutions for real-world applicati...

    Maintaining fairness across distribution shift: do we have viable solutions for real-world applicati...

  • The Case for Anticipating Undesirable Consequences of Computing Innovations Early, Often, and Across...

    The Case for Anticipating Undesirable Consequences of Computing Innovations Early, Often, and Across...

  • Regulating Artificial Intelligence: The EU AI Act - Part 1

    Regulating Artificial Intelligence: The EU AI Act - Part 1

  • Research summary: Overcoming Barriers to Cross-Cultural Cooperation in AI Ethics and Governance

    Research summary: Overcoming Barriers to Cross-Cultural Cooperation in AI Ethics and Governance

  • It’s COMPASlicated: The Messy Relationship between RAI Datasets and Algorithmic Fairness Benchmarks

    It’s COMPASlicated: The Messy Relationship between RAI Datasets and Algorithmic Fairness Benchmarks

  • AI Ethics Maturity Model

    AI Ethics Maturity Model

  • Research summary: Lexicon of Lies: Terms for Problematic Information

    Research summary: Lexicon of Lies: Terms for Problematic Information

  • Beyond Bias and Compliance: Towards Individual Agency and Plurality of Ethics in AI

    Beyond Bias and Compliance: Towards Individual Agency and Plurality of Ethics in AI

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.