• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

AI Policy Corner: New York City Local Law 144

June 9, 2025

✍️ By Vedant Thakur.

Vedant is an Undergraduate Student in Artificial Intelligence & Philosophy and a Research Assistant at the Governance and Responsible AI Lab (GRAIL), Purdue University.


📌 Editor’s Note: This article is part of our AI Policy Corner series, a collaboration between the Montreal AI Ethics Institute (MAIEI) and the Governance and Responsible AI Lab (GRAIL) at Purdue University. The series provides concise insights into critical AI policy developments from the local to international levels, helping our readers stay informed about the evolving landscape of AI governance. This piece spotlights Local Law 144 of New York City.


New York City Local Law 144

The New York City Local Law 144 (LL144) addresses the growing use of artificial intelligence in employment decisions, specifically in hiring and promotion. This law was adopted on July 5, 2023. 

An Automated Employment Decision Tool (AEDT) is a tool that substantially assists the hiring decision-making process. The tool must output a score, classification, or prediction that the computer identifies and weighs based on various inputs. LL144 considers any use of artificial intelligence, machine learning, data analytics, or statistical modeling as automation.

The law requires two conditions for employers or employment agencies to use AEDTs:

1.  Conduct Independent Bias Audits

These audits require calculating selection rates, scoring rates, and impact ratios across race/ethnicity, sex, and their intersections. These three metrics measure the disparate impact of the Automated Employment Decision Tool. 

2.  AEDT Transparency

Notice must be provided to candidates for AEDT use at least 10 business days in advance, and audit results must be published. The results must be posted for six months after the latest use of an AEDT.

Enforcement

The NYC Department of Consumer and Worker Protection (DCWP) is tasked with enforcing LL144. Violations of LL144 are subject to civil penalties ranging from $500 for the first offence and up to $1500 for subsequent offences per day. However, LL144, by itself, does not penalize discrimination as it does not require any scoring thresholds to be met. These cases will be addressed with the Uniform Guidelines on Employee Selection Procedures (29 CFR § 1607), where adverse or disparate impact is defined by the four-fifths rule. Complaints of discrimination will be referred to the NYC Human Rights Commission, while the DCWP will only address violations of LL144 (the use of AEDT without the required notices).

How Effective is the NYC Local Law 144?

LL144 defines AEDTs as tools that substantially assist or completely replace human decision-making in the hiring process. The conditions to meet this definition are open to interpretation and at the employer’s discretion. Many human-in-the-loop systems would not fit this definition, thereby excluding many employers that use automated tools from the scope of LL144. The scope is further limited by only addressing race/ethnicity and sex, but not age, disability, or other protected classes. 

Furthermore, LL144 only addresses users of the AEDTs, not the vendors or developers. Hence, the burden of correcting any disparate impact or bias then lies with the employers themselves. And without direct access to the AEDT, this may be unfeasible or ineffective. The combination of these factors, according to various auditors of these AEDTs, makes LL144 well-intentioned but ineffective.

Despite these concerns, LL144 remains a significant step towards fair artificial intelligence systems. It showcases how governance strategies, such as public disclosures and audits, can mitigate the risks of bias, discrimination, and civil rights violations. It further demonstrates the merits of transparent AI, as it allows AEDTs to be held accountable to the same standards as humans when it comes to employment decisions.

Further Reading

  1. US Equal Employment Opportunity Commission Guidance
  2. Auditing AI  
  3. Regulation of AI-Enabled Employment Decisions

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

Illustration of a coral reef ecosystem

Tech Futures: Diversity of Thought and Experience: The UN’s Scientific Panel on AI

This image shows a large white, traditional, old building. The top half of the building represents the humanities (which is symbolised by the embedded text from classic literature which is faintly shown ontop the building). The bottom section of the building is embossed with mathematical formulas to represent the sciences. The middle layer of the image is heavily pixelated. On the steps at the front of the building there is a group of scholars, wearing formal suits and tie attire, who are standing around at the enternace talking and some of them are sitting on the steps. There are two stone, statute-like hands that are stretching the building apart from the left side. In the forefront of the image, there are 8 students - which can only be seen from the back. Their graduation gowns have bright blue hoods and they all look as though they are walking towards the old building which is in the background at a distance. There are a mix of students in the foreground.

Tech Futures: Co-opting Research and Education

Agentic AI systems and algorithmic accountability: a new era of e-commerce

ALL IN Conference 2025: Four Key Takeaways from Montreal

Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

related posts

  • The Stanislavsky projects approach to teaching technology ethics

    The "Stanislavsky projects" approach to teaching technology ethics

  • Am I Literate? Redefining Literacy in the Age of Artificial Intelligence

    Am I Literate? Redefining Literacy in the Age of Artificial Intelligence

  • AI Policy Corner: Automating Licensed Professions: Assessing Health Technology and Other Industries

    AI Policy Corner: Automating Licensed Professions: Assessing Health Technology and Other Industries

  • AI Policy Corner: AI Governance in East Asia: Comparing the AI Acts of South Korea and Japan

    AI Policy Corner: AI Governance in East Asia: Comparing the AI Acts of South Korea and Japan

  • AI Policy Corner: Frontier AI Safety Commitments, AI Seoul Summit 2024

    AI Policy Corner: Frontier AI Safety Commitments, AI Seoul Summit 2024

  • Canada’s Minister of AI and Digital Innovation is a Historic First. Here’s What We Recommend.

    Canada’s Minister of AI and Digital Innovation is a Historic First. Here’s What We Recommend.

  • AI Policy Corner: Singapore's National AI Strategy 2.0

    AI Policy Corner: Singapore's National AI Strategy 2.0

  • Who's watching? What you need to know about personal data security

    Who's watching? What you need to know about personal data security

  • Open Letter: Moving Forward Together – MAIEI’s Next Chapter

    Open Letter: Moving Forward Together – MAIEI’s Next Chapter

  • Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

    Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.