• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

AI Economist: Reinforcement Learning is the Future for Equitable Economic Policy

May 14, 2020

By Richard Socher, Chief Scientist and EVP, Salesforce, and Head of Salesforce Research; and Stephan Zheng, Senior Research Scientist, Salesforce Research

Long before pandemic-related lockdowns, economic inequality has been one of the most significant issues affecting humanity. A report from the United Nations in January 2020 found that inequality is rising in most of the developed world. With so much to lose or gain, it’s no surprise that bias can influence policymaking… often to the detriment those who need the most help. This underscores the potential that AI can have for good, and why it’s important to develop tools and solutions that are simulation- and data-driven to yield more equitable policies.

The new AI Economist model from Salesforce Research is designed to address this kind of equality by identifying an optimal tax policy. By using a two-level reinforcement learning (RL) framework, training both (1) AI agents and (2) tax policies, it simulates and helps identify dynamic tax policies that best accomplish a given objective. This RL framework is model-free in that it uses zero prior world knowledge or modeling assumptions, and learns from observable data alone.

Our ultimate goal for the AI Economist is for it to be a stepping stone for future economic AI models that support real-world policy-making to improve social welfare. We believe a RL framework is well-suited for uncovering insights on how the behavior of economic agents could be influenced by pulling different policy “levers.”

The initial results are incredibly promising and prove that it’s possible to layer a simulation and data-driven approach to optimize for any social objective to more quickly create equitable, and effective, tax policies. The AI Economist model achieved results that were at least 16% better than the baseline models used in the study.

When we set out to build the AI Economist, we did so while recognizing the negative impact that inequality has on economic opportunity, health and social welfare. Economics includes numerous variables and considerations driving outcomes, but the AI Economist turns its focus to taxes. Why? Taxes are an important tool that governments use to redistribute wealth and resources – helping to address inequality. It is the first time anyone has used RL to directly learn AI tax policies in AI simulations.

A key difference between the AI Economist’s application of RL and most others is the avoidance of a “zero-sum” outcome… it’s not trying to beat anyone. “Victory” is achieved when the outcome creates an optimal balance between equality and productivity.

The simulation itself manifests into a two-dimensional world that is based on classic tax theory – modeling people who earn income by performing labor. A worker gains utility from income and incurs the cost of the labor labor effort. The simulated world features two types of resources: wood and stone. AI agents (or workers) gather and trade each resource, as well as earn income by building houses.

Each simulation also reflects different skill levels among workers. Higher-skilled workers earn more income, resulting in more utility. But the amount of effort it takes to build them reduces its utility. As learning progresses, workers develop more concrete roles based on their skill level and how each worker learned to balance income vs. effort.

Results of the AI Economist simulations were compared with three different scenarios:

Every scenario included workers that face seven income brackets, corresponding to existing U.S. federal tax brackets. Mimicking human behavior, AI agents found ways to game the system, particularly with regressive tax models, by lowering their taxes by alternating between high and low incomes rather than sustaining one or the other.

  • A free market with no taxes or efforts to redistribute via taxes
  • A progressive tax that increase with income, similar to the 2018 U.S. federal tax schedule
  • An analytical formula proposed by Emmanuel Saez that, in our specific simulation, establishes tax rates that decrease with income

Following millions of simulations, our RL approach produced effective dynamic tax policies that resulted in a better balance of equality and productivity than the other methods tested. The AI Economist achieved, at minimum, a 16% better trade-off between equality and productivity compared to the Saez model, the next best performing scenario. Compared to a free market, the AI Economist generated a 47% advantage in equality while only sacrificing 11% in productivity.

The AI Economist also proved to be more effective in accounting for and addressing aforementioned agents looking to game the system. This is especially significant because in reality, the standards for equality fluctuate over time. Modelers don’t always have the luxury of having historical data to use. The ability to adapt quickly to current socioeconomic trends is crucial to sustaining policy effectiveness in the real world, and simulation-based systems can better test what-if scenarios while mitigating the potential of real side-effects.

We’ve already started testing the impact of the AI Economist with human participants. Although experiments with real people used a simpler ruleset, the AI Economist yielded similar performance improvements as the simulations with only AI agents.

While the simulations represent a small economy for the time being, we feel it is huge step toward building a viable tool with advanced modeling capabilities for economists and governments to test and create effective policies. This iteration of the AI Economist is in some ways already more complex than other models used by economists today. As models evolve and RL grows more sophisticated, future simulations can include ‘smarter’ agents and increase the scope of each simulation to account for more complex economic processes and variables.

We invite further discussion about the AI Economist from all communities, and would love to hear your input. Please reach out to us on Twitter at @RichardSocher and @StephanZheng.


Original Paper by Stephan Zheng, Alex Trott, Sunil Srinivasa, Nikhil Naik, Melvin Gruesbeck, Kathy Baxter, David Parkes and Richard Socher

Full paper: https://arxiv.org/abs/2004.13332

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

This image shows a large white, traditional, old building. The top half of the building represents the humanities (which is symbolised by the embedded text from classic literature which is faintly shown ontop the building). The bottom section of the building is embossed with mathematical formulas to represent the sciences. The middle layer of the image is heavily pixelated. On the steps at the front of the building there is a group of scholars, wearing formal suits and tie attire, who are standing around at the enternace talking and some of them are sitting on the steps. There are two stone, statute-like hands that are stretching the building apart from the left side. In the forefront of the image, there are 8 students - which can only be seen from the back. Their graduation gowns have bright blue hoods and they all look as though they are walking towards the old building which is in the background at a distance. There are a mix of students in the foreground.

Tech Futures: Co-opting Research and Education

Agentic AI systems and algorithmic accountability: a new era of e-commerce

ALL IN Conference 2025: Four Key Takeaways from Montreal

Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

AI Policy Corner: Restriction vs. Regulation: Comparing State Approaches to AI Mental Health Legislation

related posts

  • The Paris AI Summit: Deregulation, Fear, and Surveillance

    The Paris AI Summit: Deregulation, Fear, and Surveillance

  • Jack Clark Presenting the 2022 AI Index Report

    Jack Clark Presenting the 2022 AI Index Report

  • Research summary: Challenging Truth and Trust: A Global Inventory of Organized Social Media Manipula...

    Research summary: Challenging Truth and Trust: A Global Inventory of Organized Social Media Manipula...

  • A survey on adversarial attacks and defences

    A survey on adversarial attacks and defences

  • Who to Trust, How and Why: Untangling AI Ethics Principles, Trustworthiness and Trust

    Who to Trust, How and Why: Untangling AI Ethics Principles, Trustworthiness and Trust

  • The Grand Illusion: The Myth of Software Portability and Implications for ML Progress

    The Grand Illusion: The Myth of Software Portability and Implications for ML Progress

  • Employee Perceptions of the Effective Adoption of AI Principles

    Employee Perceptions of the Effective Adoption of AI Principles

  • From Promise to Practice: A Glimpse into AI-Driven Approaches to Neuroscience

    From Promise to Practice: A Glimpse into AI-Driven Approaches to Neuroscience

  • Universal and Transferable Adversarial Attacks on Aligned Language Models

    Universal and Transferable Adversarial Attacks on Aligned Language Models

  • An Uncommon Task: participator Design in Legal AI

    An Uncommon Task: participator Design in Legal AI

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.