• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Towards Intersectionality in Machine Learning: Including More Identities, Handling Underrepresentation, and Performing Evaluation

October 30, 2022

🔬 Research Summary by Angelina Wang, a PhD student in computer science at Princeton University studying issues of machine learning fairness and algorithmic bias.

[Original paper by Angelina Wang, Vikram V. Ramaswamy, Olga Russakovsky]


Overview: We consider the complexities of incorporating intersectionality into machine learning and how this is much more than simply including more axes of identity. We consider the next steps of deciding which identities to include, handling the increasingly small number of individuals in each group, and performing a fairness evaluation on a large number of groups.


Introduction

Machine learning has typically considered demographic attributes along one axis and as binary. However, gender is not just male and female; race is not just Black and white, and many axes of identity interact to produce unique forms of discrimination. Thus, extending machine learning to intersectionality is not as easy as simply incorporating more axes of identity. In this work, we investigate three considerations that will arise along the machine learning pipeline and offer substantive steps forward: 

  • Data collection: which identities to include
  • Model training: how to handle the increasingly small number of individuals in each group
  • Evaluation: how to perform fairness evaluation on a large number of groups

Key Insights

In our work, we focus on the algorithmic effects of discrimination against demographic subgroups (rather than individuals). Specifically, we conduct empirical studies of five fairness algorithms across a suite of five tabular datasets derived from the US Census with target variables like income and travel time to work. We do so under the framework of the canonical machine learning fairness setting: supervised binary classification of a target label of social importance, which balances accuracy and one mathematical notion of fairness among a finite set of discretely defined demographic groups. These groups may result from a conjunction of identities.

Selecting which Identities to Include

As a first step, we must consider which identities are relevant to include in the task. If we include too many axes of identities and numbers of identities along each axis, we risk having a computationally intractable problem. But if we include too few, we miss critical intersectional differences in how different groups are treated. In our work, we find that a priori knowledge on which identities to include may be insufficient and recommend combining domain knowledge and empirical validation. Domain knowledge based on which identities are relevant for a particular downstream task is essential but also insufficient for understanding whether, for example, two groups share similar input distributions for one particular domain and would actually benefit from being empirically treated as the same group rather than separated.

Handling Progressively Smaller Groups

In model training, we need to consider how we can train a model when the various groups are likely to have a minimal number of individuals as a result of having added additional axes of identity. We first provide a warning that normative implications may disqualify specific existing techniques in machine learning – for example, harmful historical parallels connected to generating synthetic facial images can raise concerns. We instead suggest a potential alternative through leveraging structure in the data, such as people in different groups but with shared identities may share more features than those without shared identities.

Evaluating a Large Number of Groups

For model evaluation, we need to consider how fairness for a large number of groups can be evaluated since most existing fairness evaluation metrics are only designed for two groups, and the difference between them is on some kind of performance metric. When there are a more significant number of groups, researchers frequently use different variations of these pairwise comparisons, such as the difference between the group with the highest and lowest true positive rate. However, all of these metrics look the same if you, for example, swap the labels of the subgroup labeled Black female and white male. This is surprising in the context of intersectionality, where the historical ways different groups have been treated are vitally important. Thus, we suggest both using more deliberative, contextual pairwise comparisons, as well as incorporating additional evaluation metrics that are better able to capture existing disparities in the data. For example, we propose a metric that measures the correlation between the ranking of groups for their positive labels in the datasets and the ranking of groups for their true positive rates predicted by the model.

Between the lines

Machine learning fairness has begun to grapple with the complexities of intersectional identities, but there is still a long way to go. It’s essential to consider the different normative implications that technical decisions may have. For example, while it may be easy to think that a choice like extrapolating existing evaluation metrics to work for more than two groups is a sensible thing to do in terms of having the technical details worked out, there are additional normative implications at play when the data is about individuals of different demographic groups. In other words, more than just having an algorithm or metric technically manipulated to accommodate a more significant number of groups, we need to consider that, for example, historical injustices matter during evaluation, and there are ways we can try to encode this into the metric. In this work, we take a step toward doing this thinking by bridging existing social science work on intersectionality into machine learning.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

A rock embedded with intricate circuit board patterns, held delicately by pale hands drawn in a ghostly style. The contrast between the rough, metallic mineral and the sleek, artificial circuit board illustrates the relationship between raw natural resources and modern technological development. The hands evoke human involvement in the extraction and manufacturing processes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part I

Close-up of a cat sleeping on a computer keyboard

Tech Futures: The threat of AI-generated code to the world’s digital infrastructure

The undying sun hangs in the sky, as people gather around signal towers, working through their digital devices.

Dreams and Realities in Modi’s AI Impact Summit

Illustration of a coral reef ecosystem

Tech Futures: Diversity of Thought and Experience: The UN’s Scientific Panel on AI

This image shows a large white, traditional, old building. The top half of the building represents the humanities (which is symbolised by the embedded text from classic literature which is faintly shown ontop the building). The bottom section of the building is embossed with mathematical formulas to represent the sciences. The middle layer of the image is heavily pixelated. On the steps at the front of the building there is a group of scholars, wearing formal suits and tie attire, who are standing around at the enternace talking and some of them are sitting on the steps. There are two stone, statute-like hands that are stretching the building apart from the left side. In the forefront of the image, there are 8 students - which can only be seen from the back. Their graduation gowns have bright blue hoods and they all look as though they are walking towards the old building which is in the background at a distance. There are a mix of students in the foreground.

Tech Futures: Co-opting Research and Education

related posts

  • Probing Networked Agency: Where is the Locus of Moral Responsibility?

    Probing Networked Agency: Where is the Locus of Moral Responsibility?

  • Real talk: What is Responsible AI?

    Real talk: What is Responsible AI?

  • Sociological Perspectives on Artificial Intelligence: A Typological Reading

    Sociological Perspectives on Artificial Intelligence: A Typological Reading

  • Response to the European Commission’s white paper on AI (2020)

    Response to the European Commission’s white paper on AI (2020)

  • FeedbackLogs: Recording and Incorporating Stakeholder Feedback into Machine Learning Pipelines

    FeedbackLogs: Recording and Incorporating Stakeholder Feedback into Machine Learning Pipelines

  • Mapping value sensitive design onto AI for social good principles

    Mapping value sensitive design onto AI for social good principles

  • From Funding Crisis to AI Misuse: Critical Digital Rights Challenges from RightsCon 2025

    From Funding Crisis to AI Misuse: Critical Digital Rights Challenges from RightsCon 2025

  • The State of AI Ethics Report (Volume 4)

    The State of AI Ethics Report (Volume 4)

  • Worried But Hopeful: The MAIEI State of AI Ethics Panel Recaps a Difficult Year

    Worried But Hopeful: The MAIEI State of AI Ethics Panel Recaps a Difficult Year

  • A Generalist Agent

    A Generalist Agent

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.