• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Achieving Fairness at No Utility Cost via Data Reweighing with Influence

May 28, 2023

🔬 Research Summary by Peizhao Li, a Ph.D. Candidate at Brandeis University working on Machine Learning, with a special interest in Trustworthy and Responsible AI, Algorithmic Fairness, and Deep Learning.

[Original paper by Peizhao Li and Hongfu Liu]


Overview: With the fast development of algorithmic governance, fairness has become a compulsory property for machine learning models to suppress unintentional discrimination. In this paper, we focus on the pre-processing aspect for achieving fairness and propose a data reweighing approach that only adjusts the weight for samples in the training phase. Our algorithm computes individual weight for each training instance via influence function and linear programming, and in most cases, demonstrates cost-free fairness through vanilla classifiers.


Introduction

For artificial intelligence technology deployed in high-stakes applications like welfare distribution or school admission, it is essential to regulate algorithms and prevent unaware discrimination and unfairness in decision-making. Even though general data-driven algorithms are not designed to be unfair, the outcomes can unintentionally violate the AI principle of equality. Typically learning from historically biased data, the learner can retain or amplify the inherent bias if there is no proper constraint on data or algorithms. As a consequence, the decisions from these algorithms may disadvantage users in certain sensitive groups (e.g., women and African Americans), therefore raising societal concerns.

To mitigate unfairness algorithmically, the authors focus on the pre-processing aspect of fair learning algorithms, i.e., only transform the input training data for a machine learning model to make its decision to be fair. The pre-processing category directly diagnoses and corrects the source of bias and can be easily adapted to existing data analytic pipelines. The proposed method is to granularly compute a weight for every sample in the training set, and a vanilla model without any additional constraints can be trained on this reweighed training set and deliver fair predictions.

Key Insights

Characterizing Sample Influence

Consider a classifier trained on some training set. We can assess the contribution from one individual training sample by training the classifier two times with and without that training sample, respectively. By doing so, it is straightforward to know in a counterfactual how the model will change with regard to some typical measurements, e.g., fairness or predictive utility.

However, it can be prohibitively expensive to retrain the model many times. Influence function from robust statistics provides a first-order approximation to the contribution of a training sample. It measures the effect of changing an infinitesimal weight in a sample, then linearly extrapolates to a whole removal of a sample. The influence function offers us an option to estimate the change of a model toward any evaluative function by removing some training samples from the training set.

Fairness at No Utility Cost

Having the influence function at hand, and by writing the fairness goal into a differentiable equation, we can quantify the impact of one training sample on two objectives: predictive utility and fairness. We use the influence function to calculate how the model will change and how the utility and fairness will change by reweighing a training instance. Theoretically, under the assumption of training data sufficiency and diversity, and with a proper model dimension, the authors prove that some training data reweighing strategy always exists to improve the fairness while at least keeping the utility not decreasing.

Having that theoretical finding, the authors solve the reweighing through linear programming and compute the individual weight for each training instance. The linear programs have constraints from both utility and fairness to achieve a fair model prediction without harming the prediction performance. Solving the linear programming problem can be very fast, i.e., within a few seconds, for a tabular dataset with 30k+ training samples.

The authors demonstrate the proposed reweighing approach on multiple tabular datasets compared to several pre-processing and in-processing methods. Experimental results show that such reweighing method can achieve cost-free fairness in most cases. In contrast, other competitive methods usually get fair results by introducing non-negligible harm to the model’s utility.

Between the lines

In recent years, many algorithms have been developed to make ML models meet criteria from AI ethics. However, many algorithms bring non-trivial change to the original model, greatly changing the utility performance. Fairness at no utility cost is a favorable property since it could help to popularize fair algorithms for extensive utility-driven products and alleviate the concerns from the deployment of fair algorithms.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

ALL IN Conference 2025: Four Key Takeaways from Montreal

Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

AI Policy Corner: Restriction vs. Regulation: Comparing State Approaches to AI Mental Health Legislation

Beyond Consultation: Building Inclusive AI Governance for Canada’s Democratic Future

AI Policy Corner: U.S. Executive Order on Advancing AI Education for American Youth

related posts

  • Research summary: AI Mediated Exchange Theory by Xiao Ma and Taylor W. Brown

    Research summary: AI Mediated Exchange Theory by Xiao Ma and Taylor W. Brown

  • The Wrong Kind of AI? Artificial Intelligence and the Future of Labour Demand (Research Summary)

    The Wrong Kind of AI? Artificial Intelligence and the Future of Labour Demand (Research Summary)

  • DICES Dataset: Diversity in Conversational AI Evaluation for Safety

    DICES Dataset: Diversity in Conversational AI Evaluation for Safety

  • Bridging the Gap Between AI and the Public (TEDxYouth@GandyStreet)

    Bridging the Gap Between AI and the Public (TEDxYouth@GandyStreet)

  • Enough With “Human-AI Collaboration”

    Enough With “Human-AI Collaboration”

  • From the Gut? Questions on Artificial Intelligence and Music

    From the Gut? Questions on Artificial Intelligence and Music

  • Can We Engineer Ethical AI?

    Can We Engineer Ethical AI?

  • Analysis and Issues of Artificial Intelligence Ethics in the Process of Recruitment

    Analysis and Issues of Artificial Intelligence Ethics in the Process of Recruitment

  • Anthropomorphic interactions with a robot and robot-like agent

    Anthropomorphic interactions with a robot and robot-like agent

  • From AI Winter to AI Hype: The Story of AI in Montreal

    From AI Winter to AI Hype: The Story of AI in Montreal

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.