• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Melting contestation: insurance fairness and machine learning

December 14, 2023

🔬 Research Summary by Laurence Barry and Arthur Charpentier.

Laurence Barry is an independent actuary and a researcher at PARI (Programme de Recherche sur l’Appréhension des Risques et des Incertitudes, ENSAE/ Sciences po)

Arthur Charpentier is a professor in Montréal, and is the former director of the Data Science for Actuaries program of the French Institute of Actuaries.

[Original paper by Laurence Barry and Arthur Charpentier]


Overview: Machine learning tends to replace the actuary in the selection of features and the building of pricing models. However, avoiding subjective judgments thanks to automation does not necessarily mean that biases are removed. Nor does the absence of bias warrant fairness. This paper critically analyzes discrimination and insurance fairness with machine learning.


Introduction

Insurers have often been confronted with data-related issues of fairness and discrimination. This paper provides a comparative review of discrimination issues raised by traditional statistics versus machine learning in the context of insurance. We first examine historical contestations of insurance classification, showing that it was organized along three types of bias: pure stereotypes, non-causal correlations, or causal effects that a society chooses to protect against, which are thus the main sources of dispute. The lens of this typology then allows us to look anew at the potential biases in insurance pricing implied by big data and machine learning, showing that despite utopic claims, social stereotypes continue to plague data, thus threatening to unconsciously reproduce these discriminations in insurance. To counter these effects, algorithmic fairness attempts to define mathematical indicators of non-bias. This may prove insufficient since it assumes specific protected groups exist, which could only be made visible through public debate and contestation. These are less likely if the right to explanation is realized through personalized algorithms, which could reinforce the individualized perception of the society that blocks rather than encourages collective mobilization.

Key Insights

Insurance fairness is a dynamic concept, that depends on historical, cultural but also technical contexts. At the height of the industrial era, the veil of ignorance explained the equality of the greatest number in the face of an unknown adversity, justifying a very broad coverage in terms of

solidarity (Ewald, 1986). During the twentieth century, more segmented models were put in place with the growing capacities of data collection and calculation. Still, insurance remained based on risk classes, perceived as homogeneous groups of similar people (Barry, 2020). From the 1980s onwards, controversies arose over using this or that variable that feeds current criticisms of the biases and discriminations associated with machine learning. Examining this history allows us to identify a few main families of bias in traditional classification practices and their recent displacement with machine-learning algorithms.

Type 1 bias: pure prejudice

Some critics pointed out the prejudices that informed the statistician’s choice of variables and warned against the “myth of the actuary,” who would build supposedly objective models based on his own conception of what is risky, moral, or legitimate behavior. In principle, this type of bias should have disappeared with big data, as most are now natively digital, and thus allow us to bypass earlier manual quantification work. However, over the last twenty years, the embeddedness of social prejudices in data has been amply established; blind use of machine learning would then reproduce these biases in the models.

Type 2 bias: non-causal correlation

Another criticism pointed out the use of correlated variables that are not truly causal. For example, the use of the man/woman parameter and the credit score have provoked controversies built on this kind of argument. The solution, surely inoperable in practice, would be to limit models to purely causal variables. Interestingly, some data scientists today advocate for a shift to a new episteme, where the model’s accuracy, rather than its interpretability and its causal format, would find its legitimacy. Current algorithms indeed capture correlations without making these links explicit. This magnifies type 2 biases and further introduces a new bias due to their opacity, even if as the counterpart of greater precision.

Type 3 bias: reinforcing social injustices through classification

Another family of critics rejected the idea of a “true” classification altogether. Type 3 biases in insurance come from truly causal variables that reflect a hazard that society has chosen to mutualize. The use of genetic data in health insurance, for example, is prohibited in most countries. In this case, insurance is seen as a way not to reflect the risk but to have it borne by the entire insured population by eliminating the variable from the models. However, eliminating protected variables was very effective in traditional models but is much more difficult to implement with big data and machine learning, respectively, because protected variables are captured via their correlation with others and because the opacity of the algorithms makes highlighting this effect more complex.

Fairness through contestation

Beyond the difficulties of eliminating traditional biases with new technologies, historical studies also highlight the importance of debate and contestation to define models that could be perceived as fair. In this matter, the attempt to ensure algorithmic fairness using “absence of bias” mathematical indicators might prove insufficient. Only through discussions and sometimes legal actions could protected groups be recognized as such, which algorithmic fairness takes as a given.

While understanding the model is certainly necessary, explainability does not warrant contestability. In this matter, the locality of current explanatory algorithms is inevitable due to their non-linearity. But it might reinforce the individualized perception of the social, which blocks rather than fosters collective mobilization. Should insurance then stick to the good old pricing tables, for which all the parameters are explicit, known in advance, and therefore open to challenge? Just as any scientific theory must be falsifiable, a pricing system should be transparent to be contestable. In any case, one should be wary of replacing the myth of the actuary with the myth of the algorithm.

Between the lines

Current research on actuarial fairness is focused on the absence of biased mathematical indices. But the capacity to debate on the use of specific features that might otherwise remain unnoticed is crucial to ensure the basic premise of insurance: that the most vulnerable get protection. This promise sometimes means accepting cross-subsidies between groups (hence, statistical biases) for the sake of the common good.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

This image shows a large white, traditional, old building. The top half of the building represents the humanities (which is symbolised by the embedded text from classic literature which is faintly shown ontop the building). The bottom section of the building is embossed with mathematical formulas to represent the sciences. The middle layer of the image is heavily pixelated. On the steps at the front of the building there is a group of scholars, wearing formal suits and tie attire, who are standing around at the enternace talking and some of them are sitting on the steps. There are two stone, statute-like hands that are stretching the building apart from the left side. In the forefront of the image, there are 8 students - which can only be seen from the back. Their graduation gowns have bright blue hoods and they all look as though they are walking towards the old building which is in the background at a distance. There are a mix of students in the foreground.

Tech Futures: Co-opting Research and Education

Agentic AI systems and algorithmic accountability: a new era of e-commerce

ALL IN Conference 2025: Four Key Takeaways from Montreal

Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

AI Policy Corner: Restriction vs. Regulation: Comparing State Approaches to AI Mental Health Legislation

related posts

  • A Taxonomy of Foundation Model based Systems for Responsible-AI-by-Design

    A Taxonomy of Foundation Model based Systems for Responsible-AI-by-Design

  • Levels of AGI: Operationalizing Progress on the Path to AGI

    Levels of AGI: Operationalizing Progress on the Path to AGI

  • Deployment corrections: An incident response framework for frontier AI models

    Deployment corrections: An incident response framework for frontier AI models

  • Engaging the Public in AI's Journey: Lessons from the UK AI Safety Summit on Standards, Policy, and ...

    Engaging the Public in AI's Journey: Lessons from the UK AI Safety Summit on Standards, Policy, and ...

  • Putting AI ethics to work: are the tools fit for purpose?

    Putting AI ethics to work: are the tools fit for purpose?

  • Rewiring What-to-Watch-Next Recommendations to Reduce Radicalization Pathways

    Rewiring What-to-Watch-Next Recommendations to Reduce Radicalization Pathways

  • AI vs. Maya Angelou: Experimental Evidence That People Cannot Differentiate AI-Generated From Human-...

    AI vs. Maya Angelou: Experimental Evidence That People Cannot Differentiate AI-Generated From Human-...

  • Visions of Artificial Intelligence and Robots in Science Fiction: a computational analysis

    Visions of Artificial Intelligence and Robots in Science Fiction: a computational analysis

  • Evaluating a Methodology for Increasing AI Transparency: A Case Study

    Evaluating a Methodology for Increasing AI Transparency: A Case Study

  • Hiring Algorithms Based on Junk Science May Cost You Your Dream Job

    Hiring Algorithms Based on Junk Science May Cost You Your Dream Job

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.