• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Analysis and Issues of Artificial Intelligence Ethics in the Process of Recruitment

January 18, 2022

🔬 Research summary by Connor Wright, our Partnerships Manager.

[Original paper by V. Uday Kumar, A. Mohan, B.Srinivasa S P Kumar, Ramesh Ponnala, B Sateesh, P. Dundy Sai Maruthi]


Overview: AI is starting to become a household name in the hiring process for many businesses. While its involvement in the process varies from case to case, the attention required to tackle the problem of bias does not.


Introduction

AI being deployed in the hiring process is a common theme in today’s recruitment story. Nevertheless, the depth of this deployment varies from application to application. At times, the AI is used to simply schedule interviews, but it has been used to screen candidates. No matter the variation involved, the common thread that holds these applications together is the problem of AI bias. To explore this further, it fits to give some context on what AI in the hiring process looks like.

Key Insights

AI in the hiring process

Given the widely-touted capabilities of AI to streamline business resources, it has been well used in the hiring process. The AI’s involvement can range from screening candidates to scheduling interviews and even helping out in the interview process. However, the main inspiration can be seen in the AI lending a hand to filtering the sheer amounts of applications a job receives.

However, how this filtering is done varies on the application used. The paper details the following programmes which utilise AI to different depths:

  1. XOR interacts with a candidate through a chatbot.
  2. Paradox involves engaging with the candidate through a machine learning algorithm.
  3. Hiretual and AmazingHiring contain a database that it uses to match a particular profile.
  4. Pymetrics and Eightfold focus on cutting employee time spent on reviewing applications.
  5. HireVue, Seekout and MyInterview use the cloud for various tasks, including conducting interviews, filtering and outsourcing candidates.
  6. Humanly uses automated candidate screening.
  7. Fetcher and Loxo contact the candidate through emails and SMS.
  8. Textio helps write job descriptions, which then appeal to some candidates more than others as a form of filtering.

Despite a fruitful variation between the applications themselves, a common thread connects them all: the problem of bias.

The problem of bias

Five different types of bias are explored by the authors, which are all worth considering when deploying AI in the hiring process:

  1. Historical bias – hiring algorithms could contribute to concretizing past tendencies in a company. The company continues to look for what it already knows instead of prioritizing diversity.
  2. Representation bias – the dataset offered to the hiring algorithm must represent all different types of candidates. For example, collecting data only about people who went to university would ignore those who are also qualified for the job through other means, like internships.
  3. Measurement bias – candidate data is erroneously collected, such as being taken from a date outside the specified window.
  4. Aggregation bias – wrongly assuming the trends observed in the data apply to all individual data points. For example, assuming that all candidates from a particular area did not go to university based on a high school drop-out rate.
  5. Evaluation bias – giving more weight to specific character traits as opposed to others

Between the lines

While the problem of bias in AI is well-documented, I believe our attitude to confronting the phenomenon is equally important. The paper’s analysis shows the accuracy of the algorithms to vary significantly., at times involving 30% inaccuracy and others at 10%. From there, the authors point to how various surveys show that employers are not too worried about the 10% inaccuracy involved in AI ethics. For me, adopting the attitude where we do care for the 10% inaccuracy will be essential in the fight against bias, allowing us to take advantage of the deserved attention the AI Ethics field receives.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

An abstract spiral of dark circles appears at the centre, resembling a tornado. Several vintage magazine covers and advertisements are being drawn toward the spiral. The artworks that have already been pulled into it are becoming distorted and replaced with clusters of numbers representing their numerical embeddings.

Tech Futures: Better Imagination for Better Tech Futures

This image is a collage with a colourful Japanese vintage landscape showing a mountain, hills, flowers and other plants and a small stream. There are 3 large black data servers placed in the bottom half of the image, with a cloud of black smoke emitting from them, partly obscuring the scenery.

Tech Futures: Crafting Participatory Tech Futures

A network diagram with lots of little emojis, organised in clusters.

Tech Futures: AI For and Against Knowledge

A brightly coloured illustration which can be viewed in any direction. It has many elements to it working together: men in suits around a table, someone in a data centre, big hands controlling the scenes and holding a phone, people in a production line. Motifs such as network diagrams and melting emojis are placed throughout the busy vignettes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part II

A rock embedded with intricate circuit board patterns, held delicately by pale hands drawn in a ghostly style. The contrast between the rough, metallic mineral and the sleek, artificial circuit board illustrates the relationship between raw natural resources and modern technological development. The hands evoke human involvement in the extraction and manufacturing processes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part I

related posts

  • Bias Propagation in Federated Learning

    Bias Propagation in Federated Learning

  • Disaster City Digital Twin: A Vision for Integrating Artificial and Human Intelligence for Disaster ...

    Disaster City Digital Twin: A Vision for Integrating Artificial and Human Intelligence for Disaster ...

  • South Korea as a Fourth Industrial Revolution Middle Power?

    South Korea as a Fourth Industrial Revolution Middle Power?

  • Computer Vision’s implications for human autonomy

    Computer Vision’s implications for human autonomy

  • Harmonizing Artificial Intelligence: The role of standards in the EU AI Regulation

    Harmonizing Artificial Intelligence: The role of standards in the EU AI Regulation

  • Research Summary: The cognitive science of fake news

    Research Summary: The cognitive science of fake news

  • Human-AI Collaboration in Decision-Making: Beyond Learning to Defer

    Human-AI Collaboration in Decision-Making: Beyond Learning to Defer

  • How Cognitive Biases Affect XAI-assisted Decision-making: A Systematic Review

    How Cognitive Biases Affect XAI-assisted Decision-making: A Systematic Review

  • Responsibility assignment won’t solve the moral issues of artificial intelligence

    Responsibility assignment won’t solve the moral issues of artificial intelligence

  • Responsible Design Patterns for Machine Learning Pipelines

    Responsible Design Patterns for Machine Learning Pipelines

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.