• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Benchmark Dataset Dynamics, Bias and Privacy Challenges in Voice Biometrics Research

September 16, 2023

šŸ”¬ Research Summary by Anna Leschanowsky, a research associate at Fraunhofer IIS in Germany working at the intersection of voice technology, human-machine interaction and privacy.

[Original paper by Casandra Rusti, Anna Leschanowsky, Carolyn Quinlan, Michaela Pnacek(ova), Lauriane Gorce, and Wiebke (Toussaint) Hutiri]


Overview: Datasets lie at the heart of data-driven systems. This paper uncovers the influence of dataset practices on bias, fairness, and privacy within data-driven systems, particularly in the context of speaker recognition technology. The authors analyze usage patterns and dynamics from a decade of research in this field and demonstrate how datasets have been instrumental in shaping speaker recognition technology.Ā 


Introduction

Imagine a world where your voice is the only key you need, rendering passwords and PIN codes obsolete. This future is within reach thanks to the widespread adoption of speaker recognition technology across sectors like banking, immigration, and healthcare.

But have you ever wondered about the underlying drivers that enable this remarkable technological advancement?

The authors, who collaborated on the FairEVA project (https://faireva.org), unravel the story behind the datasets that underpin speaker recognition technology and explore their impact on bias, fairness, and privacy. The authors examined almost 700 papers from the past research decade on speaker recognition technology. They demonstrate that datasets primarily emphasize addressing technical challenges, with less attention given to demographic representation concerns. They also highlight changes in data practices since the rise of deep neural networks in speaker recognition, which have raised significant concerns regarding privacy and fairness.

Despite the rapid advancements in this field, pressing ethical questions regarding bias, fairness, and privacy within speaker recognition have remained largely unexplored. Their research underscores the need for ongoing investigations into dataset practices to address these issues. 

Key Insights

Bias in Biometrics and Data

Biometric systems are the modern-day gatekeepers of our digital world and use our unique characteristics, like our faces or voices, to safeguard our valuable assets. From unlocking smartphones to accessing services, biometric technology has become integral to our daily lives. These systems are driven by complex machine-learning models and are not immune to bias. 

The root cause often lies in the very datasets used to train and evaluate them. Datasets are the building blocks of these models, but they frequently fail to accurately represent the diversity of the real world. While previous studies have delved into dataset usage in the realm of face recognition, this paper takes a different path. The authors review nearly 700 papers presented at a prominent international speech research conference between 2012 and 2021 to investigate the usage of datasets within the speaker recognition research community.

The NIST Speaker Recognition Evaluations 

The NIST Speaker Recognition Evaluations (SREs) serve as a crucial benchmarking resource in the world of speaker recognition research. These evaluations are regularly released by the National Institute of Standards and Technology (NIST) to foster the development of speaker recognition technology. The authors explain that ā€œthe NIST SREs were both users and drivers of these dataset collections, as annual evaluation challenges required new datasets to evaluate speaker recognition technology in ever more difficult settings.ā€ 

Dataset Usage in Speaker Recognition

To identify usage patterns, the authors distinguish between datasets that were used for training speaker recognition systems and datasets used for evaluation of the systems. However, identifying training and evaluation datasets proved challenging due to naming inconsistencies. The authors created “dataset families” that represent datasets more generally to address this.

They uncovered a staggering 185 unique training and 164 unique evaluation dataset families used over the past decade in speaker recognition. Despite this diversity, a handful of datasets, particularly the NIST SRE datasets, have dominated the research field.

One standout observation is the prominence of VoxCeleb datasets. These datasets were brought to life by the Visual Geometry Group (VGG) at the University of Oxford and were crafted by scraping YouTube celebrity videos. They aimed to create a large-scale speaker recognition dataset that captures real-world speech conditions. The VoxCeleb datasets marked a significant milestone as the first large-scale, freely available datasets for speaker recognition. 

The authors’ research uncovered a concerning trend where many studies assess their systems using only one dataset, and only a few use more than three. This pattern reflects issues seen in the broader field of machine learning and indicates potential reliability issues within speaker recognition technology.

Dataset collection and bias

Data collection methods have a substantial impact on bias in speaker recognition datasets. The researchers show how the chosen data collection methods can lead to significant representation bias. For example, in some datasets, most participants were college students, creating a dataset skewed toward a younger demographic. In the case of VoxCeleb, which was scraped from YouTube, the researchers argue that the ā€œautomated processing pipeline reinforces popularity bias from search results in candidate selection.ā€

More than Bias: Privacy Threats and Ethical Questions

Finally, the research sheds light on substantial privacy concerns around these datasets. These concerns stem from the content of recorded conversations and the extensive metadata linked to them, which could enable the re-identification of participants. Additionally, web-scraped datasets like VoxCeleb lack consent from data subjects and have raised ethical questions due to the sensitive nature of voice data and its broad applications. These concerns underscore the need for rigorous data protection measures in the speaker recognition field.

Between the lines

Datasets are essential for the development of data-driven systems. Rusti et al. shine a light on dataset usage within the yet unexplored field of speaker recognition technology. Their research highlights how datasets are critical in shaping this technology and raises awareness of potential issues like bias, privacy, and ethics. The authors emphasize the importance of representative evaluation datasets and privacy-preserving voice processing to mitigate privacy risks. As speaker recognition technology becomes increasingly integrated into our lives, ensuring that it works equitably for all users while safeguarding people’s privacy is crucial.

This research serves as a wake-up call, urging the speaker recognition community to be mindful of dataset choices and ethical implications. It prompts us to ask questions about the impact of technology on society and the importance of fairness, transparency, and privacy in the development of AI systems on a broader scale.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

šŸ” SEARCH

Spotlight

Close-up of a cat sleeping on a computer keyboard

Tech Futures: The threat of AI-generated code to the world’s digital infrastructure

The undying sun hangs in the sky, as people gather around signal towers, working through their digital devices.

Dreams and Realities in Modi’s AI Impact Summit

Illustration of a coral reef ecosystem

Tech Futures: Diversity of Thought and Experience: The UN’s Scientific Panel on AI

This image shows a large white, traditional, old building. The top half of the building represents the humanities (which is symbolised by the embedded text from classic literature which is faintly shown ontop the building). The bottom section of the building is embossed with mathematical formulas to represent the sciences. The middle layer of the image is heavily pixelated. On the steps at the front of the building there is a group of scholars, wearing formal suits and tie attire, who are standing around at the enternace talking and some of them are sitting on the steps. There are two stone, statute-like hands that are stretching the building apart from the left side. In the forefront of the image, there are 8 students - which can only be seen from the back. Their graduation gowns have bright blue hoods and they all look as though they are walking towards the old building which is in the background at a distance. There are a mix of students in the foreground.

Tech Futures: Co-opting Research and Education

Agentic AI systems and algorithmic accountability: a new era of e-commerce

related posts

  • Lanfrica: A Participatory Approach to Documenting Machine Translation Research on African Languages ...

    Lanfrica: A Participatory Approach to Documenting Machine Translation Research on African Languages ...

  • Responsible Design Patterns for Machine Learning Pipelines

    Responsible Design Patterns for Machine Learning Pipelines

  • UK’s roadmap to AI supremacy: Is the ā€˜AI War’ heating up?

    UK’s roadmap to AI supremacy: Is the ā€˜AI War’ heating up?

  • Studying up Machine Learning Data: Why Talk About Bias When We Mean Power?

    Studying up Machine Learning Data: Why Talk About Bias When We Mean Power?

  • The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks (Research Summa...

    The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks (Research Summa...

  • Human-AI Interactions and Societal Pitfalls

    Human-AI Interactions and Societal Pitfalls

  • Conformity Assessments and Post-market Monitoring: A Guide to the Role of Auditing in the Proposed E...

    Conformity Assessments and Post-market Monitoring: A Guide to the Role of Auditing in the Proposed E...

  • A Holistic Assessment of the Reliability of Machine Learning Systems

    A Holistic Assessment of the Reliability of Machine Learning Systems

  • Explainable artificial intelligence (XAI) post‐hoc explainability methods: risks and limitations in ...

    Explainable artificial intelligence (XAI) post‐hoc explainability methods: risks and limitations in ...

  • Open and Linked Data Model for Carbon Footprint Scenarios

    Open and Linked Data Model for Carbon Footprint Scenarios

Partners

  • Ā 
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • Ā© 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.