• Skip to main content
  • Skip to primary sidebar
  • Skip to footer
  • Core Principles of Responsible AI
    • Accountability
    • Fairness
    • Privacy
    • Safety and Security
    • Sustainability
    • Transparency
  • Special Topics
    • AI in Industry
    • Ethical Implications
    • Human-Centered Design
    • Regulatory Landscape
    • Technical Methods
  • Living Dictionary
  • State of AI Ethics
  • AI Ethics Brief
  • 🇫🇷
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

Against Interpretability: a Critical Examination

November 5, 2021

🔬 Research summary by Dr. Andrea Pedeferri, instructional designer and leader in higher ed (Faculty at Union College), and founder at Logica, helping learners become more efficient thinkers.

[Original paper by Maya Krishnan]


Overview:  “Explainability”, “transparency, “interpretability” … these are all terms that are employed in different ways in the AI ecosystem. Indeed, we often hear that we should make AI more “interpretable” and/or more “explainable”.  In contrast, the author of this paper challenges the idea that “interpretability” and the like should be values or requirements for AI. First, it seems that these concepts are not really intuitively clear and technically implementable. Second, they are mostly proposed not as values in themselves but as means to reach some other valuable goals (e.g. fairness, respect for users’ privacy). And so the author argues that rather than directing our attention to “interpretability” or “explainability” per se, we should focus on the ethical and epistemic goals we set for AI while also making sure we can adopt a variety of solutions and tools to reach those goals.


Introduction

As we mentioned in a previous summary, back box’s opaqueness poses both epistemic (are the algorithms in fact reliable?) and ethical (are the algorithms ethical?) challenges. Relatedly, they also seem to violate people’s claim-right to know why a certain algorithm has produced some predictions or automated decisions that concern them. For many the epistemic and ethical risks back box’s opaqueness poses could be mitigated by making sure that AI is somehow interpretable, explainable and/or (as some say) transparent. Contrary to the received view on this issue, the author of this paper challenges the idea that there is a black box problem and denies that “interpretability”, “explainability” or “transparency” should be values or requirements for AI (see also here).

Key Insights

The author of this paper challenges the idea that there is a black box problem and that “interpretability”, “explainability” or “transparency” should be criteria for evaluating an AI system.

The first problem the author points out is that the terms above (“explainability”, “transparency, “interpretability”) are often unclear and poorly defined. Let’s take “interpretability”: AI is interpretable when, roughly, it is understandable to consumers. The author notices that this definition is not really helpful: it does not clarify what “understandable” means and does not offer any insight on what the term could mean when applied to algorithms. Also, there is confusion on what should be understandable. Here are a few candidates: the prediction of the algorithm itself, the inner workings of the AI that produced its prediction, or the reasons why / the justification for the algorithm making that prediction. Which one is key for human understanding?

Many seem to believe that explaining how an algorithm reached a certain outcome is tantamount to making AI understandable. However, it should be noted that causal explanations are not the same as justifications. That is, the reason or justification for a given outcome might not clearly map into the causal path that brought the algorithm to that conclusion. As the author puts it, “[t]his point is particularly apparent in the case of neural networks. The causal process by which some input triggers a certain pathway within the network does not straightforwardly map on to justificatory considerations.” Indeed it would be like asking “a person why they have given an answer to a particular question and they respond with an account of how their neurons are firing”. The causal story is not a rational explanation per se. Thus, if the explanation we look for tells us only about the causal path that gets the algorithm to a certain conclusion, this story would not provide the right level of explanation needed to rationally understand that very outcome. 

Finally, the author notices that interpretability, explainability and the like are a means to an end, i.e. ensuring that AI is ethical and trustworthy. The author recommends that we focus on those goals instead of treating interpretability and the like as they were ends in themselves. Since there might be other ways to reach those goals, it seems unhelpful to focus just on one set of solutions.

Between the lines

The paper rightly points out that we need a more coherent and precise analysis of concepts such as interpretability and explainability. And the author also clarifies that “[w]hile this paper questions both the importance and the coherence of interpretability and cognates, it does not make a decisive case for the abandonment of the concepts.” We agree with this too, as we appreciate the importance of ensuring explainability in AI as a way to both assess whether algorithms are ethical, robust and reliable, and to protect people’s right to know and understand how assessments are made. To do so, however, we first need to make sure that we agree on what we mean by ‘explanation’ and on what kind of explanation is needed for real human understanding.   

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

AI Policy Corner: Singapore’s National AI Strategy 2.0

AI Governance in a Competitive World: Balancing Innovation, Regulation and Ethics | Point Zero Forum 2025

AI Policy Corner: Frontier AI Safety Commitments, AI Seoul Summit 2024

AI Policy Corner: The Colorado State Deepfakes Act

Special Edition: Honouring the Legacy of Abhishek Gupta (1992–2024)

related posts

  • Transferring Fairness under Distribution Shifts via Fair Consistency Regularization

    Transferring Fairness under Distribution Shifts via Fair Consistency Regularization

  • Towards Algorithmic Fairness in Space-Time: Filling in Black Holes and Detecting Bias in the Presenc...

    Towards Algorithmic Fairness in Space-Time: Filling in Black Holes and Detecting Bias in the Presenc...

  • Robust Distortion-free Watermarks for Language Models

    Robust Distortion-free Watermarks for Language Models

  • Theorizing Femininity in AI: a Framework for Undoing Technology’s Gender Troubles (Research Summary)

    Theorizing Femininity in AI: a Framework for Undoing Technology’s Gender Troubles (Research Summary)

  • Research summary: Acting the Part: Examining Information Operations Within #BlackLivesMatter Discour...

    Research summary: Acting the Part: Examining Information Operations Within #BlackLivesMatter Discour...

  • Toward Responsible AI Use: Considerations for Sustainability Impact Assessment

    Toward Responsible AI Use: Considerations for Sustainability Impact Assessment

  • Measuring Disparate Outcomes of Content Recommendation Algorithms with Distributional Inequality Met...

    Measuring Disparate Outcomes of Content Recommendation Algorithms with Distributional Inequality Met...

  • Building a Credible Case for Safety: Waymo's Approach for the Determination of Absence of Unreasonab...

    Building a Credible Case for Safety: Waymo's Approach for the Determination of Absence of Unreasonab...

  • The State of AI Ethics Report (Volume 5)

    The State of AI Ethics Report (Volume 5)

  • Managing Human and Robots Together - Can That Be a Leadership Dilemma?

    Managing Human and Robots Together - Can That Be a Leadership Dilemma?

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer

Categories


• Blog
• Research Summaries
• Columns
• Core Principles of Responsible AI
• Special Topics

Signature Content


• The State Of AI Ethics

• The Living Dictionary

• The AI Ethics Brief

Learn More


• About

• Open Access Policy

• Contributions Policy

• Editorial Stance on AI Tools

• Press

• Donate

• Contact

The AI Ethics Brief (bi-weekly newsletter)

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.


Archive

  • © MONTREAL AI ETHICS INSTITUTE. All rights reserved 2024.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.