• Skip to main content
  • Skip to primary sidebar
  • Skip to footer
  • Core Principles of Responsible AI
    • Accountability
    • Fairness
    • Privacy
    • Safety and Security
    • Sustainability
    • Transparency
  • Special Topics
    • AI in Industry
    • Ethical Implications
    • Human-Centered Design
    • Regulatory Landscape
    • Technical Methods
  • Living Dictionary
  • State of AI Ethics
  • AI Ethics Brief
  • 🇫🇷
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

Contextualizing Artificially Intelligent Morality: A Meta-Ethnography of Top-Down, Bottom-Up, and Hybrid Models for Theoretical and Applied Ethics in Artificial Intelligence

June 1, 2023

🔬 Research Summary by Jennafer Shae Roberts, a researcher and writer for the non-profit Accel AI Institute. She focuses on the ethics of artificial intelligence coming from an anthropology background.

[Original paper by Jennafer Shae Roberts and Laura Montoya]


Overview: Ethics in Artificial Intelligence (AI) can emerge in many ways. This paper addresses developmental methodologies, including top-down, bottom-up, and hybrid approaches to ethics in AI, from theoretical, technical, and political perspectives. Examples through case studies of the complexity of AI ethics are discussed to provide a global perspective when approaching this challenging and often overlooked area of research.


Introduction

How can we make ethical AI when our world is so terribly unethical?

This research is about digging into the many facets of ethics in AI and how these perspectives are developed to gain an understanding of the current landscape of it. This is important because if nothing is done to create ethical AI, it will replicate questionable and sometimes outright harmful ideals, translating to autonomous outputs and actions learned from our less-than-perfect society. 

As researchers, we combed through papers and articles on the subject. We built upon the ideas of other researchers, such as Wallach et al., who first wrote about top-down, bottom-up and hybrid AI morality and included technical and theoretical angles. We added the political perspective to acknowledge where power is situated and how this comes into play when designing ethical AI systems. 

Ethics in AI works differently in varying parts of the world, suggesting that a one-size-fits-all approach won’t work for all applications or use cases, and won’t affect everyone equally.  However, having conversations about ethics in AI could radiate out and lead us to face societal ethics that are outdated and sometimes outright wrong. 

Key Insights

Could a focus on ethics in AI lead to an Ethics Revolution?

Given the inherent complexity of ethics and morality in AI, we utilized the top-down, bottom-up, and hybrid framework to address technical, theoretical, and political perspectives. 

Top-downBottom-upHybrid
TechnicalProgrammed rules 
Ex: call center chatbot
Machine learning 
Ex: Reinforcement learning, etc
Has a base of rules or instructions, but then also is fed data to learn from as it goes
Ex: autonomous vehicles employ some rule-based ethics, while also learning from other drivers and road experience
TheoreticalRule-utilitarianism and deontological ethics, principles, ie fairness
Ex: the Golden Rule, the Ten Commandments, consequentialist or utilitarian ethics, Kant’s moral imperative and other duty based theories, Aristotle’s virtues, and Asimov’s laws of robotics.
Experience-based, case-based reasoning
Ex: learn as you go from real-world consequences
Personal moral matrix combining rules and learned experience 
Ex: having some rules but also developing ethics through experiences
PoliticalCorporate and political powers
Ex: company’s list of principles
People power
Ex: groups online of individuals calling for ethics in AI
A combination of ethics from those in power who take into account ethics called for by the people
Ex: employees collaborating with their corporation on ethical AI issues

What do we mean by top-down, bottom-up, and hybrid in the context of ethics in AI?

Top-down signifies that there is some sort of rules coming from a power source at the top. Whether technical programming, theoretical philosophy, or political powers from governments or big tech companies, top-down ethics for AI can look good on paper, such as lists of principles, however, in practice, these can be too broad to be action-guiding and can cause disparities and inequalities which ultimately undermine ethics. (Roberts and Montoya, 2022)  

Bottom-up ethics, broadly speaking, is a method of learning from experience. This can be good and bad, as the world is flawed, and sometimes what seems ethical is not actually in line with commonly held values. 

Hybrid ethics in AI attempts to take the best of both top-down and bottom-up approaches and create something that is more well-roundedly ethical. However, many examples of what we would label as hybrid ethics in AI still need a lot of work, such as with contact tracing apps for COVID-19 or self-driving vehicles that balance the rules of the road and what they learn from actual drivers. 

How do we define the technical, theoretical, and political perspectives of ethics in AI?

Technical top-down and bottom-up ethics in AI primarily concerns how AI learns ethics. Machines are good at narrow tasks, such as memorization or data collection. However, AI systems can fall short in areas such as objective reasoning, which is at the core of ethics. Teaching ethics to AI is extremely difficult, both technically and socially. 

Theoretical ethics concerns philosophers and ethicists throughout the ages, who have much to argue about when it comes to right and wrong, which varies greatly across cultures. Furthermore, ethics has been historically made for people, which doesn’t directly translate to AI. 

Political ethics highlights where the power and decision-making are coming from, which then has an effect that radiates outward and influences systems, programmers, and users alike. We don’t only refer to politics as coming from the government but also from big corporations and from the bottom-up, coming from crowdsourcing, protests, and other forms of people power. (Roberts and Montoya, 2022)

Case Studies and what we learned 

Taking this framework, we applied it to real-world scenarios of applied AI that were ethically debatable. One of the examples was about contact tracing apps for COVID-19 and how differently these worked in places with different levels of government and community compliance. Another example traced an NGO which went to an Indigenous African community to use machine learning to calculate water access, which inadvertently caused harm to the community, and trust was lost. 

Contrary to what science fiction would leave us to believe about what to fear with AI, some of the most troubling ethical conundrums are when AI replicates oppressive systems, mainly disadvantaging those already marginalized and replicating racism, sexism, classism, etc. Utilizing this framework, ethics in AI can be addressed from all angles and hopefully not only create more ethical AI but a more ethical world.

Between the lines

While completing this research and engaging with ethics in AI, the lack of societal ethics cannot be ignored. It is often easy to agree on something wrong, but agreeing on solutions is a lot more challenging. We hope that at least those who intend to build and deploy ethical AI systems will consider all angles and blind spots, including those who might be marginalized or harmed by the technology, especially when it aims to help. Most AI is utilized for consumerism and capitalism, such as algorithms that decide what ads to show us as we scroll. These are ethically shifty areas. On the far end, and what we omitted from this research, is AI being used for warfare or to purposefully harm or control populations. This is an extremely unethical use of technology. As Ani Difranco says, ‘Every tool is a weapon, if you hold it right.’ However, tools can also be used to build miraculous things and change the world. We hope by exploring different angles of ethics in AI that, it can continue to improve ethically as the technology develops. 

References

[1] Allen, C., Smit, I., & Wallach, W. (2005). Artificial Morality: Top-down, Bottom-up, and Hybrid Approaches. Ethics and Information Technology, 7(3), 149–155. https://doi.org/10.1007/s10676-006-0004-4

Wallach, W., Allen, C., & Smit, I. (2008). Machine morality: bottom-up and top-down approaches for modeling human moral faculties. AI & SOCIETY, 22(4), 565–582. https://doi.org/10.1007/s00146-007-0099-0

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

AI Policy Corner: Frontier AI Safety Commitments, AI Seoul Summit 2024

AI Policy Corner: The Colorado State Deepfakes Act

Special Edition: Honouring the Legacy of Abhishek Gupta (1992–2024)

AI Policy Corner: The Turkish Artificial Intelligence Law Proposal

From Funding Crisis to AI Misuse: Critical Digital Rights Challenges from RightsCon 2025

related posts

  • Co-Designing Checklists to Understand Organizational Challenges and Opportunities around Fairness in...

    Co-Designing Checklists to Understand Organizational Challenges and Opportunities around Fairness in...

  • Whose AI Dream? In search of the aspiration in data annotation.

    Whose AI Dream? In search of the aspiration in data annotation.

  • Research summary: Health Care, Capabilities, and AI Assistive Technologies

    Research summary: Health Care, Capabilities, and AI Assistive Technologies

  • Research summary: Changing My Mind About AI, Universal Basic Income, and the Value of Data

    Research summary: Changing My Mind About AI, Universal Basic Income, and the Value of Data

  • Ethics-based auditing of automated decision-making systems: intervention points and policy implicati...

    Ethics-based auditing of automated decision-making systems: intervention points and policy implicati...

  • Sociological Perspectives on Artificial Intelligence: A Typological Reading

    Sociological Perspectives on Artificial Intelligence: A Typological Reading

  • Performative Power

    Performative Power

  • Bias Amplification Enhances Minority Group Performance

    Bias Amplification Enhances Minority Group Performance

  • Responsible Generative AI: A Reference Architecture for Designing Foundation Model-based Agents

    Responsible Generative AI: A Reference Architecture for Designing Foundation Model-based Agents

  • A collection of principles for guiding and evaluating large language models

    A collection of principles for guiding and evaluating large language models

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer

Categories


• Blog
• Research Summaries
• Columns
• Core Principles of Responsible AI
• Special Topics

Signature Content


• The State Of AI Ethics

• The Living Dictionary

• The AI Ethics Brief

Learn More


• About

• Open Access Policy

• Contributions Policy

• Editorial Stance on AI Tools

• Press

• Donate

• Contact

The AI Ethics Brief (bi-weekly newsletter)

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.


Archive

  • © MONTREAL AI ETHICS INSTITUTE. All rights reserved 2024.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.