• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Talking About Large Language Models

July 17, 2023

🔬 Research Summary by Murray Shanahan, a senior research scientist at Google DeepMind and Professor of Cognitive Robotics at Imperial College London.

[Original paper by Murray Shanahan]


Overview: The words we use to talk about current large language models (LLMs) can amplify the tendency to see them as human-like. Yet LLM-based dialogue agents are so fundamentally different from human language users that we cannot assume their behavior will conform to normal human expectations. So while it is natural to use everyday psychological terms like “believes,” “knows,” and “thinks” in the context of these systems, we should do so with caution and avoid anthropomorphism.


Introduction

Do large language models have beliefs? Do they really know anything? Do they have thoughts? These are important questions, and they tend to elicit strong opinions in people. According to this paper, the right question to ask is whether we should extend the use of familiar but philosophically contentious words like “believes,” “knows,” and “thinks” to the exotic setting of LLMs and the systems built on them. Do the nature and behavior of these artifacts warrant using such words?

Key Insights

The first important distinction the paper makes is between bare-bones language models (e.g., GPT-4) and the systems, such as dialogue agents (e.g., ChatGPT), based on those models. The bare-bones LLM is a computational object that takes a sequence of tokens (words) as input and returns a distribution over words as output. This distribution represents the model’s prediction of what the next word in the sequence is likely to be.

Although we might speak, elliptically, of the knowledge it encodes, much as we speak of the knowledge encoded in an encyclopedia, the bare-bones LLM, by itself, has such a limited repertoire of behavior that it doesn’t make conceptual sense to say it has beliefs. In particular, as the paper argues, once trained, a bare-bones LLM “has no [further] access to any external reality against which its words might be measured, nor the means to apply any other external criteria of truth, such as agreement with other language-users.”

By interacting with a user, LLM-based dialogue agents (as opposed to bare-bones LLMs) do have some contact with external reality, so at least the question of whether they have beliefs starts to make conceptual sense. However, if a dialogue agent’s only connection with the outside world is linguistic interaction with a human user, then it isn’t appropriate to ascribe beliefs to it in the fullest sense. Such a dialogue agent “cannot participate fully in the human language game of truth because it does not inhabit the world we human language-users share.” In particular, as a disembodied entity, it cannot investigate the world directly as a human or animal can.

There are many ways to move beyond such basic dialogue agents, including multi-modality (e.g., incorporating visual input), tool use (e.g., the ability to consult external websites), and embodiment (e.g., embedding a dialogue system in a robot controller). Each of these extensions provides a richer form of interaction between an LLM-based system and the external world, and each further legitimizes the ascription of beliefs to the resulting systems.

We’re not there yet, but eventually, perhaps, with further progress along these lines, there will be no need to caveat our use of words like “believes” in the context of AI systems based on LLMs. Nevertheless, we should still beware of anthropomorphizing such systems. “The sudden presence among us of exotic, mind-like entities might precipitate a shift in the way we use familiar psychological terms … [But] it may require an extensive period of interacting with, of living with, these new kinds of artifact before we learn how best to talk about them.”

Between the lines

A casual reader of the paper might interpret it as taking a skeptical stance toward artificial intelligence. However, this would be a misreading in two ways. First, the main target of the paper is disembodied large language models, not artificial intelligence more generally, and especially not AI systems that are embodied, either physically in robots or virtually in simulated worlds.

Second, the philosophical position behind the paper is not prescriptive. It is not assumed that there are facts of the matter, underwritten by metaphysics, about the nature of belief, knowledge, thought, and so on. The focus, rather, is on how words are used, and the strategy is to remind us how certain philosophically tricky words are used in their original setting, which is humans interacting with other humans.

Nevertheless, the paper is predominantly critical. It cautions against the use of certain words in the context of LLMs, and the only alternative forms of description it suggests are low-level and mechanistic. A companion paper advocates a more nuanced alternative: to frame the behavior of LLM-based dialogue systems in terms of role-play ([2305.16367] Role-Play with Large Language Models). This framing legitimizes the use of everyday psychological terms while avoiding the pitfalls of anthropomorphism.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

ALL IN Conference 2025: Four Key Takeaways from Montreal

Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

AI Policy Corner: Restriction vs. Regulation: Comparing State Approaches to AI Mental Health Legislation

Beyond Consultation: Building Inclusive AI Governance for Canada’s Democratic Future

AI Policy Corner: U.S. Executive Order on Advancing AI Education for American Youth

related posts

  • Balancing Data Utility and Confidentiality in the 2020 US Census

    Balancing Data Utility and Confidentiality in the 2020 US Census

  • Who Audits the Auditors? Recommendations from a field scan of the algorithmic auditing ecosystem

    Who Audits the Auditors? Recommendations from a field scan of the algorithmic auditing ecosystem

  • Humans, AI, and Context: Understanding End-Users’ Trust in a Real-World Computer Vision Application

    Humans, AI, and Context: Understanding End-Users’ Trust in a Real-World Computer Vision Application

  • The Design Space of Generative Models

    The Design Space of Generative Models

  • Towards Responsible AI in the Era of ChatGPT: A Reference Architecture for Designing Foundation Mode...

    Towards Responsible AI in the Era of ChatGPT: A Reference Architecture for Designing Foundation Mode...

  • Atomist or holist? A diagnosis and vision for more productive interdisciplinary AI ethics dialogue

    Atomist or holist? A diagnosis and vision for more productive interdisciplinary AI ethics dialogue

  • On the Challenges of Using Black-Box APIs for Toxicity Evaluation in Research

    On the Challenges of Using Black-Box APIs for Toxicity Evaluation in Research

  • Human-AI Interactions and Societal Pitfalls

    Human-AI Interactions and Societal Pitfalls

  • Regional Differences in Information Privacy Concerns After the Facebook-Cambridge Analytica Data Sca...

    Regional Differences in Information Privacy Concerns After the Facebook-Cambridge Analytica Data Sca...

  • Defining a Research Testbed for Manned-Unmanned Teaming Research

    Defining a Research Testbed for Manned-Unmanned Teaming Research

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.