• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
    • Tech Futures
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 7 (November 2025)
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Beyond Empirical Windowing: An Attention-Based Approach for Trust Prediction in Autonomous Vehicles

February 5, 2024

🔬 Research Summary by Zhaobo Zheng, a scientist at Honda Research Institute USA, Inc.

[Original paper by Minxue Niu, Zhaobo Zheng, Kumar Akash, and Teruhisa Misu]


Overview: The trust in autonomous driving is critical for user experience and system efficiency. This paper utilizes a selective windowing attention network to augment user trust in autonomous driving. The novel model can also analyze and visualize the more important scenarios related to user trust changes.


Introduction

SWAN: A New Way to Understand Human Trust in Autonomous Vehicles (AV)

Do you want to know how you feel about autonomous driving? Do you want to improve your human-machine interaction design? Do you want to leverage the power of attention mechanisms to analyze long time-series data?

If you answered yes to any of these questions, then have a look at SWAN: a Selective Windowing Attention Network. SWAN is a novel neural network model that can estimate human trust in AV from multimodal signals, such as speech, facial expressions, and physiological data.

Unlike traditional windowing techniques requiring manual tuning and domain knowledge, SWAN can automatically select the most relevant data intervals for trust prediction. SWAN uses window prompts and masked attention transformation to focus on the critical span of trust changes while ignoring the irrelevant or noisy parts.

SWAN has been tested on a new multimodal driving simulation dataset where it outperformed existing baselines, such as CNN-LSTM and Transformer, by a large margin. SWAN also showed robustness across different windowing ranges, demonstrating its flexibility and adaptability.

SWAN is the augmented solution for human state estimation. With SWAN, you can visualize the underlying nuances of human trust. 

Key Insights

Trust is an important factor that affects how humans interact with machines, especially in safety-critical domains like AVs. However, trust is a gradual state that changes over time, and it is difficult to label and analyze long time-series data that capture trust variations.

One common technique to deal with long time-series data is windowing, which divides the data into fixed-size, overlapping segments and applies a model to each segment. However, windowing has some drawbacks, such as:

  • The model’s performance depends on the window size, which requires manual tuning and domain knowledge.
  • The window size is fixed, which may not capture the dynamic nature of trust changes.
  • The windowing process may introduce noise or loss of information.

To overcome these limitations, the paper introduces a Selective Windowing Attention Network (SWAN), a neural network model that can automatically select the most relevant data segments for trust prediction using attention mechanisms.

SWAN consists of three main components:

  • A window prompt generator creates a set of window prompts representing different input data segments with varying lengths and positions.
  • A masked attention transformer computes the attention scores between the window prompts and the input data and selects the most informative segments based on the scores.
  • A trust predictor aggregates the selected segments and outputs a trust score for the whole input data.

The paper evaluates SWAN on a new multimodal driving simulation dataset, where participants interacted with an AV system and reported their trust levels. The dataset contains speech, facial, and physiological signals and contextual information such as driving scenarios and events.

The paper compares SWAN with several baselines, including:

  • A CNN-LSTM model that applies a convolutional neural network (CNN) and a long short-term memory network (LSTM) to the whole input data.
  • A Transformer model that applies a transformer network to the whole input data.
  • A windowing-based model that applies a CNN-LSTM model to each window segment, and uses an empirical method to select the optimal window size.

The paper shows that SWAN outperforms the baselines regarding trust prediction accuracy and demonstrates robustness across different windowing ranges. The paper also provides some qualitative analysis and visualization of the attention scores and the selected segments, which reveal some interesting insights into the trust dynamics and the factors that influence trust.

The paper demonstrates that SWAN is a novel and effective method for trust estimation in AVs. It can be extended to other applications that involve human state modeling from long time-series data.

Between the lines

This paper may pave the road for a universal model for cognitive state detection through multimodal signals. Such a model may also provide Interpretability on what triggered cognitive state changes.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

An abstract spiral of dark circles appears at the centre, resembling a tornado. Several vintage magazine covers and advertisements are being drawn toward the spiral. The artworks that have already been pulled into it are becoming distorted and replaced with clusters of numbers representing their numerical embeddings.

Tech Futures: Better Imagination for Better Tech Futures

This image is a collage with a colourful Japanese vintage landscape showing a mountain, hills, flowers and other plants and a small stream. There are 3 large black data servers placed in the bottom half of the image, with a cloud of black smoke emitting from them, partly obscuring the scenery.

Tech Futures: Crafting Participatory Tech Futures

A network diagram with lots of little emojis, organised in clusters.

Tech Futures: AI For and Against Knowledge

A brightly coloured illustration which can be viewed in any direction. It has many elements to it working together: men in suits around a table, someone in a data centre, big hands controlling the scenes and holding a phone, people in a production line. Motifs such as network diagrams and melting emojis are placed throughout the busy vignettes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part II

A rock embedded with intricate circuit board patterns, held delicately by pale hands drawn in a ghostly style. The contrast between the rough, metallic mineral and the sleek, artificial circuit board illustrates the relationship between raw natural resources and modern technological development. The hands evoke human involvement in the extraction and manufacturing processes.

Tech Futures: The Fossil Fuels Playbook for Big Tech: Part I

related posts

  • The Sociology of AI Ethics (Column Introduction)

    The Sociology of AI Ethics (Column Introduction)

  • The Nonexistent Moral Agency of Robots – A Lack of Intentionality and Free Will

    The Nonexistent Moral Agency of Robots – A Lack of Intentionality and Free Will

  • The Ethical Considerations of Self-Driving Cars

    The Ethical Considerations of Self-Driving Cars

  • Diagnosing Gender Bias In Image Recognition Systems (Research Summary)

    Diagnosing Gender Bias In Image Recognition Systems (Research Summary)

  • DC-Check: A Data-Centric AI checklist to guide the development of reliable machine learning systems

    DC-Check: A Data-Centric AI checklist to guide the development of reliable machine learning systems

  • Artificial Intelligence and Aesthetic Judgment

    Artificial Intelligence and Aesthetic Judgment

  • On the Generation of Unsafe Images and Hateful Memes From Text-To-Image Models

    On the Generation of Unsafe Images and Hateful Memes From Text-To-Image Models

  • Efficiency is Not Enough: A Critical Perspective of Environmentally Sustainable AI

    Efficiency is Not Enough: A Critical Perspective of Environmentally Sustainable AI

  • (Re)Politicizing Digital Well-Being: Beyond User Engagements

    (Re)Politicizing Digital Well-Being: Beyond User Engagements

  • Responsible Use of Technology: The IBM Case Study

    Responsible Use of Technology: The IBM Case Study

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.