• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar
  • Skip to footer
Montreal AI Ethics Institute

Montreal AI Ethics Institute

Democratizing AI ethics literacy

  • Articles
    • Public Policy
    • Privacy & Security
    • Human Rights
      • Ethics
      • JEDI (Justice, Equity, Diversity, Inclusion
    • Climate
    • Design
      • Emerging Technology
    • Application & Adoption
      • Health
      • Education
      • Government
        • Military
        • Public Works
      • Labour
    • Arts & Culture
      • Film & TV
      • Music
      • Pop Culture
      • Digital Art
  • Columns
    • AI Policy Corner
    • Recess
  • The AI Ethics Brief
  • AI Literacy
    • Research Summaries
    • AI Ethics Living Dictionary
    • Learning Community
  • The State of AI Ethics Report
    • Volume 6 (February 2022)
    • Volume 5 (July 2021)
    • Volume 4 (April 2021)
    • Volume 3 (Jan 2021)
    • Volume 2 (Oct 2020)
    • Volume 1 (June 2020)
  • About
    • Our Contributions Policy
    • Our Open Access Policy
    • Contact
    • Donate

Exploring XAI for the Arts: Explaining Latent Space in Generative Music

November 25, 2023

🔬 Research Summary by Nick Bryan-Kinns, Professor of Creative Computing at the Creative Computing Institute, University of the Arts London, where he researches the human-centred approaches to the use of AI in the Arts.

[Original paper by Nick Bryan-Kinns, Berker Banar, Corey Ford, Courtney N. Reed, Yixiao Zhang, Simon Colton, and Jack Armitage]


Overview: Deep learning AI can now generate high-quality and realistic outputs in various art forms, from visual arts to music. However, these AI models are incredibly complex and difficult to understand from a user perspective. This paper explores how explainable AI (XAI) could be used to make a generative AI music system more understandable and usable for users, especially musicians.


Introduction

Deep learning AI models can now produce realistic and high-quality artistic outputs such as music (e.g., Magenta, Jukebox) and images (e.g., Stable Diffusion, Dall-E). However, it is very difficult to understand how the output was produced or to control features of the production process. In this paper, we survey 87 generative AI research papers and find that most do not explain what the AI model is doing and hardly any real-time control of the generative process. To address this lack of transparency and control, we explore applying eXplainable AI (XAI) techniques to design a generative AI system for music. We introduce semantic dimensions to a variational autoencoder’s (VAE) latent space, which can be manipulated in real-time to generate music with various musical properties. We suggest that in doing so, we bridge the gap between the latent space and the generated musical outcomes in a meaningful way, making the model and its outputs more understandable and controllable by musicians.

Key Insights

Explainable AI and the Arts

The field of eXplainable AI (XAI) examines how machine learning models can be made more understandable to people. For example, XAI projects have focused on creating human-understandable explanations of why an AI system made a particular medical diagnosis, how the AI models in an autonomous vehicle work, and what data an AI system uses to generate insights about consumer behavior. However, current XAI research is predominantly focused on functional and task-oriented domains, such as financial modeling, so it is difficult to apply XAI techniques to artistic uses of AI directly. Moreover, in the Arts, there is typically no “right answer” or correct set of outputs that we are trying to train the AI to arrive at. In the Arts, we are often interested in surprising, delightful, or confusing outcomes, as these can spark creative engagement.

Our research explores how XAI techniques can be applied in and through the Arts to improve the use and understanding of AI in creative practice. The Arts, especially music, also provide a complex domain to test and research new AI models and approaches to explainability. Compared to domains such as healthcare and automotive industries, the arts require similar levels of robustness and reliability from their AI models but have significantly fewer ethical and life-critical implications, making the Arts a great test-bed for AI innovation.

Explainable AI and Generative Music

In our paper, we surveyed 87 recent AI music papers regarding the role of the AI in co-creation, what interaction is possible with the AI, and how much information about the AI’s state is made available to users. Our perspective is that the explainability of creative AI is a combination of the AI’s role, the interaction it offers, and the grounding that can be established with the AI. We found a small number of excellent examples of collaborative AI, interactive AI music systems, and AI models that make internal AI state available to users. However, the vast majority of AI music systems offer very little of any of these explainability features.

Explaining Latent Spaces in AI Music Generation

To explore how to make generative AI models more explainable, we build on the MeasureVAE generative model, which produces short music phrases. In our version, we force 4 of the 256 latent dimensions of the AI model to map to musical features. We then visualize these features in 2D maps in a web app that users can navigate in real-time to generate music. 

In this way, we increase the explainability of the generative AI model in two ways: i) key parts of the AI model are exposed to the user in the interface and meaningfully labeled (in this case, with relevant musical features), and ii) the real-time interaction and feedback in the user interface allows people to explore the effects of these features on the generative music and thereby implicitly learn how the model works.

In our implementation, the AI acts somewhat like a colleague – the response to the user is given in real-time, as would be done in a human-to-human musical interaction. This drives a feedback loop between a user and the AI, whereby a person’s reaction to the AI’s response informs the subsequent interaction. Thinking musically, this resembles a duet in creative improvisation, where the players make real-time decisions based on their colleague’s performance.

Between the lines

Generative AI is a source of concern and excitement in the Arts. On one hand, AI offers new tools, opportunities, and sources of inspiration for creative practice and exploration. On the other hand, there are ethical concerns about the lack of attribution and IP recognition in AI training sets, concerns about the deskilling of creative work, and concerns about bias in generative AI. We can proactively work to ensure that the artist remains key to the creative process through eXplainable AI and the design of user interfaces that embrace real-time interaction with the AI model. Indeed, working with artists to design and implement eXplainable AI systems will help mitigate concerns about the impact of AI on creativity.

Want quick summaries of the latest research & reporting in AI ethics delivered to your inbox? Subscribe to the AI Ethics Brief. We publish bi-weekly.

Primary Sidebar

🔍 SEARCH

Spotlight

ALL IN Conference 2025: Four Key Takeaways from Montreal

Beyond Dependency: The Hidden Risk of Social Comparison in Chatbot Companionship

AI Policy Corner: Restriction vs. Regulation: Comparing State Approaches to AI Mental Health Legislation

Beyond Consultation: Building Inclusive AI Governance for Canada’s Democratic Future

AI Policy Corner: U.S. Executive Order on Advancing AI Education for American Youth

related posts

  • Exploiting Large Language Models (LLMs) through Deception Techniques and Persuasion Principles

    Exploiting Large Language Models (LLMs) through Deception Techniques and Persuasion Principles

  • A Look at the American Data Privacy and Protection Act

    A Look at the American Data Privacy and Protection Act

  • Quantifying the Carbon Emissions of Machine Learning

    Quantifying the Carbon Emissions of Machine Learning

  • On the Perception of Difficulty: Differences between Humans and AI

    On the Perception of Difficulty: Differences between Humans and AI

  • HAI Weekly Seminar Series: Decolonizing AI with Sabelo Mhlambi

    HAI Weekly Seminar Series: Decolonizing AI with Sabelo Mhlambi

  • A Snapshot of the Frontiers of Fairness in Machine Learning (Research Summary)

    A Snapshot of the Frontiers of Fairness in Machine Learning (Research Summary)

  • Warning Signs: The Future of Privacy and Security in an Age of Machine Learning  (Research summary)

    Warning Signs: The Future of Privacy and Security in an Age of Machine Learning (Research summary)

  • Low-Resource Languages Jailbreak GPT-4

    Low-Resource Languages Jailbreak GPT-4

  • Ethics of AI in Education: Towards a Community-wide Framework

    Ethics of AI in Education: Towards a Community-wide Framework

  • Scientists' Perspectives on the Potential for Generative AI in their Fields

    Scientists' Perspectives on the Potential for Generative AI in their Fields

Partners

  •  
    U.S. Artificial Intelligence Safety Institute Consortium (AISIC) at NIST

  • Partnership on AI

  • The LF AI & Data Foundation

  • The AI Alliance

Footer


Articles

Columns

AI Literacy

The State of AI Ethics Report


 

About Us


Founded in 2018, the Montreal AI Ethics Institute (MAIEI) is an international non-profit organization equipping citizens concerned about artificial intelligence and its impact on society to take action.

Contact

Donate


  • © 2025 MONTREAL AI ETHICS INSTITUTE.
  • This work is licensed under a Creative Commons Attribution 4.0 International License.
  • Learn more about our open access policy here.
  • Creative Commons License

    Save hours of work and stay on top of Responsible AI research and reporting with our bi-weekly email newsletter.