Back to Research
2025

Live Improvisation with Fine-Tuned Generative AI

A Musical Metacreation Approach

Publication
Peer reviewed paper in
Proceedings of the International Conference on New Interfaces for Musical Expression (NIME)
NIME 2025, Canberra, Australia

Research Overview

Complete improvisation pipeline from text prompts to audio generation in Ableton Live

Improvisation Pipeline

Complete workflow integrating text prompts, AI model inference, and Ableton Live for dynamic musical co-creation.

Fine-tuning dataset visualization showing audio samples and spectrograms

Fine-Tuning Dataset

Custom dataset of compositions and field recordings used to fine-tune Stable Audio Open for personalized sound generation.

Inference time analysis and performance metrics during live improvisation

Inference Performance

Real-time performance analysis showing latency and generation times critical for live improvisation workflows.

Abstract

This paper presents a pipeline to integrate a fine-tuned open-source text-to-audio latent diffusion model into a workflow with Ableton Live for the improvisation of contemporary electronic music. The system generates audio fragments based on text prompts provided in real time by the performer, enabling dynamic interaction. Guided by Musical Metacreation as a framework, this case study reframes generative AI as a co-creative agent rather than a mere style imitator. By fine-tuning Stable Audio Open on a dataset of the first author's compositions and field recordings, this approach demonstrates the ethical and practical benefits of open-source solutions. Beyond showcasing the model's creative potential, this study highlights the model's significant challenges and the need for democratized tools with real-world applications.

Audio Examples

Improvisation 1

Improvisation 2

Improvisation 3

Future Directions

Our case study shows GenAI can extend musical creativity beyond mere imitation. Rapid, responsive sound generation enables AI to catalyze new ideas and open fresh improvisational pathways. Yet despite this potential, high computational and data demands may hinder broader adoption. Future work will include evaluations with audiences and artists to assess artistic impact and guide improvements. Additionally, fine- tuning more effectively with smaller datasets can reduce data needs and expand accessibility. Developing optimized pipelines that run on local hardware with limited resources will also be crucial. As this technology evolves and opens new possibilities for Musical Metacreation, our research aims to lay the groundwork for deeper human-AI collaboration in live improvisation.

This website uses cookies for functionality and analytics (only with your consent). I do not save or sell your personal data. External links may track you through their own services.