Hostname: page-component-89b8bd64d-n8gtw Total loading time: 0 Render date: 2026-05-09T09:28:21.743Z Has data issue: false hasContentIssue false

Expanding the machine: Notating generative synthesis with a state-based representation and a navigable timbre space

Published online by Cambridge University Press:  29 January 2026

Vincenzo Madaghiele*
Affiliation:
Department of Musicology, University of Oslo , Oslo, Norway
Leonard Lund
Affiliation:
KTH Royal Institute of Technology, Stockholm, Sweden
Derek Holzer
Affiliation:
Division of Media Technology and Interaction Design (MID), KTH Royal Institute of Technology, Stockholm, Sweden
Tejaswinee Kelkar
Affiliation:
Department of Musicology, University of Oslo, Oslo, Norway
Kıvanç Tatar
Affiliation:
Data Science and AI Division, Computer Science and Engineering Department, Chalmers University of Technology and University of Gothenburg, Gothenburg, Sweden
Andre Holzapfel
Affiliation:
Division of Media Technology and Interaction Design (MID), KTH Royal Institute of Technology, Stockholm, Sweden
*
Corresponding author: Vincenzo Madaghiele; Email: vincenzo.madaghiele@imv.uio.no
Rights & Permissions [Opens in a new window]

Abstract

Notating electroacoustic music can be challenging due to the uniqueness of the instruments employed. Electronic instruments can include generative components that can manipulate sound at different time levels, in which parameter variations can correlate non-linearly to changes in the instrument’s timbre. The way compositions for electronic instruments are notated depends on their interfaces and the parameter controls available to performers, which determine the state of their sound-generating system. In this article, we propose a notation system for generative synthesis based on a projection from its parameter space to a timbre space, allowing to organise synthesiser states based on their timbral characteristics. To investigate this approach, we introduce the Meta-Benjolin, a state-based notation system for chaotic sound synthesis employing a three-dimensional, navigable timbre space and a composition timeline. The Meta-Benjolin was developed as a control structure for the Benjolin, a chaotic synthesiser. Framing chaotic synthesis as a specific instance of generative synthesis, we discuss the advantages and drawbacks of the state- and timbre-based representation we designed based on the thematic analysis of an interview study with 19 musicians, who composed a piece using the Meta-Benjolin notational interface.

Information

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press
Figure 0

Figure 1. Notation for Benjolin pieces. Images by Pete Gomes, used with permission. This notation is based on the hardware Benjolin instrument, representing the eight knobs and the modular patch. The graphic signs around each knob portray different behaviours with which each knob governing synthesis parameters should be approached while performing.

Figure 1

Figure 2. A hardware Benjolin as a standalone synthesiser. Macumbista Instruments, 2025.

Figure 2

Figure 3. Diagram of signal flow and operations in the Benjolin synthesiser.

Figure 3

Figure 4. Data flow diagram of the Meta-Benjolin structure. The latent representation based on timbre learnt by the VAE is used in the screen-based interface to map three-dimensional coordinates to the corresponding synthesis parameters.

Figure 4

Figure 5. The Meta-Benjolin interface is composed of a three-dimensional navigable point cloud, a vertical composition timeline (left) and a menu (top-left).

Figure 5

Figure 6. Examples of the use of transitions to navigate long distances. D2 used a meander transition in the middle of the piece to connect two sections; within a section, neighbouring states are connected using crossfades. A5 used a crossfade and a meander transition to navigate between two neighbourhoods in the cloud, each corresponding to a section in their piece. (a) Composition by D2 (detail), Sound_example_4.m4a in the sound material. (b) Composition by A5 (detail), Sound_example_5.m4a in the sound material.

Figure 6

Figure 7. The space distribution of these two compositions gives information about how the sound evolves in time. While composer D3 created a gradual and constant evolution by navigating the whole point cloud using the meander transition, B3 was interested in exploring local variations and nuances. This difference can be seen by the fact that the viewpoint is zoomed far out in (a), while it is much closer to the cloud in (b). (a) Composition by D3 (detail), Sound_example_1.m4a in the sound material. (b) Composition by B3 (detail), Sound_example_2.m4a in the sound material.

Supplementary material: File

Madaghiele et al. supplementary material 1

Madaghiele et al. supplementary material
Download Madaghiele et al. supplementary material 1(File)
File 9.5 MB
Supplementary material: File

Madaghiele et al. supplementary material 2

Madaghiele et al. supplementary material
Download Madaghiele et al. supplementary material 2(File)
File 974.3 KB