Professional Documents
Culture Documents
OM Composer's Book 3 PDF
OM Composer's Book 3 PDF
BOOK
Volume 3
Collection Musique/Sciences
directed by Jean-Michel Bardez & Moreno Andreatta
Published works
Volume 3
Edited by
Jean Bresson, Carlos Agon, Gérard Assayag
Preface by
Roger Dannenberg
Collection Musique/Sciences
Musique/Sciences Editorial Board
Carlos Agon, UMR STMS Ircam/CNRS/UPMC, Paris
Gérard Assayag, UMR STMS Ircam/CNRS/UPMC, Paris
Marc Chemillier, Ecole des Hautes Etudes en Sciences Sociales, Paris
Ian Cross, University of Cambridge
Philippe Depalle, McGill University, Montréal
Xavier Hascher, University of Strasbourg
Alain Poirier, Conservatoire National Supérieur de Musique et de Danse, Paris
Miller Puckette, University of California, San Diego
Hugues Vinet, UMR STMS Ircam/CNRS/UPMC, Paris
Communication
Marine Nicodeau
Editions
Claire Marquet
Cover design
BelleVille
Volume editor
Jean Bresson
Editorial coordination and support
Aaron Einbond
Tous droits de traduction, d’adaptation et de reproduction par tous procédés réservés pour tous pays.
Le code de la propriété intellectuelle du 1er juillet 1992 n’autorise, aux termes de l’article L. 122-5, 2e
et 3e a), d’une part, « que les copies ou reproductions strictement réservées à l’usage du copiste et non
destinées à une utilisation collective » et, d’autre part, « que les analyses et les courtes citations dans
un but d’exemple et d’illustration ». « Toute représentation ou reproduction intégrale ou partielle, faite
sans le consentement de l’auteur ou ayant cause, est illicite » (article L.122-4). Cette représentation
ou reproduction par quelque procédé que ce soit constituerait donc une contrefao̧n sanctionnée par les
articles L. 335-2 et suivants du Code de la propriété intellectuelle.
ISBN 978-2-7521-0283-6
c 2016 by Editions DELATOUR FRANCE/Ircam-Centre Pompidou
www.editions-delatour.com
www.ircam.fr
Contents
Preface
Roger Dannenberg vii
Introduction
Jean Bresson, Carlos Agon, Gérard Assayag 1
Germination
Jean-Luc Hervé and Serge Lemouton 219
Koch’s Space
Julián Ávila 245
Appendix
OpenMusic 363
Computer music tools and technology 373
About the authors 377
vi
Preface
vii
time with rich formal symbol systems, just waiting for the right tools to enable further
exploration. This is illustrated by the work of Hiller and Isaacson, who were already
creating the first music composition programs at a time when each new computer was
considered a historic development. In this sense, CAC might be considered a traditional
practice of composition and music theory that has simply made slow progress until
recently due to its dependence on symbolic computation and ultimately automation
enabled by modern computers.
Sixty years after Hiller and Isaacson, computers and CAC have not replaced the
composer, even in situations where one might think music is highly constrained and
subject to formalisation. Ironically, CAC seems to be flourishing in experimental music
where things are the least constrained and where there is no over-arching formalism, at
least none that is known and applicable to music composition and production. Perhaps
this is the real strength and contribution of CAC in today’s musical world: CAC offers
composers a new way to construct and explore formalisms that lead to musical structure,
form, and organisation.
It is not that we do not already have such formalisms. Rather, a trend beginning in
the 20th Century has been a sort of meta-composition in which composers first create
new formalisms, new rules, new objectives, new questions, and then work within this new
framework to create new music. Here is where CAC really shines. If new formalisms can
be expressed computationally, then the ramifications of the formalisms can be explored
quickly, and either the formalism can be quickly rejected or it can be refined according
to the musical tastes of the composer. This approach allows composers both to work
faster and to extend their production to forms and sounds that might be unimaginable
through a more direct path of putting pen to paper.
If composition with computers has become meta-composition, do we have a cor-
responding meta-theory? Can we formalise the practice of CAC? There are general
studies of creativity and a few studies of computer music composition practice, but no
specific theories akin to traditional harmony or serialism. We are left at a stage of
development where the best one can do to understand the field is to study individual
cases. While the literature describes many “toolkits” and software techniques, and many
composers mention using algorithmic techniques, it is rare to find in-depth descriptions
of the intentions, decisions, and implementation details of the compositional process.
The chapters of this book, however, offer a wealth of practical and detailed case
studies of CAC in practice. The variety of concepts, strategies, and techniques that
fill these pages illustrates that CAC is not a musical genre or a musical theory. It is
more of a meta-theory, an approach to the construction, exploration, and evaluation of
musical theories and their application to composition. The role of the computer and
formalisms ranges from the most abstract, where the computer generates material to be
freely incorporated into composition, to the most specific, where the computer might
generate the final composition as digital audio. By considering all of these descriptions
as case studies, we can begin to form an overall impression of CAC practice.
This book has the common theme of OpenMusic software, which is both a visual
programming language and a variety of libraries supporting different kinds of musical
computation including selection, audio analysis, constraint satisfaction, genetic algo-
rithms, interpolation, and synthesis techniques. Thus, the chapters serve in some way
as a commentary on OpenMusic itself and the roles that programming languages and
computer music systems can play in assisting composition. One feature of OpenMusic
viii
that seems to be critical to its success is the ability to translate between the worlds of
signals and symbols. Many composers use signal analysis to obtain pitch and rhythmic
material, but the resulting data can be visualised in OpenMusic using fairly conventional
notation. This ability to see intermediate results as “music” seems to play an important
role in the exploration of musical ideas. Another feature of OpenMusic is the availability
of ready-made representations for melodies, chord sequences, and time-varying functions.
These high-level abstractions seem to facilitate the rapid construction of programs to
generate and manipulate music data in meaningful ways. Finally, OpenMusic is able to
render music as audio as well as notation from these musical representations, allowing
composers conveniently to hear as well as see the results of their models and choices.
In conclusion, computer-assisted composition is a recent phenomenon with historical
roots that can be traced back to the earliest theories and representations of music. Now
that computers are ubiquitous and enable composers easily to explore generative theories
of music, we can expect CAC to take an increasingly important role in common musical
practice. The chapters in this book provide a wealth of material on how composers think
about CAC as well as details of their individualised pursuit of new musical structures.
For both composers and non-composers, this collection is a wonderful introduction to an
emerging practice that challenges and expands our understanding of composition.
Roger Dannenberg
ix
Introduction
Ten years after publication of the first OM Composer’s Book, we are happy to present
this new volume comprising 21 contributions of contemporary music composers that
describe original works and experiments carried out using the OpenMusic computer-
aided composition environment.
OpenMusic (OM) is a visual programming language dedicated to musical modelling
and composition. It allows composers to program musical processes combining the
expressive power of a programming language with symbolic and graphical representa-
tions.1 Amongst a variety of existing works on domain-specific programming languages
for music, computer-aided composition was formalised as a specific field of research in
the early nineties.2 Today this field represents an original approach to music composition
using computer processes and programming: in contrast with most computer music en-
vironments that focus on performance and real-time interaction, it promotes an “offline”
paradigm to aid composition of musical structures. A program in OpenMusic typically
generates musical sequences, scores, sounds or any kind of musical material involved
in a compositional process. For that purpose, it brings forward advanced temporal
modelling and integrates rich structures and formalisms in music programming. The
works presented in this book provide evidence of this and illustrate these characteristics
of computer-aided compositional approaches.
Contributions to The OM Composer’s Book often hint at a future where distinctions
between composing and programming using computer-aided composition software be-
come obsolete. Indeed, computer-aided composition environments are not meant to let
computers create music, but to provide musicians with means to have the computer do
what they want; that is, to become programmers themselves.
It is interesting to note, however, that most of the authors in this book underline
a certain emancipation from the strict computational approach which can emanate
from excessive use and trust in computer systems, emphasising the interplay between
algorithmic and “intuitive” processes involved in composition. OpenMusic in some way
facilitates this orientation in compositional processes by merging programming, musical
data editing, and visualisation.
Computer-aided composition evolved along with the developments in technology and
musical æsthetics of recent decades. As the reader will notice in several chapters of
this book, there is a clear tendency to connect compositional processes in OpenMusic
1
The OM Composer’s Book. 3
with external tools and software—in particular in the field of digital sound processing.3
We have decided to organise the book from this perspective. The first chapters deal
with the traditional “symbolic” core of the computer-aided composition framework, and
later move into “mixed” approaches that utilise more and more sound generation and
processing. Of course many alternative paths could also be taken to read through the
chapters of the book.
We were glad to receive so many contribution for this volume, and are grateful to
all the authors for their participation. We would also like to thank the editorial team
of the Musique/Sciences collection for their support, and in particular Aaron Einbond
whose labours bring this book to its present form. Contributions to the OM Composer’s
Book 3 came from France, the United Kingdom, Italy, Spain, Germany, the United States,
Canada, Japan, Iran, Portugal, and Brazil. The cultural, geographic, and æsthetic
diversity of topics described in this book remind us that computer-aided composition
is a lively and continually widening domain, offering exciting perspectives for musicians
and computer music researchers around the world.
3A glossary of technology and software mentioned in different chapters is also provided in the Appendix
of the book.
2
Variazioni su AlDo ClEmenti
Michele Zaccagnini
Variazioni su AlDo ClEmenti (2013) is a piece for chamber orchestra in four movements
(“Invenzione”, “Sinfonia”, “Preludio” and “Corale”). It calls for 11 instruments: flute
(doubling piccolo), clarinet (doubling bass clarinet), glockenspiel, piano, 3 violins, 2
violas, cello, and double bass. The piece is based on a four-note set {A, D, C, E} (taken
from Aldo Clementi’s “playable” letters) and is written entirely in OpenMusic. The
processes implemented in the OpenMusic patches are mainly concerned with rhythmic
treatments, while pitch manipulations are fairly simple. The four movements express a
unified æsthetic: the exploration of repetitiveness and of the psychoacoustic boundaries
between predictability and chaos. In this paper, I start by describing the general æsthetic
that informs my compositions; the second section aims at unveiling the algorithms that
produced the first two movements of Variazioni: “Invenzione” and “Sinfonia”; the third
and final section includes conclusions and a description of current implementations of
the algorithms.
A Clementian æsthetic
The misunderstandings are generated by those who think, if subconsciously, of
music as “discourse” inadvertently making it a caricature of an arch that describes
a useless orgasm. Exaltation and depression are closed chapters: no matter how
disguised they might be, they are modest symbols of an extinct dialectic. A forte
followed by a piano, a high note followed by a low one, a “sweet” timbre followed
by a “raw” one are in themselves sonata-like cells of the greater Sonata Form.1
Aldo Clementi (author’s translation)
1 “Gli equivoci nascono solo da quanti, anche inconsciamente, pensano la Musica come discorso e
quindi, non accorgendosene, come caricature di un arco che descrive un inutile orgasmo. Esaltazione
e depressione sono capitoli chiusi: comunque camuffati, sono modesti simboli di una dialettica già
estinta. Un forte seguito da un piano, una nota acuta seguita da una grave, un timbro dolce da uno
aspro: ciò è di per sé dialettica, cellula sonatistica di una più grande Forma-Sonata.” [1], p. 48.
3
Michele Zaccagnini
In 1971, Aldo Clementi composed B.A.C.H., a short piece for solo piano. The one-
page score is supposed to be repeated at least three times; it is a sort of “musical
mechanism” which neither develops nor varies, yet it seems to possess an inner force
that propels it forward in a seeming perpetual motion. B.A.C.H. was a turning point
from which Clementi began deliberately to erase any sort of contrast that could initiate
a dialectic of tension and release. Instead, his compositions became a research of musical
stasis. Radical as this might sound, Clementi realised how avoiding any form of contrasts
was the only antidote to a sort of “narrative bias” that affects our perception of music.2
Clementi’s æsthetic philosophy is accompanied by a rich eschatological explanation that
eventually declares music a doomed art form, not too differently but possibly more
negatively than his early mentor Theodor Adorno.
My encounter with the music and æsthetic thought of Clementi has been an important
moment in my compositional journey. As idiosyncratic as his æsthetic approach might
seem, I believe Clementi’s lesson carries tremendous positive implications. His determina-
tion to abandon compositional escamotage in order to privilege the exploration of frozen
textures has been a fundamental inspiration that has pushed me to pursue my current
compositional path. Like Clementi, I am deeply interested in studying and experimenting
with the construction of static textures—textures that derive their raison d’être solely
from their constructional principles. I explore the perceptual boundary within static
textures that divides regular and irregular, predicable and chaotic. Where does one
begin and the other end? Surely, the two are not mutually exclusive, but how much
“space” divides them? The attention to textural detail rather than to the larger musical
discourse is the very aspect that links Clementi, my work, and OpenMusic.
I can describe my approach to composition as a sort of experimental process articu-
lated in four steps:
1. æsthetic premise
2. procedural hypothesis
3. design of algorithm
4. observation/assessment
In what follows I will describe steps 2 and 3 in detail. Step 4 is left for the listener.
Computer-aided composition has been a crucial addition to my compositional process
since my modus operandi heavily relies on the ability to assess speedily the outcomes
of experiments given certain hypotheses. In general, computer-aided composition allows
a composer to parse a large number of results, thereby allowing for a greater number
of hypotheses to be tested and evaluated. Variazioni su AlDo ClEmenti was written
following this experimental approach: each of the movements is a “chosen” realisation of
a hypothesis aimed at exploring the simple-complex boundary.
2 Toillustrate this point, one might think how even a single contrasting element in an otherwise uniform
texture is enough to trigger an expectation.
4
Variazioni su AlDo ClEmenti
Algorithmic procedures
Invenzione
“Invenzione”, the first movement of Variazioni, is a canon in which each instrument
plays a fixed set of four notes: the first four instruments enter by playing the set at its
original transposition A, D, C, E in four different octaves (A4, A6, A3, A5); successive
entrances introduce different transpositions that gradually undermine the original set’s
diatonic area. As instruments enter the canon, the sonic texture gradually builds up
to 11 voices; as the movement reaches its midpoint, instruments start dropping out of
the canon causing the texture to gradually thin out. Different from a traditional canon,
generally thought of as a polyphonic structure, “Invenzione” is anything but: as the
texture builds up, the different voices of the canon start blurring together producing an
increasingly amorphous texture. Voices blur with one another as the traditional elements
accounted for by a polyphonic construction do not appear and are deliberately avoided:
instruments play the same pitch-sets, with the same dynamic (always p) and with no
specific timbral treatment (e.g. ordinario); as a result, the simple diatonic fragments
played by different instruments gradually start blending together, ultimately producing
in a monophonic drone.
The minimal construction of the movement in terms of pitch, timbre, and dynamics
is counterbalanced by a fairly complex principle that informs the rhythmic placement
of each note. To determine the rhythmic placement of the four-note melodic fragments
I designed what I later named the “bouncing algorithm”. This algorithm “mimics” the
motion of an object in an enclosed space. The object moves at a given speed along
one (horizontal) dimension in a space of given width. The direction of the movement is
alternatively left-to-right and right-to-left: the object moves left-to-right until its right
edge touches the right boundary of the space, at which point it starts moving right-to-left;
conversely, the direction of its movement will revert to left-to-right whenever its left edge
touches the left boundary. Therefore, the bouncing algorithm calculates the positions
of the object at different times by tracing its movements.3 For example, in a space of
size 10, an object of size 2 and speed 6 will start moving left-to-right from position 0 at
time 0, to position 6 at time 1, then changing direction halfway through the next step
(when the right edge touches the outside wall) and landing on position 4 at time 2, then
changing direction again to land on position 2 at time 3, etc. The graphical realisation
of this example can be seen in Figure 1.
5
Michele Zaccagnini
The bouncing algorithm can be used to determine the position of different notes
within a time interval. To move from a space domain to a time domain, we can simply
“translate” the dimension of the space into a time interval and the size of the object into
the duration of a sound event.4 If, in the example from Figure 1, I assign = 1 then the
space of size 10 can be thought of as a measure of 5/4 while the object size of 2 results
in a note duration of ♩ (one quarter note). Therefore the bouncing motion of Figure 1
will look something like the score of Figure 2.
The bouncing algorithm is, in other words, a way to generate patterns. These patterns
have some interesting characteristics:
While these patterns are not necessarily easy to predict, the algorithm designed to
generate them allowed me to gain control over several perceptual variables of the textures
they can produce. First, the algorithm creates a continuous rhythmic reshuffling of
notes. The relative placement of objects in a given space will be highly variable at
different times, given that the objects’ sizes and speeds are different from one another.
Second, by dynamically tweaking some parameters, the algorithm can be implemented to
produce varying degrees of complexity in the sound texture. It is fairly easy to illustrate
this point by visualising an enclosed space, such as the one of Figure 1, and imagine
the effect of introducing more bouncing objects in the same space: the system surely
gains in complexity. On the other hand, if I expand the dimension of the space while
keeping the number of objects constant, the resulting system becomes clearer and more
predictable, i.e. objects will “spread out”. In Invenzione I decided to use these features
of the bouncing algorithm to create a sort of “perceptual paradox”: the two parameters
most responsible for increasing and decreasing the system’s complexity, space dimension
and number of objects, play one against the other. The increase of the number of voices
is counterbalanced by the increase in the dimension of the time interval in which they
move iteration after iteration. Conversely, as voices begin to exit, the size of the space
decreases compressing the rhythmic distribution of the objects.
4 The speed has no direct musical correlate: it simply determines the “shifts” of the note within the time
interval at each iteration.
5 InA New Kind of Science [2], in a section titled “System of Limited Size and Class 2 Behavior” (p. 255),
Stephen Wolfram shows a case of pattern behaviour very similar to the “bouncing algorithm”. Wolfram
observes how a dot moving in a limited space will produce highly variable patterns depending on its
speed and the size of the space. These examples, which do not include the “bouncing” feature nor do
they contemplate objects of different sizes, present a remarkable variability in terms of their periods.
6
Variazioni su AlDo ClEmenti
The formula to obtain the position of the object given the set of three variables is
obtained by first calculating the position Po of the object in “open space”:
Po = s × t
where s is the “speed” of the object and t is the iteration number. The position Pf of
the object moving forward is given by:
Pf = Po mod (L − d)
where L is the time-space interval and d is the duration-size of the object. The position
Pb while moving backwards is given by:
Pb = L − d − Pf .
θ is the variable that determines which direction the object is travelling (0 if moving
forward, 1 if moving backwards):
Po
θ= mod 2.
(L − D)
The position Px of the object within each successive time-unit is then:
Px = (1 − θ)Pf + θPb .
The patch in Figure 3 implements this bouncing formula, and reports the object
position given the same set of parameters: the speed (s), the time interval (L), the
duration (d), and the number of iterations (t). In this patch, the arithm-ser function
outputs a sequence of ascending numbers corresponding the iteration index (in the
example from Figure 1: t0 , t1 , t2 ...)
Since the algorithm outputs values that fall between 0 and L − d, each successive
value will need to be displaced in the score by a value equal to time interval times the
iteration number (L × t).6 For example, for a note at logical time t3 the value output by
the algorithm of Figure 3 needs to be added to L × 3.7 Figure 4 shows the top-level loop
generating the voices of the canon according to a set of input lists corresponding to the
speed, space dimension, duration and number of iterations. The internal bouncing loop
calls the main algorithm (Figure 3), and inner-offs iteratively adds L × t (the “inner-
offsets”) to each output value in order to shift it to its proper time interval.
If the inner offsets are added within each one of 11 the voices, there is another kind
of offset that needs to be added to the voices: the canonic, or outer, offsets. To create
the “perceptual paradox” mentioned earlier, a new voice enters whenever the size of the
time interval increases end exits whenever it decreases. To calculate the canonic offset
then, we need to observe first how L behaves over the course of the movement: L grows
arithmetically from 1 to 11 and then decreases in a palindromic fashion. Furthermore,
each interval is repeated a number of times equal to its size (L). It follows that after
7
Michele Zaccagnini
the first voice begins the next one will be delayed by 1,8 the second by 4 (L2 = 2),
the third by 9. Interestingly, while the time interval grows arithmetically, the canon
entrances follow a geometric series. These secondary offset are calculated outside of
the loop from Figure 4 and added to the raw and inner-offset values inside by means
of a simple om+ function. The right outlet of the bouncing omloop outputs a list of
durations to be paired to each of the voice’s onset lists. Eventually, the onset lists and
their corresponding durations are used to build a list of chord-seq objects. After all
the rhythmic values are ready, they are matched with the pitch cells as described at
the beginning of this section, starting from octave transpositions of the pitch-set at its
original level and gradually, as each of the canonic voices enters, transposing the pitch-
set. Each chord-seq produced is then funnelled into a multi-seq object (see Figure 5) and
eventually exported as a MIDI file.
8 The L values, as well as the offsets, onsets, and durations, are later converted to millisecond values by
means of a multiplier (250ms). This multiplier makes the smallest possible subdivision a sixteenth note
at a tempo of ♩ = 60.
8
Variazioni su AlDo ClEmenti
Figure 4. Patch generating the canonic voices. Note the inner and outer offsets added to the
main bouncing patch result.
9
Michele Zaccagnini
Sinfonia
“Sinfonia” is also based on the diatonic cell {A, D, C, E}. This second movement is
similar to the one described in the preceding sections, in that it also aims to “muddy
the waters” of its transparent diatonic cell. But if “Invenzione” blurs the material by a
gradual process of layering, rhythmic reshuffling (“bouncing algorithm”), and “centrifu-
gal”, i.e. increasingly non-diatonic, transpositions, the second movement is even more
subtle and mysterious in its inner workings and perceptual result. Throughout the entire
movement the four notes of the set sound without interruption and always in the same
register; no other transposition interferes with the diatonic cell. Subtly, the timbre of
each note changes by means of re-articulation by different instruments and intervening
microtonal inflections.
The score explicitly states how between each note and its re-articulation, there should
be no gap. Figure 6 shows how each note-end corresponds to an attack in another
instrument with the same, if slightly inflected, pitch. This connection is made clear to the
conductor by a dotted line. As a result, the entire movement produces an uninterrupted
hue of sound. Some processes subtly work under the surface to produce slight variations
in the texture, though they are hardly clear to our perception. Far from being indicative
of any developing narrative—the lack of which constitutes the “negative” premise of the
piece—these varying traits of the texture only contribute to its general sense of fragility.
The slight variations of hue come from three distinct processes. First, the re-articula-
tion of the four notes follows a specific order. Second, each re-articulation spells one of
four different microtonal inflections. Third, the onset of each re-articulation is based on
10
Variazioni su AlDo ClEmenti
Figure 6. Beginning of “Sinfonia” (dotted lines indicate the connections of the same pitch-set
elements if microtonally inflected).
a “hidden” polyrhythm. Interestingly, all three of these processes take advantage of the
same input: a matrix extracted by the OpenMusic function permutations with the set
(1 2 3 4). This function outputs all possible re-orderings of the set: in the case of a four-
element set, 4! = 24 permutations. The 24 × 4 matrix is transformed into a 6 × 16 matrix
by removing the output’s first level of parentheses and using the list-modulo function
with a modulus of 6.9 The result is a matrix of six rows:
1 2 3 4 2 1 3 4 3 1 2 4 4 1 2 3
1 2 4 3 2 1 4 3 3 1 4 2 4 1 3 2
1 3 2 4 2 3 1 4 3 2 1 4 4 2 1 3
1 3 4 2 2 3 4 1 3 2 4 1 4 2 3 1
1 4 2 3 2 4 1 3 3 4 1 2 4 3 1 2
1 4 3 2 2 4 3 1 3 4 2 1 4 3 2 1
9 list-modulo groups the elements of a list spaced at a regular interval and returns these groups as a list
of lists.
11
Michele Zaccagnini
Each of the six rows contains four instances of each element (i.e. four ones, four twos
etc.). In Figure 7 we can see how the patterns-›pitches omloop parses these lists. Four
global variables (a, c, d, e) are initiated with a list of four values, each corresponding to
a microtonal inflection of the variable’s original pitch: natural (no inflection), a quarter
tone up, a quarter tone down, and an eighth tone up or down (chosen at random). A
permut-random function contained in the micro-shift sub-patch shuffles the order of the
inflections within each variable at instantiation. Each time a variable is called a different
inflection is selected, creating microtonal oscillations around the centre pitch. Therefore,
this loop is responsible for both determining the order of notes being played, by selecting
one of the four variables depending on the order of the input list’s four elements, and the
rotation of microtonal inflections, by parsing the variable’s content.10
Figure 7. Picking the right note and inflection based on permutation lists.
10 Whenever an index “finds” its match, it parses the list stored as a global variable by means of a Lisp
nth function (in combination with loop counters). This escamotage is necessary as OpenMusic does
not provide “destructive” list operators such as the pop function. However, a simple line of code within
a Lisp object would do the job as well.
12
Variazioni su AlDo ClEmenti
Conclusions
This chapter describes one step in a path on which I embarked several years ago and that I
am still pursuing. From an æsthetic perspective, my music aims to explore different kinds
of repetitive textures propelled by perceptually ambiguous inner workings—to induce a
semi-hypnotic state of consciousness in the listener.
As I have shown in this chapter, my attempts at inducing a specific perceptual
response in the listener are not based on abstract æsthetic intuitions but rather on build-
ing tools that could, if tested and implemented properly, correctly create a connection
between the composition and its perceptual outcome. In other words, the central task of
my work is to find and harness those compositional variables responsible for effectively
creating a link between the composition and the way music is perceived.
If this goal seems too abstract and/or ambitious to be successfully carried out yet, I
have come across a few expedients that have proven effective at easing the burden. First
of all, I have looked at patterns and phenomena that exist outside of the musical context.
The bouncing algorithm was designed starting from the simple intuition that to create a
musical environment that gradually oscillates between more or less chaotic states, I could
simply try to emulate an enclosed space with bouncing objects and “extract” some of its
13
Michele Zaccagnini
Figure 8. Left: Calculating onsets based on previous onsets and rhythmic values. Right:
Computation of the next available onset using Euclidean division. If the remainder of the
division is zero then the left value of the division is simply multiplied by the next assigned
value; if the division has a nonzero remainder the “ceiling” of the division is taken.
salient features. The sort of dynamic equilibriums that environments of the sort possess
can, as the movement “Invenzione” proves, be translated into music.
Second, relying on computer-aided composition to develop my algorithm has been a
crucial addition that has sped up calculations that would otherwise be tedious and time-
consuming. A computer-aided composition environment such as OpenMusic is an ideal
space to implement these experiments since it integrates powerful tools for iteration such
as omloop and the mapcar function with score and graphic tools. It is helpful to have
different way of displaying the results of my endeavours not only by using proportional
rhythmic notation objects (chord-seq and multi-seq) but also graphical ones (bpf, etc.).
References
[1] Michela Mollia. Autobiografia della musica contemporanea. Cosenza: Lerici, 1979.
[2] Stephen Wolfram. A New Kind of Science. Champaign, IL: Wolfram Media, 2002.
Acknowledgements: I would like to thank William C. Hoston, Yu-Hui Chang, Joshua Fineberg, and
Jean Bresson for their invaluable help in developing my ideas.
14
Folk material transformations and
elaborations in A Vida é Nossa
Gonçalo Gato
In this chapter I will discuss the compositional procedures associated with the use of
OpenMusic to compose the piece A Vida é Nossa (2013) for symphonic wind band.1
The main compositional work was based on folk material, which I transcribed and/or
analysed. After proposing a terminology for some archetypical procedures of computer-
aided composition, various configurations of “manual” and “machine” composition are
explored along with æsthetic reflection. Active human decision making plays a key
role, and as a result the computer-aided compositional process features much the same
æsthetic issues as purely manual composition. This opens up the scope of algorithmic
redesign and/or manual procedures, as the composer remains focused on his primary
musical goal.
Terminology
Computer-aided composition (CAC) places human-machine interaction at its core and
beautifully show-cases the interface between art and technology. After composing this
way for some years, I found it useful to define certain terms so as to clarify the general
procedures composers carry out while composing with algorithms:
1 Composed
as part of doctoral practice-based research carried out at the Guildhall School of Music and
Drama, London.
15
Gonçalo Gato
A Vida é Nossa
Gonçalo Gato
Austero, q = 120
° 4 3 2 4 2 4
∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑
Manual implementation: using the output ofmanual an algorithm as a reference so that it is
Piccolo &4 4 4 4 4 4
complementation
manually incorporated into the∑ score. This4 ˙means J that4 it is up
- œ- œ ‰ œ. 2 ‰ œ. Œ
to theœ ‰ composer
4 œ. ‰ Œ
to
4 ∑ ∑ ∑ 3 2 4 œ™ . Œ
&4 Flauta 1 4 ∑ ∑ 4 4 ∑
f
write
& 4 the∑ notes in
4 Flauta 2 ∑ the score,
∑ generally
∑ 3
4 ∑ implying
∑ 2
4 b˙
4 œmanual
4 Jbœ ™ œ œ ‰ 4 bœ completion.
- - 2 . ‰ ‰ œ. 4 Œ ‰
4 bœ. Ó ∑
f
Automatic
4
&4 implementation:
∑Oboé 1∑ ∑ copying
∑ 4 an
3 ∑ algorithm
∑ 2˙
4
4 Ϫ output
4 J ‰ 42 ‰ into
- œ- œ- œ. œ.
Œ 4 the
4 ‰ Œ score
œ. œ.
‰ Œ using ∑ auto-
mated
4
procedures.
∑ ∑
This
∑
usually
∑
implies
3 ∑
a
∑
MIDI
2 b˙
or
- - MusicXML
. .
4 œbœ ™ œ œ ‰ 2 bœ ‰ ‰ œ 4 Œ ‰
file export.
.
bœ Ó ∑
f
Oboé 2 &4 4 4 4 J 4 4
f
In Figure 1 the first page of the score of A Vida é Nossa 44is shown
Corne Inglês
4
&4 ∑ ∑ ∑
along∑ with
∑ ∑ 3
4 ∑ ∑ 2
4 ∑ 4
4 ∑ 2
4 ∑
annotations,
4 providing
∑
Requinta em Mib
&4 ∑ an example
∑ of
∑ how ∑ these
A43Vida ∑ 4procedures
2 ∑ 4
é Nossa 4 ∑ can
2
4 ∑ shape
4
4 various
∑ materials.
∑
4 33 22 4 22 44
Trompete 2 em Sib
Saxofone Tenor & 44
& ∑∑ ∑∑ ∑∑ ∑∑ 44 ∑∑ ∑∑ 44 ∑∑ 44 ∑∑ 44 ∑∑ 4 ∑∑ ∑∑
4 33 22 4 22 44
Trompete 3 em Sib
Saxofone Barítono & 44
? ∑∑ ∑∑ ∑∑ ∑∑ 44 ∑∑ ∑∑ 44 ∑∑ 44 ∑∑ 44 ∑∑ 4 ∑∑ ∑∑
con sord.
bœ œ œ 2 ˙ œ
Trombone Tenor 1
? 44
? ∑∑ ∑∑ ∑∑ ∑∑ 33 ∑∑ ∑∑ 22 bœ œbœ 44 œbœ œ ∑J J 24 44 Œ Ó ∑∑
Fagote 1 4 44 44 ∑ 4 4 ∑ 4 ∑
mp austero f
con sord.
bœ œ œ 2 ˙
? 44
? ∑∑ ∑∑ ∑∑ ∑∑ 33 ∑∑ ∑∑ 22 44 œbœ ∑J J 424 44 œ Œ Ó ∑∑
Trombone Tenor 2
Fagote 2
¢ 4 44 44 #œ œ∑ œ 4 œ ∑ 4 ∑
manual intervention
œ œ œ b+
˙ elaboration
mp austero f
?4 ∑ ∑ ∑ ∑ 3 ∑ 2 œ œ œ 4 œ œ œ J 2.+ J 2 +
∑ œ 4 Œ ‰bœ ™ œ™ ‰ Ó
Trombone Baixo ° 44 43
∑ ∑ ∑ ∑ ∑ 24 ∑ 44 Ó 24 44 œ
Trompas 1,2 em Fá &4 4 4 mp ∑austero
4 ‰ 4˙ 4 Œ
f
Ó mf ∑
œ™ b˙ œ
Tuba
?4 ∑ ∑ ∑ ∑ 3 ∑ ∑ 2 4 jp j 2 mf 4 Œ ‰ ‰ Ó
¢ 4 4 4 4 œ 3.œ+ œ 4 b˙+ 4œ bœ™ œ™
4 3 ∑ 2 œ œ∑ œ 4 Óœ œ œ ‰ 2 4
algorithm output 4 mp
Trompas 3,4 em Fá &4 ∑ ∑ ∑ ∑ 4 ∑ 4 Œ Óf ∑
austero bœ ™ 4 ˙ 4œ f
l.v.
°? 4
∑ ∑ ∑ ∑ 3 2 4 ∑ p 2b ˙ ∑ 4œ ∑ ∑
Tímpanos
¢ 44 4 bœ ™ œ ™ bœ ™ œ ™ 4 4 4 mf 4
3 ∑ 2 ˙ 4 2 4
∑ ∑ ∑ ∑ ∑ 4 f —∑ 4 ∑ ∑ ∑ ∑
manual complementation
Trompete 1 em Sib &4 4p fff 4 4
Vibrafone
l.v. sempre
° 4¡
A
∑bw 3 2 ∑ 44
3
2 4
Trompete 2 em Sib
Percussão 1 &4 w ‘∑ ‘∑ ˙ b˙ Œ∑ b œ œ œ 4 Œ ∑Œ ˙ ˙ Œ b∑ œ œœ 4 ∑ j 4 ˙ ∑ 4˙ ∑ ∑
œœ œb œ‰ ‰ œ œ œ œ
ppp bœ bœ bb ww
ww œ ˙˙˙ ˙˙ ˙˙ bb œœœœ œœœœ œœœœ mp
mf mp f
Marimba
4 ∑ 3 ∑ ∑ ∑ 3 2 4 2 4
Trompete 3 em Sib &4 ¡ 4 ∑ ∑ 4 ∑ 4 ∑ 4 ∑ 4 ∑ ∑
4 A 3
Œ 3 2 4 2 4
Percussão 2 &4 j 4 ˙™ b˙ ™ ˙™ b˙ ™ 4 ∑ 4 ∑ 4 ∑ 4w bw ‘
bbb œœœœ bbb œœœœ ˙˙˙˙ œœœœ ˙˙˙˙ b œ b œœœ ™™™ ˙˙˙ œ b œbb œœ œœ ˙˙ ™™ ppp
con sord.
b œ œ œ
Trombone Tenor 1
?4 p
4 ∑ ∑ ∑ ∑ 3
4 ∑ ∑ 2 4 b œ
4 bœ œbœ 4 œ œ J J 42 ˙ 4œ Œ
4 Ó ∑
Wood Blocks mp austero f
4 Ó Œ Œ‰ j 3 ‰ Ó™ 2con sord. 4 bœ œ Œ œ 242 ˙ ™ j 4 j
Percussão 3 / 44 œ
? œ ∑œ œ œœœ œ œ œœ ∑ œœ ∑œ œ œ œ œ∑ œ 443 œ œ ∑œ œ 24 Œ 44 œ 444 œ œœœŒ Óœ ™
∑ 4 #œ œœ œ 4 œœ œbœ Jœ œ J 4 œ œ œ œ œœœ ∑œœœœœœœ
automatic implementation
Trombone Tenor 2 4 p
mp austero Sinos Tubulares f
Prato Suspenso
Para o Tantã Tantã Para o Glock.
arco
Percussão
Trombone Baixo4
?4 w ∑ w— ∑ ∑∑ ∑∑ 43 ∑∑ ∑∑ 24 œ˙—œ œ 44 œ œ œ ∑œJ œ œJ& 24 b˙œ 4œ Œ ∑ ‰bœ ™ œ™ ∑‰ Ó
¢ / 4 mp p
— #œ — 4
mp austero f f mf
Tuba
?4 ∑ ∑ ∑ ∑ 3 ∑ ∑ 2 4 j j2 4 Œ ‰ ‰ Ó
¢ 4 4 4
œœœ
4
œœœ
œ œ œ 4 b˙ 4œ bœ™ œ™
Figure 1. First page of the score (in C). output
algorithm Dashed boxes indicate manual procedures. mp austero
l.v.
f f
°? 4 ∑ ∑ ∑ ∑ 3 2 4 ∑ 2 ∑ 4 ∑ ∑
Tímpanos
¢ 4 4 bœ ™ œ ™ bœ ™ œ ™ 4 4 4 4
˙—
manual complementation
p fff f
Vibrafone
l.v. sempre
° 4¡
A
3 2 4 3
2 4
Percussão 1 &4 w bw ‘ ‘ ˙ b˙ Œ b œ œ œ 4 Œ Œ ˙ ˙ Œ b œ œœ 4 ∑ 4 ww j 4˙ 4˙ œœ œb œ‰ ‰ œ œ œ œ
˙˙˙ bb œœœœ œœœœ œœœœ mp
16
bœ bœ bb ww œ ˙˙ ˙˙
ppp
mf mp f
Marimba
4 A¡ 3 3
Œ 3 2 ∑ 4 ∑ 2 ∑ 4
Percussão 2 &4 œ œ ˙ œ ˙ j œ ™ ˙ œ 4 ˙™ b˙ ™ ˙™ b˙ ™ 4 4 4 4w bw ‘
bbb œœœ bbb œœœ ˙˙˙ œœœ ˙˙˙ b œ b œœ ™™ ˙˙ b œbb œœ œœ ˙˙ ™™ ppp
p
Wood Blocks
4 j 3 ™ 2 4 2 j 4 j
œ œ œ œœœ œ œ œœ Ó
Œ œœ œ
Percussão 3 /4 œ œ œ œ Œ ‰œ œ 4 œ œ œ œ‰ Ó 4 Œ œ 4 œ œ œ œ œŒ 4 œ™ œ 4 œ œœœ œ œ™ œ œ œœœ œœœœœœœ
p automatic implementation
Prato Suspenso Sinos Tubulares
Para o Tantã Tantã Para o Glock.
arco
4 w w— 3 2 4 2 œ 4
Percussão 4
¢ / 4 mp ∑ ∑ 4 ∑ ∑ 4 p˙— 4 ∑ &4 — #œ — 4 ∑ ∑
f
Folk material transformations and elaborations in A Vida é Nossa
Folk material
In their account of Claude Debussy’s (1862-1918) approach to folk material, François
Lesure and Roy Howat write:
Debussy was always consistent on the point that a folk or national music should not
be used for its themes but rather in the manner of Albéniz: “Without using actual
popular tunes he is the kind of person who has them in his blood. They have become
so natural a part of his music that one barely distinguishes a demarcation line”. To
a Hungarian friend, he wrote: “Your young musicians could usefully take inspiration
from them, not by copying them but by trying to transpose their freedom, their
gifts of evocation, colour, rhythm. [...] One should only ever use the folk music of
one’s country as a basis, never as a technique”. [7]
We can say Debussy prescribed a progressive approach for dealing with folk materials.
Béla Bartók (1881-1945) is a another composer particularly relevant to consider in this
context. Malcolm Gillies writes:
Indeed, after this early Violin Concerto none of his works escapes a strong folk
influence. In his later lecture The Relation between Contemporary Hungarian Art
Music and Folk Music (1941, in Béla Bartók Essays, 348-53), Bartók exemplified
three types of arrangement: where the folk melody is mounted like a jewel, where
melody and accompaniment are almost equal in importance, and where the folk
melody is a kind of inspirational motto to be creatively developed. [6]
Bartók was important in my own music education, and while composing A Vida é Nossa
the same æsthetic issues brought me closer to his techniques.
It is interesting to recall some of Bartók’s harmonic methods. Referring to the
Fourteen Bagatelles, Op. 6 (1908, two of which directly quote folksongs), Gillies explains
that “any sense of functional harmony is persistently undermined by the use of ostinato
figures (nos. 2, 3, 5, 10, 13), quasi-bitonal writing (nos. 1, 13), streams of parallel 5ths and
7ths (no. 4), of 4ths (no. 11), of tritones (no. 8), or of piled-up 3rds (nos. 7, 9, 10)” [6]. As
I read these accounts, I began to see ways of manipulating the folk music I was selecting,
and the techniques I developed resonate with Bartók’s approaches.
Frequently the strategy I follow is to have folk tunes as part of some texture that is
based on them, but also have other features that turn them into a new and autonomous
sonic entity (thus “transposing their freedom, their gifts of evocation, colour, rhythm [...]
as a basis, never as a technique”, as Debussy prescribed). For instance, the simple idea
of constructing heterophony could be enough to achieve this effect. This gave rise to
the “rhythm-focusing heterophony” and “proliferating canonical heterophony” textures,
on which I will go into detail further below. Other times, the folk monodic material
was subject to different kinds of transformations, either conservative (colouring) or more
destructive (resynthesis).
17
Gonçalo Gato
In much the same way, so can a set of rhythmic cells define a characteristic of a rhythm.
A process of polymerisation can then be created by concatenating cells (monomers) in
any given degree of random ordering, the results of which are similar to improvisations by
folk performers. Such processes can be carried out both in the rhythm or pitch domains.
The initial woodblock part of the score (Figure 1), for instance, is generated from a
random combination of rhythmic cells from the percussion instruments (adufes) played
by the Adufeiras de Monsanto. Figure 3 shows the OpenMusic patch implementing
such random rhythmic cell concatenation.
In the electroacoustic realm, cross-synthesis implies the combination of analysis data
from two sounds [4]. By analogy, one can term cross-resynthesis the process by which the
initial low brass melody of the piece (Figure 4) was obtained: intervals were combined
18
Folk material transformations and elaborations in A Vida é Nossa
from the previous interval set (Table 1)—altered to include ascending and descending
versions of each interval—with percussive rhythmic cells from another folk source: a Viola
da Terra performance,4 the cells of which are featured in Figure 3. The combination
process is shown in Figure 5.
4 DécioLeal toca Viola da terra, http://vimeo.com/39351351. Video directed by Tiago Pereira, 2012.
Viola da Terra is a local string instrument from Açores.
19
Gonçalo Gato
Figure 5. Folk cross-resynthesis. Rhythmic material from a folk source is combined with
intervallic material from another. The resulting melody was subsequently transposed two octaves
down and its durations were doubled.
One of the beautiful and elegant aspects of algorithmic composition of this sort lies
in the fact that it enables the composer to shift his or her attention from algorithmic
design to the aural selection of materials, provided that he or she is sure the algorithm
reproduces the desired compositional technique. Drawing an analogy to the common
20
Folk material transformations and elaborations in A Vida é Nossa
use of the piano to try out compositional ideas, algorithmic tools constitute a kind of
intelligent piano; a musical laboratory. This significant and underlying aspect is one of
the main reasons why I use OpenMusic.
Rhythm-focusing heterophony
Starting with a folk melody, I wanted to build a texture that would gradually bring into
sync various parts by progressive rhythmic diminution. The voices would start “unfo-
cused” by virtue of the different magnitudes of augmentation. They would then gradually
diminish their durations until all voices became synchronised, creating a homorhythmic
texture. The process is depicted in Figure 7.
most augmented voice
complete sync
homorhythm
neutral voice
diminution of durations
The algorithm I used is shown in Figure 8. A folk melody called Chamarrita Preta,
performed by the Grupo Folclórico da Casa do Povo de Bandeiras,5 is initially analysed
for its sequence of relative durations (using tree2ratio). It is then multiplied by 16
because a 16th note, or semiquaver, is the smallest duration present in the melody. This
makes all durations integers, with a minimum value of 1. The sequence of durations
is reversed, doubled, and fed into an iterative augmentation process of different mag-
nitudes: the leftmost mapcar function calls the interpolation lambda function, which
interpolates the initial sequence with five increasingly magnified versions. The focusing
process is thus calculated in retrograde form: the initial sequence of durations, perfectly
superimposed on the different voices at the beginning, is subject to augmentations of
different magnitudes, gradually bringing the rhythms out of sync (as if reading Figure 7
backwards). Once these augmentations are calculated, the voices are retrograded again
(retrogade_rhythms abstraction) so that the result becomes a gradual diminution of
the durations, culminating in absolute homo-rhythm with the original melody (unison
doubling). The end result of the algorithm is the poly object shown at the lower right.
21
Gonçalo Gato
Figure 8. Algorithm for melodic focusing. The bpf in the centre shows the durational contours
of melodies. The bpf-lib in the lower left shows the progressive augmentation of the durations of
each voice. The patch window on the right shows the interpolation abstraction. Note that only
the notes of the bottommost voice of the poly, the most augmented voice, are visible because
the remaining voices were subject to smaller augmentation magnitudes and so start later in the
score.
The durations were doubled in order to allow the minimum duration to be subdivided
into integer values. As the augmentation process uses only integers, it eliminates the
possibility of tuplets (which would create quantisation and playability issues). The
decision to use only integers is both operational, to prevent quantisation problems,
and aesthetic, not to allow tuplets. I define operational decisions as those taken for
purely technical reasons, while æsthetic decisions are those taken according to principles
underlying artistic creation. Hence, algorithm design is here directed by a musical goal,
which in turn determines what constraints to apply. While programming, I continuously
tested the musical results by using the playback feature so that I could modify the visual
code and/or the initial parameters to achieve a better sense of gradually coming into
sync or focus. In this way algorithms can be used to fine-tune a compositional technique.
22
Folk material transformations and elaborations in A Vida é Nossa
sync
œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ #œ œ œ œ œ
& #œ œ œ œ
œ œ
& <#>œ ™ œ œ œ œ œ ™ œ œ™ œ œ œ œ œ #œ œ œ œ Œ œ œ #œ œ œ œ œ
& œ
œ œ œ œ Œ ‰ œj œ œ™ #œ œ œ œ œ™ œ œ œ™ œ œ˙ œ œ™ œ œ œ™ #œ œ œ
sync
#œ œ œ œ
& œ œ œ œ œ œ Œ œ œ #œ œ œ œ œ œ œ œ œ #œ œ œ œ
& œ œ œ œ #œ œ œ œ œ œ Œ œ œ #œ œ œ œ œ œ œ œ œ #œ œ œ œ
complete sync
& œ
œ Œ œ œ #œ œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ
œ œ œ
& œ œ Œ œ œ #œ œ œ œ œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ
& œ
œ Œ œ œ #œ œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ
œ œ œ
œ ‰ ™ ‰ œj œ
& œ œ œ ™ œ œ™ œ œ #œ œ œ œ œ Œ
R #œ œ™ œ œ œ œ œ œ œ œ œ
œ ‰ ™ ‰ œj œ
& œ œ œ ™ œ œ™ œ œ #œ œ œ œ œ Œ
R #œ œ™ œ œ œ œ œ œ œ œ œ
œ ‰ ™ ‰ œj œ
& œ œ œ ™ œ œ™ œ œ #œ œ œ œ œ Œ
R #œ œ™ œ œ œ œ œ œ œ œ œ
œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ Œ Ó
& #œ œ œ œ #œ
œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ Œ Ó
& #œ œ œ œ #œ
& #œ œ œ œ œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ Œ Ó
#œ
œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ Œ Ó
& #œ œ œ œ #œ
& #œ œ œ œ œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ Œ Ó
#œ
& #œ œ œ œ œ œ œ œ œ #œ œ œ œ œ œ Œ œ œ Œ Ó
#œ
Figure 9. Last bars of the passage showing the synchronisation points of the voices.
23
Gonçalo Gato
Once I was satisfied with the result, I exported it as a MusicXML file and imported it
into notation software (see Figure 9). I then selected the part of the sequence beginning
when all voices are present, and ending with all voices playing the same melody at the
same time (the point of complete sync). The texture was scored for 2 oboes, 2 clarinets,
and 2 alto saxophones. As I conceived it, the instruments needed to be woodwinds— 7
I wanted
° a soft
88
∑
and stable ∑
dynamic ∑ Œ
and some timbral
. heterogeneity—and have compatible
5 6
bœ. œ#œ#œ.
œ. 3 3 3
Picc. #œ œ#œ#œ œbœbœnœœœ‰ œ ‰ #œ œ ‰ #œ. ‰ ‰ Ó ∑ Ó ‰
registers. In Figure 10, one can see œ the gradual. . synchronisation of all six parts in the
&
. . 3 ff 3
ff
-œ œ œ ™ œ œ œ™ œœ -œ œ ‰™ œ-
Ob. 2 & œ ˙ œ œ™ œ œ œ™ #œ œ ‰ œ œ- #œ œ™ œ œ œ- œœ ™ œœ™ œ ˙ œ œ #œ œ œ
œ #œ . .
Cor. Ing. & œ œ™ œ bœ œ ™ œ œ œ œœ Œ œ œ œ œ #œ Ó Œ œ. nœ œ. œ œ. #œ
œ œ™ œ #œ œ #œ œ
œ ™ #œ
f œ
f fff
3 3 3 3 3 3 nœ bœ bœ b œ b œ nœ bœ bœ 3
3 3
3 3 3 œœ
Req. Mib & œ b˙ ˙ nœ bœ œ œ œ Œ Œ œ bœ nœ
nœ #œ œ nœ <n>œ Œ nœ#œ œ œ nœ nœ#œnœ #œ. œ. œ. #œ. nœ. œ. œ. œ œ œœ œbœnœ
nœ œ ˙ ˙ nœ
3
f
Cl. 1 & œ ˙ œ œ™ œ œ œ™ #œ œ
-œ œ œ ™ œ œ œ™ œœ œ- œ ‰™ ‰ œ œ- #œ œ™ œ œ œ- œœ ™ œœ™ œ ˙ œ œ #œ -œ œ œ
#œ œ- œ œ œ -œ #œ œ-
Cl. 2 & #œ œ œ- œ œ ˙ œ œ Œ œ œ- #œ œ œ- œ œ ˙ œ œ œ œ
3 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
j . .œ . . œ. . . œ. . œ œ œ Ÿ
Cl. 3 & œ œ™ œ™
œ œ bœ ™ œ œ œ œ œ™ ‰ ∑ ∑ #œ. œ. œ. œ#œ . œ. œ. œ œ œ. œ#œ œ œ œ œ œ ˙ #œ
f 3 3
mf ff fff
œ œ œ™ bœ œbœ™ œ bœ bœ œ bœ œbœ™ ™ bœ œ #œ
? bœ ™ œ™ J œ™ bœ œ œ™ œ™ œœ bœ ™ bœ nœ Œ œ #œ #œ nœ œ
Cl. B. J ‰ #œ ™ œ œ
J fff
f f
#œ -œ œ œ œ -œ #œ -œ
Sax. Alto 1 & #œ œ œ- œ œ ˙ œ œ Œ œ œ- #œ œ œ- œ œ ˙ œ œ œ œ
˙ # œ# œ œ #œ #œ œ # œ.
Figure
? 10. Focusing in˙ the context of the score (inœ C). Shaded‰ ‰parts are not involved œ inœ the
œ œ b˙ ˙ bœ nœ ˙
œ #œ #œ bœ œ ˙
nœ #œ #œ œ #œ #œ œ œ œ. . . œ œ bœ œ nœ #œ œ .™ . . . . . . .
Fg. 1 ≈ œ #œ
3 3 3 3 3 3 3 3
process. The dashed line indicates the point from which the texture becomes homo-rhythmic.
3
f fff
œ b˙ ˙ œ œ ˙ b˙ œ œ œ. œ. œ. . . . . . œ. œ
Fg. 2
¢
? Œ Ó ∑ ∑ ? ‰ ‰ ‰ #œ #œ ‰ œ #œ œ ≈ œ œ œ œ œbœ
3 3 3 3 3 3 3
3
f fff
Virtual fundamentals
1.
° ∑ ∑ ∑ Ó Œ ‰ Œ
Tpa. 1,2 & w #ww
# œœ ˙˙ # œœ.
fp f mp f fff
in the
? shaded
Tbn. 1 ∑ parts ∑ of Figure
∑ 10).
Ó How
Œ can œ ˙one create
œ.
‰ Œ a harmony
∑ Œ that
#œ ™ is different
>œ œ
J
from, but related to, the melody’s pitch collection, and that enables some freedom and mf-pp f pp f
? ∑ ∑ ∑ Ó Œ #œ ˙ #œ. ‰ Œ ∑ Œ Œ #œ œ >œ œ
Tbn. 2
mf-pp f 3 pp f
6 See the first section of this chapter for a definition of musical complementation.
Tbn. B.
? ∑ ∑ ∑ ∑ ∑ ∑ Ó Œ ˙
¢
3 pp f
24 œ bw Para o Xil.
° Aœ
¡
Mar. & ∑ ∑ ∑ ∑ ∑ Ó œ œ w
f
Prato Susp. ¢/ ∑ ∑ ∑ ∑ ∑ ∑ w7
pp ff
Folk material transformations and elaborations in A Vida é Nossa
greater dissonance? How could that harmony be compatible with diatonicism? I chose to
explore techniques from spectral music to obtain two solutions, which were individually
programmed and tested in OpenMusic:
1. Find the virtual fundamental of the diatonic pitch collection. Complementary parts
play harmonic partials of that fundamental.
2. Consider the diatonic melody’s pitches as the same partials of changing funda-
mentals to be calculated (for instance, suppose all pitches are seventh partials
of different fundamentals). Then, obtain other partials of those fundamentals by
transposing the pitch collection by an interval corresponding to an integer-ratio in
frequency.
Techniques related to virtual fundamentals have been around for some time. Tim
Howell discusses their use in Magnus Lindberg’s Kinetics (1988-1989):
The acoustic properties of [these serially derived] chords are [then] analysed in
relation to the overtone series, as group of partials issued from one fundamental.
That spectral analysis allows all chromatic constituents to be viewed in relation to
an underlying fundamental, which may be deployed in their re-harmonisation; this
can transform initial dissonance into sonorous chords. [2], p. 239.
Figure 11. Algorithm to find the virtual fundamental of a pitch collection and calculate the
first n harmonic partials.
I was after a bell-like sonority that would transfigure the folk melody’s diatonic
harmonic field. This perceptual shift was rehearsed and tried-out using the playback
features of both notation software and algorithmic software (OpenMusic). The various
calculated harmonic series were tested by playing back their superimpositions on the folk-
melody-based heterophony. From the eight possibilities shown in Figure 11, the second
25
#œ - - #œ -
mp dolce
"Chamarrita Preta" (trad. Portugal)
∑ ∑ ∑ ∑ ∑ #œ œœ œ œ œ™ œ œ ™ œ- œ œ‰Œ ≈ œ™ œ™ œ œ
Sax. Alto 2
¢& œ œ™ - œ
- #œ œ œ œ œ œ™ œ œ œ
mp dolce
Glockenspiel
l.v. sempre
° ˙ #˙ ˙ œ #˙ ™ œ
W.B. / ∑ ∑ ∑ ∑ ∑ & ˙ ˙ #˙ ˙ œ™ J
p
Gonçalo
& Gato
Pand.
∑ ∑ ∑ ∑ Œ
Triângulo
˙—™ Para a Mar.
∑ ∑ ∑ ∑ ∑
p
Para o Estalo
Bombo / w— ∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑
Prato Susp. ¢/ ∑ ∑ w7 w7 w7 l.v.
∑ ∑ ∑ ∑ ∑
ppp f
chord of the quarter-tone-based set was selected. I then arranged some of the calculated
partials
= in the brass section, making use of the easy accessibility of quarter-tones on the
trombones (Figure 12).
36 D
° -œ œ œ ™ œ œ œ™ œ œ -œ œ ‰ ™ ‰ œ œ œ™ - ™ -
Ob. 1 & ˙ œ œ™ œ œ œ™ #œ œ - #œ - œœ
œ œ™ œ™ œ œ œ™ œ œ™ œ œ œ œ#œ ™ œ œ œ™ œ œ œ œœ œ œ
fundir-se com os metais
œ- - -
Ob. 2 &
œ ‰ Œ œ™ œ- œ#œ œ œ™ œ- œ œ œ œ œœœ œ œ œ™ œ™ œ œ œ#œ ™ œ™ œ œ œ œ™ œ™ œ œ œ œ™ œ™ œ œ œ ‰™ ‰™ œ œ œ- œ #œ ™ œ™ œ œ œ- œ œ ™
fundir-se com os metais
#œ œ -œ œ œ œ -œ œ
Cl. 1 & #œ - œ
œ œ œ œ™ œœ œ ˙ œ œ œœœ œ œ œœ Œ ‰ œ œ œ™
- #œ - œ
œ œ œ œ™ œ œœ ˙
fundir-se com os metais
- ™ - ™ -œ
Cl. 2 & ‰ ™ œ œ œ- #œ œ œ œ- œ œ œ™ œ œ œ œ™ œ œ œ œ#œ ™ œ #œ œ™ œ œ œ œ œ œ œ œ ≈Œ œ œ- #œ œ œ œ- œ œ ™ œ œ œ™ œ˙ œ œ œ™ œ #œ œ™
fundir-se com os metais
-œ -œ
Sax. Alto 2
¢& œ œ™ œ œ œœ œ œ#œ ™ œ œ™ œ œ œœœ œ œ™ œ œ™ ≈ Œ ‰ œ œ œ™
- #œ œ™ œ œ œ- œ œ ™ œ ™ œ œ œ œ™ ˙ œ œ™ œ™ œ œ œ#œ ™
fundir-se com os metais
3.
° 3
Tpa. 3,4 & ∑ ∑ ∑ Ó Œ ‰‰
#œ w w w w
pppp mf
con sord.
Tpt. 1 & ∑ ∑ ∑ ∑ ∑ Œ ˙ ˙ w w
3
pp
senza sord.
Tpt. 2 & ∑ ∑ ∑ ∑ ∑ w w w
pp mp
senza sord.
Tbn. 1
? ∑ ∑ ∑ ∑ Ó ‰ µœ œ ˙ œ ˙ w w
3 3
pppp mp
senza sord.
? ∑ ∑ ∑ ∑
Tbn. 2 Bw w w w
pppp mf
senza sord.
3
Tbn. B.
? ∑ ∑ ∑ ∑ Œ Œ
¢ Bœ ˙ w w w
pppp mf
° ˙ œ #œ ˙ œ œ™ œ˙ œ™ j j j #˙ ˙ œ œ™
Glock.
¢& J œ™ J
œ ˙ œ™ #œ ˙ œ™ œ ˙ ˙ J
To sum up, the same technical procedure used by Lindberg in Kinetics is used in
To sum up, the same technical procedure used by Lindberg is used in A Vida é
A Vida é Nossa but to carry out the opposite perceptual transformation accounted for
Nossa, but to carry out the opposite perceptual goal. As the folk tune’s pitches begin to
by superimposed
be Howell. As the onfolk
the tune's
added pitches
partials,begin
their to be superimposed
aural on the added
perception is changed: they partials,
become
theiraaural
just part perception
of a larger issonority.
changed:Furthermore,
they becomeasjust part
they doofnot
a bigger
conform sonority.
preciselyFurther-
to the
more, as series,
overtone they dotheynot render
conform theprecisely to the overtone
whole sonority series, they The
slightly inharmonic. render the iswhole
result that
the initial diatonic sonority (arguably very consonant) is changed into a new globally
sonority slightly inharmonic. The result is that the initial diatonic sonority (arguably
inharmonic sound entity. An analogy with a chemical reaction is almost irresistible:
very
A + Bconsonant)
→ C, where is changed into heterophony,
A is the folk a new globally B isinharmonic sound entity.
the added partial collectionAn and
analogy
C is
with a chemical
a new sonic entity. reaction is almost irresistible: A + B → C, where A is the folk heter-
ophony, B isnow
Turning theto
added partialsolution
the second collection and C isabove,
mentioned a new Isonic entity.a simple algorithm
developed
Turning now to solution 2 mentioned above, I developed a simple
to carry out and test the harmonic superimposition of the original algorithm
pitch to
collection
(of the folk melody) with a transposition given by an integer frequency ratio. The
carry out (and test) the harmonic superimposition of the original pitch collection (of
algorithm is shown in Figure 13. The results were approximated to semitones so as to
the folk
allow melody)
easier with aintransposition
playability the transposed given by an
parts. Theinteger-ratio
output was in frequency.sequence,
a harmonic The al-
gorithm
with eachischord
shownserving
in Figure
as a 13. The
pitch approximation
reservoir from which usedthewas the unfolds.
music semitone This so aswasto
allow easier
carried out byplayability
hand (see in the intervening
Figure parts. The
10 above, shaded output
parts) and was
can bea harmonic
seen as ansequence,
example
of mixed
with eachcomplementation:
chord serving as aitpitch involvevs manual
reservoir fromimplementation
which the music of unfolded.
algorithmic results
This was
(the harmonic progression serving as pitch reservoirs) carried out so as to coexist with
carried out by hand (see Figure
automatically-implemented 10 above,
algorithmic shaded
outputs (theparts) and can be heterophony).
rhythm-focusing said to create an
example of mixed complementation, involving manual implementation of algorithmic
results (the harmonic progression serving26 as pitch reservoirs) carried out so as to co-
exist with automatically-implemented algorithm outputs (the rhythm-focusing heter-
ophony).
Folk material transformations and elaborations in A Vida é Nossa
Figure 13. Algorithm for the calculation of pitch reservoirs related by integer frequency ratios.
Proliferating melodies
Some time before working on A Vida é Nossa, I programmed an algorithm that would
generate a series of melodies by gradually developing an initial one. Each melody would
randomly retain a melodic fragment of the previous one and add notes at the end based
on its universe of intervals and durations (therefore stimulating cognition by means of
auditory memory). The number of notes to add each time was controlled by the Fibonacci
sequence. I developed and refined the algorithm, fed in a folk melody, and decided
to display the results from subsequent cumulative evaluations7 in a poly object, which
superimposes the several voices obtained (see Figures 14 and 15). This was initially
meant only to organise the results in a logical manner, but I could not resist playing it
back!
7 Usinga feature of omloop called “accumulation”, meaning that a process can be applied several times
by acting on the previous result each time.
27
Gonçalo Gato
Figure 14. The algorithm for the proliferation of a melody. See Figure 15 for the contents of
proliferate_mel_fibo.
Finding I was pleased with the aural results, I made the æsthetic decision to turn
this “accident” into one of the core techniques of the piece. It created a kind of het-
erophonic/canonic texture: imitations were created from the same melodic fragments on
different voices at different times. This also leads to heterophony if the onsets are close
enough.8 The superimposition creates harmony from the folk melody’s pitch set but also
chromaticises it because of the newly calculated pitches. In this sense, this procedure can
be classified as an algorithmic elaboration: it cumulatively proliferates melodic segments,
intervals, and durations into an n-part texture.
After several evaluations of this indeterministic algorithm, one solution was selected
or “cherry-picked” (Figure 16). One important characteristic, which was very noticeable
aurally, is that voice activity occurs by waves: increases and decreases of the local density
of short durations, separated by moments of relative relaxation through longer durations
in the form of tied notes. It also features a light and cantabile quality in the melodies
that occur from bar 4 onwards. I was struck by the ability of this algorithm to create
such melodic lines.
8 György Ligeti’s Lontano (1967) is a good example, as heterophony therein is also created by canons.
28
Folk material transformations and elaborations in A Vida é Nossa
Different outputs from the melodic proliferation algorithm were used at various points
in the score. The technique thus created one of the main defining syntaxes of the piece.
The orchestrations that followed automatic implementation were carried out manually
(see Figure 17 for an example), along with some changes to improve playability and
musical clarity.
Algorithmic processes frequently result in elements that could be considered unmusi-
cal (or undesirable, for various reasons). In the case of this piece, I have had the chance
to hear it performed several times and could subsequently revise it to improve dynamic
balances among the voices and eliminate certain segments that sounded superfluous. For
me, these manual interventions play a key role in CAC, and show how I, as a composer,
am unwilling to sacrifice musicality even if that means carrying out “destructive” (but
musically constructive) actions on algorithmically generated materials.
29
Gonçalo Gato
Figure 16. Selected output from the patch in Figure 14. Rhythmic notation and accidentals
are shown in raw, unprocessed form. The brackets are used as visual aids to show retained
segments, which begin the next voice.
Important CAC-related reflections arose during and after the compositional process.
If algorithm design can be considered creative work and an integral part of composing—
as it naturally is—then it can allow space for much discovery to take place, without
being reduced merely to straightforward programming of a given technique or process.
It is the interaction between the algorithmic tools and the composer’s thought, mani-
fested through decisions, which makes CAC such a rewarding activity. Decisions restrict
possibilities, helping shape the workflow, but also make certain phenomena—like some
accidents—more likely to happen. This resonates with a famous statement by Pierre
Boulez [1]:
Un univers musical sans loi ne peut exister : il s’agit, sous un autre nom, de la
cohérence si chère à Webern; mais la loi seule ne permet pas à l’accident d’exister,
et prive ainsi la musique de la part la plus spontanée de ses moyens d’expression.
30
? ‰ ‰ Œ ‰ Œ ∑ ∑ ∑ ∑
Tba.
¢ j j j j j j j ™ œ œ bœ ˙™
œ bœ œ œ œ œ œ œ nœ #œ œ #œ œ œ #œ œ #œ #œ ™
f n
3 3 3 3:2q
° 3
Vib. & Ó Œ Œ #œ Œ Œ ‰ Ó ##œœœœ œœœœ ‰##œœœœŒ Œ œœ Œ Ó
œ œ ##œœ
Ó
œ #œ##œœ
Œ Ó Œ #œ#œ #œ #œ ‰ Œ Ó #œ #œ #œ œ#œ ‰ Ó
b œœ œœ œ œœ œ œ #œœ mp
œœ
p mp
œ œ
4 Para a pandeireta
Mar. & ∑ ∑ w bw ‘ ‘ ‘ ˙ ˙ w bw ‘ ‘
W.B. / œ œœœ œ œ ™ œœ œ Œ ‰ œ œ œ‰
Folk material
Œ œ œœœ transformations and elaborations in A Vida é Nossa
œ œ œ œ œ œœœ œ œ œ œ œ œœœ œ œ œ œ Œ œ œœœ œ Œ œ œœœ œ œ ™ œ œ œ Œ ‰ œ œ œœ œ ‰ Œ ‰ œ œœœ œ ‰ Œ Ó
Glockenspiel Para Sinos Tub.
Sinos Tub.
¢& ∑ ∑ Ó Œ ‰ j ™ bœ œ œ7 ˙7 Ó ∑ ∑ ∑ ∑
œ œ œ œ œ œ™ J
pp p
=
A
22 . Animato (q = 120)
° ‰œ ‰Ó ∑ ∑ ∑ ∑ ∑ ∑ ∑ ∑
Picc. &
œ œ œ œ œ #œ #œ [ ]
. ‰ ‰ œ. ‰ ‰ bœ. œ œ ˙ œ ˙ Œ #œ œ #œ œ œœ œ œ œ‰
Fl. 1 & ‰ bœ ‰ Ó ∑ ∑
J J Œ Ó
p ff p ff cantabile
b œ. œ.
œ œ œ œ œ œ#œ œ œ œ#œ ™ ˙ #w œ™ #œ.
Fl. 2 & ‰ ‰‰ ‰ ∑ ∑ ∑ Œ œ œ™ J J J Ó
p ff p
œ œ œbœ œ- \ œ. œ. œ. œ
Ob. 1 & ∑ Ó Œ bœ œ™ bœ bœ bœ œ œ œ bœ nœ œ œ œ w ˙ œœ ‰Ó ∑ ‰ œ bœ
J
p pp ppp mf
j
Ob. 2 & ∑ Ó Œ bœ bœ œj bœ œ nœ œ™
J w ˙™ Œ ∑ #w œ™ œ œ œ#œ#œ w
J
p pp
œ œ bœ bœ ™ ˙ w ˙ nœ œ œ œ œ œ #œ
Cl. 1 & ∑ Ó Œ J Œ J #œ ™ w ˙ œ Œ ∑
p mp pp
Cl. 2 ∑ Ó Œ bœ œ w w ˙ œ Œ Ó Œ œœ œ ‰Œ Ó Ó Œ #œ œ œ œ #œ #œ ˙
&
p pp p
Cl. 3
¢& ∑ Ó Œ bœ œ bœ œ œ ˙ w ˙™ Œ ∑ ∑ ∑ ∑
p
1.
°
Tpa. 1,2 & ∑ ∑ ∑ ∑ Ó Œ œ. œ. ˙ œ œ œ œ œj œ™ ˙ w ‰ œœ œœ œœœœœœœœœœ
. . . . . . .
pp n pp
3.
∑ ∑ ∑ ∑ ∑ ∑ Ó j
Tpa. 3,4 & #˙ œ™ #œ œ œ-bœ#œ w
"Senhora do Almurtão" (trad. Portugal)
pp
con sord. [ > ]
Tpt. 1 & ∑ Ó Œ œ. œ. b˙ œ œ bœ bœ œj b>œ œbœbœ œ w w ∑ ∑ ∑
J
mf cantabile p
con sord.
Tpt. 2 & ∑ Ó Œ œ. bœ ˙ bœ bœ bœ œ bœ œnœ ˙ w ˙™ Œ ∑ ∑ ∑
p
con sord.
Tpt. 3
¢& ∑ ∑ ∑ ∑ Œ ‰ œ œ œ bœ bœ w w ∑ ∑
p
Para os W.B.
~~~~~~~~~~~~ l.v.
° œ bœ
‰œ œ œ
3 œ #œnœ œ œ Ÿ˙ œ ‰ #œ#œ ‰ Ó ‰#œ#œ#œ œ œ œ
> l.v.
‰ nœbœbœ œ ‰ Ó
Vib. & #œœ œœ #œ #œ Œ nœbœbœbœ œ œ œ œ œ œ œ œ œ œ œ œ bœbœ œbœ œ œ œ œ œ œ œ œ œ œ‰ ∑ œ œ #œ#œ œ#œ#œ œ#œ
p 5
#œ
3 5
pp p pp mp pp pp
° Pandeireta
ø ø ø
Mar. & ∑ ∑ / œ œ œœ œ œ œœœ ‰ œœœ Œ œ ∑ ∑ ∑ ∑ ∑
pp
Bombo
Caixa Para o Bombo
W.B. / œ œ œœœ œœœ œœœ œ ‰ Œ Ó A*w
— ∑ w7 ∑ ∑ ∑ ∑
pp p
Sinos Tubulares 3 3 Para o Prato Susp.
Glock.
¢& ∑ ∑ bw ˙ bœ ˙ ˙ œ b˙ w ∑ ∑ ∑
mp pp
Figure 17. Score excerpt showing the implementation of the output (boxed). Notated in C.
Colouring melodies
I have always admired the way Olivier Messiaen colours melodies, whether they are
from birdsong, plainsong, or of his own composition. Béla Bartók was one of the first
composers to take this influence from Claude Debussy, as Malcolm Gillies observes about
the Bagatelles, Op. 6 (1908):
[...] streams of parallel 5ths and 7ths (no. 4), of 4ths (no. 11), of tritones (no. 8),
or of piled-up 3rds (nos. 7, 9, 10) [...] The influence of Debussy, about whose
works Bartók had recently learnt from Kodály, also lies behind several of the pieces,
notably in the use of parallel chords, and in no. 3, with its unchanging semitonal
ostinato. [6]
Stravinsky was certainly influenced in this way by Debussy too. The way Messiaen
approaches this technique is frequently more dynamic; the timbres change as the melody
progresses:
When I reproduce a birdsong, every note is provided with a chord, not a classified
chord but a complex of sounds which is designed to give the note its timbre. [3]
31
Gonçalo Gato
A particularly good and simple example is the slow first movement of Éclairs sur l’Au-
Delà: Apparition du Christ glorieux (1988-91). As the pace is slow, one can “taste” each
chord as either timbre or harmony. The fast and staggering third movement “L’oiseau-
lyre et la Ville-fiancée” shows, on the other hand, the same technique at work in a more
virtuosic manner where timbral perception dominates aurally.
I wanted to automate the colouring of melodies according to a given harmonic pro-
gression (or timbre evolution). A melody and a chord progression of the same size would
be input, plus a function (bpf) controlling the positions that the melody’s pitches would
take inside each chord (anything between the lowest and highest pitches). This algorithm
enabled me to test the technique with various melodies, timbral evolutions, and registral
evolutions. The OpenMusic patch is shown in Figures 18 and 19.
Figure 18. Patch used for melodic colouring. Timbral evolution is shown here as an
interpolation from a chord to a unison in 12 steps.
The output, used to build the main climax of A Vida é Nossa, is shown in Figure 20.
As we can see, sometimes the added pitches are higher than the melody, sometimes lower.
This was controlled by the bpf contour.
This algorithm is applied in other instances throughout the piece. For example at
the beginning the low brass melody (Figure 4) is coloured in a similar way, but this
time by using timbres that are subsets of the harmonic series. I have used this kind
of procedure in the past,9 calling it spectral enrichment. This aspect of modularity in
algorithmic composition is well documented and very appealing to me as it means that the
composer can easily create variations on processes by merely patching together different
32
Folk material transformations and elaborations in A Vida é Nossa
Figure 19. The abstraction melody_timbre_with_bpf (left) and its internal loop (right).
33
The output, used to build the main climax of the piece, is shown in Figure 20. As
Gonçalo Gato we can see, sometimes the added pitches are higher than the melody’s pitches, some-
times lower. This was controlled by the BPF contour.
Initial melody
Figure20.
Figure 20. Melodic
Melodic colouring
colouring used
used in
in the
the main
mainclimax
climaxofofthe
thepiece.
piece.
Other instances of the application of this algorithm exist throughout the piece.
For example, just at the beginning, the low brass melody (Figure 4) is coloured in a
similar way but this time by using timbres that are subsets of the harmonic series. I’ve
used this kind of procedure in the past,11 calling it spectral enrichment. This modular-
ity aspect of algorithmic composition is well documented and very appealing to me as
it means that the composer can easily create variations on processes by merely patch-
ing together different combinations of a limited set of subroutines. Accordingly, a
process of calculating harmonic partials was used and connected to the same ‘melo-
dy_timbre_with_bpf’ patch used before. The resulting algorithm is shown in Figure
21, and the output in Figure 22.
The implementation of the algorithm output was automatic but suffered
subsequent manual interventions and elaborations. In Figure 23 one can see the
respective excerpt from the final score. Interventions such as durations doubling and
discarding the first part of the algorithmic harmonization were carried out. Octave
doubling of the lowest voice plus re-harmonization of the first part constituted subse-
quent elaborations. The reason for re-harmonizing was that I wanted a denser sonority
to follow the timpani crescendo (see Figure 1 above). This aesthetic decision draws
attention to the importance of musical context in computer-aided composition.
11
In the piece Vectorial-modular (2011) for instance, an analysis of which can be found at the EarReader
website (Gato 2013).
Figure 21. Spectral enrichment patch.
34
Folk material transformations and elaborations in A Vida é Nossa
Figure 22. Spectral enrichment of a melody. The initial melody is shown on the bottom staff,
with odd-number partials added: 3, 5, 7, 9, 11, 13 and 15.
Figure 23. Score implementation of the algorithm shown in Figure 22 (in C).
Conclusion
This chapter worked as a journey through various compositional procedures used to
compose A Vida é Nossa. Various elaborations and transformations of folk material
were analysed and different types of score implementation were discussed, along with
manual procedures.
By giving special care to the use of appropriate terminology, I hope to have clarified
the procedural multiplicity of the manual/machine interface in CAC. In fact, general
æsthetic issues of compositional craft are found to be omnipresent, always informing
decisions and thus bringing CAC very close to pure manual composition. This intellectual
framework allows for an increased scope of actions in procedures such as algorithm
redesign and manual intervention, elaboration or complementation.
35
Gonçalo Gato
References
[1] Pierre Boulez. “Le système et l’idée”. In InHarmoniques, 1. Le temps des mutations.
Paris: IRCAM-Centre Pompidou/Christian Bourgois Editeur, 1986.
[2] Tim Howell. After Sibelius: Studies in Finnish Music, chapter: “Magnus Lindberg –
Rediscovering Balance”. Aldershot: Ashgate, 2006.
[3] Olivier Messiaen. Musique et couleur: Nouveaux entretiens avec Claude Samuel.
Paris: P. Belfond, 1986.
[4] Curtis Roads. The Computer Music Tutorial. Cambridge, MA: MIT Press, 1996.
Online
[7] François Lesure, Roy Howat. “Debussy, Claude (1862 - 1918), composer”. In Grove
Music Online. Oxford University Press, 2001.
http://oxfordindex.oup.com/view/10.1093/gmo/9781561592630.article.07353.
Acknowledgements: I wish to thank my doctoral supervisor Julian Anderson and co-supervisor Paul
Newland. My first OpenMusic steps were taken with Carlos Caires from the Escola Superior de Música
de Lisboa, who introduced me to the world of Computer-Aided Composition, and so deserves a special
mention and thanks. Also, I wish to thank my first composition teacher Eurico Carrapatoso for showing
me with remarkable elevation and elegance how composing always goes hand in hand with clear thinking.
This research was funded by a doctoral grant from the Fundação para a Ciência e Tecnologia (FCT),
Portugal, with funds from the POPH (Programa Operacional Potencial Humano) and Fundo Social
Europeu (European Union). I have also benefited from scholarships from the City of London Corporation
and the Guildhall School of Music and Drama, United Kingdom.
36
Tri Bhuwana for 12 voices and
Balinese gamelan:
Three worlds/tripartite world
Philippe Boivin
1 Mostof the functions used in the patches presented in this chapter have been programmed as part of
my own OpenMusic library.
37
Philippe Boivin
Systemic modes
My compositional procedures are all based on the same pattern generation system. The
core of this system uses five basic modes developing various probabilities of repetition of
any two elements a and b. A mode starts from two or more basic cells (small sequences
of a and b) and combines them randomly to create longer sequences (see Figure 2). This
random generation allows one to obtain an infinity of Markov-like chains while conserving
the proprieties of each mode.
38
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
Rhythmic generation
In Bali’s traditional music, a musical piece
can be recognised by its general musical
outline, differently treated in each melodic
part of the collective instrument. The
parts roughly correspond to pitch ranges
that are linked to different treatments of
time: long values in low strata, short
ones in high strata (stroke frequency
multiplied by 2 and by powers of 2).
Ethnomusicologist Kati Basset’s concen-
tric notation of the Dhongdhing system
(Figure 5) illustrates this principle of
binary multiplication of stroke frequency
in the different strata, starting from the
centre [2].
I developed a function called gam-ryt,
Figure 5. Dhongdhing system in concentric
which reproduces the principle of nested
notation (Kati Basset 2009).
stroke levels while adapting them to Tri
Bhuwana’s specific musical context. In Figure 6, this function uses three basic durations:
quarter note, half note, and dotted half note to generate four voices containing slower
and slower stroke levels.
The <RYT-mode> argument enables systemic modes finely to articulate alternation
between short values and long ones (1 vs. 2/3), as well as distribution between binary
(2) and ternary (3) long values. In this example, the same mode (2) is used for both
operations. The list of durations thus obtained is then restructured by an algorithm
regrouping either the single short values with the preceding long one or the succeeding
short values only, up to a maximum of two quarter notes.
39
Philippe Boivin
Melodic generation
The Bintang Tiga gamelan plays along a pelog 2 melodic scale of five notes corresponding,
given a few corrections, to the Phrygian mode transposed to C]. Clearly, however, the
40
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
notion of pitch is quite relative in Balinese music. The tuning of the instruments may
greatly vary, not only from one village to another, but also among different instruments
within a gamelan, or even from one octave to another on a given instrument. Besides,
all blade instruments (for example the gender) are played in pairs, matched to produce
beats whenever a single note is played by two musicians at the same time (Figure 7).
This is a musical world that is not equally tempered, but extremely rich; vibration
prevails over the principle of absolute pitch founding our present Western system. The
spectral analysis tools available in OpenMusic have proved particularly useful to study
these subtleties, often quite difficult to hear (Figure 8).
As with rhythmic generation and with a view to keeping a coherent approach, systemic
modes were used again, this time to generate melodies on the five notes of the pelog
scale. An algorithm controlling interval succession (Figure 9) ensures melodic outline
and colour, while leaving out over-connoted “tonal” phrases. This algorithm first controls
alternation between seconds and other types of intervals, then selection between thirds,
fourths and fifths. The gam-mel function is able to generate melodies of <tot> notes
whose outline may be modified by shifting the <low> inferior limit not to be overstepped.
Moreover, it also cuts this melody into several small loops to the return of the <start>
note.
41
Philippe Boivin
This operation of segmentation permits the endless repetition of the same loop or
the free linking of several corresponding loops together. As it happens, the principle of
repetition of more or less long cycles, punctuated by low gongs, is one of the bases of
Balinese traditional music. Moreover, redundant melodic formulas can be avoided within
a melody that may appear slightly too long as it is limited to a combination of the five
notes of the scale. Figure 10 shows how melodic-rhythmic blocks associating loops and
various stroke levels can now be built.
42
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
As may be seen in Figure 10, every stroke level finds its pitch material in the main
melody, whose corresponding notes are sent down into the lower octaves according to the
range of the selected instrument. The group of four gongs playing the block’s last stave
is the only one to use its own pitches, distributed in relation to the number of strokes in
the sequence according to an alternation principle called colotomy: Wadon (0) is struck
only once at the start of the sequence, Lanang (1) once in the middle, Kempur (2) and
Kletong (3) are distributed on the remaining strokes (see Figure 11).
In this example, the first loop returns a balanced sway visible in the third voice
that is distributed across 10 beats (3+3+2+2) and the second one a resolutely ternary
structure (3+3), two rhythmic morphologies that can never be found in Bali’s traditional
music. Even so, these novel textures fully respect this music’s usual acoustic features in
a renewed context.
What is also significant is that the rhythmic structure of the stroke levels is similar
in the two loops, although it is shortened in the second one (Figure 12). This constitutes
a basic principle to develop material with limited means in Tri Bhuwana. Rhythmic
repetition introduces organic coherence in the melodic variation brought when passing
from one loop to the other. In the following example, the three stroke levels are shown
in line so that the loops may be compared.
In addition, rhythmic variations enable the building of a new type of block, keeping
the same melodic loop. In Figure 13, at the marking RytSel.2, the regular flow of the
main melody is interrupted to give it an irregular rhythm (the rhythm of the second
stroke level played by the Calung). The resulting time dilatation in the melody, as well
as the shift of meeting points, deeply change the perception of the initial object to reveal
new aspects. Tri Bhuwana constantly plays on this dual repetition/variation concept
thanks to the modification of a few generation parameters. This allows the development
of long and non-redundant sequences from a few notes only.
43
Philippe Boivin
Proliferations
A particular playing technique called kotekan makes it possible to develop new figures,
sometimes very fast, between two successive sounds, whatever the stroke level used.
Thanks to a collection of orally transmitted formulas, Bali’s musicians know a great
many ways to pass from one note to another logically, by filling time and space at a
given speed. The four main kotekan families used in traditional music (Nyok-cag, Norot,
Telu and Empat) were analysed at length and then modelled by algorithms that are able
not only to reproduce their operating modes, but also to generate new ones.
Figure 14 first illustrates the principle of internal expansion of kotekan Nyok-Cag.
Expansion exp.2 inserts an intermediate note after each pitch of the mel melody, accord-
ing to rules based on the observation of traditional practice. It usually is a neighbour
note, superior or inferior to the expected following note. Taking exp.2 as a basis, exp.4
reproduces the same principle, and so on, each time doubling execution speed. It is
interesting to note that the algorithm tends to get more complex with the iterations in
order to improve the plasticity of the melodic lines. For instance, the choice of a simple
treatment (marked
) 1 or of a note “enjambment” (
) 2 may be made according to a set
of cumulative rules: redundancies to be avoided, an ambitus to be respected, a number
of different notes not to be exceeded, etc.
The very long resonance of the gamelan instruments must be taken into consideration.
The usual way to play the blade instruments is by muting the note just played, with one
hand, at the same time as the other hand strikes a new one with a mallet or a small
leather-covered wooden hammer. So as to play fast, there must be two musicians, one
completing the other’s part, sometimes with impressive velocity. As shown on Figure 15,
the technique of distribution of such melodic garlands between two players called Polos
and Sangsih is much simpler to write down than to apply, at least for Western musicians.
44
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
Kotekan Norot is much simpler to program, as there is only one option to set it.
First, developed from the starting note, is a sequence of superior neighbour notes that
ends three impulses before the expected following note. The latter is then struck twice
and followed by its own superior neighbour note (see Figure 16). So the predictability
of the formula is very high, since the expected note is always announced; moreover, it is
accentuated through being doubled, as shown in the lower staff of Figure 16. This type
of kotekan is particularly interesting as it can run at slow speeds as well as fast ones, but
also with possible chordal versions, such as have been developed in a novel way in the
composition of Tri Bhuwana.
Kotekan Telu is a clever mechanism: a cell of three adjacent notes of the pelog scale
is repeated several times in such a way as to obtain a succession that necessarily ends on
the expected following note (see Figure 17). The modelling process has proved extremely
complicated, due to a number of qualitative parameters to be dealt with carefully: full
melodic analysis must be carried out before selecting the cells in order to hold the same
notes as long as possible; the start of the process must be set accurately so as to retain
elegance (see additional parenthesised pitches in Figure 17); the cells must be oriented
along the melodic line, etc. The distribution between the two musicians is remarkable in
terms of both instrumental playing and sound effect: only two in three notes are played
by each, the central one always being doubled, so naturally accentuated.
Kotekan Empat is a type of extension of kotekan Telu, with the addition of a higher
fourth adjacent note. As shown in Figure 18, the additional pitch is set above the lowest
note of the three-note cells. The accentuation resulting from the simultaneous strokes
then creates a new “virtual” rhythm typical of these interlocking structures [5]. Just
making Sangsih play the two superior sounds and Polos the two inferior ones will no
doubt ensure velocity and fluidity, provided that they get along well!
These diverse techniques of expansion inspired from traditional practice can be mixed
at will to create varied and renewed polyphonic textures. Figure 19 shows an alternation
45
Philippe Boivin
of three different kotekan, as tested in a study preceding the composition of Tri Bhuwana.
Here, both kotekan Nyok-Cag and kotekan Empat follow the continuous melody in regular
quarter notes of the first stroke level played by the Panyacah (*1). The kotekan Norot
of the last two measures is adjusted on the airier rhythmic line of the second stroke level
handled by the Calung (*2).
46
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
Figure 21. Parameters generated by the macrostructure. The gam-gongan function outputs
the rhythmic structures, pitch values, as well as other parameters such as the tempo, modal
scale, link note with the next section, etc.
47
Philippe Boivin
Using this method and as shown on Figure 21, the validity of more and more complex
combinations could then be checked, which would have been difficult without the capacity
to listen almost instantaneously to the instruments sampled (a sound bank of over 200
samples I recorded blade by blade with several playing modes). This illustrates that, by
meeting the specific needs of a given work, Lisp programming can be an integral part
of the creative process of a composer, as it has indeed been for me over the past forty
years. It is then time for pencil and rubber to refine, develop, reorient, and even delete
the results returned.
This method, though appealing at first sight, was soon left aside for several reasons,
among which uncertainty as to the ability of singers to hold such intonations against the
unsettling background of the gamelan.
A more realistic approach gradually stemmed from the many simulations carried out
with synthetic sounds. A specific modal system seemed to allow the choir to sing in its
own equal temperament, taking as a reference diapason the low C] of the large Wadon
gong (427.5 Hz). The idea was to have the two sound spheres coexist, the gamelan
creating an impressive broadening of acoustic space for the voices in some cases, and
vice-versa in other cases.
In order to provide frequency areas that are common to both ensembles, the modal
scales used include all the sounds of the pelog scale. As shown in Figure 23, each of these
sounds can be paired with one of the five sounds of the pentatonic scale of A transposed
to C, which forms a second level for the generation of the modal scales. The two missing
sounds F] and B make up a third level, which completes the total chromatic range.
Each section of Tri Bhuwana is always associated with a polar note, which is both a
starting point and a link between the different loops used in the sequence. The section
features a modal scale of nine or ten sounds built in three stages, considering the polar
note of the section and that of the next section as shown in Figure 24. To the pelog scale
are added the sounds from the pentatonic scale not associated with the two polar notes,
then the possible insertion of a mobile degree with a view to filling the intervals superior
to major second.
48
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
Figure 24. Augmented pelog using notes 2 and 3 as polar notes (development in three stages).
The fifteen scales used in this opus are all built on this single unifying principle so
that through the interplay of repetitions, transformations and loop strings, each section
asserts its own modal colour. Besides, transitions from one section to the other are made
in a gradual way, through the linking of the polar notes—those of the current section
and of the next one—used to set the different scales. As shown in Figure 25, only A]
turns into B when one goes from mode 3-3 to mode 3-4.
Figure 25. Augmented pelog: the 15 scales of Tri Bhuwana. (The numbers indicate the
positions of the two polar notes for each scale.)
From a set of four pelog loops and three modal loops corresponding to section C1 of
the piece, Figure 26 shows how to isolate a melody (A), then redistribute it as a vocal
quartet in an expanded spectral layout (B), and filter it to let the gamelan punctuate it
on the pelog pitches it contains (C).
49
Philippe Boivin
Modal loops 1 and 3 actually use defective forms of the ten-sound mode, whose
complete version is only displayed in loop 2 (see Figure 27). This option makes it
easy for several intermediate modal colours to be found within a given section. Even
if the generation system is completely automated and linked to the macrostructure of
the piece, such subtleties clearly highlight the total liberty left to the composer to handle
these various materials according to his own inspiration, so as to meet musical needs of
a superior nature.
Moreover, the capacity to explore in depth the qualitative aspect of vocal generation
while conserving an organic link with the gamelan offers the last but not least advantage
of returning a material that is immediately ready to use. Just like the entirely modelled
kotekan techniques, this material may in turn be subjected to very sophisticated algorith-
50
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
mic processing, this time adapted to vocal writing: for instance, modal transpositions,
recurrences and inversions, concatenations, imitations, harmonic overlaps, rest control,
and so on. These functions make up an extensive library. They may be applicable to a
group of voices as well as to isolated voices. The flexibility of graphic programming—the
capacity to encapsulate several generation levels and to gradually refine the returned
results—makes it possible to build draft sketches that can generate musical sequences
lasting several minutes, as shown in Figure 28.
Beyond technology
This project has been a step forward in my understanding of musical languages, first and
foremost of my own. But I must now come to the heart of the matter.
It took several months for me to get attuned to the world of the gamelan [3]. I actually
had to learn how to write for this collective body, whose operating mode was utterly
unknown to me, being so far from that of a Western orchestra that is mainly focused on
the individual. While collecting relevant literature [4] and developing the architecture
of the Lisp program at the same time, I produced several musical sketches that were
immediately tested by the musicians. This to-and-fro process allowed me gradually
to grasp the usual practices, then much rarer ones sometimes reserved for initiated
musicians, for instance the art of Kendang or of Gender. Only then could I start designing
my own devices of composition for the gamelan.
However, the true question came up later when it came to choosing the title and the
texts with Kati Basset, who helped me chart my own course thanks to her judicious
51
Philippe Boivin
advice [1, 2]. She reminded me that the practice of the gamelan was first and foremost a
ritual and social gesture, far from the concept of art or pure aesthetics defining Western
concerts where music has its own intrinsic value. From the start, the work was conceived
as a triptych combining operating forces. The Balinese cosmology founding Tri Bhuwana
has been a means to give deeper sense to the speculative choices made at the beginning
of the project.
SWAH is the world of essences, principles and oneness, the origin of all things.
BUWAH is that of form and objects perceived by our senses in their everyday use.
As for BHUR, it is the world of lower reality, devoid of spirituality and tending
towards shapeless disintegration.
Kati Basset
In Tri Bhuwana, each of the three worlds is here associated with such diverse texts as
poems about meditation, knowledge, or death in Swah, verse by Jean de Baïf or François
Villon in Bhur, and even a half-genuine, half-mock version of a University governance
report for Bhuwah. Beyond the technical considerations described in this chapter, my
ability to give musical life to the basic principles of this tripartite world became the
main stake of this project. The leads were manifold and could induce treatments of
the material that I would never have imagined without the support of this cosmological
guideline. That is how willingly unfinished musical features such as armless kotekan or
headless melodies quite naturally found their way into the score, echoing the Bhur infra-
world, peopled as it is of Buta-Kala with incomplete bodies, as illustrated in Figure 29.
A last example sheds light on an original way of envisioning the text/music relation-
ship from an algorithmic outlook. It is the crossing of a text that is sacred in Java and
Bali, the “Syllabary of consonants” (visible on the left-hand side of Figure 30), with the
vibratory matter of the aksara sound-letters [e u a i o] used by Balinese musicians to sing
the names of the five notes of the pelog scale. The two voices shown in Figure 30 recite
the syllabic text which is associated with two modal transpositions of these vowels. A
new text appears, whose meaning is in no way altered by the shift effected: Once upon
a time, there were two emissaries of like power. They killed each other at the top. It is
worth noting that due to the strict logic of the device applied, the two sopranos felt very
comfortable singing this completely invented “Balinese” language.
52
Tri Bhuwana for 12 voices and Balinese gamelan: Three worlds/tripartite world
Clearly, such a long-lasting immersion in the world of Tri Bhuwana has opened my
mind to new dimensions, in particular a broader relationship to time, together with a
more physical approach to sound vibration. As I was proceeding in the writing of the
piece, I felt a deep drive to let go, to forsake the frantic nature of some of my composer
habits. Indeed, tirelessly repeating the five evocations of the name of Shiva so as to
elaborate a complete sequence (as in Section B2 of the piece) cannot be insignificant,
even with the use of variation. In this way, Tri Bhuwana will no doubt leave its deep
and long-lasting mark on my subsequent work, as was already clear in the pieces that
followed it. Even if Revenantes, a piece for female voices and cello, seems far remote
from anything Balinese, this opus composed just after Tri Bhuwana is still deeply under
its influence [7]. I have no doubt that from the spiritual experience I underwent with
Tri Bhuwana has stemmed a genuine renewal of my musical language and composition
methods.
53
Philippe Boivin
References
[1] Catherine Basset. “Au delà des apparences. Morphologie des esthétiques et cosmologie
à Bali”. Le Banian, 9, 2010.
[4] Michael Tenzer. Gamelan Gong Kebyar: The Art of Twentieth-Century Balinese
Music. Chicago: The University of Chicago Press, 2000.
[5] Wayne Vitale. “Kotekan: The technique of interlocking parts in Balinese music”.
Balungan, IV(2), 1990.
Online
[6] “Tri Bhuwana for gamelan and voices, creation in 2013”. Project presentation video
(6’32), 2013. http://www.musicatreize.org/fr/tri-bhuwana-mpx16.html.
[7] Philippe Boivin. “Choir and gamelan / Choir in prison”. Lecture recorded at IRCAM,
2015. http://medias.ircam.fr/x21af85.
Acknowledgements: The writing of this article has benefited from a work residence at the Villa Médicis
French Academy in Rome in August 2015. I am deeply grateful to Kati Basset for her expert advice and
unflinching support throughout the long musical and spiritual progress inherent to this project. I would
also like to thank Nicole Mausset for her translations and rereading of this text.
54
Programming modular progressions
in OpenMusic
Matthew Lane
This text presents the author’s modular approach to designing and using OpenMusic
patches to produce multi-dimensional progressions in instrumental music.1 Musical
passages (in one or several instruments) are passed from process to process with tailored
parameters for each process and instrument, much like an assembly line working on
a single base material. Individual progressions gradually change the passage in each
instrument, altering combinations of pitch, time, structure, and stochastic elements,
before passing the result to the next progression. The chapter introduces the overlying
organisational structure, as well as the justification for this system, and guidelines for
moving between these patches and composing by hand. Examples that make use of these
patches are presented in four compositions: Sliding Apart, Melodious Viscosity, and two
movements from Short Pieces on Falling.
Introduction
Computer-assisted composition (CAC), and specifically the use of OpenMusic, is es-
pecially prevalent in my music related to the development of progressions. I define a
progression as a directed alteration of elements of a musical passage over time. This
applies both to continuous progressions (one long line that continues to change), and
sequences (progressions where the same idea is repeated while evolving at each instance).
Sequences are essentially changing repetitions, and have long been in use, but have until
the 20th Century been mostly confined to diatonic (and sometime chromatic) modulations
of the base element (model). In the 20th Century, especially with the music of Olivier
Messiaen, rhythm began to change in a formalised way.2 Eventually, every element of
music became a possible parameter for progression, including pitch, rhythm, articulation,
timbre, speed, and all their subcategories.
My compositional process could be summarised as follows: creation of an original
idea (seed), creation of related ideas, triage of ideas, development of remaining ideas,
interpretation/reinterpretation of developed ideas where necessary, formal organisation,
55
Matthew Lane
and linking. I have used CAC in nearly all of these processes, but the present focus will
be on development. Developing different parts of a piece using the same OpenMusic
patches, with the distinct traces and colours they leave on the music, helps to create
unity between sections and to shape the æsthetic of the piece.
This text will refer throughout to four pieces: Melodious Viscosity (2013) for wind
quintet; Sliding Apart (2013) for flute, clarinet, piano, violin, and cello; and movements
“Rings” and “Float” from Short Pieces on Falling (2015) for flute, clarinet, piano, violin,
and cello. The latter in particular is an exploratory set of over a dozen small pieces
to examine which uses of computer-assisted composition I hope to extrapolate in larger
projects over the next few years; the two selected movements represent the compositional
processes I found the most appealing.
3 The applicator uses “lambda” functions applied to the passage for any number of voices at a time.
56
Programming modular progressions in OpenMusic
Applicator patches for the first treatment apply the functions directly to chord-seqs, and
entire chord-seqs are passed to the progression function through the applicator. Where
necessary, such as with discrete progressions, the applicator contains a mechanism to map
break-point functions (bpfs) or constant values to the arguments required for processing
each chord-seq. Applicator patches for the second treatment allow functions to be applied
to one note at a time, which I term musical atoms (including pitch, onset, duration,
velocity, MIDI channel, and offset information), and the application mechanism maps
different points on a temporal bpf to different arguments for the function being called.
Values can thus be sampled from a bpf for each individual atom in a line, and used as
arguments for functions that alter atoms.
57
Matthew Lane
Figure 4 shows how values from a bpf are mapped to a musical passage depending
on whether they are mapped continuously or discretely, and whether they are mapped
by chord-seqs or by atoms.
chord-seq
!
Points on bpf mapped individually to Points in the bpf mapped to each note
each note in the passage based on the based on the position of the note’s
note’s temporal position. parent cell in the passage.
atom
Figure 4. Continuous and discrete progressions using chord-seq and atom mapping.
Figure 5 shows the modular framework of these four systems implemented in Open-
Music. The music is passed from an original motive at the top to the complete altered
progression at the bottom. At the beginning, a multi-seq
!
class (A) contains an imported
motive. (B) shows bpfs for controlling the transposition, one bpf for each voice. The
next level (C) differs for continuous or discrete progressions. For continuous progressions,
the patch createbasepolypattern
!
multiplies this motive, creating and joining any number
of copies of it for any number of lines. Thus it produces one long chord-seq with several
iterations of a motive for each instrument. The patch fillmatrix from multiseq, for discrete
progressions, is similar but does not concatenate the repetitions, instead holding them
in a matrix so the cells can be treated individually. Step (D) is the application of a
patch to one or several musical passages—the primary place where these four methods
differ. For the purpose of this example, the same list of bpfs is used for all of the
progressions, one for each voice of the input data. In the first method on the left
(continuous progression/chord-seq mapping), the chord-seqs and bpfs are passed one at
a time to the patch chord-seq transpose by BPF via applydifferentlytoseveralscoreobjects,
58
Programming modular progressions in OpenMusic
which does little more than a the Lisp mapcar.4 In the second method (continuous
progression/atom mapping), the applybyatoms function maps the values from bpfs (or
constants or lists) based on the onset time of each note, and the lambda patch note
transpose only deals with one note at a time. For transpositions this is trivial, but for
more complex progressions this allows the programmer to focus only on how an argument
affects one note at a time, and saves reprogramming the bpf time-mapping each time.
Applybpfstomatrix is essentially another elaborate list iterator, working on the matrix of
chord-seqs instead of a list and sampling the supplied bpfs (or other values) to produce
the lambda function arguments. Applybpfstomatrix by atom is similar, but instead of
passing chord-seqs it processes individual notes as atoms—being careful, however, to map
the same values onto all the notes from one musical cell. This ensures that cells can be
treated independently: for example when transposing a cell this way intervals in the cell
will remain intact, but when using the applybyatoms method for continuous progressions
each interval will be altered as the transposition curve is applied (see Figure 2 for an
example).
Looking more closely at the resulting chord-seqs in Figure 5, the first and second are
identical, as are the third and fourth. So in this case application by chord-seq and by
atom yield the same results. However, the left two continuous transpositions differ from
the right two discrete transpositions: in the continuous case, even within the first three-
note cell, the initial intervallic relationships are distorted compared to the initial cell at
the upper-left of the patch. In the discrete case, entire three-note cells are transposed
4 Unlikemapcar, where each argument has its own input, applydifferentlytoseveralscoreobjects has one
input for the score objects and one for a list of all other arguments combined.
59
Matthew Lane
together, maintaining the size of the melodic intervals in the initial cell: these intervals
are repeated starting again at the fourth chord, the seventh, and the tenth.
Any number of these applicators with corresponding functions can be applied to the
lists/matrices of chord-seqs. Finally, several further functions are used:
• present list as multi-seq (E) takes the matrix of chord-seqs and reduces it to one
multi-seq, concatenating the cells in each voice.
• mergegroups (F) combines certain lines together as part of one instrument. For
example, in some cases one instrument (like a piano) may play several lines at
once—this allows them to be merged together into one staff.
• assignchannels (G) assigns each line its own MIDI channel, imperative for audi-
tioning the result through a sequencer.
A note on atoms
Atoms5 are a concept to simplify how functions are applied to chord-seqs where a
process is being applied to each note inside the chord-seq independently. Applicators like
applybyatoms manage the entire process of listing the atoms (“atomising”), passing each
atom to a function, and mapping the appropriate arguments for their transformations.
This includes mapping the y-values of bpfs supplied as arguments to different atoms
based on their onsets. Once inside a function, another function atom-›comps splits up
the elements of the atom (pitch, onset, etc.). Figure 6 shows a simple example of how a
transposition function would work in the context of applybyatoms.
Figure 6. applybyatoms maps bpf values to a function that alters each note individually.
5 The “atoms” used here are not the same as Lisp atoms, but retain the name for the same reason: they
are indivisible. They represent, for this way of working, the smallest piece of information that can still
be useful to a musical progression.
60
Programming modular progressions in OpenMusic
Types of progressions
The progressions used in my music can be grouped into three categories: pitch functions,
time functions, and stochastic functions, although with a significant degree of crossover.
Pitch functions include those that gradually raise/lower pitch or gradually expand or
compress a melody’s range. Time functions might include those that gradually slow
down/speed up a passage, apply a curve or mathematical function to the speed of a
passage, or gradually alter the lengths of the notes. Stochastic functions, usually used
in conjunction with pitch or time, include those that gradually increase/decrease ran-
domness in time or pitch, but might also include processes that gradually but randomly
remove, repeat, or freeze notes.
Most of these functions work well using atom-application tools because the various
values that affect change are applied on a temporal basis (as progressions generally are),
making the bpf mapping mechanism very useful. Some progressions, however, require
the function to consider overall low and high values for pitches or note lengths across an
entire chord-seq. This is poorly handled with atoms because they allow a function only
to manage one note at a time, disregarding other notes. In these cases, functions are
constructed to apply to whole chord-seqs using instead applicators applybpfstomatrix or
applytoseveralscoreobjects.
6 Thisnesting (possibly recursive) can be used to generate many coherent musical structures, including
and especially fractal structures. For more on this see [1].
61
Matthew Lane
A variant of this idea is the possibility easily to create progressions that go elsewhere
from the original instance. For example, the same progression could occur many times,
each transposing to a different tonal area, possibly over a different period of time. Finally,
one of the most interesting benefits of composing in this way, and one that can be
extrapolated to OpenMusic in general, is the occasional artefacts from certain processes,
the “tool marks” left by the software that often provide the unique gems from which entire
pieces can be built.
7 Several constraints programming systems exist for OpenMusic, e.g. OMClouds: see [4].
8 OpenMusic is also conducive to other forms of random distribution, using libraries like OMAlea.
62
Programming modular progressions in OpenMusic
of straight lines or simple curves. Once a progression becomes too predictable, it loses
much of its dramatic value. Even randomness, which adds a degree of instability, can
itself become predictable. One approach to dealing with this is by interrupting the
progression. I have developed tools to incorporate freezes and silences in some cases based
on direction changes in bpfs, but these also engender their own type of predictability.
Randomly (or partially controlled randomly) placed freezes, pauses, or other processes
such as flashbacks to previous sections, occasional scrubbing backwards and forwards in
the sequence, or elements as simple as sudden octave transpositions are all possible, but
this kind of formal randomness is rarely an æsthetic I seek while composing. Thus, I prefer
interpreting the OpenMusic data by hand and instead focusing on listening. As soon as
I recognise the intention of a progression, that means there is not long to change it before
it loses its musical interest. From there, the question is whether the progression should be
interrupted and continued again, interrupted and restarted, or stopped outright. When
a progression returns for the second or third time, as a recurrent idea in different parts
of a piece, the third option is often best. Next, the nature of the interruption comes into
question. It may be a slightly altered continuation of the progression already in progress
(for example, an upward scale that momentarily doubles in speed or becomes an upward
glissando). It may also be a sort of stop or break, such as a freeze or a silence. Or it may
refer to completely different material, although usually material that has already been
introduced, suggesting a flashback. Finally, only once the first interruption is written is
it possible to move forward and determine where and if another interruption is needed.
To be musically useful, this process must take into account the material and time that
precedes it.
Even beyond these two issues, the data may still lack the flow or spontaneous feeling I
seek in my music. Regardless of how far the patches go, the data itself is rarely musically
suitable for my needs. For this, more drastic measures are required: it is here that
reinterpretation becomes relevant.
Reinterpretation
What I call reinterpretation is a process like the interpretation of data discussed above,
but a step further towards impulsive intuition. The means are varied, but the principle is
the same: become a musical performer and improviser of one’s own data and document
the process. In some cases, this involves playing through the passage on an instrument
and recording or transcribing (especially the rhythm). This opens the possibility to move
back and forth in time, add occasional notes, or alter harmonies where the ear suggests.
In cases with many lines undergoing processes simultaneously, I may take what strikes
me as the leading line, reinterpret this, and then rebuild the other lines around this one.
Some computer-assisted tools for reinterpreting the passages are also useful, including
scrubbing and transcribing the result in a notation program or sequencer, or simply
tapping the tempo or chord changes and re-transcribing the new rhythm. These are not,
however, recipes to successful reinterpretation and require discretion to determine their
utility in each case.
Most important, about both interpretation of data and reinterpretation, is the time
and “space” required away from the computer-aided composition environment. Other-
wise uncomfortable passages become familiar and comfortable sounding, and undesired
rough edges become softened in our memory. Programming in OpenMusic can involve
63
Matthew Lane
laboriously going over the same passages many times for hours or days on end. This is
why, for me, it is crucial to return to the output data several days after the evaluation
and ideally to return to it in a new context. The time allows me to rehear the musical
ideas in a fresh way and to be surprised, a crucial element to my music. Either playing
the output on an instrument, in a sequencer, or in one’s head can work, but I avoid
returning to OpenMusic. For this reason, I tend more and more to work on multiple
parts of a piece at the same time, allowing me always to have something to which to
dedicate my attention (a principle of Toyota’s “lean production”, see [5]).
9 The score order in Figure 7 is: flute, clarinet, piano, violin line 1, violin line 2, cello line 1, cello line 2.
64
Programming modular progressions in OpenMusic
65
Matthew Lane
The interpretation of data for this piece involved multiple steps of reworking the
output data by hand. The major changes from the output data were the inclusion of
an interruption and another process (also generated in OpenMusic) cross-fading with
the ongoing progression. The interruption is a combination of two events: a “freeze”
in the wind instruments and piano, while the strings continue the progression by their
downwards fall (Figure 10, m. 32). The cross-fading process, a rising, somewhat spectral
piano chord progression, helps offset two problems in the first progression. It counteracts
the loss of energy created by the falling-apart progression, and it distracts from the
increasing predictability of where the first progression is headed.
Figure 10. Sliding Apart score segment, showing the gradual deceleration and downward
motion of the flute and clarinet lines along with increased randomness in both the violin and
cello lines. The small downward slashes after some notes represent short glissandi.
66
Programming modular progressions in OpenMusic
material also each undergo decreasing randomisation, hence the importance of the patch
createbasePatternWithMultipleEvaluations (Figure 11), which allows the random element
to work to its full potential, creating slightly varied instances before submitting them to
the global chain of processes.
The most important process is gradualnoteremoval (see Figure 12), again in con-
junction with adjustdurations in order to fill out the space emptied by the removal of
notes. Each instrument is given separate parameters for when to enter. There is also
an acceleration treatment through rhythmextendcompress, but in the end I deemed it
simpler for the players to read an accelerando in the score; although the OpenMusic
process helped me gauge the efficiency of the process as a whole.
67
Matthew Lane
Figure 13. Melodious Viscosity for wind quintet, score excerpt (concert pitch).
A crucial part of the interpretation of results in this piece involved performing many
evaluations. After that, bars were chosen from different sets of results that filled certain
auditory criteria, including a desire for a feeling of evenly distributed randomness, attack
points that did not sound like errors, and interesting harmonic coincidences. Then a
partial reinterpretation was done, primarily of what I perceived to be the leading lines:
bassoon (on the bottom staff of Figure 13) as well as the highest-pitched line present.
The bassoon’s first notes for several measures were altered to make the passage feel
less harmonically static. Finally, touches of colour were added: occasional harmonic
contributions by instruments where the texture was too thin, and occasional performance
techniques momentarily to draw attention away from the progression. This progression
occurs in two section of the piece, and the CAC framework allowed the creation of two
independent, but closely linked, progressions.
68
Programming modular progressions in OpenMusic
becoming clearer as the lines move higher and all of the instruments become consistently
present. Short rhythmic segments change speed, reflecting their slopes on the generating
bpf. All of this is juxtaposed with sudden stops in the fast and difficult-to-predict lines.
First, as shown in Figure 14, a rhythm
is developed from the hand-designed bpf
(createrhythmfrombpf - linelength). The
steeper the slope of a bpf segment, the
faster the notes repeat. Then the same bpf
shape is used to apply pitches to the existing
rhythm (apply notes to rhythm chord-seq).
Third, the chord-seq is “frozen” in places
(notes are suddenly held) based on changes of
direction in the bpf (and therefore melody),
using the functions get direction changes in
BPF and multi-freeze. Proceeding to part 2
in Figure 15, the function bpf-multiscale-y
scales the y-axis of the bpf for application
to different functions: while the same bpf
is mapped to rhythm and midicents, the
range of values needs to be different for each
(a difference of 2000 for 1 2/3 octaves in
midicents, and 5000 ms as an appropriate
rhythmic slope).
At this point, the result is still a single
melody (it enters Figure 15 via the connec-
tion at the top left), but is then passed to
applybyatoms to apply a ring-modulation one
note at a time, using the given note from
the chord-seq as the carrier and the mapped
value from the bpf as the modulator. Note
that the shape of the line has changed Figure 14. Patch for “Rings” (1).
somewhat due to the process of “freezing” in
certain places, and the function multi-freeze has provided a new bpf with “freezes” in the
same temporal locations as they have been placed in the chord-seq. Between the carrier,
modulator, upper sideband, and lower sideband10 there are now chords of four notes
throughout the progression, and the function split chords to melodies splits these into
four new lines (the opposite of the function merge). The final significant function, applied
to the four chord-seqs using the applytoseveralscoreobjects applicator, is noteremovalbybpf,
which uses a bpf with y-values from 0 to 100 to determine stochastically the percentage
chance that a given note stays or is removed. In this case, it is arranged so that as the
melodic line moves higher there are more notes retained. Finally MIDI channel numbers
are applied to each line, and the output is displayed in a multi-seq.
69
Matthew Lane
70
Programming modular progressions in OpenMusic
71
Matthew Lane
72
Programming modular progressions in OpenMusic
Next, only a portion of each cell is preserved, again using the same bpf mapped
to the select with intervals function and applied to individual cells. This mapping and
application of the function is done using applybpfstomatrix once again. As opposed to
a standard select function, select with intervals forces the selection only at certain time
intervals in the passage. For example, this ensures that if it received the mapped starting
and ending points of 212 and 492 ms, it actually would select from 200 to 500 ms (for
an interval of 100 ms). The function also contains safeguards to ensure that the final
resulting chord-seqs are in fact the exact length requested, even if there is empty space
at the end, and takes not only entire notes, but cuts portions of notes where necessary
at the beginning and end.
The most characteristic aspect of the progression is the repeatsegment function applied
to each cell. This function cuts out a segment from somewhere in the chord-seq and
appends it a certain number of times to the end. In this case, due to the mapping
by applybpfstomatrix, the length of the portion to extract and the number of times it is
repeated are both controlled by the same bpf shapes that have controlled everything else.
Thus, as the bpf moves higher, creating higher transpositions, intensity is also accrued
by shortening the cells unexpectedly and by obsessively repeating their final segments.
Finally, the speedchange function quantifies the note lengths so that the notation software
(in this case Sibelius) recognises them as eighth notes (for example 500 ms at a tempo
of 60). Figure 20 shows how one cell might change as it passes through all of these
functions.
Figure 20. How one cell transforms as it progresses through various functions.
Because the structural integrity is maintained from the original cell provided, this
type of progression is already much more logical to the ear, and finding a convincing
progression involves mostly work on the initial bpf and some of the scaling factors.
The work of interpreting this data into a score is less destructive, involving primarily
the addition of dynamics, to sculpt a curve of intensity reflecting the direction of the
passage, as well as articulation. Both of these additions help punctuate clear divisions
between the original cells in OpenMusic, so that the beginning of each new cell is clearly
recognisable. Figure 21 shows a segment of the final score.
73
Matthew Lane
74
Programming modular progressions in OpenMusic
Further tools
While the applications for this modular system have not been exhaustively explored, some
simple additional tools have drastically multiplied the contexts in which these patches
are effective. These include, for example, a patch to constrain musical objects to a given
mode and a patch to contract several instruments’ parts to a common temporal length.
The former passes through all the notes in a chord-seq, and if they do not belong to
the supplied mode, alters the notes so that they do. The later takes a list of chord-seqs
(corresponding to different instruments for example) and adjusts the final note of all but
the longest so they are all of equal length. While working with matrices of chord-seqs,
where all the elements in a row will eventually be concatenated, this can be crucial in
ensuring that the instruments (the different rows) line up with each other.
Other OpenMusic and sequencing resources also could allow more complex and
layered progressions. One possibility is to use multiple channels corresponding to different
playing techniques of one instrument, to allow gradual progressions where the stochastic
chances of each techniques change throughout the piece. Another is to use 3dcs (three-
dimensional curves) to map the y- and z-axes to different elements of a progression.
CAC consistently raises the question of how much “composing” the software does.
At present, no data developed in this system could be used as music in its raw form,
but analysing why helps us move closer to musical results both with OpenMusic and
by hand. The more we can formalise solutions to the problems associated with musical
data from OpenMusic, the better we are prepared to formalise solutions for issues in
our own music by other means. It becomes irrelevant whether a composer creates a
patch to rectify a problematic passage or finally composes it by hand: once the patch
is created, it means that the composer understands perfectly the cause and solution to
the problem in small steps. Consequently, should a composer never even use a note of
data taken from OpenMusic, the process of programming these progressions remains
inherently valuable. And while CAC programming may be pushed farther to interpret
our musical ideas, they remain our musical ideas, and not the computer’s.
References
[1] Michael Gogins. “Iterated Functions Systems Music”. Computer Music Journal,
15(1), 1991.
[2] Matthew Lane. “Modular Progression Programming in OpenMusic”. In Actes des
Journées d’Informatique Musicale. Montréal, 2015.
[3] Olivier Messiaen. The Technique of My Musical Language. Paris: Alphonse Leduc,
1956. Trans. John Satterfield.
[4] Charlotte Truchet, Gérard Assayag, Philippe Codognet. “OMClouds, a heuristic
solver for musical constraints”. In Metaheuristics International Conference. Kyoto,
2003.
[5] James P. Womack, Daniel T. Jones, Daniel Roos. The Machine That Changed the
World: The Story of Lean Production—Toyota’s Secret Weapon in the Global Car
Wars That Is Now Revolutionizing World Industry. New York: Free Press, 1999.
75
Musicalising sonification:
Image-to-music conversion using
OpenMusic
Luiz Castelões
David Cope [4] defines sonification as the production of sound “by translating (called nor-
malising), in one way or another, traditionally non-audible data to the audible frequency
range.” He goes on to say that “[s]ince the data typically used for data sonification has
little if any inherent musical logic, the normalised output from sonification generally has
little musical use” (p. xi, italics are mine).
Departing from Cope’s definition, as sonification becomes more and more widespread
and trivial (concerning the technologies involved) the compositional challenge at hand
shifts from the act of converting between different realms (image-to-sound, number-to-
sound), which characterises any given sonification, to effectively obtaining consistent,
original musical results through such operations.
Strategies for turning sonification into music (and finding possible intersections be-
tween musical logic and other logics in the context of sonification) have been a major
field of interest for me as a composer and scholar and will constitute the main focus of
this article, particularly with regard to image-to-music conversions.
Introduction
Image-to-sound conversions (in a broader sense than image-to-music conversions)1 con-
stitute a widespread type of sonification. One could even propose that a modern/contem-
porary tradition of image-to-sound conversion for compositional purposes began devel-
oping even prior to computer-aided composition—a tradition that includes Scriabin,
Villa-Lobos, Messiaen, and Xenakis. I will use the word “conversion” to refer to this
type of sonification, instead of Cope’s “translation”, because I believe that the latter
implies some shared meaning between the data to be sonified and its corresponding
1 Even if, in a general sense, music can make use of any sound (Cage’s “anything goes”), that is not
the same as to state that music is any sound for every composer and every musical work. In practical
terms, the amount of effort and intent that most composers put into their compositional processes
reveals that their music seeks certain sounds (potentially different for each composer, or each work), a
much more specific territory than simply any sound (“whatever”, “n’importe quoi”). This is also valid
for improvisatory, aleatoric, and “chance” approaches, which in their own ways also produce certain
sounds as opposed to any sound.
77
Luiz Castelões
sounding result (which is not necessarily the case), while the former is more neutral with
regard to obtaining a shared meaning. Also, throughout this study, whenever image-to-
sound sonification results in music it will be called “image-to-music conversion” so as to
differentiate it from Cope’s understanding of sonification (as having “little musical use”).
Clearly there is not a unique solution to the problem of obtaining music from image-
to-sound conversions, but rather multiple strategies that can serve musical solutions:
not least because of the premise that music is culturally based and individually based.
Therefore making music may constitute a different task (and, as such, imply different
methods) for different (groups of) individuals.
Following the model of a case study, I will present some compositional strategies for
making music out of sonification that have been employed in two original works: the elec-
tronic soundtrack of VIA (2013) and 3 Transcrições (2011) for solo piano. These works
highlight two separate aspects of the aforementioned compositional problem: contour-to-
music conversion and colour-to-music conversion. Although this investigation focuses on
describing a list of strategies for obtaining image-to-music conversions in these concrete
compositional examples, it is also part of a broader perspective on how to musicalise
(humanise, personalise) sonification.
2 Headed by semiotician João Queiroz and dancer Daniella Aguiar and begun in 2013. For more
information on VIA please visit the website: http://via-locativeart.com/.
3 Following Joshua Fineberg’s invaluable advice, I always transcribe CAC (Computer-Assisted
Composition) or CGA (Computer-Generated Assistance) results by hand, a slow process that however
allows for an intimate control of the final results in the score.
78
Musicalising sonification: Image-to-music conversion using OpenMusic
These preliminary decisions left me with three possibilities for a sound source:
(a) sound synthesis (via OpenMusic and CSound, for instance); (b) exporting the
generated OpenMusic scores to other software in order to generate sound; or (c) using a
sound preview provided in OpenMusic to render the musical results (typically via MIDI
playback).
The possibility of assembling bits of ready-made timbres (a sort of collage approach)
seemed closer to the project’s approach to video-dance, based on editing short images
of dance taking place on the streets: with pedestrians passing by, pigeons flying around,
street vendors, etc. So (a), sound synthesis, was discarded for its usually more abstract
approach, whereas (b), OpenMusic scores exported to other software, and (c), MIDI
rendering, were tested numerous times using hundreds of conversions. Eventually most
of the results used in the project’s final version were the product of the built-in MIDI
rendering within the OpenMusic environment.
4 Given that each MIDI note often only contributed a tiny “brick” to the overall result, the task is much
like “instrumental synthesis”, albeit carried out with MIDI samples instead of acoustic instruments—
instrumental synthesis taken back to the studio, so to speak.
5 Fora reader growing up in the 1980s or 1990s, such sounds (or their earlier counterparts) already
sounded old-fashioned then!
79
Luiz Castelões
be presented as a final music product, but only as a sketch in the early stages of any
compositional process, their use as music represents a personalised, distorted way of
exposing the compositional tool at hand.
In spite of all of these arguments, to my ears most of these MIDI sounds were simply
unbearable either sustained or in isolation. Therefore it was necessary at least to develop
algorithms to avoid two (un)musical situations: solos and sustained sounds.
Avoidance of solos, here taken to mean timbres played alone or with significant
prevalence over other timbres, was addressed and implemented by means of two simple
steps analogous to traditional music approaches: instrumentation (the choice of sound
sources, in this case particular sets of MIDI presets) and orchestration (the specific use
one makes of these sources). I structured the OpenMusic Instrumentation patches so
that any image-to-music conversion would use 15 different instruments.6 In rough terms,
15 is large enough to avoid undesirable solos. This is easily feasible by assigning different
MIDI programs (or instruments) to the different MIDI channels (see Figure 2).
The Orchestration sub-patch, in turn, divides the total pitch range in different slices
(“bands”, or registers), then assigns subsets of three MIDI channels (or instruments)
to each range-slice, and randomly chooses one MIDI channel/instrument of each subset
6 The MIDI standard provides 16 different channels, but one (channel 10) is usually reserved for
percussion.
80
Musicalising sonification: Image-to-music conversion using OpenMusic
for each note. This is similar to basic acoustic orchestration (e.g. if a certain pitch is
within the range of a flute, an oboe, and a clarinet, then it might be played by any of
them). Several different versions of orchestration sub-patches were developed and tested.
Figure 3 displays a version that segments the total pitch range into five different registers.
Figure 2. Left: sending out instrumentation settings to the MIDI system (one MIDI program
for each channel). Right: inside the Instrumentation sub-patch (random selection of programs).
The Instrumentation sub-patch is locked when an interesting combination of MIDI instruments
is reached.
Figure 3. Left: Orchestration sub-patch controlling the MIDI channel assignments in Figure 1.
Right: Inside the first omloop.
81
Luiz Castelões
In order to address the second point, avoiding sustained MIDI sounds, I systematically
favoured small values for note durations as I normalised the image contour values.
From a more global standpoint, given that resonances are not particularly attractive
in these sounds, I decided to adhere to a music of attacks by limiting myself to the very
beginning of each MIDI sound, its “richest” portion. The most straightforward way to
implement this in OpenMusic is by assigning a short maximum value for <ldur> (the
slot controlling the durations) in the chord-seq. Figure 4 illustrates the use of the Durs
sub-patch, scaling down the score durations according to a small maximum value.
Figure 4. Setting the durations from image contour data. (N.B. this figure hides the other
connections to the chord-seq shown in Figure 1).
7 In
“Estudo Capanema” from VIA, audio available at:
https://soundcloud.com/lecasteloes/estudo-capanema-2013-para.
82
Musicalising sonification: Image-to-music conversion using OpenMusic
Figure 5. Scaling onset values in VIA (e.g. to a total duration of 33 seconds). (N.B. this figure
hides the other connections to the chord-seq shown in Figure 1.)
Image-to-music relationships
The implementation of these algorithms in OpenMusic was carried out so as to allow
for maximum flexibility in the normalisation of the image data, like freely compressing
or expanding a picture before printing it. In musical terms, this means that I was able
comfortably to change normalisation values as I converted each photograph as many
times as I wished, and hear the corresponding musical results, until I was fully satisfied.
These at-times-subtle changes in the normalisation values should not be underesti-
mated because they often have a great impact on musical parameters (harmony, timbre,
melody, gesture, texture, tempo, etc.). In this way, they may be the most essential factor
(along with the above-mentioned instrumentation and orchestration strategies) in ob-
taining music from sonification—so much so that no further compositional manipulation
of the input data (such as cutting, looping, superimposing, etc.) was necessary in order
to obtain musical consistency in VIA.
In the end, hearing is the decisive factor in VIA’s image-to-music conversion method.
It determines whether (1) there should be changes in the instrumentation and orches-
tration sub-patches, (2) there should be changes in the normalisation values, or (3) a
musically consistent result has been reached, in which case the process comes to an end
(see Figure 6).
83
Luiz Castelões
Music-to-video relationships
In the early stages of our project I discussed with VIA’s team the best way to apply
music to video-dance for our purposes. We quickly rejected the possibility of adopting
a previously composed soundtrack (totally independent from dance and images) or, on
the other hand, a soundtrack that strictly followed the dancer’s bodily movements—
“Mickey Mousing”. Instead, we chose to convert the visual contours of photographs from
the locations where the dance had taken place. I then redrew these visual contours “by
hand” (using the computer mouse) by placing each photograph as a background picture
in a bpc (break-point curve) object, but always allowing a certain degree of freedom
(Figures 7 and 8).
Figure 7. Redrawing visual contours in a bpc (photograph from Beco das Cancelas).
This freedom was a means not only to personalise the whole process (as another person
would inevitably redraw it in a different way) but also to obtain consistent musical results:
as I repeated the task over and over I gradually began to perceive general connections
between the intricate visual contours and the corresponding resulting sounds, thereby
allowing myself to interfere musically in the conversion process—a consequence of what
Mikhaïl Malt calls a solfège de modèles, which in this case would mean forehearing the
potential musicality beneath the images.8
8 “Qu’entendons-nous par ‘solfège de modèles’? Nous ne parlons pas d’un solfège au sens d’un catalogue
de modèles statiques, soit d’un solfège issu d’une typologie figée. Nous parlons plutôt du développement
84
Musicalising sonification: Image-to-music conversion using OpenMusic
Figure 8. Redrawing visual contours in a bpc (photograph from Museu de Arte do Rio).
d’un savoir, de capacités intellectuelles et cognitives de la part du compositeur qui lui permettraient
soit de contrôler et maîtriser le résultat musical issu d’un modèle génératif quelconque, soit d’établir
consciemment le lien entre des représentations graphiques et/ou textuelles de certains logiciels musicaux
avec un résultat musical. Solfège au sens d’aptitude à relier le comportement de deux espaces de
caractéristiques différentes en tenant compte des particularités de chacun.” [5]
85
Luiz Castelões
intended to reduce the amount of data, including the use of only the opening musical
excerpt that resulted from a given image-to-music conversion, or the simplification of
the image contour done by hand in such a way that the conversion generated less
data. In conclusion, creative work using image-to-sound conversion implies adjusting
the magnitude of image-derived data to the magnitude of the intended musical resulting.
(3) If this magnitude is “complex”, or at least “complicated”, how does the composer
relate to this complexity, and what are the best ways to obtain it and orient it toward
an intended musical goal? VIA’s space of musicality ideally sought a level of musical
complexity comparable to handmade crafts or to human activities (even though, perhaps
paradoxically, the work’s music was entirely realised by means of a computer). In
VIA, the strategy for obtaining the desired complexity consisted in starting from data
that was already complex and working from the macro- to the micro-level, instead of
obtaining complexity by working all the way up from small building blocks. We did not
seek complexity per se, but rather the kind of complexity that possessed musical logic.
Nevertheless, it is necessary to remark that:
- The totality of the characteristics that make up such (musically logical) complexity
have not been thoroughly formalised through the image-to-music algorithms devel-
oped for VIA and would constitute an incommensurable challenge; instead, such
desired complexity is simply “encountered” and “chosen” by ear among the several
musical results furnished by the converting algorithms. This complexity is partially
analysed, understood, and managed, and these stages contribute to the continuous
refinement of the original algorithms, an ongoing and potentially endless work in
progress.
- In VIA the composer does not control the whole process from the start, but
rather exerts gradual and partial control over the encountered, and chosen by ear,
complexities generated by image-to-sound conversions.
- The production of musical results in this case is therefore significantly open to
unforeseen structures.
- The final musical result is as much, or even more, “found” than “created”—in
this context, music composition becomes the intellectual-auditory discovery of a
pre-musical object, an abstract object with a highly musical potential.
(4) How does the magnitude of music making relate to independently developed
images of choreography, as sound and image are united within an audio-visual product?
There is certainly a re-signification that takes place as the finished musical component,
which—let us suppose—was satisfactory enough when heard alone, joins the video com-
ponent, which in turn contains the images that served as abstract data sources for the
musical component. In VIA there was limited space for feedback between music and
image; however exchanges of early mixes allowed us to learn about the particular twofold
association between music and video as well as between music and dance. One of the
main conclusions drawn from this dialogue was that music and choreography seemed to
be more clearly related when the musical component was dense and intricate. In early
mixes that had soundtracks with more empty spaces, more silences, the artificial aspect
of the audio-visual collage became more apparent and it looked and sounded like an
unintentional mistake, a technical flaw. This finding has further motivated the search
for complexity within the musical component, as previously delineated in (3).
86
Musicalising sonification: Image-to-music conversion using OpenMusic
(5) How can one obtain human-like musical results while making music exclusively
through a non-human device such as a computer? How can one make computer music that
sounds as if humans had performed or even improvised it? Here the computer’s role as a
musical tool is emphasised, but without allegiance to a machine æsthetic: one in which the
“watermark” [1] of the employed technology becomes too apparent, rendering the musical
work that employs it obsolete even before the technology’s own programmed obsolescence.
Strategies for obtaining a human-like yet machine-made sonic profile include:
- ongoing refinement of the musical problems’ formalisations and the implementation
of algorithms that provide human-like musical results (which inevitably involve
sound-to-sound intensity variation and frequent tempo fluctuation);
- emphasis on complexity (the more dense and intricate the algorithm output is, the
less often one perceives the small-scale limitations of the technology employed);
- the search for asymmetrical rhythmic sequences, obtained effortlessly and at will
as one copies an image by hand (or with a computer mouse);
- the search for varied instrumentation models (e.g. with 15 different musical instru-
ments) combined with the use of aleatoric orchestration, which simulates spontane-
ity and surprise; and
- orchestration by registers, simulating instrumental chamber music practices.
87
Luiz Castelões
88
Musicalising sonification: Image-to-music conversion using OpenMusic
Figure 10. Examples of colour-to-music conversions. The same image is converted (a)
using RGB: R=midicents, G=onsets, B=dynamics; (b) using HSV: H=midicents, S=onsets,
V=dynamics; (c) using CMYK: C=midicents, M=onsets, Y=dynamics.
89
Luiz Castelões
Figure 11. Overview of the patch used for the second movement of 3 Transcrições. Selection
of four horizontal lines from the picture and colour-to-music conversions.
90
Musicalising sonification: Image-to-music conversion using OpenMusic
Figure 14. chord-seq of the beginning of 3 Transcrições used for musical transcription.
91
Luiz Castelões
References
[1] Rodolfo Caesar. Círculos Ceifados. Rio de Janeiro: 7 Letras, 2008.
[2] Luiz E. Castelões. “Conversão de imagens para sons através de três classes do
OpenMusic”. Revista DAPesquisa, 8, 2011.
[3] Luiz E. Castelões, Talita De Oliveira, Yago Franco. “Conversores de parámetros del
color a parámetros sonoros cuantificables usando los sistemas RGB, HSV y CMYK”.
Sonic Ideas, 7(14), 2015.
[4] David Cope. “Preface”. In Jean Bresson, Carlos Agon, Gérard Assayag (eds.) The
OM Composer’s Book 2. Editions Delatour France/IRCAM-Centre Pompidou, 2008.
[5] Mikhaïl Malt. “Concepts et modèles, de l’imaginaire à l’écriture dans la composition
assistée par ordinateur”. In Bruno Bossis, Anne Veitl, Marc Battier (eds.)
Musique, instruments, machines: Autour des musiques électroacoustiques. Actes
actes du séminaire du MINT (vol. 2). Musicologie, Informatique et Nouvelles
Technologies/Observatoire Musical Français/Université de Paris-Sorbonne, 2006.
92
On “slow” computer-aided
composition
Julien Vincenot
Introduction
I began to consider myself a composer relatively late in my life as a musician, first
approaching the discipline through electroacoustic composition and improvisation. I now
count myself among the increasing number who learned composition with the help of
computer music tools, and CAC in particular. I completed my studies in the composition
Cursus of the Conservatory of Montbéliard, France with Jacopo Baboni Schilingi and
Frédéric Voisin. In this unusual context, CAC was not merely one more tool to master—
something “on the side”—among all those that composers need to know today. Within
this class we were encouraged to think all together with CAC. The software, PWGL
in this case, was not only a “toolbox” for the composer, but a hyper-expressive means
to learn how to conceive and describe our musical ideas and, above all, a way to share
them. Patches were for us a medium to exchange and confront ideas, and to solve musical
problems lying in front of us that we might not have been able to express in words at
first.
CAC made us entirely reconsider our way of thinking about detail while creating
music: written detail does not necessarily have to be realised by the composer because
its properties can be declared algorithmically, modelled before it even exists. On the
other hand, much more space and time is left for compositional detail: to make artistic
decisions on a multitude of time scales and musical parameters, as is more often the case
in electroacoustic composition.
Some may see in this a crutch, a “fig leaf”. On the contrary, I consider CAC a
prosthesis, the “augmented arm” I need to extend my level of consciousness, to multiply
my abilities for thinking, and to realise my musical ideas. In some situations making a
specific patch is the only way to help me better understand an idea I ultimately write
93
Julien Vincenot
out by hand. Without the bridge that CAC offered me I never could have reached the
level of confidence and freedom I feel today with musical writing; indeed my music would
hardly exist.
In the following sections I present some developments I have made with PWGL and
OpenMusic1 since 2009, which are gathered today in the jv-components library.2
1I discovered computer-aided composition with PWGL, and was immediately seduced by the possibilities
of constraint programming. During my year in IRCAM’s Cursus, I felt instantly on familiar ground
when I discovered OpenMusic, as both OpenMusic and PWGL share the same family tree with
PatchWork [2]. Today, I continue to use PWGL on a daily basis, for reasons of habit and comfort,
but also because constraints are fundamental to the way I work. Yet I find OpenMusic especially useful
to work in conjunction with other IRCAM software such as Audiosculpt, for instance to extract data
from SDIF files, and it is now natural for me to have both programs in my toolbox.
2 This library was originally a part of my masters research on musical morphogenesis and transposition
of non-musical materials into music with the help of CAC [8]. Some of these applications are already
available in OpenMusic; the rest of the sources are completely open and can be adapted to any Lisp-
based environment.
3 Emblematic examples are Esquisse (and later OMTristan, initially developed by Tristan Murail and
Laurent Pottier in PatchWork, then ported to OpenMusic by Rozalie Hirs); Combine developed by
Brian Ferneyhough and Mikhaïl Malt for PatchWork, then OpenMusic; or Profile initially developed
by Jacopo Baboni Schilingi, Nicola Evangelisti and Mikhaïl Malt for OpenMusic, then adapted and
augmented as JBS-Profile in PWGL.
94
On “slow” computer-aided composition
Figure 1. Interpolations of two chords in five steps with different list ordering: (60 62 67) to
(49 60 70) vs. (62 67 60) to (49 60 70).
If we take the example of lists of values evolving through time, representing for
instance melodies or more complex objects, it is even more clear that the forms of the
start and endpoint are not taken into account during the process. Even if the notion of
an element or point in those lists is clearly defined by the algorithm, this is not the case
for the relations between these elements (in the Gestalt sense [6]), for possible groupings
that we might take for granted, and especially for functions (musical, formal, etc.) that
these elements can hold in our perception within a given object. Interpolating such forms
in the standard way amounts necessarily to making a sort of “cross-fade” between them
(see Figure 2).
The first requirement to achieve a convincing metamorphosis is to find how to match
characteristic points or features—holding particular functions, including perceptually
salient features—of two distant morphological structures.4 For instance, we might want
to transition from an elementary form, containing few salient features, to another one
much more complex and articulated. In this case it is necessary to determine which zones
precisely will be matched and which details can emerge in a more unpredictable way (see
Figure 3).
The first tool I present addresses this issue. It is an algorithm for supervised interpo-
lation inspired by morphing techniques in image processing. The most obvious example
is a transformation of one human face to another that maintains, throughout the steps
of processing, consistency in the location of the different parts of the face—contributing
to the realism of the result. This coherence is ensured by the fact that different features
(eyes, nose, ears, etc.) are preliminarily identified in the images by a supervising operator
who remains active during the process.
4 Comprising all the properties that can used to describe the identity of an object as well as its form
(elements and relations between them, functions, contrasts, directionalities, intensities, etc.) and that
can be identified at various scales in space and time.
95
Julien Vincenot
Figure 2. Simple interpolation of two curves, regularly resampled, in seven steps. OpenMusic
patch (top) and the seven interpolation steps (bottom).
In our case, dealing with bpf profiles, we need to start with a first step of seg-
mentation. This step can be achieved in different ways: automatically—unsupervised
classification, detection of transients or unstable zones, etc.—or by hand, in an arbi-
trary or exploratory way—by putting markers along the objects we want to morph (see
Figure 4).
96
On “slow” computer-aided composition
Figure 4. Placing markers along profiles for segmentation prior to interpolation (for this
example, both profiles are cut into 11 segments).
At this point the profiles can be interpolated according to the raw segments that are
defined by the markers. But it is also possible, and even more interesting, to gather
some of these first-level segments into subgroups. The interface I developed in PWGL
(shown in Figure 5) allows one to compose, through the use of parentheses around indices
representing the segments, the zones that will be put into correspondence during the
interpolation process. This operation is indispensable when the two initial profiles of
the interpolation contain a different number of segments: in this case the problem can
be bypassed by creating an identical number of subgroups. Therefore it is possible for
the user to elaborate precise, realistic metamorphoses or totally unexpected ones by
determining in parallel the nature of the segments and the groups to which they belong.
Figure 5. Placing markers along a profile for segmentation and grouping segments prior to
interpolation (in PWGL). In this example, a profile is cut into 11 segments then regrouped in 5
parts. The two bpfs at the bottom show the second and fifth groups, gathering segments with
indices (1 2) and (7 8 9 10) respectively.
The engine of the supervised interpolator that I developed then iterates on the
successive subgroups determined by the user. Two standard interpolators5 are employed
to deal with both the form and the length of the profiles through different phases
of resampling. Each step of the metamorphosis is then flattened—the subgroups are
discarded—and the result can be visualised directly as a set of bpfs.
5 Using the well-known interpolation function from PatchWork, OpenMusic, and PWGL.
97
Julien Vincenot
In Figure 6 we can see two possible interpolations made with this approach, corre-
sponding to different groupings of the segments in the same pair of profiles. The first, on
the left, might be more obvious if we take into account the relative similarity of features
in both profiles. In the second most of the target profile emerges, progressively, from a
single segment of the starting profile. Both propositions are valid in the sense that they
were actively “composed” with the system.
Figure 6. Two possibilities of metamorphosis based on the same segmentation but with different
groupings.
I experimented with this process for the first time while writing Ascidiacea for solo
bass flute (2010), whose formal skeleton strictly follows the metamorphosis of a relatively
simple figure—a re-encoding, as a bpf, of a musical phrase written by hand—toward a
more complex morphology extracted from arbitrary non-musical material.
In this context, my approach to the profiles was a bit different from what we might
be used to when manipulating such graphical objects: the amplitude of the curve was
not taken as the continuous expression of a given musical or sonic parameter, but as a
discrete representation of contrast inside a musical object (see Figure 7). The value 0
represents a given category of musical materials sharing some common aspects; value 1
another category, and so on. After evaluating the metamorphosis itself intuitively, as
shown previously, I discretised all the successive steps into a number of categories.6
The piece strictly follows the plan given in Figure 8, which must be read one line
after the other: the first line being the initial phrase (shown in Figure 7), the second
following it immediately, and so on. The plan is an interpolation between the phrases 1
and 7. The connected fragments, belonging to different phrases, are shown by vertical
dashed lines. Similar categories (obtained after discretisation of the profiles) are labelled
with the same letter. Starting from this, I determined the nature of the different musical
materials, their evolution throughout the piece, and their mutual relations: resemblance,
discontinuity, amplification, dissipation, budding, hybridation, corruption, absorption,
etc.
6 This operation was performed with the class-num (previously class-1 ) function from the Morphologie
library. This clustering function, similar to k-means, allows the classification of a given number of points
(in any number of dimensions) by optimising progressively the position of n centres whose quantity is
determined by the user. For instance, if I want to classify the list (8 5 8 4 5 2 3 2 9 10) into three
categories, the function will easily stabilise itself around the following classes: (0 1 0 1 1 2 2 2 0 0).
98
On “slow” computer-aided composition
iv
Ascidiacea
morphological creature for bass flute
Julien Vincenot
ó ñ ó ï ó ï
“anomaly” only key clicks
? ??
bisb. più forte possib.
???
bisb. ad lib.
(timbral trill)
(slow) (moderate)
q ! 50-60 (fast) 3 fingerings 2 fing.
, ≈
pizz. (molto secco)
q!
œo o
freely alternate between idem
n. vibr. p. vibr. n. vibr.
o
pitch and air sound
nœ #œ
& Lœ œ™ ÍÍÍÍÍÍÍÍÍÍ œ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍ œ O #O nœ ÍÍÍÍÍÍÍÍÍ ˙ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍ œ ‰
Jœ nœ œ™ ˙ B> #> n> O ¿ ¿ ¿ ¿ ¿ ¿ ¿ ¿ ¿
ÆJ
bœ b>
Æ
Bass Flute
œ >
> > >
ppp p mf p ppp p mp o
3
2 2
1 1 1
ó ï ñ ó ï
0 ¿
bisb.
?? ¿ 0
?? ? ???
µ ≈
bisb. (slower possib.)
, œo
æ ≈ its re-encoding
2 fing. 3 fing.
# œo
sempre frullato
6
œæ ≈ œ #œ O
inhaling 5 n. vibr. vibr. n. vibr.
™ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍ œ ÍÍÍÍÍÍÍÍÍÍprofile.
Ϫ
r
2
#œ
bœfi œ™
#œ
#. #. #. #. #œ
œ ÍÍÍÍÍÍ œ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍ
5
& Figure 7.#œ nœ Initialj phrase #œ of Ascidiacea for bass flute and
œ œ œ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍ œ™ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍ œ ÍÍÍÍÍÍ œ ÍÍÍÍÍÍÍÍÍÍÍÍÍ œ™
as a nœcontrast O
J
The gesture with pizzicati was added later to the score as an intentional
p pp p
3
mp
“anomaly”. The final poco
pp mp pp
sfz
silence seems also to contradict the properties of class 0—a choice I cannot remember or explain
today.
ñ Phrase 1ï ñ ó
only key clicks
Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
ff
??? ? ??
bisb.
ad. lib.
, , ≈ ,
p. vibr.
q™!
vibr. n. vibr.
°
# œo
frullato
#œo ææ ææ #œo
3
&
12 seconds
nœ ÍÍÍÍÍÍÍÍÍÍ ÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍÍ œ ÍÍÍÍÍÍÍÍÍÍÍÍÍ œ™ <µ œ> ‰ #œ
#O nœ #O œ™ œ < µ œ> ¿ ¿ ¿ ¿ ¿ ¿ ¿ ¿ ¿ ¿
>
#œ
µœ
#œ œ #O bœ #. #. #. #. #. #. #. #.
RÔ >
mf o poco
sfz
p ppp
poco
sfz
pp
fij
t. ram
ì ì
pizz.
Phrase 2
¢& 16 seconds
¥
> nO
mf #
ff
Phrase 3
19 seconds
Phrase 4
14 seconds
Phrase 5
14 seconds
Phrase 6
22 seconds
Phrase 7
23 seconds
Figure 8. Formal plan of Ascidiacea, showing the connections between similar classes of musical
materials throughout the piece.
The elements of the initial sequence—gestures in grace notes, sustained notes, har-
monics, key clicks—were therefore used as a starting point to determine the material of
the following phrases. The different instances of materials belonging to the same class
were consequently written with similar instrumental techniques. Beside these primary
elements, others appeared as the initial figure became more complex, giving me the
possibility to imagine other categories of materials or to hybridise existing ones. Outside
of these recurring categories a certain number of “anomalies” were added locally, not
directly related to the structure given by the interpolation.
99
Julien Vincenot
Follow a drawn
& œ
œ #œ œ œ œ #œ
profile (heuristic)
? œ œ œ # œ œ œ œ # œ
œ
& œ #œ œ #œ No repetition
? œ œ œ # œ n œ œ #œ n œ œ #œ (heuristic)
Allowed intervals
œ
& œ #œ œ œ
? œ œ #œ n œ #œ œ n œ œ #œ #œ (1 2 5 7, up or down)
Figure 9. Several melodic generations computed with the PWGL constraints solver (PMC
engine) adding rules progressively.
7 PWConstraints consists of two engines, PMC and Score-PMC. Both were developed for
PatchWork [7] then expanded into PWGL. PMC was brought to OpenMusic by Örjan Sandred
as part of the OMCS library.
100
On “slow” computer-aided composition
is, to try to understand something one might want to hear but without knowing how to
write it: only how it should and should not be. In this sense constraints are exploratory
means to create new musical materials. But the true challenge was to formalise new rules,
to learn the language to write them (in this case an interface in Lisp), and above all to
realise the logic behind the very notion of a rule. This means understanding the difference
between musical properties (of single elements) and relations (between elements) and how
to constrain both on a local, detailed scale, and on a more global, morphological scale.
The jv-components library contains numerous examples of rules I designed over the
years. Some of them concern interval structures, others control metronomic modulations
(inspired by the works of Elliott Carter and José Manuel López López). A few of
them focus on instrumental idiomaticity, allowing a given result, already constrained by
many musical rules, to adapt its complexity to the technical restrictions of one specific
instrument: maximum spacing between fingers, impossible notes, multiphonic chaining,
etc. The last category of rules I investigated are morphological rules, which allow one to
generate results that reproduce the global behaviour of another musical object or model,
with a variable degree of similarity, after the extraction of its morphological properties.
Why is it necessary to valorise these techniques today? The first reason is they are
yet not well known and especially not taught in composition classes. Second, the speed
of curent machines trivialises a number of usual techniques to the point where they are
now computable in real time. On the contrary, using constraints, especially as the level of
exigency increases, remains relatively incompatible with real time and relies intensively
on machine resources. Despite an interactive dimension—with constant feedback between
the definition of the search space, the setting of rules, the result, and its perception—
constraints still allow access to a form of “slowness”, generating material with a “rarity”
which otherwise seems to have disappeared. This rarity, even if it is controlled by the
user who defines the rules to be applied, often leads to unexpected and surprising results.
Another aspect of constraint systems worthy of interest is that they can help to
understand and control the notion of musical style, whether it concerns music from the
past, today’s music, or especially one’s own style. In particular it is possible to extract,
from a given musical entity, not necessarily the rules that were used to write it but the
rules that might have been used.8 This kind of knowledge can be precious to understand
the writing techniques and formalisations that a composer learned or designed him or
herself—his or her craftsmanship—and what resulted from accident and artistic intuition.
The extraction of rules from a given musical object is also a powerful asset for music
composition. Of course, in the field of creation, it may not be much more interesting
to generate hours of music “in the manner of” than it was previously to do it by hand.
But it is more expressive to pull such rules out of their initial musical context and apply
them to a new one where the properties they convey are more meaningful. This applies
also within the work of a composer. I noticed for instance, from one piece to another,
that I had a tendency to reuse a certain number of rules in specific settings while others
disappeared after a while. Rules have become a way to remain consistent within my own
work, even while keeping for each new project a significant margin for experimentation.
8 The information might be lost, or maybe the composer had no conscious notion of any rules while
composing. In this respect analysis is necessarily a posture of construction of a specific point of view,
or even a posture of creation.
101
Julien Vincenot
In my work today the privileged system for musical exploration and composition with
constraints is the Cluster-Engine library conceived by the composer Örjan Sandred.
It succeeds his previous libraries OMRC (for OpenMusic) and PWMC (for PWGL),
themselves following premises from Laurson’s PWConstraints. This new engine is
remarkable for various reasons. It allows one to control several musical voices in parallel
by using a network of semi-independent constraint engines—one for each parameter of
each voice, including metrical structure, pitches, and rhythms—that can solve a musical
problem collaboratively. Therefore the rules, through the use of accessors of various
scopes, can be applied to each one of these parameters, between parameters within a
single voice, and of course between voices. Backtracking 9 procedures were also optimised,
allowing a spectacular gain in speed and flexibility compared to previous approaches. As
Sandred explains:
The intricate network of dependencies that exist in a score makes it hard to develop
a good strategy for how to find an acceptable solution. After having struggled
with some of the existing constraint solving systems, it became necessary for me
to develop a new, more efficient constraint solver that could handle more complex
musical formalisations. I needed a system that could approach the problem from
an angle where the nature of a music structure is taken into account. [10]
An example of this awareness is found in the vast possibilities left to the composer while
determining the search space. Here it is constituted by separate domains, one for each
parameter. This way we can define which time-signatures, notes or chords, or durations
values are allowed. But much more interesting is the possibility to insert already well-
formed musical entities or motifs. A list representing a rhythmic motif, a pitch motif, or
even an intervalic motif (freely transposable by the engine) might be, as much as single
isolated values, picked up and controlled by rules thus reinforcing the emergence of a
musically consistent result.
In Figure 10 we can see the generation of a simple result in PWGL using the cluster-
engine for one voice with such motifs. Only one rule is applied, r-only-m-motifs provided
by the library, which makes sure that only transposable pitch motifs will be selected to
build the result (except for the initial starting point, MIDI note 60 in this case). The
rhythm domain includes two isolated values (1/4 and -1/8) and five different motifs.
The pitch domain defines a mandatory starting point and five transposable pitch motifs,
starting with the “m” symbol and followed by a series of intervals. It is interesting to pay
attention to the last two motifs that each comprise only one element, a list of intervals.
This allows one to define transposable chords, which could also appear within a larger
list of intervals. For instance the motif (m (-2 5 5)) starting from MIDI note 60 would
yield, after a descending major second, the chord (58 63 68).
I used this engine intensively for the composition of my recent pieces silent_data
_corrupt for alto saxophone and live computer (see Figure 11)10 and Mémoire de l’eau
9 Backtracking is the capacity for a constraint engine to move backwards during computation and rewrite
a part of a solution that might contradict the given rules. Within the Cluster-Engine, backtracking
was optimised by the use of backjumping: instead of erasing the solution step by step to try and find
the error (classic backtracking), the engine “jumps” directly to the variable which failed a specific rule,
corrects it, and then resumes the computation.
10 silent_data_corrupt(2014) was presented at the end of my year in IRCAM’s Cursus, in collaboration
with saxophonist Nikita Zimin.
102
On “slow” computer-aided composition
Figure 10. Rhythm and pitch domains using motifs (in PWGL).
for accordion and live computer (see Figure 12).11 In both cases I had in mind to generate
virtuosic passages inspired by improvisation in general and free jazz in particular. To
reach this result I paid specific attention, even before defining the rules, to the elaboration
of complex search spaces that comprised freely transposable melodic motifs (and chords
in the case of the accordion). In particular, I carefully designed the representations
of my transposable chords, allowing efficient and idiomatic chaining for the accordion
player. Already at this point, without defining any rule, the results were promising: the
candidates were already determining a musical context as fully-fledged rules would do.
Figure 11. Excerpt from silent_data_corrupt for alto saxophone and live computer.
Figure 12. Excerpt from the introduction to Mémoire de l’eau for accordion and live computer.
11 Mémoire de l’eau (2015) is the result of a collaboration with accordeonist Yohann Juhel.
103
Julien Vincenot
The second set of rules was dedicated mostly to rhythm, and also controlled some aspects
of the pitch engines relative to the rhythmic context:
[...] a measure of dissimilarity, based on the search for the minimal cost of the
transformation of one sequence into another, by means of the following three op-
erations applied step by step: substitution of one symbol by another, insertion of
a new symbol, deletion of a symbol. Thus, for example, the edit distance between
the list (C D E G A G G) and the list (E D E G A G A G) equals 3, the minimal
number of operations needed for the transformation. [9] (author’s translation)
12 Morphologie was initially developed in 1997 by Frédéric Voisin and Jacopo Baboni Schilingi for
OpenMusic, then recast for PWGL and any Lisp environment in 2009.
See http://www.fredvoisin.com/spip.php?article28
104
On “slow” computer-aided composition
How does this apply concretely to our problem, the metamorphosis of musical entities?
The original edit-distance function of the Morphologie library, following the algorithm
as expressed by Levenshtein, only returns the minimal cost to transform one sequence
into another but does not specify the successive steps of the process. This is the purpose
of the function s-interpolation 13 from the library jv-components, which can deduce the
possible path. In most cases there are different possible minimal paths among which the
result is selected randomly. An exemple of a path between the two words “AMPHORA”
and “AMORPHA” is as follows:
((AMPHORA)(AMOPHORA)(AMORPHORA)(AMORPHRA)(AMORPHA))
It is interesting to note that going from one word to the other or back can produce slightly
different results. This is a consequence of the order of operations in this implementation,
which always starts from the beginning of each sequence:
((AMORPHA)(AMRPHA)(AMPHA)(AMPHOA)(AMPHORA))
This algorithm allows one to move from one sequence to another in a relatively
convincing way, whether acting on pitches, durations, or any other parameter. The
result is all the more relevant, from a perceptual point of view, as common elements are
maintained between the successive steps. The interpolation is “informed” by the simple
nature of the algorithm: it does not add, delete, or replace elements by accident but
takes into account both the content and form of the initial sequences.
I used this simple approach to generate progressions between different melodic frag-
ments for Mémoire de l’eau (see Figure 13). Unlike numeric interpolation, this approach
does not allow one to decide a priori how many steps will be returned between the
two sequences: the number of elements is necessarily equal to the edit distance plus 1.
I therefore manually selected afterwards—as one often does with classic interpolations—
the steps I considered to be more musically interesting and relevant.
The advantage of a purely symbolic approach is the possibility to morph sequences
containing any kind of musical materials, as long as they can be encoded as symbols
(melodic or rhythmic motifs, chord sequences, etc.) or sequences of symbols representing
more complex materials at a higher level. All depends on the encoding chosen to represent
these materials or forms. And if we use the appropriate predicates for testing equalities,
even lists of any length and nesting level become symbols in their own right.14
To this approach I have applied the concepts that were already expressed in my
algorithm for supervised interpolation dedicated to numeric profiles (see Figures 5 and 6).
13 Thisfunction is a translation I made from Javascript to Lisp of the algorithm proposed by Jean-Paul
Davalan on his website. The section dedicated to the Levenshtein distance is particularly informative
and allows one to experiment with the algorithm directly from a web browser.
See http://jeux-et-mathematiques.davalan.org/
14 The = function in Lisp will accept for its arguments only values of type number (mathematical
equality). Equal or equalp, on the other hand, will allow us to check for instance whether symbols or
lists (even nested ones) are identical or not. Lisp is particularly well-suited to express any properties
or relations in the form of custom predicates. The jv-components library contains some proposed
predicates based on morphological properties. For instance, the simple function close-p tests if the edit
distance between two lists is inferior to a given threshold; that is if these two lists are “close enough”
to be considered approximately equal.
105
Julien Vincenot
Figure 13. Melodic process using symbolic interpolation (Mémoire de l’eau for accordion and
live computer). The different steps of the interpolation are identified by beaming.
106
On “slow” computer-aided composition
Figure 14. PWGL patch for the supervised symbolic interpolation of two melodic profiles.
distance paradigm), we need as a last step to resize each according to the longest
one. This implies repeating some steps between certain pairs, in the most homogeneous
manner possible, in order to obtain exactly the same number of steps for each process.
Figure 15 shows a simple example of such a metamorphosis between to melodic profiles,
highlighting the different parallel interpolations.
This approach, as mentioned before, can be applied to any kind of musical parameter
or even to complex encodings aggregating multiple parameters. In the specific case
of rhythm the results are already promising, perceptually speaking, especially when
interpolating simple lists of durations. Still, in the metric domain the process often
107
Julien Vincenot
Conclusions
Within the field of computer music today, computer-aided composition as we know it is
going through a phase of transition with the emergence of new tools oriented toward real-
time, interactivity, and performance like the bach library for Max [1], OpusModus,15
or the new “reactive mode” in OpenMusic [4]. All of these recent innovations are
stimulating and we cannot yet imagine the applications that will result: especially for
the development of open/interactive scores and forms in composition, and for interactive
installations in particular.
At the same time, for those who believe in musical writing, and especially in a possible
association of writing with technology, it is necessary not to allow the disappearance of
the knowledge and techniques accumulated on the side of deferred-time in the domains
of composition, analysis, and formalisation in general. For me the applications of tools
like OpenMusic or PWGL and the reasons to keep them in our arsenal are numerous.
First of all is the possibility to work in parallel with various modes of representation
(scores, bpfs, maquettes, sound files, etc.) and to organise all of them inside the
topological environment of the patch. There is also the flexibility and expressivity of
the Lisp language which, even with its age, allows custom modes of representation
and manipulation that are still unattainable with a Max patch, for instance. And
finally there is the capacity of these tools fully to exploit the machine’s computing
resources to run complex (and often richer) musical algorithms, and not only a graphical
interface. For my own work, CAC is still a privileged way to introduce materials
derived from sound or extra-musical models in composition, to try to solve the conflict
between algorithmic complexity and perception of the listener, to favour expressivity and
instrumental idiomaticity without sacrificing complexity and experimentation, to link
instrumental and electronic composition, and even to think of instrumental composition
as electronic composition.
One might remember that these tools, which were designed for the fast temporality
of music production, are still among the rare tools allowing a focus on slow temporality:
that of composition—but also of programming, interpretation, analysis, and critique—in
other words, of musical thinking.
15 http://opusmodus.com/
108
On “slow” computer-aided composition
References
[1] Andrea Agostini, Daniele Ghisi. “Real-Time Computer-Aided Composition with
bach”. Contemporary Music Review, 32(1), 2013.
[2] Gérard Assayag, Camilo Rueda, Mikael Laurson, Carlos Agon, Olivier Delerue.
“Computer-Assisted Composition at IRCAM: From PatchWork to OpenMusic”.
Computer Music Journal, 23(3), 1999.
[3] Antoine Bonnet, Camilo Rueda. Situation. IRCAM documentation, Paris, 1999.
[4] Jean Bresson, Jean-Louis Giavitto. “A Reactive Extension of the OpenMusic Visual
Programming Language”. Journal of Visual Languages and Computing, 25(4), 2014.
[5] Grégoire Carpentier. “Global Constraints in Orchestration”. In Charlotte Truchet,
Gérard Assayag (eds.) Constraint Programming in Music. London: Wiley-ISTE, 2011.
Online
109
OM-Darwin: Generative and
descriptive aspects of genetic
algorithms
Geof Holbrook
111
Geof Holbrook
Figure 1. Typical OM-Darwin patch: generation of a chord sequence on the left, description
of desired properties on the right.
For our purposes, the search space is often all possible musical sequences of a given
type, with some limits on dimensions and ranges. An example of a search space might
be “15 chords with cardinalities between 3 and 5, and a pitch range of 4800 to 8400
midicents”. In OM-Darwin, I use the term species to indicate the definition of a search
space. The library provides built-in species that represent things like melodies, chord
sequences, rhythm trees, and polyphonic passages. Custom species can also be defined
by the user.
The fitness function is a measurement of a single candidate, returning a numeric value
that represents its suitability or “fitness”. The GA makes no restrictions on the nature
of this function—it can measure any property that the user wishes to control, as long
as the output is expressed as a number. The convention in OM-Darwin is that higher
numbers are interpreted as worse than lower numbers, and a fitness of zero means the
candidate is perfect (the fitness can been seen as a penalty, or distance from an ideal
state). So, the GA searches for solutions that give as low values as possible when passed
to the fitness function. If it finds one that gives a value of zero, it stops.
Each iteration of the GA might be called a “generation” in which the population
is duplicated several times with slight alterations, sorted by fitness, and then cut back
to its original size, preserving the most suitable solutions. These slight alterations are
made either by mutation (a small change to one detail of the object) or by recombination
(combining two existing candidates into a single new one). At any given time there is
a current best solution, and the current best fitness will improve over time, especially
after hundreds or thousands of generations. It can, perhaps, be seen intuitively that this
improvement will occur; but to put it more exactly, it occurs because the system has
three required features: transmission, variation, and selection [4]. For a more in-depth
discussion of genetic algorithms and evolutionary computing, see [2].
112
OM-Darwin: Generative and descriptive aspects of genetic algorithms
In Figure 2, we can say that the two rightmost inputs (the leftmost <self> input of
ga-engine is seldom used) represent the generative and descriptive sides of the algorithm,
although there is not much in the way of generative processes going on here. Each solution
is simply a sequence of chords, with dimensions and ranges given as arguments to the
built-in make-ga-chord-seq function. The fitness function (detail in Figure 3) makes a
measurement of the sensory dissonance of each chord, using a function written by Sean
Ferguson [5]. It returns a sum of differences between the measurement and a goal value,
provided as an argument. Running the GA will produce solutions whose dissonance
values get progressively closer to the goal value, as this sum of differences approaches
zero.
This example is instructive, as it demonstrates two common features of fitness func-
tions for musical objects: 1) the application of the same measurement over the elements
of a given structure (chords, in this case) and 2) a measurement of proximity to a
desired value (the absolute difference between the actual and target values). Because
these and other kinds of descriptive strategies are common, the library includes a set
of helper functions called criteria. There are different criteria for iteration over notes,
chords, melodic intervals, simultaneously sounding pairs of notes, etc. They can be given
arguments for expressing various kinds of static or dynamic goal values, specifying rates
of occurrence, or supplying a custom measurement function to be applied to individual
elements. Criteria can be combined by joining them into lists, which amounts to adding
their outputs to determine an aggregate fitness value. Looking back at Figure 1, one can
see the dissonance function combined with some other measurements, expressed using
the built-in functions, to create an aggregate fitness function.
113
Geof Holbrook
114
OM-Darwin: Generative and descriptive aspects of genetic algorithms
values can be easily made into a species, by simply replacing all instances of om-random
with the OM-Darwin function om-gene, and replacing instances of nth-random with
nth-gene. Expressed as a function (lambda mode), and passed through define-species,
this patch can then be used to set the species argument for ga-engine, in the same way
as a built-in species (see Figure 4).
What define-species is doing, in short, is running a test on the function given to it.
It counts the number of calls to om-gene (and nth-gene), recording the inputs as well
(representing ranges of values, just like om-random). Because the convention in OM-
Darwin is that genotypes are lists of integers of a fixed length, it is important that the
patch be designed so that the calls to om-gene and nth-gene are made in a predictable
order. This is a necessary limitation of the library.
This establishes a crucial division between the generative and descriptive parts of the
algorithm. If the phenotype is defined in such a way that a certain property is guaranteed,
then that property results from a generative procedure in the mapping function. Other
properties, which have to be achieved through adaptation, are the desired outcome of the
descriptive part of the system. Seen this way, the composer has two ways of controlling
properties of the output: phenotypic controls and adaptive controls.
Take the example in Figure 4. The pitches are expressed in the genotype by their
integer rank within a harmonic series, with the mapping function translating these into
actual pitches; so, the property of belonging to a harmonic series is a phenotypic control.
That means that all candidates in the search space have this property; there is no way
for mutation or recombination to alter the fact. Other properties are alterable by the
GA if they are dependent on the specific harmonic ranks chosen—a measurement of such
115
Geof Holbrook
properties can then have an impact on the selection part of the algorithm. So, if we wish
also to control the melodic intervals of the sequence (for example), we include that in
the fitness function, making it an adaptive control.
The reversal of the assignment of these two properties to phenotypic and adaptive
controls is also possible. We might write a generative function that chooses melodic
intervals within a certain range, but then measure the membership or non-membership
of each note to a harmonic series in the fitness function. This means that there are at
least two approaches in this case, and the user must decide which is more efficient, either
by intuition or trial and error.
Figure 5. The generator patch for s-oboes. The genes encode, from left to right along the top
of the patch: long note value, scale note values, starting note, scalar intervals, scale length, and
scale direction.
116
OM-Darwin: Generative and descriptive aspects of genetic algorithms
3 3 + 7
! ! # %! ! ! ! $ ! $! ! ! ! !
3 7
! ! ' ! % ! #! " ! !
3 3
$ ! ! $! %! ! ! '! $! ! % ! %!
* 43 !
140
( &
3
, )
,
7
# # " 5
! % ! %! +
3
# , , ! !"
3
/
5
! ! ! ! ! ! ! '!
3 3
* % !# '! ! 3 ! ! % ! !. -
143
! ! !) !
, 4 , 5
, %! %! # !3
3 5 3
" ! , ! !
5 5
! % ! % ! %!
146
* 0 # - 4 # ,
5
#" - 3
3
7 5
5
One of the instances of om-gene encodes the length of a scalar run, which can be
as long as nine notes; another is called multiple times to encode the melodic intervals.
Note that the latter instance is always evaluated nine times per melodic unit—but if
the run is shorter, then the extra information is discarded. If it were evaluated only
as many times as needed for each scale, then the genotype would be of different length
for different candidates, which would be nonsensical to the ga-engine box. There is a
safeguard against this: define-species makes multiple test evaluations of the generating
patch and signals an error if the genotype is not the same length every time.
The single input to this patch allows the user to decide on the number of units the
assembled melody will have. However, the ranges of the genetic functions could also be
specified as inputs to the patch. Then we would have a whole set of parameters separate
from the genotype, as we often do, which we might call species parameters. They would
be part of the definition of the species for any given trial, and could not be altered by
mutation or recombination.
117
Music engraving by LilyPond 2.19.26—www.lilypond.org
Geof Holbrook
Although the GA did not arrive at a perfect solution, the effect of these colliding
criteria is powerful enough to give the desired effect. In the best-case scenario, the GA
outputs a perfect result (a fitness of zero) in a reasonable amount of time. But it is
possible, as in this case, that a satisfactory result is not necessarily perfect.1 For a
given problem, it is actually possible that no perfect solution exists, which means that
the global optimum fitness will be some positive value. In such cases it is generally
unnecessary to find out what that value is, as there are likely many solutions which are
close enough to the optimum fitness to be acceptable, or even “new and interesting”.
However, it is also possible that the GA will fail to make significant progress toward
a decent solution. In this case there are a number of steps the composer can take: 1)
restructure the fitness function, 2) rewrite the mapping function, or 3) redistribute the
phenotypic and adaptive controls.
The distribution of work between generative and descriptive programming seems to
be entirely appropriate in the case of s-oboes, although one could imagine the phenotypic
and adaptive controls being reversed. Some method that uses diatonic fragments, along
with a pitch mapping applied to 80% of the notes, could be used to assemble a phenotype
that automatically adheres to the rules given above. Conversely, a fitness function can be
devised that measures the pattern of long and short notes and limits changes of direction
to give the quick passages a scalar quality.
1 “The goal is to identify new and interesting solutions—normally more than one is desirable—and these
solutions must be good. However, finding global optima may be undesirable, impractical, or even
impossible” [3].
118
OM-Darwin: Generative and descriptive aspects of genetic algorithms
References
[1] Carlos Agon, Moreno Andreatta. “Modeling and Implementing Tiling Rhythmic
Canons in the OpenMusic Visual Programming Language”. Perspectives of New
Music, 49(2), 2011.
[2] Peter J. Bentley. Evolutionary Design by Computers. San Francisco: Morgan
Kaufmann, 1999.
[3] Peter J. Bentley. “Exploring Component-based Representations—The Secret
of Creativity by Evolution?”. In Ian Parmee (ed.) Evolutionary Design and
Manufacture (Selected Papers from ACDM’00). London: Springer, 2000.
[4] Richard Dawkins. “Universal Darwinism”. In D. S. Bendall (ed.) Evolution from
Molecules to Men. Cambridge, UK: Cambridge University Press, 1983.
[5] Sean Ferguson, Richard Parncutt. “Composing in the Flesh: Perceptually-Informed
Harmonic Syntax”. In Proceedings of the Sound and Music Computing Conference.
Paris, 2004.
119
Computer-aided composition in the
creation of As I ride the late night
freeways
Matthew Schumaker
This article presents a composer case study documenting the use of OpenMusic software
in the compositional processes of As I ride the late night freeways, a work for soprano and
orchestra. The introduction lays out the work’s æsthetic environment, considering Cathy
Park Hong’s poem Census and how its theme of freeway driving can suggest composi-
tional approaches and parameters that unify the musical work. Architect Zaha Hadid’s
parametric æsthetic is also considered as a “streamlining” influence. The following section
focuses on the development of the musical processes, which take a variety of forms: from
harmonic and textural elaborations drawn from the analysis of car sounds, to a poetic
suggestion of aerodynamism, given by the algorithmic shaping of musical gestures and by
the smooth morphing of one gesture into the next. Other related OpenMusic patches
illustrate how these ideas are being developed further for use in future works.
1 Composedin large part during 2014-2015 in Paris, where I was supported by a Georges Ladd Fellowship
from the Music Department of the University of California, Berkeley.
121
Matthew Schumaker
122
Computer-aided composition in the creation of As I ride the late night freeways
Figure 2. Extracting frequency contents from the car sounds using the chord-sequence analysis.
Two approaches were used for each of the sound sequences: one to create static
overarching pitch fields; the second to create chord sequences that suggest the harmonic
transition within each sound. The sound files were all processed using the chord-seq-
analysis function from OM-pm2 (see Figure 2). Due to the noisy character of these
sounds, a “standard” spectral analysis would necessarily report a large amount of chaotic
2 DaitoManabe’s exciting project Sound of Honda - Ayrton Senna 1989 - (2013), influenced my choice to
use Ayrton Senna driving recordings. See http://www.daito.ws/en/work/sound_of_honda_senna.html
123
Matthew Schumaker
The chord sequences obtained from the partial-tracking analyses were also used to
recreate the frequency transitions and the Doppler effect of the car sounds in orchestral
string textures. The extracted chords were exported to a notation program, divided into
different lines of the divisi string section, and given durations roughly proportional to
their duration as components in the overarching sound. Glissandi were often employed
to connect these chords in order, giving the impression of the continuous frequency tran-
sitions in the aggressive acceleration or deceleration of the original car sound. Additional
sculpting of dynamics and the insertion of occasional rhythmic accents and tremolos
contribute to the orchestral emulation (see Figure 4).
The music in the opening section of the piece derives its harmony from chords found
in the analyses of the seven race car sound edits. Here the chords unearthed in the
partial-tracking analyses are freely combined to form a new ordering (Figure 5). This
new sequence, in turn, plays an important structural role in the piece, appearing in
different guises as a sequence of musical pillars that signal the beginning, middle, and
ending of the work.
124
Computer-aided composition in the creation of As I ride the late night freeways
Figure 5. “Pillar” harmonic sequence freely derived from the seven race car sound analyses.
125
Matthew Schumaker
A predetermined pitch field, derived from the initial analyses of the car sounds, is
loaded into the patch as a MIDI file that was processed to yield a list of sorted pitches.
The om-scale and om-round objects are then used in series to map the samples from
the bpf into values between the lowest and highest pitches of this sorted pitch list. A
chord-seq (visible at the bottom left of the figure) provides a preliminary display of the
transcription of the bpf into a musical line. After this, the notes-change object of the
Profile library refines the pitches in this preliminary chord-seq by finding the closest
pitches from the chosen pitch field. Each sampled pitch is then given a 32nd -note duration
in a voice object to produce the final musical line.
The process above created a significant musical statement, referred to as the “aero
line” (Figure 7). From here, the musical texture is thickened through imitation of the
line in another voice with a certain delay. The two lines are then combined and displayed
in a poly object where they can be previewed and exported as MusicXML files into a
notation program. Once in the notation program, these lines were distributed between
orchestral wind instruments to make the sequence.
After establishing the aero line, I chose to elaborate it through a series of variations.
One straightforward way to do this was to make circular rotations of the list of points in
the original bpf line. To do this, I routed the values from the original line into a rotate
function, which was also given a random seed number from 0 to 100. The resulting
random circular permutation of the list, as before, was sampled against the given pitch
field using the notes-change function. An imitative line again was generated with a delay
for the start of the second voice. This process generated new lines that were, nonetheless,
still linked to the contours and proportions of the original aero line. Seven variations
were made through trial and error along with other variations created from a retrograde
of the aero line.
126
Computer-aided composition in the creation of As I ride the late night freeways
127
Matthew Schumaker
This morphing process takes a source melodic fragment and a target melodic fragment
as inputs, loaded into the patch as separate MIDI files. The getpitch sub-patch then
collects the pitches from each fragment into separate lists, converted into bpf objects that
become the inputs to the “morphing” process. The bpf-interpol function morphs between
the two lines, creating a series of interpolated curves between the source bpf and the
target bpf. Using this function, the composer has some control over the smooth transition
of the overall interpolation in two ways: first, by setting the number of interpolated curves
and, second, by using an interpolation exponential factor which distorts the resulting
interpolation process toward the starting or ending lines as desired.
Pitches are assigned to these lines in two different ways. The MIDI pitches in the
intervening curves resulting from the interpolation were in some cases used “as-is” with
a simple rounding to the nearest equal-tempered pitch. In other cases, where tighter
control was desired, a Lisp function (closest-pitch-match) was used to ensure that all
resulting pitches were derived from either the target or source lines.
The interpolation sequences are then notated musically in voice objects. While the
pitches of the interpolations are automatically derived from the process, the rhythms of
128
Computer-aided composition in the creation of As I ride the late night freeways
the interpolated lines are specified by the composer. In some cases, a uniform rhythm
in the resulting lines was desired in order to draw attention to the smooth unfolding
of the interpolation in the pitch domain. In these cases, one rhythm list was specified
and applied to all the curves. In other cases where the rhythmic profiles of the first line
and the second line differed significantly, a list of rhythm lists, one for each curve, was
supplied. The idea of smooth transitioning from one rhythmic profile to the next was a
leading concern in intuitively drafting these lists. As a final step, the individual voice
objects representing the musical sequences of the interpolation are joined together using
OpenMusic’s concat function. Figure 10 shows the result of an example interpolation
process.
129
Matthew Schumaker
lists. These rhythm lists grow gradually from a short sequence of 5 notes to one of 21
notes. The last rhythm list is taken from the original MIDI file and corresponds to the
full length of the aero line.
Having defined the list of attack times for this growth process, what remain to be
determined are the corresponding pitches. A custom Lisp function, choose-set-pitch-or-
random-pitch, was made to carry this out. This function takes four lists as inputs: two
pitch lists and two lists controlling probabilities. Each of the four lists, in turn, has as
many elements as there are attacks in the rhythm lists. For each rhythm list, or “motif”,
one list of pitches is created that retains the pitch ordering from the original aero line.
A second pitch list is generated that contains a random sampling and ordering of pitches
taken from the whole of the line. The first of the probability lists is made from a master
bpf object that shapes the probability that a given attack will come from the first pitch
list, the one that preserves the recognisable ordering of pitches in the aero line. For every
attack in the rhythm lists, the bpf returns an increasingly high number up to 100. In
contrast, the second probability list always generates a random number between 0 and
100 for each attack in the master rhythm list. For each note in the sequence then the
Lisp function compares a value in the bpf probability list with the random probability
list. If the value of the random probability list is the higher number, then the pitch from
the corresponding random pitch permutation list is chosen; otherwise, the pitch from the
list preserving the aero line ordering is chosen. Since the values of the bpf probability
list become increasingly higher, as the process continues it becomes more and more likely
that pitches preserving the aero line ordering will be chosen. The perceptual result of
the finished line corresponds nicely to the desired goal—the aero line appears to emerge
almost imperceptibly from a chaotic mass of pitches.
130
Computer-aided composition in the creation of As I ride the late night freeways
This process was used in the piece primarily in an extended section written for divisi
strings. The resulting motifs are intuitively overlaid and imitated to create a web of
polyphonic sound divided amongst the different string voices. In addition, the form of
this section is elaborated and extended by concatenating sequences of different processes.
For instance, this process of creating the aero line from random pitches is quickly followed
by one of the morphing processes described earlier, now enabling the string lines to
transition smoothly from the newly formed aero line to a single, pulsing high pitch.
Figure 12 (a-b) presents corresponding sections of the score.
131
Matthew Schumaker
[...]
[...]
Figure 12a. As I ride the late night freeways, mm. 330-333, 356-358, and 362-365 (showing
only the process in the string parts).
132
Computer-aided composition in the creation of As I ride the late night freeways
Figure 12b. As I ride the late night freeways, mm. 376-385 (showing only the string parts).
133
Matthew Schumaker
References
[1] Eric Daubresse, Philippe Hurel. “Creating Polyphonic Sequences in the Form of
Rhythmic and Melodic Canons”. In Jean Bresson, Carlos Agon, Gérard Assayag (eds.)
The OM Composer’s Book 2. Editions Delatour France/IRCAM-Centre Pompidou,
2008.
[2] Patrik Schumacher. “The Parametricist Epoch: Let the Style Wars Begin”. AJ—The
Architects’ Journal, 231(16), 2010.
134
Materials and techniques in
D’improvviso da immobile
s’illumina for bass clarinet, two
orchestras, piano, and percussion
Federico Bonacossa
This paper describes how the computer was used to inform large-scale aspects as well
as to generate much of the pitch and rhythmic materials of D’improvviso da immobile
s’illumina (2013), a concerto for bass clarinet, two orchestras, piano, and percussion. In
particular it will discuss how OpenMusic was used to suggest possible transformations
of melodic ideas composed intuitively and to create a rhythmicon based on the frequency
information gathered from the analysis of a bass clarinet sample and the various com-
promises that had to be made along the way.
While composing D’improvviso da immobile s’illumina I attempted to reconcile two
seemingly opposing approaches to composition: a purely intuitive one, representing the
fascinating ability of our subconscious to naturally distill and combine elements from
the most diverse experiences, and a more “scientific” one, in which the material was
generated either by algorithms or some type of analysis with the help of a computer. The
two approaches influenced one another; the computer was used to generate variations of
the material created intuitively, and intuition guided a number of changes to the original
data collected.
Form
The form of the concerto is based on the shape of the waveform of the short sound
sample shown in Figure 1. The sample is a recording of a low B[ (58.2 Hz) played on
bass clarinet. In order to use the waveform as an outline for the form, I interpreted
the timeline of the sample so that each second would correspond to approximately one
minute of the piece. The envelope of the waveform provided an outline for dynamic
shape and density and consequently suggested a general plan for the use of instrumental
forces. I chose to divide the work into five main sections, corresponding more or less to
the various distinct stages of the envelope:
1. Introduction – bass clarinet solo.
2. Orchestral gradual crescendo.
3. Bass clarinet cadenza.
4. Short dialogue between bass clarinet, bassoons, and cellos.
5. Orchestral recapitulation.
135
Federico Bonacossa
Pitch material
The concerto’s pitch material can be reduced to two main pitch-collections. The first
consists of various set-theoretical manipulations of a basic hexachord. The second was
derived from the spectral analysis of the original sound sample from Figure 1. For the
sake of clarity, I will refer to the two pitch-collections as “Collection 1” and “Collection 2”.
The first step in the compositional process was to write the introduction for solo bass
clarinet. Rather than establishing a clear set of pitches or rhythms to use, I defined a
few general parameters:
Pitch centres: I chose D as the initial central pitch because it was one of the loudest
partials in the spectral analysis (shown in Figure 2) from which I planned to derive
Collection 2, and could act as a pivot point between the two collections.
136
Materials and techniques in D’improvviso da immobile s’illumina
Figure 2. Spectral analysis of the bass clarinet sample used in the concerto (first 13 pitches).
Pitches are rounded to the nearest quarter-tone. Note the dynamic levels.
Collection 1 was derived a posteriori from the five pitches used in the introduction.
The pitches were arranged to form “Hexachord 1” shown in Figure 4 by adding an
additional pitch (C5) to fill in the gap between A and D.
The prime form of hexachord 1 is the symmetrical set {0,1,3,6,8,9} (Forte number
6-Z29)[2].1 This set is inversionally symmetrical; its inversion corresponds to a transpo-
sition of the original set (T3 ). Hexachord 1 corresponds to T6 of the prime form. After
some experimentation with different transpositions of the set, I decided to combine two
transpositions, the original one (T6 ) and T9 , to generate a larger eight-note set that
forms Collection 1 (last measure in Figure 4).
1 Notethat in the Forte catalogue 6-Z29 is sometimes listed as {0,2,3,6,7,9}, but {0,1,3,6,8,9} is equivalent
and actually more compact. See for instance http://solomonsmusic.net/pcsets.htm.
137
Federico Bonacossa
The MathTools package in OpenMusic includes several useful objects for working
with sets. I created several simple additional tools (such as the abstraction midic-›set)
that I used in combination with the library to accomplish specific tasks. Figure 5
shows a patch that can quickly calculate prime forms and inversions. Figure 6 shows
a patch used to combine sets to form larger ones (in this case Collection 1). In order
to generate additional material I also used OpenMusic to calculate and notate subsets
and transpositions of the initial hexachord. This then served as a pool of material from
which I could draw when necessary (Figure 7).
Figure 5. Using the objects from the MathTools package to calculate the prime form and
inversion of a set. The n-cercle objects provide a graphic representation of the sets.
Figure 6. Transpositions of the hexachord and set combinations (generation of Collection 1).
138
Materials and techniques in D’improvviso da immobile s’illumina
The patch shown in Figure 7 allows the user to enter a set and calculate all chromatic
transpositions. This is a simple task, but the advantage here is the ability to notate the
results in a usable layout. The abstraction transp-chrom outputs directly to a voice object
and subdivides each transposition into measures of the appropriate length, which makes
it much easier to read (the user can also enter a beat value). This can then be exported
as a MusicXML file into notation software (or be printed directly in OpenMusic).
Figure 8 shows the same approach applied to subsets. In the patch the user can
specify the length of the subsets to notate. The function sub-power gives a list of all
the subsets of the indicated length. The abstraction notate-divid-chrom calculates all
chromatic transpositions of each individual set and notates them in the specified metre.
There are some slight but important differences between transp-chrom and notate-
div-chrom. In transp-chrom the internal loop simply adds each value generated from the
arithm-ser to the list of pitches. In notate-div-chrom each individual element of the list
is transposed chromatically before moving on to the next.
139
Federico Bonacossa
The octachord Collection 1 has several features that I find compelling and that, not
coincidentally, characterise the pitch material of several of my earlier compositions, for
example Rivir for solo viola, shown in Figure 9.
140
Materials and techniques in D’improvviso da immobile s’illumina
Figure 11. Some recurring sets derived from Collection 1 (see Figure 5).
In my recent compositions, I have often experimented with the expansion and con-
traction of sets. For this purpose I created two simple patches that enable me to try
out different transformations. The two patches are based on two different approaches:
expansion or contraction by scaling, and expansion or contraction by addition or subtrac-
tion of a constant value. The patch shown in Figure 12 is based on the first approach. It
allows one to scale the intervals by a factor and then see and hear the result. The scaling
factor can be a non-integer (Figure 12 uses a factor 0.5), in which case the results might
need to be rounded (om-round). The resulting set has the same contour as the original
one, starting at a specified pitch (in this case D4).
Figure 12. OpenMusic patch for expanding and contracting sets by interval scaling. A: input
a sequence of notes, B: scaling factor, C: starting pitch.
141
Federico Bonacossa
The patch shown in Figure 13 is based on the second approach. It calculates intervallic
expansions by adding or subtracting a specified number of half steps to each interval. In
contrast Figure 14 shows a short passage from the concerto consisting of an expansion
of the first tetrachord from Figure 11 (1) calculated using the interval scaling approach.
Figure 13. Half-step-expander. A: starting set, B: number of half-steps to add to each interval,
C: starting pitch of the resulting set.
Figure 14. Motivic expansion (intervals scaled by a factor of 3). D’improvviso da immobile
s’illumina, excerpt from m. 143, Bass clarinet.
142
Materials and techniques in D’improvviso da immobile s’illumina
Figure 15. Left: Spectral analysis of the sound from Figure 1 in Spear. Right: OpenMusic
patch used to filter spectral data.
Although this process can be done “manually” by reading the information directly
from the SDIF file, OpenMusic provides a flexible way to experiment with different
settings, and, most important, hear the results immediately. In other words, it allows
the composer to make decisions based on the actual sound.
In the patch shown, I specified the pitch range according to the instruments I planned
to use in the concerto (A). As most of the frequencies in the analysis are non-tempered
pitches, I rounded the data to the nearest quarter-tone (B) and then eliminated any
duplicates that might result (C). The resulting 36-note series is shown in Figure 16 (the
first 13 pitches were already shown in Figure 2).
Figure 16. Pitch collection derived from spectral data (Collection 2).
143
Federico Bonacossa
A parallel can be drawn between the ratio of rhythmical beats and the ratio of
musical tones by virtue of the common mathematical basis of both musical time
and musical tone. The two times, in this view, might be said to be “in harmony”,
the simplest possible. [1] (pp. 50–51)
Cowell proposes the application of the “principles of relating time to musical tone”
not only to durations, but also to metre, dynamics, form, and tempo. In the table shown
in Figure 17 he charts the rhythmic relationships inherent to a major triad that can be
derived from the mathematical relationship between the fundamental and all the other
tones. The first column indicates the tone’s location in the harmonic series. Columns two
and three indicate the interval of the tone in relation to the fundamental and the name
of the tone assuming C as the starting pitch. The last column shows the mathematical
ratios of the vibrations using 16 Hz as the fundamental frequency.
Figure 17. Rhythmic relationships based on the overtone series, from [1], p. 47.
Cowell stresses the similarities between “consonant” rhythmic relationships and con-
sonant intervals, according to how many repetitions of a tone were needed before it
would realign with the fundamental: “the smaller the number of units that must be
passed over before that coincidence is re-established, the more consonant is the interval.”
He also points out that the “the higher one goes in the overtone series, the greater the
number of units that must be passed over, and the further is the effect from one of simple
consonance.” The reason why the tones result in harmony instead of a “chaos of sounds”,
he writes, “is that at regular intervals the vibrations coincide”.
Cowell utilises the ratios of the intervals in the harmonic series to create a “rhythmic
harmony” so that, for example, a melody with a pulse of three half notes in the space of
two would correspond to the interval of a pure perfect fifth, five quarters in the space of
four to a major third and so on. He then gives various examples of their use, many of
144
Materials and techniques in D’improvviso da immobile s’illumina
them quite complex, admitting that most musicians would probably find them difficult
to perform.
Cowell also makes the important point that complex rhythmic relationships can be
easily performed by a player piano and even suggests the creation of an instrument that
would assign specific rhythms to each pitch (p. 64). His dream became reality through
support by Charles Ives, who commissioned the Russian inventor Leon Theremin to build
a rhythmicon, a machine that would convert pitch into rhythm and vice-versa. Theremin
built two models in 1932, effectively creating the first rhythm machine, but neither was
ever reliable enough to be used in performance [6]. In 1960 Theremin built another
rhythmicon while at the Moscow Conservatory or possibly the USSR Sound Recording
Institute. The instrument could perform complex polyrhythms such as 7 against 9 or 5
against 13. The model to this day is still in working condition [7].2
In New Musical Resources, Cowell always utilises ratios based on the overtone series
even though he realises that they are in most cases significantly different from the ratios
found in equal temperament. The reason for his choice is easy to understand if we assign
durations to two pitches a major third apart using just intonation and equal temperament
and compare the results. For instance, taking A4 as the fundamental, if we calculate the
duration of C]5 based on its frequency in the tempered scale (≈554.3653 Hz) and relate
it to the fundamental (440 Hz), we end up with an irrational ratio. (Even rounding the
top frequency to 554 Hz this ratio is still 277/220.) If, however, we use a justly tuned
C]5 (550 Hz),3 we obtain the much simpler ratio 5/4. In terms of durations, this means
that for every four repetitions of the fundamental there will be five repetitions of the
third (see Figure 18).
Figure 18. Durations derived from the ratio of the frequencies of an A4 and a justly tuned
C]5. The duration of the C] is essentially a quintuplet spanning four bars.
James Tenney uses a similar process in Spectral CANON for CONLON Nancarrow in
which, not coincidentally, he tunes the player piano to the first 24 partials of the harmonic
series [3].4 In this piece Tenney calculated durations for the piece by defining an arbitrary
initial duration k and calculating successive durations with ratios from the overtone series,
starting with the ninth partial (9:8 ratio). The details of the algorithm used by Tenney
are discussed at length for instance by De Paiva et al. in [4]. An implementation of the
algorithm is shown in Figure 19.
145
Federico Bonacossa
Figure 19. Algorithm used to calculate durations in James Tenney’s Spectral CANON for
CONLON Nancarrow, adapted from [4].
5 Aspreviously mentioned Tenney chose an arbitrary initial duration, although as de Paiva observes, 4
seconds seems to be an ideal length in terms of how it allows the relationships to be perceived [4].
146
Materials and techniques in D’improvviso da immobile s’illumina
results, although I knew that this would present some practical problems once the values
were converted to standard notation. The main patch, shown in Figure 21, also calculates
how many repetitions of the note would be necessary to fill 5 minutes (B). Figure 22 shows
a list of all midicent values with their calculated durations in milliseconds. Each result
was then stored as a voice (see the set of “instance” boxes at the bottom of the figure).
These voices were then consolidated into a single score in a poly object (Figure 23) and
exported to Finale.
Figure 21. Patch to assign a duration to a midicent value based on its frequency and determine
the number of repetitions needed.
Figure 22. List of pitches and durations used in the rhythmicon sections. Each column
represents the chord tone number followed by, in parenthesis, its pitch in midicents and duration
in milliseconds.
147
Federico Bonacossa
While Cowell advocated notating each part of a polyrhythm in its own metre to
eliminate the need for performers to “negate” the natural accent of the metre ([1], p. 70),
doing so in the concerto would have resulted in more than 20 different simultaneous
metres. I instead chose to notate all parts in the same metre so that the conductor
would at least be able to provide regular reference points for the performers.
The resulting score (Figure 24) was used as a reference to establish durations in
the actual piece. Because standard music notation is ill-suited for this sort of process,
and because the relatively high precision level I chose led to results that were often
unnecessarily complicated and impractical, I often simplified the notation to make it
easier to read and perform. Indeed, D’improvviso da immobile s’illumina was written for
a student orchestra so it was important for me to keep the notation relatively simple.
As an example let’s take the pitch A]4. Its frequency (466.1638 Hz) corresponds to
a period of 1000 / 466.1638 = 2.1451688 ms, yielding 2150 ms when multiplied by 1000
and rounded to the nearest multiple of 50 (1/20 of the beat). If we input this value into
the voice object we obtain the result shown in Figure 25 (A). The resulting notation
is not consistent and is also unnecessarily difficult to read. This problem could have
been avoided by rounding to a different value, for example 250, but I decided to use the
148
Materials and techniques in D’improvviso da immobile s’illumina
Figure 24. First measure of the raw score used to create the rhythmicon in D’improvviso da
immobile s’illumina.
149
Federico Bonacossa
same beat division for all pitches and make modifications as necessary if the results were
impractical. This particular pitch (A]4) was assigned to clarinet 1. I decided to choose
a simpler rhythmic notation that was fairly close in length to the original duration. The
duration I chose (2333 ms instead of 2150 ms) corresponds to 7/12 of a measure, two
beats and a third, which allowed me to make the notation easier to read while remaining
relatively close to the original length, as visible in Figure 25 (A)—the original—and
(B)—the new duration.
In my view this process is similar to rounding the pitches from the spectral analysis.
Naturally if we were to align both sets of durations the difference would become quite
noticeable within a few measures, but the result is still very effective. In addition to
simplifying the notation, I also periodically interrupted the pattern to ensure that the
performer could breathe regularly (but always for an full duration unit).
Additional operations
I ultimately decided to reduce the size of the chord to 24 pitches which allowed me to
split or reinforce voices when needed. I diverged further from the original results by
offsetting some of the entrances to avoid having the voices align too soon, especially as
the rhythmicon unfolded. Lastly, to make the parts more interesting I placed regularly
recurring accents in each voice; as a result, most voices have a macro as well as a micro
pulse (Figures 26 and 27).
> >
&Œ Œ œJ ˙ ˙ œœ œ œ ‰Œ Ó Ó Œ Œ œJ
J
f
3 3 3 3
Once I collected the raw material for the rhythmicon sections, I separated the chord
into several smaller sonorities and plotted the entrances carefully, leaving the entrance
of the fundamental tone last (Figure 28).
150
Materials and techniques in D’improvviso da immobile s’illumina
I then segmented the score into smaller sections and inserted between them material
based on Collection 1 so that the rhythmicon would only gradually become prominent.
Measures 59–70 of the concerto (see mm. 59–62 in Figure 29) are an example of how
I delayed the crescendo of the rhythmicon in this way. In m. 71 the rhythmicon is
interrupted by the percussion, which recalls mm. 4–7 of the concerto. The rhythmicon
becomes prominent at m. 91, continuing on for nearly 50 measures, its texture steadily
thickening until the fundamental tone of the chord is finally heard in the tubas in m. 117
(Figure 30), signaling the arrival of the climactic point. The low B[1 is played by
both tubas and is also doubled by double bass 2, while being reinforced one octave
higher by double bass 1 and two octaves higher by bassoon 2 and horn 1. By suddenly
becoming the loudest pitch, the fundamental tone provides a sense of meaning to the
otherwise chaotic texture, revealing a large-scale hierarchy that was hitherto disguised.
The instruments occasionally align with one another, but are otherwise asynchronous.
But when the rhythmicon comes back after the long cadenza in m. 176 all voices are
finally aligned, quickly dissipating again into an asynchronous texture. The return of the
chord structurally defines the beginning of the final section.
Conclusion
This article has demonstrated how OpenMusic may be used as a practical tool for
composers to try out different ideas and generate material while still leaving plenty of
room for intuitive choices. Often compositions have a way of taking on a life of their
own and naturally deviate in various ways from the original plan. This unpredictability
is one of the most fascinating and important aspects of a composer’s work.
151
Federico Bonacossa
152
Materials and techniques in D’improvviso da immobile s’illumina
153
Federico Bonacossa
References
[1] Henry Cowell. New Musical Resources. Cambridge: Cambridge University Press,
1958.
[2] Allen Forte. The Structure of Atonal Music. New Haven: Yale University Press, 1973.
[3] Larry Polansky. “The Early Works of James Tenney”. In Peter Garland (ed.)
Soundings 13: The Music of James Tenney, chapter: “XI: Spectral CANON for
CONLON Nancarrow”. Santa Fe, NM: Soundings Press, 1983.
[4] Charles De Paiva Santana, Jean Bresson, Moreno Andreatta. “Modeling and
Simulation: The Spectral CANON for CONLON Nancarrow by James Tenney”. In
Proceedings of the Sound and Music Computing Conference. Stockholm, 2013.
[5] Robert Wannamaker. “Rhythmicon Relationships, Farey Sequences, and James
Tenney’s Spectral CANON for CONLON Nancarrow (1974)”. Music Theory
Spectrum, 34(1), 2012.
Online
154
Musique instrumentale concrète:
Timbral transcription in What the
Blind See and Without Words
Aaron Einbond
Reproduction
Mimesis in music may date to the beginning of music itself, and its reach extends both
historically and geographically, from Plato’s Republic to Tuvan throat singing ([13],
p. 58). It is also central to common-practice music, whether the imitation of imitative
counterpoint, the programme of programme music, or the affect of the Baroque doctrine
of the affections. In the 20th century mimesis has been reframed by “technological
reproducibility”, as in Walter Benjamin’s landmark essay. In technology’s wake, Pierre
Schaeffer treats reproduction by juxtaposing “abstract” and “concrete” listening modes;
it is with their synthesis that he defines “reduced listening” ([2], p. 37). Composers
who combine instrumental and electronic technologies of reproduction include François-
Bernard Mâche and his “phonography”, Clarence Barlow’s “synthrumentation”, Gérard
Grisey’s “instrumental synthesis”, Trevor Wishart’s sonic “transformation”, and Peter
Ablinger’s “phonorealism” [15, 5]. Surveying these perspectives as well as those of the
early 21st century, Nicolas Donin writes: “‘composition’ [...] now includes as well the
navigation between different ‘reproducibilities’ through operations of translation such as
155
Aaron Einbond
Transcription
A potential definition of musical transcription is the projection of material from one
musical space to another—a space of frequencies or pitches, timings or rhythms, or other
acoustic parameters. Especially when working with noise-based instrumental sounds, a
suggestive metaphor is timbre space, a model of timbre as a multidimensional percept
[21, 8]. Such timbral dimensions have been associated with audio features, or descriptors:
“characteristics extracted from the source sounds, or higher level descriptors attributed to
them” [17]. The notion of the descriptor is the perfect technological correlate of reduced
listening: like Schaeffer’s phenomenological époché [2], the computer takes no account
156
Musique instrumentale concrète...
of the sound’s physical origins as it performs its analysis. Of course this does not apply
to the human listener, who must set the parameters of the algorithm beforehand and
interpret the results after, in both cases based on his or her own listening experience.
So rather than transparent measures of perception, timbral descriptors are contextually
sensitive materials upon which the composer necessarily exercises a subjective influence.
Decisions include what samples to use, how to segment them, what descriptors to analyse,
among many others. Far from automatic, timbral transcription can be a rich expressive
resource.
In What the Blind See for viola, bass clarinet, harp, piano, percussion, and electronics
(2009), sources for transcription include instrumental materials, physical gestures per-
formed with an electronic controller, and field recordings of rain, ice, and snow. The
process is termed Corpus-Based Transcription [7] as it takes advantage of corpus-based
concatenative synthesis (CBCS) with the CataRT package for Max, and transcription
with OpenMusic. While the details have been described elsewhere [7, 5], an example
can be outlined as follows: a large database of instrumental samples (the corpus, in this
case made up of samples of the five performers of the ensemble) is compared to a field
recording (the target, in this case an arctic glacier) using CataRT and a purpose-built
Max patch for analysis. After segmenting both corpus and target into short grains, and
matching those segments with the most similar descriptor values, the glacier recording
is reproduced with an audio mosaic of instrumental samples.
This mosaic is then used as a model for an instrumental score. The timings, filenames,
and associated data are stored in an SDIF file, which can be loaded by the OpenMusic
SDIFfile object. The data are extracted, organised, and rhythmically quantified in
OpenMusic (Figure 1), then the contents of the resulting poly object are exported
as a MusicXML file. This is imported into Finale and edited, with the aid of the
descriptors and metadata stored in the SDIF file, by annotating playing techniques,
dynamics, and articulations, as well as adjusting or simplifying rhythms and techniques
to facilitate playability. The resulting transcription is incorporated into the work as the
compositional process continues (Figure 2), and finally the ensemble performs the score
live, reinterpreting the transcribed field recording in concert.
In the finished work, the original field recording of a glacier is not performed, but
its shadow remains through the fusion of the ensemble into a colourful imprint. Even
without an explicit concrete referent, the individual timbral details and identities of each
instrument are subsumed into an “instrumental ambient”2 soundscape. At the same time
an association with other field recordings of rain and snow that are directly revealed later
in the work may retrospectively conjure a secondary denotation of the texture’s origin.
In other examples transcription sources are made more or less explicit, opening up a
powerful expressive resource with a decisive effect on perception.
One of the keys to this multi-valence is the simulation of reduced listening through
close miking, sampling, and amplification. In What the Blind See, contact microphones
are placed on the piano frame, piano soundboard, bass drum skin, and tam-tam, and
miniature lavalier microphones are used on the bridge of the viola, near the embouchure
of the bass clarinet, and beneath the soundboard of the harp. These microphones permit
the amplification of sounds distorted out-of-scale to their mode of production, flattening
157
Aaron Einbond
158
Musique instrumentale concrète...
The same microphones are used for recording the pre-compositional sample databases
as well as amplifying the instruments in concert, assuring the closest possible reproduc-
tion of the audio mosaic live. More important, the same musicians are called upon to
record the database as will premiere the work, in this case Ensemble L’Instant Donné,
and the imprint of their personalities and instruments becomes a collaborative part of
the compositional process. This process is not only “radically idiomatic”3 but radically
personalised. Like the audio descriptors themselves, the performers and instruments too
become sites of individual expressive variation, privileging the specific, concrete, and
personal over general, abstract, or universal.
While using OpenMusic to transcribe samples of a performer is not new,4 a signif-
icant difference here is the representation of noise-rich audio in its full detail without
recourse to the time-frequency grid of a spectrogram. To reproduce this sonic timbral
palette in symbolic music notation is a significant challenge. During sampling sessions,
performers’ playing techniques are documented with verbal explanations, photographs,
and video. When the sessions are edited into sample databases, the filename of each
sample is annotated by its playing technique in as much detail as possible, which is then
used to edit the score in Finale. Nonetheless the noise-rich timbres push the limits of
music notation: for many works the score is supplemented by sending the performers
a selection of their own audio samples. The goal is to trigger their sonic and physical
memories, so the symbolic notation serves as a mnemonic for personal oral history going
back to the sampling session.
Notably, for many playing techniques, pitch may not be a salient feature of either
the sound or its notational representation. Regardless, CataRT estimates the pitch
of each grain using the yin∼ object in Max. However the yin∼ algorithm may not
give a reliable estimate, or independently, the user may decide not to weight pitch
in constructing the audio mosaic. When the mosaic is imported to OpenMusic, the
estimated pitch is used in the poly display merely as a placeholder for each grain. As
the score is edited into its final form this pitch information may be retained, adjusted,
or eliminated entirely according to the playing technique needed. Comparing Figures
1 and 2, the viola pitches are replaced by x-shaped note heads (pizzicato on dampened
strings) and harp pitches by rectangular note heads (fingernails sur la table), while those
of the breathy low bass clarinet and marimba with Superball mallets are retained with
adjustments and elaborations.
At the heart of the OpenMusic patch omquantify quantises the durations of grains
in the audio mosaic, expressed in milliseconds, into notated rhythms (in Figure 1,
omquantify is included inside the make-voice sub-patch). Significantly, this use of detailed
rhythmic notation does not have a metrical function in itself, despite its “complex”
appearance. As such it could be analogised to an elaborated form of graphic notation
where the metrical grid serves a reference purpose only. Precedents include scores of
György Ligeti, Gérard Grisey, and Tristan Murail.5 But while the notated metre is not
intended to be heard directly, the quantisation parameters are chosen finely enough to
159
Aaron Einbond
Gesture
A physical gesture rather than a field recording may be used as a target: in another
example from What the Blind See, a Wacom tablet was used to “improvise” a short ges-
ture mapped directly to the Max lcd object that CataRT employs as a two-dimensional
representation of the sample corpus. As in the previous example, the exported SDIF
file—including sample timings, textual metadata, and descriptor values—was imported
and transcribed in OpenMusic (Figure 3) and a MusicXML file was exported to Finale
for further editing (Figure 4). The gesture was drawn to connect the approximate spectral
region of a preceding passage, based on an audio mosaic of rain, to other material that fol-
lows, with the goal of a smooth timbral transition. The interlocking timbral associations
that result—including nails and fingers on the harp sur la table, plectrum glissando along
the low piano strings, and wire brushes on the vibraphone—would have been difficult to
conceive with pencil and paper alone. Indeed the smooth trajectory in timbral parameters
does not necessarily match a symbolic pitch trajectory: although short segments comprise
rising scales or arpeggios, they do not connect the instrumental registers of the ensemble
in the most predictable way. Hearkening back to Lachenmann, the re-contextualisation
of extended instrumental techniques in a timbre-driven composite gesture stimulates the
composer’s—and listener’s—imagination unconventionally.
Figure 3. OpenMusic poly transcribed from a gesture improvised with a Wacom tablet.
160
Musique instrumentale concrète...
Transformation
A third example is taken from the opening of What the Blind See, where transcription
is applied to instrumental sample targets themselves instead of “extra-musical” sources.
By taking a short instrumental figure from one instrument and transcribing it in varied
form for another, the technique extends a classical form of musical imitation: motivic
transformation. In this case, the audio recording of one instrument, the viola, is taken as
the target for an audio mosaic using the samples of another, either the harp or vibraphone
(Figure 5). This is closer to Lachenmann’s musique concrète instrumentale than musique
instrumentale concrète, as there is no “extra-musical” referent. However by transforming
an acoustic target, rather than symbolic music notation, associations between materials
including rhythms and playing techniques may suggest themselves beyond the composers’
notational habits, conditioned and distorted by personal and historical practice.
Even more than in the previous examples, this one introduces details of the perform-
ers’ collaborative interpretations on several levels. The viola part is derived from the
earlier work Beside Onself, performed and recorded from the notated score by the violist
of the ensemble. So the performer’s interpretation of the written material, complete with
shades of timbral variation and expressive microtiming, becomes the target of the audio
161
Aaron Einbond
mosaic of harp or vibraphone samples. The transcription then takes into account these
variations, as well as the variations of attack time and responsiveness of the viola playing
techniques. The resulting transcriptions differ in significant detail from the original
notated viola part, but preserve its rhythmic and timbral flavour. This transcription takes
advantage of a further layer of rhythmic choice: the OpenMusic patch is used to filter the
audio mosaic output by CataRT according to a range of successive loudness thresholds,
producing transcriptions of different rhythmic densities. The final harp and vibraphone
parts in Figure 5 are drawn freely from these different rhythmic layers, permitting greater
control over the result.
Figure 5. What the Blind See mm. 11-21, showing the viola part transcribed for harp and
vibraphone.
c Edition Gravis Verlag, Brühl, Germany. Printed with kind permission.
Of course all three parts, the viola target and harp and vibraphone mosaics, are
subject to similar variations in timbre and microtiming when reinterpreted in live per-
formance. Performed simultaneously in the finished work, the result is a homophonic
doubling of the motivic material that fuses in terms of timbre and density. The live
electronics (indicated by the pedal marking for cue 3 in Figure 5) respond through a
similar process, complementing the viola with a real-time audio mosaic drawn from the
harp and vibraphone sample databases. In effect this is another version of the notated
harp and vibraphone parts, re-synthesised in parallel with the performance by the live
players, complementing with a successive layer of heterophony.
In a related passage later in the work, a dense ensemble texture is transcribed from
solo instrumental material. In this case, rather than responding to a notated score, the
bass clarinetist improvised on a repeated low B[ tongued as fast as possible: the natural
162
Musique instrumentale concrète...
Figure 6. Arnold Schoenberg, Fünf Orchesterstücke Op. 16, No. 3. (reduction with C[ pedal
tone omitted for clarity).
variations in speed and response of the instrument produced a subtly shifting rhythm.
A mosaic was then transcribed for the rest of the ensemble using sample banks combined
from all four instruments excluding bass clarinet. In the finished score the bass clarinet
accompanies with a similar figure, ideally producing a timbral fusion between the soloist
and the ensemble texture.
Instead of the multiple-stranded heterophony of the opening viola solo, this passage
unfolds a virtual timbral polyphony. The mosaic made with CataRT is conceptually
monophonic, matching one sample at a time of the corpus to the target. However, when
the notes of this line are distributed to the instruments of the ensemble each note is
sustained or let ring longer than its monophonic value. The effect could be likened to the
verticalisation of a melodic line into a sustained harmony, for example in Schoenberg’s
Fünf Orchesterstücke Op. 16, No. 3 (Figure 6). The result suggests a “timbral pedal”
accompanying the soloist (Figure 7).
The rhythmic details of the bass clarinet part were derived from a different process.
An audio recording of an improvisation with CataRT, using the mouse as controller, was
segmented in OpenMusic using the transient-detection object from the OM-SuperVP
library (Figure 8). By applying incremental transient detection threshold values, rhyth-
mic patterns of decreasing density are produced. By using OM-SuperVP instead of
AudioSculpt6 the parameters can be flexibly adjusted directly within the OpenMusic
patch, and successive analyses can be applied iteratively. The use of a list of threshold
values to produce a rhythmic matrix resembles the varying loudness thresholds used to
compose Figure 5, however now the results are used formally: each level of the matrix is
composed out successively, to produce a carefully-controlled decrease in rhythmic density
across a long-term section of the work. The bass clarinet rhythm in Figure 7 is drawn
from the fourth line in Figure 8, partway through the process.
6 The OM-SuperVP library provides the means to run most of the AudioSculpt sound analysis and
processing features as part of OpenMusic visual programs. This has the advantage of connecting them
directly to the compositional workflow and iterative procedures.
163
Aaron Einbond
164
Musique instrumentale concrète...
Imaginary Gardens
In Without Words for soprano, 11 instruments, and electronics (2012), the sample data-
bases and field recordings that have informed my work since What the Blind See are
augmented by a database of vocal texts. The database was drawn from fragments,
translations, and paraphrases of writers including Marianne Moore, Wallace Stevens,
Matsuo Bashō, Douglas Huebler, Kenneth Goldsmith, and many other prose and poetic
texts—chosen because of their connection to transcription, mimesis, metaphor, and
place. They were recorded in collaboration with soprano Amanda DeBoer Bartlett who
improvised with the texts as well as with various vocal “preparations”—objects to be
sung through, including whistle, kazoo, coffee mug, cardboard tube, and spring drum.9
In analogy to the examples above, here audio recordings of the texts are privileged over
their symbolic printed versions.
This vocal database is treated as a mobile resource, ranging from isolated phonemes
to full words, phrases, or occasional sentences. Accordingly semantic references may be
entirely lost or more or less intelligible depending on vocal technique and whether or not
a preparation is present. This continuum of semanticity is exploited in the transcription
process, with smaller units available for audio mosaics suggesting other references, and
longer units conveying intact linguistic denotations of their own. An example of the
165
Aaron Einbond
former is the opening mosaic, performed by the live soprano, based on a target field
recording of frogs. The work opens with the field recording alone, then gradually cross-
fades it with a recorded audio mosaic of soprano samples, and eventually integrates the
live soprano herself in a seamless continuum. The tissue of whispered phonemes (square
note heads in Figure 9), principally fricatives, sometimes sung through a police whistle
(upper staff), imitates the high spectral range of the frogs, with breathy sung high D5
(diamond note heads) pointing to a prominent frequency of their call. The same field
recording receives a contrasting transcription for the full instrumental ensemble a few
minutes later in the work, by a similar technique to Figures 1 and 2.
In the vocal transcription shown in Figure 9, most of the phonemes are far too short
for their source texts to be identifiable, while a few open out into words or phrases
(“more in”, “okay yeah”), hinting at the semantic origins of the utterances. Instead the
connotation of the underlying field recording is more salient. Its compositional sense is
later revealed in a soprano citation of Marianne Moore: “imaginary gardens with real
toads in them”.
Over the course of the 18-minute work, these textual fragments are hesitantly aug-
mented, like folding fans that expand and contract. Eventually some vocal samples are
exposed in their entirety, like the Marianne Moore text above. For another text, a haiku
by Matsuo Bashō translated by Robert Hass, the simplest possible transcription process
is applied: the original soprano sample is re-notated as faithfully as possible with the
aid of AudioSculpt and OpenMusic. A chord-sequence analysis is exported as an
SDIF file, imported to OpenMusic, and the most salient monophonic pitch is chosen
from each segment. Nevertheless the “accuracy” of the result is belied by the subjective
decisions throughout the process, essential to capture the fine shades of rhythm and pitch
(Figure 10).
When the same soprano who created the original sample reinterprets the transcrip-
tion, she reproduces the historical/autobiographical moment of the sampling session—
the sample becomes a field recording. Yet as the soprano relearns to perform “herself”
166
Musique instrumentale concrète...
she underlines the tension between performance, transcription, and referent. As Peter
Ablinger says of his Weiss/Weisslich 36, “the same is not the same. There is a difference.
At least the difference between just being here and: listening”.10 Perhaps the same
difference lies behind Basho’s riddle-like haiku.
Traces
Textual fragments are taken as targets for audio mosaics as well as corpora. Brief excerpts
of Wallace Stevens’s own recorded reading of Credences of Summer are cited, placing
historical recordings of the poet in dialogue with the live soprano. One recording was
taken as the basis of an audio mosaic using a database of bass flute samples. The
patch in Figure 11 adds another feature to the CBCS workflow: textual metadata
output by CataRT are recorded in the SDIF file in name-value tables. Then using
OpenMusic objects getnvtlist and find-in-nvtlist the values can be read and used to
inform the manual editing of the score in Finale, as seen in Figures 11-12. For example,
evaluating the mapcar box labelled “SoundFile names” (Figure 11) yields the filename
“bfl-closed_o_harmonics-C3” for the second sample (in both the poly object in Figure 11
and the score in Figure 12). The pitch calculated by yin∼ is discarded and replaced with
fingered pitch C3 documented in the filename. To indicate the flutist’s closed embouchure
(lips sealed around the mouthpiece), a black circle is added over a hollow rectangular
note head. Finally, the flutist is directed to shape her mouth into the phoneme [o]. This
process is repeated for each sample to build the bass flute part.
Reminiscent of the technique of Figure 8, this audio mosaic was made in CataRT
with a range of different segmentations based on loudness threshold to produce mosaics of
different densities. Beyond the contextual variations in rhythmic transcription produced
by omquantify, there are also variations in the units chosen by CataRT according to the
segmentation. As CataRT calculates an average of descriptor values over each segment,
segments of different lengths corresponding to the same position in the target will have
slightly different average values, and as a result the best match in the database might
be different. So successive rhythms and playing techniques in the final matrix do not
align precisely, even though derived from the same mosaic. This is a case in point of
how slight variations in parameters can produce significantly different musical outcomes.
The durations of the bass flute samples themselves also come into play, for example
the 4-second tongued overtone glissando that concludes the third line of Figure 12.
During transcription a subjective compromise is made between the durations of the
target segments and the durations of the samples segments in the database.
These bass flute lines are reinterpreted in the finished work accompanied by repeti-
tions of the Wallace Stevens target sample itself: “Trace the gold sun about the whitened
sky/Without evasion by a single metaphor”. As the sample loops, the varying bass flute
lines are deployed in a gradual densification, in contrast to a gradual descrescendo by the
voice of Stevens. Instead of a syllable-by-syllable transcription, the bass flute traces a
more subtle mosaic privileging the weight and density of the grainy historical recording,
and emphasising the timbre of the poet reading over the sense of the printed text.
10 See http://ablinger.mur.at/docu1515.html
167
Aaron Einbond
Figure 11. Patch for Without Words with sub-patch get-SoundFiles on the left.
Figure 12. Transcriptions of the same Wallace Stevens sample for bass flute at increasing
loudness segmentation thresholds, subjectively edited in Finale.
168
Musique instrumentale concrète...
At the end of the work, the field recording, the Stevens quotation, and their varied
transcriptions return in a collage-like coda. While the recapitulation is perceptible on
a semantic level, it also suggests an abstract musical return. The recapitulation of the
same constellation of timbres, playing techniques, and spectral descriptors defined by a
concrete recording and/or its transcription can be heard in analogy to the tonal return of
common-practice music. In fact, given the statistical interpretation of tonality by scholars
like Carol Krumhansl [10] and David Temperley [19], as well as the spatial tonal models
advanced by Fred Lerdahl [11] and others, a direct connection could be made between the
pitch distribution of a tonal centre and the statistical distribution of descriptor weights
in a soundscape. Tonality as a place: a topic that invites further exploration.
Translation
Since 2012, transcription with CataRT and OpenMusic has advanced in several di-
rections: the bach package for Max has been integrated into the CataRT workflow to
facilitate real-time interaction with musical notation. Bach can also expedite exchange of
data with OpenMusic, which remains useful for high-level computer-aided composition
tasks. CataRT itself is available in a new version incorporating the MuBu package
of Max externals, improving its portability, clarity, and especially permitting access to
larger audio databases by taking advantage of 64-bit Max.
Posing the question of how mimesis can be generalised beyond strict transcription,
I integrated CataRT with a machine-learning algorithm that permits the association
of audio not only by descriptor similarity, but by shared context. Inspired by the
existing program OMax [1] and the PyOracle package for Python [18], the result-
ing tool CatOracle is available in the MuBuForMax-1.8.5 distribution and later.
It combines the powerful Audio Oracle algorithm for musical pattern matching with
CBCS for applications including computer improvisation, high-level control of synthesis,
computer-assisted composition, and musicological analysis. CatOracle extends the
notion of transcription beyond a reproduction coordinated linearly in time, permitting
an “improvisation” based on a target that can backtrack, repeat, or skip. Yet the shared
contextual relationships compel the result to bear a degree of resemblance of the original.
This could be described as behavioural transcription or “style imitation” [3]—or more
generally “translation”,11 a term chosen to suggest displacement of elements along the
time axis, as well as expression through a new improvisational language. Applied to
machine improvisation, it is especially complementary to other sites of spontaneity in
the transcription process: collaborative improvisation during sampling sessions, gestural
control with the mouse or Wacom tablet, and live reinterpretation of expressive micro-
timing.
A first compositional application in Xylography for violoncello and electronics (2015)
begins with an acoustic improvisation by cellist Pierre Morlet, transcribed in detail
in the score and reinterpreted in performance, translated live into multiple strands of
computer improvisation, the courses of which are further guided by ongoing transcription
of the live performer. This layered process of mimetic and creative feedback points to
promising territory at the intersection of composition, transcription, improvisation, and
interpretation.
169
Aaron Einbond
References
[1] Gérard Assayag, Georges Bloch, Marc Chemillier, Arshia Cont, Schlomo Dubnov.
“OMax Brothers: A Dynamic Topology of Agents for Improvisation Learning”. In
Proceedings of the ACM Multimedia Workshop on Audio and Music Computing for
Multimedia. Santa Barbara, 2006.
[2] Michel Chion. Guide des Objets Sonores: Pierre Schaeffer et la recherche musicale.
Paris: Institut National de l’Audiovisuel & Editions Buchet-Chastel, 1983-1995.
English translation: Guide To Sound Objects. Pierre Schaeffer and Musical Research
by John Dack and Christine North, 2009.
[3] Arshia Cont, Schlomo Dubnov, Gérard Assayag. “A Framework for Anticipatory
Machine Improvisation and Style Imitation”. In Anticipatory Behavior in Adaptive
Learning Systems. Rome, 2006.
[4] Francis Crick. “Central Dogma of Molecular Biology”. Nature, 227, 1970.
[6] Aaron Einbond. “Subtractive Synthesis: noise and digital (un)creativity”. In Aaron
Cassidy, Aaron Einbond (eds.) Noise In And As Music. University of Huddersfield
Press, 2013.
[9] Vijay Iyer. “Embodied Mind, Situated Cognition, and Expressive Microtiming in
African-American Music”. Music Perception, 19(3), 2002.
[11] Fred Lerdahl. Tonal Pitch Space. Oxford University Press, 2001.
[12] Fred Lerdahl, Ray Jackendoff. Generative Theory of Tonal Music. Cambridge, MA:
MIT Press, 1983.
[13] Theodore Levin. Where Rivers and Mountains Sing. Bloomington: Indiana
University Press, 2010.
170
Musique instrumentale concrète...
[19] David Temperley. The Cognition of Basic Musical Structures. Cambridge, MA: MIT
Press, 2001.
[20] Tolga Tüzün. “Maquette As Data Structure and Synthesis Agent in Metathesis”.
In Jean Bresson, Carlos Agon, Gérard Assayag (eds.) The OM Composer’s Book 2.
Editions Delatour France/IRCAM-Centre Pompidou, 2008.
[21] David Wessel. “Timbre Space as a Musical Control Structure”. Computer Music
Journal, 3(2), 1979.
Acknowledgements: I thank the musicians of Ensemble L’Instant Donné with whom I developed
the instrumental samples and techniques for What the Blind See and who premiered the work at the
Centquatre in Paris on June 12th, 2009. I thank soprano Amanda DeBoer Bartlett and the musicians
of Ensemble Dal Niente with whom I developed the samples and techniques for Without Words and
who premiered and recorded the work in Chicago on my portrait album released by Carrier Records
in 2012 (http://carrierrecords.com/index.php?album=einbond). I gratefully acknowledge researchers
at IRCAM including Diemo Schwarz and members of the ISMM team for developing and supporting
the CataRT package, Eric Daubresse for production support on What the Blind See, Jean Bresson,
Yan Maresz, Alexis Baskind, Emmanuel Jourdan, Mikhaïl Malt, Jean Lochard, and Cyril Beros. My
participation in two years of the Cursus in Composition and Music Technology was made possible by
the Fulbright Scholarship and Georges Ladd Prix de Paris.
171
Recomposing Beethoven with Music
Neurotechnology
Eduardo Reck Miranda and Anders Vinjar
1 Theterm “Music Neurotechnology”, coined by Miranda and colleagues, appeared in print for the first
time in 2009 [8].
2 fMRI stands for “functional magnetic resonance imaging” and can be used to show which parts of the
brain are involved in a particular mental process.
3 IRCAM’s ATO-MS software was also used to compose Corpus Callosum, and Modalys and
AudioSculpt were used in addition to ATO-MS and OpenMusic to compose Shockwaves.
173
Eduardo Reck Miranda and Anders Vinjar
introduce the compositions Symphony of Minds Listening and Corpus Callosum. Next,
we focus on the technical details of collecting and handling fMRI data, followed by
an overview of the OpenMusic patches that we developed for this project, and an
explanation of how ATO-MS software was used to generate orchestrations based on
fMRI information.
174
Recomposing Beethoven with Music Neurotechnology
The Apollonian brain is concerned with unilateral meanings, whereas the Dionysian
brain tends to forge connections between allegedly unrelated concepts. This is what our
minds do all the time when composing: the further our Apollonian brains push us to
perceive the world according to its agenda, the stronger the pull of our Dionysian brains
to perceive the world differently. Hence, computer technology is of foremost importance
for our métier, because it allows us to stretch our Apollonian musical side far beyond our
ability to do so by hand, prompting our Dionysian side to counteract accordingly. This
cognitive push and pull seems to be the driving force behind our musical creativity.
175
Eduardo Reck Miranda and Anders Vinjar
Each scanning session generated sets of fMRI data, each of which we associated
to a measure of the second movement of Beethoven’s 7th symphony. This is shown
schematically in Figure 3 and will be explained in more detail later.
Figure 3. The result of a scanning section is a set of fMRI data for each measure of Beethoven’s
work. (Note: this is only a schematic representation; the brain images do not correspond to the
actual music shown.)
The score of Beethoven’s movement (using a MIDI representation) was cut into
measures and analysed with Artificial Intelligence (AI) software developed at ICCMR4
[2, 10]. This software extracted statistical information about the structure of the music
such as melodic direction, tonal signature, intervals between notes, rhythmic structure,
176
Recomposing Beethoven with Music Neurotechnology
and so on. This information was then used by the AI software to reconstruct the
Beethoven movement, but with fMRI data influencing the process. Not surprisingly,
the fMRI scans differed amongst the three listeners. Therefore, brain activity from three
different minds yielded three different movements for the resulting composition, each of
which displaying varied degrees of resemblance to the original symphony.
In order to test the composition methods (or the “reconstruction” method) thor-
oughly, the composer decided to remain faithful to the original form of Beethoven’s
movement and avoid as much as possible editing the musical reconstructions generated
automatically by the system. The objective was to gain a good understanding of the
outcomes in a practical real-world, large-scale compositional task. However, this decision
heavily constrained his musical imagination during the composition of Symphony of
Minds Listening. One could say that the compositional process as whole was more
Apollonian here than Dionysian. Even though the piece caught the imagination of the
audience and received positive press reviews,5 the composer felt that the æsthetic results
lacked his creative touch and musical experience. Nevertheless, it was a useful experiment
that was followed up with Corpus Callossum and subsequently Shockwaves.
Corpus Callosum
In Corpus Callosum the composer allowed more freedom in his handling of the materials
produced by the computer. The compositional process was more Dionysian here than
Apollonian. And this time he worked with the fMRI data from his own brain only.
The title of the composition refers to the part of the brain that connects the left and
right hemispheres and facilitates communication between them. As already mentioned,
the left hemisphere is largely engaged in processing details. It is often associated with a
more objective, or scientific, knowledge of the world. Conversely, the right hemisphere is
largely engaged in taking a holistic view and is often associated with a more subjective,
or poetic, interpretation of the world [14]. One should bear in mind that this distinction
is controversial because it has been difficult to observe it clearly. In a normal brain
the two hemispheres work tightly together: they are highly interconnected and interact
through the corpus callosum. Nevertheless this notion is inspiring as philosophical and
artistic metaphor.
During the scanning session Miranda recalls that sometimes he lost concentration on
the music and his mind wandered off. We reckon that this is a typical manifestation of
brain asymmetry at work: while one side of his brain was striving to pay attention to
musical detail, the other was making mental associations, producing imageries, eliciting
feelings, and so on.
The orchestra for Corpus Callosum is divided into two groups: one to be placed on the
left side of the stage and the other on the right side, representing respectively the left and
right hemispheres of the brain. The composition develops as an interaction between these
two realms. The instruments on the right (2 flutes, 2 oboes, 2 clarinets, bass clarinet, 2
bassoons, 2 French horns, 2 trumpets, trombone, bass trombone, percussion—2 players,
5 Thesymphony was premiered on 23 February 2013 by Ten Tors Orchestra, conducted by Simon Ible, at
Peninsula Arts Contemporary Music Festival, Plymouth, UK. See a review in Gramophone magazine
and news story in The Times: http://goo.gl/FDM0iX and http://goo.gl/GtacVu
177
Eduardo Reck Miranda and Anders Vinjar
Figure 4. Excerpt from the score of Corpus Callosum, showing music representing the right
hemisphere.
Figure 5. Excerpt from the score of Corpus Callosum, showing a string quartet representing
the left hemisphere.
178
Recomposing Beethoven with Music Neurotechnology
2 violins, viola, and violoncello) play segments that were composed emphasising orches-
tration (Figure 4). These passages do not handle rhythm and melody explicitly. Rather,
the instruments play clusters of sustained notes; the focus is on timbre. Conversely,
the instruments on the left (marimba or piano, 2 violins, viola, and violoncello) play
modifications of passages pinched from Beethoven’s score (Figure 5). These segments
are noticeably rhythmic and melodic, and timbre is secondary.
In addition to programming OpenMusic patches to make transformations informed
by fMRI data, in Corpus Callosum we also used computer-aided orchestration software
to generate the orchestrations for the “right side” of the composition. Before we discuss
the orchestrations and transformations, below is brief explanation of how the fMRI data
were handled.
6 In psychology, the p-value is the probability of obtaining the same result by chance.
179
Eduardo Reck Miranda and Anders Vinjar
Table 1. The values of the strongest 5 ICA components for the first 10 measures of the
Beethoven movement yielded by the fMRI from Miranda.
In order to enable easy access to the ICA components during the composition work-
flow, the resulting set of ICA components were stored in a class-array object in Open-
Music (Figure 6). This facilitated accessing individual components, and specific lookups
based on time (Figures 7 and 8). Having the plots of the ICA components visually
available in the class-array also provided a valuable overview of the data during the
compositional process.
180
Recomposing Beethoven with Music Neurotechnology
Figure 8. Selecting components from total set of ICA components. This enables the composer
to select ICA values from a total of 25 options.
7 ATO-MS is unreleased software for computer-aided orchestration developed at IRCAM, preceding the
current Orchids system.
181
Eduardo Reck Miranda and Anders Vinjar
The system takes a given audio file, in this case a section from the recording of the
Beethoven movement, and analyses its spectrum using a Fast Fourier Transform (FFT)
with a window lasting two seconds, which roughly corresponds to one measure of the
musical score. This analysis enabled us to extract a set of 25 frequency values (in Hz)
for the most prominent partials of the spectrum at every two seconds of audio.
The partials of a sound spectrum can be described by frequency and amplitude
values; however, we discarded the amplitudes produced by the FFT analysis. Instead,
we replaced the amplitudes by the intensity of the 25 ICA components associated to the
respective measures of the Beethoven movement. This resulted in what we refer to as
the fMRI-modulated spectrum.
The ATO-MS audio database contains spectral analyses of recordings from the
instruments of a full symphonic orchestra. It contains information for the entire range
of each instrument, including different dynamics, articulations, and playing techniques.
Given a list of instruments, for which one wants the system to produce an orchestration,
the system searches the database for combinations of sounds whose blended spectra best
approximate the target fMRI-modulated spectrum. The system produces a number of
suggestions for the composer to work with.
As an example, let us examine how the first measure of the excerpt shown in Figure 4
was composed. The audio segment (Figure 10) that was used as a target for the
orchestration of the four measures of the score shown in Figure 4 corresponds to measures
139-147 of the Beethoven movement. The 25 frequencies generated by the FFT analysis of
the first 2 seconds of this sound are shown in Table 2, next to the ICA components of the
fMRI taken at the moment Miranda was in the scanner listening to the corresponding
measure. The values of the ICA components replaced the original amplitudes from
the FFT analysis, producing the fMRI-modulated spectral target for the orchestration
process.
182
Recomposing Beethoven with Music Neurotechnology
Figure 10. The waveform of the audio recording of measures 139-147 of the Beethoven
movement, lasting 8 seconds.
In this case the instruments specified were 2 flutes, 2 oboes, 2 clarinets, bass clarinet,
and 2 bassoons. Among the various suggestions generated by the system, the results
shown in the first measure of Figure 4 were selected. Note that oboe 1 was not used, and
that individual notes have different dynamics, which are necessary to obtain the desired
spectral behaviour.
183
Eduardo Reck Miranda and Anders Vinjar
Figure 11. Interpolator patch for two lists of values using a scaled modifier.
In the following section we introduce the four transformation patches which were
primarily used in Corpus Callosum, describing two of them in detail.
184
Recomposing Beethoven with Music Neurotechnology
185
Eduardo Reck Miranda and Anders Vinjar
Let us examine what happens if we set the fMRI_index = 0.6. In this case, we would
expect an output approximately half way between the original pitch and the inversion;
in other words, an almost neutral set of intervals. The result when applied to the pitch
sequence on the left of Figure 13 is {62.6, 62.6, 62, 62, 62.6}, yielding {63, 63, 62, 62,
63} when rounded up to the nearest whole number, and corresponding to D]4, D]4, D4,
D4, D]4 as shown in Figure 14.
Figure 14. Sequence after inversion with fMRI_index = 0.6, giving a nearly neutral set of
pitch intervals.
Pitch-scrambling algorithm
An algorithm to reorder the pitches of the input signal randomly provides an simple
stochastic transformation. Using the musical example in Figure 13 as a starting point,
the result of applying this transformation four times could produce, for instance, the
sequences illustrated in Figure 15.
Figure 15. The result of applying the pitch-scrambling algorithm four times on the same input.
As with the pitch inversion algorithm, the fMRI_index can be used to create a control
signal with which the amount of transformation can be varied. In order to illustrate this,
let us assume an fMRI_index = 0.4. Considering the same input measure as before and
the transformed values from the first pitch scramble shown in Figure 15, the value of d,
between the first scramble and the input sequence is calculated as follows:
d = {(74 − 71), (72 − 71), (71 − 74), (71 − 72), (71 − 71)} = {3, 1, −3, −1, 0}
The scaled modifier values are then calculated by multiplying the difference values
by fMRI_index = 0.4:
SMV = {1.2, 0.4, −1.2, −0.4, 0}
Finally, the SMV values are summed with the values of the original input to give a
transformed set of output values:
As with the previous example, the resulting values are rounded up to the nearest
whole number, giving a transformed set of MIDI values equal to {72, 71, 73, 72, 71},
which is rendered as {C5, B4, C]5, C5, B4}, as shown in Figure 16. Note that the
output is significantly closer, in overall structure, to the unscrambled input than to the
first scrambled transformation.
186
Recomposing Beethoven with Music Neurotechnology
Figure 16. Transformed output created by the pitch scrambling algorithm assuming
fMRI_index = 0.4.
Figure 17. Generating variations based on fMRI-modulated style analysis using the LZ library
(the fMRI index corresponds to the variable niter in the patch).
187
Eduardo Reck Miranda and Anders Vinjar
Figure 18. Patch for stylistic analysis passing through analysis of audio. Results from sound
analyses are used as inputs to the LZ library functions visible at the bottom of the patch.
188
Recomposing Beethoven with Music Neurotechnology
Then each section was processed separately and the compositions (or recompositions)
were reconstructed according to the same number of measures as in the original. In other
words, the 13 sections provided a template for the new compositions. They informed the
overarching forms of each of the three movements of Symphony of Minds Listening, and
to some extent the forms of the other two orchestral works. More information on how
this process was carried out is available in [10].
At the lowest level of deconstructing the piece, processing the materials on a section-
by-section basis, and re-assembling the sections, the process tended to be Apollonian,
particularly in Symphony of Minds Listening. During the process of reassembling—or
recomposing—the sections, Miranda deliberately avoided interfering with the computer-
generated materials (informed by fMRI transformations) as much as possible. But of
course he produced a wealth of materials to choose from by experimenting with different
transformation algorithms and running each many times. From these, the composer’s
choices of the particular materials that were used in the finished compositions have no
objective explanation; the process was Dionysian.
As already mentioned, the composer felt that the æsthetic quality of the resulting
work was compromised by the lack of further Dionysian interventions on the computer-
generated transformations and on the musical form. A different approach was adopted for
Corpus Callosum and Shockwaves. Again the processing of materials for these two pieces
was organised according to the same 13 sections shown above. However, the recomposi-
tion process saw no barriers with respect to these sections: some of them were manually
shortened, others enlarged, and others still overlapped. Further, another dimension to the
compositional process was added by the inclusion of computer-generated orchestration,
which was also generated on a section-by-section basis. Miranda then meddled with the
computer-generated materials extensively before they made their way into the final score.
These processes were fundamentally Dionysian and were not documented.
We tend to think of software such as OpenMusic as an aid to an Apollonian compo-
sitional processes. However, Miranda’s Dionysian processes were very much supported by
one of the key features of OpenMusic: the possibility to archive multiple variations of
a process, letting the composer store those that provide potentially interesting material
with which to work further down the road.
189
Eduardo Reck Miranda and Anders Vinjar
Figures 19-20 below show one of the patches we made for the recomposition of different
sections using a maquette.
190
Recomposing Beethoven with Music Neurotechnology
Concluding remarks
This chapter introduces the algorithms and compositional methods that we designed—
informed and inspired by neuroscience—and provides a glimpse of the potential of Music
Neurotechnology for creativity, and more generally, for musical research. Our approach
is inspired by the way in which the brain listens to music: sounds are deconstructed as
soon as they enter the ear. Different streams of neuronally coded data travel through
distinct auditory pathways towards cortical structures, such as the auditory cortex and
beyond, where the data are mingled with data from other senses and memories, and
reconstructed into what is perceived as music.
Computing technology is an invaluable tool for modelling creative processes such
as the ones we introduce above. However, the authors suggest that composers should
not always approach computer-aided composition as a substitute for their own creative
hand and musical intuition. In theory Symphony of Minds Listening could be considered
more elegant than the other two pieces because the composer followed our methods
“by the book”. However, from the three pieces mentioned in this chapter, the one that
the composer considers most æsthetically successful is the one in which he allowed the
most flexibility with respect to amending the outcomes from the transformations and
prescribed musical form: Shockwaves.
The beauty of a tool such as OpenMusic is that it allows composers to model creative
processes and explore them in flexible ways. In addition to being a powerful programming
tool for developing sophisticated models “of” musical creativity, OpenMusic provides
useful support for exploring the potential of Music Neurotechnology to develop models
“for” musical creativity.
References
[1] Gerold Baier, Thomas Hermann, Ulrich Stephani. “Event-based sonification of EEG
rhythms in real time”. Clinical Neurophysiology, 118(6), 2007.
[3] Mick Grierson, Chris Kieffer. “Better Brain Interfacing for the Masses: Progress in
Event-Related Potential Detection using Commercial Brain Computer Interfaces”.
In CHI’11 Extended Abstracts on Human Factors in Computing Systems. Vancouver,
2011.
[4] Thilo Hinterberger. “Orchestral Sonification of Brain Signals and its Application
to Brain-Computer Interfaces and Performing Arts”. In Proceedings of the 2nd
International Workshop on Interactive Sonification. York, 2007.
191
Eduardo Reck Miranda and Anders Vinjar
[11] Friedrich Nietzsche. The Birth of Tragedy. Oxford University Press, 2000.
[12] Ramaswamy Palaniappan. “Electroencephalogram-based Brain-Computer Interface:
An Introduction”. In Eduardo R. Miranda, Julien Castet (eds.) Guide to Brain-
Computer Music Interfacing. London: Springer, 2011.
Online
[15] The FIL Methods Group at UCL. “SMP8 Manual”. Institute of Neurology,
University College London. http://www.fil.ion.ucl.ac.uk/spm/.
Acknowledgements: The authors would like to thank the following colleagues for their contribution in
a way of another at various stages of this project. Zoran Josipovic (Cognitive Neurophysiology Lab, New
York University, USA) and Dan Lloyd (Trinity College, Hartford, USA) for their valuable assistance
with the fMRI scans and analysis. Duncan Williams and Aurelien Antoine (Plymouth University’s
Interdisciplinary Centre for Computer Music Research) for their help with the design of generative
algorithms. Philippe Esling (IRCAM) for granting us access to the ATO-MS code and helping us to
adapt the software to generate the fMRI-based orchestrations. Miranda is grateful to Ten Tors Orchestra
and its artistic director Simon Ible for performing the three compositions discussed in this chapter in
public concerts and festivals in Plymouth. Finally the authors are grateful for institutional and financial
support from Plymouth University’s School of Humanities and Performing Arts and EPSRC, in the UK,
and BEK and NOTAM in Norway.
192
Composing for the resonance:
Finding new relationships between
architecture and musical composition
Ambrose Field
This chapter documents my search for a new compositional method that could be in-
formed by the acoustic response of the venue in which a piece is to be performed.
Although composers have written pieces for particular spaces for centuries, this is tra-
ditionally a process informed by aural memory. I had two main aims: to make a tight
bond between score and acoustic result, and to be able to design pieces where the acoustic
contribution of a venue would be a known entity at composition time—rather than an
after-effect of a performance. Both of these factors create new musical possibilities as
they permit the space itself to become an integrated part of the composition. This body
of work would not have been possible without OpenMusic, and the role of the system
is explained with example patches demonstrating sound analysis, harmonic selection,
rhythmic generation, and texture manipulation.
I have always been attracted to the sound of spaces. The sparkling sheen of the Vienna
Konzerthaus; the short, muted slap-back effect you hear when singing outdoors in dense
forests; the huge, billowing acoustics of large Gothic European cathedrals; or the effects
of the weather on sound transmission in humid Asian cities in monsoon season. I wish
to access these acoustic effects directly as part of my music, as the accepted disjunction
between space and piece is a little odd—it is all sound after all that the audience will
hear, so why compose only half of the experience? The separation of reverberation in
the minds of composers, listeners, and performers from the musical material on the stave
itself is largely a Western phenomenon inherited from the late 18th century. It simply
does not exist in Eastern cultures where music historically was not performed in concert
halls. Performing in a large building, shopping mall, concert hall, or parking lot is an
entirely different acoustic experience from performing the same music in a small room
or domestic setting. So why do we expect one piece to be flexible enough to cope with
such a wide array of performance venues? My goal in researching this idea was to come
up with a compositional method that could help make pieces more specific to the venues
in which they are performed, regardless of what types of venue they might be. In order
to do this, it is necessary to take account of the acoustic at the pre-compositional stage
in the process of making a work.
193
Ambrose Field
Composers for centuries have written music for particular spaces. Just as it is hard to
imagine the early polyphony of Perotin being performed today in Notre Dame without
consideration for the acoustic, it is equally hard to imagine that Perotin would have
written his music without the acoustic effects of Notre Dame Cathedral in mind—for
this was a building with which he was intimately familiar. The link between performance
venue and composition space in this example from early music is significantly close: it is
not until the 19th century that this becomes fractured, to accommodate new marketing
propositions of touring orchestras and performers. From the 19th century on, the space
in which a musical work would be performed ceased to matter so much to composers—
particularly in central Europe, largely due to the construction of concert halls as public
entertainment venues. The space instead needed to serve the music, and concert halls
are today still judged on their clarity as much as their added resonance.1
This bond between acoustic resonance and composition has previously been negoti-
ated within the aural memory of the composer: few composers have precisely measured
the spaces in which their work is to be performed. So, I was prepared to sacrifice the idea
that a piece can be performed anywhere regardless of the acoustic; instead, I adopted an
approach where the composition becomes tuned to a particular space for these pieces.
This has allowed a much more detailed investigation of this idea.
OpenMusic offers an amazing opportunity: the combination of sound analysis with
the ability to feed that analysis back into a compositional strategy, gives rise to new
possibilities. It would be easy to see this method as a kind of parametric version of Alvin
Lucier’s I am sitting in a room: where an audible room analysis is fed-back into the piece,
as discussed in [1]. What I would like to accomplish poses a somewhat harder problem
than establishing a feedback loop: to try and gauge, or assess before performance, how
a room might respond to particular types of musical construction that at this point in
the composers’ working processes, have no realisation in sound. A further goal is to
understand how the response of the room could actually be composed for, enabling the
room itself to take an active part in the performance.
1 The effect of this relationship on the perception of sound localisation in contemporary music is discussed
in [3].
194
Composing for the resonance...
2 Quantaform Series performed by Jos Zwaanenburg, flute. Sargasso Records, London, SCD28071, 2013.
3 The RT-60 is a measurement of the reverberation time of a room. It indicates how long reverberation
takes to decay by 60 dB [2].
195
Ambrose Field
Figure 1. Generating a stack of rhythmic fragments based on specific reverb times. The sub-
patch make_first_fragment generates an initial rhythm. As the purpose of the omloop is to
generate further variations on this rhythm, I lock the graph of the fragment when I am happy
with the aural result.
196
Composing for the resonance...
Within the inner omloop of the patch, a rhythm is created by the m-rhythm function
from the OMTimePack library, for reverb time using simple 0-order Markov probability
functions (see Figure 3). In The OM Composer’s Book 1 [6], Paul Nauert notes that
the simpler probability tables are in fact the ones which most often offer the greatest
compositional potential. In experimenting with the generative process for this piece, I
tried constraints-based methods using the OMRC library and although they would have
been well suited to generating a layered substructure, m-rhythm provided faster solutions
of equal quality.
Figure 3. Sub-patch inside the omloop from Figure 1 to generate rhythmic fragments.
197
Ambrose Field
The same patch was re-used in a variety of contexts in making this piece, from
forming local-level rhythm to providing high-level structure. It generates slightly different
rhythms each time, as each fragment is assembled from a random process, albeit one
subject to design through the use of a probability table. However, each rhythm output is
related to the other materials generated in that it is a product of the same process. The
rhythms generated are then folded into the piece using a series of structural time markers
derived from the initial rhythm generated (from Figure 2), creating another unified
musical bond between the contents of the fragments and their use in-time within the piece.
As a result of this folding process, and the rescaling of sections and structures, the audible
relationship between acoustic proportion and notation becomes progressively abstracted.
Thus Quantaform Series does not have an absolute one-to-one correspondence between
reverb time and notation: one has absolutely influenced the other, but by way of choice
and modification.
198
Composing for the resonance...
Figure 4. Score extract: Quantaform 16 for solo flute and reverberant acoustics.
199
Ambrose Field
Figure 5. Patch analysing how an impulse response affects harmony. (The input and output
multi-seq are shown in Figures 6 and 7.)
Figure 7. First seven measures of Quantaform 16 showing the prolongation effect the
reverberation has on the input material. The texture is spread onto three staves for clarity,
with the longest resonances on the lowest stave.
200
Composing for the resonance...
4 From the Department of Electronics at the University of York, led by my colleague Dr. Jude Brereton.
5 The <delta-time> parameter in as-›om specifies a window during which all notes will be gathered and
considered as part of the same chord.
201
Ambrose Field
Figure 8. Extracting the pitch of resonances which occur on a short-term time scale from an
impulse response.
text-file box as calculate occurrences can take a while to compute and the data may be
required in other patches (and then simply copied from this text file). Figure 9 shows a
detail of the calculate occurrences sub-patch.
By extracting information about the short-term resonant profile of a space, valuable
information is generated influencing how an acoustic might instantly respond to, and
help to inform, my choice of harmonic colours appropriately. As these harmonies are not
necessarily based on the same pitch centres as those derived from the longer-term analysis
(see Figure 10), they provide a means of informing the pre-compositional planning for
the moment-to-moment evolution of the piece.
Having ascertained which pitches occur most frequently on a very short-term timescale
in the acoustic, I then also wanted to obtain information that might match a general,
longer-term perception of the space. The patch in Figure 10 analyses the impulse response
to find the resonances with the longest durations, as opposed to the most frequently
202
Composing for the resonance...
Figure 10. Patch to find which pitches occur over longer time scales. Note the longer analysis
window of 500 ms, and also the display of only the pitches with the n longest durations.
203
Ambrose Field
Conclusion
OpenMusic has played an evolving part in my compositional practice. It is perhaps at
its most helpful to me in presenting a wide set of possibilities, rather than suggesting
any one specific solution. Whilst many of the methods I have used could be extended
through stricter, algorithmic approaches, my working practice instead foregrounds human
intervention and a questioning of computed output. In short, I need choice.
Singers working on my piece Architexture 1 commented that they felt liberated: the
acoustic clarity that was generated through appropriately spacing out the notes through
time was very welcome, allowing them to concentrate on articulation and diction. Open-
Music has permitted me to develop a new compositional approach: the investigation of
what the most effective compositional strategies might be for particular spaces based on
measured, architectural acoustics. It has played a significant part in making these pieces
a reality, and in helping performers approach the material with clarity.
References
[1] Densil Cabrera. “Acoustical, psychoacoustical and subjective assessment of Alvin
Lucier’s I am sitting in a room”. In Australasian Computer Music Association
Conference. Sydney, 2001.
[2] F. Alton Everest, Ken C. Pohlmann. Master Handbook of Acoustics. New York:
McGraw-Hill, 5th edition, 2009.
[3] Jean-Pascal Jullien, Olivier Warusfel. “Technologies et perception auditive de
l’espace”. In Les Cahiers de I’IRCAM, 5. Espaces. IRCAM-Centre Pompidou, 1994.
[4] David G. Malham, Anthony Myatt. “3-D Sound Spatialization using Ambisonic
Techniques”. Computer Music Journal, 19(4), 1995.
[5] Robert J. McAulay, Thomas F. Quatieri. “Speech analysis/synthesis based on a
sinusoidal representation”. IEEE Transactions on Acoustics, Speech, and Signal
Processing, 34(4), 1986.
204
Sketching, synthesis, spatialisation:
The integrated workflow of Cognitive
Consonance
Christopher Trapani
Composed in 2009-2010, Cognitive Consonance is a single piece divided into two: two
plucked-string soloists from two disparate worlds, two different sets of tools for composing
with electronics, and two different approaches to the practicality of writing non-tempered
ensemble music.
This article outlines the role of computer-aided composition in the first part of
Cognitive Consonance, an 11-minute movement entitled “Disorientation”, written for an
unusual solo instrument with extreme limitations: a custom-designed microtonal qanûn.
Three distinct stages of the process can be singled out. First, the pre-compositional
phase, where OpenMusic served as a practical tool for organising microtonal pitch
content. Second, the synthesis stage, where physical models of plucked strings were
created via a customised Modalys patch, alongside the batch retuning of sample libraries
(using the phase vocoder SuperVP). As a final step, the OMPrisma library was used
to generate polished eight-channel sound files combining samples and synthesis, ready to
be exported to the performance patch.
Sketching
The qanûn (kanun in Turkish orthography) is a trapezoidal zither, common throughout
North Africa, Turkey, and the Middle East. It is capable of being retuned while played
thanks to a series of short levers underneath the strings known as mandals, which can
be raised or lowered to change the length of the string.
The layout of these mandals varies by region, with multiple degrees of microtonality:
chromatic semitones in Armenia, tempered quarter-tones in North Africa, and a finer
gradation in Turkey, where most models divide the semitone into six parts. Pictures of
the instrument are included as Figures 1 and 2.
“Disorientation” was composed for a specific instrument, designed by the late French-
Swiss qanûn player Julien Jalâl Eddine Weiss. Rather than equidistant mandals, Julien
has created an intricate system of microtonal tuning with the goal of representing a
maximal number of just intonation ratios. Strings have as many as fourteen mandals,
for a total or fifteen possible microtonal accidentals on a given string (notated with a
personal series of symbols derived from Western, Arabic, Turkish, and Persian prac-
tice), representing the just intonation ratios outlined in Julien’s handwritten schema in
Figure 3.
205
Christopher Trapani
206
Sketching, synthesis, spatialisation: The integrated workflow of Cognitive Consonance
Figure 4 shows the eight gradations of pitch available between a C and C], using
the intervals in cents taken from the lower right-hand corner of Julien’s diagram. The
challenge of tracking the combinatorial possibilities of this detailed pitch network fit
naturally with OpenMusic modelling at the pre-compositional stage. It was necessary,
first, to gain a comprehensive idea of the practical applications of this theory—to explore
which pitches were actually available on the instrument—and second, to have a means
of comparing Julien’s approximations to independently derived pitch content to better
understand how closely his microtones were capable of approximating given passages.
207
Christopher Trapani
The results are sometimes surprising: we learn that the closest available pitch to a
tempered middle C is actually a B + 6 mandals, and not the qanûn’s C, which would be
two cents lower (thanks to the instrument’s Pythagorean tuning). Figure 6 gives a look
at the loops and operations inside this patch.
208
Sketching, synthesis, spatialisation: The integrated workflow of Cognitive Consonance
The next step was to compare the pitches on Julien’s instrument to an ideal harmonic
spectrum. As shown in Figure 7 we can visualise the deviations in cents within a bpf,
discovering that the largest divergence between the instrument and the first 25 pitches of
the harmonic series falls on the 17th partial, which as a C] on Julien’s qanûn is 10 cents
higher than a tempered C]. (Compare this to the difference between a just major third
and a tempered major third, an approximation central to nearly all tonal music, at 14
cents.)
209
Christopher Trapani
Finally, Figure 9 shows a score excerpt with the first appearance of this material,
underpinned by the low C fundamental in the harp and contrabass.
3
#
+0, / 0,
Qanun $ " 6
,- , ,- ,
+ + % % % +
3
6 " - "
(, , ,
)/
p ppp , pp f
3
mf pp f
Elec. $ "" !
"
"
"
0 , , ', , , 0 ', , / .
,, ,, 2
0/ , (,
" ! + - "" ,
#$ "
Gtr. % % " % % % . + . .
)((
!
3 ppp sfz mf
pp p
/
+0,&, 0/ . &/ . 0/ . &/ . 0/ . &/ . 0/ . &/ .
(put chopstick down)
" * % ! " 0/ &/ . 0, &,
"* * *
3
$" "
sfz mp pppp p pppp pppp
Hp.
)" ! "
pp hit with chopstick
" * % " % % % % % " % %
/
* )/
( VIBRAPHONE
$ "" ! " , 0,
with bow
Perc. % * $ 0/ " /. /. , 0/ /. /. " / 8
con Ped. n
" " 3. . 3. 3. 3.
pp p
3 1
!3 3
(irregular tremolando)
)" 1
"1 3. . "! 3. " 33 3 1
3. 1
% % % . 3 " 33 3
Vc.
3
n n mp
col legno battuto col legno battuto
punta d'arco 7 pizz. punta d'arco
ricochet l. v.
)" . ! ",
#+ "
ricochet
Cb. + 6 . % " % % % % % "' + 6 . %
, ,,,,,,,,
pp p , ,,,,,,,,
pppp pp pppp
Synthesis
In generating the electronic material for “Disorientation”, physical modelling synthesis
with Modalys also played a large role. The goal was to create a “virtual qanûn”, whose
capabilities complement and exceed the actual instrument, by using vibrato, glissandi,
and by pushing beyond the capabilities of a human performer. The “mono-string” and
“pluck” models seemed to mimic the freely-resonating triple strings of the qanûn. This
synthesis also was controlled using OpenMusic, via the Modalys library, with precise
microtonal pitch information and onsets described within chord-seq objects.
The Modalys sub-patch itself, the result of extensive experimentation (and a bit too
large and labyrinthine to be included here) consists of three embedded patches. At the
lowest level, there is the Modalys synthesis patch that creates, for each midicent value
in the chord-seq, two or three strings (depending on the register, as on the actual qanûn
or indeed the modern piano), which may also be attached to a resonating membrane.
To these synthesised strings a pitch-variance control applies a random variation in cents
within a given threshold, recreating the natural variance in acoustic tuning that gives a
slight “chorus” effect to doubled strings. The distance between the virtual strings can
also be controlled, as well as the height and velocity of the attack, lending a further sense
of tangible physicality to the final synthesis.
210
Sketching, synthesis, spatialisation: The integrated workflow of Cognitive Consonance
The intermediate sub-patch, shown in Figure 10, is a loop that creates, for each
midicent pitch in the chord-seq being synthesised, breakpoint functions capable of piloting
Modalys. Onset data is translated into bpfs that describe the up-and-down motion of
a Modalys bi-two-mass object (mimicking the plectrum). Access points (the position
where the plectrum touches the string) and listening points (the position on the string
whose vibrations are calculated) are both programmed to change dynamically around a
given threshold for a bit of automated timbral variety.
The top layer of the Modalys sub-patch, the control panel (see Figure 11), contains
all of the relevant data for the later synthesis. A key object is chord-seq-spread, which
reorganises the chord-seq’s data according to pitch, listing all of the onsets, durations
(irrelevant to the freely-ringing strings here, which are given a fade-out after the final
attack of the length given to the last-onset-plus-dur sub-patch), and velocities associated
with each midicent value in the chord-seq. Each virtual string (or, properly speaking, each
set of two or three strings, as explained above) can thus be re-attacked, with the physical
model of the vibrating string taking resonance into account. After much experimentation,
a range of scaled parameters were linked to velocity, each contributing toward a more
realistic synthesis: the height from which a virtual plectrum attacks the string, the drop
time, the stiffness of the plectrum, and the force of the pluck in Newtons. The upper and
lower limits of each of these synthesis parameters is given in the control panel. There
is also an option to make every note in the chord-seq sound as an octave harmonic, by
calculating strings an octave lower and placing a bi-two-mass object at their midpoint for
211
Christopher Trapani
212
Sketching, synthesis, spatialisation: The integrated workflow of Cognitive Consonance
213
Christopher Trapani
214
Sketching, synthesis, spatialisation: The integrated workflow of Cognitive Consonance
Spatialisation
The final sound files for “Disorientation” are in eight channels, in a circular configuration
around the audience. The generation of these eight-channel files was integrated into
OpenMusic as well, via a set of custom abstractions built around the Csound spatial-
isation library OMPrisma. The examples that follow focus on Modalys synthesis,
but OMPrisma was also used extensively to spatialise the batch-processed samples
mentioned above.
OMPrisma comprises a set of tools for
creating multichannel sound files accord-
ing to several different spatialisation algo-
rithms (including VBAP, DBAP, and Am-
bisonics) and types of movement (discrete,
continuous, trajectories) capable of adapt-
ing to customised speaker configurations.
Taking a list of mono sound files as an
input, an OMPrisma class object (named
for the type of spatialisation to be applied)
calculates a Csound score according to a
series of matrix data. Parameters can be
added as keywords: durations, entrance
delays, gain, x- and y-positions, x- and y-
envelopes, among others. Each parameter
can be input as a list or as a single value
that will be applied to all columns of the
matrix. The object can also be opened in
an editor for a look at all of the matrix
values, and the possibility of editing these
by hand; Figure 15 shows an editor view of
a VBAP.discrete object containing values
corresponding to seven sound files. Figure 15. Matrix data inside of an OM-
Figure 16 displays a typical musical ap- Prisma VBAP object.
plication of a discrete OMPrisma process.
The chord-seq contains an algorithmically generated passage which will be used to pilot
Modalys synthesis, where each midicent value will be synthesised as a separate string.
The angle-panning-vbap abstraction then acts as an intermediate step between the list of
sound files and the OMPrisma objects. Its simplified controls calculate a list of positions
along a given arc (described in degrees) so that each sound file is given a unique position
in space for a truly vivid synthesis. These positions can also be randomised by toggling
the value 0 or 1. In Figure 16 a light ring of high strings positioned in the back of the
hall (95 to 265 degrees) can be seen in the eight-channel waveform.
Figure 17 provides a look inside a slightly expanded version of the abstraction which
also applies a random distance value to each of the sound files. Note the bpc object
(“speaker setup”) connected to the vbap-setup function near the end of the synthesis
chain; this box contains speaker position data, and can be customised to any configura-
tion.
215
Christopher Trapani
216
Sketching, synthesis, spatialisation: The integrated workflow of Cognitive Consonance
Figure 18. From algorithmically generated passage to spatialised eight-channel sound file.
Rather than a discrete position, each synthesised string is here given its own trajec-
tory. This is accomplished by using a VBAP.continuous object along with a set of tools
for manipulating trajectories, as displayed in Figure 19. An initial trajectory, described
in a bpc object, is reflected using the traj-mirror function. Traj-interpol then calculates
a number of trajectories (corresponding to the length of the incoming list of sound files)
between these two extremes. The result is a smooth movement over the length of the
sound file from a single point, front and centre, to a splayed arrangement at the back of
the hall, where each virtual string again has a unique position.
217
Christopher Trapani
References
[1] Aaron Einbond, Christopher Trapani, Diemo Schwarz. “Precise Pitch Control in Real
Time Corpus-Based Concatenative Synthesis”. In Proceedings of the International
Computer Music Conference. Ljubljana, 2012.
218
Germination
Germination was created for IRCAM’s Manifeste Festival in 2013. The work originates
from the architectural characteristics of IRCAM itself and its location in Paris. The
research and production activities of the institute are located underground beneath place
Igor Stravinsky, which is traversed daily by thousands of people who generally are not
aware of their existence. The idea of Germination was to make emerge, on the surface
of the pavement, a work of music performed in the concert hall beneath.
The work is in two parts: an ensemble piece for 13 musicians and electronics performed
in the concert hall is followed on place Igor Stravinsky by an electroacoustic piece diffused
by 50 MP3 players, each connected to a miniature loudspeaker (see Figure 1-b). After
the first part, the audience is invited to emerge from the building and to listen to the
second part of the piece outside. This path, from below the ground up into the air, is
reminiscent of a plant sprouting in two phases: the development of a seed in the ground
followed by the growth of the plant on the surface. A vegetal installation emphasised this
idea of plant germination, and allowed better integrating of the diffusion system on site
(see Figure 1-a). The electroacoustic piece was performed on three successive evenings,
each time in a different version: starting from instrumental sounds the first night to more
abstract and noisy sounds the last.
Figure 1. Left: Germination vegetal installation on place Igor Stravinsky. Right: MP3 player
and its loudspeaker on-site. Photographs
c Astrid Verspieren and Serge Lemouton.
219
Jean-Luc Hervé and Serge Lemouton
The organic idea of vegetal growth was the compositional model, both for the instrumen-
tal and electroacoustic parts. OpenMusic was used as a computer-aided composition
tool for the entire piece; however in this text we will focus on the creation of the second,
electroacoustic part, which was played on the 50 MP3 players. Every “agent” (MP3
player and speaker) played a different sound file, and the whole installation produced
an enormous polyphony where each part had its own rhythmic, harmonic, and timbral
specificity.1 Generating such a large number of sound files, one for each voice but all
derived from the same model, was a difficult task to perform manually. OpenMusic
allowed easy iterative implementation through “batch” processing, turning out to be a
perfect tool to achieve this goal.
Each sound file or voice is made up of several successive sections, organised according
to a process of transformation of sound material in the domains of pitch, timbre, and
rhythm. Every section is developed in 50 variants, corresponding to the 50 voices. These
variants have slightly different durations, so that the transition from one section to the
other is usually asynchronous among the different voices. However, some of the sections
(e.g. 1, 2, and 5 as shown in Figure 2) start at the exact same time in all tracks: in this
case silences are inserted between the sections in order to synchronise the next starting
point.
Figure 2. The 50 voices are divided into sections of varying lengths. A number of
synchronisation points (bold arrows in this graph) group the beginnings of specific sections.
The electroacoustic part of Germination was composed in several steps (see Figure 3).
Sound files were selected from a database following specific criteria and then processed
using OM-SuperVP in order to multiply and vary the sounds in the corpus. They were
then ordered according to a process of evolution of sonic parameters, and put together
rhythmically in order to generate the 50 variants of each section. Then the sections were
concatenated in an OpenMusic maquette to produce the final 50 tracks.
1 Like
a population of frogs living in a pond, which all belong to the same species and emit the same
sound, but each at a different pitch and rhythm.
220
Germination
Figure 3. Workflow and software used for the electroacoustic part of Germination.
A set of sounds from the corpus is selected according to a given criterion (instrument,
pitch, playing mode, etc.) and enriched by applying the same modifications to all sounds
in the set. For each section of the piece, a different sound transformation is applied.
In the example shown in Figure 4, a section of “sliding” sounds is processed with pitch
and time modifications.2 A set of similar sounds (for instance, sustained string sounds)
is processed in a loop that produces for each sound n variants (10 in this case), selecting
random time-stretch and transposition values within predetermined ranges. The sounds
are processed one-by-one within the loop trsp-ts-loop2, and each time 10 variants are
produced by the loop trsp-ts-loop3. The contents of trsp-ts-loop3 are visible in Figure 5.
The sub-patch make_glis_alea selects random start and end values within a specified
pitch range for the dynamic transposition process (generating artificial glissandi).
221
Jean-Luc Hervé and Serge Lemouton
Figure 5. Sound treatments using OM-SuperVP. The input sound (<son>) is processed n
times (<nb calculs>) by dynamic transposition and time stretching.
222
Germination
Figure 6. Ordering a set of sound files according to descriptor values computed by CataRT.
223
Jean-Luc Hervé and Serge Lemouton
Figure 8. Rhythmic process to determine of the time-positioning of sound files. Left: random
variation of the value range using a Gaussian distribution. Right: internal patch to compute
the exponential curve.
224
Germination
Figure 9. Main patch to generate the maquettes and sound files corresponding to the 50
tracks of the piece.
With each evaluation of the savesounds omloop, sound files are generated from each
maquette to be loaded on the MP3 players. The full set of maquettes can also be
collected in an other maquette, the maquette of maquettes (see Figure 11).
225
Jean-Luc Hervé and Serge Lemouton
Figure 10. Inside the maquette generation process. The arithmetics computes and adjusts
the onsets of the different sounds (sections) according to the specified pivots.
226
Germination
Figure 11. The maquette of maquettes. Each rectangle represents a track: a maquette
containing its own sequence of sections, as shown in the open window.
Conclusion
Sound processing software is generally designed and adapted to commercial musical
practice. OpenMusic, on the contrary, is not primarily dedicated to sound processing
and therefore presents a number of particularities and constraints in this domain. But
constraints for a composer often present opportunities to imagine other ways forward; to
explore alternatives and finally open new horizons.
In Germination, OpenMusic permitted the development of an algorithmic approach
to sound processing and editing, where the composer defines a prototype actualised
each time in a different way thanks to aleatoric processes. This approach is perfectly
suited to an organic notion of music, taking living species as a model, where the genetic
code defining the species is transcribed in each individual of a population with specific
variations. The same musical sequence can be repeated, yet always modified. This
characteristic allowed us in Germination to compose hyper-heterophonies: where each
voice realises the same sequence differently, reminiscent of the plastic versatility of living
things.
227
Dialogue with OpenMusic in the
process of composing Nothing that is
not there and the Nothing that is
Takéshi Tsuchiya
In Nothing that is not there and the Nothing that is for violoncello and electronics1
OpenMusic was used at several stages of the compositional process: in particular in the
construction of the overall structure, in the pitch structure of the violoncello part, and in
various aspects of sound design for the electroacoustic part. In this paper, I will explain
some important details of these different aspects.
Introduction
Nothing that is not there and the Nothing that is reflects the underlying theme of a
philosophical hypothesis, a kind of “simulation hypothesis” [1]. I also feel that this
theme is close to the Zen philosophy that is part of the traditional culture of Japan.
When I compose a work accompanied by electronic techniques, my first consideration
is the relationship of the instrument and the electroacoustic sound. In Nothing that is not
there and the Nothing that is, this relationship is not only that of a mere accompaniment.
The electronics are internalised as an extension of the violoncello. This extension is not
only timbral: it has various repercussions for the structure as well. But for this, it is
necessary to reconsider traditional musical parameters such as pitch, rhythm, dynamics,
tempo, and extend them to become new structures and transformations of timbre at the
micro-level of sound synthesis and processing.
Since I began to think about and write music with the use of computer-aided com-
position, OpenMusic has become one of the most significant tools among the variety of
methods involved in my compositional process. It is of course applied differently in each
work, but one important common point is structure-building and its coordination with
sketches written by hand.
The role of handwriting in my compositions can be roughly divided into two kinds
of thought processes. One of them is the idea of the existence of a whole structure.
I illustrate this with the calligraphy shown in Figure 1, which had an important role
in inspiring Nothing that is not there and the Nothing that is. At first glance this
1 Written for Ensemble Contemporary α and premiered at the 20th Seoul International Computer Music
Festival in 2013.
229
Takéshi Tsuchiya
calligraphy may look like “avant-garde” drawing, but it actually represents traditional
Kanji characters. These characters are written with Japanese black ink called sumi.
Their meaning is “meditation”.
The drawing has a time dimension (the order and speed of the successive paintbrush
strokes), just like a musical performance. It can be seen as an expression/interpretation
the Kanji characters. It also creates a space that is uninterrupted. At the right in
Figure 1 is the “block body” of the calligraphy, with numbers indicating the order of the
strokes. This order is important: it determines the busyu (or “radicals”) that constitute
the internal structure of the Kanji. The radicals have meanings on their own, and bring
up new meanings when they are combined (see Figure 2). This specific structure of Kanji
can be said to be a characteristic of thought in Asia and particularly in Japan.
I was inspired by this calligraphic structure to construct Nothing is not there and the
Nothing that is. However its reflection in the piece is quite different from the technical
approach seen in Impressionistic music: instead I have aimed to represent the concept
of time in Japanese or Asian cultures, as well as the relationship between the whole
structure and its parts.
Ryo Yanagi2 explains Japanese structural beauty using the Western concept of the
“golden section” [2]. For instance, he uses this ratio to describe the relationship of whole
and parts of the structure of the Tea Room. Although I have not used his specific
230
...Nothing that is not there and the Nothing that is
Figure 3. Sketch of Section IX of Nothing that is not there and the Nothing that is.
Construction
The overall structure of this work is implemented in the maquette shown in Figure 4.
The whole can be divided into twelve sections (I-XII). These twelve sections are aggre-
gated into five groups (or “scenes”), each colour-coded within the maquette and marked
-
1
.
5 Figure 5 shows the patch generating this maquette.
231
Takéshi Tsuchiya
The groups
-
1
5 have the following features:
1 (Section I, XII) [Static] Violoncello part played on the bridge and on the tailpiece.
No clear pitch and very low volume level.
2 (Section II, VII) [Kinetic] Various glissandi over a wide pitch range of the violon-
cello.
3 (Section III, VI, VIII) [Static] Symmetrical with scene
.
2 Uses violoncello noise
material.
4 (Section IV, V) [Kinetic] Complete contrast with the other four. Focus on the
violoncello (normal and extended playing techniques). Electronics consist of spatial
processing, remaining in the background.
5 (Section IX, X, XI) [Kinetic] Uses various techniques of interpolation. Comprises a
process staring from the lowest tone and reaching the highest tone of the violoncello.
Figure 6 details the internal constitution of the maquette for Section II. Its contents
are four temporal objects (patches inside the maquette). II-A1 is the principal object
determining the structure of the section. It is connected functionally with the other
objects II-A2, II-A3 and II-B. Figure 7 shows the contents of the patch in II-A1: the
pitch profile is filtered using the posn-match function, and then transmitted to the other
objects.
Such processes using OpenMusic (and particularly the maquette) allow the con-
struction of substructures with non-periodic and unpredictable timings. The conception
of these substructures is “narrative” and does not rely on traditional compositional logic.
232
...Nothing that is not there and the Nothing that is
233
Takéshi Tsuchiya
Figure 8. Opening of Nothing that is not there and the Nothing that is.
234
...Nothing that is not there and the Nothing that is
The process begins with the analysis of a violoncello recording using the transient-
detection tool from the OM-superVP library. This analysis generates approximately
a dozen temporal markers that define segments of the original waveform. The sound is
then processed using four different combinations of time stretching or pitch transposition,
also using OM-SuperVP. Finally the original sound and the four processed sounds are
concatenated into a single sound file.
Figure 10 shows a detailed view of one of these processes, corresponding to one of
the four sub-patches in Figure 9. The OMAlea library is used to determine stretching
factors for supervp-timestretch using pseudorandom number generation (zoom1 ), zoom-
ing up to double the duration of the segments. Other processing on the original sound
is applied using the OM-Sox library (left part of the patch); the two results are then
concatenated (sound-seq) and normalised.
235
Takéshi Tsuchiya
Some sounds of the electroacoustic part of this section were also prepared using the
CSound engine through the OM2CSound library (see Figures 11 and 12). OM2Csound
allows the control of Csound by converting data to Csound scores, but also the visual
programming of Csound orchestras. In this case, most of the control process actually
takes place in the orchestra (Figure 12). The Csound instrument processes fragments
of the recorded sounds (recordings of violoncello played with the bow on the tail-piece
and on the bridge) using finely tuned feedback effects.
236
...Nothing that is not there and the Nothing that is
Figure 12. Csound instrument definition (contents of inst-1 from Figure 11).
237
Takéshi Tsuchiya
The kind of glissando presented in this passage is one of the most distinctive features
of the piece (see also Section VIII, mm. 66 to 79 in Figure 14).
238
...Nothing that is not there and the Nothing that is
These passages are not based on the general pitch system of the other parts of the
work. Pitches are determined following the process implemented in Figure 15 where
glissando time and width are computed from a graph drawn with a bpf object. Each
iteration, implemented as an omloop, allows the pitch intervals determined by the bpf
points to be calculated, and the timing of the glissando to be controlled piecewise for each
bpf segment. Note that this loop uses tak.acum3, a module of the Tak-Tools library
(a collection of my own composition tools) adding a special feature for accumulation in
OpenMusic loops.
Figure 15. Patch for generation of glissandi (II-A1_patch in Figure 7). The iterations (omloop
at the right of the figure) generate glissandi from bpf segments.
Sections IX to XI form one group developing pitch material with the goal of reaching
the highest E[ of the violoncello. In Section IX the specific notation of a curve is used
to represent the beams of long grace notes sequences. These notes are played a piacere
(see Figure 16). The patch used to create these structures is shown in Figure 17.
239
Takéshi Tsuchiya
Figure 17. Main patch for Section IX. The pitch range of each beam (start and end pitch for
the sequence of grace notes) is determined by the set of chords at the top of the figure.
The main loop generating the pitch sequence is shown in Figure 18 (left). A random-
walk procedure is used (internal loop ran-w, displayed in Figure 19) to generate a sequence
of pitches within each specified range, given the number of elements required (specified
in the preliminary sketches—see Figure 3) and a pair of weights for the probability of
ascending vs. descending intervals (1:1, 3:1, 1:5, 1:1, 7:3, 1:4). In this process the internal
loop l-rem-d (also visible in Figure 19) discards tied notes and repetitions.
240
...Nothing that is not there and the Nothing that is
Figure 18. omloop for generating sequences of grace notes and its main sub-patch.
Figure 19. Internal loops from Figure 18. Left: random walk. Right: removing repeated notes.
241
Takéshi Tsuchiya
Eventually the melodic lines from Section IX continue to develop into figures including
ascending accelerandi (in Section X), then toward the final stage of a stable pitch profile
in Section XI (see Figure 20). This process is implemented using the interpolation tools
from the Profile library (see Figures 21 and 22).
Conclusion
I have described some relationships of OpenMusic and compositional process in my work
Nothing that is not there and the Nothing that is for violoncello and electronics. With
the work’s background in the culture of Japan, as well as my own identity, it has aspects
related to Japanese tradition as a frame of thinking. As a result, while Kanji characters
or calligraphy are not directly represented, they do appear implicitly, for example, in the
relationship between the whole and parts of the structure of the work. One could say that
these strong ties are a unique cultural and artistic property, and become fundamental
propositions for musical creation.
Since I became acquainted with the field of computer-aided composition in the early
1990s, OpenMusic has become an invaluable tool for the control of temporal structures
in the field of sound synthesis, instrumental writing, and form. The concept of time in
composition, and its creative meaning, is a major concern for every composer. In this
respect the maquette offers a shift away from stereotypes, and allows for the creation
of new temporal frameworks—establishing a relationship between the internal structure
and the evolution of the piece in time. In addition, a close coordination with sound
generation and processing engines such as CSound and SuperVP, as well as the output
to Max, are now indispensable for the construction of my work.
242
...Nothing that is not there and the Nothing that is
Figure 22. Visualisation of the chords inside the chord-seq from Figure 21. First and last steps
of the interpolation.
References
[1] Keiji Hashimoto. “A Dream of a Butterfly of Zhuangzi: A Meaning and Structure of
Wuhua”. Tetsugaku: The Journal of Hiroshima Philosophical Society, 51, 1999.
[2] Ryo Yanagi. Golden Section – The Ratio of Japan. Tokyo: Bijutsu Shuppan-Sha,
1977.
243
Koch’s Space
Julián Ávila
When we listen to a sound, our brain analyses its characteristics. This process of
extracting sound information in our brain always includes a description of spatial hearing;
that is, for any sound we hear, or brain determines its location so there is no “non-spatial
hearing” [2]. This behaviour could be an anthropological phenomenon with practical
application in life; but if source location is such an important element for our brains, it
seems reasonable that it should also be considered by composers.
Space is one of the most important elements in my music, and it is handled and
composed in the same way I work with other elements related to time, structure, process,
or the vertical and horizontal organisation of sound. For this reason I try to balance both
space and time (which in other disciplines are closely related concepts, for instance in
physics, which could be said to search for a “theory of everything”), a concern similar to
the connection that Karlheinz Stockhausen made between pitch, rhythm, and structure;
or later to spectral techniques.
An important issue is how to relate the two basic elements of time and space in order
to write music. In my music, fractals and self-similarity are highly useful tools to control
the processes that take place in a work. They allow me to apply the same idea in different
compositional strata without losing unity. The main goal is to choose a system applicable
to all the elements included in the composition of a space-time field.
1A topographic filter is a filter that depends on spatial location of the signal. For instance, location
could control cut-off frequency, Q factor, or any other filter parameters.
2 The Koch curve, also known as Koch Island or Koch Snowflake, “is a mathematically defined
coastline with triangular promontories whatever the scale of magnification. [. . . ] The Koch fractal
is the prototype of an extensive family of fractals based on the repetition of a simple geometrical
transformation”[3].
245
Julián Ávila
The score is divided into three pages located on three stands, positioned in a semicircle
(centre, left, and right) around the interpreter. The work is performed in darkness and it
is the interpreter, playing with a head lantern switched on throughout the performance,
who illuminates the staves.
The light is captured by photocells that allow electronic spatialisation using Max,
behaving as if it were the result of the illumination. The sound material is generated in
the form of an increasing and decreasing process of constructing the Koch curve so that
each performer’s movement, from left to right or from right to left, corresponds to an
iteration in this process. In this way the work attempts to “fractalise” not only sound
material and temporal process, but also the concert space and the electronics.
This work is closely related to the topic of spectral diffusion.3 Even though Koch’s
Space is not a “spectral diffusion piece”, as it is more related to topographic filtering and
spatial processes, it begins a compositional practice that moves in this direction and in
which research is currently open.4
Spatial development
There is only one sound source in this piece, the saxophone, along with its spatialisation
through the topographic filter, a “shadow” sound. This filter, parameterised using values
from the Koch curve, works as a shadow of the direct sound of the saxophone, meaning
that the public hears the direct sound coming from the stage—where the saxophonist
is positioned—as well as the filtered amplification coming from a surrounding array of
speakers. The filtered sound is spatialised precisely to the location that the performer
lights up in order to view the score.
The first step in the compositional process was the implementation of Koch’s fractal
shape in its first, second, and third iterations. These curves have been calculated as bpc
objects in OpenMusic using the OMChaos library and exported to an interpolation
table in Max. Figure 1 shows the fractal shape implementation that will be the main
source for all the other patches. The function make-w is used to define the horizontal and
vertical translation and angular shifting of the first iteration in order to build iterations 2
and 3.
Usually a filter graph is defined in dB (in the vertical, y-axis) as a function of frequency
(in the horizontal, x-axis). The filter used in this work is an FFT-based band-pass filter5
in which all the frequencies outside the band are completely attenuated, so there is
no need to graph the level. This makes it more convenient to present how the filter
3 Spectraldiffusion is the main topic of my Ph.D. research at NOVARS, University of Manchester, Spectral
Diffusion and Spectral Energy in Electroacoustic Composition: Sculpting Space in Time Informed by
Applications in Biophysics.
4 SpectralDiffusion and Spectral Energy research employing spectroscopy measurements has been widely
applied in the field of biophysics, using the Fourier Transform as its main analytical tool. Studies of
spectral diffusion in proteins [1] have identified behaviour, which composers can observe and reinvent
in the language of spectromorphology [6] and space-form [7] in electroacoustic composition. This work
with proteins may well be transposed and placed under the umbrella of spectral analysis of aesthetics
and music composition with computers, with a special emphasis on spectral and physical space.
5 The use of an FFT-based filter helps avoid a slope (dB/octave) at the cut-off frequency that exists in
traditional filters, as well as a change of phase in the filtered signal.
246
Koch’s Space
evolves throughout the space: the band-pass graphs show a frequency range on the y-
axis and represent space as an azimuth angle (in degrees) on the x-axis: hence the name
“topographic filter”. In other words, the band-pass filter limits vary depending on the
amplified sound position inside the performance space.
The shape of this topographic filter is derived from Koch’s fractal curves. In order to
keep the band-pass limits well-defined as a function of space, and to avoid holes in the
frequency bands, the original curves are transformed by taking the absolute value of the
x-increment as shown in Figure 2.
Figure 3 shows the original fractal shapes (iterations 1, 2, and 3), and the correspond-
ing transformed shapes for the topographic filter. The high-cut frequencies are scaled
between 200Hz and 500Hz, and the low-cut frequencies are an inversion of the same shape
scaled between 25Hz to 100Hz (see Figure 2). Figure 4 complements Figure 2, showing
the topographic filters that correspond to the first and third iterations.
247
Julián Ávila
248
Koch’s Space
The outcome of the patch from Figure 3 is exported to Max to modulate the limits
of the band-pass filter depending on the position of the performance illumination. The
three iteration curves produce three different filters that are applied to different sections
of the piece.
These filters have to be spread and “located” around the audience in the concert
space in order to achieve a spatial sound structure that changes when the light of the
performer is pointed towards a specific direction in the hall. As a many concert halls have
the configuration of an “Italian” theatre, the spatial distribution of the filtering process
is organised in a similar way (even though other kinds of configurations are possible).
The surrounding matrix of speakers6 cover the left, right, and back areas of the hall,
whilst the stage is reserved for the saxophone. So the direct sound will come from the
front, and the fractal-shaped filters will spread over the left, right and back, as shown in
Figure 5.
6 The number of speaker depends on the hall size, but at least four are needed to implement the filter
distribution in the room.
249
Julián Ávila
There exist many examples of music that has been composed and afterwards spatialised
or distributed in space. However, in this case, the aim of giving the spatial design
equal importance to other musical element leads the compositional process towards an
organisation of all the elements around the space. The organisation of the score is
therefore conditioned by the spatial form: as the topographic filter is divided into three
“sections”, the fractal curve that shapes the sound material is divided into the same three
parts and distributed in three score pages.
The musical process consists of 12 constructions and deconstructions of the fractal
states, as well as mixtures or deformations between fractal iterations. Each complete
fractal shape is distributed across the three score pages (each one placed on one stand),
and the score of Koch’s Space is read horizontally across the pages, rather than from
the top to the bottom of each page. The first iteration (corresponding to the first stave
of each page) is read from left to right, but the second iteration is read from right to
left, and so on. Reading the score in this way will produce movements of the lighting
during the performance, which are tracked to control the amplified sound filtering and
spatialisation processes. Figure 6 shows the correspondence between the topographic
filter and the disposition of the score.
250
Koch’s Space
1 4n
L∞ = lim =0 P∞ = lim =∞
n→∞ 3n n→∞ 3n−1
In other words, whilst the area of this shape is finite, the perimeter is infinite. This
peculiarity is utilised in this work to define pitch range limits and register. The perimeter
of the shape is applied to the pitch range (scaled to fractions of an octave) and the length
of the small segments of the curve is applied to the register or starting note (also scaled
within an octave). Figure 7 shows the first nine numbers that determine the perimeter
and segment length of each of the first nine iterations of Koch’s fractal and their scaling
within an octave range.
• Direct mapping.
• Mixing one curve iteration for pitches and another for rhythms.
• Interpolation between two different curve iterations.
• Superposition of curve iterations for use in different layers.
7 Koch’s snowflake is similar to the fractals presented in this chapter, but applied to the edges of an
equilateral triangle [4].
251
Julián Ávila
Direct mapping is the simplest translation between Koch’s fractal and musical ma-
terial, consisting of resizing the x-axis according to time and scaling the y-axis according
to register. This correlation between the fractal curve and music is used in Sections 3,
6, and 9. Figure 8 shows the patch for Section 6. Note that the align-chords function is
used to group nearby points from the bpc into chords. Figure 9 shows the corresponding
score extract.
252
Koch’s Space
Mixing uses different iterations for mapping rhythm and notes in order to obtain more
rhythmical content within an iteration. This rhythmical information is used not only for
repeated notes, but is translated into music using many different strategies. In fact, this
internal subdivision is used to change any parameter within the same note, for instance
timbre, dynamic, etc. The example in Figure 10 uses values from the first iteration inside
the rhythm of the second iteration of the curve, using the x-transfer function. Figure 11
shows the corresponding score extract from Section 4.
253
Julián Ávila
Interpolation allows one to obtain a shape “between” two iterations (see Figure 12).
This method is used to generate additional material from the three iterations used for
this piece; the only requirement is that the two curves be scaled within the same range.
As different iterations have a different number of points, the om-sample function is also
used to resample both bpc objects before connecting them to bpf-interpol.
254
Koch’s Space
Superposition allows the use of points of two different fractal iterations for the same
section (see for instance the patch and score of Section 7, Figures 13 and 14). In this
case, lower notes are used as fingerings to produce non-ordinary harmonics in contrast
to the same notes produced with normal fingering. Due to the complexity of iteration 3,
the function omquantify is used here to avoid small rhythmical subdivisions, as well as
align-chord to group coincident points in a chord, and merger to mix both shapes into
the final voice object.
255
Julián Ávila
These four methods for mapping material are combined to compose the structure
of the piece. Sections 1 to 9 comprise a process in which the fractal grows little by
little—from the simplest shapes to the most complex. Section 9 contains the shape that
corresponds to iteration 3 of the fractal—the furthest stage of the fractal used in this
piece—and after this point iterations decrease down to the first one again. Table 1 shows
how the musical material is organised to create the macro-structure.
Rhythms: K1
1 → Pitches: K0
7400 25”
Rhythms: K2
2 ← Pitches: K0
7400 25”
3 → 7400–7350 33”
K1
Rhythms: K2
4 ← Pitches: K1
8200–8100 33”
5 → 8200–8000 33”
Interpolation
K1-K2
8 ← 8450–7850 45”
Interpolation
K2-K3
Table 1. Overall structure of Koch’s Space. Direction indicates the reading direction (i.e.
spatial movement) of the section. Iteration indicates the curves used to determine the rhythm
and pitch material. Filter is the “topographic filter” used in the section.
256
Koch’s Space
System design
The live electronic system in Koch’s Space works in a similar way to procedural audio—
for instance as used in game engines—in which it is not possible to know in advance
the exact final result of the sound, but the behaviour of the sound is clearly defined.
Max is used to handle live interactions, and Arduino serves as an interface between
the photocells and Max.8 No score follower or event trigger is used; the system reacts to
the performance and leaves the performer absolutely free to take control of the electronic
system with her or his movements. Figure 15 shows a diagram of the different parts of
the system.
257
Julián Ávila
On stage
This chapter points out the importance of space and describes a methodology to work
in time-space composition balancing these two basic musical elements. I try to make no
separation between time and space elements, so that almost all the elements in this work
have properties in both fields. In fact this particular treatment of space is not only a
compositional idea: the performance also puts forward relationships between sound and
space, intensified by the visual effect of illumination which takes an important role in the
transmission of the musical idea.
Spatiality in music is not a new invention. There are numerous ancient examples,
such as Ancient Greek theatre; the works of Alessandro Striggio, Giovanni Gabrieli,
Thomas Tallis; the 17th -century cori spezatti [5] in San Marcos of Venice; and 20th -
century examples such as Karlheinz Stockhausen’s Gruppen (1955–57) or Luigi Nono’s
Prometeo (1981–85), to name just a few. But today’s technology allows us to think
differently about spatiality in music, so that new methods can be implemented as a
potential basis for new music. The initial idea of Koch’s Space was developed through
recent compositional tools, and most important, the idea of space has transcended the
score or the compositional process to be present in the performance.
References
[1] Jürgen Baier, Mads Gabrielsen, Silke Oellerich, Hartmut Michel, Marin van Heel,
Richard J. Cogdell, Jürgen Köhler. “Spectral Diffusion and Electron-Phonon
Coupling of the B800 BChl a Molecules in LH2 Complexes from Three Different
Species of Purple Bacteria”. Biophysical Journal, 97(9), 2009.
[2] Jens Blauert. Spatial Hearing: The Psychophysics of Human Sound Localization
(Revised Edition). Cambridge, MA: MIT Press, 1996.
[3] Hans Lauwerier. Fractals: Images of Chaos. London: Penguin Books, 1991.
[4] Benoit Mandelbrot. Fractals: Form, Chance, and Dimension. San Francisco: W. H.
Freeman & Co., 1977.
[5] Davitt Moroney. “Alessandro Striggio’s Mass in Forty and Sixty Parts”. Journal of
the American Musicological Society, 60(1), 2007.
[6] Denis Smalley. “Spectromorphology: explaining sound-shapes”. Organised Sound,
2(2), 1997.
[7] Denis Smalley. “Space-form and the acousmatic image”. Organised Sound, 12(1),
2007.
258
Electronic sound creation in
Balænoptera for bass clarinet,
electronic sounds, and live electronics
Fabio De Sanctis De Benedictis
On March 19th , 2013 a fin whale was beached south of Livorno. As, I suppose, in
recent times many Italians and perhaps Europeans—surely I—may feel confused like that
cetacean, this was the pretext to begin to plan Balænoptera, a quadraphonic composition
for bass clarinet, electronic sounds, and live electronics.1 The present paper is focused
on the use of OpenMusic to generate electronic sounds, or Computer Generated Sounds
as expressed by David Cope [2]. One of the strongest aspects of OpenMusic is its
integration with other IRCAM software; therefore it is in this direction that it has been
used in Balænoptera.2 A concise description will be given of formal and compositional
techniques in the work as they relate to algorithmic composition.
0 1 2 3 4 5 6 7 8 9 10 11
A B C D E F G H I L M N
O P Q R S T U V Z
1 Thiswork was requested by my friend and colleague Carlo Failli and is dedicated to him. The premiere
took place in 2015 during the sixth edition of the contemporary and electronic music concert series
Suoni Inauditi (Unheard Sounds) at the Istituto Superiore di Studi Musicali “Pietro Mascagni”.
2 The main OpenMusic libraries used in this work are OM-SuperVP, OM-Spat, and OMChroma, as
well as chord-sequence analyses obtained from Audiosculpt. PWGL and the PWCsound library have
been used as well; however in the course of this paper the examples of algorithmic processes, even if
realised with PWGL patches, will be presented in the form of analogous OpenMusic patches.
259
Fabio De Sanctis De Benedictis
3 See [4]. The interested reader can find fruitful information on Pitch-Class Set Theory there as well as
in [8] and [9].
4 (1,1,2,6)
is used here instead of (1,1,2,4) accepting Donatoni’s theory of the unconscious and error [3],
and following my initial feeling about this subdivision even if incoherent with other decisions.
5 84a indicates the first beat of measure 84, 84b the second beat, and so on.
260
Electronic sound creation in Balænoptera
6 Thesefrequency bands are approximately proportional to the intervals of contiguous pitches in the first
chord extracted from the analysis of multiphonic #4. The threshold of 2123 Hz corresponds to the
highest frequency of the multiphonic.
7 The choice of starting times is based on the number five, which is another important structural element
of the work. Five is also the number of formant filters which are used in other sound transformations
later in the piece.
261
Fabio De Sanctis De Benedictis
Figure 1. Spatialising the audio files of the first section using OM-Spat.
262
Electronic sound creation in Balænoptera
Four audio files (one for each channel) were created using CataRT by performing
four different movements in the descriptor space (Figure 3).8
The recorded audio was then stretched to exactly one minute. Each sound was assigned
to a separate channel, with no further spatialisation or movements. The onset times
were set to 0, 2, 4, and 8 seconds in this section (0, 1, 2, and 4 multiplied by 2). As a
final compositional decision for this section, I decided to begin the electronic part eight
seconds before the instrumental part.
263
Fabio De Sanctis De Benedictis
Figure 5. The same patch as in Figure 4, now using the function sdif-›chord-seq to extract all
the pitches of the multiphonic chord sequence, sorted and without repetitions.
264
Electronic sound creation in Balænoptera
In Figure 5 we can see the same process applied to all the pitches of the multiphonic
chord sequence. Clearly the number of invariants increases as a consequence of the
large number of pitches. A different version of this algorithm can be realised using
the function as-›om as a substitute for sdif-›chord-seq. The former permits one to set
the maximum polyphony and so to control the complexity of the chords and generated
combinations. During the composition of Balænoptera both functions were utilised and
the most satisfactory solutions were selected.10
10 The use of invariants of chords deduced from audio analysis represents a topic of interest for the
development of compositional tools and a research step toward a possible integration of Spectralism
and Pitch-class Set Theory.
265
Fabio De Sanctis De Benedictis
On the left of Figure 6 the mean fundamental frequency and the duration of each
sound are extracted. The difference of pitch is used in supervp-transposition to homogen-
ise the pitches, and the ratio of durations is used to set supervp-timestretch to obtain
sounds of the same duration prior to cross-synthesis.
Then the resulting sounds are distributed with attack times derived from a rhythmic
series obtained using the patch shown in Figure 7: the 120-second-long section is divided
proportionally into 1+1+2+1+1+2 parts (112112 series), and then each part is subdi-
vided into 1+1+2+4 (1124 series) as well as 4+2+1+1 (retrograde of 1124), according to
Boulez’s concept of “de-multiplied rhythm” ([1], pp. 145–146). The two rhythmic series
based upon 1124 and 4211 sub-divisions (from the second and third staff of the last poly
object in Figure 7) are shown in Figure 8.
The cross-synthesised audio files were also time-stretched, by factors of two and four,
and arranged in four channels, deployed cyclically from channel one to four until all files
were used. They were ordered according to the attack times of the two rhythmic series of
Figure 8, as well as a third series whose values were determined, in order, by the average
of the respective attack times of the first two series. In other words, the first attack time
of this series is 0, the second attack time is the average of the second attack times in the
two original series, and so on.
The mono mix of the quadraphonic file was analysed and transformed into a chord
sequence, whose rhythm offered the basis for the rhythmic figures of the bass clarinet
part in this section (see Figure 9).
266
Electronic sound creation in Balænoptera
267
Fabio De Sanctis De Benedictis
Figure 10. Formant filter with random selection of vowels and voice type.
268
Electronic sound creation in Balænoptera
random vowel of a random voice type. SuperVP performs interpolations between the
filter values, resulting in smooth transitions from each state to the following one. In
addition, the functions supervp-timestretch and supervp-transposition apply stretching
and transposition, independently, to the final sound.
In Figure 11 we can see the patch corresponding to the recursive formant filter. In
each sub-patch there is an instance of the filter from Figure 10, with the number of
vowels/voices set to 5. Filtered sounds can be transposed ad libitum using the supervp-
transposition function.
The sounds in the electronic part are constituted by mixing the sounds produced
by the filters in Figure 10 and Figure 11, the latter with transpositions set to 0, -12,
and -24 semitones. The sounds were selected in retrograde order with respect to their
appearance in the previous section. In total, 24 audio files were distributed across a
duration of 45 seconds. The previous recursive rhythm algorithm (Figure 7) was used
once again, dividing the overall duration into 2+1+1+2+1+1 parts (retrograde of series
112112) and subdividing each part into 4+2+1+1 parts (retrograde of series 1124). The
result can be observed in Figure 12.
Figure 12. Rhythm of audio file attack times in fourth section from 4’45” to 5’30”.
269
Fabio De Sanctis De Benedictis
Other sounds in these sections were synthesised in Csound using the OMChroma
library. Two instruments were created using the STKClarinet 11 opcode (without and
with an additional amplitude envelope) and converted to OMChroma classes.
The synthesis was parameterised using the pitches of chord sequences obtained from
the bass clarinet multiphonic analyses. So natural sounds are put into a dialectic with
the corresponding artificial ones, a sort of mirror game. The patch implementing this
process is shown in Figure 13. This patch also allows a duration multiplier to be set, in
this case to the values five or ten (again using five as structural number). The synthesised
sounds were processed by the formant filter as in Figure 10.
The analysis data from the four multiphonics were used and re-synthesised succes-
sively with the two Csound instruments and the two different duration multipliers. So
a total of 16 different sounds were created, then submitted to the formant filter with
transpositions of -24, -12, 0, +12, and +24 semitones, producing a total of 16 × 5 = 80
different sounds. The transposed versions of the same sounds were mixed in succession,
with a global fade-in/out, according to attack times of 0, 1, 2, 4, and 5 seconds deduced
from the series +1, +1, +2, +1 (a part of 112112). The order of the files in this
mix was derived from the melodic profiles of bass clarinet figures played just before,
and developed by inversion, retrograde, and retrograde-inversion according to Morris’s
P-Space theory.12 Finally the mixed sounds were distributed along a duration of 90
seconds using the same recursive rhythm generation method as before.
270
Electronic sound creation in Balænoptera
Final duet
The last part (actually the first to be composed, and the initial inspiration for the work)
proposes a duet between the bass clarinet and the whale sounds, as if the instrument swam
side by side with the great mammals, an episode of calm resignation and reconciliation.
The whale sounds are almost unprocessed (except for reverb applied in Max). So in the
end the source material is clearly revealed, and serves to reconfigure and feed back on
the memory of what was heard until this moment (see [6]). The section lasts about one
minute and a half. Ten whale sounds have been selected and distributed along a duration
of 90 seconds, approximately one every nine seconds.
The whale sounds were analysed as chord sequences then translated into musical
notation in OpenMusic (see Figure 14). The chords, approximated to quarter tones
and transposed down by an octave, give a pitch reservoir from which the bass clarinet
13 Paulstretch is a program for extreme audio stretching. In contrast with AudioSculpt, it does not
preserve attack transients when stretching the sounds, but smooths the resulting sounds to obtain
particular textures. See http://hypermammut.sourceforge.net/paulstretch/.
14 PWCsound is a library created by Giorgio Zucco for controlling Csound in the PWGL environment.
See http://pwcsound.jimdo.com/.
271
Fabio De Sanctis De Benedictis
pitches are derived. The chord sequences obtained by the function as-›om have also
been transformed into rhythmic sequences as in Figure 9 and mixed together to form the
clarinet rhythm following the succession of the corresponding whale sounds.
Figure 15 shows the patch related to this process. This patch gathers in a maquette
the voices resulting from the sub-patches cb01, cb02, etc. with different whale sound
analyses. These patches are similar to the patch in Figure 9, here using as-›om instead
of sdif-›chord-seq. Each voice is assigned an onset time corresponding to the onset of
the corresponding sound in the electronic part. Figure 16 shows the contents of the
maquette.
Figure 15. Extraction of the overall rhythmic polyphony from whale sounds.
The maquette allowed me to listen to and verify the overall rhythm, with every
rhythmic sequence transposed to a different pitch. By exporting the individual sequences
in a format suitable to be read by Finale, and rebuilding the whole structure as a
polyphonic score, it was possible to print sheet music to be used as a reference for the
rhythm of the bass clarinet part. Figure 17 shows the beginning of the instrumental part,
transposed, as presented in the final duet.
272
Electronic sound creation in Balænoptera
Figure 17. The beginning of the final duet: bass clarinet part, transposed.
Conclusions
OpenMusic and its libraries allowed me to structure the audio material in this com-
position in a way consistent with the form, as well as to produce strong connections
between the instrumental and electronic parts. In particular, it permitted a gradation in
the formal unfolding of the types of electronic sounds used, both in terms of their inner
composition and the process of their implementation. OpenMusic also allowed for the
use of the same material and processes both for the audio part and the instrumental part:
multiphonics, formant frequencies, numeric series, and so on. This careful calibration of
processes and materials ensured a high level of formal control, and the same criteria
were applied to the live electronics. Balaenoptera goes a long way beyond personal and
expressive content to mark a turning point in my composition operations, laying the
foundation for further future developments toward a significant integration of acoustic
and electronic sound.
273
Fabio De Sanctis De Benedictis
References
[1] Pierre Boulez. Note di apprendistato. Torino: Einaudi, 1968.
[2] David Cope. “Preface”. In Jean Bresson, Carlos Agon, Gérard Assayag (eds.) The
OM Composer’s Book 2. Editions Delatour France/IRCAM-Centre Pompidou, 2008.
[3] Franco Donatoni. Questo. Milano: Adelphi, 1970.
[4] Allen Forte. The Structure of Atonal Music. New Haven: Yale University Press, 1973.
[8] Susanna Pasticci. “Teoria degli insiemi e analisi della musica post-tonale”. Bollettino
del GATM, Anno II(1), 1995.
Online
[9] Gary Tucker. “A brief Introduction to Pitch-Class Set Analysis”. Mount Allison
University, 2001.
http://www.mta.ca/faculty/arts-letters/music/pc-set_project/pc-set_new/.
Acknowledgements: I would like to thank Daniele Salvini and Fabrizio Broglia of the Mathematics
Faculty of Pisa University, through whom I had the opportunity to meet Carlos Agon and Moreno
Andreatta. I wish to thank Carlos Agon, Moreno Andreatta, and their colleagues, whose work
has expanded my compositional horizons; and Jean Bresson, whose support, keen suggestions, and
corrections to this paper have brought it greater clarity and lucidity. Last but not least I thank Carlo
Failli for his excellent performance of Balænoptera.
274
Modelling a gesture: Tak-Sı̄m for
string quartet and live electronics
Alireza Farhang
This chapter offers a general outline of the processes of gesture modelling which were
devised for, and applied to, Persian music in the composition of Tak-Sı̄m, written for
The Kronos String Quartet. After a brief elucidation of the background to the work’s
composition, the various strategies that allowed me to integrate elements from Persian
musical culture will be presented, in an attempt to illustrate how conflicts between two
ontologically opposed musical cultures become a source of enrichment. Using analyses
of recordings by master setār player Ahmad Ebādi, I created models of ornamentations,
melodic fluctuations, timbral evolutions, extended techniques, and other parameters of
performance.
One cannot avoid the acquisitions of the past, except by regressing to a truly
primitive state. [...] Far from adding constraints, these acquisitions, or in other
words, our very culture, our mental functioning forms part of our musical material,
just as much as known or imagined sounds, and can be integrated with every degree
of freedom into a new musical discourse.
Tristan Murail [7]
Introduction
As a composer myself, this statement from Tristan Murail is significant to me. The sound
universe of a composer is not born out of a vacuum; rather, it is a reverberation of the
sounds with which that composer has lived. For a composer educated in both the Western
and Persian classical traditions, the process of composition takes on new proportions.
The composer must deal with problems related to an ontological gap between these two,
distant musical cultures. When it comes to rich and ancient cultures, the composer is
torn between two different ways of thinking that are sometimes radically opposed, and
that prevent him from seeking an original and personal language. On the one hand
he/she weaves emotional ties with his or her culture of origin, which places him/her in
the golden cage of eternal and timeless wonders, taking away all freedoms; on the other,
the composer lives in a modern world where discourse is based on a questioning of the
past.
As a composer who is familiar with the rationality of the music of Beethoven, Debussy
and Ferneyhough, the heavenly serenity and spirituality of Persian music, as well as the
noise of war and chaos of modern life, my career path reflects a perpetual challenge:
how to address the gap between the content of a pre-Galilean conscience, and the
275
Alireza Farhang
276
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
The radif
Persian music places great importance upon improvisation, but only according to a strict
set of rules. The radif 3 is a “systematic, sequential organisation of all melodies, motifs
and variations” [2]. This collection of types of melodies, organised according to a special
logic, is the basis of musical material that is memorised and transmitted orally from
master to student; its use implies a structured representation of pre-melodies which are
freely connected by improvised ornamentations. The standard radif is the declension of
12 sequences of which 5 are called āvāz (songs) and 7 are called dastgāh (which literally
means position of the hand).4 The abstract, succinct nature of these melodies gives
considerable freedom to the musician, allowing him or her to improvise with ornaments
that are characteristic of Persian music. In this way the performer may add his or her
own personal touches to a performance whilst still respecting the tradition of the radif.
Intervals and their relationships to one another constitute the basis for musicologi-
cal/theoretical research. In general, they are performed, checked, and analysed on string
instruments, in particular the oud (lute). The strings on the oud are tuned in perfect
277
Alireza Farhang
fourths; thus in Persian music the fourth is the key interval which the performer may
play without changing hand-position. The octave, the fifth, and the fourth correspond
to the fixed frets on the instrument.
Two superimposed dāng are known as a māyeh 6 and create a certain modal colour.
According to Talāı̄, a dastgāh or a āvāz may contain several māyeh. The performer must
be able to navigate between various māyeh in an appropriately idiomatic manner.
Circular representation
Circular representation, originating from astronomy, is one means of visually analysing
modal/tonal systems and their relationships to each other. In Iran and Greece, this type
of representation, used for star-charts and music alike, has existed for many centuries
(see Figure 2).
In OpenMusic the class n-cercle (circular representation) and the c2chord function
(to convert the circular representation to a chord), make it possible to create models of
combinations of modal and tonal transposable systems; furthermore, they allow the user
to analyse non-octavian scales which originate from the superimposition of two ore more
modes. In order to facilitate the representation of intervals, an approximation to 1/8th
tones was used. The superimposition of two dāng of shour and tchārgāh is a means of
constructing the māyeh of homāyun. Once the pitch approximation has been applied,
intervals are organised by 1/8th -tone as shown in Figure 3.
Figure 4 shows an OpenMusic patch containing a circular representation of the
māyeh of homāyun. In this example, the space between each point on the circle cor-
responds to 1/8th of a tone (the octave is therefore divided by 48). We may observe
that the range of the māyeh of homāyun is limited to a minor seventh (in this case from
D to C). Thus, in order to complete the circle (or octave), a tone is added at the end
of the māyeh. The patch superimposes the intervals of two dāng (in this case, shour
and tchārgāh) specifying an interval (also in 1/8th tones) between them. It outputs the
scale-degrees of the māyeh in the form of a chord.
6 Although māyeh is the result of superimposed dāng, it is described as “mode” in Western musical
terminology (i.e., whereby no distinction is made between the concepts of mode and māyeh).
278
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Figure 3. Intervals of the radif modes (dāng) in midicents and1/8th tone intervals.
Figure 4. Representation of the structure of the homāyun using the class n-cercle.
279
Alireza Farhang
Each combination of the four modes yields a different acoustical result. On occasion,
the modes overlap to produce new modes with a greater range. In these cases, it
is possible to round out micro-intervallic deviations in order to avoid discrepancies.
The phenomenon of overlapping modes can give rise to flexible-interval modes which
necessitate the use of accidentals. In Figure 5 we see the intervallic structure of āshour-
āvand, a sub-division of the radif which results from superimposing the māhour and
shour modes. In this example, the starting point of the māhour mode is C, and that
of the shour mode is D, a difference of 200 cents. The result of superimposing these
two modes is a new, third mode which spans C to G, in which E is inflected both an
1/8th -tone lower and a 1/4th -tone lower (E[ a 1/4th -tone higher).
The patch shown in Figure 6 helps to understand the intervallic structure of the
modes. Also it allows one to make new combinations of modes by superimposing two
dāng while specifying respective starting points separately. The sub-patch scale_maker-2
receives the intervals, the starting-point, a reference-pitch, the degree of approximation
(25 midicents, or 1/8th tone, in this case) and the range of the scale (for instances to
visualise it over several octaves). The combination of modes is calculated and subse-
quently displayed using a circular representation (n-cercle—visible here at the bottom of
the patch).
280
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Figure 6. OpenMusic patch to combine two dāng, specifying the starting point for each.
281
Alireza Farhang
Form
In Iran, the close relationship between poetry and music is considered self-evident. Music
from the Persian radif tradition, which is learned through oral transmission, is from
the first moment of its dissemination associated with poetry. It contains “melodies
which originate from ancient sung-poems, often of sequences which commemorate the
ta’zieh;8 their quality is melancholic, even poignant, and their rhythm is free. In addition,
instrumental demonstrations that highlight the technique have become an element of the
radif in its own right” [3]. Instrumental music is based upon song; in order to learn
the radif by heart, instrumentalists sing the poems which are associated with a given
melody. Melodic/rhythmic figures, dynamics, pauses, the tahrir (a word describing the
particular vibrations in the throat), etc. are all closely associated with poetry, which is
itself considered to be a “perfect musical form”.
As for the notion of time in Persian culture, Jacques Le Goff suggests that Western
culture, and by extension the Western musical tradition has, since the Middle Ages,
become subordinate to the time of the marketplace, “measured, or in other words, clearly
directed and predictable, [...] both eternally re-beginning and perpetually unpredictable
from any natural perspective” [6]. For the Iranian musician, time is perceived in a
fundamentally different manner: it is not chronometric and metres are not based on
regular strong and weak beats.
7 Following the development of polyphony, for combinatorial reasons, symphonic instruments of the 19th
century tend to produce purer, quasi-sinusoidal sounds. This phenomenon has never occurred in non-
European music, or more specifically, in non-symphonic music; the monophonic character of Persian
music and the importance of timbre does not necessitate a harmonic tone. Non-symphonic instruments
(i.e. those used in rituals, battles, therapy, etc.) have a more complex sound and display a higher degree
of inharmonicity. The spectral richness of an instrument like setār or ney allows the instrumentalist to
control the timbre as the expressive parameter of his performance.
8 Ta’azieh is a sacred theatrical genre that primarily commemorates the martyrdom of Imam Hossein
(the third Imam of the Shiites).
282
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Tak-Sı̄m
Two anecdotes
Having heard the Arditti Quartet perform my second string quartet, Echo-Chaos (2007),
David Harrington, the Kronos Quartet’s first violinist, professed to have detected sonori-
ties that originated from a distant musical culture. Although Echo-Chaos belonged to
a period during which I avoided using materials coming from Persian musical culture,
the influence was nonetheless present in this piece. After this meeting, I became acutely
aware of the importance of my own personal sonic-universe, of the identity of raw sound
material and its role in works of music.
One year later, following a research trip to New York, a serendipitous event which
reinforced my reflections upon musical identity convinced me to compose a work for
Kronos, which emphasised the synthesis of my reflections upon my own musical identity.
During a composition seminar at Columbia University in 2009, a young Turkish composer
suggested that we listen to an instrumental work of Turkish classical music. Having
played the work in question, he asked those present to identify the instrument that was
featured. The general consensus was that it must be some traditional Turkish instrument.
Some justified this hypothesis on grounds of the instrument’s timbre and intonation, but
no one had realised that it was, in fact, a cello. Thus, I discovered the extent to which
playing technique, intonation, intervallic content, articulation, ways of sustaining a note,
etc. characterise an instrument, its sonic image, and its identity.
283
Alireza Farhang
However, the musical universe of Tak-Sı̄m, is intimately associated with Iranian music.
In order to render the raw, initial compositional material more neutral in nature, it
was necessary to remove its overtly cultural connotations. It was therefore desirable to
deconstruct this material. To this end, one must possess a sound understanding of the
musical and aesthetic structure of Persian music.
9 Ahmad Ebādi (1906-1992) is a famous Master of setār whose unique style, virtuosity, and improvisation
technique make him one of the best setār players of his time.
10 “When Ahmad Ebādi was playing in live radio broadcasts, he asked members of his entourage about
his performance. Some remarked that there was too much noise. Technical means of radio broadcasts
at the time were limited; thus Ebādi, in order to minimise the degradation of sound quality of
his instrument during radio transmissions, began to avoid the drone string and favour the tak-sı̄m
technique.” Houshang Ebtehāg, translated by the author; see https://youtu.be/Zm2_Gf2asVM.
284
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
The composition of Tak-Sı̄m was undertaken using a recording of the radif, performed
by Ahmad Ebādi. Samples were first segmented into simple musical phrases by ear.
Each phrase, or neume, may be considered as a large gesture, or a combination of
micro-gestures. After transcription of the samples (Figure 7), further dissection into
smaller fragments was necessary in order to isolate individual, single, “pure” gestures
(see Figure 8); this was the basis for the creation of a catalogue of gestures used as a
“reservoir” from which the work’s key material is derived.
285
Alireza Farhang
11 Segāh is one of the modes of radif. The intervallic structure of segāh is similar to shour.
286
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Figure 11. The melodic contour is rendered visually using a bpf. Subtle variations in pitch
may therefore be tracked here.
In order to extract spectral information from the sound, it was necessary to deter-
mine the precise moments which were—from a spectral point of view—of most interest.
Once again using AudioSculpt, in this case the chord-sequence analysis and transient
detection functions, I generated a series of markers that identified moments of spectral
contrast. This process requires considerable fine-tuning (in terms of setting the variables
appropriately for the sample in question) in order to avoid an analysis containing digital
artefacts.
12 Infact we detect the pitch or fundamental frequency via the frequency differences between the
components of a harmonic sound rather than via a weak or sometimes absent fundamental frequency.
287
Alireza Farhang
Figure 12. Sonogram of the neume (above) and its intensity (below).
Figure 13. The spectral content of the neume represented as a series of chords.
288
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Harmonic texture
Harmonic texture in Tak-Sı̄m is also derived from chord-sequence analyses. Both the
instrumental and electronic content of the piece were based upon a harmonic grid created
from a sequence of chords obtained using this method. As the degree of inharmonicity of
the setār is considerable, the chords obtained were accordingly inharmonic (in a psycho-
acoustic sense). Nonetheless, the inharmonicity alone was not enough to yield a harmonic
progression.13 To this end, it was necessary to “enrich” the spectral-content of the chords
using a simple algorithm which added a series of notes to those already present in the
analyses (see Figure 14). In order to facilitate the composition of the instrumental
material, the aforementioned chords were arpeggiated and printed. Figure 15 illustrates
the chords in their original form, and Figure 16 represents them after application of the
process of harmonic enrichment; the latter was ultimately used for the Section V of the
work.
Figure 14. In this patch a simple algorithm adds a series of notes to those already present in
the analyses.
Beyond manipulating the degree of harmonicity of the chords obtained in the process
just described, other procedures were used to obtain the harmonic texture of the piece:
inversion, change of range or density, interpolation between two or more chords, and
transposition. Figure 17 shows a patch used to invert a chord while applying an algorithm
that increases the spacing between the components (while keeping adjacent pitches less
than or equal to one octave apart). Through this process, although the new chord is
wider, the texture remains close to that of the original chord but increases its range. The
pivot note represents the centre of inversion (here F), and all other components of the
chord are inverted around it. Note that in the inverted chord, the intensity of each note
is identical to that of the corresponding original note.
13 Thedissonance-consonance rule of tonal music is perfectly valid in music whose harmonic structure is
based on timbre. This creates tension-relaxation, and thus a sensation of movement.
289
Alireza Farhang
290
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
291
Alireza Farhang
292
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
34
4
12
÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷
. . œ.
œ. œ. œ. œ. œ. œ œ
1
. . œ.
1-8-2
œ. œ. œ. œ. œ. œ œ
3
1-8-1
4
4 4
œ œ
1-9-1
œœœ œœœ
œœ œœ
interaction with dynamics of cello
1-8-3
# œ# œ
arco ord.
– –. –. –. –. –. –. –. –. –. –. –. –. –. –. –. –. œ# O
s.p. s.t. s.p. s.p. ord.
arco ord.
s.p.
œ
s.p. ord.
æ Bœ œ #O O
#O #O
O#œ O O O#O#O#O
#O O O
& ( –– ) O O
O O O œ #O
Vln. 2
OO
f F p f
O# œ
s.p. ord.
O O
B–––––––––––––––– O # O III– IV
Vla.
–. –. –. –. –. –. –. –. –. –. –. –. –. –. –. –. O
O O
O O # œ O (O)
F p
f
? æO
s.p. ord.
Vlc.
(– )
f p
293
Alireza Farhang
8 88 6
410
5
15 4 e = 72 q = 60
4
÷ ÷ ÷ ÷ ÷÷÷ ÷
• Samplor
÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷ ÷÷÷÷÷÷÷÷÷÷ ÷ ÷ ÷ ÷ ÷ ÷
œ. œ. œ. œ. œ. œ. .
2
œ. œ. œ. œ. œ. œ. œ.
œ œ.
3
1 4 1
œ
5-1
œœœ 3-1
œœ
2-1 3-2
. . 3˘ . ord. (h )
# # Oœ ‰
3
r # œ # # Oœ
con legno tratto
œ #œ #O O
tremolo arco
& – ‰ O O O œO –
II III
Ȯ
IV
Vln. 1
æ # O O æ
" J J
p P p P ! F
.
œ #œ œ (q )
ord.
(h )
#O #O O #O #O O
& # œ œ OO OO OO
#O #O
O O O
Vln. 2
" P " P !F
˘
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ . . #œ Ÿ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
œ. # œ œ RÔ ® #
ord. vib.
1/4 ! con legno tratto
B – & Œ # Ȯ # œO B
–O ( O )
Vla.
O
p F P ! F
3
±>
s.p. s.t.
s.t. s.p.
? O Oæ
arco ord.
O
con legno tratto
J
Vlc. O – O
æ æ
ƒ " P "
294
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Sound synthesis
Except for some real-time processes, almost all of the electronic sounds used in Tak-Sı̄m
are made with additive synthesis. In order to build a dialogue between the electronic
sound and the expressive instrumental sound of the string quartet, all the materials
generated by computer must be sculpted carefully. Although sinusoidal waves, as the
most basic component of sound, offer a large palette of possibilities for creating new
colours, the lack of the human musical expression in synthesised sounds is a major concern
for many composers. In Tak-Sı̄m, modelling an expressive gesture generated by a human
performer is an efficient method to make the electronic sound more expressive.
Figure 21 shows the patch designed to control the electronic synthesis in Tak-Sı̄m.
This patch incorporates of five types of control: a harmonic grid (1 and 2), tremolo-like
gestures (3), and melodic gestures (4 and 5). It integrates most of the processes for ma-
terial generation and transformation described in the previous sections, and constitutes
the first level of synthesis control (generation of pitches, amplitudes, and onsets of the
oscillators of the additive synthesiser).
In Figure 21 an inverted chord derived from the analysis of the neume segāh provides
material for the synthesis process. Sections 1 and 2 of the patch are similar; however, they
generate two different types of sound. The parameters in the corresponding synthesis
processes are set in order to produce a sound with more vibrato and beating for the first,
and a dry and steady sound for the second. The global dynamic envelope of the sounds
is drawn in a bpf (labelled “wave shape”).
Along with melodic gestures, tremolo, vibrato, and beating are the gestures that
are most often used in the quartet. Section 3 generates a tremolo-like gesture via an
algorithm. A global pitch (freq), dynamic (amp) and tempo can be controlled using three
bpfs. The patch tremolo produces a series of repeated chords, which may be plotted
visually with a chord-seq (see Figure 22). The precise content of this data may then be
manipulated by hand where deemed necessary by the composer.
295
Alireza Farhang
The melodic contour of the neume segāh, imported from AudioSculpt, is edited and
modelled in the section 4 of the patch. It is “coloured” with a group of pitches derived
from the aforementioned harmonic grid. The proximity of timbres between this group of
pitches—represented here as a chord—and the harmonic grid, as well as the morphologic
similarity between instrumental gestures and modelled gestures—all derived from the
neume segāh, make the emergence of the electronic part sound natural.
Finally, section 5 of the patch generates vibrati that are used to extend the neume.
Figure 23 represents different methods for generating these vibrati in OpenMusic. Du-
ration and frequency deviations are among the parameters which allow one to design a
whole new gesture from the data extracted from the previous analyses. Depending upon
the context, beating and vibrato may be considered gestures in and of themselves.
All of the control data are then transferred to a synthesis patch that generates five
sounds (see Figures 24 and 25). For each of the five sounds it is possible to define a global
shape using bpf objects (see Figure 25). For instance, a sound can start quickly and
fade out slowly, which is suitable where a short attack with a long release is necessary.
When a sound has a long attack it is more suitable for a harmonic nap or grid. At
the top of the figure, the sub-patch titled sounds defines the actual timbre of the sound
by controlling the components of the sound synthesis process (see Figure 26). In fact,
this patch instantiates synthesis classes from the OMChroma library: the choice of a
particular class determines the contents of the Csound instrument that will be used for
synthesis.
296
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
Figure 24. Sound synthesis (1): this patch takes the data from the first-level control patch
(Figure 21) and outputs five different sounds (see Figure 25).
Figure 25. Sound synthesis (2): the sub-patch om-composers_synthesis, shown in Figure 24.
In this patch the global shape of the sounds is editable.
297
Alireza Farhang
Conclusion
As a composer who is cognisant of the centuries-old tradition in which he is situated,
I feel that the act of composing is one of projecting, onto paper, personal and unique
musical thought. However, a composer’s originality may also come from the fact that
he or she is affected by the sensorial and spiritual stimulus of an external event. On
the one hand, as Hegel maintains, “a work of art is superior to anything which exists in
nature” [5]; on the other, Persian musical thought dictates that the highest form of art
is nature itself, and all works of art are merely incomplete imitations of nature.
In the West, composition is purely an intellectual pursuit. What makes it, in Hegel’s
view, superior to nature is the spiritual content the composer instils within his or her
musical world in order that the work might demonstrate originality. Such a notion is
incompatible with the process by which a musician playing in an Eastern tradition sets
out to create music. For the traditional Iranian musician, a supreme work of music exists
only beyond the capacities for human perception. It belongs to an eternal and immutable
centre which is indifferent to conditions in the material world. The role of the performer
is therefore that of intermediary, albeit one who is by necessity equipped with ability
and knowledge, between the metaphysical and the real worlds. It is therefore necessary
to draw a distinction between composer (in the Western sense) and musician (in the
298
Modelling a gesture: Tak-Sı̄m for string quartet and live electronics
References
[1] Abd al-Qāder Mohamadtaghi Binesh b. Ghaybı̄ Marāghi. Maqāsed al-alhān. Tehran:
Bongāh-e tardjomeh va nashr-e ketāb, 1978.
[2] Joël Bastenaire. “Panorama des musiques savants et populaires d’Iran”. Écouter
Voir, 130, 2002.
[3] Nelly Caron, Daryush Safvat. Iran. Les tradition musicales. Paris: Editions Buchet-
Chastel, 1966.
[4] Jean During. “The ‘Imaginal’ Dimension and Art of Iran”. The world of music,
19(3-4), 1977.
[5] Georg Wilhelm Friedrich Hegel. Aesthetics: Lectures on Fine Arts. Oxford
University Press, 1831/1975.
[6] Jacques Le Goff. “Au Moyen Age : temps de l’Eglise et temps du marchand”.
Annales. Economies, Sociétés, Civilisations, 15(3), 1960.
[7] Tristan Murail. “Questions de cible”. In Pierre Michel (ed.) Modèles et Artifices.
Presses Universitaires de Strasbourg, 2004.
[8] Daryush Shayegan. Le regard mutilé. Paris: Albin Michel, 1989.
[9] Daryush Talāı̄. A new approach to the theory of Persian art music. Tehran:
Mahour Cultural Foundation, 1993. Translated from Persian Negareshi no be teori-e
mousighi-e irāni.
Acknowledgements: I thank Paul Clift for his translation of this text. Without his generous help this
chapter could not have been published.
299
Electronic dramaturgy and
computer-aided composition
in Re Orso
Re Orso (King Bear) is an opera merging acoustic instruments and electronics.1 The
electronics were realised at IRCAM with the assistance of Carlo Laurenzi. The libretto,
written by Catherine Ailloud-Nicolas and Giordano Ferrari, is based on a fable by Arrigo
Boito.
Every moment of the opera is exclusively inspired by and tightly related to the
prescriptions of the libretto and the intimate structure of the drama: there are no vocal,
instrumental or electronic sounds that do not have a deep connection to and a musical
justification in the dramaturgy. In addition, an important compositional objective was
that the electronic material be endowed with a clear dramaturgic role, in order to be
perceived as a character on its own (actually, several characters) with a personality that
develops during the action.2
Preceded by a short exordium, Re Orso is divided in two parts of approximately 45’
(five scenes) and 30’ (three scenes) separated by an intermezzo storico. The ensemble
leaves the pit at the end of this first part and the singers remain alone with the accom-
paniment of electronic sounds. Voice and electronics are therefore essential elements
of the dramaturgy and of the composition. Both have been written and organised
with computer-aided compositional tools. This text explores some of the representative
OpenMusic patches developed for this project.
1 Commissioned by Françoise and Jean-Philippe Billarant for IRCAM, the Ensemble Intercontemporain,
the Opéra Comique in Paris, La Monnaie in Brussels, and a French Commande d’Etat. It was premiered
at the Opéra Comique in May 2012.
See http://www.opera-comique.com/fr/saisons/saison-2011-2012/mai/re-orso
2 For this reason the word “electronics” does not appear in the subtitle of the opera, and is replaced by
its respective roles: Re Orso, Musical Legend for 4 singers, 4 actors, ensemble, invisible voices and
sounds, spatialisation, and acoustic totem.
301
Marco Stroppa and Jean Bresson
Figure 1. Re Orso. Opéra Comique, Paris (May, 2012). Photos: Elisabeth Carecchio
Compositional environment
Re Orso is organised around a “compositional environment” consisting, at the most basic
level, of 72 “formant chords”3 and 3 scale structures,4 as well as of several rhythmic cells
and dynamic profiles. The music deals with this environment according to a panoply of
rules that will not be analysed here.
This pre-compositional environment is used to generate the basic musical structures
of the piece, which the composer calls Leitgestalt. A Leitgestalt is defined as a cognitive
morphology [11] playing a dramaturgic role; that is, an acoustic and theatrical event
that impinges upon our cognition and hence can be recognised in other parts of the piece
when it comes back, even if it is transformed. For instance, a downward glissando always
means a kind of fall (moral, physical, psychological, etc.), large or small depending on
the extent and details of its realisation.
Re Orso brings a variety of electronic materials into play:
• Imaginary voices (FOF-based synthesis using the OM-Chant library [3]);
• Imaginary sounds (synthesised using the OMChroma library [1]);
• Imaginary instruments (physical modelling synthesis using Modalys);
• Imaginary ensemble of the intermezzo storico (recorded and remixed ensemble);
• Mysterious voices (recorded voices processed with AudioSculpt’s cross-synthesis);
• Mechanic doubles of the two buffo roles:
– Hand-controlled puppet (vocal sounds processed with SuperVP-Trax),
– Computer-controlled, singing Disklavier;
• Real-time processing of the main female role’s voice (Chromax spectral delay [5]);
• Overall coordination and concert setup (Max with Antescofo).
3 Chords algorithmically derived from databases of formant data (see footnote 11) transcribed from sung
vowels (5 or 8 formants) approximated in a chromatic space. In total, Re Orso uses a data base of 44
5-note chords and 28 7-to-10-note chords.
4A “scale structure” is a succession of absolute pitches that encompasses the playing range of the
ensemble. These structures are based on some intervallic and symmetric properties.
302
Electronic dramaturgy and computer-aided composition in Re Orso
The voices (4 singers and 4 actors) and instruments are amplified.5 The sound projec-
tion uses an 8-channel frontal setup (3 loudspeakers placed on the stage, 3 loudspeakers
hung above the stage, plus one hidden below the stage and one oriented downward from
the ceiling) and an acoustic totem, a column of eight loudspeakers that appears from
below the floor in the middle of the stage at the end of the piece.
OpenMusic has been used both for writing instrumental parts and for the generation
of electronic sounds. In particular, the OM-Chant library allowed the connection of
formalised compositional processes to vocal sound synthesis using the FOF technique
and the Chant synthesiser.
Sound models
The sound model is a fundamental concept in Re Orso as well as in most previous works
by Marco Stroppa. It is, with respect to the composition of sound, something similar to
the idea of Leitgestalt, or more precisely, to the concept of Musical Information Organism
developed by the composer in the 1980s for his musical material: a group of sounds
“that consists of several components and properties of varying complexity, maintaining
certain relationships and giving rise to a particular form” [9]. This concept implies that
different sounds corresponding to the same model can be recognised as possessing the
same identity and, therefore, can express the same sonic potential [10]. Technically, this
concept of sound model is implemented in the OMChroma library as a class (called
cr-model), which aims at facilitating the generation of families of sounds [4].
5 The amplification has three main purposes: first, it allows for a better fusion between the acoustic
sounds and the electronic materials; second, it gives the stage director total freedom of movement
on stage, without risking that the singers are not heard during a tutti. Finally, used as an acoustic
microscope, it can bring to the audience softly whispered sounds that would otherwise be inaudible.
6 Thisis an homage to Jonathan Harvey’s Mortuos plango, vivos voco, which uses the same bell at the
beginning of the piece.
303
Marco Stroppa and Jean Bresson
Figure 2. Patch synthesising the bell strokes played at the end of the intermezzo storico.
domains (leading to another cr-model labelled “model 2”) and finally converted into a
set of sound synthesis control parameters.
The expand-model function performs the conversion of the successive segments from
model 2 into a list of instances of a given class of “synthesis event” from the OMChroma
library (in this case, the class add-1 is used, which corresponds to a simple additive
synthesis). The function therefore returns one add-1 instance for each segment in the
input cr-model. Along with the cr-model, the second argument of expand-model is a patch
(cr-control-patch), which determines the details of the mapping performed between the
cr-model segments and the sound synthesis parameters.
The editor window of cr-control-patch is also visible in the figure. In this special
patch, the synthesis parameters (or “slots” of the class add-1 ) are represented by the
out-slot arrow boxes at the bottom. The mappings here are of several kinds:
• The number of “components” of each synthesis event (numcols—in this case with
the class add-1, the components corresponds to partials in the additive synthesis
process), as well as their respective amplitudes (amp), are derived from the two
“standard” patch inputs.
• The durations (durs) and frequencies (freq) of the components are determined
according to programmed relations with the data contained in the cr-model segment
(represented by the model-data box at the top).
304
Electronic dramaturgy and computer-aided composition in Re Orso
• The amplitude envelopes of the components (aenv) are not dependent on any
external data and are specified statically by the bpf visible at the bottom right of
the patch.
The expand-model function applies this cr-control-patch iteratively, updating the
model-data at each iteration with the values from the current segment, and producing a
new instance of add-1.7 The synthesize function at the bottom of the figure collects this
list of add-1 instances and runs the actual synthesis process. For each state of the spectral
evolution of the original stroke sound (that is, for each segment of the cr-model, and each
iteration within expand-model) a new, increasingly dark bell stroke is synthesised.
7 Infunctional programming a function like expand-model, taking another function as one of its
arguments, is called a higher-order function.
8 See for instance Peter Ablinger: http://ablinger.mur.at/speaking_piano.html
305
Marco Stroppa and Jean Bresson
Figure 4. Processing a sound model to generate a score (MIDI file) for the Disklavier.
306
Electronic dramaturgy and computer-aided composition in Re Orso
The final MIDI editing was done with Digital Performer (see Figure 5).
Pedal automations are added (una corda or mute), especially in the pp passages. This
not only changes the sound, but also creates a theatrical effect, as the piano on stage has
no lid and viewers can see the entire mechanics shifting closer to the strings.
Figure 5. MIDI file of the beginning of the first song of the Disklavier in Digital Performer.
Notice the spectral difference between f passages (e.g. in the first four 3/4 bars, with more notes
in the high register) and pp passages (e.g. bars 5 to 8, with reduced pitch range), as well as
the pedal controllers: short sustain-pedal strokes (represented by the dark rectangles) in the
f passages, followed in the pp passages by a combination of una corda (wider rectangles) and
shorter sustain-pedal strokes.
9 Itis known that the auditory perception of a high sound dynamic is accompanied not only by a higher
amplitude, but especially by a wider spectrum. In addition, due to the semitonal approximation, the
higher part of the spectrum tends to sound like a chromatic cluster; so filtering out some notes in the
soft passages proves to be helpful musically.
307
Marco Stroppa and Jean Bresson
308
Electronic dramaturgy and computer-aided composition in Re Orso
309
Marco Stroppa and Jean Bresson
<wdur>, <wout> (which determine the shape of the envelope of the FOF grains), or
<bw-scale> (an adjustable formant bandwidth scaler).
At the right of the figure the ch-f0 object controls the fundamental frequency over
the duration of the sound synthesis process. The base pitch is here modulated (using
param-process) by both vibrato (controlled by frequency and amplitude envelopes) and
jitter (aleatoric perturbation, also controlled by frequency and amplitude values).
The synthesize function at the bottom of the patch collects a list of Chant synthesis
events (in this case, one ch-fof and one ch-f0 ), formats an SDIF control file, and performs
an external call to the Chant synthesiser in order to produce the sound file.
Figure 6. Synthesis of a messa di voce with OM-Chant. Imaginary voices (Part I, Scene 2).
This patch is actually quite generic, and leaves many open parameters (input arrow
boxes in Figure 6) that are set in Figure 7 in order to produce a series of different sound
files. This series of sounds implements a progressive evolution from “realistic” sounds
(formants and fundamental frequency in the range of human voices, natural vibrato
curves, etc.) to more synthetic/imaginary ones, all produced by the same program.
310
Electronic dramaturgy and computer-aided composition in Re Orso
Figure 7. Iterative call of the patch from Figure 6 with variable set of control parameters.
The last sounds and set of parameters in Figure 7 are of particular interest. Very low
fundamental frequencies, on the order of a few Hertz, are perceived as trains of impulses
rather than pitches, while the short attacks and growing resonance times of the FOF
grains produce the effect of percussive, almost bell-like sounds.
In FOF synthesis the control of the formants’ bandwidths (and their shape in general)
is important for the perceived result. At the limiting case, an infinitely narrow bandwidth
(0Hz) is a simple partial (sinusoidal signal), and the FOF synthesis can then produce
spectra resembling additive-synthesis sounds. We will get back to these interesting
properties in our last example.
311
Marco Stroppa and Jean Bresson
• It contains both the pitches D (Leitgestalt of the King, D=Re=King) and B[ (which
is in the pitch-class range of the Leitgestalt of the Worm, A4-C]5). However, both
pitches are located one octave higher, that is, in the vocal range of the King’s
coerced wife (Oliba, high soprano). A lot of theatrical meanings for a simple ring!
• The range and vocal speed were compatible with those of the Queen of the Night
sound synthesis process.
The patch generating one of the sounds played in this passage is presented in Figure 8.
The omloop named notes-to-chant contains a patch similar to the one in Figure 6.
Instead of synthesising a sound at each iteration, it generates timed Chant events (ch-
fof and ch-f0 ) corresponding to the short tones of the cell phone (at the beginning) or
to slightly longer sung vowels with an expressive vibrato (at the end). The collected
sequence of events is synthesised as a whole phrase (outside the omloop). Automatic
interpolations are therefore performed between the successive Chant parameter changes,
which produce smooth progressions between the distinct sound features, and implement
the gradual embodiment of the initial ring into a human voice. While the short tones
sound quite synthetic and mechanical at first, as soon as a vocal-like vibrato is added
they suddenly acquire more organic and natural characteristics.
Figure 8. Generating a sung sequence (or “phrase”) from a list of Chant events: cell-
phone/queen-of-the-night at the beginning of Re Orso.
312
Electronic dramaturgy and computer-aided composition in Re Orso
Figure 9. Main OpenMusic patch generating the “Death of the King’s Voice”. The contents
of the framed boxes are detailed in Figures 11–15.
12 Inthis and the following figures the patch has been simplified for the sake of clarity. The presented
values are also not the exact ones used for the sounds played during the opera.
313
Marco Stroppa and Jean Bresson
The resulting sonogram (Figure 10) reveals a number of interesting details of this
process. The fundamental frequency of the FOF generator starts at a high D (D5, 585
Hz) and plunges into sub-audio values (0.3 Hz, that is, one stroke each 3.3 s). This
fundamental frequency is easy to track (or hear) at the beginning, and then disappears
in the low register. The “plunge” was structurally divided into octaves (D5-D4, D4-
D3, etc.). Within some of these octaves (D4-D3/tenor register, and D2-D1/deep bass
register) a vibrato was applied, conferring a strong vocal connotation to the glissando in
these selected parts of the sound.13 At the same time five vocal formants, that are clearly
visible at the beginning, seem to collapse into independent strokes from approximately
40” after the beginning of the process. The bandwidths of the formants progressively
decrease, from approximately 80-120 Hz down to very small values (less that 1 Hz at the
end), and make the formants increasingly more resonant. The strokes (or duration of the
grains) get longer and longer, from 7 ms to several seconds, in order to let the resonance
be heard. At the end the spectrum looks much like an additive synthesis sound. The
spectral characteristics (frequencies, amplitudes) of this final part are derived from the
bell-sound model described earlier.14
13 However, as classical singers do not often use such long glissandi, and normally do not add vibrato
while performing a glissando, this example already suggests a kind of imaginary voice even for a more
“vocal” moment.
14 Onemight also perceive an indirect reference to the famous passage in Karlheinz Stockhausen’s
Kontakte where a high pitch (E5) starts a downward glissando ending with slow, isolated, reverberated
impulses tuned to E3.
314
Electronic dramaturgy and computer-aided composition in Re Orso
The sub-patch at the upper-left of Figure 9 (gen-fql) produces the sequence of formant
data. This part of the process is detailed in Figures 11 and 12.
The sequence starts with formant values selected from a vowel database (Phase 1).
During this initial phase the synthesis process renders smooth transitions between the
vowel “u” of a soprano (labelled “su”), an alto (“au”), a tenor (“tu”), and, finally, a
bass (“bu”). The function duplicate-formants does exactly what it says: it duplicates the
formant data so that the total number of formants equals the target number of additive
partials required at the end of the process.
The sequence ends with the values of the bell-sound spectrum (Phase 3). These values
are generated from spectral analysis data in the west_bell sub-patch visible in Figure 12.
A cr-model object is created from a partial tracking analysis and a sequence of hand-
positioned markers (two separate SDIF files exported from AudioSculpt). Only one of
the segments is selected as a source for frequency and pitch material, and bandwidths
are computed from the frequencies using the autobw function of the OM-Chant library.
In between (Figure 11, Phase 2) the interpolation function generates 12 intermediate
states of the FOF generators. Figure 13 shows the contents of the resulting cr-model
object, considered the “knowledge base” (or “skeleton”) of the synthesis process.
Figure 11. Generation of the sequence of formant data for the “Death of the King’s Voice”
synthesis process. Left: formant values extracted from a vowel database. Right: formant values
derived from the spectral analysis of the Winchester Cathedral bell recording (see Figure 12).
Middle: interpolation of the formant values.
315
Marco Stroppa and Jean Bresson
Figure 12. Extracting formant values from the analysis of the bell recording. Note that the
original sound has a spectrum similar to C minor, but for the opera, a spectrum similar to A
minor was needed, hence the multiplication of the frequency values (“Tune to A”).
316
Electronic dramaturgy and computer-aided composition in Re Orso
Figure 14. Control patch mapping the cr-model data to the parameters of the ch-fof synthesis
events in Figure 9.
The “BW scalers” curve defined in the main patch is used to scale the bandwidth
values of the formants during this mapping process. When the sound is clearly vocal
(at the beginning) this factor is 1.0 (that is, the initial values are used), then it rapidly
decreases from 0.02 to very small values (until 0.01, which is 1/100th of the original
bandwidth) for the last part of the process (remember that narrow formants get closer
to additive partials). Using a multiplication factor rather than absolute values has the
advantage that the wider formants will still have, proportionally, a larger bandwidth and
therefore decrease more rapidly, which is exactly what happens with the resonances of
a bell sound. As mentioned previously, the duration of the FOF grains (wdur) also
increases during the process.
317
Marco Stroppa and Jean Bresson
The most difficult task in the design of this synthesis process was to find musically
and perceptually convincing values for the successive intermediate states of the changing
parameters. The interesting moments (for example, when the sound begins perceptually
to become a bell) are very sensitive and usually the optimal values for these parameters
are located within small, cross-dependent intervals. We had to proceed through intensive
trial-and-error to find them.
The linear interpolation of the fundamental frequency, for instance, did not work
well in the logarithmic space of pitches (the glissando is too slow at the beginning, and
quickly moves down to “unmusical” values at the end). A linear pitch interpolation
yields better results, but then the process tends to sound quite mechanical. Using linear
frequency interpolation on portions of the glissando within each octave (see “F0 profile”
on Figure 9) seemed to give the more interesting musical results.
The ch-f0 events controlling the evolution of the fundamental frequency of the syn-
thesis process are generated on the right of Figure 9. Here expand-model and the gen_F0
mapping patch only use the time information from the cr-model and take data from the
fundamental frequency curve. The details of gen_F0 are visible in Figure 15. This patch
mostly consists of a choice of whether vibrato is applied to the current segment—see the
selection list (2 3 4) in Figure 9. If the current segment falls within the selection, the
frequency value is turned into a linear ramp to the next frequency in the list, to which
vibrato is applied using param-process.
Figure 15. Mapping between values of the cr-model of Figure 9, the frequency curve, and the
ch-f0 slots. This patch applies (or not) the vibrato function depending on whether the current
segment is among the selected indices allowed to pass from <select-vib>.
318
Electronic dramaturgy and computer-aided composition in Re Orso
Finally, the ch-transition function visible in Figure 9, associated with the transition
patch, controls the smooth evolution of the whole process, ensuring transitions between
the “continuous” vibrato curves and static values of the fundamental frequency. The
transition patch performs a number of tests on the successive pairs of ch-f0 events
generated in the previous part of the process (are the two events continuous/bpf objects?
do they overlap? etc.) and makes decisions about the data to be produced as an output
to the synthesis process (see [2] for more details on this process).
Conclusions
Re Orso allowed us to explore several exciting æsthetic and scientific concepts. During
the performances we realised that the electronics could, indeed, play the role of a dramatic
character, certainly invisible to the audience, but likely to tell a “story” and to evolve
emotionally as the opera unfolds.
The Chant and OMChroma paradigms, both present in the compositional process,
proved to be complementary: while OMChroma mainly deals with sound structure
at the (discrete) level of the notes, OM-Chant was naturally suited to address issues
of phrasing, which are crucial for the synthesis of sung voice and for conferring to the
synthetic sounds a high degree of expressivity. With these tools, we were able to produce
sounds of high acoustic quality within a reasonable computation time and, especially, to
give to the whole sonic realm the unity for which the composer was searching.
319
Marco Stroppa and Jean Bresson
We are aware that we only skimmed over certain aspects of the synthesis system,
and that much more experience is needed fully to exploit its full potential. The constant
guidance of an æsthetic perspective (that is, synthesising a certain sound family because
of the needs of the libretto) helped us to focus on the most salient musical directions.
This was a crucial aspect of the experience.
References
[1] Carlos Agon, Jean Bresson, Marco Stroppa. “OMChroma: Compositional Control
of Sound Synthesis”. Computer Music Journal, 35(2), 2010.
[2] Jean Bresson, Raphaël Foulon, Marco Stroppa. “Reduction as a Transition
Controller for Sound Synthesis Events”. In FARM—Workshop on Functional Art,
Music, Modeling and Design, ICFP’13. Boston, 2013.
[3] Jean Bresson, Marco Stroppa. “The Control of the CHANT Synthesizer in
OpenMusic: Modelling Continuous Aspects in Sound Synthesis”. In Proceedings
of the International Computer Music Conference. Huddersfield, 2011.
[4] Jean Bresson, Marco Stroppa, Carlos Agon. “Generation and Representation of
Data and Events for the Control of Sound Synthesis”. In Proceedings of the Sound
and Music Computing Conference. Lefkada, 2007.
[5] Arshia Cont, Carlo Laurenzi, Marco Stroppa. “Chromax, the other side of the
spectral delay between signal processing and composition”. In Proceedings of the
International Conference on Digital Audio Effects (DAFx). Maynooth, 2013.
[6] Xavier Rodet. “Time-domain Formant-wave Function Synthesis”. Computer Music
Journal, 8(3), 1984.
[7] Xavier Rodet, Pierre Cointe. “FORMES: Composition and Scheduling of Processes”.
Computer Music Journal, 8(3), 1984.
[8] Xavier Rodet, Yves Potard, Jean-Baptiste Barrière. “The CHANT Project: From
the Synthesis of the Singing Voice to Synthesis in General”. Computer Music Journal,
8(3), 1984.
[9] Marco Stroppa. “Musical Information Organisms: An approach to composition”.
Contemporary Music Review, 4(1), 1989.
[10] Marco Stroppa. “High-Level Musical Control Paradigms for Digital Signal
Processing”. In Proceedings of the International Conference on Digital Audio Effects
(DAFx). Verona, 2000.
[11] Marco Stroppa. “Auf der Suche nach formalen Polyphonien. Zwischen Musiktheorie
und Neurowissenschaft”. Musik & Ästhetik, Heft 01(1), 2012.
Acknowledgements: The authors would like to thank Thibaut Carpentier, Arshia Cont, Nicholas Ellis,
José Fernandez, Raphaël Foulon, Gilbert Nouno and Xavier Rodet from IRCAM, Marlon Schumacher
from CIRMMT/McGill University, and Jean-Baptiste Barrière for their invaluable help and advice.
320
Rima Flow: Oral tradition and
composition
Alessandro Ratoci
Two genres of improvised musical expression, so distant in time and space as the tradi-
tional ottava rima chant of 19th -century rural Italy and the contemporary “beatboxing”
practice of urban hip-hop music, have been the inspiration for my piece Rima Flow for
tuba and electronics.1
Computer-assisted composition techniques have been intensively employed to derive
all of the symbolic (notational) material and the electronic (concrete) sounds directly
from recorded sources of those oral-tradition repertoires. This provided the opportunity
to experience formalisation as a kind of cultural perspective to discover new possibilities
from a traditional source of material.
1 Written during the IRCAM Cursus in composition and music technologies in 2015.
321
Alessandro Ratoci
Beside the use of this form for heroic literature, a parallel use of ottava rima is
consistently found in traditional popular chants of rural areas in the centre of Italy,
passed from generation to generation by word of mouth.
These relations between “high” (literary) and “popular” (oral) culture have been part
of the social identity of Italy from the Middle Ages to the advent of marketed mass-media
culture. As of the beginning of the 20th century, oral poetry still had an important role in
the education of moral and religious values for the lower classes, which were still largely
illiterate. The divulgation of spiritual poems, ranging from Tasso’s Gerusalemme Liberata
to the entirety of Dante’s Divina Commedia, was carried out by popular storytellers with
the encouragement of religious authorities. It was still common in the late 1980s and 90s
to find old rural poets who, despite their semi-illiterate condition, were able perfectly to
sing entire poems dating from the Renaissance or even entire chants of Dante’s Divina
Commedia. As an aid to memory, and to catch the listener’s attention, the poems where
often sung over stereotypical melodies of a simple syllabic character. These melodies were
the base for improvised embellishments and ornamentations which were often exaggerated
to demonstrate the storyteller’s virtuosic abilities. These melismas, extracted from some
original recordings by different performers, are the most characteristic source materials
used for the composition of Rima Flow.
Beatboxing is one of the distinctive elements of the hip-hop culture that formed during
the late 1960s among African American youth of the American suburbs, rapidly becoming
a widespread part of identity in the metropolitan areas of the U.S. and Europe. Important
pioneers of the genre have been, among many others Kyle Jones (a.k.a Scratch), Marcel
Theo Hal (a.k.a Biz Markle), Darren Robinson (a.k.a Buffy) and Kenny Muhammad.
There have already been interesting examples of cross-fertilisation between beatbox-
ing style and extended techniques of contemporary instrumental music. One of the first
instruments to be involved in this practice has been the flute, surely for the peculiar
embouchure which leaves the mouth and the lips complete freedom of movement, and
322
Rima Flow: Oral tradition and composition
probably because of the rich percussive repertoire of effects that where already present
in the Jazz and Rock flute tradition.
It is unknown when beatboxing was first transferred to the flute, though it seems to
have occurred sometime in the early 2000s. Many of the articulatory and extended
techniques involved in beatbox flute performance have been used by composers for
decades. Examples include the “ch” articulations, key-slaps, flutter tonguing, spit
tonguing, and audible breathing in the performances of Jethro Tull’s front man, Ian
Anderson.
Christopher Kuhns [2]
The most relevant exponents of this experience are, among others, Tim Barsky,
Nathan “Flutebox” Lee, and Greg Pattillo. The large and widespread favour that
this genre of performance has encountered from the public through the internet has
encouraged many young instrumentalists to confront themselves with this kind of sound
palette: on the internet is possible to find examples of beatboxing using the most
disparate of wind instruments: saxophones, clarinets, or oboe, even if the presence of
the reed inside the oral cavity of the performer limits the possible percussive techniques.
The large mouthpiece of the tuba is also suitable for these kinds of sound effects and
many of them are already present in the contemporary repertoire of extended techniques.
The idea to include beatboxing elements in Rima Flow came also from the collaboration
with the instrumentalist Jean-Baptiste Renaux, who is both a brilliant classical tuba
player and a virtuoso beatboxer. This collaboration was very inspiring to me and the
final results represented a perfect depiction of the performer’s personality and musical
identity.
A range of heterogeneous sounds was employed to establish a continuum between two
dialectic opposites of sonic identity: the pure sound of the tuba and the singing voice
on the one hand, and harmonic sound and breath noise on the other (see Figure 1).
Ordinary tuba playing techniques have a clear vocal quality, and the very low register
exhibits homologies between the glottal pulses of vocal phonation and the mechanisms of
sound production on brass instruments. At the same time, the large palette of beatbox
vocal percussion, especially if performed inside the mouthpiece of the instrument, can be
employed as an intermediary sound morphology between un-pitched percussive extended
techniques and the unvoiced phonemes of the human voice.
• Singing popular ottava rima poems, commonly know in Tuscany as cantar di poesia,
was often improvised as part of public contests, exactly as in the modern “slam”.
323
Alessandro Ratoci
• Definition of the general form and the sections of the piece, modelled after the
stanza structure of the ottava rima poems.
• Extraction of melodic profiles from audio recordings as a source for melodic material
of both instrumental and electronic parts.
324
Rima Flow: Oral tradition and composition
The density of ornamentation also follows this general distribution, with a more
melismatic sequence at the beginning and end, and a more syllabic setting toward the
centre of the stanza. An approximation of the line durations to natural numbers was
used to structure the internal proportions of the different sections of the piece according
to the following numerical series: (10 7 4 7 4 6 5 8) (7 7 4 5 5 5 5 7) (8 6 4 6 5 7 4 7) (7 7 4 6
5 6 4 7) (8 6 4 5 5 6 4 8) (7 6 5 5 4 5 3 8) (7 5 3 5 5 5 4 6) (7 6 4 4 4 5 4 8) (7 6 5 5 5 6 5 8) (8 7
4 6 5 7 4 6) (7 5 4 5 5 6 5 7) (7 5 4 6 5 6 4 7) (7 6 4 6 5 5 4 7) (7 5 4 5 5 6 4 8) (8 5 4 5 5 5 4 7).
The proportions between the large-scale sections of the piece are also given by the
average duration of each of the eight lines of the stanzas (7 6 4 5 5 6 4 7), as if the whole
piece itself were a section of an ottava rima poem.
325
Alessandro Ratoci
326
Rima Flow: Oral tradition and composition
(a)
(b)
Figure 5. MIDI transcription of a fundamental frequency curve: (a) with maximum precision,
and (b) using syllables.
327
Alessandro Ratoci
Instead of searching for the perfect transcription (in the sense of a resemblance to the
original) I started to use the transcription tools as a first stage of symbolic transformation
of the materials. Figure 7 for instance shows several different transcriptions of the same
musical passage.
2 Thereactive feature of OpenMusic 6.9 (dark-framed boxes in Figure 6) was useful to have instantaneous
visual feedback from the quantisation parameters and their impact on musical notation.
328
Rima Flow: Oral tradition and composition
329
Alessandro Ratoci
• To store all of the original sound files, their segmentation, melodic profile curves,
and transcribed notation.
• To access the materials according to the line segmentation of the poetic text, listen-
ing and confronting the original sound files with a “virtual instrument” rendering
of the transcribed fragments.
• To perform different kinds of filtering and the basic operations of inversion and
retrograde on the melodic profile curves.
• To perform different kinds of symbolic transformations on transcribed materials.
• To re-synthesise the transformed profiles into new sound files.
• To generate vector graphics from the melodic profile curves that can be imported
in music notation software (Finale, Sibelius, etc.)
The filtering operations shown in Figure 8 are necessary to control the complexity
of pitch transcription and to produce either a smoother or less stable quality in re-
synthesised audio. The algorithms are taken from the OM-Fil library by Mikhail
Malt. They include the mean-filter function with different values for <window-size>
and <recursion-level> (respectively, the amount of data that is taken into account for
averaging purposes, and the number of times the process is repeated).
The symbolic operations on transcribed notation are also simple, as I wanted to keep a
degree of resemblance to the original melos, and because the most radical transformations
of the material are already done during the transcription process. These operations can
be summarised in the following categories:
330
Rima Flow: Oral tradition and composition
Figure 8. Profile server: filtering and other operations on fragments of melodic profiles.
Performing the common operations of inversion and retrograde over small subgroups
of notes is a simple and effective way to generate new melodies where the degree of
resemblance to the original material can be precisely controlled. The interval-mult-groups
or retrograde-groups patches in Figure 9 both take a list of chords in their first input and
a list of numbers, corresponding to the segmentation of original melody, in the second.
Figure 10 shows the detail of the interval-mult-group sub-patch. In this example a
multiplication ratio of -1 produces a simple interval inversion using the first element
of each group as the pivot pitch (see Figure 11). In this way an internal hierarchy of
melodic material can be established in which some important notes of the profile are
left recognisable, as they do not vary over successive presentation of the transformed
material.
331
Alessandro Ratoci
332
Rima Flow: Oral tradition and composition
The melodic profiles control the pitch parameter of the synthesis processes and are
altered by other parameters used to obtain transformed versions of the original melisma.
These parameters are:
• <grain-dur>: the length of granular fragments that are obtained by sampling the
melodic curve at regular intervals. Longer grains naturally produce a more “fat”,
but also “blurred”, sound while shorter grains produce a clearer pitch profile, but
with a somewhat “thinner” timber.
• <stretch>: the melodic profile can be stretched (> 1) or compressed (< 1) in time
before re-synthesis.
• <transp>: the melodic fragments can be freely transposed by a given number of
midicents (mc).
333
Alessandro Ratoci
Finally, another possible output related to the analysis and elaboration of melodic
contours is the creation of vector graphic objects (SVG files) to be imported in notation
software for semi-aleatoric notation of instrumental gestures (see Figure 13).
334
Rima Flow: Oral tradition and composition
3 The reduce function takes a list of elements and combines them using an arbitrary binary operation (in
this case joining two sound objects using sound-seq).
335
Alessandro Ratoci
Figure 14. Slicing a sound file. Right: Inside the slicer loop.
336
Rima Flow: Oral tradition and composition
Figure 16. Construction of hybrid sound material (beatboxing sounds and extracted phonemes
from ottava rima chant) by probabilistic juxtaposition.
Conclusions
More than a powerful instrument to speed up long and repetitive tasks while dealing with
musical composition and audio processing, OpenMusic also opens new and interesting
perspectives of a conceptual nature. By unifying in the same workspace aspects of the
compositional process that are generally considered separate—like the generation of a
score, the synthesis of sounds, and electroacoustic transformations—it inspires the sense
of a higher conception of the work as a whole.
337
Alessandro Ratoci
Transcript of the ottava rima poem “Poesia del sogno di un nipote e del nonno morto”,
as recorded by Francesco Ugolini, Firenze, 2006.
Natura dammi soccorso ai miei bisogni, Ora vi è un dazio che non si può soffrire...
che alla fonte della Musa io mi bagni Il prezo è il macinato di mulino;
Quindici ottave vi farò sui sogni Per riportare a casa il sacco pieno
e un numero all’ambo s’accompagni. Ci vuol tre franchi e ottanta, non di neno.
Di rose, gelsomini e catalogni, Rincarato è le biade, paglia e fieno;
di zanzare scorpion piattole e ragni Rincarato è il dormire e la porzione;
Da un sonno tardo su il letto rivolto Rincarato è le case ed il terreno,
Ecco un Nipote che sogna il Nonno morto. Per tre volte è più cara la pigione
E disse: « Come mai? chi vi ci ha porto? Nonno, ci fu la guerra sopra il Reno,
Mentre un dì voi passaste all’altra vita? La Prussia si batté con Napoleone...
Or siete diritto e voi moriste torto; E Pio Nono ha perduto il temporale.
Ditemi, come la sta questa partita? » Nonno, Roma è venuta capitale,
Disse: « Nipote mio, stai bene accorto: Per quanto sia nelle dimostrazioni...
Sono un’ombra terrena in via smarrita; Nonno, ti ho raccontato il bene e il male,
Il pensier mi guidò, se ti contenti, Qua de’ viventi tutti i paragoni.
Di saper le notizie dei viventi. » Nonno, ’gli è un mondo, sai: chi scende e sale,
Le mura non son più sui fondamenti, L’uomo giusto convien così ragioni,
Com’eran prima, che voi bensapete: Tutto t’ho detto in questa parte esterna
Quelli zecchini d’oro e quelli argenti Dimmi come sis ta la in vita eternal...
Ora son fogli, e c’è poche monete! Tristo è colui che l’umano scherna
Son diradati i frati nei conventi, Principiò a dir cosi, parlò al nipote,
E pochissima stima gode il Prete; l’inferno l’è un orribile caverna,
Non è la Religione come una volta... ombre notturne per le vie remote.
« Seguita, gli dicea, che il Nonno ascolta... » Tristo è colui che l’immortale scherna
La campagna ci dà buona raccolta: L’ingresso è tetro che il terror percuote
Per grazia del Ciel, sono più belle: Gole profonde e tempestosi venti
S’empiono i tini, le bigonvìce e sporta. Ripiene d’urli, strepiti e lamenti.
Ma son tanto più care le gabelle! Più giù c’è l’orlo e il cerchio dei serpenti,
Un miglio intero, sai, fuor d’ogni porta, Di mostri di demoni scatenati;
Se tu vedessi, l’hanno fatte belle: Che riguardano tutti i malviventi
Si pesano i barrocci alla stadera, Ch’ebbero al mondo i poveri strapazzati.
E il nome è intitolato la barriera. Ci sono degli avari qui presenti,
Rispose il Nonno: « Ma Nipote, è vera Ci sono fattori milordi preti e frati;
Quel che mi dici tu ne’ tuoi pensieri? » Di spie di ladri di ruffiani d’ogni veleno
Disse il Nipote: « Un altro n’è in carriera: La terza parte dell’ inferno è pieno
Gli hanno disfatto, sai, tanti poderi. La discendenza tua là troverai,
E vi è un vapore da mattina a sera, Genitori, fratelli e le sorelle,
’gli è quello che trasporta i passeggeri... Nonni, bisnonni e figli se ce n’hai
Tranta vagoni s’accatena attorno, Son costì fra gli Arcangeli e le ancelle.
Da Firenze in due ore va a Livorno. Il Padreterno tu saluterai,
Poi c’è le guardie con trombetta e corno, Quel divino che fece opere belle;
Acciò non attraversi il viandante, Cerca di star con lui di notte e giorno,
E un fil di ferro lungo all’intorno, Ti lascio in terra e in paradiso torno.
Che si chiama il telegrafo volante. Si risveglia il nipote tutto adorno,
In un quarto d’ra, sia di notte o giorno, Di fede di speranza e religione;
Si dà notizie per le città tante, Ogni vizio mondano gli fa scorno,
Nella chiamata, come al referire... » Solo confida alla sua vocazione.
Risponde il Nonno: « Tu mi fai stordire! » I dritti del cantore interpretonno,
Un’altra, Nonno, ve ne voglio dire, I segni dati all’immaginazione;
Di quando voi facevi il contadino, L’indice mi svegliò qui all’improvviso,
E che il grano valeva dodici lire, Per risvegliar pensieri in paradiso.
Il più bello e il più caro uno zecchino:
338
Rima Flow: Oral tradition and composition
References
[1] Edward Hirsch. A Poet’s Glossary. Boston: Houghton Mifflin Harcourt, 2014.
[2] Christopher Kuhns. “Beatboxing and the Flute: Its History, Repertoire, and
Pedagogical Importance”. Theses, Treatises and Dissertations, Florida State
University, 2014.
[3] Dan Stowell, Mark D. Plumbley. Characteristics of the beatboxing vocal style.
Technical report, Centre for Digital Music, Department of Electronic Engineering,
Queen Mary, University of London, 2008.
Acknowledgements: I would like to thank Mikhaïl Malt and Jean Bresson for their continuous support
during my time in the Cursus 1 at IRCAM, Jérémie Garcia for the development of the graphic export
routines, and Florent Jacquemard for his interest in the question of rhythmic quantification.
339
Ab-Tasten: Atomic sound modelling
with a computer-controlled grand
piano
Marlon Schumacher
This chapter describes concepts and techniques for the composition of the piece Ab-
Tasten for computer-controlled grand piano and electronics. It will discuss some of
the conceptual implications of sound representations for music creation and introduce a
model for corpus-based atomic decomposition, which served for the composition of both
acoustic and electronic materials. The accurate control of timing and dynamics on the
computer-controlled piano allowed me to compose in a continuum between instrumental
writing and sound modelling, and exploit principles of auditory organisation to create
the illusion of spatial sound synthesis with an acoustic instrument.
Introduction
The piece Ab-Tasten for computer-controlled piano and electronics was commissioned
in 2011 for a live@CIRMMT concert, dedicated to works reflecting the historic, cul-
tural, and technological transformations of the clavier.1 With a history of over four
hundred years of development, its latest embodiment, the modern grand piano, is one of
today’s most versatile instruments. In an organological sense, it stands out as a hybrid
instrument associated with different instrument families, based on its sound producing
medium (vibrating string), excitation mechanism (striking action of hammers), or playing
technique (keyboard interface). Indeed, composers have interpreted the piano in various
ways, e.g. as percussion instrument (Béla Bartók, Szabadban), string instrument (Helmut
1 http://www.cirmmt.org/activities/live-cirmmt/clavisphere/
341
Marlon Schumacher
342
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
343
Marlon Schumacher
each iteration the globally best matching atom (over all possible temporal locations) is
selected, a model is built from the most significant to least significant element (in terms
of energy), rather than from beginning to end of the target sound (in terms of time).
Consequently, the final selection of atoms as well as their temporal positions in the
model are determined in the matching process. A dictionary-based model can be further
analysed to derive higher-level structures (similar to partials derived from STFTs). For
instance, atoms can be grouped based on their own parameters (e.g. amplitude or
duration) or depending on their context (e.g. proximity of atoms in time or frequency).
See also [20] for an overview of dictionary-based methods for analysis, visualisation, and
transformation of audio signals.
344
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
The use of a target sound as a template for sound synthesis based on collections of
concrete sound material relates to a number of techniques, such as audio mosaicing [22],
adaptive micromontage [19], or data-driven concatenative sound synthesis [9].2 Although
these systems offer powerful sound analysis/synthesis techniques, few of them provide
possibilities for linking the models to other compositional processes and symbolic repre-
sentations. One notable example is the work by Einbond et al. using the CataRT system
for feature-based transcription of audio targets in OpenMusic [7]. This representation is
based on a Euclidean vector space of audio descriptors populated with atoms (or units),
which are selected based on their proximity to a target position. One difference between
this approach and CBAD is that its k-nearest-neighbour matching does not iterate over
a residual in an out-of-time context and thus has no model for polyphony or temporality.
Figure 1. Visual artworks by Arcimboldo, Dalí, Pras, Silver. Note the simultaneous perception
of local objects and larger-scale forms.
2 Inanother perspective, the possibility of modelling a sound in several stages with different sound
elements shares similarities with spectral modelling synthesis [16].
345
Marlon Schumacher
The visual analogy lends itself well to illustrating the notion of different modes of
perception as a function of context and resolution: in the case of a target image that
is modelled with only a few smaller-scale images, it will hardly be possible to recognise
the underlying larger-scale form. Instead, the characteristics of the smaller-scale images
themselves and the relationships between them (e.g. relative position, organisation) are
perceived as the content of the image. In the opposite case, with a high number and
density of smaller-scale images (as in a photomosaic), their cumulative features become
perceptually more salient and perception shifts to a more holistic mode, focusing on
macro-scale characteristics. At this resolution, details of the smaller-scale elements will
be barely visible, shifting the mode of perception from recognition of individual objects to
qualities of the “fabrics” of the modelled image (think of individual threads in a carpet).
A well-known example for a similar effect in the musical domain is György Ligeti’s or-
chestral work Atmosphères, in which the dense superposition of instrumental parts shifts
perception from recognition of individual notes and patterns to the textural qualities of
a larger-scale morphology. It is hardly possible to follow the characteristics of individual
instruments, which in this context do not carry individual musical meaning themselves
but rather become properties of a global timbre (a whole that is other than the sum of
its parts).
In Ab-Tasten, the first step for modelling of target sounds was the creation of a dictionary
that is representative of the acoustic grand piano. Simply stated, this means recording
individual piano notes from the acoustic grand piano to be used as atoms, which represent
the smallest indivisible elements of the model. Sampling a keyboard instrument such as
the piano is comparatively straightforward due to its percussive sound production and
structure of discrete keys with orthogonal dimensions of pitch and dynamics. Since
the sound production of the Disklavier can be controlled via MIDI, this sampling process
could be automated via a computer program that triggered individual notes and recorded
the sounds with a microphone. Sampling the entire combinatoric space of possible MIDI
key numbers and velocities would require recording 11176 (88×127) samples, which at an
average duration of 10 seconds would correspond to a corpus of about 31 hours of sound.
Atomic decomposition with such a large dictionary would be impractical in terms of
computing time. Initial tests with the Disklavier revealed that sampling the keys at 127
individual MIDI velocities was unnecessary, as the JND (just-noticeable difference) for
dynamics was at a value of about 4. Moreover, MIDI velocities greater than 120 sounded
unnaturally harsh, while values below 20 would not always accelerate the piano hammers
enough to reach the string and produce a sound. Thus, the 88 keys of the piano were
sampled at 25 distinct MIDI velocities (in the range of 20 to 120 in steps of 4), resulting in
a total of 2200 individual recordings. Since the decay of a piano note varies with its pitch
the duration of the notes was fixed to 15 seconds and the recording trimmed off earlier
if the rms amplitude fell below a threshold of -24dB. The program for carrying out this
sampling process was realised in OpenMusic using the library OM-SoX, which provides
a suite of objects and functions for audio recording and processing [24]. Figure 2 shows
the OpenMusic patch sending MIDI events to the acoustic grand piano for triggering
individual notes and recording the resulting acoustic sounds as audio files.
346
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
On the top left of this figure we see the generation of the lists for MIDI key numbers
and velocities, a number specifying the maximum duration of the recording in seconds,
and a string used as prefix for naming the files (a). Visible on the top right is the outer
loop (sample-robot), which iterates over the list of key numbers (b). On the bottom
left we see the inner loop, which iterates over the list of MIDI velocities and generates
unique filenames (c). The abstraction sox-samplebot (d) performs the sampling: the
function sox-process starts the audio recording. After 0.4 seconds a MIDI “note on”
event is sent, and the program sleeps for 15 seconds. In the meantime, the function
sox-trimsilence trims off “silence” (i.e. audio with an RMS amplitude value below -24dB)
from the beginning and end of the recording, before the “note off” event is sent. The
resulting sound is written to disk as an audio file with a unique name.
Figure 2. The patch used for automated sampling of individual piano sounds.
347
Marlon Schumacher
Figure 3. A patch illustrating atomic sound modelling with the library OM-Pursuit.
3 Thistext describes the library at the time of composing Ab-Tasten (2011). It has since then been
further developed and extended with spectral audio descriptors and a constraint programming system.
348
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
the file path, pitch in midicents, and MIDI velocity. The function soundgrain-decomp
then takes these data, carries out the matching process, and returns three values: a
sound file of the audio synthesis of the model (E), a sound file of the residual (F), and
an SDIF file containing a parametric description of the model (G). At the bottom we
can see OpenMusic’s SDIF editor, which displays information about the model; each
atom is represented as an array with fields for onset time, duration, magnitude and norm
(amplitude), corpus index, file index, and file path.
This parametric description (G) can be converted into OpenMusic objects and
integrated into compositional processes like any other musical data. Figure 4 shows an
OpenMusic patch containing objects and functions for representing and manipulating
a model. The SDIFfile object at the top left (containing the model) is converted to a
sgn-array object, a tabulated structure in which columns represent individual atoms (in
OM-Pursuit referred to as “soundgrains”) and rows represent different parameters of
the atoms (A). This object can be directly used to drive audio synthesis or spatialisation
processes, e.g. using the libraries OM-SoX or OMPrisma [20]. Since the sgn-array
object internally stores pitch and velocity information for the individual soundgrains, it
can also be converted into a score-array object, a similar structure, in which columns
represent MIDI notes (B). This score-array object can be directly converted into a chord-
seq object (visible at the bottom left of the figure), and eventually exported to a MIDI
file: a “score” which can be performed by the computer-controlled piano. OM-Pursuit
includes two higher-order functions which allow connecting a Lisp function or patch in
lambda mode for vertical and horizontal manipulation of these objects: process-array-
comp iterates over columns, e.g. for filtering soundgrains that fall outside a specified pitch
range (C), and process-array-slot allows selecting a row to perform global processing, such
as rescaling of a parameter for the entire object (D).
Figure 4. Left (A, B): Conversion from SDIFfile to sgn-array and score-array. Right (C, D):
two examples for vertical and horizontal processing before converting to chord-seq objects.
349
Temporal Dictionary
Locations
(SDIF) Model
(SDIF) (Audio)
MarlonSound
Schumacher
Target
Atomic Model Sgn-Array Score-Array Chord-Seq MIDI file
Decomposition (SDIF) (Matrix) (Matrix) (Symbolic) (MIDI)
(Audio)
Temporal Dictionary
Locations
(SDIF) Model
(SDIF) (Audio)
Sound
Atomic Model Sgn-Array Score-Array Chord-Seq MIDI file
Target
Decomposition (SDIF) (Matrix) (Matrix) (Symbolic) (MIDI)
(Audio)
Modelling a birdsong
Dictionary with an acoustic grand piano
Temporal
Locations
(SDIF) (SDIF)
Let us look at a concrete example of the creation of musical materials for the piece.
Figure Sound
6 shows a sonogram
Target
Atomic (LPC analysis)
Model of an excerpt
Sgn-Array Score-Array of the birdsong MIDI
Chord-Seq thatfile served as
Decomposition (SDIF) (Matrix) (Matrix) (Symbolic) (MIDI)
a source for developing most of the materials. Besides being an hommage to Messiaen,
(Audio)
the sound was chosen because of its inherent musical qualities including its rhythmic and
motivic structure. Visible as vertical lines in the sonogram are the markers used to specify
Figure 6. Top: Sonogram of birdsong. Bottom: Three models with 10, 50, 850 atoms,
respectively. Grayscale represents dynamics. Note how the matching process determines
polyphony and temporal structure.
possible temporal locations for atoms, defined as an evenly spaced grid with an interval
of 16 milliseconds (256 audio samples at 16 kHz sampling rate). Below the sonogram we
can see three horizontally aligned models whose parameters for the decomposition are
identical except for the number of iterations (10, 50, and 850, respectively). The grayscale
values in the chord-seq objects represent dynamics (as in the sonogram). This example
demonstrates the perceptual shift discussed before as a function of density and resolution;
for the first model consisting of 10 iterations (10 piano notes) it will be hardly possible
to recognise an underlying target sound. Instead, we perceive an abstract musical phrase
350
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
in which each of the individual piano notes are clearly identifiable. At 850 iterations,
in contrast, the same piano notes become micro-scale elements, integrated into a cloud
of sounds whose global features (granularity, pitch contour, etc.) become perceptually
salient. Instead of the relationships between individual notes we perceive an emergent
form on a larger scale.
Once converted into a symbolic representation, models of the birdsong were then
further processed (e.g. quantised, filtered) to create musical materials for the piece. For
the development of these materials I often switched between alternative representations
(e.g. continuous vs. metric time), as each offers different possibilities for manipulation.
For instance, in some situations the model was treated as a continuous signal using
sound-processing techniques (filtering, stretching, resampling), while in other situations
the model was treated as a symbolic musical structure and developed based on intervallic
relationships, harmony, rhythm, etc. Figure 7 shows a model of the complete birdsong
(4 phrases) in the chord-seq object and its quantification via the function omquantify
into a metric representation in a voice object. This is the main theme of the piece as it
appears on the acoustic piano at 1’20”.
Figure 7. Piano model of the birdsong as a chord-seq object (top) and its metric representation
in a voice object (bottom).
351
Marlon Schumacher
352
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
piano parts. This creates an ambiguous situation in which cues for sequential and
simultaneous grouping compete with each other, and where other cognitive processes—
such as attention and expectation—can influence the forming of auditory streams.
Figure 8. Examples of musical structures in 60TET (chord-seq objects) dispatched into five
distinct parts in 12TET (multi-seq objects).
353
Marlon Schumacher
the audience and developing an approach that offers heterogeneous, equally privileged
listening positions, rather than an idealised “projection” of the music which is degraded
for the majority of listeners.
To realise this idea, an unconventional setup was used. The acoustic grand piano
was positioned at the centre of the hall. The audience was seated around it, turned
towards the instrument. Four loudspeakers were placed around the audience, at the
corners of the hall. Each virtual piano part was projected from a dedicated loudspeaker,
which avoided the use of phantom sources and thus the distortion of the spatial image
depending on listener position. The result was an intimate listening situation with
individual auditory perspectives, fusing performance space and listening space. This
setup is shown in Figure 9.
445.11Hz 450.28Hz
440Hz
460.81Hz 455.52Hz
Figure 9. Performance setup for the piece. Note the physical piano at the centre of the
audience and the four loudspeakers at the corners of the hall representing the virtual pianos.
The numbers next to the pianos indicate their diapason in Hertz. (The speakers placed in
between the audience members were used for other electronic parts which are not discussed in
this article.)
354
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
The spatial disposition and the assignment of micro-tuned chromatic scales to differ-
ent loudspeakers result in a morphological correlation between pitch and spatial position.
If we recall the microtonal structures from Figure 8, it can be seen how adjacent micro-
tonal pitches translate to adjacent positions in space, whereas vertical pitch structures
(chords/harmonies) result in spatial constellations. Each listening position provides an
individual auditory perspective for an acoustic situation in which instrumental sounds
emanating from the centre are complemented by microtonal and spatial extensions in the
electronics.
355
Marlon Schumacher
resolution and scale changes the mode of perception: the short-time morphological
features of the birdsong (e.g. pitch contour, amplitude envelope) are changing so slowly
that they shift into the macro-scale of the score, while the decomposition at 15 times
higher temporal resolution brings the micro-scale details in the fine structure of the sound
to the foreground, akin to a “zoom” effect. Figure 10 shows an excerpt from the piece
(6’47”–7’00”) displayed in five chord-seq objects (representing the physical piano and the
four virtual pianos). The rapid chirps from the original birdsong now resemble slowly
modulating, granular glissandi, traversing the individual piano parts.
Figure 10. Atomic model of a time-stretched birdsong displayed in 5 staves (chord-seq objects)
representing the physical and virtual piano parts. The labelled brackets indicate individual
glissandi.
This perceptual situation produces a striking effect: although the individual piano
sounds emanate from discrete spatial locations, their high repetition rate, as well as
the proximity and continuity of pitch, create the perception of distinct streams of piano
sounds that seem to move continuously through the hall and in between the acoustic and
virtual pianos. The global impression of this sonority can be described as an amorphous
granular sound texture in which individual sound sources seem to emerge, move through
space and disappear again. This auditory illusion can be explained by the general
preference of the perceptual system for organising sound events in a way that yields
the simplest interpretation of an auditory scene [14]. As in the Glissando Illusion, the
sounds produced at the different physical locations are timbrally quasi identical, making
it ambiguous whether sound events have been produced by distinct, static sound sources,
or by a single source that has changed its position. As a result, the perceptual system
356
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
prioritises other cues for forming auditory streams. Rather than parsing the acoustic
information in a consistent way that would create complicated temporal/harmonic pat-
terns for each physical sound source, it trades off the misinterpretation of spatial cues in
favour of forming more plausible, simple streams based on continuity and proximity in
pitch and time. The resulting effect is an illusion of spatial movement correlated with
pitch.
The perception of spatial depth and distance might be attributed to the similarity
between the characteristic changes of sound features as a function of distance to the
listener, and as an effect of dynamics on the piano. Indeed, studies of piano timbre have
revealed semantic congruencies and similar pianistic performance strategies to express
the labels “distant” (a spatial attribute), “dark”, and “muddled” (timbral attributes) [1].
Two important perceptual cues to determine the distance of a sound source are its overall
level (which decreases with distance due to the spreading of the wavefront) and relative
high-frequency energy (which decreases with distance due to atmospheric absorption by
water molecules) [21]. A similar change in sound characteristics can be observed as a
function of dynamics on the piano: the lower the velocity when striking a key, the lower
will be the overall level as well as the relative high-frequency energy (spectral centroid)
of the sound. These correlated spectral and level differences of piano tones might be
interpreted by the perceptual system as distance cues, such as a pp note on the piano
sounding more distant as compared with a ff note.4
Closing remarks
In this chapter I discussed some of the conceptual implications of abstract sound repre-
sentations as compositional models and introduced a corpus-based, atomic representation
that establishes a direct link between a concrete sound phenomenon and a collection of
sounds. Combined with the possibilities offered by the robotic grand piano, this model
allows an integrated approach for the composition of materials for both acoustic and
electronic parts by extracting structures at different resolutions and time scales, and
with different sound corpora.
The electronics were conceived as virtual microtonal copies of the acoustic instru-
ment, creating a perceptual ambiguity between acoustic and electronic sounds, between
cumulative whole and constituent parts, and between reality and imagination. Using a
specific spatial disposition of loudspeakers, acoustic instrument, and listeners, together
with a morphological correlation between pitch and spatial location, it was possible to
exploit principles of auditory organisation to create the illusion of depth and spatial
movement and transfer concepts of spatial sound synthesis to an electroacoustic setting.
The libraries OM-Pursuit and OM-SoX provided the functionalities to realise artistic
ideas that would have otherwise been difficult to achieve.
Although it might seem that CBAD is merely another analysis/synthesis technique,
I believe its true potential lies in its conceptual and compositional implications. The
process of atomic modelling raises interesting questions related to context, resolution, and
scale as constructors of modes of perception and, consequently, semantics of materials.
357
Marlon Schumacher
Similar to how spectral representations paved the way for new approaches, the atomic
model offers an alternative paradigm that may inspire new directions for compositional
thinking.
The German title Ab-Tasten has multiple meanings: its literal translation means
“sampling”, as in sampling an audio signal. In this spelling, it also refers to the piano’s
performance interface, meaning “removing the keys”, in the sense of a renewed interpre-
tation of the piano as an acoustic synthesiser. Lastly, it can be interpreted as “touching”,
“sensing”, or “exploring”, as in exploring unknown lands.
References
[1] Michel Bernays. The expression and production of piano timbre: gestural control
and technique, perception and verbalisation in the context of piano performance and
practice. Ph.D. thesis, Université de Montréal, 2013.
[3] Albert S. Bregman. Auditory Scene Analysis: The Perceptual Organization of Sound.
Cambridge, MA: MIT Press, 1994.
[5] Jean Bresson, Carlos Agon, Gérard Assayag (eds.) The OM Composer’s Book 2.
Editions Delatour France/IRCAM-Centre Pompidou, 2008.
[6] Diana Deutsch. “Grouping Mechanisms in Music”. In Diana Deutsch (ed.) The
Psychology of Music. San Diego: Academic Press, second edition, 1999.
[8] Willis D. Ellis (ed.) A Source Book of Gestalt Psychology. Abingdon: Routledge,
1999.
[11] Sune Hye Park. Elements of Impressionism evoked in Debussy and Ravel’s “Reflets
dans l’eau” and “Jeux d’eau”: The theme of water. Ph.D. thesis, University of
Washington, 2012.
358
Ab-Tasten: Atomic sound modelling with a computer-controlled grand piano
[12] Jean-Claude Risset. “Pitch Control and Pitch Paradoxes Demonstrated with
Computer-Synthesized Sounds”. Journal of the Acoustical Society of America,
46(1A), 1969.
[13] Winfried Ritsch. “Robotic Piano Player Making Pianos Talk”. In Proceedings of the
Sound and Music Computing Conference. Padova, 2011.
[14] Angélique A. Scharine, Tomasz R. Letowski. “Auditory Conflicts and Illusions”. In
Clarence E. Rash, Michael B. Russo, Tomasz R. Letowski, Elmar T. Schmeisser
(eds.) Helmet-Mounted Displays: Sensation, Perception and Cognition Issues. Fort
Rucker, AL: US Army Aeromedical Research Laboratory, 2009.
[15] Marlon Schumacher, Jean Bresson. “Spatial Sound Synthesis in Computer-Aided
Composition”. Organised Sound, 15(03), 2010.
[16] Xavier Serra, Julius Smith. “Spectral Modeling Synthesis: A Sound Analy-
sis/Synthesis System Based on a Deterministic Plus Stochastic Decomposition”.
Computer Music Journal, 14(4), 1990.
[17] Robert Silver. “Digital composition of a mosaic image”. US Patent Office, 2000.
[18] Karlheinz Stockhausen, Jerome Kohl. “Clavier Music 1992”. Perspectives of New
Music, 31(2), 1993.
[19] Bob L. Sturm. “Adaptive Concatenative Sound Synthesis and Its Application to
Micromontage Composition”. Computer Music Journal, 30(4), 2006.
[20] Bob L. Sturm, Curtis Roads, Aaron McLeran, John J. Shynk. “Analysis, Visu-
alization, and Transformation of Audio Signals Using Dictionary-Based Methods”.
Journal of New Music Research, 38(4), 2009.
[21] Pavel Zahorik, Douglas S. Brungart, Adelbert W. Bronkhorst. “Auditory distance
perception in humans: A summary of past and present research”. Acta Acustica
United with Acustica, 91(3), 2005.
[22] Aymeric Zils, Francois Pachet. “Musical Mosaicing”. In Proceedings of the
International Conference on Digital Audio Effects (DAFx). Limerick, 2001.
Online
Acknowledgements: This work would not have been possible without the analytical spirit and creative
input of Graham Boyes. I would like to express my gratitude to Philippe Depalle for enlightening
discussions on the signal processing aspects and to Jean Bresson for support with the implementation
in OpenMusic.
359
Appendix
OpenMusic
Visual programs/Patches
A patch is a visual graph representing a program. This is the main programming
interface displayed in the figures of this book. In a patch, boxes are connected, or
“patched” together, in order to define the functional layout of the program. Each box
represents a function and has a set of inputs (at the top of the box) that correspond
to the function’s arguments (or parameters) and a set of outputs (at the bottom of
the box) that correspond to the returned/computed value(s).
The evaluation of a box refers to the operation of executing this function to compute
and return values. If an input of the box is connected to the output of another box, it
means that the value of the second box must be computed and used as an argument for
the first function.
Figure 1 shows a simple patch using the functions for the addition (+) and multipli-
cation (×) of numbers. It is equivalent to the expression (3 + 6) × 100. The value of the
box × at the bottom is the result of the execution of the function × with two arguments:
the first argument is the value of the box +, and the second argument is 100. The value
of the box + is the result of the execution of + with two arguments, 3 and 6.
363
OpenMusic
Abstractions
If the number 3 in the expression (3 + 6) × 100 is made into a variable, the previous
program becomes a single-parameter function defined as: f (x) = (x + 6) × 100. This
simple procedure is called functional abstraction.
Abstraction in OpenMusic consists of creating custom functions graphically. In
Figure 2 (a) the variable x is represented by an arrow-shaped input-box visible at the
top-left of the window. The output-box (at the bottom) corresponds to the value that
will be returned.
The patch (or abstraction) can then be used as an internal function in other Open-
Music patches, as shown in Figure 2 (b). The value(s) connected to the patch box
input(s) will then be bound to the corresponding variable(s).
(a) (b)
Figure 2. (a) The patch patch1 corresponds to the function f (x) = (x + 6) × 100. (b) patch1
is used as a sub-patch in another patch. The value 5 is bound to the patch input. This second
patch corresponds to the expression f (5)/10 = (5 + 6) × 100/10.
364
OpenMusic
Objects
Data structures can be used in OpenMusic patches in order to create and manipulate
objects (particularly musical ones). In terms of object-oriented programming, the
definition of an object is referred to as a class, and the actual object created from a
given class is called an instance.
Object constructors are also represented as graphical boxes in OpenMusic. The
inputs and outputs of these boxes represent the attributes of the object (called the slots
of the class). The evaluation of an object box produces an instance of the corresponding
class: inputs are used as arguments for the construction of such an instance, while outputs
allow access to the values of these attributes for the last created instance.
The patch in Figure 3 contains two boxes corresponding to the class note. The first
input (and the first output) of an object box always corresponds to the instance itself
(the input allows the creation of a new instance by copying an existing one, and the
output simply returns the created instance). The other inputs (and outputs), from left
to right, correspond to the pitch, duration, intensity, and MIDI channel of the note. The
pitch input of the upper note box in Figure 3 is connected to the value 6700 (which
corresponds to G4).1 This value is used by the box +, which increments it by 1200 (12
semitones = one octave). The note box at the bottom of the patch uses this new pitch
value and creates a new note instance (G5).
Object boxes are generally associated with a graphical editor that enables the display
and manual editing of musical data.
1 Pitchesin OpenMusic are expressed in midicents, that is, using the MIDI conventional values (one
unit = one semitone, 60 = middle C) multiplied by 100, in order to allow for the manipulation of
micro-intervals as small as 1/100th semitone. Throughout this book we follow the convention MIDI
note 60 = C4.
365
OpenMusic
The main types of objects available and generally cited in this book are:
• chord: A set of simultaneous notes. A chord is defined with lists of values for
each of its attributes (list of pitches, list of durations, list of dynamics, etc.).
• chord-seq: A sequence of chords labelled with absolute time values in millisec-
onds. The chord-seq is also defined with separate lists for each of its attributes
(lists of lists of pitches, durations, dynamics, etc.) and with a list of onsets for
setting the time-position of each chord.
• voice: A sequence of chords expressed using traditional rhythmic notation. The
voice is defined with a list of chords and a rhythm tree determining the metrical
and rhythmic structure.
366
OpenMusic
Higher-order functions
In functional programming, the term “higher-order” designates a function that generates
another function, or uses a function as one of its arguments. A typical example is the
sort function, which sorts a list according to a given criterion. This criterion can be
any function allowing the comparison of two elements of the list: it can be a simple
test of magnitude (< or >) for comparing numbers, or any other more or less complex
test, depending on the nature of the elements in the list. (Imagine, for instance, a
comparison of harmonicity between chord objects.) This test function could therefore be
(and actually is) a parameter of the generic sort function.
Other frequent higher-order functions are mapping utilities such as mapcar. In
functional programming a mapping consists of applying the same function to all of the
elements of a list: the arguments of the mapcar function are therefore another function
followed by a list of parameters (see Figure 4).
In OpenMusic patches, this notion of higher-order functions is realised through the
use of lambda boxes. A lambda box is a standard function or abstraction box set in a
special state (called “lambda”). When a box is in this state, a small λ icon is displayed
in its upper-left corner. The evaluation of a lambda box does not return the result from
the box computation, but the function defined by this box. The patch in Figure 4 shows
a function box used in the “lambda” mode as an argument of mapcar. The mapcar box
applies this function to the elements of its second argument (the list of pitches at the top
of the patch).
Figure 4. Mapping the function om+ to the elements of a list. The function om+ generates
a list of pitches by adding its two arguments, which can be either numbers or lists of numbers.
In the “lambda” mode this box, connected to (0 700 1200), represents “the function applying
om+ with (0 700 1200)”; that is, f (x) = om + (x, [0, 700, 1200]). The mapcar evaluation returns
the list of resulting lists (a list of lists of pitches, that is, a sequence of chords). The two object
boxes in this patch refer to the class chord-seq.
367
OpenMusic
Figure 5. The omloop box carries out an iteration on a pitch list in order to enrich each
element by two additional pitches. A test (using omif ) decides whether a given pitch should be
processed or not.
368
OpenMusic
Maquettes
The maquette is an original interface designed for the purpose of unifying the concepts
of program and score in OpenMusic. It is a 2-dimensional space in which graphic boxes
(called temporal boxes) can be laid out and organised in time. A temporal box may
contain an isolated musical object, a program (patch) that generates an object, or it
can contain an internal maquette (which enables the user to construct hierarchical
structures).
OpenMusic users typically use this interface to sketch and program the general
structure or form of musical sequences or pieces.
The horizontal axis of the maquette represents time in such a way that the position
and size of the graphic boxes correspond to offset and duration values. The vertical axis
is a dimension that can be used freely in the calculation of the temporal boxes. Temporal
boxes may also be connected to each other and evaluated as standard boxes in a visual
program. Figure 6 is an example of a maquette in which temporal boxes arranged
along the time axis are functionally connected.
Figure 6. Example of a maquette. Each box contains a program computing a musical object.
369
OpenMusic
370
OpenMusic
Online resources
Please consult the OpenMusic website for more information, links to scientific publica-
tions, and download/installation instructions:
http://repmus.ircam.fr/openmusic/
371
Computer music tools and technology
Computer music tools used to complement OpenMusic are cited at different places in
this book. This section provides a brief description and pointers to these technologies
and software.
Software cited
Max [Cycling’74]. Real-time graphical environment for music and audio processing,
widely used for the development of interactive music applications.
https://cycling74.com/products/max/
A number of external objects and libraries for Max are also cited, such as:
• bach, a library of patches and externals for music notation, sequencing, and
computer-aided composition.
http://www.bachproject.net/
• MuBu (Multi-Buffer), a container for sound and motion data providing structured
memory for the processing of recorded data.
http://forumnet.ircam.fr/product/mubu-en/
• Antescofo, a score following system and synchronous programming language for
computer music composition and performance.
http://forumnet.ircam.fr/product/antescofo-en/
Patchwork. Graphical programming environment designed for musical applications.
Antecedent of OpenMusic, developed at IRCAM by Mikael Laurson, Jacques Duthen,
Camilo Rueda, Gérard Assayag and Carlos Agon. Not available anymore on modern
computer platforms.
Csound. Open-source programming language designed and optimised for sound render-
ing and signal processing.
http://www.csounds.com/
CataRT [IRCAM]. Real-time concatenative sound synthesis system, which selects and
plays grains from a corpus of recorded audio, segmented and analysed by descriptors,
according to proximity to a target position in descriptor space. Available both as a
standalone application and as a library for Max.
http://ismm.ircam.fr/catart/
373
Computer music tools and technology
AudioSculpt [IRCAM]. Sound editing and processing program based on the short-term
Fourier transform representation. Provides graphical editing tools for sound analysis,
filtering, cross synthesis, time stretching, transposition, segmentation, etc.
http://forumnet.ircam.fr/product/audiosculpt-en/
The main use of AudioSculpt (and Spear, see below) with OpenMusic consists of
generating spectral analyses that are exported and loaded as input data in the computer-
aided composition environment. The analysis can be more or less precise and faithful
with regard to the original sound, and yield more or less dense sets of functions of time,
frequency, and amplitude (called partials), imported in OpenMusic as SDIF or text files
(see p. 375). Two principal types of analysis are used:
Spear. An application for audio analysis, editing, and synthesis based on the “partial
tracking” technique, representing a sound with many individual sinusoidal tracks.
Developed by Michael Klingbeil.
http://www.klingbeil.com/spear/
Modalys [IRCAM]. Physical modelling sound synthesiser. Modalys is also the name
of the OpenMusic library that works as a front-end to the sound synthesis kernel.
http://forumnet.ircam.fr/product/modalys-en/
374
Computer music tools and technology
OSC. Open Sound Control. Protocol for communication of music and media data
through UDP among computers, sound synthesisers, and other multimedia devices.
Originally developed at CNMAT, UC Berkeley.
http://opensoundcontrol.org/
SDIF. Sound Description Interchange Format. Standard for the interchange of sound
description data, consisting of a basic data format framework and an extensible set of
standard sound descriptions (including spectral, sinusoidal, time-domain, and higher-
level models). Created by IRCAM, CNMAT, and IUA-UPF.
http://www.ircam.fr/sdif
MusicXML [MakeMusic]. Open file format for music notation, based on XML (Exten-
sible Markup Language). MusicXML is used as an import/export format for scores in
many music notation programs.
http://www.musicxml.com/
375
About the authors
377
About the authors
378
About the authors
379
About the authors
380
About the authors
381
About the authors
382
About the authors
383
About the authors
384
Dépôt légal : 3ème trimestre 2016
Imprimé en France