Professional Documents
Culture Documents
The Oxford Handbook of Music and The Brain (2019)
The Oxford Handbook of Music and The Brain (2019)
The Oxford Handbook of Music and The Brain (2019)
Edited by
MICHAEL H. THAUT
and
DONALD A. HODGES
Great Clarendon Street, Oxford, 26 , United Kingdom
Oxford University Press is a department of the University of Oxford. It furthers the University’s
objective of excellence in research, scholarship, and education by publishing worldwide. Oxford is a
registered trade mark of Oxford University Press in the UK and in certain other countries
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or
transmitted, in any form or by any means, without the prior permission in writing of Oxford
University Press, or as expressly permitted by law, by licence or under terms agreed with the
appropriate reprographics rights organization. Enquiries concerning reproduction outside the scope of
the above should be sent to the Rights Department, Oxford University Press, at the address above
You must not circulate this work in any other form and you must impose this same condition on any
acquirer
ISBN 978–0–19–880412–3
ebook ISBN 978–0–19–252613–7
Links to third party websites are provided by Oxford in good faith and for information only. Oxford
disclaims any responsibility for the materials contained in any third party website referenced in this
work.
T C
List of Contributors
SECTION I INTRODUCTION
1. The Neuroscientific Study of Music: A Burgeoning
Discipline
D A. H M H. T
Index
L C
Eckart Altenmüller, Institute of Music Physiology and Musicians’ Medicine (IMMM), University
of Music, Drama and Media, Germany
Aaron L. Berkowitz, Department of Neurology, Brigham and Women’s Hospital, Harvard Medical
School, USA
Emmanuel Bigand, CNRS, UMR5022, Laboratoire d’Etude de l’Apprentissage et du
Développement, Université de Bourgogne, France and Institut Universitaire de France, France
Anthony Brandt, The Shepherd School of Music, USA
Elvira Brattico, Center for Music in the Brain (MIB), Department of Clinical Medicine, Aarhus
University, Denmark and The Royal Academy of Music, Aarhus/Aalborg, Denmark
Thenille Braun Janzen, Music and Health Science Research Collaboratory (MaHRC), University of
Toronto, Canada
Steven M. Demorest, Northwestern University, USA
Michael G. Erkkinen, Department of Neurology, Brigham and Women’s Hospital, USA
Laura Ferreri, Cognition and Brain Plasticity Group, Bellvitge Biomedical Research Institute,
Hospitalet de Llobregat, Barcelona and Department of Cognition, Development and Educational
Psychology, University of Barcelona, Spain. Laboratoire d’Etude des Mécanismes Cognitifs,
Université Lumière Lyon 2, 69676 Lyon, France
Shinichi Furuya, Sony Computer Science Laboratories Inc., Japan
Molly Gebrian, University of Wisconsin-Eau Claire, Department of Music and Theatre Arts, USA
Jessica A. Grahn, Brain and Mind Institute, Western University, Canada
Rachel E. Guetta, The National Center for PTSD, VA Boston Healthcare System, USA
Shantala Hegde, Clinical Neuropsychology and Cognitive Neuroscience Center and Music
Cognition Laboratory, Department of Clinical Psychology, National Institute of Mental Health and
Neurosciences, Bengaluru, India
Donald A. Hodges, University of North Carolina at Greensboro, USA
Timothy L. Hubbard, Arizona State University, USA and Grand Canyon University, USA
Christos I. Ioannou, Institute of Music Physiology and Musicians’ Medicine (IMMM), University
of Music, Drama and Media, Germany
Lutz Jäncke, Division of Neuropsychology, Institute of Psychology, University of Zurich,
Switzerland
Irma Järvelä, Department of Medical Genetics, University of Helsinki, Finland
Patrik N. Juslin, Department of Psychology, Uppsala University, Sweden
Stefan Koelsch, Department for Biological and Medical Psychology, University of Bergen, Norway
Yuko Koshimori, Music and Health Research Collaboratory (MaHRC), University of Toronto,
Canada
Yune S. Lee, Department of Speech and Hearing Science, The Ohio State University, USA
Psyche Loui, Northeastern University, USA
Susan Marsh-Rollo, Auditory Development Lab, McMaster University, Canada
Bjorn Merker, Independent Scholar, Kristianstad, Sweden
Benjamin Morillon, Institut de Neurosciences des Systèmes, Aix-Marseille Université & INSERM,
Marseille, France
Steven J. Morrison, University of Washington, USA
Aline Moussard, Centre de Recherche de l’Institut Universitaire de Gériatrie de Montréal
(CRIUGM), Canada
Adam Ockelford, School of Education, University of Roehampton, London, UK
Sébastien Paquette, International Laboratory for Brain, Music and Sound Research (BRAMS),
Université de Montréal, Québec, Canada
Marcus T. Pearce, Queen Mary University of London, UK and Aarhus University, Denmark
Virginia B. Penhune, Department of Psychology Concordia University, Canada
David Peterson, Institute for Neural Computation, University of California San Diego, USA
Vesa Putkinen, Turku PET Centre, University of Turku, Turku, Finland
Isabelle Royal, Département de psychologie, Université de Montréal, Québec, Canada
Frank Russo, Ryerson University, Canada
Laura S. Sakka, Department of Psychology, Uppsala University, Sweden
Charlene Santoni, Faculty of Music, University of Toronto, Canada
E. Glenn Schellenberg, Department of Psychology, University of Toronto Mississauga, Canada
Daniel S. Scholz, Institute of Music Physiology and Musicians’ Medicine (IMMM), University of
Music, Drama and Media, Germany
Daniele Schön, Institut de Neurosciences des Systèmes, Aix-Marseille Université & INSERM,
France
Michael Schutz, Institute for Music and the Mind, McMaster University, Canada
L. Robert Slevc, Department of Psychology, University of Maryland, USA
Klaus Martin Stephan, SRH Gesundheitszentrum Bad Wimpfen, Germany
Swathi Swaminathan, Rotman Research Institute, Baycrest Health Sciences, Canada
J. Eric T. Taylor, Brain and Mind Institute, Western University, Canada
Mari Tervaniemi, Cognitive Brain Research Unit, Department of Psychology and Logopedics,
Faculty of Medicine, University of Helsinki, Helsinki, Finland and Cicero Learning, Faculty of
Educational Sciences, University of Helsinki, Helsinki, Finland
Corene Thaut, Faculty of Music, University of Toronto, Canada
Michael H. Thaut, Music and Health Science Research Collaboratory (MaHRC), University of
Toronto, Canada
Barbara Tillmann, CNRS, Lyon Neuroscience Research Center, Auditory Cognition and
Psychoacoustics team, France and University of Lyon, France
Laurel J. Trainor, Department of Psychology, Neuroscience & Behavior, McMaster University,
Canada
Pauline Tranchant, Département de psychologie, Université de Montréal, Canada
Christina M. Vanden Bosch der Nederlanden, Brain and Mind Institute, Western University,
Canada
Robin W. Wilkins, University of North Carolina at Greensboro, USA
SECTION I
IN T R OD U C T ION
CHAPT E R 1
THE NEUROSCIENTIFIC
STUDY OF MUSIC: A
BURGEONING DISCIPLINE
D O N A L D A . H O D G E S A N D MI C H A E L H . T H A U T
V B
D
From these earliest explorations into music and the brain, neuromusical
research has exploded in recent decades, as indicated in Fig. 2. What began
as fledgling, pioneering efforts from the 1940s to the 1960s has burgeoned
into a relative flood of publications in the 2000s.
FIGURE 2. The number of published articles obtained from a simple “music and brain” search in
PubMed (https://www.ncbi.nlm.nih.gov/pubmed/).
The six chapters comprising Section IV delve into the ways the brain
responds to music. Once again, we see multiple overlapping and mutually
reinforcing domains. All meaningful musical experiences involve memory
in one way or another. Lutz Jäncke explores discrete, music-only, and
shared memory systems that involve auditory processing, episodic,
autobiographic, semantic, and implicit memories, as well as motor
programs, emotion, and motivation. Each of these components has neural
correlates designed for encoding, storing, and retrieving musical memories.
Such a diffuse and distributed network may help explain commonly
reported musical influences on non-musical memory formation. Psyche
Loui and Rachel Guetta tackle relationships between music and attention,
executive function, and creativity. The topic of attention in music can be
informed by general theories of attention, as well as those specifically
applied to musical stimuli. Passive music listening experiences are less
likely to affect executive functions, but research is ongoing concerning
whether and to what extent active musicing affects executive functions in
terms of near and far transfers and in terms of relevant neural mechanisms.
Attention and executive functions, along with their attendant brain
networks, are both connected to musical creativity.
Patrik Juslin and Laura Sakka provide a thorough and detailed review of
neuroimaging studies related to music and emotion. Although certain brain
regions have been more or less consistently implicated in the processing of
musical emotions, much is still unclear. For example, it is not always
certain in some experimental designs whether participants are “merely”
perceiving or actually experiencing musical emotions. Juslin and Sakka
provide methodological recommendations for moving the field forward.
Neurochemical responses are the basis for musical emotions and Yuko
Koshimori reviews recent work in this emerging field. Musical experiences
induce the release of neurotransmitters (e.g., dopamine, serotonin, and
acetylcholine), neuropeptides (e.g., beta-endorphin, oxytocin, and arginine
vasopressin), steroid hormones (e.g., cortisol), and peripheral immune
biomarkers. In addition to the main area of research concerning
neurochemical responses in music listening and music performance
experiences, another primary course of investigation involves the
intentional manipulation of neurochemicals via music in a variety of health
and wellness issues (e.g., Parkinson’s disease, chronic pain, and stress).
Elvira Brattico’s discussion of neuroaesthetics combines but also moves
beyond the previous chapters in this section; this is another emerging field
that demonstrates the maturing of neuromusical research. Building on
decades of previous work in music perception, cognition, and more recently
emotion, neuroaesthetics investigates matters such as brain areas involved
in liking, preference, and aesthetic judgments. While this undoubtedly
introduces more subjectivity into the discussion, it also moves us closer to a
core human experience that lies at the root of music’s importance.
Music and language are both ubiquitous aspects of the human
experience and questions about the nature of and relationships between the
two have been asked and the answers debated for centuries. Now
neuroscientists are posing new questions, such as “to what extent are music
and language processed in distinct, shared, or homologous networks?”
Daniele Schön and Benjamin Morillon give answers to this and related
questions based on current evidence. They also discuss the effects of
musical experiences on language acquisition and skills.
As was the case with Section III, Section IV demonstrates the
tremendous complexity of human musical experiences from a
neuroscientific standpoint. Steadily, patiently, over a period of time and
with new technologies and methodologies, a clearer picture is emerging.
Authors of the six chapters comprising Section V are all concerned with
unraveling knotty issues surrounding the ways musicianship and brain
function interact with each other. Virginia Penhune begins with the notion
that musical training affects numerous brain structures, including gray and
white matter, auditory cortex and association areas, motor regions, frontal
regions, and parietal cortex. Some variances between adult musicians and
non-musicians may be due to pre-existing differences, but sufficient
research exists to support the contention that long-term musical training
produces many of these changes. Penhune also discusses reasons why
music has such strong effects on brain plasticity.
Irma Järvelä takes us on a tour of genomics, specifically the role of
genetics in human musicality. Genes influencing inner ear development,
auditory pathways, and cognition are all linked to musical aptitude. In
addition, genomics research suggests that music and language have a
common evolutionary heritage and that genes play a role in the effects
music has on the body. Eckart Altenmüller, Shinichi Furuya, Daniel Scholz,
and Christos Ioannou examine the contributions that prolonged extensive
goal-directed practice, multisensory-motor integration, high arousal, and
emotional and social rewards make toward inducing brain plasticity. They
discuss motor planning and control, and finally musician’s dystonia, that is,
plasticity-induced loss of skills or what they call de-expertise.
Michael Erkkinen and Aaron Berkowitz review neuroimaging studies of
music improvisation. Using PET, fMRI, tDCS (transcranial direct current
stimulation), and EEG, researchers have implicated numerous brain regions
involved in the spontaneous creation of music. Overall, improvisation
activates a broad network of brain regions involving cognitive control and
monitoring, motor planning and execution, multimodal sensation,
motivation, emotional/limbic processing, and language regions. Timothy
Hubbard describes and discusses auditory and motor neural mechanisms
supporting musical imagery. Involuntary musical imagery includes
anticipatory musical imagery, musical hallucinations, schizophrenia,
earworms, and synesthesia. Embodied musical imagery is covered in such
examples as spatial and force metaphors, the role of mimicry, the distinction
between the inner ear and inner voice, the effects of mental practice on
performance, musical imagery and dance, and musical affect.
Vesa Putkinen and Mari Tervaniemi are concerned with neural plasticity
in music learning. Focusing primarily on studies employing ERPs derived
from EEG and MEG, they found evidence to support the contention that
musical training enhances domain-general auditory processing skills,
though far transfer to executive functions is less certain. They also contend
that training alone does not account for all the differences between
musicians and non-musicians, as self-selection is a confound in terms of
predisposing factors.
These six chapters push beyond the nature of passive music listening
situations into the realm of active musicing experiences. While we cannot
pretend that we fully understand what is transpiring in the brain of Daniel
Barenboim as he conducts a Mahler symphony, by fits and starts, patient
marching, and occasional leaping, we are moving forward.
MU S IC , T HE B R A IN ,
A N D C U LT U R A L
C ON T E X T S
CHAPT E R 2
DONALD A. HODGES
I C
N
S have long recognized the co-equal roles biology and culture play
in the phenomenon we call music (e.g., Blacking, 1973). Fifty years ago,
Gaston (1968), quoting Dobzhansky, expressed the idea clearly and
succinctly. In asking how we developed characteristics of humanness, he
wrote:
To begin to answer this question, it is not necessary to separate the biology from the culture
of man [italics in the original]. They go hand in hand. “The fact which must be stressed,
because it has frequently been missed or misrepresented, is that the biological and cultural
evolutions are parts of the same process” (Dobzhansky, 1962, p. 22). This means that the
part of man’s culture we call music has a biological as well as a cultural basis. (p. 11)
G I M
B
N P
FIGURE 3. This working model illustrates that development and learning exist on a continuum, as
each independently and simultaneously influence neural plasticity. While development is largely
guided by experience-expectant mechanisms, it also receives input from experience-dependent
mechanisms. Similarly, learning is mostly guided by experience-dependent mechanisms, but also
receives experience-expectant input.
Reprinted from Human Brain Mapping 31(6), Adriana Galván, Neural plasticity of
development and learning, pp. 879–90, Figure 1+, doi.org/10.1002/hbm.21029, Copyright ©
2010, John Wiley and Sons.
Looking for explanations of how these changes occur in the brain leads
us to two basic brain development processes, neural pruning and
myelination, both of which have been implicated in musical studies. Each
process is driven by both genetic instructions and lived experiences.
Neural Pruning
Early on in development, the brain overproduces synapses, the connections
between neurons (Berk, 2017). Different brain regions peak at different
times, but by age 2 there may be as many as 50 percent more synapses in a
given area than will be present during adulthood (Stiles, Reilly, Levine,
Trauner, & Nass, 2012). Following the peak of this rapid proliferation of
synapses, a protracted period of decline extends throughout childhood and
into early adulthood. Operating on a “use it or lose it” basis, unused
synapses are selectively pruned, leaving a sculpted brain (Gogate, Giedd,
Janson, & Rapoport, 2001). The number of possible connections—100,000
trillion (1014) synapses in the cerebral cortex—is far too great to be
determined by genetics alone. Rather, the general outlines are genetically
programmed, with selective pruning guided by sensory and motor
experience, psychoactive drugs, gonadal hormones, parent–child
relationships, peer relationships, stress, intestinal flora, and diet (Kolb &
Gibb, 2011). Changes in cortical thickness as a result of pruning are
associated with behavior.
Sculpting the brain is not simply a matter of deleting unused cells and
synapses. At the same time this is happening, new synapses are being
formed throughout the lifetime. Synapses formed early in life are
“expecting” certain life experiences that will prune them into optimal
networks. Later forming synapses are more localized and specific to
particular learning experiences. “Thus, experiences are changing neural
networks by both adding and pruning synapses” (Kolb & Gibb, 2011, p.
268).
Myelination
As neurons communicate among themselves by creating neural networks,
they have numerous dendrites for input but only one axon for output. Over
time, axons are covered in a fatty sheath called myelin that enhances
transmission speed up to 100 times and improves efficiency (Zull, 2002).
Genetic instructions drive myelination in a process that moves through the
brain from bottom to top and back to front. Thus, it is only in one’s early to
mid-20s that the frontal lobes are fully myelinated, and increasing
myelination is related to enhanced cognitive functioning (Webb, Monk, &
Nelson, 2001). Because myelin is white in appearance, the core of the brain
is called white matter; here, billions of fibers connect different regions of
gray matter into neural networks (Filley, 2005).
Although genetic instructions are essential, myelination is also
responsive to learning experiences, as “neurons that wire together fire
together” and “neurons that fire apart wire apart—or neurons out of sync
fail to link” (Doidge, 2007, pp. 63–64). In other words, when we engage
repeatedly in a thought or action (e.g., practicing scales), the neural
network(s) supporting those processes becomes stronger with repetitive
stimulation (Fields, 2009). Specifically, learning experiences elicit more
wrappings of the axon, making message transmission increasingly efficient.
Thus, for example, Bengtsson et al. (2005) found that practicing the piano
induced changes in white matter plasticity. Changes were greater during
childhood than during adolescence or adulthood.
Improved efficiency comes at a cost, as myelination decreases flexibility
in neural responses. That is, the brain places restrictions on itself such that
what is learned limits what can be learned (Quartz, 2003). The more attuned
to surrounding cultural expressions (e.g., language, music, etc.) children
become, the less responsive they are to other cultural expressions (Pons,
Lewkowicz, Soto-Faraco, & Sebastián-Gallés, 2009). Responding
appropriately to unfamiliar tonal and rhythmic structures becomes more
difficult once one has learned the music of the surrounding culture (Patel,
Meltznoff, & Kuhl, 2004).
C I I I
R M
There is significant evidence that the fetus responds to sounds during the
last trimester before birth, as activations in the primary auditory cortex were
recorded using fMRI in the left hemisphere of fetuses at 33 weeks gestation
(Jardi et al., 2008). Newborns as early as 1–3 days old responded to music
with activations in both hemispheres (Perani et al., 2010). Excerpts of
Western tonal music registered primarily in the right hemisphere (RH),
while altered or dissonant versions reduced RH responses and activated left
inferior frontal cortex and limbic structures. “These results demonstrate that
the infant brain shows a hemispheric specialization in processing music as
early as the first postnatal hours” (Perani et al., 2010, p. 4758). Similarly,
using near-infrared spectroscopy (NISR), researchers found that neonates
registered speech and music sounds in both hemispheres, with more
coherent responses to speech in the left hemisphere (LH) (Kotilahti et al.,
2010). Regarding music specifically, researchers used event-related
potentials (ERP) to determine that newborns can process musical pitch
intervals (Stefanics et al., 2009), distinguish pitch from timbre (Háden et
al., 2009), detect the beat in music (Winkler et al., 2009), and create
expectations for tonal patterns (Carral et al., 2005). While one cannot rule
out the effects of learning entirely, it seems clear that we come into the
world prepared to process musical sounds.
The foregoing suggests inborn proclivities for musical processing, but
not predetermined responses to specific styles of music. After reviewing the
literature, Hannon and Trainor (2007) concluded that neural networks
“become increasingly specialized for encoding the musical structure of a
particular culture” (p. 470). As an example, Shahin, Roberts, & Trainor
(2004) found that auditory evoked potentials in 4- and 5-year-olds were
larger in those who received Suzuki music lessons compared to controls
who did not; even larger responses were generated by tones from the
instrument studied (i.e., piano or violin).
To conclude this section, we look at two studies in which researchers
examined the effects of enculturation more closely. Mehr and colleagues
conducted several experiments designed to explore the ways in which
infants imbue music with social meanings. Five-month-old infants heard
one of two novel songs presented by a parent, by a toy, or by a friendly, but
unfamiliar adult in person and subsequently via video (Mehr, Song, &
Spelke, 2016). Later, these infants heard two novel individuals sing the
familiar and then the unfamiliar song. Those infants who had previously
heard a parent sing the familiar song preferred (i.e., looked longer at) the
new person singing it rather than the new person singing the unfamiliar
song. The amount of exposure to the song received at home was correlated
with the length of selective attention. These effects were not found in the
infants who initially heard the familiar song emanating from a toy or a
socially-unrelated person. Thus, songs sung by caretakers embody social
meanings for five-month-old infants.
In an extension, eleven-month-old infants were randomly assigned to
one of two groups; one group listened to one of two novel songs sung by a
parent, while the others heard a song that emanated from a toy activated by
a parent (Mehr & Spelke, 2017). Subsequently, they viewed a video of two
new people, each singing one of the songs. In a following silent condition,
two people appeared next to each other, each presenting and endorsing an
object, such as a small stuffed toy or models of an apple or pear, and the
infant was allowed to reach toward the objects. Preference was indicated by
eye gaze and touching. Infants in both groups chose the object presented by
the singer of the familiar song. Clearly, infants preferred familiar songs
regardless of whether they were learned by hearing the parent sing or by
playing with a musical toy. Even though both groups chose the familiar
song, infants who heard parents sing the song gazed longer at the object
than those who heard it coming from a toy. Again, music was imbued with
social meanings.
T S M U
The first three categories are based strongly on Nettl’s (2000) gradient-
of-universality approach. The authors then provide a list of seventy items
related to music’s sound structures (i.e., pitch, rhythm, melodic structure
and texture, form, vocal style, expressive devices, and instruments) and
extra-musical features (i.e., contexts, contents, and behavior). Just as the
twelve tones of Western music’s chromatic scale provide for an infinite
number of realizations, so might these putative universals provide the
structure of human music within which the cultural variations are also
infinite. Continued research, however, is critically necessary.
Some candidates for relatively universal functions and roles of music in
worldwide cultures have also been offered (Table 2).
Table 2. Functions and roles of music
Viewed through the lens of cultural neuroscience, the central thesis of this
chapter is that biological and cultural aspects of musical experiences are
inextricably intertwined. Virtually nothing about musical experiences is
purely biological or purely cultural. We might consider a tree with its root
system as a visual analogy (Fig. 4). Let the trunk represent musicality as a
universal aspect of being human. Let the branches represent major cultural
traditions and the smaller twigs and leaves stand for particular musical
genres and styles.1 The cultural distance hypothesis (Demorest & Morrison,
2016) suggests that leaves on the same branch (i.e., nearby musical styles)
are more understandable to listeners than leaves on the opposite side of the
tree. Supporting the visible part of the tree is a dense, deep-seated root
system. These roots represent the supporting biological and cultural
underpinnings of music. Each root is an amalgam of biological and cultural
aspects, such that it is impossible to disentangle the Gordian knot.
FIGURE 4. A visual analogy for human musicality. The roots represent biological and cultural
underpinnings. The trunk represents musicality as a universal aspect of humankind. The branches
represent different cultural traditions and the twigs and leaves represent particular musical genres
and styles.
R
Adachi, M., Trehub, S., & Abe, J.-I. (2004). Perceiving emotion in children’s songs across age and
culture. Japanese Psychological Research 46(4), 322–336.
Allman, W. (1994). The stone age present. New York: Simon & Schuster.
Ambady, N., & Bharucha, J. (2009). Culture and the brain. Current Directions in Psychological
Science 18(6), 342–345.
Arnett, J. (2008). The neglected 95 percent: Why American psychology needs to become less
American. American Psychologist 63(7), 602–614.
Bahat, A. (1980). The musical traditions of the oriental Jews. The World of Music 22(2), 46–55.
Balkwill, L.-L., & Thompson, W. (1999). A cross-cultural investigation of the perception of emotions
in music: Psychophysical and cultural cues. Music Perception 17(1), 43–64.
Balkwill, L.-L., Thompson, W., & Matsunaga, R. (2004). Recognition of emotion in Japanese,
Western, and Hindustani music by Japanese listeners. Japanese Psychological Research 46(4),
337–349.
Bangert, M., & Schlaug, G. (2006). Specialization of the specialized in features of external human
brain morphology. European Journal of Neuroscience 24(6), 1832–1834.
Bengtsson, S., Nagy, Z., Skare, S., Forsman, L., Forssberg, H., & Ullén, F. (2005). Extensive piano
practicing has regionally specific effects on white matter development. Nature Neuroscience 8(9),
1148–1150.
Berk, L. (2017). Development through the lifespan. New York: Pearson Education.
Bermudez, J., & Zatorre, R. (2005). Differences in gray matter between musicians and nonmusicians.
The neurosciences and music II: From perception to performance. Annals of the New York
Academy of Sciences 1060, 395–399.
Blacking, J. (1973). How musical is man? Seattle: University of Washington Press.
Blake, E., & Cross, I. (2008). Flint tools as portable sound-producing objects in the upper paleolithic
context: An experimental study. In P. Cunningham, J. Heeb, & R. Paardekooper (Eds.),
Experiencing archaeology by experiment (pp. 1–19). Oxford: Oxbow Books.
Boiles, C. (1984). Universals of musical behavior: A taxonomic approach. The World of Music 26(2),
50–64.
Brown, S., & Jordania, J. (2011). Universals in the world’s musics. Psychology of Music 41(2), 229–
248.
Carral, V., Huotilainen, M., Ruusuvirta, T., Fellman, V., Näätänen, R., & Escera, C. (2005). A kind of
auditory “primitive intelligence” already present at birth. European Journal of Neuroscience
21(11), 3201–3204.
Chiao, J. (2009). Cultural neuroscience: A once and future discipline. Progress in Brain Research
178, 287–304.
Chiao, J., Li, S., Seligman, R., & Turner, R. (Eds.). (2016). The Oxford handbook of cultural
neuroscience. Oxford: Oxford University Press.
Clayton, M. (2009). The social and personal functions of music in cross-cultural perspective. In S.
Hallam, I. Cross, & M. Thaut (Eds.), The Oxford handbook of music psychology (pp. 35–44).
Oxford: Oxford University Press.
Conard, N., Malina, M., & Münzel, C. (2009). New flutes document the earliest musical tradition in
southwestern Germany. Nature 460(7256), 737–740.
Constable, G. (1973). The Neanderthals. New York: Time-Life Books.
Creel, S., Newport, E., & Aslin, R. (2004). Distant melodies: Statistical learning of nonadjacent
dependencies in tone sequences. Journal of Experimental Psychology 30(5), 1119–1130.
Cross, I. (2007). Music and cognitive evolution. In L. Barrett & R. Dunbar (Eds.), The Oxford
handbook of evolutionary psychology (pp. 649–667). Oxford: Oxford University Press.
Cross, I. (2009–2010). The evolutionary nature of musical meaning. Musicæ Scientiæ, Special Issue
2009–2010, 179–200.
Cross, I., & Morley, I. (2009). The evolution of music: Theories, definitions and the nature of the
evidence. In S. Malloch & C. Trevarthen (Eds.), Communicative musicality (pp. 61–81). Oxford:
Oxford University Press.
Cross, I., Zubrow, E., & Cowan, F. (2002). Musical behaviours and the archaeological record: A
preliminary study. In J. Mathieu (Ed.), Experimental archaeology. British Archaeological Reports
International Series 1035 (pp. 25–34). Oxford: BAR Publishing.
Curtis, M. E., & Bharucha, J. J. (2009). Memory and musical expectation for tones in cultural
context. Music Perception 26(4), 365–375.
Dams, L. (1985). Paleolithic lithophones: Descriptions and comparisons. Oxford Journal of
Archaeology 4(1), 31–46.
Davies, S. (2010). Emotions expressed and aroused by music: Philosophical perspectives. In P. Juslin
& J. Sloboda (Eds.), Handbook of music and emotion: Theory, research, applications (pp. 15–43).
Oxford: Oxford University Press.
DeFelipe, J. (2011). The evolution of the brain, the human nature of cortical circuits, and intellectual
creativity. Frontiers in Neuroanatomy 5(29), 1–17.
Demorest, S., & Morrison, S. (2003). Exploring the influence of cultural familiarity and expertise on
neurological responses to music. Annals of the New York Academy of Sciences 999, 112–117.
Demorest, S., & Morrison, S. (2016). Quantifying culture: The cultural distance hypothesis of
melodic expectancy. In J. Chiao, S.-C. Li, R. Seligman, & R. Turner (Eds.), The Oxford handbook
of cultural neuroscience (pp. 183–196). Oxford: Oxford University Press.
Demorest, S., Morrison, S., Beken, M., & Jungbluth, D. (2008). Lost in translation: An enculturation
effect in music memory performance. Music Perception 25(3), 213–223.
Demorest, S., Morrison, S., Nguyen, V., & Bodnar, E. (2016). The influence of contextual cues on
cultural bias in music memory. Music Perception 33(5), 590–600.
Demorest, S., Morrison, S., Stambaugh, L., Beken, M., Richards, T., & Johnson, C. (2010). An fMRI
investigation of the cultural specificity of music memory. Social Cognitive and Affective
Neuroscience 5(2–3), 282–291.
Demorest, S., & Osterhout, L. (2012). ERP responses to cross-cultural melodic expectancy
violations. Annals of the New York Academy of Sciences 1252, 152–157.
Dissanayake, E. (2009). Root, leaf, blossom, or bole: Concerning the origin and adaptive function of
music. In S. Malloch & C. Trevarthen (Eds.), Communicative musicality: Exploring the basis of
human companionship (pp. 17–30). Oxford: Oxford University Press.
Dobzhansky, T. (1962). Mankind evolving. New Haven, CT: Yale University Press.
Doidge, N. (2007). The brain that changes itself. New York: Penguin.
Drake, C., & Ben El Heni, J. (2003). Synchronizing with music: Intercultural differences. Annals of
the New York Academy of Sciences 999, 429–437.
Drake, C., & Bertrand, D. (2001). The quest for universals in temporal processing in music. Annals
of the New York Academy of Sciences 930, 17–27.
Dubos, R. (1981). Celebrations of life. New York: McGraw-Hill.
Elbert, T., Pantev, C., Wienbruch, C., Rockstroh, B., & Taub, E. (1995). Increased cortical
representation of the fingers of the left hand in string players. Science 270(5234), 305–307.
Fields, D. (2009). The other brain. New York: Simon & Schuster.
Filley, C. (2005). White matter and behavioral neurology. In J. Ulmer, L. Parsons, M. Moseley, & J.
Gabrieli (Eds.), White matter in cognitive neuroscience. Annals of the New York Academy of
Sciences 1064, 162–183.
Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., … Koelsch, S. (2009).
Universal recognition of three basic emotions in music. Current Biology 19(7), 573–576.
Galvan, A. (2010). Neural plasticity of development and learning. Human Brain Mapping 31(6),
879–890.
Gaser, C., & Schlaug, G. (2003). Gray matter differences between musicians and nonmusicians.
Annals of the New York Academy of Sciences 999, 514–517.
Gaston, E. (1968). Man and music. In E. Gaston (Ed.), Music in therapy (pp. 7–29). New York:
Macmillan.
Gogate, N., Giedd, J., Janson, K., & Rapoport, J. (2001). Brain imaging in normal and abnormal
brain development: New perspectives for child psychiatry. Clinical Neuroscience Research 1(4),
283–290.
Gregory, A. (1997). The roles of music in society: The ethnomusicological perspective. In D.
Hargreaves & A. North (Eds.), The social psychology of music (pp. 123–140). Oxford: Oxford
University Press.
Gregory, A., & Varney, N. (1996). Cross-cultural comparisons in the affective response to music.
Psychology of Music 24(1), 47–52.
Háden, G., Stefanics, G., Vestergaard, M., Denham, S., Sziller, I., & Winkler, I. (2009). Timbre-
independent extraction of pitch in newborn infants. Psychophysiology 46(1), 69–74.
Han, S., & Ma, Y. (2015). A culture–behavior–brain loop model of human development. Trends in
Cognitive Sciences 19(11), 666–676.
Han, S., Northoff, G., Vogeley, K., Wexler, B., Kitayama, S., & Varnum, M. (2013). A cultural
neuroscience approach to the biosocial nature of the human brain. Annual Review of Psychology
64, 335–359.
Hannon, E. E., & Trainor, L. J. (2007). Music acquisition: Effects of enculturation and formal
training on development. Trends in Cognitive Sciences 11(11), 466–472.
Hannon, E. E., & Trehub, S. E. (2005). Metrical categories in infancy and adulthood. Psychological
Science 16(1), 48–55.
Hodges, D., Burdette, J., & Hairston, D. (2005). Aspects of multisensory perception: The integration
of visual and auditory information processing in musical experiences. In G. Avanzini, L. Lopez, S.
Koelsch, & M. Majno (Eds.), The neurosciences and music II: From perception to performance.
Annals of the New York Academy of Sciences 1060, 175–185.
Hodges, D., Hairston, W., Maldjian, J., & Burdette, J. (2010). Keeping an open mind’s eye:
Mediation of cross-modal inhibition in music conductors. In S. M. Demorest, S. J. Morrison, & P.
S. Campbell (Eds.), Proceedings of the 11th International Conference on Music Perception and
Cognition (ICMPC 11) (pp. 415–416). Seattle, Washington.
Hoffman, D., Standish, C., Garcia-Diez, M., Pettitt, P., Milton, J., Zilhão, J., … Pike, A. (2018). U-
Th dating of carbonate crusts reveals Neandertal origin of Iberian cave art. Science 359(6378),
912–915.
Hornbostel, E., & Sachs, C. (1992). Classification of musical instruments. In H. Meyers (Ed.),
Ethnomusicology: An introduction (pp. 444–461). New York: W. W. Norton.
Hutchinson, S., Lee, L., Gaab, N., & Schlaug, G. (2003). Cerebellar volume of musicians. Cerebral
Cortex 13(9), 943–949.
Hyde, K., Lerch, J., Norton, A., Forgeard, M., Winner, E., Evans, A., & Schlaug, G. (2009). Musical
training shapes brain development. Journal of Neuroscience 20(10), 3019–3025.
Jardi, R., Pins, D., Houfflin-Debarge, V., Chaffiotte, C., Rocourt, N., Pruvo, J.-P., … Thomas, P.
(2008). Fetal cortical activation to sound at 33 weeks of gestation: A functional MRI study.
NeuroImage 42(1), 10–18.
Jonaitis, E., & Saffran, J. (2009). Learning harmony: The role of serial statistics. Cognitive Science
33(5), 951–968.
Kanduri, C., Kuusi, T., Ahvenainen, M., Philips, A., Lähdesmäki, H., & Järvelä, I. (2015a). The
effect of music performance on the transcriptome of professional musicians. Scientific Reports 5,
9506. doi:10.1038/srep09506
Kanduri, C., Raijas, P., Ahvenaninen, M., Phillips, A., Ukkola-Vuoti, L., Lähdesmäki, H., & Järvelä,
I. (2015b). The effect of listening to music on human transcriptome. PeerJ 3, e830.
doi:10.7717/peerj.830
Karuza, E., Newport, E., Aslin, R., Starling, S., Tivarus, M., & Bavelier, D. (2013). The neural
correlates of statistical learning in a word segmentation task: An fMRI study. Brain and Language
127(1), 46–54.
Kendig, F., & Levitt, G. (1982). Overture: Sex, math and music. Science Digest 90(1), 72–73.
Kessler, E. J., Hansen, C., & Shepard, R. N. (1984). Tonal schemata in the perception of music in
Bali and the West. Music Perception 2(2), 131–165.
Kirham, N., Slemmer, J., & Johnson, S. (2002). Visual statistical learning in infancy: Evidence for a
domain general learning mechanism. Cognition 83(2), B35–B42.
Koelsch, S., Schröger, E., & Tervaniemi, M. (1999). Superior pre-attentive auditory processing in
musicians. NeuroReport 10(6), 1309–1313.
Kolb, B., & Gibb, R. (2011). Brain plasticity and behaviour in the developing brain. Journal of
Canadian Child Adolescent Psychiatry 20(4), 265–276.
Kotilahti, K., Nissilä, I., Näsi, T., Lipiäinen, L., Noponen, T., Meriläinen, P., … Fellman, V. (2010).
Hemodynamic responses to speech and music in newborn infants. Human Brain Mapping 31(4),
595–603.
Krumhansl, C. L., Toivanen, P., Eerola, T., Toiviainen, P., Järvinen, T., & Louhivuori, J. (2000).
Cross-cultural music cognition: Cognitive methodology applied to North Sami yoiks. Cognition
76(1), 13–58.
Kuhl, P., & Rivera-Gaxiola, M. (2008). Neural substrates of language acquisition. Annual Review of
Neuroscience 31, 511–534.
Kunej, D., & Turk, U. (2000). New perspectives on the beginnings of music: Archaeological and
musicological analysis of a middle Paleolithic bone “flute.” In N. Wallin, B. Merker, & S. Brown
(Eds.), The origins of music (pp. 235–268). Cambridge, MA: MIT Press.
Lauuka, P., Eerola, T., Thingujan, N., & Yamasaki, T. (2013). Universal and culture-specific factors
in the recognition and performance of musical affect expressions. Emotion 13(3), 434–449.
List, G. (1984). Concerning the concept of the universal and music. The World of Music 26(2), 40–
47.
Lomax, A. (1968). Folk song style and culture. New Brunswick, NJ: Transaction Books.
McDermott, J. (2008). The evolution of music. Nature 453(7193), 287–288.
Manuck, S., & McCaffery, J. (2014). Gene–environment interaction. Annual Review of Psychology
65, 41–70.
Manzano, O., & Ullén, F. (2018). Same genes, different brains: Neuroanatomical differences between
monozygotic twins discordant for musical training. Cerebral Cortex 28(1), 1–8, 387–394.
Marek, G. (1975). Toscanini. London: Vision Press.
Mason, O. (1897). Geographical description of the musical bow. American Anthropologist 10(11),
377–380.
Matsunaga, R., Yokosawa, K., & Abe, J. (2012). Magnetoencephalography evidence for different
brain subregions serving two musical cultures. Neuropsychologia 50(14), 3218–3227.
Mehr, S., Song, L., & Spelke, E. (2016). For 5-month-old infants, melodies are social. Psychological
Science 27(4), 486–501.
Mehr, S., & Spelke, E. (2017). Shared musical knowledge in 11-month-old infants. Developmental
Science 21(2), e12542. doi:10.1111/desc.12542
Merriam, A. (1964). The anthropology of music. Chicago, IL: Northwestern University Press.
Mithen, S. (2006). The singing Neanderthals: The origins of music, language, mind, and society.
Cambridge, MA: Harvard University Press.
Morley, I. (2006). The evolutionary origins and archaeology of music: An investigation into the
prehistory of human musical capacities and behaviors (Doctoral dissertation). University of
Cambridge, Cambridge. Darwin College Research Reports, DCRR-002. Retrieved from
https://www.darwin.cam.ac.uk/drupal7/sites/default/files/Documents/publications/dcrr002.pdf
Morrison, S., & Demorest, S. (2009). Cultural constraints on music perception and cognition. In J. Y.
Chiao (Ed.), Progress in brain research, Vol. 178: Cultural neuroscience: Cultural influences on
brain function (pp. 67–77). Amsterdam: Elsevier.
Morrison, S., Demorest, S., Aylward, E., Cramer, S., & Maravilla, K. (2003). fMRI investigation of
cross-cultural music comprehension. NeuroImage 20(1), 378–384.
Morrison, S., Demorest, S., Campbell, P., Bartolome, S., & Roberts, J. (2013). Effect of intensive
instruction on elementary students’ memory for culturally unfamiliar music. Journal of Research
in Music Education 60(4), 363–374.
Morrison, S., Demorest, S., & Stambaugh, L. (2008). Enculturation effects in music cognition: The
role of age and music complexity. Journal of Research in Music Education 56(2), 118–129.
Mosing, M., Madison, G., Pedersen, N., Kuja-Halkola, R., & Ullén, F. (2014). Practice does not
make perfect: No causal effect of music practice on music ability. Psychological Science 25(9),
1795–1803.
Mosing, M., Peretz, I., & Ullén, F. (2018). Genetic influences on music expertise. In D. Hambrick, G.
Campitelli, & B. Macnamara (Eds.), The science of expertise: Behavioral, neural, and genetic
approaches to complex skill (pp. 272–282). New York: Routledge.
Mumford, L. (1967). The myth of the machine. New York: Harcourt Brace Jovanovich.
Münte, T., Altenmüller, E., & Jäncke, L. (2002). The musician’s brain as a model of neuroplasticity.
Nature Reviews Neuroscience 3(6), 473–478.
Münte, T., Kohlmetz, C., Nager, W., & Altenmüller, E. (2001). Superior auditory spatial tuning in
conductors. Nature 409(6820), 580.
Nan, Y., Knösche, T., & Frederici, A. (2006). The perception of musical phrase structure: A cross-
cultural ERP study. Brain Research 1094(1), 179–191.
Nan, Y., Knösche, T., Zysset, S., & Frederici, A. (2008). Cross-cultural music phrase processing: An
fMRI study. Human Brain Mapping 29(3), 312–328.
Nettl, B. (1977). On the question of universals. The World of Music 19, 2–13.
Nettl, B. (1983). The study of ethnomusicology. Urbana, IL: University of Illinois Press.
Nettl, B. (2000). An ethnomusicologist contemplates universals in musical sound and culture. In N.
Wallin, B. Merker, & S. Brown (Eds.), The origins of music (pp. 463–472). Cambridge, MA: MIT
Press.
Nettl, B. (2005). The study of ethnomusicology: Thirty-one issues and concepts. Champaign, IL:
University of Illinois Press.
Neuhaus, C. (2003). Perceiving musical scale structures: A cross-cultural event-related brain
potentials study. Annals of the New York Academy of Sciences 999, 184–188.
Nketia, J. (1984). Universal perspectives in ethnomusicology. The World of Music 26(2), 3–20.
Norton, A., Winner, E., Cronin, K., Overy, K., Lee, D., & Schlaug, G. (2005). Are there pre-existing
neural, cognitive, or motoric markers for musical ability? Brain and Cognition 59(2), 124–134.
Oikkonen, J., & Järvelä, I. (2014). Genomics approaches to study musical aptitude. Bioessays 36(11),
1102–1108.
Pantev, C., Oostenveld, R., Engelien, A., Ross, B., Roberts, L., & Hoke, M. (1998). Increased
auditory cortical representation in musicians. Nature 392(6678), 811–814.
Pantev, C., Roberts, L., Schultz, M., Engelien, A., & Ross, B. (2001). Timbre-specific enhancement
of auditory cortical representations in musicians. Neuroreport 12(1), 169–174.
Patel, A., Meltznoff, A., & Kuhl, K. (2004). Cultural differences in rhythm perception: What is the
influence of native language? In S. Lipscomb, R. Ashley, R. Gjerdingen, & P. Webster (Eds.),
Proceedings of the 8th International Conference on Music Perception and Cognition. Evanston,
IL: Northwestern University. CD-ROM.
Perani, D., Saccuman, M., Scifo, P., Spada, D., Andreolli, G., Rovelli, R., … Koelsch, S. (2010).
Functional specializations for music processing in the human newborn brain. Proceedings of the
National Academy of Sciences 107(10), 4758–4763.
Peretz, I., Brattico, E., Järvenpää, M., & Tervaniemi, M. (2009). The amusic brain: In tune, out of
key, and unaware. Brain 132(5), 1277–1286.
Peretz, I., Saffran, J., Schön, D., & Gosselin, N. (2012). Statistical learning of speech, not music, in
congenital amusia. Annals of the New York Academy of Sciences 1252, 361–367.
Pons, F., Lewkowicz, D., Soto-Faraco, S., & Sebastián-Gallés, N. (2009). Narrowing of intersensory
speech perception in infancy. Proceedings of the National Academy of Sciences, 106(26), 10598–
10602.
Prideaux, T. (1973). Cro-Magnon man. New York: Time-Life Books.
Quartz, S. (2003). Learning and brain development: A neural constructivist perspective. In P. Quinlan
(Ed.), Connectionist models of development (pp. 279–309). New York: Psychology Press.
Rampon, C., Jiang, C., Dong, H., Tang, Y.-P., Lockhart, D., Schultz, P., … Hu, Y. (2000). Effects of
environmental enrichment on gene expression in the brain. Proceedings of the National Academy
of Sciences 97(23), 12880–12884.
Saffran, J. (2003). Absolute pitch in infancy and adulthood: The role of tonal structure.
Developmental Science 6(1), 35–47.
Saffran, J., Aslin, R., & Newport, E. (1996). Statistical learning by 8-month-old infants. Science
274(5294), 1926–1928.
Saffran, J., & Griepentrog, G. (2001). Absolute pitch in infant auditory learning: Evidence for
developmental reorganization. Developmental Psychology 37(1), 74–85.
Saffran, J., Johnson, E., Aslin, R., & Newport, E. (1998). Statistical learning of tone sequences by
human infants and adults. Cognition 70(1), 27–52.
Schlaug, G., Forgeard, M., Zhu, L., Norton, A., Norton, A., & Winner, E. (2009). Training-induced
neuroplasticity in young children. The neurosciences and music III. Annals of the New York
Academy of Sciences 1169, 205–208.
Schlaug, G., Norton, A., Overy, K., & Winner, E. (2005). Effects of music training on the child’s
brain and cognitive development. The neurosciences and music II: From perception to
performance. Annals of the New York Academy of Sciences 1060, 219–230.
Shahin, A., Roberts, L., & Trainor, L. (2004). Enhancement of auditory cortical development by
musical experience in children. NeuroReport 15(12), 1917–1921.
Smith, H. (1953). From fish to philosopher. Garden City, NY: Doubleday Anchor.
Soley, G., & Hannon, E. (2010). Infants prefer the musical meter of their own culture: A cross-
cultural comparison. Developmental Psychology 46(1), 286–292.
Stefanics, G., Háden, G., Sziller, I., Balázs, L., Beke, A., & Winkler, I. (2009). Newborn infants
process pitch intervals. Clinical Neurophysiology 120(2), 304–308.
Stiles, J., Reilly, J., Levine, S., Trauner, D., & Nass, R. (2012). Neural plasticity and cognitive
development: Insights from children with perinatal brain injury. Oxford: Oxford University Press.
Tervaniemi, M., Kujala, A., Alho, K., Virtanen, J., Ilmoniemi, R., & Näätänen, R. (1999). Functional
specialization of the human auditory cortex in processing phonetic and musical sounds: A
magnetoencephalographic (MEG) study. NeuroImage 9(3), 330–336.
Tillmann, B., & McAdams, S. (2004). Implicit learning of musical timbre sequences: Statistical
regularities confronted with acoustical (dis)similarities. Journal of Experimental Psychology:
Learning, Memory, and Cognition 30(5), 1131–1142.
Trainor, L., Marie, C., Gerry, D., Whiskin, E., & Unrau, A. (2012). Becoming musically enculturated:
Effects of music classes for infants on brain and behavior. Annals of the New York Academy of
Sciences 1252, 129–138.
Trehub, S., Unyk, A., & Trainor, L. (1993). Adults identify infant-direct music across cultures. Infant
Behavior and Development 16(2), 193–211.
Turk-Browne, N., Jungé, J., & Scholl, B. (2005). The automaticity of visual statistical learning.
Journal of Experimental Psychology: General 134(4), 552–564.
Turner, R., & Ioannides, A. (2009). Brain, music and musicality: Inferences from neuroimaging. In S.
Malloch & C. Trevarthen (Eds.), Communicative Musicality (pp. 147–181). Oxford: Oxford
University Press.
Ullén, F., Hambrick, D., & Mosing, M. (2016). Rethinking expertise: A multifactorial gene–
environment interaction model of expert performance. Psychological Bulletin 142(4), 427–446.
Wade, B. (2009). Thinking musically: Experiencing music, expressing culture (2nd ed.). New York:
Oxford University Press.
Wakin, D. (2006). John Cage’s long music composition in Germany changes a note. New York Times,
May 6. Retrieved September 26, 2017 from
http://www.nytimes.com/2006/05/06/arts/music/06chor.html
Webb, S., Monk, C., & Nelson, C. (2001). Mechanisms of postnatal neurobiological development:
Implications for human development. Developmental Neuropsychology 19(2), 147–171.
Wilson, F. (1998). The hand: How its use shapes the brain, language, and human culture. New York:
Vintage Books.
Winkler, I., Háden, G., Ladining, O., Sziller, I., & Honing, H. (2009). Newborn infants detect the
beat in music. Proceedings of the National Academy of Sciences 106(7), 2468–2471.
Wong, P., Roy, A., & Margulis, E. (2009). Bimusicalism: The implicit dual enculturation of cognitive
and affective systems. Music Perception 27(2), 81–88.
Yi, T., McPherson, G., Peretz, I., Berkovic, S., & Wilson, S. (2014). The genetic basis of music
ability. Frontiers in Psychology 5(658), 1–19.
Yi, T., McPherson, G., & Wilson, S. (2018). The molecular genetic basis of music ability and music-
related phenotypes. In D. Hambrick, G. Campitelli, & B. Macnamara (Eds.), The science of
expertise: Behavioral, neural, and genetic approaches to complex skill (pp. 283–303). New York:
Routledge.
Zull, J. (2002). The art of changing the brain: Enriching the practice of teaching by exploring the
biology of learning. Sterling, VA: Stylus Publishing.
1
To be more accurate, each leaf should have a different shape to represent the individuality of
various musical styles.
CHAPT E R 3
C U LT U R A L D I S TA N C E : A
C O M P U TAT I O N A L
A P P R O A C H TO E X P L O R I N G
C U LT U R A L I N F L U E N C E S
ON MUSIC COGNITION
S T E V E N J. MO R R I S O N, S T E V E N M. D E MO R E S T, A N D
MA R C U S T. P E A R C E
R L
C D
Limitations
The analysis of two or more types of music along any given musical
parameter (for example, pitch as in the illustration above) or combination of
parameters imposes the assumption that such an analysis is valid within
each music type. While a music tradition such as Western art music (at least
that from or deriving from the common practice period of approximately
the mid-seventeenth to early twentieth centuries) has a well-established
history of analysis and interpretation based, in part, on both sequential and
concurrent pitch interval relationships, the same may not be said of other
traditions. Tools such as IDyOM offer the flexibility to examine cultural
distance according to a variety of individual or combinations of musical
parameters. Nevertheless, any specific configuration runs the risk of
privileging one parametric hierarchy over another. Thus, in terms of cross-
cultural research, such statistical models will virtually always impose the
perspective of a particular music tradition, at least to some degree.
This limitation has ramifications for fully comparative studies in that the
degree to which a parameter holds primacy for one set of participants may
not hold true for the other. Much as emotion recognition, so familiar to the
experience of westernized listeners, did not figure meaningfully in the
music tradition of the Mafa (Fritz, 2013), the statistical likelihood of
patterns of pitch may contribute less to musical thinking among Rwandans
and more to North Americans (as in Cameron et al., 2015) than does the
complexity of patterns of rhythm. In this way, cultural distance is a tool
through which one can isolate norms for one or more musical parameters as
well as provide a particular perspective on musical meaning-making.
A related limitation is that IDyOM currently requires symbolic score-
like input in which notes are represented as discrete events with discrete
properties (e.g., onset time, pitch). This does not readily accommodate
musical cultures which depend heavily on timbral, dynamic, or textural
changes. The same is true of musical cultures that have no written tradition,
where the distinction between composition and performance is blurred or
nonexistent or where music is inextricably combined with other modes of
communication (Cross, 2014).
Despite the emphasis here on the advantageous aspects of familiarity,
without question novelty is an attractive characteristic of music. Models of
musical expectancy (e.g., Huron, 2006; Meyer, 1956) describe the interest
inherent in and stimulation derived from that which is unfamiliar and
surprising in music. The constant curiosity for new musical ideas suggests
ongoing willingness to explore less “predictable” musical scenarios. With
much of the world’s music readily—and in many cases instantly—
accessible, such willingness leads as easily to unfamiliar music traditions as
to the remoter corners of one’s own. We have used cultural distance as a
means of explaining processing difficulties (as operationalized by
recognition memory); however, it is equally viable as a tool to examine
such positive aspects of music experience as interest and surprise. Although
Cook (2008) was referring specifically to musicologists, his description can
arguably be construed more broadly: “Practically all of us are at least to
some degree musically multilingual … as a result one understands even the
tradition(s) in which one is most ‘at home’ as options amongst other
options, understands them in relation to other traditions rather than as
absolutes” (p. 63).
R
Agres, K., Abdallah, S., & Pearce, M. T. (2018). Information-theoretic properties of auditory
sequences dynamically influence expectation and memory. Cognitive Science 42(1), 43–76.
Baek, Y. M. (2015). Relationship between cultural distance and cross-cultural music video
consumption on YouTube. Social Science Computer Review 33(6), 730–748.
Balkwill, L. L. (2006). Perceptions of emotion in music across cultures. Paper presented at Emotional
Geographies: The Second International & Interdisciplinary Conference, May, Queen’s University,
Kingston, Canada.
Balkwill, L. L., & Thompson, W. F. (1999). A cross-cultural investigation of the perception of
emotion in music: Psychophysical and cultural cues. Music Perception 17(1), 43–64.
Balkwill, L. L., Thompson, W. F., & Matsunaga, R. (2004). Recognition of emotion in Japanese,
Western, and Hindustani music by Japanese listeners. Japanese Psychological Research 46(4),
337–349.
Begleiter, R., El-Yaniv, R., & Yona, G. (2004). On prediction using variable order Markov models.
Journal of Artificial Intelligence Research 22, 385–421.
Bell, T. C., Cleary, J. G., & Witten, I. H. (1990). Text compression. Englewood Cliffs, NJ: Prentice
Hall.
Bozkurt, B., Ayangil, R., & Holzapfel, A. (2014). Computational analysis of Makam music in
Turkey: Review of state-of-the-art and challenges. Journal of New Music Research 43(1), 3–23.
Brown, S., & Jordania, J. (2013). Universals in the world’s musics. Psychology of Music 41(2), 229–
248.
Bunton, S. (1997). Semantically motivated improvements for PPM variants. The Computer Journal
40(2–3), 76–93.
Burns, E. M., & Campbell, S. L. (1994). Frequency and frequency-ratio resolution by possessors of
absolute and relative pitch: Examples of categorical perception. Journal of the Acoustical Society
of America 96(5), 2704–2719.
Cameron, D. J., Bentley, J., & Grahn, J. A. (2015). Cross-cultural influences on rhythm processing:
Reproduction, discrimination, and beat tapping. Frontiers in Psychology 6, 366. Retrieved from
https://doi.org/10.3389/fpsyg.2015.00366
Castellano, M. A., Bharucha, J. J., & Krumhansl, C. L. (1984). Tonal hierarchies in the music of
north India. Journal of Experimental Psychology: General 113(3), 394–412.
Chiao, J. Y., Iidaka, T., Gordon, H. L., Nogawa, J., Bar, M., Aminoff, E., … Ambady, N. (2008).
Cultural specificity in amygdala response to fear faces. Journal of Cognitive Neuroscience 20(12),
2167–2174.
Cleary, J. G., & Teahan, W. J. (1997). Unbounded length contexts for PPM. The Computer Journal
40(2–3), 67–75.
Conklin, D., & Witten, I. H. (1995). Multiple viewpoint systems for music prediction. Journal of
New Music Research 24(1), 51–73.
Cook, N. (2008). We are all (ethno)musicologists now. In H. Stobart (Ed.), The new (ethno)-
musicologies (pp. 48–70). Lanham, MD: Scarecrow Press.
Cross, I. (2008). Musicality and the human capacity for culture. Musicae Scientiae 12(1 Suppl.),
147–167.
Cross, I. (2014). Music and communication in music psychology. Psychology of Music 42(6), 809–
819.
Curtis, M. E., & Bharucha, J. J. (2009). Memory and musical expectation for tones in cultural
context. Music Perception 26(4), 365–375.
Demorest, S. M., & Morrison, S. J. (2016). Quantifying culture: The cultural distance hypothesis of
melodic expectancy. In J. Y. Chiao, S.-C. Li, R. Seligman, & R. Turner (Eds.), The Oxford
handbook of cultural neuroscience (pp. 183–194). Oxford: Oxford University Press.
Demorest, S. M., Morrison, S. J., Beken, M. N., & Jungbluth, D. (2008). Lost in translation: An
enculturation effect in music memory performance. Music Perception 25(3), 213–223.
Demorest, S. M., Morrison, S. J., Beken, M. N., Stambaugh, L. A., Richards, T. L., & Johnson, C.
(2010). Music comprehension among western and Turkish listeners: fMRI investigation of an
enculturation effect. Social Cognitive and Affective Neuroscience 5, 282–291.
Demorest, S. M., Morrison, S. J., Nguyen, V. Q., & Bodnar, E. N. (2016). The influence of contextual
cues on cultural bias in music memory. Music Perception 33(5), 590–600.
Demorest, S. M., & Osterhout, L. (2012). ERP responses to cross-cultural melodic expectancy
violations. Annals of the New York Academy of Sciences 1252, 152–157.
Demorest, S. M., & Schultz, S. J. (2004). Children’s preference for authentic versus arranged
versions of world music recordings. Journal of Research in Music Education 52(4), 300–313.
Deva, B. C., & Virmani, K. G. (1975). A study in the psychological response to ragas. (Research
Report II of Sangeet Natak Akademi). New Delhi, India: Indian Musicological Society.
Drake, C., & Ben El Heni, J. (2003). Synchronizing with music: Intercultural differences. Annals of
the New York Academy of Sciences 999, 429–437.
Egermann, H., Fernando, N., Chuen, L., & McAdams, S. (2015). Music induces universal emotion-
related psychophysiological responses: Comparing Canadian listeners to Congolese Pygmies.
Frontiers in Psychology 5, 1341. Retrieved from https://doi.org/10.3389/fpsyg.2014.01341
Egermann, H., Pearce, M. T., Wiggins, G. A., & McAdams, S. (2013). Probabilistic models of
expectation violation predict psychophysiological emotional responses to live concert music.
Cognitive, Affective & Behavioral Neuroscience 13(3), 533–553.
Flowers, P. J. (1980). Relationship between two measures of music preference. Contributions to
Music Education 8, 47–54.
Frith, S. (1996). Music and identity. In S. Hall & P. Du Gay (Eds.), Questions of cultural identity (pp.
108–127). London: Sage Publications.
Fritz, T. (2013). The dock-in model of music culture and cross-cultural perception. Music Perception:
An Interdisciplinary Journal 30(5), 511–516.
Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., … Koelsch, S. (2009).
Universal recognition of three basic emotions in music. Current Biology 19(7), 573–576.
Fung, C. V. (1994). Undergraduate nonmusic majors’ world music preference and multicultural
attitudes. Journal of Research in Music Education 42(1), 45–57.
Gingras, B., Pearce, M. T., Goodchild, M., Dean, R. T., Wiggins, G., & McAdams, S. (2015).
Linking melodic expectation to expressive performance timing and perceived musical tension.
Journal of Experimental Psychology: Human Perception & Performance 42(4), 594–609.
Giuliano, R. J., Pfordresher, P. Q., Stanley, E. M., Narayana, S., & Wicha, N. Y. (2011). Native
experience with a tone language enhances pitch discrimination and the timing of neural responses
to pitch change. Frontiers in Psychology 2, 146. Retrieved from
https://doi.org/10.3389/fpsyg.2011.00146
Golby, A. J., Gabrieli, J. D., Chiao, J. Y., & Eberhardt, J. L. (2001). Differential responses in the
fusiform region to same-race and other-race faces. Nature Neuroscience 4, 845–850.
Gregory, A. H., & Varney, N. (1996). Cross-cultural comparisons in the affective response to music.
Psychology of Music 24(1), 47–52.
Hannon, E. E. (2009). Perceiving speech rhythm in music: Listeners classify instrumental songs
according to language of origin. Cognition 111(3), 403–409.
Hannon, E. E., & Trehub, S. E. (2005a). Metrical categories in infancy and adulthood. Psychological
Science 16(1), 48–55.
Hannon, E. E., & Trehub, S. E. (2005b). Tuning in to musical rhythms: Infants learn more readily
than adults. Proceedings of the National Academy of Sciences 102(35), 12639–12643.
Hansen, N. C., & Pearce, M. T. (2014). Predictive uncertainty in auditory sequence processing.
Frontiers in Psychology 5, 1–17. Retrieved from https://doi.org/10.3389/fpsyg.2014.01052
Heingartner, A., & Hall, J. V. (1974). Affective consequences in adults and children of repeated
exposure to auditory stimuli. Journal of Personality and Social Psychology 29(6), 719–723.
Henrich, J., Heine, S. J., & Norenzayan, A. (2010). The weirdest people in the world? Behavioral
and Brain Sciences 33(2–3), 61–83.
Hofstede, G. (1983). National cultures in four dimensions: A research-based theory of cultural
differences among nations. International Studies of Management & Organization 13(1–2), 46–74.
Huron, D. B. (2006). Sweet anticipation: Music and the psychology of expectation. Cambridge, MA:
MIT Press.
Iversen, J. R., Patel, A. D., & Ohgushi, K. (2008). Perception of rhythmic grouping depends on
auditory experience. Journal of the Acoustical Society of America 124, 2263–2271.
Juslin, P. N. (2000). Cue utilization in communication of emotion in music performance: Relating
performance to perception. Journal of Experimental Psychology: Human Perception and
Performance 26(6), 1797–1812.
Juslin, P. N., & Laukka, P. (2003). Communication of emotions in vocal emotion and music
performance: Different channels, same code? Psychological Bulletin 129(5), 770–814.
Karaosmanoğlu, M. K. (2012). A Turkish Makam music symbolic database for music information
retrieval: Symbtr. In Proceedings of the 13th ISMIR Conference, Porto, Portugal, 223–228.
Keil, A., & Keil, C. (1966). A preliminary report: The perception of Indian, Western, and Afro-
American musical moods by American students. Ethnomusicology 10(2), 153–173.
Kessler, E. J., Hansen, C., and Shepard, R. N. (1984). Tonal schemata in the perception of music in
Bali and the West. Music Perception 2(2), 131–65.
Kohavi, R. (1995). A study of cross-validation and bootstrap for accuracy estimation and model
selection. In Proceedings of the Fourteenth International Joint Conference on Artificial
Intelligence (Vol. 2, pp. 1137–1145). San Mateo, CA: Morgan Kaufmann.
Krumhansl, C. L. (1995). Music psychology and music theory: Problems and prospects. Music
Theory Spectrum 17(1), 53–80.
Krumhansl, C. L., Louhivuori, J., Toiviainen, P., Jarvinen, T., & Eerola, T. (1999). Melodic
expectation in Finnish spiritual folk hymns: Convergence of statistical, behavioral, and
computational approaches. Music Perception 17(2), 151–195.
Krumhansl, C. L., & Shepard, R. N. (1979). Quantification of the hierarchy of tonal functions within
a diatonic context. Journal of Experimental Psychology: Human Perception and Performance
5(4), 579–594.
Krumhansl, C. L., Toivanen, P., Eerola, T., Toiviainen, P., Järvinen, T., & Louhivuori, J. (2000).
Cross-cultural music cognition: Cognitive methodology applied to North Sami Yoiks. Cognition
76(1), 13–58.
Kuhl, P. K. (1991). Human adults and human infants show a “perceptual magnet effect” for the
prototypes of speech categories, monkeys do not. Attention, Perception, & Psychophysics 50(2),
93–107.
Laukka, P., Eerola, T., Thingujam, N. S., Yamasaki, T., & Beller, G. (2013). Universal and culture-
specific factors in the recognition and performance of musical affect expressions. Emotion 13(3),
434–449.
LeBlanc, A. (1982). An interactive theory of music preference. Journal of Music Therapy 19(1), 28–
45.
Lynch, M. P., & Eilers, R. E. (1991). Children’s perception of native and nonnative musical scales.
Music Perception 9(1), 121–131.
Lynch, M. P., & Eilers, R. E. (1992). A study of perceptual development for musical tuning.
Perception & Psychophysics 52(6), 599–608.
Lynch, M. P., Eilers, R. E., Oller, D. K., & Urbano, R. C. (1990). Innateness, experience, and music
perception. Psychological Science 1(4), 272–276.
Lynch, M. P., Eilers, R. E., Oller, K. D., Urbano, R. C., & Wilson, P. (1991). Influences of
acculturation and musical sophistication on perception of musical interval patterns. Journal of
Experimental Psychology: Human Perception and Performance 17(4), 967–975.
Lynch, M. P., Short, L. B., and Chua, R. (1995). Contributions of experience to the development of
musical processing in infancy. Developmental Psychobiology 28(7), 377–398.
MacKay, D. J. C. (2003). Information theory, inference, and learning algorithms. Cambridge:
Cambridge University Press.
Merton, R. K. (1972). Insiders and outsiders: A chapter in the sociology of knowledge. American
Journal of Sociology 78(1), 9–47.
Meyer, L. B. (1956). Emotion and meaning in music. Chicago, IL: University of Chicago Press.
Morrison, S. J., & Demorest, S. M. (2009). Cultural constraints on music perception and cognition.
Progress in Brain Research 178, 67–77.
Morrison, S. J., Demorest, S. M., Aylward, E. H., Cramer, S. C., & Maravilla, K. R. (2003). fMRI
investigation of cross-cultural music comprehension. NeuroImage 20(1), 378–384.
Morrison, S. J., Demorest, S. M., Campbell, P. S., Bartolome, S. J., & Roberts, J. C. (2012). Effect of
intensive instruction on elementary students’ memory for culturally unfamiliar music. Journal of
Research in Music Education 60(4), 363–374.
Morrison, S. J., Demorest, S. M., & Stambaugh, L. A. (2008). Enculturation effects in music
cognition: The role of age and music complexity. Journal of Research in Music Education 56(2),
118–129.
Morrison, S. J., & Yeh, C. S. (1999). Preference responses and use of written descriptors among
music and nonmusic majors in the United States, Hong Kong, and the People’s Republic of China.
Journal of Research in Music Education 47(1), 5–17.
Nan, Y., Knösche, T. R., & Friederici, A. D. (2006). The perception of musical phrase structure: A
cross-cultural ERP study. Brain Research 1094(1), 179–191.
Nan, Y., Knösche, T. R., & Friederici, A. D. (2009). Non-musicians’ perception of phrase boundaries
in music: A cross-cultural ERP study. Biological Psychology 82(1), 70–81.
Nan, Y., Knösche, T. R., Zysset, S., & Friederici, A. D. (2008). Cross-cultural music phrase
processing: An fMRI study. Human Brain Mapping 29(3), 312–328.
Narmour, E. (1990). The analysis and cognition of basic melodic structures: The implication-
realization model. Chicago, IL: University of Chicago Press.
Neuhaus, C. (2003). Perceiving musical scale structures: A cross-cultural event-related brain
potentials study. Annals of the New York Academy of Sciences 999, 184–188.
Omigie, D., Pearce, M. T., & Stewart, L. (2012). Tracking of pitch probabilities in congenital amusia.
Neuropsychologia 50(7), 1483–1493.
Omigie, D., Pearce, M. T., Williamson, V. J., & Stewart, L. (2013). Electrophysiological correlates of
melodic processing in congenital amusia. Neuropsychologia 51(9), 1749–1762.
Patel, A. D., & Daniele, J. R. (2003). An empirical comparison of rhythm in language and music.
Cognition 87(1), B35–B45.
Patel, A. D., & Demorest, S. M. (2013). Comparative music cognition: Cross-species and cross-
cultural studies. In D. Deutsch (Ed.), The psychology of music (3rd ed., pp. 647–681). London:
Academic Press.
Pearce, M. T. (2005). The construction and evaluation of statistical models of melodic structure in
music perception and composition (Doctoral dissertation). Department of Computing, City
University, London.
Pearce, M. T., & Müllensiefen, D. (2017). Compression-based modelling of musical similarity
perception. Journal of New Music Research 46(2), 135–155.
Pearce, M. T., Müllensiefen, D., & Wiggins, G. A. (2010). Melodic grouping in music information
retrieval: New methods and applications. In Z. W. Ras & A. Wieczorkowska (Eds.), Advances in
music information retrieval (pp. 364–388). Berlin: Springer.
Pearce, M. T., Ruiz, M. H., Kapasi, S., Wiggins, G. A., & Bhattacharya, J. (2010). Unsupervised
statistical learning underpins computational, behavioural and neural manifestations of musical
expectation. NeuroImage 50(1), 302–313.
Perlman, M., & Krumhansl, C. L. (1996). An experimental study of internal interval standards in
Javanese and Western musicians. Music Perception 14(2), 95–116.
Pfordresher, P. Q., & Brown, S. (2009). Enhanced production and perception of musical pitch in tone
language speakers. Attention, Perception, & Psychophysics 71(6), 1385–1398.
Polak, R., London, J., & Jacoby, N. (2016). Both isochronous and non-isochronous metrical
subdivision afford precise and stable ensemble entrainment: A corpus study of Malian djembe
drumming. Frontiers in Neuroscience 10, 285. Retrieved from
https://doi.org/10.3389/fnins.2016.00285
Raman, R., & Dowling, W. J. (2016). Real-time probing of modulations in South Indian classical
(Carnatic) music by Indian and Western musicians. Music Perception 33(3), 367–393.
Raman, R., & Dowling, W. J. (2017). Perception of modulations in south Indian classical (Carnatic)
music by student and teacher musicians: A cross-cultural study. Music Perception 34(4), 424–437.
Renninger, L. B., Wilson, M. P., & Donchin, E. (2006). The processing of pitch and scale: An ERP
study of musicians trained outside of the western musical system. Empirical Musicology Review
1(4), 185–197.
Sauvé, S., Sayad, A., Dean, R. T., & Pearce, M. T. (2017). Effects of pitch and timing expectancy on
musical emotion. arXiv Preprint, 1708.03687.
Schaffrath, H. (1992). The ESAC databases and MAPPET software. Computing in Musicology 8, 66.
Schaffrath, H. (1994). The ESAC electronic songbooks. Computing in Musicology 9, 78.
Schaffrath, H. (1995). The Essen folksong collection. In D. Huron (Ed.), Database containing 6,255
folksong transcriptions in the Kern format and a 34-page research guide [computer database].
Menlo Park, CA: CCARH.
Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy.
Music Perception 14(3), 295–318.
Shehan, P. K. (1981). Student preferences for ethnic music styles. Contributions to Music Education
9, 21–28.
Shehan, P. K. (1985). Transfer of preference from taught to untaught pieces of non-Western music
genres. Journal of Research in Music Education 33(3), 149–158.
Small, C. (1998). Musicking: The meanings of performing and listening. Middletown, CT: Wesleyan
University Press.
Soley, G., & Hannon, E. E. (2010). Infants prefer the musical meter of their own culture: A cross-
cultural comparison. Developmental Psychology 46(1), 286–292.
Stobart, H., & Cross, I. (2000). The Andean anacrusis? Rhythmic structure and perception in Easter
songs of northern Potosi, Bolivia. British Journal of Ethnomusicology 9(2), 63–92.
Thompson, W. F., & Balkwill, L. L. (2010). Cross-cultural similarities and differences. In P. N. Juslin
& J. A. Sloboda (Eds.), Handbook of music and emotion: Theory, research, applications (pp. 755–
790). New York: Oxford University Press.
Trulsson, Y. H., & Burnard, P. (2016). Insider, outsider or cultures in-between. In P. Burnard, E.
Mackinlay, & K. Powell (Eds.), The Routledge international handbook of intercultural arts
research (pp. 115–125). New York: Routledge.
Wong, P. C. M., Chan, A. H. D., Roy, A., & Margulis, E. H. (2011). The bimusical brain is not two
monomusical brains in one: Evidence from musical affective processing. Journal of Cognitive
Neuroscience 23(12), 4082–4093.
Wong, P. C., Ciocca, V., Chan, A. H., Ha, L. Y., Tan, L. H., & Peretz, I. (2012). Effects of culture on
musical pitch perception. PloS ONE 7(4), e33424.
Wong, P. C. M., Roy, A. K., & Margulis, E. H. (2009). Bimusicalism: The implicit dual enculturation
of cognitive and affective systems. Music Perception 27(2), 81–88.
Yoshida, K. A., Iversen, J. R., Patel, A. D., Mazuka, R., Nito, H., Gervain, J., & Werker, J. F. (2010).
The development of perceptual grouping biases in infancy: A Japanese-English cross-linguistic
study. Cognition 115(2), 356–361.
1
The Essen Folk Song Collection was retrieved from: http://kern.humdrum.org/cgi-bin/browse?
l=/essen. The SymbTR database was retrieved from: https://github.com/MTG/SymbTr.
CHAPT E R 4
W H E N E X T R AVA G A N C E
IMPRESSES: RECASTING
ESTHETICS IN
E VO L U T I O N A RY T E R M S
B JO R N ME R K E R
S A , I: T
E L
Among songbirds with vocal production learning (Janik & Slater, 1997)
there is an association between a high duty cycle for song (i.e., a large
amount of continuous singing per day), large song repertoires, and high
pattern variety among songs (Baylis, 1982). This correlation is presumably
driven by the fact that protracted production of monotonous singing loses
the attention of its audience by the ubiquitous mechanism of habituation
(Hartshorne, 1956; Kroodsma, 1978; Sachs, 1967; Sokolov, 1963).
Persistent singing is energetically costly and takes place at the expense of
useful activities such as foraging. Why, then, prolong the song display
beyond the boredom threshold, thus incurring the additional cost of
acquiring the means to produce elaborate song? Whence the waste and
frivolity of virtuoso performance?
Because the costs of signaling are paid for by the same metabolic engine
that foots the bill for survival, the very fact of surviving with the added
burden of exaggerated signaling is in itself informative. It supplies proof
positive that the signaler is capable of sustaining that additional burden. The
capacity is therefore necessarily an aspect of signaler quality, a
circumstance Amotz Zahavi codified in what he named the “handicap
principle” (Zahavi, 1975), a principle that completes the Darwinian theory
of sexual selection (Darwin, 1871).
The logic of the handicap principle is quite general, and is by no means
limited to promoting potential genetic benefits to offspring. A male in an
agonistic interaction with a conspecific needs to assess the actual fighting
ability of the rival, and not his genetic potential. Similarly a female in a
species with obligate bi-parental care must assess the extent of a
prospective mate’s capacity to invest in the care of offspring. For a variety
of reasons that capacity can and does vary independently of the genes he
contributes to those offspring. In these cases and others, a display of excess
capacity in the form of elaborate signaling can indicate capacity in the
relevant behavioral dimension, provided such signaling actually relates,
directly or indirectly, to abilities and resources employed in the behavioral
dimension of interest to the receiver.
An example involving a direct relationship between signaling and a
desired or feared quality in the signaler is physical fitness itself. Loud
singing for many hours on a daily basis proves that the singer has the
energy reserves, predator vigilance, stamina, and foraging ability to sustain
such behavior without succumbing. For a receiver this means that those
same resources are available for other uses should the animal’s
circumstances or needs require it. Indirect relationships between signaling
and signaler qualities can be quite remote, as illustrated by numerous
laboratory and field studies inspired by the “developmental stress
hypothesis” over the past two decades (Hasselquist et al., 1996; Nowicki et
al., 2002a).
The learned acquisition of elaborate song is a protracted and demanding
sequence of intertwined perceptual, attentional, memory, and motor
challenges that unfolds after hatching in a still developing brain. The
sequence of passive song memorization, followed by stage-wise practice of
vocal skill spanning over weeks and months, interacts with and feeds back
upon the development and neural maturation of an elaborate system of
interconnected forebrain nuclei dedicated to song learning and production
(Iwaniuk & Nelson, 2003; reviewed in Nowicki et al., 2002a). Thus, the
size of the song control nuclei of the mature songbird forebrain correlates
not only with average song repertoire size across species, but with the
repertoire size and song proficiency achieved by individuals within a
species (Gahr, 2000; Garamszegi & Eens, 2004). The latter circumstance is
of central biological significance, because repertoire size and song
proficiency are factors used by females in choosing a mate (Nowicki,
Searcy, & Peters, 2002b).
Each sequential stage of this delicately tuned two-way interaction
between neural development and behavior is susceptible to perturbation by
a variety of external stressors and disturbances (hence the name
“developmental stress hypothesis”). They include, but are not limited to,
immune challenges, disease and parasites, nutritional status dependent on
parental provisioning and later the bird’s own foraging ability,
environmental pollutants, and disruptions at the nest (reviewed by
MacDougall-Shackleton & Spencer, 2012). The management of such
encumbrances consumes developmental resources which otherwise would
have been available for the practice-dependent growth of the song system.
A large repertoire and proficient song performance accordingly can only
be acquired by an individual who as a nestling was cared for by well-
functioning parents, who grew up in a secure nest, was subsequently
unencumbered by disease and parasites, and—in possession of sharp
faculties, memory capacity, foraging ability, and predator vigilance—
engaged in hundreds of hours of successful singing practice. Whatever
impairs the post-hatching growth of a bird’s system of song nuclei, and
whatever keeps the bird from attending to and practicing song is later
evident as deficits in the size and perfection of its mature song repertoire.
This makes a large repertoire of complex song a direct causal reflection of
an individual’s successful passage through a demanding and varied
developmental obstacle course.
The more demanding the performance to be acquired, the more
comprehensive a measure of an individual’s personal history and qualities
lies implicit in the perfected, mature songbout. In effect, then, an
individual’s level of song proficiency sums up, in a single performance, the
entire developmental history of the singer, and as such provides an all-
round certificate of competence, of all-around individual phenotypic
quality. It tells its audience, in a way impossible to counterfeit, that the
singer comes from, as it were, “a good background.” Potential mates and
rivals thus do well to take a singer displaying mastery and virtuosity
seriously.
Though none of this is likely to be accomplished without an adequate
genetic background, it is the finished phenotype and not the genotype that
fights with rivals and helps a bonded female provision her offspring and
defend the nest. Hence the importance of markers for phenotypic quality
when decisions in these regards have to be made on the spot during a brief
breeding season. That is what the expert songbout provides, conferring on
the singer high priority as mate or rival. Provided, that is, that there are ears
competent to assess the quality of the songbout, and to discriminate an
outstanding performance from a middling one. This in turn leads us to the
crux of cultural esthetics, namely the means by which receivers judge
performance quality and the critical dependence of those means on the
cultural song tradition within which the performance takes place.
S A , II: A B
B
S A , III:
S M
T H C
We are now ready to make a swift transition to human arts and esthetics,
and we do so via human music on the plausible assumption that the first
form of human music proper was song. In fact, song may have preceded
speech in our evolutionary history, perhaps in the form of song and dance in
a group setting as a first form of the human arts (Merker, 2005, 2008). A
strong reason to make these assumptions is provided by the fact that
humans, unlike our closest relatives among the apes, indeed unlike any
other primate, are vocal learners, and more specifically, are vocal
production learners (Janik & Slater, 1997; see also Doupé & Kuhl, 1999).
Among non-human animals, this capacity for learning to reproduce by
voice novel sound patterns originally received by ear has most commonly
evolved to serve learned song. Therefore, the default assumption regarding
the function for which our own capacity for vocal learning originated would
be song as well. If so, learned song preceded speech in our evolutionary
ancestry (see Merker, 2012, 2015 for details), and we have landed squarely
in the constellation of factors outlined in previous sections as critical for the
origin and maintenance of complex cultural traditions of ritual lore in
animals with learned song. As a biological trait, this would include the
motivational mechanism of a conformal motive ensuring fidelity to tradition
(Merker, 2005), the role of prior familiarity in appreciation (cf. Madison &
Schiölde, 2017), as well as the ultimate purpose for taking on the burden of
acquisition, namely to impress a competent audience with one’s command
of the shared lore. As we saw in the section “Squandering as Asset, II,”
fidelity to tradition coupled with a shared exposure history furnishes a
standard of judgment short of which the tradition eventually collapses into
idiosyncratic caprice without grounds for comparing one performance with
another.
Trends in Western art over the past century have tended to obscure the
fundamental nature of this connection. In fact, it has been actively
combatted as a fetter on the exercise of untrammeled creativity. In good
agreement with the present perspective, the history of contemporary art
accordingly abounds in examples of idiosyncratic caprice exercised in the
absence of shared criteria for comparing one performance or creation with
another. Proof of this assertion surfaces from time to time in the form of
adventitious revelations that expose the arbitrary nature of the judgments
involved (Cheston, 2015; Jordan-Smith, 1960; Museum of Hoaxes, 2005.
Also: Wikipedia entries for Disumbrationism, Pierre Brassau).
Each step toward such a state of affairs typically meets with opposition
when it first occurs. Presumably this trend in the serious arts of Western
culture (poetry, painting, and music first and foremost, though not limited to
these) would not have proceeded as far as it has were it not for a more
general cultural ambience in the West emphasizing the inherent value of
novelty and the sanctity of artistic freedom, buttressed by the Romanticism
myth of artistic genius. That myth emphasizes the role of rare artistic
endowment over that of diligent mastery of a tradition in the genesis of
great art (Smith, 1924; Waterhouse, 1926). This cultural ambience
eventually ripened into outright celebration of iconoclasm in the course of
the twentieth century. Yet even then, with each advance of idiosyncratic
license, voices were raised in protest, sometimes trenchantly so.
One illustrative example occurred when a faction of musicians in the
modern jazz genre abandoned all reliance on traditional form in what they
styled “free form jazz.” The jazz bassist and band leader Charles Mingus, a
creative musician by no means a stranger to innovation, witnessed a key
event in this development. It was Ornette Coleman’s controversial 1960
performances at the New York City “Five Spot” jazz club. Mingus
commented: “… if the free-form guys could play the same tune twice, then
I would say they were playing something … Most of the time they use their
fingers on the saxophone and they don’t even know what’s going to come
out. They’re experimenting” (Wikipedia entry “Charles Mingus”). On
another occasion he noted, “They don’t even know their Parker” (B.
Merker, personal observation).
Note that Mingus’ comments by no means are directed against creativity
or innovation as such. They remind us, rather, of the necessity, under
circumstances where freedom is in fact possible because the means of
artistic expression are learned, of a shared exposure history to ground
substantive assessment of artistic merit. It is that shared background that
supplies the crucial “common currency” by which alone the informational
emotion of “being impressed” serves as an index of comparative value
across different performances. Without that anchoring in a shared tradition,
the emotional reaction of being impressed becomes as arbitrary and
idiosyncratic as the performances themselves. The bulwark against bluff has
been broken.
The reaction of being impressed by outstanding artistic creations for
which one has been prepared by an appropriate exposure history is
ubiquitous across the arts. It must not be confounded with the kind of
emotional responses that originate in personal associations forged in the
course of significant life events. A tune that figured prominently in a
teenage romantic infatuation may, when encountered years or even decades
later, compel strong feelings on an associative basis without reflecting on
the tune’s artistic merits (Konečni, 2005; Rauhe, 2003; Scherer & Zentner,
2001).
It is otherwise when we encounter a piece of music, perhaps for the first
time, for which our listening history of the genre to which it belongs has
equipped us to appreciate its masterfully patterned content, and we groan
and even weep in admiration (Gabrielsson, 2011; Scherer et al., 2001–2002;
see also Konečni, 2005). We may even feel our skin covered in goose-
bumps, and a shiver or chill traverse our spine (reviewed in Hunter &
Schellenberg, 2010; see further Gabrielsson, 2011; Scherer & Zentner,
2001; Silvia & Nusbaum, 2011; Vickhoff, Åström, & Theorell, 2012). But
what a peculiar way to express our admiration, by sighing, groaning, chills,
and even tears!
Our analysis of animal cultural esthetics allows us to make sense of
these peculiar behavioral and physiological tokens of being impressed. An
ordinary trigger for bodily reactions such as shivers, goose-bumps
(piloerection), or chills is genuine fear (Marks, 1969, pp. 2, 39). They are
the peripheral expressions of the central fear state, as it engages the
autonomic (sympathetic) nervous system on an automatic, involuntary
basis. These low-level autonomic (involuntary) reactions apparently remain
patent even under circumstances where an esthetic stimulus taps the fear
range of the information dimension, but on contextual grounds undergoes a
hedonic reversal, as already covered. Thus the shivers, chills, and goose-
bumps betray the origin of the emotional impact of strong esthetic
experiences in the fear range of the informational dimension depicted in
Figs. 1 and 2, in good agreement with the present interpretive framework
(cf. Benedek & Kaernbach, 2011).
Similarly for the sighing, groaning, and weeping elicited by strong
esthetic experiences. Tears appear to be the most common bodily response
to strong experiences of music (Gabrielsson, 2011; Scherer et al., 2001–
2002; see also Konečni, 2005). A prominent ordinary setting for such
reactions is the experience of personal loss, for which such reactions serve a
largely involuntary expressive function (e.g., Averill, 1979; Frijda, 1988).
As we saw in the previous section, the action tendency contingent on being
esthetically moved should be a readiness, indeed an urge, to yield, submit,
surrender, or capitulate to the source of an impressive performance, be it a
rival who has bested us by a masterly performance, or a suitor who has
penetrated our defenses by the same.
In either case, loss is implicit in the act of surrender. Being bested by a
rival is attended by a direct loss of status and its perquisites. What hovers in
the evolutionary background, as we have seen, is the potential for physical
attack from an agonist whose masterful performance, according to the
developmental stress hypothesis, advertises his all-round superior
phenotypic characteristics. In surrendering to a suitor, one loses freedom of
choice in matters as important as the parentage of one’s offspring, along
with loss of personal independence for the considerable stretch of time that
the partnership will last.
More abstractly conceived, a certain giving up (loss) of self is implicit in
every act of submission. Arthur Schopenhauer emphasized “forgetfulness of
self” in discussing esthetics, and its special relation to experiences of the
sublime which he illustrated by way of landscape painting (Schopenhauer,
1844/1966, vol. I, pp. 200ff., vol. II. pp. 369ff.). For a recent discussion of
this important (and once celebrated) topic in esthetics, see Konečni (2005,
2011). Absorption in the pattern-stream of a musical performance promotes
forgetfulness of anything extraneous, including one’s sense of self. Such
absorption, given the requisite level of background familiarity, will be all
the more complete and compelling in the case of outstanding performances,
not only because their masterly patterning invites it, even compels it, but
because they tax our powers of apprehension. Then self-surrender and
forgetfulness of self may reach a peak, a circumstance that may bear on the
psychology of transcendental and religious experiences that are a prominent
aspect of strong experiences of music (Gabrielsson, 2011).
What drives tears to our eyes even though we are not actually sad or
grieving, then, is the tacit sense of loss coupled to the action tendency of
surrender promoted by an outstanding performance. The phenomenon is not
even strictly confined to arts and esthetics: similar responses can occur on
witnessing an outstanding performance in, say, sports. To prevent
misunderstanding, note that none of this is to be taken to mean that the
connection between such reactions and surrender is directly present to the
minds of those experiencing them. The listening mind is typically absorbed
in the pattern of the performance, far from the sadness of loss or the cold
hand of fear. In keeping with the hedonic reversal invoked here, happiness
and joy are typical of these intense experiences (Gabrielsson, 2011).
These caveats regarding what might be present to the mind of the
listener/beholder do not mean, however, that the evolutionary logic of
capitulating to the originator of a masterful display necessarily is a matter
of our distant ancestry only. There is no dearth of examples of strangers
soliciting casual amorous liaisons with famous creators of art on the basis
of encountering their creations alone (Lipsius, 1919; Miller, 2000, p. 331;
see also Nettle & Clegg, 2006).
In sum, only where artistry is embedded in a tradition that poses a
challenge of acquisition for its practitioners and also has shaped the
sensibilities of the intended audience does the latter’s emotional response of
being impressed provide a measure of artistic merit. It is only when both
conditions are met that a causal connection between intuitive response and
artistic merit is in fact patent, according to the psychological account given
in the section “Squandering as Asset, III.” Such was typically the state of
affairs throughout human cultures until the advent of modernity in the West,
and even there it still holds for its popular culture within any of its given
subcultures. By the same token, where anything can be art, nothing in fact
is.
T P I M :
N “M ” “E ”
The thesis that art generally, and music specifically, exerts its effect via the
informational dimension defined in the section “Squandering as Asset, III”
has obvious consequences for the much discussed issue of “music and
meaning” and its subdomain “music and emotion” (Davies, 1994; Juslin &
Sloboda, 2001; Meyer, 1956; Robinson, 1997). It does so by allowing us to
draw a principled distinction between the undoubted psychological impact
of music on the one hand, and questions of its carrying meaning as well as
of its portraying or inducing emotions on the other.
Strictly speaking only the sentences of language “mean” at all, as most
trenchantly argued by Staal (Staal, 1989). The multilevel combinatorics of
phonemes and morphemes by which language performs its arbitrary (in the
sense of conventional) mapping between the form of utterances and their
meaning (compare “bord,” “Tisch,” and “table” for the selfsame type of
object in Swedish, German, and English, respectively) constitutes a bona
fide code for representing and conveying meaning. This code is so detailed
and comprehensive that virtually every difference in the strings of
phonemes that make up sentences makes a difference in the information
conveyed by those sentences. This lexically semanticized syntactic code
turns sequential patterns of vocally produced sounds into statements about
things that bear not the slightest resemblance to those sound sequences
themselves (see the “table” example above). Thus we think and
communicate about objects, events, matters of fact, states of the world,
ideas, intentions, beliefs and desires, without limit, using the same few
dozen phonemes to do so. This is what it means to “mean,” namely that
something encodes something other than itself, which is its meaning.
Much of what compels our interest, carries significance, and recruits our
psychological engagement—in short, has psychological impact—does so
without the detour of meaning in this sense. Our non-linguistic perception
and cognition quite generally operates on patterns of sensory input by
discriminating, segmenting, grouping, classifying and generalizing within
and across them, and not by using them as vehicles for encoding matters
other than themselves. Something need not, in other words, mean in order
to be meaningful: witness the experience of a magnificent sunset as but one
of an infinitude of cases in point.
Viewed in this light, the patterns of music define themselves as
perceptual objects that engage the informational dimension of our
perceptual/cognitive capacities in the manner of auditory analogs of
visually presented arabesques or the shifting patterns of a turning
kaleidoscope, to use Hanslick’s felicitous metaphors (Hanslick, 1854). As
such they need to “sound good” (and “better,” and “best”), not to refer to
circumstances other than themselves. In so doing they exploit the limitless
pattern-generativity music conquers for itself by a combinatorics of
“particulate” elements drawn from the discretized continua of pitch and
duration (for which see Abler, 1989; Merker, 2002; Merker, Morley, &
Zuidema, 2015). This limitless generativity fits ill with a conception of
music as a device for portraying or evoking the limited set of subjective
states that make up our emotions. Not only is the empirical evidence
supporting that conception weak (Konečni, 2003, 2008; Konečni et al.,
2008; Scherer, 2003), but weighty arguments have been leveled against it,
arguments for which Hanslick’s 1854 essay is still the unsurpassed locus
classicus (Hanslick, 1854; see also Davies, 1994; Zangwill, 2004).
It is the common experience of having been moved by music, even to
the point of tears or chills in some instances, that has lent credence to the
notion that music, somehow, is “about” emotions, or exerts its effects by
inducing them. This “being moved” or being impressed by music is indeed
an emotional response. But as we saw in the section “Squandering as Asset,
III,” that response moves up and down the intensity dimension of a single
one of the basic emotions, rather than across them. The fact that music has
an emotional impact in this sense must accordingly be sharply distinguished
from the claim that music is about emotions in the plural, or aims at
evoking emotions, again in the plural, which in either case it would have to
do to fit the metaphor of being a “language of emotions” (Spencer, 1911).
To the extent that the patterns of music refer to circumstances other than
themselves (e.g., storms, battles, or dancing peasants in programmatic
music) they tend to do so by dynamically or otherwise mimicking,
resembling, or caricaturing the things to be evoked (Hanslick, 1854). That
is not how language carries meaning except in the special cases of
onomatopoeia and some of the uses of prosody, both of which lie outside of
the central coding device that gives the sentences of language their unique
and unbounded capacity to mean. So even when music is intended to mean
—which is far from always the case—it does not mean, it mimics.
In song and music without lyrics it is the vocal or instrumental patterns
themselves that are the information conveyed. As detailed in previous
sections, our emotional response to these patterns concerns the inducing
patterns themselves as they interact with the background of our prior
musical familiarity. When that background relates to the genre to which the
patterns belong, the specificity, scope, and intricacy of the interaction is
commensurately enhanced. It is here that the infinite pattern generativity of
music comes into its own. It furnishes the makings of the untold structural
devices (variation and repetition, various symmetries and asymmetries, etc.)
needed to create temporal trajectories capable of sustaining our interest in
the face of the ubiquitous habituability of our cognitive equipment, a
habituability that converts every novelty to “old hat” in the course of a few
encounters. It is not to our emotions that this content is addressed in the first
place, but to our imagination, as Eduard Hanslick, following Arthur
Schopenhauer, insisted (Hanslick, 1854; Schopenhauer, 1844/1966, vol. II,
pp. 447ff.).1
Engaging the contours of our recognition memory the ever-varied
patterns of music trigger a variety of familiarity-based expectancies which
their temporally unfolding melodic, rhythmic, and harmonic patterns
confirm, violate, or complement, generating tensions, their resolution, and
new expectancies in ever-shifting peregrinations across the sensibility
landscape sculpted by the listener’s history of prior exposure (Meyer, 1956;
Narmour, 1977; Schopenhauer, 1844/1966, vol. II, p. 455). For a given
musical listening experience, it is the cumulative effect—presumably in
“leaky integrator” fashion—of the particular sequence of twists and turns
along the temporal trajectory of this interaction that determines how far up
the information dimension toward “awe” a given experience of music takes
us and thus the extent to which it moves or impresses us. In this sense
“being moved” or “impressed” is a specifically esthetic emotion. It may
even be deemed the esthetic emotion (Konečni, 2005, 2011, 2015),
generated by hedonic reversal in the danger zone of the information
dimension.
In both cultural history and the listening history of individuals, the
infinite space of musical combinatorics differentiates into occupied
subregions according to genre (cf. Merker, 2002, pp. 11–12). Individual
specimens that make up any given region of this space will exhibit greater
or lesser elegance with greater or lesser degrees of well-formedness (see,
e.g., Lerdahl & Jackendoff, 1983) and greater or lesser efficacy in stirring a
given listener’s imagination on encounter. And just as in other perceptual
and cognitive systems, pattern invariants are bound to be extracted across
the sampled space in accordance with a variety of shared structural
characteristics, clustering musical impressions under high-level descriptors.
Thus the categories “nostalgia” (sentimental, dreamy, melancholic),
“power” (energetic, triumphant, heroic), and “tension” (agitated, nervous,
impatient, irritated) extracted by Zentner and colleagues from responses to a
diverse sample of European classical music (Zentner, Grandjean, & Scherer,
2008). The authors interpret their results in terms of a model of music-
specific emotions. They might also be construed in terms of high-level
intuitive (statistical) pattern-classification, ranging across the vast and
multiform world of musical patterns that have accumulated in a given
musical culture and whose uneven sampling has shaped the musical
familiarity and sensibilities of any given listener.
The perspective on the psychological impact of music presented here by
no means relegates music to an abstract domain of formalist
connoisseurship. Some of its patterns—say of rhythmic music meant to
support dancing—access a presumably species-specific predisposition for
bodily entrainment to isochrony-based auditory patterns, and help optimize
such entrainment (Merker, 2014; Merker, Madison, & Eckerdal, 2009). The
central role of music in youth and popular culture suffices to dispel any
overly formalist notion of its nature, and fits well with the evolutionary
perspective on esthetics presented here.
R
Abler, W. L. (1989). On the particulate principle of self-diversifying systems. Journal of Social and
Biological Structures 12(1), 1–13.
Apter, M. J. (1982). The experience of motivation: The theory of psychological reversals. New York:
Academic Press.
Averill, J. R. (1979). The functions of grief. In C. Izard (Ed.), Emotions in personality and
psychopathology (pp. 339–368). New York: Plenum Press.
Baylis, J. R. (1982). Avian vocal mimicry: Its function and evolution. In D. E. Kroodsma & E. H.
Miller (Eds.), Acoustic communication in birds (pp. 51–83). New York: Academic Press.
Benedek, M., & Kaernbach, C. (2011). Physiological correlates and emotional specificity of human
piloerection. Biological Psychology 86(3), 320–329.
Berlyne, D. E. (1960). Conflict, arousal, and curiosity. New York: McGraw-Hill.
Berlyne, D. E. (1971). Aesthetics and psychobiology. New York: Appleton Century.
Bindra, D. (1959). Stimulus change, reactions to novelty, and response decrement. Psychological
Review 66(2), 96–103.
Bloom, P. (2010). How pleasure works. New York: W. W. Norton.
Bradley, M. M., & Lang, P. J. (2000). Measuring emotion: Behavior, feeling and physiology. In R. D.
Lane, L. Nadel, & G. Ahern (Eds.), Cognitive neuroscience of emotion (pp. 242–276). New York:
Oxford University Press.
Cheston, P. (2015). Artist in legal row claims “former workshop sold her paint-spattered carpet as
genuine works.” Retrieved from http://www.standard.co.uk/news/london/artist-in-legal-row-
claims-former-workshop-sold-her-paint-spattered-carpet-as-genuine-works-a2947666.html
Clark, A. (2013). Whatever next? Predictive brains, situated agents, and the future of cognitive
science. Behavioral and Brain Sciences 36(3), 181–204.
Darwin, C. (1871). The descent of man and selection in relation to sex. New York: D. Appleton &
Company.
Davies, S. (1994). Musical meaning and expression. Ithaca, NY: Cornell University Press.
Doupé, A. J., & Kuhl, P. K. (1999). Birdsong and human speech: Common themes and mechanisms.
Annual Review of Neuroscience 22, 567–631.
Dowsett-Lemaire, F. (1979). The imitative range of the song of the Marsh Warbler, Acrocephalus
palustris, with special reference to imitations of African birds. Ibis 121(4), 453–468.
Ekman, P. (1999). Basic emotions. In T. Dalgleish & M. Power (Eds.), Handbook of cognition and
emotion (pp. 45–60). Chichester: John Wiley and Sons.
Fontaine, J. J. R., & Scherer, K. R. (2013). Emotion is for doing: The action tendency component. In
J. J. R. Fontaine, K. R. Scherer, & C. Soriano (Eds.), Components of emotional meaning: A
sourcebook (Chapter 11). Oxford Scholarship Online. Oxford: Oxford University Press.
doi:10.1093/acprof:oso/9780199592746.001.0001
Frijda, N. H. (1987). Emotion, cognitive structure, and action tendency. Cognition and Emotion 1(2),
115–143.
Frijda, N. H. (1988). Laws of emotion. American Psychologist 43(5), 349–358.
Friston, K. (2002). Functional integration and inference in the brain. Progress in Neurobiology 68(2),
113–143.
Gabrielsson, A. (2011). Strong experiences with music. Oxford: Oxford University Press.
Gahr, M. (2000). Neural song control system of hummingbirds: Comparison to swifts, vocal learning
(songbirds) and nonlearning (suboscines) passerines, and vocal learning (budgerigars) and
nonlearning (dove, owl, gull, quail, chicken) nonpasserines. Journal of Comparative Neurology
426(2), 182–196.
Garamszegi, L. Z., & Eens, M. (2004). Brain space for a learned task: Strong intraspecific evidence
for neural correlates of singing behavior in songbirds. Brain Research Reviews 44(2–3), 187–193.
Grassberger, P. (1986). Toward a quantitative theory of self-generated complexity. International
Journal of Theoretical Physics 25(9), 907–938.
Hanslick, E. (1854). Vom musikalisch Schönen. Beiträge zur Revision der Ästhetik der Tonkunst.
Leipzig: Weigel.
Hartshorne, C. (1956). The monotony threshold in singing birds. Auk 73, 176–192.
Hasselquist, D., Bensch, S., & von Schantz, T. (1996). Correlation between song repertoire, extra-
pair paternity and offspring survival in the great reed warbler. Nature 381(6579), 229–232.
Heinrichs, M., von Dawans, B., & Domes, G. (2009). Oxytocin, vasopressin, and human social
behavior. Frontiers in Neuroendocrinology 30(4), 548–557.
Hinton, G. E., & Zemel, R. S. (1994). Autoencoders, minimum description length, and Helmholtz
free energy. In J. D. Cowan, G. Tesauro, & J. Alspector (Eds.), Advances in neural information
processing systems 6 (pp. 3–10). San Mateo, CA: Morgan Kaufmann.
Hunter, P. G., & Schellenberg, E. G. (2010). Music and emotion. In M. R. Jones, R. R. Fay, & A. N.
Popper (Eds.), Music perception (pp. 129–164). New York: Springer.
Huron, D. (2001). Is music an evolutionary adaptation? Annals of the New York Academy of Sciences
930, 43–61.
Iwaniuk, A. N., & Nelson, J. E. (2003). Developmental differences are correlated with relative brain
size in birds: A comparative analysis. Canadian Journal of Zoology 81(12), 1913–1928.
Izard, C. E. (2007). Basic emotions, natural kinds, emotion schemas, and a new paradigm.
Perspectives on Psychological Science 2(3), 260–280.
Janik, V. M., & Slater, P. J. B. (1997). Vocal learning in mammals. Advances in the Study of Behavior
26, 59–99.
Jarvis, E. D. (2007). Neural systems for vocal learning in birds and humans: A synopsis. Journal of
Ornithology 148, 35–44.
Jordan-Smith, P. (1960). The road I came; some recollections and reflections concerning changes in
American life and manners since 1890. Caldwell, Idaho: Caxton Printers.
Juslin, P. N., & Sloboda, J. A. (Eds.). (2001). Music and emotion: Theory and research. Oxford:
Oxford University Press.
Kamin, L. J. (1969). Predictability, surprise, attention, and conditioning. In R. Church & B. Campbell
(Eds.), Punishment and aversive behavior (pp. 279–296). New York: Appleton-Century-Crofts.
Kelly, A. M., & Goodson, J. L. (2014). Social functions of individual vasopressin–oxytocin cell
groups in vertebrates: What do we really know? Frontiers in Neuroendocrinology 35(4), 512–529.
Konečni, V. J. (2003). Review of P. N. Juslin and J. A. Sloboda (Eds.), Music and emotion: Theory
and research. Music Perception 20, 332–341.
Konečni, V. J. (2005). The aesthetic trinity: Awe, being moved, thrills. Bulletin of Psychology and the
Arts 5(2): 27–44.
Konečni, V. J. (2008). Does music induce emotion? A theoretical and methodological analysis.
Psychology of Aesthetics, Creativity, and the Arts 2(2), 115–129.
Konečni, V. J. (2011). Aesthetic trinity theory and the sublime. Philosophy Today 55, 64–73.
Konečni, V. J. (2015). Being moved as one of the major aesthetic emotional states: A commentary on
“Being moved: linguistic representation and conceptual structure.” Frontiers in Psychology 6, 343.
Konečni, V. J., Brown, A., & Wanic, R. (2008). Comparative effects of music and recalled life-events
on emotional state. Psychology of Music 36(3), 289–308.
Konishi, M. (2004). The role of auditory feedback in birdsong. In H. P. Ziegler & P. Marler (Eds.),
The behavioral neurobiology of birdsong. Annals of the New York Academy of Sciences 1016, 463–
475.
Kroodsma, D. E. (1978). Continuity and versatility in birdsong: Support for the monotony threshold
hypothesis. Nature 274(5672), 681–683.
Kroodsma, D. E., & Parker, L. D. (1977). Vocal virtuosity in the brown thrasher. Auk 94, 783–785.
Kuehnast, M., Wagner, V., Wassiliwizky, E., Jacobsen, T., & Mennighaus, W. (2014). Being moved:
Linguistic representation and conceptual structure. Frontiers in Psychology: Emotion Science 5,
1242.
Lerdahl, F., & Jackendoff, R. (1983). A generative theory of tonal music. Cambridge, MA: MIT
Press.
Lim, M. M., & Young, L. J. (2006). Neuropeptidergic regulation of affiliative behavior and social
bonding in animals. Hormones and Behavior 50(4), 506–517.
Lipsius, I. M. (1919). Liszt und die Frauen. Leipzig: Breitkopf & Härtel.
MacDougall-Shackleton, S. A., & Spencer, K. A. (2012). Developmental stress and birdsong:
Current evidence and future directions. Journal of Ornithology 153(Suppl. 1), S105–S117.
Madison, G., & Schiölde, G. (2017). Repeated listening increases the liking for music regardless of
its complexity: Implications for the appreciation and aesthetics of music. Frontiers in
Neuroscience 11, 147.
Marks, I. M. (1969). Fears and phobias. New York: Academic Press.
Merker, B. (2002). Music: The missing Humboldt system. Musicae Scientiae 6(1), 3–21.
Merker, B. (2005). The conformal motive in birdsong, music and language: An introduction. In G.
Avanzini, L. Lopez, S. Koelsch, & M. Majno (Eds.), The neurosciences and music II: From
perception to performance. Annals of the New York Academy of Sciences 1060, 17–28.
Merker, B. (2007a). Consciousness without a cerebral cortex: A challenge for neuroscience and
medicine. Behavioral and Brain Sciences 30(1), 63–134.
Merker, B. (2007b). Music at the limits of the mind. In G. Kugiumutzakis (Ed.),Sympantiki Armonia,
Musike kai Epistimi. Ston Miki Theodoraki [Universal harmony, music and science. In honour of
Mikis Theodorakis]. Heraklion: Crete University Press.
Merker, B. (2008). Ritual foundations of human uniqueness. In S. Malloch & C. Trevarthen (Eds.),
Communicative musicality (pp. 45–59). Oxford: Oxford University Press.
Merker, B. (2012). The vocal learning constellation: Imitation, ritual culture, encephalization. In N.
Bannan & S. Mithen (Eds.), Music, language and human evolution (pp. 215–60). Oxford: Oxford
University Press.
Merker, B. (2013a). The efference cascade, consciousness, and its self: Naturalizing the first person
pivot of action control. Frontiers in Psychology 4, article 501, 1–20.
Merker, B. (2013b). Cortical gamma oscillations: The functional key is activation, not cognition.
Neuroscience & Biobehavioral Reviews 37(3): 401–417.
Merker, B. (2014). Groove or swing as distributed rhythmic consonance: Introducing the groove
matrix. Frontiers in Human Neuroscience 8, article 454, 1–4.
Merker, B. (2015). Seven theses on the biology of music and language. Signata 6, 195–213.
Merker, B., Madison, G., & Eckerdal, P. (2009). On the role and origin of isochrony in human
rhythmic entrainment. Cortex 45(1): 4–17.
Merker, B., Morley, I., & Zuidema, W. (2015). Five fundamental constraints on theories of the
origins of music. Philosophical Transactions of the Royal Society of London: Biology 370(1664):
20140095. doi:10.1098/rstb.2014.0095
Meyer, L. B. (1956). Emotion and meaning in music. Chicago, IL: University of Chicago Press.
Miller, G. F. (2000). The mating mind: How sexual choice shaped the evolution of human nature.
New York: Doubleday.
Museum of Hoaxes (2005). Monkey art fools expert. Retrieved from:
http://hoaxes.org/weblog/comments/monkey_art_fools_expert
Narmour, E. (1977). Beyond Schenkerism: The need for alternatives in music analysis. Chicago, IL:
University of Chicago Press.
Nettle, D., & Clegg, H. (2006). Schizotypy, creativity and mating success in humans. Proceedings of
the Royal Society of London B: Biological Sciences 273, 611–615. doi:10.1098/rspb.2005.3349
Nowicki, S., Searcy, W. A., & Peters, S. (2002a). Brain development, song learning and mate choice
in birds: A review and experimental test of the “nutritional stress hypothesis.” Journal of
Comparative Physiology A: Sensory, Neural, and Behavioral Physiology 188: 1003–1004.
Nowicki, S., Searcy, W. A., & Peters, S. (2002b). Quality of song learning affects female response to
male bird song. Proceedings of the Royal Society of London B: Biological Sciences 269, 1949–
1954.
Okanoya, K. (2004). Song syntax in Bengalese finches: Proximate and ultimate analyses. Advances
in the Study of Behavior 34, 297–345.
Patel, A. D. (2008). Music, language, and the brain. Oxford: Oxford University Press.
Pearce, E., Launay, J., & Dunbar, R. I. M. (2015). The ice-breaker effect: Singing mediates fast social
bonding. Royal Society Open Science 2, 150221. Retrieved from
http://dx.doi.org/10.1098/rsos.150221
Pinker, S. (1997). How the mind works. New York: Penguin Putnam.
Rauhe, H. (2003). Musik heilt und befreit. In H. G. Bastian & G. Kreutz (Eds.), Musik und
Humanität. Interdiziplinäre Grundlagen für (musikalische) Erzhiehung und Bildung (pp. 182–
191). Mainz: Schott.
Rescorla, R. A., & Wagner, A. R. (1972). A theory of Pavlovian conditioning: Variations in the
effectiveness of reinforcement and non-reinforcement. In A. H. Black & W. F. Prokasy (Eds.),
Classical conditioning II: Current research and theory (pp. 64–99). New York: Appleton-Century-
Crofts.
Riebel, K. (2003). The “mute” sex revisited: Vocal production and perception learning in female
songbirds. Advances in the Study of Behavior 33, 49–86.
Robinson, J. (Ed.). (1997). Music and meaning. Ithaca, NY: Cornell University Press.
Rohrmeier, M. A., & Koelsch, S. (2012). Predictive information processing in music cognition. A
critical review. International Journal of Psychophysiology, 83, 164–175.
Sachs, E. (1967). Dissociation of learning in rats and its similarities to dissociative states in man. In
J. Zubin & H. Hunt (Eds.), Comparative psychopathology: Animal and human (pp. 249–304). New
York: Grune and Stratton.
Scherer, K. R. (2003). Why music does not produce basic emotions. In R. Breslin (Ed.), Proceedings
of the Stockholm Music Acoustic Conference, 2 vols., Vol. 1 (pp. 25–28). Retrieved from
http://www.speech.kth.se/smac03
Scherer, K. R., & Zentner, M. R. (2001). Emotional effects of music: Production rules. In P. N. Juslin
& J. A. Sloboda (Eds.), Music and emotion: Theory and research (pp. 361–392). Oxford: Oxford
University Press.
Scherer, K. R., Zentner, M. R., & Schacht, A. (2001–2002). Emotional states generated by music: An
exploratory study of music experts. Musicae Scientiae, Special Issue: Current trends in the study of
music and emotion, 149–171.
Schopenhauer, A. (1844/1966). The world as will and representation (2nd ed.; orig. ed. 1819). Trans.
E. F. J. Payne, 2 vols. New York: Dover.
Shields, S. A., MacDowell, K. A., Fairchild, S. B., & Campbell, M. L. (1987). Is mediation of
sweating cholinergic, adrenergic, or both? A comment on the literature. Psychophysiology 24(3),
312–319.
Silvia, P. J., & Nusbaum, E. C. (2011). On personality and piloerection: Individual differences in
aesthetic chills and other unusual aesthetic experiences. Psychology of Aesthetics, Creativity, and
the Arts 5(3), 208–214.
Smith, L. P. (1924). Four words: Romantic, originality, creative, genius. Oxford: Clarendon Press.
Sokolov, E. N. (1963). Higher nervous functions: The orienting reflex. Annual Review of Physiology
25, 545–580.
Spencer, H. (1911). On the origin and function of music. In Essays on education and kindred subjects
(pp. 312–330). London: J. M. Dent & Sons.
Staal, F. (1989). Rules without meaning. New York: Peter Lang.
Strohminger, N. S. (2013). The hedonics of disgust (Doctoral dissertation). University of Michigan.
Retrieved from https://deepblue.lib.umich.edu/handle/2027.42/97960
Taylor, H. (2009). Towards a species songbook: Illuminating the vocalisations of the Australian pied
butcherbird (Cracticus nigrogularis) (Doctoral dissertation). University of Western Sydney.
Tribus, M. (1961). Thermodynamics and thermostatics: An introduction to energy, information and
states of matter, with engineering applications. New York: Van Nostrand.
Vickhoff, B., Åström, R., & Theorell, T. (2012). Musical piloerection. Music and Medicine 4, 82–89.
Waterhouse, F. A. (1926). Romantic “originality.” The Sewanee Review 34, 40–49.
Zahavi, A. (1975). Mate selection: A selection for a handicap. Journal of Theoretical Biology 53(1),
205–214.
Zangwill, N. (2004). Against emotion: Hanslick was right about music. British Journal of Aesthetics
44(1), 29–43.
Zentner, M., Grandjean, D., & Scherer, K. R. (2008). Emotions evoked by the sound of music:
Characterization, classification, and measurement. Emotion 8(4), 494–521.
1
For Schopenhauer’s influence on Hanslick, see Merker (2007b), to which can be added the fact
that the “smoking gun” of that influence in Hanslick’s final paragraph was eliminated from all but the
first edition of Hanslick’s famous essay.
SECTION III
MU S IC P R OC E S S IN G IN
T HE HU MA N B R A IN
CHAPT E R 5
CEREBRAL
O R G A N I Z AT I O N O F M U S I C
PROCESSING
T H E N I L L E B R A U N JA N Z E N A N D MI C H A E L H . T H A U T
N B M P
H B
Auditory-Frontal Networks
The transformation of the auditory information into a musically meaningful
tonal context involves several areas of the frontal cortex. Studies of music
syntax, utilizing primarily expectancy violation paradigms, have
demonstrated that regions of the inferior frontal gyrus respond to harmonic
expectancy violations (Bianco et al., 2016; Janata, Birk, et al., 2002;
Koelsch et al., 2002; Koelsch, Fritz, Schulze, Alsop, & Schlaug, 2005;
Maess, Koelsch, Gunter, & Friederici, 2001; Seger et al., 2013; Tillmann,
Janata, & Bharucha, 2003). Reports have repeatedly indicated that the
cortical network comprising the inferior frontolateral cortex (corresponding
to BA44), inferior frontal gyrus, the anterior portion of the superior
temporal gyrus, and the ventral premotor cortex, is involved in the
processing of musical structure (for review, see Koelsch, 2006, 2011). This
network appears to be specialized in establishing syntactic relationships by
evaluating the harmonic relationship between incoming tonal information
and a preceding harmonic sequence, thus detecting musical-structural
irregularities and organizing fast short-term predictions of upcoming
musical events (Koelsch, 2006). Recent imaging research has also
suggested that rhythmic and melodic deviations in musical sequences may
recruit different cortical areas—pitch deviations engage a neural network
comprising auditory cortices, inferior frontal and prefrontal areas, whereas
rhythmic deviations of a musical sequence recruit neural networks
involving the posterior parts of the auditory cortices and parietal areas
(Lappe, Lappe, & Pantev, 2016; Lappe, Steinsträter, & Pantev, 2013).
These findings are in accordance with the dual-pathway model of
auditory processing, which hypothesizes that two auditory processing
pathways originate from the primary auditory cortex, each contributing to
processing different higher-order aspects of auditory stimuli (Belin &
Zatorre, 2000; Bizley & Cohen, 2013; Hickok & Poeppel, 2007;
Rauschecker & Scott, 2009). The anterior-ventral auditory pathway—which
projects from anterior superior temporal gyrus to anterior inferior frontal
gyrus and prefrontal areas—is predominantly involved in perceiving
auditory objects and processing auditory spectral features. For instance, it
has been shown that the inferior frontal gyrus and related areas of the
ventrolateral prefrontal cortex are activated during phonological and
semantic processing, non-verbal auditory sound detection (Kiehl, Laurens,
Duty, Forster, & Liddle, 2001), discrimination and auditory feature
detection (Gaab, Gaser, Zaehle, Jancke, & Schlaug, 2003; Zatorre,
Bouffard, & Belin, 2004), and auditory working memory (Kaiser, Ripper,
Birbaumer, & Lutzenberger, 2003), which reinforces the assumption that
these areas play a more fundamental role in auditory processing. On the
other hand, the posterior-dorsal stream—which connects posterior superior
temporal gyrus with posterior inferior frontal gyrus, posterior parietal
cortex, and premotor cortex—has been implicated in extracting spectral
motion and temporal components of an auditory stimulus, thus processing
how frequencies change over time (review: Plakke & Romanski, 2014;
Zatorre & Zarate, 2012). Recent evidence indicates that the dorsal pathway
of auditory processing also plays an important role in calculating and
comparing pitch or temporal manipulations within a context and using this
auditory information to select and prepare appropriate motor responses
(Belin & Zatorre, 2000; Chen, Rae, & Watkins, 2012; Foster, Halpern, &
Zatorre, 2013; Hickok & Poeppel, 2007; Loui, 2015; Saur et al., 2008;
Warren, Wise, & Warren, 2005).
Frontal cortex activity has also been associated with cognitive demands
or the stimulus properties within a task. Tasks that require maintenance and
rehearsal of musical information activate the working memory functional
network, comprising the ventrolateral premotor cortex (encroaching Broca’s
area), dorsal premotor cortex, the planum temporale, inferior parietal lobe,
the anterior insula, and subcortical structures (Koelsch et al., 2009; Royal et
al., 2016; Schulze, Zysset, Mueller, Friederici, & Koelsch, 2011). The
medial prefrontal cortex (primarily the medial orbitofrontal region) appears
to be particularly engaged in tasks requiring self-referential judgments
(Alluri et al., 2013; Zysset, Huber, Ferstl, & von Cramon, 2002), musical
semantic memory (Groussard et al., 2010; Platel, Baron, Desgranges,
Bernard, & Eustache, 2003), and music-evoked autobiographical memory
(Janata, 2009; Von Der Heide, Skipper, Klobusicky, & Olson, 2013). Areas
in the frontal lobes, such as parietal and ventrolateral prefrontal regions, are
differentially activated depending on the relative attentional demands of the
tasks (Alho, Rinne, Herron, & Woods, 2014; Janata, Tillmann, & Bharucha,
2002; Maidhof & Koelsch, 2011; Satoh, Takeda, Nagata, Hatazawa, &
Kuzuhara, 2001).
Involuntary musical imagery—that is, the spontaneous experience of
having music looping in one’s head—is associated with cortical thickness in
regions of the right frontal and temporal cortices as well as the anterior
cingulate and left angular gyrus (Farrugia, Jakubowski, Cusack, & Stewart,
2015). On the other hand, voluntary musical imagery—the generation of
mental representation of music or musical attributes in the absence of real
sound input—engages secondary auditory cortices, the parietal cortex,
inferior frontal regions, the supplementary motor area (SMA) and pre-SMA
(Brown & Martinez, 2007; Halpern, Zatorre, Bouffard, & Johnson, 2004;
Harris & De Jong, 2014; Peretz et al., 2009; Zatorre, Halpern, Perry, Meyer,
& Evans, 1996). Neural activity in motor areas during perception or mental
imagery of sounds have been repeatedly reported when musicians listen to a
well-rehearsed musical sequence (Bangert et al., 2006; D’Ausilio,
Altenmüller, Olivetti Belardinelli, & Lotze, 2006; Harris & De Jong, 2014;
Haueisen & Knösche, 2001) or when pianists watch silent video recordings
of hands playing a silent keyboard (Baumann et al., 2007; Bianco et al.,
2016; Hasegawa et al., 2004). Activation of the fronto-parietal motor-
related network (comprising Broca’s area, the premotor region, intraparietal
sulcus, and inferior parietal region) was also found when non-musicians
listened to a piano piece they learned to play (Lahav, Saltzman, & Schlaug,
2007). These studies collectively show that the mere perception or mental
imagery of sounds (which would normally be associated with a specific
action) can automatically trigger representations of the movement necessary
to produce these sounds, providing strong evidence that perception and
action are intrinsically coupled in the human brain and in cognition (for
review, see Keller, 2012; Maes, Leman, Palmer, & Wanderley, 2014;
Novembre & Keller, 2014).
Auditory-Motor Networks
Projections from motor cortex to the auditory cortex are an architectural
feature common to many animal species (Schneider, Nelson, & Mooney,
2014). Animal research has indeed proven to be an important model to
investigate the synaptic and circuit mechanisms by which the motor cortex
interacts with auditory cortical activity (e.g., Merchant, Perez, Zarco, &
Gamez, 2013; Nelson et al., 2013; Roberts et al., 2017; Schneider &
Mooney, 2015). For instance, a recent study in mice found that axons from
the secondary motor cortex make synapses onto both excitatory and
inhibitory neurons in deep and superficial layers of the auditory cortex and
that a subset of these neurons extends axons to various subcortical areas
important for auditory processing (Nelson et al., 2013). The analysis of
local field potentials of behaving macaques has also provided valuable
insight regarding the neural underpinnings for beat synchronization,
showing, for example, that beta-band oscillations may enable
communication between distributed circuits involving the striato-thalamo-
cortical network during rhythm perception and production (for a review, see
Merchant & Bartolo, 2018; Merchant, Grahn, Trainor, Rohrmeier, & Fitch,
2015; see also Chapter 8).
Recent research has also identified fiber projections transmitting
auditory signals into motor regions in the human brain (Fernández-Miranda
et al., 2015). Fernández-Miranda and colleagues demonstrated that the left
and right arcuate fascicle, a white matter fiber tract that links lateral
temporal cortex with frontal areas, is segmented into subtracts with distinct
fiber terminations (Fig. 2). One set of fibers terminates at the ventral
precentral and caudal middle frontal gyri (BA4, BA6), providing direct
projections from auditory cortex to motor areas (primary motor cortex,
premotor cortex).
FIGURE 2. Subtracts of the left arcuate fascicle with terminations on primary motor cortex and
premotor cortex, corresponding to Brodmann areas BA6 and BA4 (ventral precentral and caudal
middle frontal gyri).
Reprinted by permission from Brain Structure and Function 220 (3), Asymmetry,
connectivity, and segmentation of the arcuate fascicle in the human brain, Juan C. Fernández-
Miranda, Yibao Wang, Sudhir Pathak, Lucia Stefaneau, Timothy Verstynen, and Fang-Cheng
Yeh, pp. 1665–1680, https://doi.org/10.1007/s00429-014-0751-7 © Springer-Verlag Berlin
Heidelberg, 2014.
Cortico-Cerebellar Network
The cerebellum receives segregated projections from prefrontal, frontal,
parietal, and superior temporal regions via the pontine nuclei in the
brainstem (Fig. 3). Output projections are then sent from the cerebellar
cortex to specialized deep cerebellar nuclei, which in turn project back, via
the thalamus, to the region of the cerebral cortex from which the initial
projection originated (Koziol, Budding, & Chidekel, 2011; Schmahmann &
Pandya, 1997). These parallel cortico-cerebellar loops place the cerebellum
in a unique position to use all the information it receives from the neocortex
to build, through a learning process, an internal “model” that contains all of
the dynamic processes that are required to perform a specific movement or
behavior. This feedforward information (or efferent copy) is used to
generate a representation of the expected sensory consequences of that
command, and to compute error signals that can produce online changes to
adjust its execution and/or to improve future predictions (for review, see
Sokolov, Miall, & Ivry, 2017; Wolpert, Miall, & Kawato, 1998). Indeed,
research has demonstrated that the cerebellum is key in establishing sensory
prediction errors by processing signal discrepancies between the expected
sensory consequences of a stimulus/movement and the actual sensory input
(Baumann et al., 2015; Koziol et al., 2014; Manto et al., 2012; Tseng,
Diedrichsen, Krakauer, Shadmehr, & Bastian, 2007). These error signals are
essential for sensorimotor control, motor adaptation, and learning because
they allow rapid adjustments in the motor output and refinement of future
sensory predictions in order to reduce the variability of subsequent actions
(Doyon, Penhune, & Ungerleider, 2003; Petter, Lusk, Hesslow, & Meck,
2016; Shadmehr, Smith, & Krakauer, 2010; Sokolov et al., 2017).
FIGURE 3. Diagram of cortico-cerebellar and basal ganglia-thalamo-cortical networks and the
intricate connectivity between these circuits. The basal ganglia-thalamo-cortical timing network
normally involves the SMA, PFC, Striatum, PPC, GPe, Th, STN, VTA, and SN. The cerebellar
network involves the Cb Cortex, PN, DN, and IO. Note that the cerebellum is also connected to
multiple cortical and subcortical regions, and that reciprocal connections between the basal ganglia
and the cerebellum are not illustrated. Abbreviations: PFC, prefrontal cortex; SMA, supplementary
motor area; PPC, posterior parietal cortex; Th, thalamus; GPe, globus pallidus; STN, sub thalamic
nuclei; SN, substantia nigra; VTA, ventral tegmental area; PN, pontine nuclei; DN, dentate nucleus;
Cb, cerebellar cortex; IO, inferior olive.
Reproduced with permission from Petter et al. (2016).
R
Adamos, D. A., Laskaris, N., & Micheloyannis, S. (2018). Harnessing functional segregation across
brain rhythms as a means to detect EEG oscillatory multiplexing during music listening. Journal of
Neural Engineering 15, 036012.
Agostino, P. V., & Cheng, R. K. (2016). Contributions of dopaminergic signaling to timing accuracy
and precision. Current Opinion in Behavioral Sciences 8, 153–160.
Akkal, D., Dum, R. P., & Strick, P. L. (2007). Supplementary motor area and presupplementary
motor area: Targets of basal ganglia and cerebellar output. Journal of Neuroscience 27(40),
10659–10673.
Alho, K., Rinne, T., Herron, T. J., & Woods, D. L. (2014). Stimulus-dependent activations and
attention-related modulations in the auditory cortex: A meta-analysis of fMRI studies. Hearing
Research 307, 29–41.
Allman, M. J., & Meck, W. H. (2012). Pathophysiological distortions in time perception and timed
performance. Brain 135(3), 656–677.
Allman, M. J., Teki, S., Griffiths, T. D., & Meck, W. H. (2014). Properties of the internal clock: First-
and second-order principles of subjective time. Annual Review of Psychology 65, 743–771.
Alluri, V., Brattico, E., Toiviainen, P., Burunat, I., Bogert, B., Numminen, J., & Kliuchko, M. (2015).
Musical expertise modulates functional connectivity of limbic regions during continuous music
listening. Psychomusicology 25(4), 443–454.
Alluri, V., Toiviainen, P., Burunat, I., Kliuchko, M., Vuust, P., & Brattico, E. (2017). Connectivity
patterns during music listening: Evidence for action-based processing in musicians. Human Brain
Mapping 38(6), 2955–2970.
Alluri, V., Toiviainen, P., Jääskeläinen, I. P., Glerean, E., Sams, M., & Brattico, E. (2012). Large-
scale brain networks emerge from dynamic processing of musical timbre, key and rhythm.
NeuroImage 59(4), 3677–3689.
Alluri, V., Toiviainen, P., Lund, T. E., Wallentin, M., Vuust, P., Nandi, A. K., … Brattico, E. (2013).
From Vivaldi to Beatles and back: Predicting lateralized brain responses to music. NeuroImage 83,
627–636.
Amunts, K., Morosan, P., Hilbig, H., & Zilles, K. (2012). Auditory system. In J. K. Mai & G.
Paxinos (Eds.), The human nervous system (3rd ed., pp. 1270–1300). London: Elsevier.
Andoh, J., & Zatorre, R. J. (2011). Interhemispheric connectivity influences the degree of modulation
of TMS-induced effects during auditory processing. Frontiers in Psychology 2, 161.
Ashe, J., & Bushara, K. (2014). The olivo-cerebellar system as a neural clock. In H. Merchant & V.
de Lafuente (Eds.), Neurobiology of interval timing: Advances in experimental medicine and
biology (pp. 155–166). New York: Springer.
Bajo, V. M., Nodal, F. R., Moore, D. R., & King, A. J. (2010). The descending corticocollicular
pathway mediates learning-induced auditory plasticity. Nature Neuroscience 13(2), 253–260.
Bangert, M., Peschel, T., Schlaug, G., Rotte, M., Drescher, D., Hinrichs, H., … Altenmüller, E.
(2006). Shared networks for auditory and motor processing in professional pianists: Evidence from
fMRI conjunction. NeuroImage 30(3), 917–926.
Bartolo, R., Prado, L., & Merchant, H. (2014). Information processing in the primate basal ganglia
during sensory-guided and internally driven rhythmic tapping. Journal of Neuroscience 34(11),
3910–3923.
Baumann, O., Borra, R. J., Bower, J. M., Cullen, K. E., Habas, C., Ivry, R. B., … Sokolov, A. A.
(2015). Consensus paper: The role of the cerebellum in perceptual processes. Cerebellum 14(2),
197–220.
Baumann, S., Griffiths, T. D., Sun, L., Petkov, C. I., Thiele, A., & Rees, A. (2011). Orthogonal
representation of sound dimensions in the primate midbrain. Nature Neuroscience 14(4), 423–425.
Baumann, S., Koeneke, S., Schmidt, C. F., Meyer, M., Lutz, K., & Jancke, L. (2007). A network for
audio-motor coordination in skilled pianists and non-musicians. Brain Research 1161(1), 65–78.
Beauchamp, M. S., Argall, B. D., Bodurka, J., Duyn, J. H., & Martin, A. (2004). Unraveling
multisensory integration: Patchy organization within human STS multisensory cortex. Nature
Neuroscience 7(11), 1190–1192.
Beauchamp, M. S., Nath, A. R., & Pasalar, S. (2010). fMRI-guided transcranial magnetic stimulation
reveals that the superior temporal sulcus is a cortical locus of the McGurk effect. Journal of
Neuroscience 30(7), 2414–2417.
Beauchamp, M. S., Yasar, N. E., Frye, R. E., & Ro, T. (2008). Touch, sound and vision in human
superior temporal sulcus. NeuroImage 41(3), 1011–1020.
Belin, P., & Zatorre, R. J. (2000). “What,” “where” and “how” in auditory cortex. Nature
Neuroscience 3(10), 965–966.
Bhattacharya, J., & Petsche, H. (2005). Phase synchrony analysis of EEG during music perception
reveals changes in functional connectivity due to musical expertise. Signal Processing 85(11),
2161–2177.
Bianco, R., Novembre, G., Keller, P. E. E., Kim, S.-G. G., Scharf, F., Friederici, A. D. D., …
Sammler, D. (2016). Neural networks for harmonic structure in music perception and action.
NeuroImage 142, 454–464.
Bizley, J. K., & Cohen, Y. E. (2013). The what, where and how of auditory-object perception. Nature
Reviews Neuroscience 14(10), 693–707.
Blood, A. J., & Zatorre, R. J. (2001). Intensely pleasurable responses to music correlate with activity
in brain regions implicated in reward and emotion. Proceedings of the National Academy of
Sciences 98(20), 11818–11823.
Bostan, A. C., Dum, R. P., & Strick, P. L. (2013). Cerebellar networks with the cerebral cortex and
basal ganglia. Trends in Cognitive Sciences 17(5), 241–254.
Brattico, E., Alluri, V., Bogert, B., Jacobsen, T., Vartiainen, N., Nieminen, S., & Tervaniemi, M.
(2011). A functional MRI study of happy and sad emotions in music with and without lyrics.
Frontiers in Psychology 2, 308.
Brechmann, A., & Scheich, H. (2005). Hemispheric shifts of sound representation in auditory cortex
with conceptual listening. Cerebral Cortex 15(5), 578–587.
Brown, S., & Martinez, M. J. (2007). Activation of premotor vocal areas during musical
discrimination. Brain and Cognition 63(1), 59–69.
Brown, S., Martinez, M. J., & Parsons, L. M. (2006). The neural basis of human dance. Cerebral
Cortex 16(8), 1157–1167.
Buckner, R. L. (2013). The cerebellum and cognitive function: 25 years of insight from anatomy and
neuroimaging. Neuron 80(3), 807–815.
Buhusi, C. V., & Meck, W. H. (2005). What makes us tick? Functional and neural mechanisms of
interval timing. Nature Reviews Neuroscience 6(10), 755–765.
Burunat, I., Alluri, V., Toiviainen, P., Numminen, J., & Brattico, E. (2014). Dynamics of brain
activity underlying working memory for music in a naturalistic condition. Cortex 57, 254–269.
Caligiore, D., Pezzulo, G., Baldassarre, G., Bostan, A. C., Strick, P. L., Doya, K., … Herreros, I.
(2017). Consensus paper: Towards a systems-level view of cerebellar function: The interplay
between cerebellum, basal ganglia, and cortex. Cerebellum 16(1), 203–229.
Cammoun, L., Thiran, J. P., Griffa, A., Meuli, R., Hagmann, P., & Clarke, S. (2015).
Intrahemispheric cortico-cortical connections of the human auditory cortex. Brain Structure &
Function 220(6), 3537–3553.
Cha, K., Zatorre, R. J., & Schönwiesner, M. (2016). Frequency selectivity of voxel-by-voxel
functional connectivity in human auditory cortex. Cerebral Cortex 26(1), 211–224.
Chapin, H. L., Zanto, T., Jantzen, K. J., Kelso, S. J. A. A., Steinberg, F., & Large, E. W. (2010).
Neural responses to complex auditory rhythms: The role of attending. Frontiers in Psychology 1,
547–558.
Chauvigné, L. A. S., Gitau, K. M., & Brown, S. (2014). The neural basis of audiomotor entrainment:
An ALE meta-analysis. Frontiers in Human Neuroscience 8, 776.
Chen, C. H., Fremont, R., Arteaga-Bracho, E. E., & Khodakhah, K. (2014). Short latency cerebellar
modulation of the basal ganglia. Nature Neuroscience 17(12), 1767–1775.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008a). Listening to musical rhythms recruits motor
regions of the brain. Cerebral Cortex 18(12), 2844–2854.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008b). Moving on time: Brain network for auditory-
motor synchronization is modulated by rhythm complexity and musical training. Journal of
Cognitive Neuroscience 20(2), 226–239.
Chen, J. L., Rae, C., & Watkins, K. E. (2012). Learning to play a melody: An fMRI study examining
the formation of auditory-motor associations. NeuroImage 59(2), 1200–1208.
Chen, J. L., Zatorre, R. J., & Penhune, V. B. (2006). Interactions between auditory and dorsal
premotor cortex during synchronization to musical rhythms. NeuroImage 32(4), 1771–1781.
Choppin, S., Trost, W., Dondaine, T., Millet, B., Drapier, D., Vérin, M., … Grandjean, D. (2016).
Alteration of complex negative emotions induced by music in euthymic patients with bipolar
disorder. Journal of Affective Disorders 191, 15–23.
Coull, J. T., Cheng, R. K., & Meck, W. H. (2011). Neuroanatomical and neurochemical substrates of
timing. Neuropsychopharmacology 36(1), 3–25.
Coull, J. T., Hwang, H. J., Leyton, M., & Dagher, A. (2012). Dopamine precursor depletion impairs
timing in healthy volunteers by attenuating activity in putamen and supplementary motor area.
Journal of Neuroscience 32(47), 16704–16715.
Coull, J. T., Vidal, F., & Burle, B. (2016). When to act, or not to act: That’s the SMA’s question.
Current Opinion in Behavioral Sciences 8, 14–21.
Cunnington, R., Bradshaw, J. L., & Iansek, R. (1996). The role of the supplementary motor area in
the control of voluntary movement. Human Movement Science 15(5), 627–647.
D’Ausilio, A., Altenmüller, E., Olivetti Belardinelli, M., & Lotze, M. (2006). Cross-modal plasticity
of the motor cortex while listening to a rehearsed musical piece. European Journal of
Neuroscience 24(3), 955–958.
Da Costa, S., van der Zwaag, W., Marques, J. P., Frackowiak, R. S. J., Clarke, S., & Saenz, M.
(2011). Human primary auditory cortex follows the shape of Heschl’s gyrus. Journal of
Neuroscience 31(40), 14067–14075.
de Heer, W. A., Huth, A. G., Griffiths, T. L., Gallant, J. L., & Theunissen, F. E. (2017). The
hierarchical cortical organization of human speech processing. Journal of Neuroscience 37(27),
6539–6557.
Del Olmo, M. F., Cheeran, B., Koch, G., & Rothwell, J. C. (2007). Role of the cerebellum in
externally paced rhythmic finger movements. Journal of Neurophysiology 98(1), 145–152.
Diedrichsen, J., Criscimagna-Hemminger, S. E., & Shadmehr, R. (2007). Dissociating timing and
coordination as functions of the cerebellum. Journal of Neuroscience 27(23), 6291–6301.
Doyon, J., Penhune, V., & Ungerleider, L. G. (2003). Distinct contribution of the cortico-striatal and
cortico-cerebellar systems to motor skill learning. Neuropsychologia 41(3), 252–262.
Dum, R. P. (2002). An unfolded map of the cerebellar dentate nucleus and its projections to the
cerebral cortex. Journal of Neurophysiology 89(1), 634–639.
Durstewitz, D. (2003). Self-organizing neural integrator predicts interval times through climbing
activity. Journal of Neuroscience 23(12), 5342–5353.
Farrugia, N., Jakubowski, K., Cusack, R., & Stewart, L. (2015). Tunes stuck in your brain: The
frequency and affective evaluation of involuntary musical imagery correlate with cortical structure.
Consciousness and Cognition 35, 66–77.
Fatemi, S. H., Aldinger, K. A., Ashwood, P., Bauman, M. L., Blaha, C. D., Blatt, G. J., …Welsh, J. P.
(2012). Consensus paper: Pathological role of the cerebellum in autism. Cerebellum 11(3), 777–
807.
Fernández-Miranda, J. C., Wang, Y., Pathak, S., Stefaneau, L., Verstynen, T., & Yeh, F. C. (2015).
Asymmetry, connectivity, and segmentation of the arcuate fascicle in the human brain. Brain
Structure & Function 220(3), 1665–1680.
Foster, N. E. V. V, Halpern, A. R., & Zatorre, R. J. (2013). Common parietal activation in musical
mental transformations across pitch and time. NeuroImage 75, 27–35.
Friston, K. J. (2011). Functional and effective connectivity: A review. Brain Connectivity 1(1), 13–
36.
Froud, K. E., Wong, A. C. Y., Cederholm, J. M. E., Klugmann, M., Sandow, S. L., Julien, J.-P., …
Housley, G. D. (2015). Type II spiral ganglion afferent neurons drive medial olivocochlear reflex
suppression of the cochlear amplifier. Nature Communications 6(1), 7115.
Frühholz, S., Trost, W., & Grandjean, D. (2014). The role of the medial temporal limbic system in
processing emotions in voice and music. Progress in Neurobiology 123, 1–17.
Frühholz, S., Trost, W., & Kotz, S. A. (2016). The sound of emotions: Towards a unifying neural
network perspective of affective sound processing. Neuroscience & Biobehavioral Reviews 68,
96–110.
Fujioka, T., Trainor, L. J., Large, E. W., & Ross, B. (2012). Internalized Timing of isochronous
sounds is represented in neuromagnetic beta oscillations. Journal of Neuroscience 32(5), 1791–
1802.
Gaab, N., Gaser, C., Zaehle, T., Jancke, L., & Schlaug, G. (2003). Functional anatomy of pitch
memory: An fMRI study with sparse temporal sampling. NeuroImage 19(4), 1417–1426.
Gao, J. H., Parsons, L. M., Bower, J. M., Xiong, J., Li, J., & Fox, P. T. (1996). Cerebellum implicated
in sensory acquisition and discrimination rather than motor control. Science 272(5261), 545–547.
Giovannelli, F., Banfi, C., Borgheresi, A., Fiori, E., Innocenti, I., Rossi, S., … Cincotta, M. (2013).
The effect of music on corticospinal excitability is related to the perceived emotion: A transcranial
magnetic stimulation study. Cortex 49(3), 702–710.
Grahn, J. A., & Brett, M. (2007). Rhythm and beat perception in motor areas of the brain. Journal of
Cognitive Neuroscience 19(5), 893–906.
Grahn, J. A., Henry, M. J., & McAuley, J. D. (2011). fMRI investigation of cross-modal interactions
in beat perception: Audition primes vision, but not vice versa. NeuroImage 54(2), 1231–1243.
Grahn, J. A., & Rowe, J. B. (2009). Feeling the beat: Premotor and striatal interactions in musicians
and nonmusicians during beat perception. Journal of Neuroscience 29(23), 7540–7548.
Grahn, J. A., & Rowe, J. B. (2013). Finding and feeling the musical beat: Striatal dissociations
between detection and prediction of regularity. Cerebral Cortex 23(4), 913–921.
Griffiths, T. D., & Warren, J. D. (2002). The planum temporale as a computational hub. Trends in
Neurosciences 25(7), 348–353.
Griffiths, T. D., & Warren, J. D. (2004). What is an auditory object? Nature Reviews Neuroscience
5(11), 887–892.
Grothe, B. (2000). The evolution of temporal processing in the medial superior olive, an auditory
brainstem structure. Progress in Neurobiology 61(6), 581–610.
Groussard, M., Viader, F., Hubert, V., Landeau, B., Abbas, A., Desgranges, B., … Platel, H. (2010).
Musical and verbal semantic memory: Two distinct neural networks? NeuroImage 49(3), 2764–
2773.
Grube, M., Cooper, F. E., Chinnery, P. F., & Griffiths, T. D. (2010). Dissociation of duration-based
and beat-based auditory timing in cerebellar degeneration. Proceedings of the National Academy of
Sciences 107(25), 11597–11601.
Grube, M., Lee, K. H., Griffiths, T. D., Barker, A. T., & Woodruff, P. W. (2010). Transcranial
magnetic theta-burst stimulation of the human cerebellum distinguishes absolute, duration-based
from relative, beat-based perception of subsecond time intervals. Frontiers in Psychology 1, 171.
Haber, S. N., & Knutson, B. (2010). The reward circuit: Linking primate anatomy and human
imaging. Neuropsychopharmacology 35(1), 4–26.
Halpern, A. R., & Zatorre, R. J. (1999). When that tune runs through your head: A PET investigation
of auditory imagery for familiar melodies. Cerebral Cortex 9(7), 697–704.
Halpern, A. R., Zatorre, R. J., Bouffard, M., & Johnson, J. A. (2004). Behavioral and neural
correlates of perceived and imagined musical timbre. Neuropsychologia 42(9), 1281–1292.
Halsband, U., Ito, N., Tanji, J., & Freund, H. J. (1993). The role of premotor cortex and the
supplementary motor area in the temporal control of movement in man. Brain 116(1), 243–266.
Harrington, D. L., Castillo, G. N., Greenberg, P. A., Song, D. D., Lessig, S., Lee, R. R., & Rao, S. M.
(2011). Neurobehavioral mechanisms of temporal processing deficits in Parkinson’s disease. PLoS
ONE 6(2), e17461.
Harrington, D. L., & Jahanshahi, M. (2016). Reconfiguration of striatal connectivity for timing and
action. Current Opinion in Behavioral Sciences 8, 78–84.
Harris, R., & De Jong, B. M. (2014). Cerebral activations related to audition-driven performance
imagery in professional musicians. PLoS ONE, 9(4), e93681.
Hasegawa, T., Matsuki, K. I., Ueno, T., Maeda, Y., Matsue, Y., Konishi, Y., & Sadato, N. (2004).
Learned audio-visual cross-modal associations in observed piano playing activate the left planum
temporale: An fMRI study. Cognitive Brain Research 20(3), 510–518.
Haueisen, J., & Knösche, T. R. (2001). Involuntary motor activity in pianists evoked by music
perception. Journal of Cognitive Neuroscience 13(6), 786–792.
Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews
Neuroscience 8(5), 393–402.
Horn, A. K. E. (2006). The reticular formation. Progress in Brain Research 151, 127–155.
Huffman, R. F., & Henson, O. W. (1990). The descending auditory pathway and acousticomotor
systems: Connections with the inferior colliculus. Brain Research Reviews 15(3), 295–323.
Humphries, C., Liebenthal, E., & Binder, J. R. (2010). Tonotopic organization of human auditory
cortex. NeuroImage 50(3), 1202–1211.
Huron, D. B. (2006). Sweet anticipation: Music and the psychology of expectation. Cambridge, MA:
MIT Press.
Hyde, K. L., Peretz, I., & Zatorre, R. J. (2008). Evidence for the role of the right auditory cortex in
fine pitch resolution. Neuropsychologia 46(2), 632–639.
Ito, M. (2008). Control of mental activities by internal models in the cerebellum. Nature Reviews
Neuroscience 9(4), 304–313.
Iversen, J. R., & Balasubramaniam, R. (2016). Synchronization and temporal processing. Current
Opinion in Behavioral Sciences 8, 175–180.
Ivry, R. B., Spencer, R. M., Zelaznik, H. N., & Diedrichsen, J. (2002). The cerebellum and event
timing. Annals of the New York Academy of Sciences 978, 302–317.
Jahanshahi, M., Jenkins, I. H., Brown, R. G., Marsden, C. D., Passingham, R. E., & Brooks, D. J.
(1995). Self-initiated versus externally triggered movements: I. An investigation using
measurement of regional cerebral blood flow with PET and movement-related potentials in normal
and Parkinson’s disease subjects. Brain 118(4), 913–933.
Jahanshahi, M., Jones, C. R. G., Zijlmans, J., Katzenschlager, R., Lee, L., Quinn, N., … Lees, A. J.
(2010). Dopaminergic modulation of striato-frontal connectivity during motor timing in
Parkinson’s disease. Brain 133(3), 727–745.
Janak, P. H., & Tye, K. M. (2015). From circuits to behaviour in the amygdala. Nature 517(7534),
284–292.
Janata, P. (2009). The neural architecture of music-evoked autobiographical memories. Cerebral
Cortex 19(11), 2579–2594.
Janata, P. (2015). Neural basis of music perception. In G. G. Celesia & G. Hickok (Eds.), Handbook
of clinical neurology: The human auditory system (Vol. 129, pp. 187–205). Amsterdam: Elsevier.
Janata, P., Birk, J., Van Horn, J., Leman, M., Tillmann, B., & Bharucha, J. J. (2002). The cortical
topography of tonal structures underlying Western music. Science 293(5539), 2425–2430.
Janata, P., Tillmann, B., & Bharucha, J. J. (2002). Listening to polyphonic music recruits domain-
general attention and working memory circuits. Cognitive, Affective & Behavioral Neuroscience
2(2), 121–140.
Jäncke, L., Loose, R., Lutz, K., Specht, K., & Shah, N. (2000). Cortical activations during paced
finger-tapping applying visual and auditory pacing stimuli. Cognitive Brain Research 10(1–2), 51–
66.
Juslin, P. N., & Västfjäll, D. (2008). Emotional responses to music: The need to consider underlying
mechanisms. Behavioral and Brain Sciences 31(5), 559–575.
Kaiser, J., Ripper, B., Birbaumer, N., & Lutzenberger, W. (2003). Dynamics of gamma-band activity
in human magnetoencephalogram during auditory pattern working memory. NeuroImage 20(2),
816–827.
Keller, P. E. (2012). Mental imagery in music performance: Underlying mechanisms and potential
benefits. Annals of the New York Academy of Sciences 1252(1), 206–213.
Kelly, R. M., & Strick, P. L. (2003). Cerebellar loops with motor cortex and prefrontal cortex of a
nonhuman primate. Journal of Neuroscience 23(23), 8432–8444.
Keren-Happuch, E., Chen, S. H. A., Ho, M. H. R., & Desmond, J. E. (2014). A meta-analysis of
cerebellar contributions to higher cognition from PET and fMRI studies. Human Brain Mapping
35(2), 593–615.
Kiehl, K. A., Laurens, K. R., Duty, T. L., Forster, B. B., & Liddle, P. F. (2001). Neural sources
involved in auditory target detection and novelty processing: An event-related fMRI study.
Psychophysiology 38(1), 133–142.
Klein, C., Liem, F., Hänggi, J., Elmer, S., & Jäncke, L. (2016). The “silent” imprint of musical
training. Human Brain Mapping 37(2), 536–546.
Klein, M. E., & Zatorre, R. J. (2011). A role for the right superior temporal sulcus in categorical
perception of musical chords. Neuropsychologia 49(5), 878–887.
Klein, M. E., & Zatorre, R. J. (2015). Representations of invariant musical categories are decodable
by pattern analysis of locally distributed BOLD responses in superior temporal and intraparietal
sulci. Cerebral Cortex 25(7), 1947–1957.
Koelsch, S. (2006). Significance of Broca’s area and ventral premotor cortex for music-syntactic
processing. Cortex 42(4), 518–520.
Koelsch, S. (2011). Toward a neural basis of music perception: A review and updated model.
Frontiers in Psychology 2, 110.
Koelsch, S. (2014). Brain correlates of music-evoked emotions. Nature Reviews Neuroscience 15(3),
170–180.
Koelsch, S., Fritz, T., Schulze, K., Alsop, D., & Schlaug, G. (2005). Adults and children processing
music: An fMRI study. NeuroImage 25(4), 1068–1076.
Koelsch, S., Fritz, T., v. Cramon, D. Y., Müller, K., & Friederici, A. D. (2006). Investigating emotion
with music: An fMRI study. Human Brain Mapping 27(3), 239–250.
Koelsch, S., Gunter, T. C., v. Cramon, D. Y., Zysset, S., Lohmann, G., & Friederici, A. D. (2002).
Bach speaks: A cortical “language-network” serves the processing of music. NeuroImage 17(2),
956–966.
Koelsch, S., Schulze, K., Sammler, D., Fritz, T., Müller, K., & Gruber, O. (2009). Functional
architecture of verbal and tonal working memory: An fMRI study. Human Brain Mapping 30(3),
859–873.
Koelsch, S., & Skouras, S. (2014). Functional centrality of amygdala, striatum and hypothalamus in a
“small-world” network underlying joy: An fMRI study with music. Human Brain Mapping 35(7),
3485–3498.
Koelsch, S., Skouras, S., Fritz, T., Herrera, P., Bonhage, C., Küssner, M. B., & Jacobs, A. M. (2013).
The roles of superficial amygdala and auditory cortex in music-evoked fear and joy. NeuroImage
81, 49–60.
Koelsch, S., Skouras, S., & Lohmann, G. (2018). The auditory cortex hosts network nodes influential
for emotion processing: An fMRI study on music-evoked fear and joy. PLoS ONE 13(1),
e0190057.
Kornysheva, K., & Schubotz, R. I. (2011). Impairment of auditory-motor timing and compensatory
reorganization after ventral premotor cortex stimulation. PLoS ONE 6(6), e21421.
Kotz, S. A., Brown, R. M., & Schwartze, M. (2016). Cortico-striatal circuits and the timing of action
and perception. Current Opinion in Behavioral Sciences 8, 42–45.
Kotz, S. A., Stockert, A., & Schwartze, M. (2014). Cerebellum, temporal predictability and the
updating of a mental model. Philosophical Transactions of the Royal Society B: Biological
Sciences 369(1658), 20130403.
Koziol, L. F., Budding, D., Andreasen, N., D’Arrigo, S., Bulgheroni, S., Imamizu, H., …Yamazaki,
T. (2014). Consensus paper: The cerebellum’s role in movement and cognition. Cerebellum 13(1),
151–177.
Koziol, L. F., Budding, D. E., & Chidekel, D. (2011). Sensory integration, sensory processing, and
sensory modulation disorders: Putative functional neuroanatomic underpinnings. Cerebellum
10(4), 770–792.
Lahav, A., Saltzman, E., & Schlaug, G. (2007). Action representation of sound: Audiomotor
recognition network while listening to newly acquired actions. Journal of Neuroscience 27(2),
308–314.
Langers, D. R. M. (2014). Assessment of tonotopically organised subdivisions in human auditory
cortex using volumetric and surface-based cortical alignments. Human Brain Mapping 35(4),
1544–1561.
Lappe, C., Lappe, M., & Pantev, C. (2016). Differential processing of melodic, rhythmic and simple
tone deviations in musicians: An MEG study. NeuroImage 124, 898–905.
Lappe, C., Steinsträter, O., & Pantev, C. (2013). Rhythmic and melodic deviations in musical
sequences recruit different cortical areas for mismatch detection. Frontiers in Human
Neuroscience 7, 260.
Large, E. W., Herrera, J. A., & Velasco, M. J. (2015). Neural networks for beat perception in musical
rhythm. Frontiers in Systems Neuroscience 9, 159.
Large, E. W., & Snyder, J. S. (2009). Pulse and meter as neural resonance. Annals of the New York
Academy of Sciences 1169, 46–57.
LeDoux, J. (2007). The amygdala. Current Biology 17(20), R868–R874.
Lee, K.-H., Egleston, P. N., Brown, W. H., Gregory, A. N., Barker, A. T., & Woodruff, P. W. R.
(2007). The role of the cerebellum in subsecond time perception: Evidence from repetitive
transcranial magnetic stimulation. Journal of Cognitive Neuroscience 19(1), 147–157.
Lee, Y. S., Janata, P., Frost, C., Hanke, M., & Granger, R. (2011). Investigation of melodic contour
processing in the brain using multivariate pattern-based fMRI. NeuroImage 57(1), 293–300.
Lehéricy, S., Ducros, M., Krainik, A., Francois, C., Van De Moortele, P. F., Ugurbil, K., & Kim, D. S.
(2004). 3-D diffusion tensor axonal tracking shows distinct SMA and pre-SMA projections to the
human striatum. Cerebral Cortex 14(12), 1302–1309.
Lehne, M., Rohrmeier, M., & Koelsch, S. (2013). Tension-related activity in the orbitofrontal cortex
and amygdala: An fMRI study with music. Social Cognitive and Affective Neuroscience 9(10),
1515–1523.
Leow, L. A., & Grahn, J. A. (2014). Neural mechanisms of rhythm perception: Present findings and
future directions. Advances in Experimental Medicine and Biology 829, 325–338.
Lima, C. F., Krishnan, S., & Scott, S. K. (2016). Roles of supplementary motor areas in auditory
processing and auditory imagery. Trends in Neurosciences 39(8), 527–542.
Loui, P. (2015). A dual-stream neuroanatomy of singing. Music Perception 32(3), 232–241.
Lusk, N. A., Petter, E. A., Macdonald, C. J., & Meck, W. H. (2016). Cerebellar, hippocampal, and
striatal time cells. Current Opinion in Behavioral Sciences 8, 186–192.
Maes, P.-J., Leman, M., Palmer, C., & Wanderley, M. M. (2014). Action-based effects on music
perception. Frontiers in Psychology 4, 1008.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in
Broca’s area: An MEG study. Nature Neuroscience 4(5), 540–545.
Maidhof, C., & Koelsch, S. (2011). Effects of selective attention on syntax processing in music and
language. Journal of Cognitive Neuroscience 23(9), 2252–2267.
Manto, M., Bower, J. M., Conforto, A. B., Delgado-García, J. M., Da Guarda, S. N. F., Gerwig, M.,
… Timmann, D. (2012). Consensus paper: Roles of the cerebellum in motor control: The diversity
of ideas on cerebellar involvement in movement. Cerebellum 11, 457–487.
Martínez-Molina, N., Mas-Herrero, E., Rodríguez-Fornells, A., Zatorre, R. J., & Marco-Pallarés, J.
(2016). Neural correlates of specific musical anhedonia. Proceedings of the National Academy of
Sciences 113(46), E7337–E7345.
Marvel, C. L., & Desmond, J. E. (2010). Functional topography of the cerebellum in verbal working
memory. Neuropsychology Review 20(3), 271–279.
Mas-Herrero, E., Dagher, A., & Zatorre, R. J. (2017). Modulating musical reward sensitivity up and
down with transcranial magnetic stimulation. Nature Human Behaviour 2(1), 27–32.
Matell, M. S., & Meck, W. H. (2004). Cortico-striatal circuits and interval timing: Coincidence
detection of oscillatory processes. Cognitive Brain Research 21(2), 139–170.
Mauk, M. D., & Buonomano, D. V. (2004). The neural basis of temporal processing. Annual Review
of Neuroscience 27, 307–340.
Mayville, J. M., Jantzen, K. J., Fuchs, A., Steinberg, F. L., & Kelso, J. A. S. (2002). Cortical and
subcortical networks underlying syncopated and synchronized coordination revealed using fMRI.
Human Brain Mapping 17(4), 214–229.
Medina, J. F., & Mauk, M. D. (2000). Computer simulation of cerebellar information processing.
Nature Neuroscience 3(Suppl.), 1205–1211.
Menon, V., & Levitin, D. J. (2005). The rewards of music listening: Response and physiological
connectivity of the mesolimbic system. NeuroImage 28(1), 175–184.
Merchant, H., & Bartolo, R. (2018). Primate beta oscillations and rhythmic behaviors. Journal of
Neural Transmission 125, 461–470.
Merchant, H., Grahn, J., Trainor, L., Rohrmeier, M., & Fitch, W. T. (2015). Finding the beat: A
neural perspective across humans and non-human primates. Philosophical Transactions of the
Royal Society B: Biological Sciences 370(1664), 20140093.
Merchant, H., Harrington, D. L., & Meck, W. H. (2013). Neural basis of the perception and
estimation of time. Annual Review of Neuroscience 36, 313–336.
Merchant, H., Perez, O., Zarco, W., & Gamez, J. (2013). Interval tuning in the primate medial
premotor cortex as a general timing mechanism. Journal of Neuroscience 33(21), 9082–9096.
Michaelis, K., Wiener, M., & Thompson, J. C. (2014). Passive listening to preferred motor tempo
modulates corticospinal excitability. Frontiers in Human Neuroscience 8, 252.
Middleton, F. A., & Strick, P. L. (2001). Cerebellar projections to the prefrontal cortex of the primate.
Journal of Neuroscience 21(2), 700–712.
Mitterschiffthaler, M. T., Fu, C. H. Y., Dalton, J. A., Andrew, C. M., & Williams, S. C. R. (2007). A
functional MRI study of happy and sad affective states induced by classical music. Human Brain
Mapping 28(11), 1150–1162.
Molinari, M., Leggio, M. G., Filippini, V., Gioia, M. C., Cerasa, A., & Thaut, M. H. (2005).
Sensorimotor transduction of time information is preserved in subjects with cerebellar damage.
Brain Research Bulletin 67, 448–458.
Molinari, M., Leggio, M. G., & Thaut, M. H. (2007). The cerebellum and neural networks for
rhythmic sensorimotor synchronization in the human brain. Cerebellum 6(1), 18–23.
Moore, E., Schaefer, R. S., Bastin, M. E., Roberts, N., & Overy, K. (2014). Can musical training
influence brain connectivity? Evidence from diffusion tensor MRI. Brain Sciences 4(2), 405–427.
Morillon, B., & Baillet, S. (2017). Motor origin of temporal predictions in auditory attention.
Proceedings of the National Academy of Sciences 114(42), E8913–E8921.
Mueller, K., Fritz, T., Mildner, T., Richter, M., Schulze, K., Lepsien, J. J., … Möller, H. E. (2015).
Investigating the dynamics of the brain response to music: A central role of the ventral
striatum/nucleus accumbens. NeuroImage 116, 68–79.
Nachev, P., Kennard, C., & Husain, M. (2008). Functional role of the supplementary and pre-
supplementary motor areas. Nature Reviews Neuroscience 9, 856–869.
Narayanan, N. S., Land, B. B., Solder, J. E., Deisseroth, K., & DiLeone, R. J. (2012). Prefrontal D1
dopamine signaling is required for temporal control. Proceedings of the National Academy of
Sciences 109(50), 20726–20731.
Nayagam, B. A., Muniak, M. A., & Ryugo, D. K. (2011). The spiral ganglion: Connecting the
peripheral and central auditory systems. Hearing Research 278(1–2), 2–20.
Nelson, A., Schneider, D. M., Takatoh, J., Sakurai, K., Wang, F., & Mooney, R. (2013). A circuit for
motor cortical modulation of auditory cortical activity. Journal of Neuroscience 33(36), 14342–
14353.
Norman-Haignere, S., Kanwisher, N., & McDermott, J. H. (2013). Cortical pitch regions in humans
respond primarily to resolved harmonics and are located in specific tonotopic regions of anterior
auditory cortex. Journal of Neuroscience 33(50), 19451–19469.
Novembre, G., & Keller, P. E. (2014). A conceptual review on action-perception coupling in the
musicians’ brain: What is it good for? Frontiers in Human Neuroscience 8, 603.
Nozaradan, S., Schwartze, M., Obermeier, C., & Kotz, S. A. (2017). Specific contributions of basal
ganglia and cerebellum to the neural tracking of rhythm. Cortex 95, 156–168.
O’Reilly, J. X., Mesulam, M. M., & Nobre, A. C. (2008). The cerebellum predicts the timing of
perceptual events. Journal of Neuroscience 28(9), 2252–2260.
Ohnishi, T., Matsuda, H., Asada, T., Aruga, M., Hirakata, M., Nishikawa, M., … Imabayashi, E.
(2001). Functional anatomy of musical perception in musicians. Cerebral Cortex 11(8), 754–760.
Pallesen, K. J., Brattico, E., Bailey, C., Korvenoja, A., Koivisto, J., Gjedde, A., & Carlson, S. (2005).
Emotion processing of major, minor, and dissonant chords: A functional magnetic resonance
imaging study. Annals of the New York Academy of Sciences 1060, 450–453.
Palomar-García, M. Á., Zatorre, R. J., Ventura-Campos, N., Bueichekú, E., & Ávila, C. (2017).
Modulation of functional connectivity in auditory-motor networks in musicians compared with
nonmusicians. Cerebral Cortex 27(5), 2768–2778.
Pannese, A., Grandjean, D., & Frühholz, S. (2016). Amygdala and auditory cortex exhibit distinct
sensitivity to relevant acoustic features of auditory emotions. Cortex 85, 116–125.
Paquette, S., Fujii, S., Li, H. C., & Schlaug, G. (2017). The cerebellum’s contribution to beat interval
discrimination. NeuroImage 163, 177–182.
Parsons, L. M. (2012). Exploring the functional neuroanatomy of music performance, perception, and
comprehension. The Cognitive Neuroscience of Music 930(1), 211–231.
Parsons, L. M., Petacchi, A., Schmahmann, J. D., & Bower, J. M. (2009). Pitch discrimination in
cerebellar patients: Evidence for a sensory deficit. Brain Research 1303, 84–96.
Patterson, R. D., Uppenkamp, S., Johnsrude, I. S., & Griffiths, T. D. (2002). The processing of
temporal pitch and melody information in auditory cortex. Neuron 36(4), 767–776.
Pecenka, N., Engel, A., & Keller, P. E. (2013). Neural correlates of auditory temporal predictions
during sensorimotor synchronization. Frontiers in Human Neuroscience 7, 380.
Pelzer, E. A., Melzer, C., Timmermann, L., von Cramon, D. Y., & Tittgemeyer, M. (2017). Basal
ganglia and cerebellar interconnectivity within the human thalamus. Brain Structure and Function
222(1), 381–392.
Perani, D. (2012). Functional and structural connectivity for language and music processing at birth.
Rendiconti Lincei 23(3), 305–314.
Peretz, I., Gosselin, N., Belin, P., Zatorre, R. J., Plailly, J., & Tillmann, B. (2009). Music lexical
networks: The cortical organization of music recognition. Annals of the New York Academy of
Sciences 1169, 256–265.
Peretz, I., & Zatorre, R. J. (2005). Brain organization for music processing. Annual Review of
Psychology 56, 89–114.
Petter, E. A., Lusk, N. A., Hesslow, G., & Meck, W. H. (2016). Interactive roles of the cerebellum
and striatum in sub-second and supra-second timing: Support for an initiation, continuation,
adjustment, and termination (ICAT) model of temporal processing. Neuroscience & Biobehavioral
Reviews 71, 739–755.
Pfordresher, P. Q., Mantell, J. T., Brown, S., Zivadinov, R., & Cox, J. L. (2014). Brain responses to
altered auditory feedback during musical keyboard production: An fMRI study. Brain Research
1556, 28–37.
Plakke, B., & Romanski, L. M. (2014). Auditory connections and functions of prefrontal cortex.
Frontiers in Neuroscience 8, 199.
Platel, H., Baron, J. C., Desgranges, B., Bernard, F., & Eustache, F. (2003). Semantic and episodic
memory of music are subserved by distinct neural networks. NeuroImage 20(1), 244–256.
Proverbio, A. M., Orlandi, A., & Pisanu, F. (2016). Brain processing of consonance/dissonance in
musicians and controls: A hemispheric asymmetry revisited. European Journal of Neuroscience
44(6), 2340–2356.
Rao, S. M., Harrington, D. L., Haaland, K. Y., Bobholz, J. A., Cox, R. W., & Binder, J. R. (1997).
Distributed neural systems underlying the timing of movements. Journal of Neuroscience 17(14),
5528–5535.
Rauschecker, J. P., & Scott, S. K. (2009). Maps and streams in the auditory cortex: Nonhuman
primates illuminate human speech processing. Nature Neuroscience 12(6), 718–724.
Reser, D. H., Burman, K. J., Richardson, K. E., Spitzer, M. W., & Rosa, M. G. P. (2009). Connections
of the marmoset rostrotemporal auditory area: Express pathways for analysis of affective content
in hearing. European Journal of Neuroscience 30(4), 578–592.
Reybrouck, M., & Brattico, E. (2015). Neuroplasticity beyond sounds: Neural adaptations following
long-term musical aesthetic experiences. Brain Sciences 5(1), 69–91.
Roberts, T. F., Hisey, E., Tanaka, M., Kearney, M. G., Chattree, G., Yang, C. F., … Mooney, R.
(2017). Identification of a motor-to-auditory pathway important for vocal learning. Nature
Neuroscience 20(7), 978–986.
Rogenmoser, L., Zollinger, N., Elmer, S., & Jäncke, L. (2016). Independent component processes
underlying emotions during natural music listening. Social Cognitive and Affective Neuroscience
11(9), 1428–1439.
Ross, B., Barat, M., & Fujioka, T. (2017). Sound-making actions lead to immediate plastic changes
of neuromagnetic evoked responses and induced β-band oscillations during perception. Journal of
Neuroscience 37(24), 5948–5959.
Ross, J. M., Iversen, J. R., & Balasubramaniam, R. (2016). Motor simulation theories of musical beat
perception. Neurocase 22(6), 558–565.
Rossignol, S., & Melvill Jones, G. (1976). Audio-spinal influence in man studied by the H-reflex and
its possible role on rhythmic movements synchronized to sound. Electroencephalography and
Clinical Neurophysiology 41(1), 83–92.
Royal, I., Vuvan, D. T., Zendel, B. R., Robitaille, N., Schönwiesner, M., & Peretz, I. (2016).
Activation in the right inferior parietal lobule reflects the representation of musical structure
beyond simple pitch discrimination. PLoS ONE 11(5), e0155291.
Sachs, M. E., Ellis, R. J., Schlaug, G., & Loui, P. (2016). Brain connectivity reflects human aesthetic
responses to music. Social Cognitive and Affective Neuroscience 11(6), 884–891.
Salimpoor, V. N., Benovoy, M., Larcher, K., Dagher, A., & Zatorre, R. J. (2011). Anatomically
distinct dopamine release during anticipation and experience of peak emotion to music. Nature
Neuroscience 14(2), 257–264.
Salimpoor, V. N., Van Den Bosch, I., Kovacevic, N., McIntosh, A. R., Dagher, A., & Zatorre, R. J.
(2013). Interactions between the nucleus accumbens and auditory cortices predict music reward
value. Science 340(6129), 216–219.
Salimpoor, V. N., Zald, D. H., Zatorre, R. J., Dagher, A., & McIntosh, A. R. (2015). Predictions and
the brain: How musical sounds become rewarding. Trends in Cognitive Sciences 19(2), 86–91.
Sänger, J., Müller, V., & Lindenberger, U. (2012). Intra- and interbrain synchronization and network
properties when playing guitar in duets. Frontiers in Human Neuroscience 6, 312.
Santoro, R., Moerel, M., De Martino, F., Goebel, R., Ugurbil, K., Yacoub, E., & Formisano, E.
(2014). Encoding of natural sounds at multiple spectral and temporal resolutions in the human
auditory cortex. PLoS Computational Biology 10(1), e1003412.
Satoh, M., Takeda, K., Nagata, K., Hatazawa, J., & Kuzuhara, S. (2001). Activated brain regions in
musicians during an ensemble: A PET study. Cognitive Brain Research 12(1), 101–108.
Saur, D., Kreher, B. W., Schnell, S., Kummerer, D., Kellmeyer, P., Vry, M.-S., … Weiller, C. (2008).
Ventral and dorsal pathways for language. Proceedings of the National Academy of Sciences
105(46), 18035–18040.
Schindler, A., Herdener, M., & Bartels, A. (2013). Coding of melodic gestalt in human auditory
cortex. Cerebral Cortex 23(12), 2987–2993.
Schmahmann, J. D., & Pandya, D. N. (1997). The cerebrocerebellar system. International Review of
Neurobiology 41, 31–38, 38a, 39–60.
Schneider, D. M., & Mooney, R. (2015). Motor-related signals in the auditory system for listening
and learning. Current Opinion in Neurobiology 33, 78–84.
Schneider, D. M., Nelson, A., & Mooney, R. (2014). A synaptic and circuit basis for corollary
discharge in the auditory cortex. Nature 513(7517), 189–194.
Schön, D., Gordon, R. L., & Besson, M. (2005). Musical and linguistic processing in song
perception. Annals of the New York Academy of Sciences 1060(1), 71–81.
Schonwiesner, M., & Zatorre, R. J. (2009). Spectro-temporal modulation transfer function of single
voxels in the human auditory cortex measured with high-resolution fMRI. Proceedings of the
National Academy of Sciences 106(34), 14611–14616.
Schubotz, R. I. (2007). Prediction of external events with our motor system: Towards a new
framework. Trends in Cognitive Sciences 11(5), 211–218.
Schulze, K., Zysset, S., Mueller, K., Friederici, A. D., & Koelsch, S. (2011). Neuroarchitecture of
verbal and tonal working memory in nonmusicians and musicians. Human Brain Mapping 32(5),
771–783.
Schwartze, M., Keller, P. E., & Kotz, S. A. (2016). Spontaneous, synchronized, and corrective timing
behavior in cerebellar lesion patients. Behavioural Brain Research 312, 285–293.
Schwartze, M., Rothermich, K., Schmidt-Kassow, M., & Kotz, S. A. (2011). Temporal regularity
effects on pre-attentive and attentive processing of deviance. Biological Psychology 87(1), 146–
151.
Seger, C. A., Spiering, B. J., Sares, A. G., Quraini, S. I., Alpeter, C., David, J., & Thaut, M. H.
(2013). Corticostriatal contributions to musical expectancy perception. Journal of Cognitive
Neuroscience 25(7), 1062–1077.
Shadmehr, R., Smith, M. A., & Krakauer, J. W. (2010). Error correction, sensory prediction, and
adaptation in motor control. Annual Review of Neuroscience 33, 89–108.
Sokolov, A. A., Miall, R. C., & Ivry, R. B. (2017). The cerebellum: Adaptive prediction for
movement and cognition. Trends in Cognitive Sciences 21(5), 313–332.
Spencer, R. M. C., Ivry, R. B., & Zelaznik, H. N. (2005). Role of the cerebellum in movements:
Control of timing or movement transitions? Experimental Brain Research 161(3), 383–396.
Stewart, L., Overath, T., Warren, J. D., Foxton, J. M., & Griffiths, T. D. (2008). fMRI evidence for a
cortical hierarchy of pitch pattern processing. PLoS ONE 3(1), e1470.
Stoodley, C. J., & Schmahmann, J. D. (2009). Functional topography in the human cerebellum: A
meta-analysis of neuroimaging studies. NeuroImage 44(2), 489–501.
Stoodley, C. J., & Schmahmann, J. D. (2010). Evidence for topographic organization in the
cerebellum of motor control versus cognitive and affective processing. Cortex 46(7), 831–844.
Stupacher, J., Hove, M. J., Novembre, G., Schütz-Bosbach, S., & Keller, P. E. (2013). Musical
groove modulates motor cortex excitability: A TMS investigation. Brain and Cognition 82(2),
127–136.
Suga, N., & Ma, X. (2003). Multiparametric corticofugal modulation and plasticity in the auditory
system. Nature Reviews Neuroscience 4(10), 783–794.
Teki, S., Grube, M., & Griffiths, T. D. (2012). A unified model of time perception accounts for
duration-based and beat-based timing mechanisms. Frontiers in Integrative Neuroscience 5, 90.
Teki, S., Grube, M., Kumar, S., & Griffiths, T. D. (2011). Distinct neural substrates of duration-based
and beat-based auditory timing. Journal of Neuroscience 31(10), 3805–3812.
Tervaniemi, M., Medvedev, S. V., Alho, K., Pakhomov, S. V., Roudas, M. S., Van Zuijen, T. L., &
Näätänen, R. (2000). Lateralized automatic auditory processing of phonetic versus musical
information: A PET study. Human Brain Mapping 10(2), 74–79.
Tesche, C. D., & Karhu, J. J. T. (2000). Anticipatory cerebellar responses during somatosensory
omission in man. Human Brain Mapping 9(3), 119–142.
Thaut, M. H., Demartin, M., & Sanes, J. N. (2008). Brain networks for integrative rhythm formation.
PLoS ONE 3(5), e2312.
Thaut, M. H., McIntosh, G. C., Prassas, S. G., & Rice, R. R. (1992). Effect of rhythmic auditory
cuing on temporal stride parameters and EMG patterns in normal gait. Neurorehabilitation and
Neural Repair 6(4), 185–190.
Thaut, M. H., Stephan, K. M., Wunderlich, G., Schicks, W., Tellmann, L., Herzog, H., … Hömberg,
V. (2009). Distinct cortico-cerebellar activations in rhythmic auditory motor synchronization.
Cortex 45(1), 44–53.
Thaut, M. H., Trimarchi, P., & Parsons, L. (2014). Human brain basis of musical rhythm perception:
Common and distinct neural substrates for meter, tempo, and pattern. Brain Sciences 4(2), 428–
452.
Tillmann, B., Janata, P., & Bharucha, J. J. (2003). Activation of the inferior frontal cortex in musical
priming. Annals of the New York Academy of Sciences 999, 209–211.
Toiviainen, P., Alluri, V., Brattico, E., Wallentin, M., & Vuust, P. (2014). Capturing the musical brain
with Lasso: Dynamic decoding of musical features from fMRI data. NeuroImage 88, 170–180.
Tollin, D. J. (2003). The lateral superior olive: A functional role in sound source localization.
Neuroscientist 9(2), 127–143.
Tramo, M. J., Shah, G. D., & Braida, L. D. (2002). Functional role of auditory cortex in frequency
processing and pitch perception. Journal of Neurophysiology 87(1), 122–139.
Trost, W., Ethofer, T., Zentner, M., & Vuilleumier, P. (2012). Mapping aesthetic musical emotions in
the brain. Cerebral Cortex 22(12), 2769–2783.
Tseng, Y., Diedrichsen, J., Krakauer, J. W., Shadmehr, R., & Bastian, A. J. (2007). Sensory prediction
errors drive cerebellum-dependent adaptation of reaching. Journal of Neurophysiology 98(1), 54–
62.
Von Der Heide, R. J., Skipper, L. M., Klobusicky, E., & Olson, I. R. (2013). Dissecting the uncinate
fasciculus: Disorders, controversies and a hypothesis. Brain 136(6), 1692–1707.
Warren, J. D., Jennings, A. R., & Griffiths, T. D. (2005). Analysis of the spectral envelope of sounds
by the human brain. NeuroImage 24(4), 1052–1057.
Warren, J. D., Uppenkamp, S., Patterson, R. D., & Griffiths, T. D. (2003). Separating pitch chroma
and pitch height in the human brain. Proceedings of the National Academy of Sciences 100(17),
10038–10042.
Warren, J. E., Wise, R. J. S., & Warren, J. D. (2005). Sounds do-able: Auditory-motor
transformations and the posterior temporal plane. Trends in Neurosciences 28(12), 636–643.
Warrier, C., Wong, P., Penhune, V., Zatorre, R., Parrish, T., Abrams, D., & Kraus, N. (2009). Relating
structure to function: Heschl’s gyrus and acoustic processing. Journal of Neuroscience 29(1), 61–
69.
Wessinger, C. M., VanMeter, J., Tian, B., Van Lare, J., Pekar, J., & Rauschecker, J. P. (2001).
Hierarchical organization of the human auditory cortex revealed by functional magnetic resonance
imaging. Journal of Cognitive Neuroscience 13(1), 1–7.
Wilkins, R. W., Hodges, D. A., Laurienti, P. J., Steen, M., & Burdette, J. H. (2014). Network science
and the effects of music preference on functional brain connectivity: From Beethoven to Eminem.
Scientific Reports 4(1), 6130.
Wilson, E. M. F., & Davey, N. J. (2002). Musical beat influences corticospinal drive to ankle flexor
and extensor muscles in man. International Journal of Psychophysiology 44(2), 177–184.
Witt, S. T., Laird, A. R., & Meyerand, M. E. (2008). Functional neuroimaging correlates of finger-
tapping task variations: An ALE meta-analysis. NeuroImage 42(1), 343–356.
Wolpert, D. M., Miall, R. C., & Kawato, M. (1998). Internal models in the cerebellum. Trends in
Cognitive Sciences 2(9), 338–347.
Wu, J., Zhang, J., Ding, X., Li, R., & Zhou, C. (2013). The effects of music on brain functional
networks: A network analysis. Neuroscience 250, 49–59.
Wu, J., Zhang, J., Liu, C., Liu, D., Ding, X., & Zhou, C. (2012). Graph theoretical analysis of EEG
functional connectivity during music perception. Brain Research 1483, 71–81.
Zatorre, R. J. (2002). Auditory cortex. In V. S. Ramachandran (Ed.), Encyclopedia of the Human
Brain (pp. 289–301). Amsterdam: Elsevier.
Zatorre, R. J. (2015). Musical pleasure and reward: Mechanisms and dysfunction. Annals of the New
York Academy of Sciences 1337(1), 202–211.
Zatorre, R. J., Bouffard, M., & Belin, P. (2004). Sensitivity to auditory object features in human
temporal neocortex. Journal of Neuroscience 24(14), 3637–3642.
Zatorre, R. J., Halpern, A. R., Perry, D. W., Meyer, E., & Evans, A. C. (1996). Hearing in the mind’s
ear: A PET investigation of musical imagery and perception. Journal of Cognitive Neuroscience
8(1), 29–46.
Zatorre, R. J., & Salimpoor, V. N. (2013). From perception to pleasure: Music and its neural
substrates. Proceedings of the National Academy of Sciences 110(Suppl. 2), 10430–10437.
Zatorre, R. J., & Zarate, J. (2012). Cortical processing of music. In D. Poeppel, T. Overath, A. N.
Popper, & R. R. Fay (Eds.), The human auditory cortex: Springer handbook of auditory research
(Vol. 43, pp. 261–294). New York: Springer.
Zysset, S., Huber, O., Ferstl, E., & von Cramon, D. Y. (2002). The anterior frontomedian cortex and
evaluative judgment: An fMRI study. NeuroImage 15(4), 983–991.
CHAPT E R 6
NETWORK
NEUROSCIENCE: AN
I N T R O D U C T I O N TO G R A P H
T H E O RY N E T W O R K - B A S E D
TECHNIQUES FOR MUSIC
AND BRAIN IMAGING
RESEARCH
R O B I N W. WI L K I N S
FIGURE 2. Demonstration of the Network Statistic Degree. This figure depicts the degree of a
network node. In this network, Node 9 has edge connections to four other nodes in the network.
Thus, Node 9 has a degree of four. Note that Node 9 also connects to Nodes 6, 7, 8, and 11 within
the network but does not connect to Node 10 or Node 12.
FIGURE 3. Depiction of three networks: regular, small-world, and random. This figure
demonstrates differences in connections within three networks that have the same number of nodes.
The regular network has connections with all neighboring nodes but no long-range connections. The
random network has haphazard connections throughout the network. In contrast, the small-world
network has primarily nearest neighbor connections but also some long-range connections across the
network. This is referred to as the “small-world” effect. Small-world networks have been revealed to
be a property of the brain.
As shown in Fig. 3, in the regular network we can see that each node is
connected to each and every other neighboring node, but does not have
long-range connections to nodes across the network. The regular network is
considered completely connected. However, in a random network, node
connections are arbitrary. Thus, in contrast to both the regular and the
random network, the small-world network depicted in the center of Fig. 3
shows that most nodes connect to neighboring nodes. However, this
network has a few nodes with long-range connections to other network
nodes. Thus, while the regular network has a lot of node-to-nearest-
neighbor connections, the small-world network also has a few distinct long-
range nodal connections that, in turn, generate close proximity through
direct connectivity (Amaral et al., 2000). These direct connections are
found regardless of node location (i.e., regional proximity). This
phenomenon of a “small-world” effect is a widely recognized characteristic
of complex brain networks (Bassett & Bullmore, 2006; Watts & Strogatz,
1998).
In random networks, all node degree connections are possible. In most
complex systems however, high degree nodes tend to connect to other high
degree nodes. In other words, the network does not scale regularly. A scale-
free network is a network where the degree distribution follows a power
law. Thus, in complex systems, rather than high degree nodes exhibiting
random connection to any particular node, high degree nodes tend to self-
select by connecting to other high degree nodes and therefore generate a
non-Gaussian distribution that is scale-free. Intuitively, when considered as
a characteristic framework for understanding the brain, this makes sense.
The brain selectively utilizes its high degree connections as resources in an
efficient fashion in order to coordinate a host of widely distributed system-
level functions. These complex networks are termed “scale-free.” To recap,
nodes in complex systems, such as the brain, generally have a non-Gaussian
degree distribution, often with a long tail toward a high degree. Complex
brain networks exhibit characteristics of small-world networks where nodes
tend to connect to other nodes in disparate regions of the network
(Bullmore & Sporns, 2012). Finally, the degree distributions of nodes in
complex networks are scale-free and follow a power law (Barabási &
Albert, 1999).
If the nearest neighbors of a node are also directly connected to each
other they form a cluster (Watts & Strogatz, 1998). Nodes that tend to
cluster are considered hubs (see Fig. 4). As the term implies, hubs function
as connection “interchanges” within the network. The clustering coefficient
quantifies the number of connections that exist between the nearest
neighbors of a node as a proportion of the maximum number of possible
connections. Random networks have a low average clustering whereas
complex networks typically have high clustering. Those nodes with high
degrees, as hubs, are considered central to the network and can demonstrate
their importance to the overall functioning brain network. This is important
when considering application to the brain. Understanding brain function,
and how it may be structurally or functionally altered or remediated via
musical experiences, has important implications for understanding the
effects of music and musical training as well as treating a variety of
neurological conditions and disorders (El Haj, Fasotti, & Allain, 2012;
Hodges & Wilkins, 2015; Hyde et al., 2009; Schlaug, 2009a; Wilkins, 2015;
Wilkins et al., 2012, 2014; Wilkins et al., 2018; Wong, Skoe, Russo, Dees,
& Kraus, 2007).
FIGURE 4. Demonstration of a hub. Node 7, shown as a darker circle, is central to all the other
nodes in the network and is therefore a hub. Note that Node 7 has a degree of five, but due to its high
centrality, Node 7 is also considered a hub within the entire network.
I M B
R
Although there are still fundamental questions about music and the brain
that remain unresolved, network science offers key tools that hold promise
for providing answers about complex systems in new ways. As the field
continues to advance, network neuroscience and the study of brain
connectivity, through network-based statistics, will expand new
experimental and theoretical avenues for understanding how structural brain
connectivity leads to dynamic brain function. The discussion in this chapter,
in particular, illustrates how network-based approaches may advance
fundamental questions surrounding the promising effects of music in
neurological research and rehabilitation (Hodges & Wilkins, 2015;
Kotchoubey, Pavlov, & Kleber, 2015; Thaut et al., 2008). As a
computationally robust field, network neuroscience provides a new
mathematical framework for investigating complex systems that goes
beyond previously conventional approaches to experimental design and
neuroimaging research. Methods from graph theory provide a robust, well-
established framework for assessing brain connectivity, both locally and
globally, offering a rigorous opportunity to expansively and non-invasively
explore the entire human brain under whole brain activity experiences
(Bullmore & Sporns, 2009; Rubinov & Sporns, 2010). Analyses can reveal
patterns of both structural and functional brain connectivity. A network
neuroscience approach provides unprecedented opportunities for examining
the effects of musical experiences on the human brain. The methods and
techniques presented here provide an opportunity for researchers to pursue
questions that may further advance the field of music and brain research,
deepening our scientific understanding surrounding the effects of music on
the brain.
R
Albert, R., Jeong, H., & Barabási, A.-L. (2000). Error and attack tolerance of complex networks.
Nature 406(6794), 378–382.
Alluri, V., Toivianen, P., Jaaskelainen, J. P., Glerean, E., Sams, M., & Brattico, E. (2012). Large-scale
brain networks emerge from dynamic processing of musical timbre, key and rhythm. NeuroImage
59(4), 3677–3689.
Alluri, V., Toiviainen, P., Lund, T. E., Wallentin, M., Vuust, P., Nandi, A. K., … Brattico, E. (2013).
From Vivaldi to Beatles and back: Predicting lateralized brain responses to music. NeuroImage 83,
627–636.
Amaral, L. A., Scala, A., Barthelemy, M., & Stanley, H. E. (2000). Classes of small-world networks.
Proceedings of the National Academy of Sciences 97(21), 11149–11152.
Barabási, A.-L. (2002). Linked: The new science of networks. Cambridge, MA: Perseus Publishing.
Barabási, A.-L., & Albert, R. (1999). Emergence of scaling in random networks. Science 286(5439),
509–512.
Bassett, D. S., & Bullmore, E. (2006). Small-world brain networks. Neuroscientist 12(6), 512–523.
Bassett, D. S., & Bullmore, E. (2009). Human brain networks in health and disease. Current Opinion
in Neurology 22(4), 340–347.
Bassett, D. S., Khambhati, A. N., & Grafton, S. T. (2017). Emerging frontiers of neuroengineering: A
network science of brain connectivity. Annual Review of Biomedical Engineering 19, 327–352.
Bassett, D. S., & Sporns, O. (2017). Network neuroscience. Nature Neuroscience 20(3), 353–364.
Betzel, R. F., Erickson, M. A., Abell, M., O’Donnell, B. F., Hetrick, W. P., & Sporns, O. (2012).
Synchronization dynamics and evidence for a repertoire of network states in resting EEG.
Frontiers in Computational Neuroscience 6. Retrieved from
https://doi.org/10.3389/fncom.2012.00074
Bigand, E., Tillmann, B., Peretz, I., Zatorre, R. J., Lopez, L., & Majno, M. (Eds.). (2015). The
neurosciences and music V: Cognitive stimulation and rehabilitation. Annals of the New York
Academy of Sciences 1337.
Biswal, B. B., Kylen, J. V., & Hyde, J. S. (1997). Simultaneous assessment of flow and BOLD
signals in resting-state functional connectivity maps. NMR in Biomedicine 10(45), 165–170.
Biswal, B. B., Yetkin, F. Z., Haughton, V. M., & Hyde, J. S. (1995). Functional connectivity in the
motor cortex of resting human brain using echo-planar MRI. Magnetic Resonance in Medicine
34(4), 537–541.
Blondel, V. D., Guillaume, J. L., Lambiotte, R., & Lefebvre, E. (2008). Fast unfolding of
communities in large networks. Journal of Statistical Mechanics 2008. Retrieved from
https://doi.org/10.1088/1742-5468/2008/10/P10008
Blum, K., Simpatico, T., Febo, M., Rodriquez, C., Dushaj, K., Li, M., … Badgaiyan, R. D. (2017).
Hypothesizing music intervention enhances brain functional connectivity involving dopaminergic
recruitment: Common neuro-correlates to abusable drugs. Molecular Neurobiology 54(5), 3753–
3758.
Borgatti, S. (2005). Centrality and network flow. Social Networks 27(1), 55–71.
Broyd, S. J., Demanuele, C., Debener, S., Helps, S. K., James, C. J., & Sonuga-Barke, E. J. S. (2009).
Default-mode brain dysfunction in mental disorders: A systematic review. Neuroscience &
Biobehavioral Reviews 33(3), 279–296.
Buckner, R. L., Andrews-Hanna, J. R., & Schacter, D. L. (2008). The brain’s default mode network:
Anatomy, function, and relevance to disease. Annals of the New York Academy of Sciences 1124,
1–38.
Bullmore, E., & Sporns, O. (2009). Complex brain networks: Graph theoretical analysis of structural
and functional systems. Nature Reviews Neuroscience 10(3), 186–198.
Bullmore, E., & Sporns, O. (2012). The economy of brain network organization. Nature Reviews
Neuroscience 13(5), 336–349.
Butts, C. T. (2009). Revisiting the foundations of network analysis. Science 325(5939), 414–416.
Cohen, A. L., Fair, D. A., Dosenbach, N. U. F., Miezin, F. M., Dierker, D., & Van Essen, D. C.
(2008). Defining functional areas in individual human brains using resting functional connectivity
MRI. NeuroImage 41, 45–57.
Craddock, R. C., James, G. A., Holtzheimer, P. E., Hu, X. P., & Mayberg, H. S. (2012). A whole
brain fMRI atlas generated via spatially constrained spectral clustering. Human Brain Mapping
33(8), 1914–1928.
El Haj, M., Fasotti, L., & Allain, P. (2012). The involuntary nature of music-evoked autobiographical
memories in Alzheimer’s disease. Consciousness and Cognition 21(1), 238–246.
Euler, L. (1736). Solutio problematis ad geometriam situs pertinentis. Commentarii Academiae
Scientiarum Imperialis Petropolitanae 8, 128–140. Reprinted and translated in N. L. Biggs, E. K.
Lloyd, & R. J. Wilson, Graph Theory 1736–1936 (pp. 3–8). Oxford: Oxford University Press,
1976.
Fauvel, B., Groussard, M., Chetelat, G., Fouquet, M., Landeau, B., Eustache, F., … Platel, H. (2014).
Morphological brain plasticity induced by musical expertise is accompanied by modulation of
functional connectivity at rest. NeuroImage 90, 179–188.
Fortunato, S. (2010). Community detection in graphs. Physics Reports 486(3–5), 75–174.
Fox, M. D., Snyder, A. Z., Vincent, J. L., Corbetta, M., Van Essen, D. C., & Raichle, M. E. (2005).
The human brain is intrinsically organized into dynamic, anticorrelated functional networks.
Proceedings of the National Academy of Sciences 102(27), 9673–9678.
Fox, M. D., Zhang, D., Snyder, A. Z., & Raichle, M. E. (2009). The global signal and observed
anticorrelated resting state brain networks. Journal of Neurophysiology 101(6), 3270–3283.
Friston, K. J., Frith, C. D., Turner, R., & Frackowiak, R. S. (1995). Characterizing evoked
hemodynamics with fMRI. NeuroImage 2(2), 157–165.
Gaser, C., and Schlaug, G. (2003). Brain structures differ between musicians and non-musicians.
Journal of Neuroscience 23(27), 9240–9245.
Girvan, M., & Newman, M. E. (2002). Community structure in social and biological networks.
Proceedings of the National Academy of Sciences 99(12), 7821–7826.
Greicius, M. (2008). Resting-state functional connectivity in neuropsychiatric disorders. Current
Opinion in Neurology 21(4), 424–430.
Greicius, M., Krasnow, B., Reiss, A. L., & Menon, V. (2003). Functional connectivity in the resting
brain: A network analyses of the default mode hypothesis. Proceedings of the National Academy of
Sciences 100(1), 253–258.
Gusnard, D. A., Akbudak, E., Shulman, G. L., & Raichle, M. E. (2001a). Medial prefrontal cortex
and self-referential mental activity: Relation to a default mode of brain function. Proceedings of
the National Academy of Sciences 98(7), 4259–4264.
Gusnard, D. A., Akbudak, E., Shulman, G. L., & Raichle, M. E. (2001b). Role of medial prefrontal
cortex in a default mode of brain function. NeuroImage 13(6), S414.
Guye, M., Bettus, G., Bartolomei, F., & Cozzone, P. (2010). Graph theoretical analysis of structural
and functional connectivity MRI in normal and pathological brain networks. Magnetic Resonance
Materials in Physics, Biology and Medicine 23(5–6), 409–421.
Hagberg, A. A., Schult, D. A., & Swart, P. J. (2008). Exploring network structure, dynamics, and
function using NetworkX. In G. Varoquaux, T. Vaught, & J. Millman (Eds.), Proceedings of the
7th Python in Science Conference (SciPy2008) (pp. 11–15). Pasadena, CA.
Hayasaka, S., & Laurienti, P. J. (2010). Comparison of characteristics between region- and voxel-
based network analyses in resting-state fMRI data. NeuroImage 50(2), 499–508.
He, Y., & Evans, A. (2010). A review of structural and functional brain connectivity. Current
Opinion in Neurology 23(4), 341–350.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Hodges, D. A., & Wilkins, R. W. (2015). How and why does music move us? Answers from
psychology and neuroscience. Music Education Journal 101(4), 41–47.
Hyde, K. L., Lerch, J., Norton, A., Forgeard, M., Winner, E., Evans, A. C., & Schlaug, G. (2009).
Musical training shapes structural brain development. Journal of Neuroscience 29(10), 3019–3025.
Joyce, K. E., Laurienti, P. J., Burdette, J. H., & Hayasaka, S. (2010). A new measure of centrality for
brain networks. PLoS ONE 5(8), e12200.
Karmonik, C., Brandt, A., Anderson, J. R., Brooks, F., Lytle, J., Silverman, E., & Frazier, J. T (2016).
Music listening modulates functional connectivity and information flow in the human brain. Brain
Connectivity 6(8), 632–641.
Koelsch, S. (2009). A neuroscientific perspective on music therapy. Annals of the New York Academy
of Sciences 1169, 374–384.
Koelsch, S., Skouras, S., & Lohmann, G. (2018). The auditory cortex hosts network nodes influential
for emotion processing: An fMRI study on music-evoked fear and joy. PLoS ONE 13(1),
e0190057.
Kotchoubey, B., Pavlov, Y. G., & Kleber, B. (2015). Music in research and rehabilitation of disorders
of consciousness: Psychological and neurophysiological foundations. Frontiers in Psychology 6,
1763. Retrieved from https://doi.org/10.3389/fpsyg.2015.01763
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11(8), 599–605.
Kruschwitz, J. D., List, D., Waller, L., Rubinov, M. & Walter, H. (2015). GraphVar: A user-friendly
toolbox for comprehensive graph analyses of functional brain connectivity. Journal of
Neuroscience Methods 245, 107–115.
Lindquist, K. A., Pendl, S., Brooks, J. A., Wilkins, R. W., Kraft, R. A., & Gao, W. (2018). Dynamic
functional connectivity of intrinsic networks during emotions. NeuroImage. Under review.
Liu, C., Abu-Jamous, B., Brattico, E., & Nandi, A. K. (2017). Towards tunable consensus clustering
for studying functional brain connectivity during affective processing. International Journal of
Neural Systems 27(2), 1650042. doi:10.1142/S0129065716500428
Liu, C., Brattico, E., Abu-Jamous, B., Pereira, C. S., Jacobsen, T., & Nandi, A. K. (2017). Effect of
explicit evaluation on neural connectivity related to listening to unfamiliar music. Frontiers in
Human Neuroscience 11, 611. Retrieved from https://doi.org/10.3389/fnhum.2017.00611
Logothetis, N. K. (2008). What we can do and what we cannot do with fMRI. Nature 453(7197),
869–878.
Magee, W. L., Clark, I., Tamplin, J., & Bradt, J. (2017). Music interventions for acquired brain injury.
Cochrane Database of Systematic Reviews 1, CD006787. doi:10.1002/14651858.CD006787.pub3
Mitchell, M. (2009). Complexity: A guided tour. Oxford: Oxford University Press.
Morcom, A. M., & Fletcher, P. C. (2007). Does the brain have a baseline? Why we should be
resisting a rest. NeuroImage 37(4), 1073–1082.
Moussa, M. N., Vechlekar, C. D., Burdett, J. H., Steen, M. R., Hugenschmidt, C. E., & Laurienti, P. J.
(2011). Changes in cognitive state alter human functional brain networks. Frontiers in Human
Neuroscience 5, 1–15. Retrieved from https://doi.org/10.3389/fnhum.2011.00083
Mucha, P. J., Richardson, T., Macon, K., Porter, M. A., & Onnela, J. P. (2010). Community structure
in time-dependent, multiscale, and multiplex networks. Science 328(5980), 876–878.
Mumford, J. A., Horvath, S., Oldham, M. C., Langfelder, P., Geschwind, D. H., & Poldrack, R. A.
(2010). Detecting network modules in fMRI time series: A weighted network analysis approach.
NeuroImage 52(4), 1465–1476.
Newman, M. E. J. (2003). The structure and function of complex networks. SIAM Review 45, 167–
256.
Newman, M. E. J. (2005). Power laws, Pareto distributions and Zipf’s law. Contemporary Physics
46(5), 323–351.
Newman, M. E. J. (2006). Modularity and community structure in networks. Proceedings of the
National Academy of Sciences 103(23), 8577–8582.
Newman, M. E., & Girvan, M. (2004). Finding and evaluating community structure in networks.
Physical Review E 69(2 Pt. 2), 026113.
Power, J. D., Cohen, A. L, Nelson, S. M., Wig, G. S., Barnes, K. A., Church, J. A., … Petersen, S. E.
(2011). Functional network organization of the human brain. Neuron 72(4), 665–678.
Raglio, A., Attardo, L., Gontero, G., Rollino, S., Groppo, E., & Granieri, E. (2015). Effects of music
and music therapy on mood in neurological patients World Journal of Psychiatry 5(1), 68–78.
Raichle, M. E. (2001). A default mode of brain function. Proceedings of the National Academy of
Sciences 98(2), 676–682.
Rubinov, M., & Sporns, O. (2010). Complex network measures of brain connectivity: Uses and
interpretations. NeuroImage 52(3), 1059–1069.
Sachs, M. E., Ellis, R. J., Schlaug, G., & Loui, P. (2016). Brain connectivity reflects aesthetic
responses to music. Social Cognitive and Affective Neuroscience 11(6), 884–891.
Savoy, R. A. (2005). Experimental design in brain activation MRI: Cautionary tales. Brain Research
Bulletin 67, 361–365.
Schlaug, G. (2001). The brain of musicians. A model for functional and structural adaptation. Annals
of the New York Academy of Sciences 930, 281–299.
Schlaug, G. (2009a). Listening to music facilitates brain recovery processes. Annals of the New York
Academy of Sciences 1169, 372–373.
Schlaug, G. (2009b). Music, musicians, and brain plasticity. In S. Hallam, I. Cross, & M. Thaut
(Eds.), The Oxford handbook of music psychology (pp. 197–207). Oxford: Oxford University
Press.
Schlaug, G., Marchina, S., & Norton, A. (2009). Evidence for plasticity in white matter tracts of
chronic aphasic patients undergoing intense intonation-based speech therapy. Annals of the New
York Academy of Sciences 1169, 385–394.
Shirer, W. R., Ryali, S., Rykhlevskaia, E., Menon, V., & Greicius, M. D. (2012). Decoding subject-
driven cognitive states with whole-brain connectivity patterns. Cerebral Cortex 22(1), 158–165.
Sihvonen, A. J., Sarkamo, T., Leo, V., Tervaniemi, M., Altenmuller, E., & Soinila, S. (2017). Music-
based interventions in neurological rehabilitation. The Lancet Neurology 16(8), 648–660.
Sporns, O., Chialvo, D., Kaiser, M., & Hilgetag, C. (2004). Organization, development and function
of complex brain networks. Trends in Cognitive Sciences 8(9), 418–425.
Sporns, O., & Kotter, R. (2004). Motifs in brain networks. PLoS Biology 2(11), e369.
Sporns, O., Tononi, G., & Kötter, R. (2005). The human connectome: A structural description of the
human brain. PLoS Computational Biology 1(4), e42.
Stam, C. J. (2014). Modern network science of neurological disorders. Nature Reviews Neuroscience
15, 683–695.
Stam, C. J., & Reijneveld, J. P. (2007). Graph theoretical analysis of complex networks in the brain.
Nonlinear Biomedical Physics 1, 3. doi:10.1186/1753-4631-1-3
Stanley, M. L., Moussa, M. N., Paolini, B. M., Lyday, R., Burdette, J. H., & Laurienti, P. J. (2013).
Defining nodes in complex networks. Frontiers in Computational Neuroscience 7, 169. Retrieved
from https://doi.org/10.3389/fncom.2013.00169
Steen, M., Hayasaka, S., Joyce, K., & Laurienti, P. (2011). Assessing the consistency of community
structure in complex networks. Physical Review E 84(1–2), 016111.
Strogatz, S. H. (2001). Exploring complex networks. Nature 410(6825), 268–276.
Telesford, Q. K., Simpson, S. L., Burdette, J. H., Hayasaka, S., & Laurienti, P. J. (2011). The brain as
a complex system: Using network science as a tool for understanding the brain. Brain Connectivity
1(4), 295–308.
Thaut, M. H., Demartin, M., & Sanes, J. N. (2008). Brain networks for integrative rhythm formation.
PLoS ONE 3, e2312.
Thaut, M. H., Gardiner, J. C., Holmberg, D., Horwitz, J., Kent, L., Andrews, G., … McIntosh, G. R.
(2009). Neurologic music therapy improves executive function and emotional adjustment in
traumatic brain injury rehabilitation. Annals of the New York Academy of Sciences 1169, 406–416.
Towlson, E., Vertes, P. E., Ahnert, S., Schafer, W. R., & Bullmore, E. T. (2013). The rich club of the
C. elegans neuronal connectome. Journal of Neuroscience 33(15), 6380–6387.
Tuch, D. S., Reese, T. G., Wiegell, M. R., & Wedeen, V. J. (2003). Diffusion MRI of complex neural
architecture. Neuron 40(5), 885–895.
Van den Heuvel, M. P., de Lange, S. C., Zalesky, A., Seguin, C., Yeo, B. T. T., & Schmidt, R. (2017).
Proportional thresholding in resting-state fMRI functional connectivity networks and
consequences for patient-control connectome studies: Issues and recommendations. NeuroImage
152, 437–449.
Van Wijk, B. C., Stam, C. J., & Daffertshofer, A. (2010). Comparing brain networks of different size
and connectivity density using graph theory. PloS ONE 5, e13701.
Wang, J., Zuo, X., & He, Y. (2010). Graph-based network analysis of resting-state functional MRI.
Frontiers in Systems Neuroscience 4, 16. Retrieved from https://doi.org/10.3389/fnsys.2010.00016
Wang, M., González, C. A., Hidalgo, & Barabási, A.-L. (2009). Understanding the spreading patterns
of mobile phone viruses. Science 324(5930), 1071–1076.
Watts, D. J. (2003). Six degrees: The science of a connected age. New York: W. W. Norton.
Watts, D. J., & Strogatz, S. H. (1998). Collective dynamics of “small-world” networks. Nature
393(6684), 440–442.
Wedeen, V. J., Hagmann, P., Tseng, W. Y., Reese, T. G., & Weiskoff, R. M. (2005). Mapping complex
tissue architecture with diffusion spectrum magnetic resonance imaging. Magnetic Resonance in
Medicine 54(6), 1377–1386.
West, B. J. (2011). Overview 2010 of ARL program on network science for human decision making.
Frontiers in Physiology 2, 76. Retrieved from https://doi.org/10.3389/fphys.2011.00076
Whitfield-Gabrielli, S., & Nieto-Castanon, A. (2012). Conn: A functional connectivity toolbox for
correlated and anticorrelated brain networks. Brain Connectivity 2(3).
doi:10.1089/brain.2012.0073
Wilkins, R. W. (2015). Network science and the effects of music on the human brain (Doctoral
dissertation). University of North Carolina at Greensboro.
Wilkins, R. W., Giridharan, S., Johnston, M., Brooks, J. A., Lindquist, K. A., & Kraft, R. A. (2018).
Changes in resting-state functional brain networks during naturalistic music listening. In
preparation.
Wilkins, R. W., Hodges, D. A., Laurienti, M., Steen, M., & Burdette, J. H. (2012). Network science:
A new method for investigating the complexity of musical experiences in the brain. Leonardo
45(3), 282–283.
Wilkins, R. W., Hodges, D. A., Laurienti, M., Steen, M., & Burdette, J. H. (2014). Network science
and the effects of music preference on functional brain connectivity: From Beethoven to Eminem.
Scientific Reports 4, 6130. doi: 10.1038/srep06130
Wong, P. C., Skoe, E., Russo, N. M., Dees, T., & Kraus, N. (2007). Musical experience shapes human
brainstem encoding of linguistic pitch patterns. Nature Neuroscience 10(4), 420–422.
Wu, J., Zhang, J., Ding, X., Liu, D., & Zhou, C. (2013). The effects of music on brain functional
networks: A network analyses. Neuroscience 250, 49–59.
Wu, J., Zhang, J., Liu, C., Liu, D., Ding, X., & Zhou, C. (2012). Graph theoretical analysis of EEG
functional connectivity during music perception. Brain Research 1483, 71–81.
Zatorre, R., Evans, A., Meyer, E., & Gjedde, A. (1992). Lateralization of phonetic and pitch
discrimination in speech processing. Science 256(5058), 846–849.
Zatorre, R., & Samson, S. (1991). Role of the right temporal neocortex in retention of pitch in
auditory short-term memory. Brain 114(6), 2403–2417.
Zuo, X., Ehmke, R., Mennes, M., Imperati, D., Castellanos, X., Sporns, O., & Milham, M. (2011).
Network centrality in the human functional connectome. Cerebral Cortex 22(8), 1862–1875.
CHAPT E R 7
ACOUSTIC STRUCTURE
AND MUSICAL FUNCTION:
MUSICAL NOTES
I N F O R M I N G A U D I TO RY
RESEARCH
MI C H A E L S C H U T Z
I O
G N : D
C H
G H : D
I N
The complexities in composers’ grouping of individual notes into chords
are well known (Aldwell et al., 2002), yet the musical importance of
individual harmonics is less transparent, even though single notes produced
by musical instruments contain incredible sophistication and nuance
(Hjortkjaer, 2013). Musical instruments produce sounds rich in overtones,
which for pitched instruments generally consist of harmonics at integer
multiples of the fundamental (Dowling & Harwood, 1986; Tan et al., 2010),
as well as other non-harmonic energy (particularly during a sound’s onset).
The lawful structure of these overtones serves as an important binding cue,
triggering a decision by the perceptual system to blend overtones such that
“the listener is not usually directly aware of the separate harmonics”
(Dowling & Harwood, 1986, p. 24). Although some musicians develop the
ability to “hear out” individual components of their instruments (Jourdain,
1997, p. 35), in general this collection of frequencies fuses into a single
musical unit. Consequently for practical matters the complex structure of
individual notes is of less musical interest than the composer’s complex
selection of structural cues (Broze & Huron, 2013; Huron & Ollen, 2003;
Patel & Daniele, 2003; Poon & Schutz, 2015), or the performer’s
interpretation of those cues (Chapin, Jantzen, Kelso, Steinberg, & Large,
2010).
Although the musical importance of small note-to-note variations in
amplitude with respect to phrasing and expressivity (Bhatara, Tirovolas,
Duan, Levy, & Levitin, 2011; Repp, 1995) is widely recognized, the small
moment-to-moment amplitude variations in individual overtones have
received less research attention. Musical sounds contain overtones shifting
in their relative strength over time (Jourdain, 1997, p. 35), and some
textbooks explicitly note the importance of these dynamic changes
(Thompson, 2009, p. 59). Yet the role of spectra is often presented as time-
invariant and described through summaries of spectral content irrespective
of temporal changes in a note’s spectra.
Musical instruments produce notes rich in temporal variation—not only
in their overall amplitudes, but even with respect to the envelopes of
individual harmonics. For example, Fig. 1 visualizes a musical note
performed on the trumpet (left panel) and clarinet (right panel), based on
instrument sounds provided by the University of Iowa Electronic Music
studios (Fritts, 1997). The intensity (z axis) of energy extracted from each
harmonic (x axis) is graphed over time (y axis). These 3D visualizations
illustrate the temporal complexity of harmonics bound into the percept of a
single note. In fact, divorced from its context in a melody, expressive
timings in musical passages, discussion of performer’s intentions regarding
phrasing and numerous other considerations, analysis of isolated notes
affords invaluable insight. Small temporal variations in each overtone play
a key role in the degree to which synthesized notes sound “real” rather than
“artificial.” Highly trained musicians can routinely produce different
variations on a single note (“brighter” or “more legato,” “shimmery,” etc.),
which involve intentionally varying both the balance and temporal changes
in a note’s overtones.
FIGURE 1. Visualization of single notes produced by a trumpet (left) and clarinet (right),
illustrating their complex temporal structure. Although the trumpet spectrum changes more
dynamically than the clarinet, each partial is in constant flux.
The goal of these 3D figures is to illustrate the dynamic nature of the harmonic structure of
musical tones. Consequently they are not complete acoustical analyses (which are readily
available elsewhere), but serve to highlight information lost in temporally invariant power
spectra.
A S M
T
FIGURE 2. Power spectra of trumpet and clarinet. These plots accurately convey the trumpet’s
energy at many harmonics in contrast to the clarinet’s energy primarily at odd numbered harmonics.
However, power spectra fail to convey any information about the temporal changes in harmonic
amplitude so crucial to a sound’s timbre.
T U T V
S M P R
The most surprising outcome from this survey was that although most
articles included a wealth of technical information on spectral structure,
duration, and the exact model of headphones or speakers used to present the
stimuli, about 35 percent failed to define the stimuli’s temporal structure.
This finding is not unique to Music Perception—my team found similar
problems with under-specification in the journal Attention, Perception &
Psychophysics (Gillard & Schutz, 2013). More important than under-
specification, both surveys revealed a strong bias against sounds with the
kinds of temporal variations common to musical instruments. Although flat
tones lend themselves well to tight experimental control and consistent
replication amongst different labs, they fail to capture the richness of the
sounds forming the backbone of the musical listening experience. Yet they
remain prominent in a wide range of research on auditory perception on
tasks purportedly designed to illuminate generalizable principles of auditory
perception.
Prominent researchers have noted that the world is “[not] replete with
examples of naturally occurring auditory pedestals [i.e., flat amplitude
envelopes]” (Phillips, Hall, & Boehnke, 2002, p. 199). Yet flat tones appear
to be the normative approach to research on auditory perception, which are
clearly far removed from the complexity of natural musical sounds—as
shown in Fig. 5. Note that each of the three musical instruments visualized
not only exhibits constant temporal changes, but temporal changes in the
amplitudes of each individual harmonic. This dynamic fluctuation contrasts
starkly with the flat tones favored in auditory perception research shown in
the bottom right panel. This over-fixation on sounds lacking meaningful
amplitude variation is not confined to behavioral work; a large-scale review
of auditory neuroscience research concluded with a note of caution that
important properties of functions of the auditory system will only be fully
understood when researchers begin employing envelopes that “involve
modulation in ways that are closer to real-world tasks faced by the auditory
system” (Joris, Schreiner, & Rees, 2004, p. 570). The acoustic distance
between the temporally dynamic musical sounds and temporally
constrained flat tones common in auditory perception and neuroscience
research raises important questions about the degree to which theories and
models derived from these experiments generalize to musical listening. The
complexities of balancing competing needs for experimental control and
ecological relevance are significant, and will serve as the focus of the
following section.
FIGURE 5. Single notes produced by an oboe (upper left), French horn (upper right), and viola
(lower left) illustrate their temporal complexity. Although their specific mix of harmonics varies,
these instruments all exhibit constant changes in the strength of each harmonic over the tone’s
duration. This temporal complexity contrasts strongly with the temporal simplicity of the flat tone
depicted in the lower right panel, which lacks temporal variation beyond abrupt onsets/offsets, and
no change in relative strength of harmonics.
O M C
S S
A
Funding supporting this research was provided by the Natural Sciences and
Engineering Research Council of Canada (NSERC), Social Sciences and
Humanities Research Council of Canada (SSHRC), and the Ontario Early
Researcher Award (ERA). I would like to thank Maxwell Ng for his
assistance in creating the visualizations of the instrument sounds used
throughout this chapter.
R
Abry, C., Cathiard, M. A., Robert-Ribes, J., & Schwartz, J. L. (1994). The coherence of speech in
audio-visual integration. Current Psychology of Cognition 13, 52–59.
Acoustical Society of America Standards Secretariat (1994). Acoustical Terminology ANSI S1.1–
1994 (ASA 111-1994). American National Standard. ANSI/Acoustical Society of America.
Alais, D., & Burr, D. (2004). The ventriloquist effect results from near-optimal bimodal integration.
Current Biology 14(3), 257–262.
Aldwell, E., Schachter, C., & Cadwallader, A. (2002). Harmony & voice leading (3rd ed.). Boston,
MA: Schirmer.
Alexander, P. L., & Broughton, B. (2008). Professional orchestration: The first key. Solo instruments
& instrumentation note, volume 1 (3rd ed.). Petersburg, VA: Alexander Publishing.
Bedford, F. L. (2004). Analysis of a constraint on perception, cognition, and development: One
object, one place, one time. Journal of Experimental Psychology: Human Perception and
Performance 30(5), 907–912.
Bhatara, A., Tirovolas, A. K., Duan, L. M., Levy, B., & Levitin, D. J. (2011). Perception of emotional
expression in musical performance. Journal of Experimental Psychology: Human Perception and
Performance 37(3), 921–934.
Bonath, B., Noesselt, T., Martinez, A., Mishra, J., Schwiecker, K., Heinze, H.-J., & Hillyard, S. A.
(2007). Neural basis of the ventriloquist illusion. Current Biology 17(19), 1697–1703.
Boulez, P. (1987). Timbre and composition—timbre and language. Contemporary Music Review
2(1), 161–171.
Broze, Y., & Huron, D. (2013). Is higher music faster? Pitch–speed relationships in Western
compositions. Music Perception: An Interdisciplinary Journal 31(1) 19–31.
Caclin, A., McAdams, S., Smith, B. K., & Winsberg, S. (2005). Acoustic correlates of timbre space
dimensions: A confirmatory study using synthetic tones. Journal of the Acoustical Society of
America 118(1), 471–482.
Chapin, H., Jantzen, K., Kelso, J. A. S., Steinberg, F., & Large, E. W. (2010). Dynamic emotional and
neural responses to music depend on performance expression and listener experience. PLoS ONE
5, 1–14.
Chuen, L., & Schutz, M. (2016). The unity assumption facilitates cross-modal binding of musical,
non-speech stimuli: The role of spectral and amplitude cues. Attention, Perception, &
Psychophysics 78(5), 1512–1528.
Clough, J., & Conley, J. (1984). Basic harmonic progressions. New York: W. W. Norton.
Corrigall, K. A., & Trainor, L. J. (2010). Musical enculturation in preschool children: Acquisition of
key and harmonic knowledge. Music Perception: An Interdisciplinary Journal 28(2), 195–200.
Corrigall, K. A., & Trainor, L. J. (2014). Enculturation to musical pitch structure in young children:
Evidence from behavioral and electrophysiological methods. Developmental Science 17(1), 142–
158.
Cox, T. J. (2008). Scraping sounds and disgusting noises. Applied Acoustics 69(12), 1195–1204.
Dowling, W. J., & Harwood, D. L. (1986). Music cognition. Orlando, FL: Academic Press.
Eerola, T., Friberg, A., & Bresin, R. (2013). Emotional expression in music: Contribution, linearity,
and additivity of primary musical cues. Frontiers in Psychology 4, 1–12. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00487
Erickson, R. (1975). Sound Structure in Music. Berkeley, CA: University of California Press.
Ernst, M. O., & Banks, M. S. (2002). Humans integrate visual and haptic information in a
statistically optimal fashion. Nature 415(6870), 429–433.
Fendrich, R., & Corballis, P. M. (2001). The temporal cross-capture of audition and vision.
Perception & Psychophysics 63(4), 719–725.
Ferrand, C. T. (2002). Harmonics-to-noise ratio: An index of vocal aging. Journal of Voice 16(4),
480–487.
Fritts, L. (1997). University of Iowa Electronic Music Studios. University of Iowa. Retrieved from
http://theremin.music.uiowa.edu/MIS.html
Gaver, W. (1993). What in the world do we hear? An ecological approach to auditory event
perception. Ecological Psychology 5(1) 1–29.
Gillard, J., & Schutz, M. (2013). The importance of amplitude envelope: Surveying the temporal
structure of sounds in perceptual research. In Proceedings of the Sound and Music Computing
Conference (pp. 62–68). Stockholm, Sweden.
Gordon, J. W. (1987). The perceptual attack time of musical tones. Journal of the Acoustical Society
of America 82(1) 88–105.
Goswami, U. (2011). A temporal sampling framework for developmental dyslexia. Trends in
Cognitive Sciences 15(1) 3–10.
Grassi, M., & Casco, C. (2009). Audiovisual bounce-inducing effect: Attention alone does not
explain why the discs are bouncing. Journal of Experimental Psychology: Human Perception and
Performance 35(1), 235–243.
Grey, J. M. (1977). Multidimensional perceptual scaling of musical timbres. Journal of the
Acoustical Society of America 61(5), 1270–1277.
Grey, J. M., & Gordon, J. W. (1978). Perceptual effects of spectral modifications on musical timbres.
Journal of the Acoustical Society of America 63(5), 1493–1500.
Guerrieri, M. (2012). The first four notes: Beethoven’s Fifth and the human imagination. New York:
Alfred A. Knopf.
Hamberger, C. L. (2012). The evolution of Schoenberg’s Klangfarbenmelodie: The importance of
timbre in modern music. The Pennsylvania State University. Retrieved from
https://etda.libraries.psu.edu/files/final_submissions/8130
Heinlein, C. P. (1928). The affective characters of the major and minor modes in music. Journal of
Comparative Psychology 8, 101–142.
Hevner, K. (1935). The affective character of the major and minor modes in music. American
Journal of Psychology 47(1), 103–118.
Hjortkjaer, J. (2013). The musical brain. In J. O. Lauring (Ed.), An introduction to neuroaesthetics:
The neuroscientific approach to aesthetic experience, artistic creativity, and arts appreciation (pp.
211–244). Copenhagen: Museum Tusculanum Press.
Houtsma, A. J. M., Rossing, T. D., & Wagennars, W. M. (1987). Auditory demonstrations on
compact disc. Journal of the Acoustical Society of America. New York: Acoustical Society of
America/Eindhoven: Institute for Perception Research.
Huron, D., & Ollen, J. (2003). Agogic contrast in French and English themes: Further support for
Patel and Daniele (2003). Music Perception: An Interdisciplinary Journal 21(2), 267–271.
Iverson, P. (1995). Auditory stream segregation by musical timbre: Effects of static and dynamic
acoustic attributes. Journal of Experimental Psychology: Human Perception and Performance 21,
751–763.
Iverson, P., & Krumhansl, C. L. (1993). Isolating the dynamic attributes of musical timbre. Journal of
the Acoustical Society of America 94, 2594–2603.
Joris, P. X., Schreiner, C. E., & Rees, A. (2004). Neural processing of amplitude-modulated sounds.
Physiological Reviews 84, 541–577.
Jourdain, R. (1997). Music, the brain, and ecstasy: How music captures our imagination. New York:
William Morrow and Company.
Kendall, R. A. (1986). The role of acoustic signal partitions in listener categorization of musical
phrases. Music Perception 4(2), 185–213.
Koelsch, S., & Friederici, A. D. (2003). Toward the neural basis of processing structure in music.
Annals of the New York Academy of Sciences 999, 15–28.
Krimphoff, J., McAdams, S., & Winsberg, S. (1994). Caractérisation du timbre des sons complexes.
II. Analyses acoustiques et quantification psychophysique. Journal de Physique IV Colloque 4,
625–628.
Kumar, S., Forster, H. M., Bailey, P., & Griffiths, T. D. (2008). Mapping unpleasantness of sounds to
their auditory representation. Journal of the Acoustical Society of America 124(6), 3810–3817.
Lowis, M. J. (2002). Music as a trigger for peak experiences among a college staff population.
Creativity Research Journal 14(3–4), 351–359.
McAdams, S., Winsberg, S., Donnadieu, S., de Soete, G., & Krimphoff, J. (1995). Perceptual scaling
of synthesized musical timbres: Common dimensions, specificities, and latent subject classes.
Psychological Research 58(3), 177–192.
McDermott, J. (2012). Auditory preferences and aesthetics: Music, voices, and everyday sounds. In
R. J. Dolan & T. Sharot (Eds.), Neuroscience of preference and choice: Cognitive and neural
mechanisms (pp. 227–257). London: Academic Press.
Miller, J. R., & Carterette, E. C. (1975). Perceptual space for musical structures. Journal of the
Acoustical Society of America 58(3), 711–720.
Moore, B. C. J. (1997). An introduction to the psychology of hearing (4th ed.). London: Academic
Press.
Neuhoff, J. G. (2004). Ecological psychoacoustics (J. G. Neuhoff, Ed.). Amsterdam:
Elsevier/Academic Press.
Neuhoff, J. G., & McBeath, M. K. (1996). The Doppler illusion: The influence of dynamic intensity
change on perceived pitch. Journal of Experimental Psychology: Human Perception and
Performance 22(4), 970–985.
Pallesen, K. J., Brattico, E., Bailey, C., Korvenoja, A., Koivisto, J., Gjedde, A., & Carlson, S. (2005).
Emotion processing of major, minor, and dissonant chords: A functional magnetic resonance
imaging study. Annals of the New York Academy of Sciences 1060, 450–453.
Pastore, R. E., Flint, J., Gaston, J. R., & Solomon, M. J. (2008). Auditory event perception: The
source–perception loop for posture in human gait. Perception & Psychophysics 70(1), 13–29.
Patel, A. D., & Daniele, J. R. (2003). Stress-timed vs. syllable-timed music? A comment on Huron
and Ollen (2003). Music Perception: An Interdisciplinary Journal 21(2), 273–276.
Phillips, D. P., Hall, S. E., & Boehnke, S. E. (2002). Central auditory onset responses, and temporal
asymmetries in auditory perception. Hearing Research 167(1–2), 192–205.
Poon, M., & Schutz, M. (2015). Cueing musical emotions: An empirical analysis of 24-piece sets by
Bach and Chopin documents parallels with emotional speech. Frontiers in Psychology 6, 1–13.
Retrieved from https://doi.org/10.3389/fpsyg.2015.01419
Repp, B. H. (1995). Quantitative effects of global tempo on expressive timing in music performance:
Some perceptual evidence. Music Perception: An Interdisciplinary Journal 13(1), 39–57.
Rimsky-Korsakov, N. (1964). Principles of orchestration (M. Steinberg, Ed.). New York: Dover.
Risset, J.-C., & Wessel, D. L. (1999). Exploration of timbre by analysis and synthesis. In D. Deutsch
(Ed.), The Psychology of Music (pp. 113–169). San Diego, CA: Gulf Professional Publishing.
Rossing, T. D., Moore, R. F., & Wheeler, P. A. (2013). The science of sound (3rd ed.). London:
Pearson Education.
Saldanha, E. L., & Corso, J. F. (1964). Timbre cues and the identification of musical instruments.
Journal of the Acoustical Society of America 36(11), 2021–2026.
Schellenberg, E. G. (2002). Asymmetries in the discrimination of musical intervals: Going out-of-
tune is more noticeable than going in-tune musical intervals. Music Perception: An
Interdisciplinary Journal 19(2), 223–248.
Schellenberg, E. G., & Trehub, S. E. (2003). Good pitch memory is widespread. Psychological
Science 14(3), 262–266.
Schenker, H. (1971). Analysis of the first movement. In E. Forbes (Ed.), Beethoven Symphony No. 5
in C minor (pp. 164–182). New York: W. W. Norton.
Schulkind, M. D., Hennis, L. K., & Rubin, D. C. (1999). Music, emotion, and autobiographical
memory: They’re playing your song. Memory & Cognition 27(6), 948–955.
Schutz, M. (2008). Seeing music? What musicians need to know about vision. Empirical Musicology
Review 3(3), 83–108.
Schutz, M. (2009). Crossmodal integration: The search for unity (Dissertation). University of
Virginia.
Schutz, M., & Kubovy, M. (2009a). Causality and cross-modal integration. Journal of Experimental
Psychology: Human Perception and Performance 35(6), 1791–1810.
Schutz, M., & Kubovy, M. (2009b). Deconstructing a musical illusion: Point-light representations
capture salient properties of impact motions. Canadian Acoustics 37(1), 23–28.
Schutz, M., & Lipscomb, S. (2007). Hearing gestures, seeing music: Vision influences perceived tone
duration. Perception 36(6), 888–897.
Schutz, M., & Manning, F. (2012). Looking beyond the score: The musical role of percussionists’
ancillary gestures. Music Theory Online 18, 1–14.
Schutz, M., Stefanucci, J., Baum, S. H., & Roth, A. (2017). Name that percussive tune: Associative
memory and amplitude envelope. Quarterly Journal of Experimental Psychology 70(7), 1323–
1343.
Schutz, M., & Vaisberg, J. M. (2014). Surveying the temporal structure of sounds used in music
perception. Music Perception: An Interdisciplinary Journal 31(3), 288–296.
Sekuler, R., Sekuler, A. B., & Lau, R. (1997). Sound alters visual motion perception. Nature
385(6614), 308.
Skarratt, P. A., Cole, G. G., & Gellatly, A. R. H. (2009). Prioritization of looming and receding
objects: Equal slopes, different intercepts. Attention, Perception, & Psychophysics 71(4), 964–970.
Sloboda, J. (1991). Music structure and emotional response: Some empirical findings. Psychology of
Music 19(2), 110–120.
Strong, W., & Clark, M. (1967). Perturbations of synthetic orchestral wind-instrument tones. Journal
of the Acoustical Society of America 41(2), 277–285.
Suzuki, M., Okamura, N., Kawachi, Y., Tashiro, M., Arao, H., Hoshishiba, T., … Yanai, K. (2008).
Discrete cortical regions associated with the musical beauty of major and minor chords. Cognitive,
Affective, & Behavioral Neuroscience 8(2), 126–131.
Tan, S.-L., Pfordresher, P. Q., & Harré, R. (2007). Psychology of music: From sound to significance.
New York: Psychology Press.
Terhardt, E. (1974). On the perception of periodic sound fluctuations (roughness). Acta Acustica
United with Acustica 30, 201–213.
Tervaniemi, M., Schröger, E., Saher, M., & Näätänen, R. (2000). Effects of spectral complexity and
sound duration on automatic complex-sound pitch processing in humans: A mismatch negativity
study. Neuroscience Letters 290, 66–70.
Thompson, W. F. (2009). Music, thought, and feeling: Understanding the psychology of music. New
York: Oxford University Press.
Tirovolas, A. K., & Levitin, D. J. (2011). Music perception and cognition research from 1983 to
2010: A categorical and bibliometric analysis of empirical articles in Music Perception. Music
Perception: An Interdisciplinary Journal 29(1), 23–36.
Tovey, D. F. (1971). The Fifth Symphony. In E. Forbes (Ed.), Beethoven Symphony No. 5 in C minor
(pp. 143–150). New York: W. W. Norton.
Trehub, S. E., Endman, M. W., & Thorpe, L. A. (1990). Infants’ perception of timbre: Classification
of complex tones by spectral structure. Journal of Experimental Child Psychology 49(2), 300–313.
Vallet, G., Shore, D. I., & Schutz, M. (2014). Exploring the role of amplitude envelope in duration
estimation. Perception 43(7), 616–630.
Vatakis, A., Ghazanfar, A. A., & Spence, C. (2008). Facilitation of multisensory integration by the
“unity effect” reveals that speech is special. Journal of Vision 8(9), 1–11.
Vatakis, A., & Spence, C. (2007). Crossmodal binding: Evaluating the “unity assumption” using
audiovisual speech stimuli. Perception & Psychophysics 69(5), 744–756.
Vuoskoski, J. K., & Eerola, T. (2012). Can sad music really make you sad? Indirect measures of
affective states induced by music and autobiographical memories. Psychology of Aesthetics,
Creativity, and the Arts 6, 1–10.
Walker, J. T., & Scott, K. J. (1981). Auditory-visual conflicts in the perceived duration of lights,
tones and gaps. Journal of Experimental Psychology: Human Perception and Performance 7(6),
1327–1339.
Wang, S., Liu, B., Dong, R., Zhou, Y., Li, J., Qi, B., … Zhang, L. (2012). Music and lexical tone
perception in Chinese adult cochlear implant users. The Laryngoscope 122, 1353–1360.
Warren, R. M. (2013). Auditory perception: A new synthesis. Amsterdam: Elsevier.
Welch, R. B. (1999). Meaning, attention, and the “unity assumption” in the intersensory bias of
spatial and temporal perceptions. In G. Aschersleben, T. Bachmann, & J. Musseler (Eds.),
Cognitive contributions to the perception of spatial and temporal events (pp. 371–387).
Amsterdam: Elsevier.
Welch, R. B., & Warren, D. H. (1980). Immediate perceptual response to intersensory discrepancy.
Psychological Bulletin 88(3), 638–667.
1
All analyses of notes in this chapter are based on additional samples from the University of
Iowa Electronic Music studios (Fritts, 1997).
2
However, presenting notes without transients as part of a melodic sequence (rather than as
isolated tones) may mitigate this confusion (Kendall, 1986).
3
Provided that the acoustic information is of sufficient quality (Alais & Burr, 2004; Ernst &
Banks, 2002).
CHAPT E R 8
NEURAL BASIS OF
RHYTHM PERCEPTION
C H R I S T I N A M. VA N D E N B O S C H D E R N E D E R L A N D E N, J.
E R I C T. TAY L O R, A N D JE S S I C A A . G R A H N
Despite the ease with which humans pick up on the beat and synchronize
their movements to music, it is not trivial to understand how human brains
perceive and process rhythm. Musical rhythms are beat-based, which means
that the pattern of onsets gives rise to the feeling of an underlying pulse or
framework. Perceiving a beat can make it easier to predict and act on
upcoming events in a rhythmic sequence. However, many other naturally
occurring rhythms in our environment do not have a regular pulse or beat,
such as walking, talking, or the car engine turning over. Such rhythms are
called non-beat-based. Different mechanisms have been proposed to
account for the way that humans encode beat- and non-beat-based rhythms.
Absolute timing mechanisms encode exact durations of all intervals in a
sequence, whereas relative timing mechanisms encode when intervals start
and stop in relation to the beat. If there is no regular beat, then absolute
timing is likely necessary to encode the rhythm. However, with a beat,
relative timing may be used. There is evidence for distinct neural networks
associated with absolute and relative timing (Teki, Grube, Kumar, &
Griffiths, 2011), with participants relying on either mechanism depending
on the nature of the rhythm and task demands.
A number of approaches are used to understanding rhythm processing,
incorporating methodologies based on behavior, neuroimaging, and patient
studies. Measuring behavior is fundamental to our understanding of rhythm
because it can provide a direct measure of how we move to music.
However, much of behavioral research is correlational—that is, distinct
measures of stimulus characteristics and tapping variability may be related
to one another, but the stimulus characteristics may not cause tapping
variability as there may be a third unmeasured variable that is the true
influencer of performance. Although neuroimaging approaches are ideal for
discovering more about when and where in the brain rhythm is processed,
some neuroimaging studies fail to include (or are unable to include for
methodological reasons) behavioral measures of rhythm processing, which
makes it difficult to determine exactly how differences in neural activation
relate to real-world outcomes. That is, simply because there are differences
in activation for two different rhythms does not necessarily mean that
participants will also perceive them differently. Finally, studies of patients
with brain damage or dysfunction provide significant insights, but rely on
natural accidents, which do not lead to the same amount or location of
damage in each individual. This makes it difficult to understand what
lesioned areas are truly necessary for rhythm processing or whether a
particular combination of areas is required to perceive rhythm. Of course, a
combination of these approaches, combined with methods that focally
disrupt ongoing neural processing, such as transcranial magnetic
stimulation (TMS), have given rise to a rich literature on human rhythm
perception. This chapter will review the current literature on the neural
basis of rhythm perception, which highlights important brain areas for
perceiving a beat, and how the human brain entrains to rhythms in music.
F B
O M
L M
D R
C P
E R P
Thus far, this chapter has focused on rhythm perception in the auditory
modality. However, rhythms include temporally patterned stimuli in any
modality. For example, the isochronous blinking of a car’s turn signal is a
visual rhythm, and your phone’s vibrating notification is a tactile rhythm. In
this section, we will discuss how rhythm is perceived in non-auditory
modalities, focusing on vision. Predictably, the neural correlates of rhythm
perception differ between modalities, but some are also shared. These
shared substrates might be a clue to the neural representation of rhythm in a
pure, temporal sense, uncontaminated by modality-specific processing: the
sine qua non of rhythm perception.
Like audition, vision is sensitive to temporal regularities in the
environment. For example, visual-spatial attention is biased toward reliably
repeating patterns (Zhao, Al-Aidroos, & Turk-Browne, 2013). Unlike
audition, rhythmic visual stimuli, such as a blinking dot, do not lead to a
strong sense of beat: Auditory rhythms are reproduced and remembered
better than visual ones (Glenberg, Mann, Altman, Forman, & Procise, 1989;
Collier & Logan, 2000, respectively). While it is true that audition generally
has better temporal sensitivity than vision (e.g., Goldstone & Lhamon,
1972) this does not explain why auditory rhythms give rise to a sense of
beat and visual ones do not. Recently, researchers have instantiated visual
rhythms with more dynamic stimuli in an attempt to capitalize on the visual
system’s sensitivity to motion and acceleration. A blinking stimulus isn’t
visually natural, but a moving one is. Concordantly, rotating bars and
bouncing balls can give rise to a sense of beat in a manner similar to
auditory stimuli (Grahn, 2012; Hove, Iversen, Zhang, & Repp, 2013;
Iversen, Patel, Nicodemus, & Emmorey, 2015). Even more naturalistic
stimuli, like watching a dancer or following a conductor’s baton, give rise
to timing advantages illustrative of beat perception (Luck & Sloboda, 2009;
Su & Salazar-Lopez, 2016). The message from this new literature is that
although audition wins over other modalities for temporal processing
superiority, rhythm processing is possible in other modalities when the
stimuli are crafted to follow that sense’s priorities.
Given that visual rhythm processing is possible, how does the brain do
it? One possibility is that visual rhythm processing piggy-backs on the
rhythmically-superior auditory and motor resources. According to this view,
visual rhythm perception involves the creation of an internal auditory
rhythm to accompany visual stimuli. Evidence for this perspective was
demonstrated in an fMRI task where participants watched or heard
rhythmic stimuli in counterbalanced blocks of a tempo adaptation task;
visual sequences produced a stronger sense of beat and stronger bilateral
putamen activity when preceded by the auditory task block versus with no
prior auditory experience with the task (Grahn, Henry, & McAuley, 2011).
This change in brain response during the visual task following the auditory
block resembled the activation observed in auditory tasks alone (Grahn &
Brett, 2007). When the visual task preceded the auditory block, there was
no enhancement to rhythm perception or brain response in the basal
ganglia, indicating that it was not simply a practice effect. This study used
blinking visual rhythms that do not readily elicit rhythm perception, so the
authors suggested that the observed behavior and brain responses reflected
the co-opting of typical auditory rhythm perception to achieve the
perception of a visual rhythm. In a later fMRI study with discrete and
moving visual and auditory stimuli, this putamen activity was shown to
reflect a supra-modal rhythm perception response: Activity in the putamen
corresponded to the strength of synchrony with an ongoing rhythm,
regardless of the modality and without prior auditory experience with the
stimuli (Hove, Fairhurst, Kotz, & Keller, 2013). This idea of a supra-modal,
or modality-general, process underpinning rhythm perception received
further support from a study measuring ERPs in response to temporal
expectancy violations in an adaptive tempo task with auditory and visual
stimuli (Pasinski, McAuley, & Snyder, 2016). They found larger amplitude
ERPs for the auditory task, but similar patterns as the visual response,
suggesting again the presence of a modality-general rhythm perception
network, likely rooted in the basal ganglia and the motor system.
I D M
T
Rhythm processing abilities vary widely in the general population. It is not
difficult to run into someone who proclaims that she has two left feet, and
there is evidence of individuals who are actually “beat deaf.” These
individuals cannot align their movements to the beat of a musical piece
despite being able to synchronize to a metronome (Phillips-Silver et al.,
2011). Differences in experience, such as music training, can enhance
rhythm perception and production. But individual differences in abilities
that are associated with beat perception can lead people to encode, store,
and act on auditory information in different ways. For instance, individual
differences in auditory short-term memory (STM) and regularity detection
are associated with better rhythm abilities, especially when reproducing
longer rhythms (Grahn & Schuit, 2012). Music training also accounts for
unique variance in rhythm reproduction abilities compared to auditory STM
and regularity detection, although musical training may only influence
rhythm perception abilities in certain tasks (Bauer, Kreutz, & Herrmann,
2015; Grahn & Brett, 2007; Geiser, Ziegler, Jancke, & Meyer, 2009).
Regularity detection is also correlated with activation in auditory-motor
areas, including left SMA and left dorsal and ventral premotor areas, which
may indicate that people who are better at detecting the beat in music also
rely more heavily on transforming rhythms into auditory-motor
representations instead of relying purely on auditory cues (Grahn & Schuit,
2012). These findings are similar to previous work showing that strong beat
perceivers showed greater activation in SMA compared to weak beat
perceivers, when listening to an ambiguous rhythm (Grahn & McAuley,
2009). Individual motor abilities may also be important for predicting
individual differences in preferred tempo (120 bpm or 2 Hz), which is the
rate at which listeners feel most comfortable tapping to music or a
metronome (McAuley et al., 2006). An individual’s specific peak frequency
in the beta range, assessed during a motor tapping task, predicts preferred
tempo (Bauer et al., 2015), providing additional evidence that auditory-
motor interactions can lead to differences in the way that people prefer to
entrain to music.
Although much of the literature on rhythm and beat perception makes
claims about commonalities across individuals in the neural processing of
rhythms, there is considerable variation in the way humans respond to
rhythms. These individual differences are particularly important to consider
when trying to use rhythm as a therapeutic tool, as in patients with PD.
Although rhythmic stimulation may have seemingly miraculous effects for
some individuals, there are many others for whom rhythmic stimulation
may have no effect or perhaps even a negative effect on gait (Leow, Parrott,
& Grahn, 2014; Nombela et al., 2013). Further research is necessary to
characterize what factors lead to these individual differences in neural
processing of rhythm, including auditory-motor interactions, musical
background, and biological differences to better target interventions to the
individual.
M J A
C
As we have seen, the neuroscience of rhythm is a vibrant field of study that
can be approached from many angles. Despite this variety, there has been a
common thread throughout this chapter: the involvement of motor
processes during the perception of rhythm and beat. Given the role of the
motor system in fundamental timing processes, it is not surprising that
similar networks should become involved with the perception of rhythm.
What interests us is the reliability and the variety of motor system
participation in rhythm processing, whether it is the recruitment of the basal
ganglia during the perception of strong beats as revealed by fMRI (see
section on “Feeling the Beat”), the heightened corticospinal excitability of
toe-tapping as revealed by TMS (see section on “Feeling the Beat”), the
auditory-motor coordination of beta-band patterns recorded from EEG (see
section on “Oscillatory Mechanisms”), the co-development of movement
and rhythm production in children (see section on “Development of
Rhythm”), or the use of others’ actions to guide joint action (see section on
“Comparative Psychology and Evolution of Rhythm Perception”). These
are just a few of the many ways in which the auditory and motor systems
interact to produce the rich experience of rhythm perception and
production. Promising new research offers these auditory-motor interactions
as the basis for therapies to help patients with neurodegenerative diseases of
the basal ganglia (e.g., Spaulding et al., 2013), or patients with
developmental language disorders, such as dyslexia (e.g., Flaugnacco et al.,
2015). Our ability to neurally process rhythms is not only important for
being able to clap along to our favorite song, but is important for examining
fundamental psychological questions ranging from individual differences in
perception and production to what makes humans unique from other
species. Future work on the neural bases of rhythm perception has the
potential to inform a wide range of domains including aesthetics, evolution,
and human perception and production.
R
Bartolo, R., Prado, L., & Merchant, H. (2014). Information processing in the primate basal ganglia
during sensory-guided and internally driven rhythmic tapping. Journal of Neuroscience 34(11),
3910–3923.
Bauer, A. K., Kreutz, G., & Herrmann, C. S. (2015). Individual musical tempo preference correlates
with EEG beta rhythm. Psychophysiology 52(4), 600–604.
Bekius, A., Cope, T., & Grube, M. (2016). The beat to read: A cross-lingual link between rhythmic
regularity perception and reading skill. Frontiers in Human Neuroscience 10, 425. Retrieved from
https://doi.org/10.3389/fnhum.2016.00425
Brochard, R., Abecasis, D., Potter, D., Ragot, R., & Drake, C. (2003). The “ticktock” of our internal
clock: Direct brain evidence of subjective accents in isochronous sequences. Psychological
Science 14(4), 362–366.
Brown, S., Pfordresher, P. Q., & Chow, I. (2017). A musical model of speech rhythm.
Psychomusicology: Music, Mind, and Brain 27(2), 95–112.
Cameron, D. J., Pickett, K. A., Earhart, G. M., & Grahn, J. A. (2016). The effect of dopaminergic
medication on beat-based auditory timing in Parkinson’s disease. Frontiers in Neurology 7, 19.
Retrieved from https://doi.org/10.3389/fneur.2016.00019
Cameron, D. J., Stewart, L., Pearce, M. T., Grube, M., & Muggleton, N. G. (2012). Modulation of
motor excitability by metricality of tone sequences. Psychomusicology: Music, Mind, and Brain
22(2), 122–128.
Chemin, B., Mouraux, A., & Nozaradan, S. (2014). Body movement selectively shapes the neural
representation of musical rhythms. Psychological Science 25(12), 2147–2159.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008). Listening to musical rhythms recruits motor
regions of the brain. Cerebral Cortex 18(12), 2844–2854.
Chen, J. L., Zatorre, R. J., & Penhune, V. B. (2006). Interactions between auditory and dorsal
premotor cortex during synchronization to musical rhythms. NeuroImage 32(4), 1771–1781.
Cirelli, L. K., Bosnyak, D., Manning, F. C., Sinelli, C., Marie, C., Fujioka, T., … Trainor, L. J.
(2014). Beat-induced fluctuations in auditory cortical beta-band activity: Using EEG to measure
age-related changes. Frontiers in Psychology 5, 742. Retrieved from
https://doi.org/10.3389/fpsyg.2014.00742
Cirelli, L. K., Einarson, K. M., & Trainor, L. J. (2014). Interpersonal synchrony increases prosocial
behavior in infants. Developmental Science 17(6), 1003–1011.
Cirelli, L. K., Spinelli, C., Nozaradan, S., & Trainor, L. J. (2016). Measuring neural entrainment to
beat and meter in infants: Effects of music background. Frontiers in Neuroscience 10, 229.
Retrieved from https://doi.org/10.3389/fnins.2016.00229
Collier, G. L., & Logan, G. (2000). Modality differences in short-term memory for rhythms. Memory
& Cognition 28(4), 529–538.
Collier, G. L., & Wright, C. E. (1995). Temporal rescaling of simple and complex ratios in rhythmic
tapping. Journal of Experimental Psychology: Human Perception and Performance 21(3), 602–
627.
Cook, P., Rouse, A., Wilson, M., & Reichmuth, C. (2013). A California sea lion (Zalophus
californianus) can keep the beat: Motor entrainment to rhythmic auditory stimuli in a non vocal
mimic. Journal of Comparative Psychology 127(4), 412–427.
Cooper, G., & Meyer, L. B. (1960). The rhythmic structure of music. Chicago, IL: University of
Chicago Press.
Cross, E. S., Hamilton, A. F. D. C., & Grafton, S. T. (2006). Building a motor simulation de novo:
Observation of dance by dancers. NeuroImage 31(3), 1257–1267.
D’Ausilio, A., Altenmuller, E., Olivetti Belardinelli, M., & Lotze, M. (2006). Cross-modal plasticity
of the motor cortex while listening to a rehearsed musical piece. European Journal of
Neuroscience 24(3), 955–958.
Ding, N., Patel, A. D., Chen, L., Butler, H., Luo, C., & Poeppel, D. (2017). Temporal modulations in
speech and music. Neuroscience & Biobehavioral Reviews 81(Part B), 181–187.
Drost, U. C., Rieger, M., Brass, M., Gunter, T. C., & Prinz, W. (2005). Action-effect coupling in
pianists. Psychological Research 69(4), 233–241.
Drost, U. C., Rieger, M., & Prinz, W. (2007). Instrument specificity in experienced musicians.
Quarterly Journal of Experimental Psychology 60(4), 527–533.
Epstein, D. (1995). Shaping time: Music, the brain, and performance. New York: Macmillan.
Essens, P. J. (1986). Hierarchical organization of temporal patterns. Perception & Psychophysics
40(2), 69–73.
Fairhurst, M. T., Janata, P., & Keller, P. E. (2012). Being and feeling in sync with an adaptive virtual
partner: Brain mechanisms underlying dynamic cooperativity. Cerebral Cortex 23(11), 2592–2600.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schön, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A randomized
control trial. PLoS ONE 10(9), e0138715.
Freedberg, D., & Gallese, V. (2007). Motion, emotion and empathy in esthetic experience. Trends in
Cognitive Sciences 11(5), 197–203.
Fujioka, T., Trainor, L. J., Large, E. W., & Ross, B. (2009). Beta and gamma rhythms in human
auditory cortex during musical beat processing. Annals of the New York Academy of Sciences 1169,
89–92.
Fujioka, T., Trainor, L. J., Large, E. W., & Ross, B. (2012). Internalized timing of isochronous sounds
is represented in neuromagnetic beta oscillations. Journal of Neuroscience 32(5), 1791–1802.
Gallese, V., & Goldman, A. (1998). Mirror neurons and the simulation theory of mind-reading.
Trends in Cognitive Sciences 2(12), 493–501.
Geiser, E., Ziegler, E., Jancke, L., & Meyer, M. (2009). Early electrophysiological correlates of meter
and rhythm processing in music perception. Cortex 45(1), 93–102.
Glenberg, A. M., Mann, S., Altman, L., Forman, T., & Procise, S. (1989). Modality effects in the
coding reproduction of rhythms. Memory & Cognition 17(4), 373–383.
Goldstone, S., & Lhamon, W. T. (1972). Auditory-visual differences in human temporal judgment.
Perceptual and Motor Skills 34(2), 623–633.
Grabe, E., & Low, L. (2002). Durational variability in speech and the rhythm class hypothesis. In N.
Warner & C. Gussenhoven (Eds.), Papers in laboratory phonology 7 (pp. 515–546). Berlin:
Mouton de Gruyter.
Grahn, J. A. (2012). See what I hear? Beat perception in auditory and visual rhythms. Experimental
Brain Research 220(1), 51–61.
Grahn, J. A., & Brett, M. (2007). Rhythm and beat perception in motor areas of the brain. Journal of
Cognitive Neuroscience 19(5), 893–906.
Grahn, J. A., & Brett, M. (2009). Impairment of beat-based rhythm discrimination in Parkinson’s
disease. Cortex 45(1), 54–61.
Grahn, J. A., Henry, M. J., & McAuley, J. D. (2011). fMRI investigation of cross-modal interactions
in beat perception: Audition primes vision, but not vice versa. NeuroImage 54(2), 1231–1243.
Grahn, J. A., & McAuley, J. D. (2009). Neural bases of individual difference in beat perception.
NeuroImage 47(4), 1894–1903.
Grahn, J. A., & Rowe, J. B. (2009). Feeling the beat: Premotor and striatal interactions in musicians
and nonmusicians during beat perception. Journal of Neuroscience 29(23), 7540–7548.
Grahn, J. A., & Schuit, D. (2012). Individual difference in rhythmic ability: Behavioral and
neuroimaging investigations. Psychomusicology: Music, Mind, and Brain 22(2), 105–121.
Grube, M., Cooper, F. E., & Griffiths, T. D. (2013). Auditory temporal-regularity processing
correlates with language and literacy skill in early adulthood. Cognitive Neuroscience 3(3–4), 225–
230.
Habib, M., Lardy, C., Desiles, T., Commeiras, C., Chobert, J., & Besson, M. (2016). Music and
dyslexia: A new musical training method to improve reading and related disorders. Frontiers in
Psychology 7, 26. Retrieved from https://doi.org/10.3389/fpsyg.2016.00026
Harrington, D. L., Haaland, K. Y., & Hermanowitz, N. (1998). Temporal processing in the basal
ganglia. Neuropsychology 12(1), 3–12.
Hasegawa, A., Okanoya, K., Hasegawa, T., & Seki, Y. (2011). Rhythmic synchronization tapping to
an audio-visual metronome in budgerigars. Scientific Reports 1, 120. doi:10.1038/srep00120
Hattori, Y., Tomonaga, M., & Matsuzawa, T. (2013). Spontaneous synchronized tapping to an
auditory rhythm in a chimpanzee. Scientific Reports 3, 1566. doi:10.1038/srep01566
Haueisen, J., & Knösche, T. R. (2001). Involuntary motor activity in pianists evoked by music
perception. Journal of Cognitive Neuroscience 13(6), 786–792.
Henry, M. J., & Herrmann, B. (2014). Low-frequency neural oscillations support dynamic attending
in temporal context. Timing and Time Perception 2(1), 62–86.
Henry, M. J., Herrmann, B., & Grahn, J. A. (2017). What can we learn about beat perception by
comparing brain signals and stimulus envelopes? PLoS ONE 12(2), e0172454.
Henry, M. J., & Obleser, J. (2012). Frequency modulation entrains slow neural oscillations and
optimizes human listening behavior. Proceedings of the National Academy of Sciences 109(49),
20095–20100.
Honing, H., Merchant, H., Háden, G. P., Prado, L., & Bartolo, R. (2012). Rhesus monkeys (Macaca
mulatta) detect rhythmic groups in music, but not the beat. PloS ONE 7(12), e51369.
Hove, M. J., Fairhurst, M. T., Kotz, S. A., & Keller, P. E. (2013). Synchronizing with auditory and
visual rhythms: An fMRI assessment of modality differences and modality appropriateness.
NeuroImage 67, 313–321.
Hove, M. J., Iversen, J. R., Zhang, A., & Repp, B. H. (2013). Synchronization with competing visual
and auditory rhythms: Bouncing ball meets metronome. Psychological Research 77(4), 388–398.
Iversen, J. R., Patel, A. D., Nicodemus, B., & Emmorey, K. (2015). Synchronization to auditory and
visual rhythms in hearing and deaf individuals. Cognition 134, 232–244.
Jones, M. R., & Boltz, M. (1989). Dynamic attending and responses to time. Psychological Review
96(3), 459–491.
Kirschner, S., & Tomasello, M. (2010). Joint music making promotes prosocial behavior in 4-year-
old children. Evolution and Human Behavior 31(5), 354–364.
Large, E. W. (2008). Resonating to musical rhythm: Theory and experiment. In S. Grondin (Ed.),
Psychology of time (pp. 189–231). Bingley: Emerald.
Large, E. W., & Gray, P. M. (2015). Spontaneous tempo and rhythmic entrainment in a bonobo (Pan
paniscus). Journal of Comparative Psychology 129(4), 317–328.
Large, E. W., & Jones, M. R. (1999). The dynamics of attending: How people track time-varying
events. Psychological Review 106(1), 119–159.
Large, E. W., & Snyder, J. S. (2009). Pulse and meter as neural resonance. Annals of the New York
Academy of Sciences 1169, 46–57.
Leder, H., Bär, S., & Topolinski, S. (2012). Covert painting simulations influence aesthetic
appreciation of artworks. Psychological Science 23(12), 1479–1481.
Lehongre, K., Ramus, F., Villiermet, N., Schwartz, D., & Giraud, A.-L. (2011). Altered low-gamma
sampling in auditory cortex accounts for the three main facets of dyslexia. Neuron 72(6), 1080–
1090.
Leow, L.-A., Parrott, T., & Grahn, J. A. (2014). Individual differences in beat perception affect gait
responses to low- and high-groove music. Frontiers in Human Neuroscience 8, 1–12. Retrieved
from https://doi.org/10.3389/fnhum.2014.00811
Lerdahl, F., & Jackendoff, R. (1983). A generative theory of tonal music. Cambridge, MA: MIT
Press.
Lotze, M., Scheler, G., Tan, H. R., Braun, C., & Birbaumer, N. (2003). The musician’s brain:
Functional imaging of amateurs and professionals during performance and imagery. NeuroImage
20(3), 1817–1829.
Luck, G., & Sloboda, J. A. (2009). Spatio-temporal cues for visually mediated synchronization.
Music Perception: An Interdisciplinary Journal 26(5), 465–473.
McAuley, J. D., Jones, M. R., Holub, S., Johnston, H. M., & Miller, N. S. (2006). The time of our
lives: Life span development of timing and event tracking. Journal of Experimental Psychology:
General 135(3), 348–367.
Merchant, H., & Honing, H. (2014). Are non-human primates capable of rhythmic entrainment?
Evidence for the gradual audiomotor evolution hypothesis. Frontiers in Neuroscience 7, 274.
Retrieved from https://doi.org/10.3389/fnins.2013.00274
Merchant, H., Pérez, O., Zarco, W., & Gámez, J. (2013). Interval tuning in the primate medial
premotor cortex as a general timing mechanism. Journal of Neuroscience 33(21), 9082–9096.
Nombela, C., Rae, C. L., Grahn, J. A., Barker, R. A., Owen, A. M., & Rowe, J. B. (2013). How often
does music and rhythm improve patients’ perception of motor symptoms in Parkinson’s disease?
Journal of Neurology 260(5), 1404–1405.
Novembre, G., Ticini, L. F., Schütz-Bosbach, S., & Keller, P. E. (2013). Motor simulation and the
coordination of self and other in real-time joint action. Social Cognitive and Affective
Neuroscience 9(8), 1062–1068.
Nozaradan, S., Peretz, I., Missal, M., & Mouraux, A. (2011). Tagging the neuronal entrainment to
beat and meter. Journal of Neuroscience 31(28), 10234–10240.
Pasinski, A. C., McAuley, J. D., & Snyder, J. S. (2016). How modality specific is processing of
auditory and visual rhythms? Psychophysiology 53(2), 198–208.
Patel, A. D. (2003). Rhythm in language and music. Annals of the New York Academy of Sciences
999, 140–143.
Patel, A. D. (2006). Musical rhythm, linguistic rhythm, and human evolution. Music Perception: An
Interdisciplinary Journal 24(1), 99–104.
Patel, A. D., Iversen, J. R., Bregman, M. R., & Schulz, I. (2009). Experimental evidence for
synchronization to a musical beat in a nonhuman animal. Current Biology 19(10), 827–830.
Patel, A. D., Iversen, J. R., Chen, Y., & Repp, B. H. (2005). The influence of metricality and
modality on synchronization with a beat. Experimental Brain Research 163(2), 226–238.
Peelle, J. E., Gross, J., & Davis, M. H. (2013). Phase-locked responses to speech in human auditory
cortex are enhanced during comprehension. Cerebral Cortex 23(6), 1378–1387.
Phillips-Silver, J., Toiviainen, P., Gosselin, N., Piche, O., Nozaradan, S., Palmer, C., & Peretz, I.
(2011). Born to dance but beat deaf: A new form of congenital amusia. Neuropsychologia 49(5),
961–969.
Picton, T. W., John, M. S., Dimitrijevic, A., & Purcell, D. (2003). Human auditory steady-state
responses. International Journal of Audiology 42, 177–219.
Poole, J. H., Tyack, P. L., Stoeger-Horwath, A. S., & Watwood, S. (2005). Animal behaviour:
Elephants are capable of vocal learning. Nature 434(7032), 455–456.
Povel, D. J., & Essens, P. (1985). Perception of temporal patterns. Music Perception: An
Interdisciplinary Journal 2(4), 411–440.
Povel, D. J., & Okkerman, H. (1981). Accents in equitone sequences. Perception & Psychophysics
30(6), 565–572.
Power, A. J., Colling, L. J., Mead, N., Barnes, L., & Goswami, U. (2016). Neural encoding of the
speech envelope by children with developmental dyslexia. Brain and Language 160, 1–10.
Press, C., Cook, J., Blakemore, S. J., & Kilner, J. (2011). Dynamic modulation of human motor
activity when observing actions. Journal of Neuroscience 31(8), 2792–2800.
Ravignani, A., Delgado, T., & Kirby, S. (2017). Musical evolution in the lab exhibits rhythmic
universals. Nature Human Behaviour 1(1), 0007. doi:10.1038/s41562-016-0007
Rizzolatti, G., & Craighero, L. (2004). The mirror-neuron system. Annual Review of Neuroscience
27, 169–192.
Schachner, A., Brady, T. F., Pepperberg, I. M., & Hauser, M. D. (2009). Spontaneous motor
entrainment to music in multiple vocal mimicking species. Current Biology 19(10), 831–836.
Schroeder, C. E., Lakatos, P., Kajikawa, Y., Partan, S., & Puce, A. (2008). Neuronal oscillations and
visual amplification of speech, Cell Press 12(3), 106–113.
Schwartze, M., Keller, P. E., Patel, A. D., & Kotz, S. A. (2011). The impact of basal ganglia lesions
on sensorimotor synchronization, spontaneous motor tempo, and the detection of tempo changes.
Behavioural Brain Research 216(2), 685–691.
Selezneva, E., Deike, S., Knyazeva, S., Scheich, H., Brechmann, A., & Brosch, M. (2013). Rhythm
sensitivity in macaque monkeys. Frontiers in Systems Neuroscience 7. Retrieved from
https://doi.org/10.3389/fnsys.2013.00049
Snyder, J. S., & Large, E. W. (2005). Gamma-band activity reflects the metric structure of rhythmic
tone sequences. Cognitive Brain Research 24(1), 117–126.
Spaulding, S. J., Barber, B., Colby, M., Cormack, B., Mick, T., & Jenkins, M. E. (2013). Cueing and
gait improvement among people with Parkinson’s disease: A meta-analysis. Archives of Physical
Medicine and Rehabilitation 94(3), 562–570.
Stupacher, J., Hove, M. J., Novembre, G., Schutz-Bosbach, S., & Keller, P. E. (2013). Musical
groove modulates motor cortex excitability: A TMS investigation. Brain and Cognition 82(2),
127–136.
Su, Y. H., & Salazar-López, E. (2016). Visual timing of structured dance movements resembles
auditory rhythm perception. Neural Plasticity 2016, 1678390. doi:10.1155/2016/1678390
Taylor, J. E. T., & Witt, J. K. (2015). Listening to music primes space: Pianists, but not novices,
simulate heard actions. Psychological Research 79(2), 175–182.
Taylor, J. E. T., Witt, J. K., & Grimaldi, P. J. (2012). Uncovering the connection between artist and
audience: Viewing painted brushstrokes evokes corresponding action representations in the
observer. Cognition 125(1), 26–36.
Teki, S., Grube, M., Kumar, S., & Griffiths, T. D. (2011). Distinct neural substrates of duration-based
and beat-based auditory timing. Journal of Neuroscience 31(10), 3805–3812.
Winkler, I., Haden, G. P., Ladinig, O., Sziller, I., & Honing, H. (2009). Newborn infants detect the
beat in music. Proceedings of the National Academy of Sciences 106(7), 2468–2471.
Woodruff Carr, K., White-Schwoch, T., Tierney, A. T., Strait, D. L., & Kraus, N. (2014). Beat
synchronization predicts neural speech encoding and reading readiness in preschoolers.
Proceedings of the National Academy of Sciences 111(40), 14559–14564.
Zarco, W., Merchant, H., Prado, L., & Mendez, J. C. (2009). Subsecond timing in primates:
Comparison of interval production between human subjects and rhesus monkeys. Journal of
Neurophysiology 102(6), 3191–3202.
Zhao, J., Al-Aidroos, N., & Turk-Browne, N. B. (2013). Attention is spontaneously biased toward
regularities. Psychological Science 24(5), 667–677.
Zuk, J., Bishop-Liebler, P., Ozernov-Palchik, O., Moore, E., Overy, K., Welch, G., & Gaab, N.
(2017). Revisiting the “enigma” of musicians with dyslexia: Auditory sequencing and speech
abilities. Journal of Experimental Psychology: General 146(4), 495–511.
CHAPT E R 9
S T E FA N K O E L S C H
W D N O H O
C
The auditory system evolved phylogenetically from the vestibular system.
Interestingly, the vestibular nerve contains a substantial number of
acoustically responsive fibers. The otolith organs (saccule and utricle) are
sensitive to sounds and vibrations (Todd, Paillard, Kluk, Whittle, &
Colebatch, 2014), and the vestibular nuclear complex in the brainstem
exerts a major influence on spinal (and ocular) motoneurons in response to
loud sounds with low frequencies, or with sudden onsets (Todd et al., 2014;
Todd & Cody, 2000). Moreover, both the vestibular nuclei and the auditory
cochlear nuclei in the brainstem project to the reticular formation (also in
the brainstem), and the vestibular nucleus also projects to the parabrachial
nucleus, a convergence site for vestibular, visceral, and autonomic
processing in the brainstem (Balaban & Thayer, 2001; Kandler & Herbert,
1991). Such projections initiate and support movements and contribute to
the arousing effects of music. Thus, subcortical processing of sounds does
not only give rise to auditory sensations, but also to muscular and
autonomic responses, and the stimulation of motoneurons and autonomic
neurons by low-frequency beats might contribute to the human impetus to
“move to the beat” (Grahn & Rowe, 2009; Todd & Cody, 2000). In addition
to vibrations of the vestibular apparatus and cochlea, sounds also evoke
resonances in vibration receptors, that is, in the Pacinian corpuscles (which
are sensitive from 10 Hz to a few kHz, and located mainly in the skin, the
retroperitoneal space in the belly, the periosteum of the bones, and the sex
organs), and maybe even responses in mechanoreceptors of the skin that
detect pressure. The famous international concert percussionist Dame
Evelyn Glennie is profoundly deaf and hears mainly through vibrations felt
in the skin (personal communication with Dame Glennie), and probably in
the vestibular organ. Thus, we do not only hear with our cochlea, but also
with the vestibular apparatus and mechanoreceptors distributed throughout
our body.
A F E
B T
E M G
F
While auditory features are extracted, the acoustic information enters the
auditory sensory memory (or “echoic memory”), and representations of
auditory Gestalten (Griffiths & Warren, 2004) or “auditory objects” are
formed. The auditory sensory memory (ASM) retains information only for a
few seconds, and information stored in the ASM fades quickly. The ASM is
thought to store physical features of sounds (such as pitch, intensity,
duration, location, timbre, etc.), sound patterns, and even abstract features
of sound patterns (e.g., Paavilainen, Simola, Jaramillo, Näätänen, &
Winkler, 2001). Operations of the ASM are at least partly reflected
electrically in the mismatch negativity (MMN, e.g., Näätänen, Tervaniemi,
Sussman, Paavilainen, & Winkler, 2001). The MMN is an ERP with
negative polarity and a peak-latency of about 100–200 ms and appears to
receive its main contributions from neural sources located in the PAC and
adjacent auditory (belt) fields, with additional (but smaller) contributions
from frontal cortical areas (for reviews, see Deouell, 2007; Schönwiesner et
al., 2007).
Auditory sensory memory operations are indispensable for music
perception; therefore, practically all MMN studies are inherently related to,
and relevant for, the understanding of the neural correlates of music
processing. As will be outlined below, numerous MMN studies have
contributed to this issue (a) by investigating different response properties of
the ASM to musical and speech stimuli, (b) by using melodic and rhythmic
patterns to investigate auditory Gestalt formation, and/or (c) by studying
effects of long- and short-term musical training on processes underlying
ASM operations. Especially the latter studies have contributed substantially
to our understanding of neuroplasticity (i.e., to changes in neuronal
structure and function due to experience), and thus to our understanding of
the neural basis of learning (for a review see Tervaniemi, 2009). Here,
suffice it to say that MMN studies showed effects of long-term musical
training on the processing of sound localization, pitch, melody, rhythm,
musical key, timbre, tuning, and timing (e.g., Koelsch, Schröger, &
Tervaniemi, 1999; Putkinen, Tervaniemi, Saarikivi, de Vent, & Huotilainen,
2014; Rammsayer & Altenmüller, 2006; Tervaniemi, Castaneda, Knoll, &
Uther, 2006; Tervaniemi, Janhunen, Kruck, Putkinen, & Huotilainen, 2016).
Auditory oddball paradigms were also used to investigate processes of
melodic and rhythmic grouping of tones occurring in tone patterns (such
grouping is essential for auditory Gestalt formation, see also Sussman,
2007), as well as effects of musical long-term training on these processes.
These studies showed effects of musical training (a) on the processing of
melodic patterns (Fujioka, Trainor, Ross, Kakigi, & Pantev, 2004;
Tervaniemi, Ilvonen, Karma, Alho, & Näätänen, 1997; Tervaniemi,
Rytkönen, Schröger, Ilmoniemi, & Näätänen, 2001; Zuijen, Sussman,
Winkler, Näätänen, & Tervaniemi, 2004; in these studies, patterns consisted
of four or five tones), (b) on the encoding of the number of elements in a
tone pattern (Zuijen, Sussman, Winkler, Näätänen, & Tervaniemi, 2005),
and (c) on the processing of patterns consisting of two voices (Fujioka,
Trainor, Ross, Kakigi, & Pantev, 2005).
The formation of auditory Gestalten entails processes of perceptual
separation, as well as processes of melodic, rhythmic, timbral, and spatial
grouping. Such processes have been summarized under the concepts of
auditory scene analysis and auditory stream segregation (Bregman, 1994).
Grouping of acoustic events follows Gestalt principles such as similarity,
proximity, and continuity (for acoustic cues used for perceptual separation
and auditory grouping see Darwin, 1997, 2008). In everyday life, such
operations are not only important for music processing, but also, for
instance, for separating a speaker’s voice during a conversation from other
sound sources in the environment. That is, these operations are important
because their function is to recognize and to follow acoustic objects, and to
establish a cognitive representation of the acoustic environment. It appears
that the planum temporale (which is part of the auditory association cortex)
is a crucial structure for auditory scene analysis and stream segregation,
particularly due to its role for the processing of pitch intervals and sound
sequences (Griffiths & Warren, 2002; Patterson, Uppenkamp, Johnsrude, &
Griffiths, 2002; Snyder & Elhilali, 2017).
M E F :
P L D
M S B :
P N D
FIGURE 2. (a) Examples of chord functions: The chord built on the first scale tone is denoted as
the tonic, the chord on the second tone as the supertonic, and the chord on the fifth tone as the
dominant. (b) The dominant-tonic progression represents a regular ending of a harmonic sequence
(top), the dominant-supertonic progression is less regular and unacceptable as a marker of the end of
a harmonic progression (bottom sequence, the arrow indicates the less regular chord). (c) ERPs
elicited in a passive listening condition by the final chords of the two sequence types shown in (b).
Both sequence types were presented in pseudorandom order equiprobably in all twelve major keys.
Brain responses to irregular chords clearly differ from those to regular chords (best to be seen in the
black difference wave, regular subtracted from irregular chords). The first difference between the
two waveforms is maximal around 200 ms after the onset of the fifth chord (ERAN, indicated by the
long arrow) and taken to reflect processes of music-syntactic analysis. The ERAN is followed by an
N5 taken to reflect processes of harmonic integration (short arrow). (d) Activation foci (small
spheres) reported by functional imaging studies on music-syntactic processing using chord sequence
paradigms (Koelsch, Fritz, et al., 2005; Maess et al., 2001; Koelsch et al., 2002; Tillmann et al.,
2003) and melodies (Janata et al., 2002). Large gray disks show the mean coordinates of foci
(averaged for each hemisphere across studies, coordinates9 refer to standard stereotaxic space).
Reprinted from Trends in Cognitive Sciences, 9(12), Stefan Koelsch and Walter A. Siebel, Towards a
neural basis of music perception, pp. 578–584, Copyright © 2005 Elsevier Ltd. All rights reserved.
At the same time, the final tonic “prolongs” the initial tonic, whereas the
final supertonic does not. Therefore, the nonlocal dependency between
initial and final chord is fulfilled in the upper sequence and violated in the
bottom sequence. Figure 2c shows brain-electric responses to the final
chords of the sequences shown in Figure 2b: the irregular supertonics elicit
an ERAN (early right anterior negativity, indicated by the arrow) compared
to the regular tonic chords. Importantly, as described earlier, the ERAN
elicited here is a conglomerate of the sMMN (due to processing the local
dependency violation) and the “hierarchical ERAN” (due to the processing
of the nonlocal dependency violation). A study by Zhang and colleagues
(Zhang, Zhou, Chang, & Yang, 2018), however, nicely showed effects of
nonlocal context effects on local harmonic processing using the ERAN.
The ERAN has a larger amplitude in individuals with musical training,
is reduced by strong attentional demands, but can be elicited even if
participants ignore the musical stimulus (for a review see Koelsch, 2012).
Most studies reporting an ERAN used harmonies as stimuli, but the ERAN
can also be elicited by melodies (e.g., Carrus, Pearce, & Bhattacharya,
2013; Fiveash, Thompson, Badcock, & McArthur, 2018; Miranda &
Ullman, 2007; Zendel, Lagrois, Robitaille, & Peretz, 2015). Moreover, a
study by Sun and colleagues (Sun, Liu, Zhou, & Jiang, 2018) reported that
the ERAN can also be elicited by rhythmic syntactic processing.
Interestingly, a study by Przysinda and colleagues (Przysinda, Zeng, Maves,
Arkin, & Loui, 2017) showed differential ERAN responses in classical and
jazz musicians depending on their preferences for irregular, or unusual
harmonies. The ERAN is relatively immune against predictions: the ERAN
latency, but not amplitude, is influenced by veridical expectations (Guo &
Koelsch, 2016). However, Vuvan and colleagues (Vuvan, Zendel, & Peretz,
2018) reported that random feedback (including false feedback) on
participants’ detection of out-of-key tones in melodies modulated the
ERAN amplitude, possibly suggesting that attention-driven changes in the
confidence in predictions (i.e., changes in the precision of predictions)
might alter the ERAN amplitude. Recent studies also report that the ERAN
is absent in individuals with “amusia” (Sun, Lu, et al., 2018), or that pitch-
judgment tasks can eliminate the ERAN in amusics (Zendel et al., 2015).
In children, the ERAN becomes visible around the age of 30 months
(Jentschke, Friederici, & Koelsch, 2014), and several studies have reported
ERAN responses in pre-school children (Corrigall & Trainor, 2014;
Jentschke, Koelsch, Sallat, & Friederici, 2008; Koelsch, Grossmann,
Gunter, Hahne, & Friederici, 2003). Children with specific language
impairment show a reduced (or absent) ERAN (Jentschke et al., 2008),
whereas neurophysiological correlates of language-syntactic processing are
developed earlier, and more strongly in children with musical training
(Jentschke & Koelsch, 2009).
Functional neuroimaging studies using chord sequences (similar to those
shown in Figure 2b, e.g., Koelsch et al., 2002; Koelsch, Fritz, Schulze,
Alsop, & Schlaug, 2005; Maess, Koelsch, Gunter, & Friederici, 2001;
Tillmann, Janata, & Bharucha, 2003; Villarreal, Brattico, Leino, Østergaard,
& Vuust, 2011) or melodies (Janata, Tillmann, & Bharucha, 2002) suggest
that music-syntactic processing involves the pars opercularis of the inferior
frontal gyrus (corresponding to BA 44v; Amunts et al., 2010) bilaterally,
but with right-hemispheric weighting (see the spheres in Figure 2d). It
seems likely that the involvement of BA 44v in music-syntactic processing
is mainly due to the hierarchical processing of (syntactic) information: This
part of Broca’s area is involved in the hierarchical processing of syntax in
language (e.g., Friederici et al., 2006; Makuuchi, Bahlmann, Anwander, &
Friederici, 2009), the hierarchical processing of action sequences (e.g.,
Fazio et al., 2009; Koechlin & Jubault, 2006), and possibly also in the
processing of hierarchically organized mathematical formulas and termini
(Friedrich & Friederici, 2009; although activation in the latter study cannot
clearly be assigned to BA 44 or BA 45). Finally, using an artificial musical
grammar, a recent study by Cheung and colleagues (Cheung, Meyer,
Friederici, & Koelsch, 2018) reported activation of BA 44v associated with
the processing of nonlocal (nested) dependencies (however, note that
dependencies in that study were not hierarchically organized).
It appears that inferior BA 44 is not the only structure involved in
music-syntactic processing: additional structures include the superior part
of the pars opercularis (Koelsch et al., 2002), ventral premotor cortex
(PMCv; Janata et al., 2002; Koelsch, Fritz, et al., 2005; Parsons, 2001), and
the anterior portion of the STG (Koelsch, Fritz, et al., 2005). The PMCv
possibly contributes to the processing of local music-syntactic dependencies
(i.e., information based on a finite state grammar): activations of PMCv
have been reported in a variety of functional imaging studies on auditory
processing using musical stimuli, linguistic stimuli, auditory oddball
paradigms, pitch discrimination tasks, and serial prediction tasks,
underlining the importance of these structures for the sequencing of
structural information, the recognition of structure, and the prediction of
sequential information (Janata & Grafton, 2003). With regard to language,
Friederici (2004) reported that activation foci of functional neuroimaging
studies on the processing of hierarchically organized long-distance
dependencies and transformations are located in the posterior IFG (with the
mean of the coordinates reported in that article being located in the inferior
pars opercularis), whereas activation foci of functional neuroimaging
studies on the processing of local dependency violations are located in the
PMCv (see also Friederici et al., 2006; Makuuchi et al., 2009; Opitz &
Kotz, 2011). Moreover, patients with lesions in the PMCv show disruption
of the processing of finite state, but not phrase-structure grammar (Opitz &
Kotz, 2011).
That is, in the abovementioned experiments that used chord sequence
paradigms to investigate the processing of harmonic structure, the music-
syntactic processing of the chord functions probably involved processing of
both finite state grammar (local dependencies) and phrase-structure (or
“context-free”) grammar (hierarchically organized nonlocal dependencies).
The music-syntactic analysis involved a computation of the harmonic
relation between a chord function and the context of preceding chord
functions (phrase-structure grammar). Such a computation is more difficult
(and less common) for irregular than for regular chord functions, and this
increased difficulty is presumably reflected in a stronger activation of
(inferior) BA 44 in response to irregular chords. In addition, the local
transition probability from the penultimate to the final chord is lower for the
dominant–supertonic progression than for the dominant–tonic progression
(finite state grammar), and the computation of the (less predicted) lower-
probability progression is presumably reflected in a stronger activation of
PMCv in response to irregular chords. The stronger activation of both BA
44 and PMCv appears to correlate with the perception of a music-
syntactically irregular chord as “unexpected” (although emotional effects of
irregular chords probably originate from BA 47, discussed below).
Note that the ability to process context-free grammar is available to
humans, whereas non-human primates are apparently not able to master
such grammars (Fitch & Hauser, 2004). Thus, it is highly likely that only
humans can adequately process music-syntactic information at the phrase-
structure level. It is also worth noting that numerous studies showed that
even “non-musicians” (i.e., individuals who have not received formal
musical training) have a highly sophisticated (implicit) knowledge about
musical syntax (e.g., Tillmann, Bharucha, & Bigand, 2000). Such
knowledge is presumably acquired during listening experiences in everyday
life.
Finally, it is important to note that violations of musical expectancies
also have emotional effects, such as surprise, or tension (Huron, 2006;
Koelsch, 2014; Lehne & Koelsch, 2015; Meyer, 1956). Consequently,
musical irregularity confounds emotion-eliciting effects, and it is difficult to
disentangle cognitive and emotional effects of music-syntactic irregularities
in neuroscientific experiments. For example, a study by Koelsch and
colleagues (Koelsch, Ftiz, et al., 2005) reported activation foci in both BA
44 and BA 47 (among other structures) in response to musical expectancy
violations, and a study by Levitin and Menon (2005) reported activation of
BA 47 (without BA 44) in response to scrambled (unpleasant) vs. normal
music. BA 47 is paralimbic, five-layered palaeocortex (not neocortex), and
activation of this region with musical irregularities is most likely due to
emotional effects (this is also consistent with an fMRI study reporting that
musical tension correlates with neural activity in BA 47; Lehne, Rohrmeier,
& Koelsch, 2014). Note that, because BA 47 is not neocortex, it is
problematic to consider this region as a “language area.” Moreover, BA 47
is adjacent to BA 44/45/46, thus activation foci originating in Broca’s area
can easily be misplaced in BA 47. Based on receptorarchitectonic (and
cytoarchitectonic) data, a study by Amunts et al. (2010) showed that BA 47
does not cluster together with BA 44/45/46 (Broca’s area in the wider
sense), nor with BA 6 (PMC).
As mentioned earlier, hierarchical processing of syntactic information
from different domains (such as music and language) requires contributions
from neural populations located in BA 44. However, it is still possible that,
although such neural populations are located in the same brain area, entirely
different (non-overlapping) neural populations serve the syntactic
processing of music and language within the same area. That is, perhaps the
neural populations mediating language-syntactic processing in BA 44 are
different from neural populations mediating music-syntactic processing in
the same area. Therefore, the strongest evidence for shared neural resources
for the syntactic processing of music and language stems from experiments
that revealed interactions between music-syntactic and language-syntactic
processing (Carrus et al., 2013; Fedorenko, Patel, Casasanto, Winawer, &
Gibson, 2009; Koelsch, Gunter, Wittfoth, & Sammler, 2005; Patel, Iversen,
Wassenaar, & Hagoort, 2008; Slevc, Rosenberg, & Patel, 2009; Steinbeis &
Koelsch, 2008). In these studies, chord sequences or melodies were played
simultaneously with (visually presented) sentences, and it was shown, for
example, that the ERAN elicited by irregular chords interacted with the left
anterior negativity (LAN) elicited by linguistic (morpho-syntactic)
violations (Koelsch, Gunter, et al., 2005; Steinbeis & Koelsch, 2008). Thus,
music-syntactic processes can interfere with language-syntactic processes.
In summary, neurophysiological studies show that music- and language-
syntactic processes engage overlapping resources (presumably located in
the inferior frontolateral cortex), and evidence showing that these resources
underlie music- and language-syntactic processing is provided by
experiments showing interactions between ERP components reflecting
music- and language-syntactic processing (in particular LAN and ERAN).
Importantly, such interactions are observed in the absence of interactions
between LAN and MMN, that is, in the absence between language-syntactic
and acoustic deviance processing (reflected in the MMN), and in the
absence of interactions between the ERAN and the N400 (i.e., in the
absence of music-syntactic and language-semantic processing). Therefore,
the reported interactions between LAN and ERAN are syntax-specific and
cannot be observed in response to any kind of irregularity.
C R
R
Abla, D., Katahira, K., & Okanoya, K. (2008). On-line assessment of statistical learning by event-
related potentials. Journal of Cognitive Neuroscience 20(6), 952–964.
Amunts, K., Lenzen, M., Friederici, A. D., Schleicher, A., Morosan, P., Palomero-Gallagher, N., &
Zilles, K. (2010). Broca’s region: Novel organizational principles and multiple receptor mapping.
PLoS Biology 8(9), e1000489.
Balaban, C. D., & Thayer, J. F. (2001). Neurological bases for balance–anxiety links. Journal of
Anxiety Disorders 15(1), 53–79.
Bard, P. (1934). On emotional expression after decortication with some remarks on certain theoretical
views: Part II. Psychological Review 41(5), 424.
Bendor, D., & Wang, X. (2005). The neuronal representation of pitch in primate auditory cortex.
Nature 436(7054), 1161–1165.
Bregman, A. S. (1994). Auditory scene analysis: The perceptual organization of sound. Cambridge,
MA: MIT Press.
Cardoso, S. H., Coimbra, N. C., & Brandão, M. L. (1994). Defensive reactions evoked by activation
of NMDA receptors in distinct sites of the inferior colliculus. Behavioural Brain Research 63(1),
17–24.
Carrus, E., Pearce, M. T., & Bhattacharya, J. (2013). Melodic pitch expectation interacts with neural
responses to syntactic but not semantic violations. Cortex 49(8), 2186–2200.
Cheung, V., Meyer, L., Friederici, A. D., & Koelsch, S. (2018). The right inferior frontal gyrus
processes hierarchical non-local dependencies in music. Scientific Reports 8, 3822.
doi:10.1038/s41598-018-22144-9
Chomsky, N. (1995). The minimalist program. Cambridge, MA: MIT Press.
Cleeremans, A., Destrebecqz, A., & Boyer, M. (1998). Implicit learning: News from the front. Trends
in Cognitive Sciences 2(10), 406–416.
Conway, C. M., & Christiansen, M. H. (2005). Modality-constrained statistical learning of tactile,
visual, and auditory sequences. Journal of Experimental Psychology: Learning, Memory, and
Cognition 31(1), 24–39.
Corrigall, K. A., & Trainor, L. J. (2014). Enculturation to musical pitch structure in young children:
Evidence from behavioral and electrophysiological methods. Developmental Science 17(1), 142–
158.
Daikoku, T., Yatomi, Y., & Yumoto, M. (2014). Implicit and explicit statistical learning of tone
sequences across spectral shifts. Neuropsychologia 63, 194–204.
Daikoku, T., Yatomi, Y., & Yumoto, M. (2015). Statistical learning of music-and language-like
sequences and tolerance for spectral shifts. Neurobiology of Learning and Memory 118, 8–19.
Darwin, C. J. (1997). Auditory grouping. Trends in Cognitive Sciences 1(9), 327–333.
Darwin, C. J. (2008). Listening to speech in the presence of other sounds. Philosophical Transactions
of the Royal Society B: Biological Sciences 363(1493), 1011–1021.
Deouell, L. Y. (2007). The frontal generator of the mismatch negativity revisited. Journal of
Psychophysiology 21(3/4), 188–203.
Dibben, N. (1994). The cognitive reality of hierarchic structure in tonal and atonal music. Music
Perception 12(1), 1–25.
Dienes, Z. (2012). Conscious versus unconscious learning of structure. In P. Rebuschat & J. Williams
(Eds.), Statistical learning and language acquisition (pp. 337–364). Berlin: Walter de Gruyter.
Ettlinger, M., Margulis, E. H., & Wong, P. C. (2011). Implicit memory in music and language.
Frontiers in Psychology 2. Retrieved from https://doi.org/10.3389/fpsyg.2011.00211
Fazio, P., Cantagallo, A., Craighero, L., D’Ausilio, A., Roy, A. C., Pozzo, T., … Fadiga, L. (2009).
Encoding of human action in Broca’s area. Brain 132(7), 1980–1988.
Fedorenko, E., Patel, A., Casasanto, D., Winawer, J., & Gibson, E. (2009). Structural integration in
language and music: Evidence for a shared system. Memory & Cognition 37(1), 1–19.
Fitch, W. T., & Hauser, M. D. (2004). Computational constraints on syntactic processing in a
nonhuman primate. Science 303(5656), 377–380.
Fiveash, A., Thompson, W. F., Badcock, N. A., & McArthur, G. (2018). Syntactic processing in
music and language: Effects of interrupting auditory streams with alternating timbres.
International Journal of Psychophysiology 129(1), 31–40.
François, C., Chobert, J., Besson, M., & Schön, D. (2013). Music training for the development of
speech segmentation. Cerebral Cortex 23(9), 2038–2043.
Francois, C., & Schön, D. (2011). Musical expertise boosts implicit learning of both musical and
linguistic structures. Cerebral Cortex 21(10), 2357–2365.
François, C., & Schön, D. (2014). Neural sensitivity to statistical regularities as a fundamental
biological process that underlies auditory learning: The role of musical practice. Hearing Research
308, 122–128.
Friederici, A. D. (2004). Processing local transitions versus long-distance syntactic hierarchies.
Trends in Cognitive Sciences 8(6), 245–247.
Friederici, A. D., Bahlmann, J., Heim, S., Schubotz, R. I., & Anwander, A. (2006). The brain
differentiates human and non-human grammars: Functional localization and structural
connectivity. Proceedings of the National Academy of Sciences 103(7), 2458–2463.
Friedrich, R., & Friederici, A. D. (2009). Mathematical logic in the human brain: Syntax. PLoS ONE
4(5), e5599.
Friston, K. (2010). The free-energy principle: A unified brain theory? Nature Reviews Neuroscience
11(2), 127–138.
Friston, K., & Kiebel, S. (2009). Predictive coding under the free-energy principle. Philosophical
Transactions of the Royal Society B: Biological Sciences 364(1521), 1211–1221.
Fujioka, T., Trainor, L. J., Ross, B., Kakigi, R., & Pantev, C. (2004). Musical training enhances
automatic encoding of melodic contour and interval structure. Journal of Cognitive Neuroscience
16(6), 1010–1021.
Fujioka, T., Trainor, L. J., Ross, B., Kakigi, R., & Pantev, C. (2005). Automatic encoding of
polyphonic melodies in musicians and nonmusicians. Journal of Cognitive Neuroscience 17(10),
1578–1592.
Furl, N., Kumar, S., Alter, K., Durrant, S., Shawe-Taylor, J., & Griffiths, T. D. (2011). Neural
prediction of higher-order auditory sequence statistics. NeuroImage 54(3), 2267–2277.
Geisler, C. D. (1998). From sound to synapse: Physiology of the mammalian ear. New York: Oxford
University Press.
Grahn, J. A., & Rowe, J. B. (2009). Feeling the beat: Premotor and striatal interactions in musicians
and nonmusicians during beat perception. Journal of Neuroscience 29(23), 7540–7548.
Griffiths, T. D., & Warren, J. D. (2002). The planum temporale as a computational hub. Trends in
Neurosciences 25(7), 348–353.
Griffiths, T. D., & Warren, J. D. (2004). What is an auditory object? Nature Reviews Neuroscience
5(11), 887–892.
Guo, S., & Koelsch, S. (2016). Effects of veridical expectations on syntax processing in music:
Event-related potential evidence. Scientific Reports 6, 19064. doi:10.1038/srep19064
Hackett, T. A., & Kaas, J. (2004). Auditory cortex in primates: Functional subdivisions and
processing streams. In M. S. Gazzaniga (Ed.), The cognitive neurosciences (pp. 215–232).
Cambridge, MA: MIT Press.
Hauser, M. D., Chomsky, N., & Fitch, W. T. (2002). The faculty of language: What is it, who has it,
and how did it evolve? Science 298(5598), 1569–1579.
Hay, J. F., Pelucchi, B., Estes, K. G., & Saffran, J. R. (2011). Linking sounds to meanings: Infant
statistical learning in a natural language. Cognitive Psychology 63(2), 93–106.
Hofstadter, D. R. (1979). Gödel, Escher, Bach. New York: Basic Books.
Huffman, R. F., & Henson, O. W. (1990). The descending auditory pathway and acousticomotor
systems: Connections with the inferior colliculus. Brain Research Reviews 15(3), 295–323.
Hunt, R. H., & Aslin, R. N. (2010). Category induction via distributional analysis: Evidence from a
serial reaction time task. Journal of Memory and Language 62(2), 98–112.
Huron, D. B. (2006). Sweet anticipation: Music and the psychology of expectation. Cambridge, MA:
MIT Press.
Hyde, K. L., Peretz, I., & Zatorre, R. J. (2008). Evidence for the role of the right auditory cortex in
fine pitch resolution. Neuropsychologia 46(2), 632–639.
Jackendoff, R., & Lerdahl, F. (2006). The capacity for music: What is it, and what’s special about it?
Cognition 100(1), 33–72.
Janata, P., & Grafton, S. T. (2003). Swinging in the brain: Shared neural substrates for behaviors
related to sequencing and music. Nature Neuroscience 6(7), 682–687.
Janata, P., Tillmann, B., & Bharucha, J. J. (2002). Listening to polyphonic music recruits domain-
general attention and working memory circuits. Cognitive, Affective, & Behavioral Neuroscience
2(2), 121–140.
Jentschke, S., Friederici, A. D., & Koelsch, S. (2014). Neural correlates of music-syntactic
processing in two-year old children. Developmental Cognitive Neuroscience 9, 200–208.
Jentschke, S., & Koelsch, S. (2009). Musical training modulates the development of syntax
processing in children. NeuroImage 47(2), 735–744.
Jentschke, S., Koelsch, S., Sallat, S., & Friederici, A. D. (2008). Children with specific language
impairment also show impairment of music-syntactic processing. Journal of Cognitive
Neuroscience 20(11), 1940–1951.
Johnsrude, I. S., Penhune, V. B., & Zatorre, R. J. (2000). Functional specificity in the right human
auditory cortex for perceiving pitch direction. Brain 123(1), 155–163.
Kaas, J. H., & Hackett, T. A. (2000). Subdivisions of auditory cortex and processing streams in
primates. Proceedings of the National Academy of Sciences 97(22), 11793–11799.
Kaas, J. H., Hackett, T. A., & Tramo, M. J. (1999). Auditory processing in primate cerebral cortex.
Current Opinion in Neurobiology 9(2), 164–170.
Kandler, K., & Herbert, H. (1991). Auditory projections from the cochlear nucleus to pontine and
mesen-cephalic reticular nuclei in the rat. Brain Research 562(2), 230–242.
Koechlin, E., & Jubault, T. (2006). Broca’s area and the hierarchical organization of human behavior.
Neuron 50(6), 963–974.
Koelsch, S. (2009). Music-syntactic processing and auditory memory: Similarities and differences
between ERAN and MMN. Psychophysiology 46(1), 179–190.
Koelsch, S. (2012). Brain and music. Chichester: Wiley-Blackwell.
Koelsch, S. (2014). Brain correlates of music-evoked emotions. Nature Reviews Neuroscience 15(3),
170–180.
Koelsch, S., Busch, T., Jentschke, S., & Rohrmeier, M. (2016). Under the hood of statistical learning:
A statistical MMN reflects the magnitude of transitional probabilities in auditory sequences.
Scientific Reports 6, 19741. doi:10.1038/srep19741
Koelsch, S., Fritz, T., Schulze, K., Alsop, D., & Schlaug, G. (2005). Adults and children processing
music: An fMRI study. NeuroImage 25(4), 1068–1076.
Koelsch, S., Grossmann, T., Gunter, T. C., Hahne, A., & Friederici, A. D. (2003). Children
processing music: Electric brain responses reveal musical competence and gender differences.
Journal of Cognitive Neuroscience 15(5), 683–693.
Koelsch, S., Gunter, T. C., Cramon, D. Y. von, Zysset, S., Lohmann, G., & Friederici, A. D. (2002).
Bach speaks: A cortical “language-network” serves the processing of music. NeuroImage 17(2),
956–966.
Koelsch, S., Gunter, T. C., Wittfoth, M., & Sammler, D. (2005). Interaction between syntax
processing in language and in music: An ERP study. Journal of Cognitive Neuroscience 17(10),
1565–1577.
Koelsch, S., Rohrmeier, M., Torrecuso, R., & Jentschke, S. (2013). Processing of hierarchical
syntactic structure in music. Proceedings of the National Academy of Sciences 110(38), 15443–
15448.
Koelsch, S., Schröger, E., & Tervaniemi, M. (1999). Superior pre-attentive auditory processing in
musicians. Neuroreport 10(6), 1309–1313.
Koelsch, S., & Siebel, W. A. (2005). Towards a neural basis of music perception. Trends in Cognitive
Sciences 9(12), 578–584.
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11(8), 599–605.
Krumhansl, C. L., & Cuddy, L. L. (2010). A theory of tonal hierarchies in music. Music Perception
36, 51–87.
Lamprea, M. R., Cardenas, F. P., Vianna, D. M., Castilho, V. M., Cruz-Morales, S. E., & Brandão, M.
L. (2002). The distribution of Fos immunoreactivity in rat brain following freezing and escape
responses elicited by electrical stimulation of the inferior colliculus. Brain Research 950(1–2),
186–194.
Langner, G., & Ochse, M. (2006). The neural basis of pitch and harmony in the auditory system.
Musicae Scientiae 10(1), 185.
LeDoux, J. E. (2000). Emotion circuits in the brain. Annual Review of Neuroscience 23, 155–184.
Lehne, M., & Koelsch, S. (2015). Toward a general psychological model of tension and suspense.
Frontiers in Psychology 6. Retrieved from https://doi.org/10.3389/fpsyg.2015.00079
Lehne, M., Rohrmeier, M., & Koelsch, S. (2014). Tension-related activity in the orbitofrontal cortex
and amygdala: An fMRI study with music. Social Cognitive and Affective Neuroscience 9(10),
1515–1523.
Lerdahl, F., & Jackendoff, R. (1983). A generative theory of tonal music. Cambridge, MA: MIT
Press.
Lerdahl, F., & Krumhansl, C. L. (2007). Modeling tonal tension. Music Perception 24(4), 329–366.
Levitin, D. J., & Menon, V. (2005). The neural locus of temporal structure and expectancies in music:
Evidence from functional neuroimaging at 3 tesla. Music Perception: An Interdisciplinary Journal
22(3), 563–575.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in the
area of Broca: An MEG-study. Nature Neuroscience 4(5), 540–545.
Makuuchi, M., Bahlmann, J., Anwander, A., & Friederici, A. D. (2009). Segregating the core
computational faculty of human language from working memory. Proceedings of the National
Academy of Sciences 106(20), 8362–8367.
Malmierca, M. S., Anderson, L. A., & Antunes, F. M. (2015). The cortical modulation of stimulus-
specific adaptation in the auditory midbrain and thalamus: A potential neuronal correlate for
predictive coding. Frontiers in Systems Neuroscience 9, 19. Retrieved from
https://doi.org/10.3389/fnsys.2015.00019
Marcus, G. F., Vijayan, S., Rao, S. B., & Vishton, P. M. (1999). Rule learning by seven-month-old
infants. Science 283(5398), 77–80.
Merchant, H., & Honing, H. (2014). Are non-human primates capable of rhythmic entrainment?
Evidence for the gradual audiomotor evolution hypothesis. Frontiers in Neuroscience 7, 274.
Retrieved from https://doi.org/10.3389/fnins.2013.00274
Merker, B., Morley, I., & Zuidema, W. (2015). Five fundamental constraints on theories of the
origins of music. Philosophical Transactions of the Royal Society B: Biological Sciences
370(1664), 20140095.
Meyer, L. B. (1956). Emotion and meaning in music. Chicago, IL: University of Chicago Press.
Miranda, R. A., & Ullman, M. T. (2007). Double dissociation between rules and memory in music:
An event-related potential study. NeuroImage 38(2), 331–345.
Moore, B. C. J. (2008). An introduction to the psychology of hearing (5th ed.). Bingley: Emerald.
Näätänen, R., Tervaniemi, M., Sussman, E., Paavilainen, P., & Winkler, I. (2001). “Primitive
intelligence” in the auditory cortex. Trends in Neurosciences 24(5), 283–288.
Nevins, A., Pesetsky, D., & Rodrigues, C. (2009). Pirahã exceptionality: A reassessment. Language
85(2), 355–404.
Öngür, D., & Price, J. L. (2000). The organization of networks within the orbital and medial
prefrontal cortex of rats, monkeys and humans. Cerebral Cortex 10(3), 206–219.
Opitz, B., & Kotz, S. A. (2011). Ventral premotor cortex lesions disrupt learning of sequential
grammatical structures. Cortex 48(6), 664–673.
Paavilainen, P., Simola, J., Jaramillo, M., Näätänen, R., & Winkler, I. (2001). Preattentive extraction
of abstract feature conjunctions from auditory stimulation as reflected by the mismatch negativity
(MMN). Psychophysiology 38(2), 359–365.
Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., & Pantev, C. (2012). Statistical learning
effects in musicians and non-musicians: An MEG study. Neuropsychologia 50(2), 341–349.
Parsons, L. (2001). Exploring the functional neuroanatomy of music performance, perception, and
comprehension. Annals of the New York Academy of Sciences 930, 211–231.
Patel, A. D. (2008). Music, language, and the brain. Oxford: Oxford University Press.
Patel, A. D., & Balaban, E. (2001). Human pitch perception is reflected in the timing of stimulus-
related cortical activity. Nature Neuroscience 4(8), 839–844.
Patel, A. D., Iversen, J. R., Wassenaar, M., & Hagoort, P. (2008). Musical syntactic processing in
agrammatic Broca’s aphasia. Aphasiology 22(7), 776–789.
Patterson, R. D., Uppenkamp, S., Johnsrude, I. S., & Griffiths, T. D. (2002). The processing of
temporal pitch and melody information in auditory cortex. Neuron 36(4), 767–776.
Perani, D., Saccuman, M. C., Scifo, P., Spada, D., Andreolli, G., Rovelli, R., … Koelsch, S. (2010).
Functional specializations for music processing in the human newborn brain. Proceedings of the
National Academy of Sciences 107(10), 4758–4763.
Perruchet, P., & Pacton, S. (2006). Implicit learning and statistical learning: One phenomenon, two
approaches. Trends in Cognitive Sciences 10(5), 233–238.
Petkov, C. I., Kayser, C., Augath, M., & Logothetis, N. K. (2006). Functional imaging reveals
numerous fields in the monkey auditory cortex. PLoS Biology 4(7), e215.
Pickles, J. O. (2008). An introduction to the physiology of hearing (3rd ed.). Bingley: Emerald.
Przysinda, E., Zeng, T., Maves, K., Arkin, C., & Loui, P. (2017). Jazz musicians reveal role of
expectancy in human creativity. Brain and Cognition 119, 45–53.
Putkinen, V., Tervaniemi, M., Saarikivi, K., de Vent, N., & Huotilainen, M. (2014). Investigating the
effects of musical training on functional brain development with a novel melodic MMN paradigm.
Neurobiology of Learning and Memory 110, 8–15.
Rammsayer, T., & Altenmüller, E. (2006). Temporal information processing in musicians and
nonmusicians. Music Perception 24(1), 37–48.
Rohrmeier, M. (2011). Towards a generative syntax of tonal harmony. Journal of Mathematics and
Music 5(1), 35–53.
Rohrmeier, M., & Cross, I. (2008). Statistical properties of tonal harmony in Bach’s chorales. In
Ken’ichi Miyazaki, Mayumi Adachi, Yuzuru Hiraga, Yoshitaka Nakajima, and Minoru Tsuzaki
(Eds.), Proceedings of the 10th International Conference on Music Perception and Cognition.
ICMPC (CD-ROM).
Rohrmeier, M., & Rebuschat, P. (2012). Implicit learning and acquisition of music. Topics in
Cognitive Science 4(4), 525–553.
Rohrmeier, M., Zuidema, W., Wiggins, G. A., & Scharff, C. (2015). Principles of structure building
in music, language and animal song. Philosophical Transactions of the Royal Society B: Biological
Sciences 370(1664), 20140097.
Romberg, A. R., & Saffran, J. R. (2010). Statistical learning and language acquisition. Wiley
Interdisciplinary Reviews: Cognitive Science 1(6), 906–914.
Saffran, J. R. (2001). Words in a sea of sounds: The output of infant statistical learning. Cognition
81(2), 149–169.
Saffran, J. R., Aslin, R. N., & Newport, E. L. (1996). Statistical learning by 8-month-old infants.
Science 274(5294), 1926–1928.
Saffran, J. R., Newport, E. L., & Aslin, R. N. (1996). Word segmentation: The role of distributional
cues. Journal of Memory and Language 35(4), 606–621.
Salzer, F. (1962). Structural hearing: Tonal coherence in music (Vol. 1). New York: Dover
Publications.
Schenker, H. (1956). Neue musikalische theorien und phantasien: Der freie satz (2nd ed.). Vienna:
Universal Edition.
Schön, D., & François, C. (2011). Musical expertise and statistical learning of musical and linguistic
structures. Frontiers in Psychology 2, 167. Retrieved from https://doi:10.3389/fpsyg.2011.00167
Schönwiesner, M., Novitski, N., Pakarinen, S., Carlson, S., Tervaniemi, M., & Näätänen, R. (2007).
Heschl’s gyrus, posterior superior temporal gyrus, and mid-ventrolateral prefrontal cortex have
different roles in the detection of acoustic changes. Journal of Neurophysiology 97(3), 2075–2082.
Serafine, M. L., Glassman, N., & Overbeeke, C. (1989). The cognitive reality of hierarchic structure
in music. Music Perception 6(4), 397–430.
Sethares, W. A. (2005). The gamelan. In W. A. Sethares, Tuning, timbre, spectrum, scale (pp. 165–
187). Berlin: Springer.
Sinex, D. G., Guzik, H., Li, H., & Henderson Sabes, J. (2003). Responses of auditory nerve fibers to
harmonic and mistuned complex tones. Hearing Research 182(1–2), 130–139.
Slevc, L. R., Rosenberg, J. C., & Patel, A. D. (2009). Making psycholinguistics musical: Self-paced
reading time evidence for shared processing of linguistic and musical syntax. Psychonomic
Bulletin & Review 16(2), 374–381.
Snyder, J. S., & Elhilali, M. (2017). Recent advances in exploring the neural underpinnings of
auditory scene perception. Annals of the New York Academy of Sciences 1396, 39–55.
Song, J. H., Skoe, E., Wong, P. C. M., & Kraus, N. (2008). Plasticity in the adult human auditory
brainstem following short-term linguistic training. Journal of Cognitive Neuroscience 20(10),
1892–1902.
Steedman, M. J. (1984). A generative grammar for jazz chord sequences. Music Perception 2(1), 52–
77.
Steinbeis, N., & Koelsch, S. (2008). Shared neural resources between music and language indicate
semantic processing of musical tension-resolution patterns. Cerebral Cortex 18(5), 1169–1178.
Strait, D. L., Kraus, N., Skoe, E., & Ashley, R. (2009). Musical experience and neural efficiency:
Effects of training on subcortical processing of vocal expressions of emotion. European Journal of
Neuroscience 29(3), 661–668.
Sun, L., Liu, F., Zhou, L., & Jiang, C. (2018). Musical training modulates the early but not the late
stage of rhythmic syntactic processing. Psychophysiology 55(2), e12983.
Sun, Y., Lu, X., Ho, H. T., Johnson, B. W., Sammler, D., & Thompson, W. F. (2018). Syntactic
processing in music and language: Parallel abnormalities observed in congenital amusia.
NeuroImage: Clinical 19, 640–651.
Sussman, E. S. (2007). A new view on the MMN and attention debate: The role of context in
processing auditory events. Journal of Psychophysiology 21(3), 164–175.
Tarr, B., Launay, J., & Dunbar, R. I. (2014). Music and social bonding: “Self–other” merging and
neurohormonal mechanisms. Frontiers in Psychology 5, 1096. Retrieved from
https://doi.org/10.3389/fpsyg.2014.01096
Teinonen, T., Fellman, V., Näätänen, R., Alku, P., & Huotilainen, M. (2009). Statistical language
learning in neonates revealed by event-related brain potentials. BMC Neuroscience 10(1), 21.
Teinonen, T., & Huotilainen, M. (2012). Implicit segmentation of a stream of syllables based on
transitional probabilities: An MEG study. Journal of Psycholinguistic Research 41(1), 71–82.
Terhardt, E. (1991). Music perception and sensory information acquisition: Relationships and low-
level analogies. Music Perception: An Interdisciplinary Journal 8(3), 217–239.
Tervaniemi, M. (2009). Musicians—same or different? Annals of the New York Academy of Sciences
1169, 151–156.
Tervaniemi, M., Castaneda, A., Knoll, M., & Uther, M. (2006). Sound processing in amateur
musicians and nonmusicians: Event-related potential and behavioral indices. Neuroreport 17(11),
1225–1228.
Tervaniemi, M., Ilvonen, T., Karma, K., Alho, K., & Näätänen, R. (1997). The musical brain: Brain
waves reveal the neurophysiological basis of musicality in human subjects. Neuroscience Letters
226(1), 1–4.
Tervaniemi, M., Janhunen, L., Kruck, S., Putkinen, V., & Huotilainen, M. (2016). Auditory profiles
of classical, jazz, and rock musicians: Genre-specific sensitivity to musical sound features.
Frontiers in Psychology 6, 1900. Retrieved from https://doi.org/10.3389/fpsyg.2015.01900
Tervaniemi, M., Rytkönen, M., Schröger, E., Ilmoniemi, R. J., & Näätänen, R. (2001). Superior
formation of cortical memory traces for melodic patterns in musicians. Learning & Memory 8(5),
295–300.
Thiessen, E. D., Kronstein, A. T., & Hufnagle, D. G. (2013). The extraction and integration
framework: A two-process account of statistical learning. Psychological Bulletin 139(4), 792–814.
Tillmann, B., Bharucha, J., & Bigand, E. (2000). Implicit learning of tonality: A self-organized
approach. Psychological Review 107(4), 885–913.
Tillmann, B., Janata, P., & Bharucha, J. J. (2003). Activation of the inferior frontal cortex in musical
priming. Cognitive Brain Research 16(2), 145–161.
Todd, N. P. M., & Cody, F. W. (2000). Vestibular responses to loud dance music: A physiological
basis of the “rock and roll threshold”? Journal of the Acoustical Society of America 107(1), 496–
500.
Todd, N. P. M., Paillard, A., Kluk, K., Whittle, E., & Colebatch, J. (2014). Vestibular receptors
contribute to cortical auditory evoked potentials. Hearing Research 309, 63–74.
Tramo, M. J., Shah, G. D., & Braida, L. D. (2002). Functional role of auditory cortex in frequency
processing and pitch perception. Journal of Neurophysiology 87(1), 122–139.
Villarreal, E. A. G., Brattico, E., Leino, S., Østergaard, L., & Vuust, P. (2011). Distinct neural
responses to chord violations: A multiple source analysis study. Brain Research 1389, 103–114.
Vuvan, D. T., Zendel, B. R., & Peretz, I. (2018). Random feedback makes listeners tone-deaf.
Scientific Reports 8(1), 7283.
Warren, J. D., Uppenkamp, S., Patterson, R. D., & Griffiths, T. D. (2003). Separating pitch chroma
and pitch height in the human brain. Proceedings of the National Academy of Sciences 100(17),
10038–10042.
Whitfield, I. (1980). Auditory cortex and the pitch of complex tones. Journal of the Acoustical
Society of America 67(2), 644–647.
Winkler, I. (2007). Interpreting the mismatch negativity. Journal of Psychophysiology 21(3–4), 147–
163.
Wong, P. C. M., Skoe, E., Russo, N. M., Dees, T., & Kraus, N. (2007). Musical experience shapes
human brainstem encoding of linguistic pitch patterns. Nature Neuroscience 10(4), 420–422.
Zatorre, R. J. (1988). Pitch perception of complex tones and human temporal-lobe function. Journal
of the Acoustic Society of America 84, 566–572.
Zatorre, R. J. (2001). Neural specializations for tonal processing. Annals of the New York Academy of
Sciences 930, 193–210.
Zatorre, R. J., Belin, P., & Penhune, V. B. (2002). Structure and function of auditory cortex: Music
and speech. Trends in Cognitive Sciences 6(1), 37–46.
Zendel, B. R., Lagrois, M.-É., Robitaille, N., & Peretz, I. (2015). Attending to pitch information
inhibits processing of pitch information: The curious case of amusia. Journal of Neuroscience
35(9), 3815–3824.
Zhang, J., Zhou, X., Chang, R., & Yang, Y. (2018). Effects of global and local contexts on chord
processing: An ERP study. Neuropsychologia 109, 149–154.
Zuijen, T. L. von, Sussman, E., Winkler, I., Näätänen, R., & Tervaniemi, M. (2004). Grouping of
sequential sounds: An event-related potential study comparing musicians and nonmusicians.
Journal of Cognitive Neuroscience 16(2), 331–338.
Zuijen, T. L. von, Sussman, E., Winkler, I., Näätänen, R., & Tervaniemi, M. (2005). Auditory
organization of sound sequences by a temporal or numerical regularity: A mismatch negativity
study comparing musicians and non-musicians. Cognitive Brain Research 23(2–3), 270–276.
1
For similar results obtained from patients with (right) PAC lesions see Johnsrude, Penhune, and
Zatorre (2000) and Zatorre (2001).
2
Note that a finite state automaton will only (mis)understand that “Peter kissed Mary”!
CHAPT E R 10
M U LT I S E N S O RY
PROCESSING IN MUSIC
FRANK RUSSO
M P
It has often been noted that speech is perceived by eye and by ear. This is
normally characterized as an opportunity to minimize uncertainty as it
allows the brain to capitalize on convergences. However, it can also
represent a sensory-processing challenge in that information from across
two channels must somehow be bound together into a common
representation. This challenge may be even greater in music given the
additional channels of sensory information that are routinely involved and
the intentional use of uncertainty as a compositional device. Nevertheless,
under most conditions, multisensory information in music is successfully
integrated yielding a coherent and stable multisensory percept.
Information from across the senses may be integrated in a manner that is
cognitive or perceptual (Schutz, 2008). Cognitive integration takes place
after information from two or more channels has been processed
independently (see review and meta-analysis concerning audio-visual music
by Platz & Kopiez, 2012). A classic musical example of this type of
integration is the influence of performer attractiveness on judgments of
performance quality (Wapnick, Mazza, & Darrow, 2000). In this example,
information from one channel does not so much alter perception in another
as much as influence how those perceptions are evaluated.
Another musical example that reflects cognitive multisensory integration
concerns the “blue note” in live jazz and blues. Blue notes are often
accompanied by a visual display that conveys negatively valenced emotion
(e.g., wincing of the eyes, shaking, or rolling the head back). Thompson,
Graham, and Russo (2005) sought to assess the effect of this practice by
using twenty clips of a blues concert performed by B.B. King. Although all
of the selected clips possessed some level of dissonance, half were
performed with a relatively neutral facial expression. Two groups of
participants were asked to provide judgments of dissonance. One group
made judgments in an auditory-only condition and the other made
judgments in an auditory-visual condition. Results revealed that visual
information influenced judgments of dissonance, such that the difference
between dissonant and neutral performances was greater in the audio-visual
condition. However, it would be erroneous to conclude that information
from the visual and auditory channel had been integrated at the level of
perceptual representation.
Integration at the perceptual level is said to take place when information
from across the senses is integrated in a manner that is automatic and pre-
attentive (Arieh & Marks, 2008; Spence, 2011). All of the multisensory
examples considered in the rest of this chapter meet these simple criteria.
However, the neural mechanisms allowing for perceptual integration are by
no means uniform. To foreshadow, there are at least three main types of
mechanisms that have been implicated. The mechanisms vary with respect
to network size but all involve some form of direct or indirect
communication between primary sensory areas of the brain (see Fig. 1).
FIGURE 1. Schematic diagram of brain circuitry underpinning three mechanisms of multisensory
integration (STS = Superior Temporal Sulcus; IFG = Inferior Frontal Gyrus; S = Somatosensory
Cortex; A = Auditory Cortex; V = Visual Cortex). Top panel diagrams first mechanism involving
primary sensory areas only. Second panel diagrams second mechanism involving the first
mechanism in addition to a known multisensory area, superior-temporal sulcus (STS). Bottom panel
diagrams third mechanism that may be described as sensorimotor. It builds on the second
mechanism adding feedback connections from a known motor planning area, inferior frontal gyrus
(IFG). Subcortical contributions from the superior colliculus not diagrammed.
Pathways
Much like the “what” (ventral) and “where” (dorsal) visual pathways
originally proposed to explain functional organization in the visual system
(Goodale & Milner, 1992), there are two main auditory pathways leading
out of auditory cortex and terminating in frontal areas (Zatorre, Chen, &
Penhune, 2007). A ventral auditory pathway is thought to be involved
primarily with category-based representations (e.g., phonemes). A dorsal
“auditory-motor” pathway is thought to be specialized for sensorimotor
translations of time-varying information that is not categorical. This
pathway may be particularly important in the context of learning a new
piece of music (Lahav, Saltzman, & Schlaug, 2007; Schalles & Pineda,
2015), perceiving emotion in music (McGarry, Pineda, & Russo, 2015;
Thompson et al., 2005; Vines, Krumhansl, Wanderley, Dalca, & Levitin,
2011), and in the type of feedback monitoring required for performance,
particularly in continuous pitch instruments like voice or violin (Loui, 2015;
Zatorre et al., 2007). The auditory-motor pathway involves reciprocal
connections between inferior frontal gyrus and posterior subdivisions of the
superior temporal gyrus (auditory parabelt) and superior temporal sulcus
(multisensory area).
M P P
Visuomotor Influences
Numerous studies have demonstrated that the size of a sung melodic
interval can be judged directly through the visual system. When videos of
sung melodic intervals are presented to observers without audio, they are
able to accurately scale them according to size (Thompson & Russo, 2007).
This ability does not appear to require music or vocal training, which argues
against a cognitive account based on long-term memory associations, and
further suggests that some aspects of the visual information provide reliable
cues for judging interval size. Video-based tracking has shown that larger
intervals possess more head movement, eyebrow raising, and mouth
opening. The influence of visual information on perception of size in sung
melodic intervals persists even under point-light presentation conditions in
which the dynamic information in the display is retained while eliminating
static visual cues (Abel, Li, Russo, Schlaug, & Loui, 2016).
The visual channel continues to influence the perceived size of sung
melodic intervals even when audio is present (Russo, Sandstrom, &
Maksimowski, 2011; Thompson et al., 2005; Thompson, Russo, &
Livingstone, 2010). The mouth area may be particularly important in
judging the size of sung melodic intervals as reducing the level of audibility
in an audio-visual presentation (by increasing level of background noise)
causes observers to increase the proportion gaze directed toward the mouth
(Russo et al., 2011). However, the visual influence on auditory judgments
has been found to be mitigated for participants with a young onset of
musical training (Abel et al., 2016). One interpretation of this finding is that
early-trained musicians possess a stronger audio-motor representation of
sung melodic intervals. This enhancement in priors may allow them to
focus on auditory input or rely less heavily on non-auditory input when
presented with multisensory musical stimuli. This prioritization may be
further reinforced through experience playing in groups where orthogonal
streams of audio and visual information may co-exist. But how can we be
sure that vision influences melodic pitch processing at a perceptual (vs.
cognitive) level?
One behavioral means of assessing whether multisensory integration is
perceptual is to utilize a dual-task paradigm. Thompson et al. (2010)
presented participants with sung melodic intervals accompanied by facial
expressions used to perform a small or large interval (two and nine
semitones, respectively). Participants were asked to count the number of 0’s
and 1’s that were superimposed over the singer’s face during performance
of each interval. The conditions were blocked by digit speed (300 or 700
msec per digit) as well as task demand (single or dual task). Results
revealed that the influence of the visual information on auditory judgments
of sung melodic interval size was not moderated by cognitive load. These
findings suggest that the integration was automatic and pre-attentive.
The cortical underpinnings of this example of multisensory integration
in music may originate in motion selective areas of the dorsal visual
pathway, such as the medial temporal and the medial superior temporal
areas. Both of these areas are adjacent to the posterior bank of the superior
temporal sulcus, a known multisensory area that projects to premotor areas
allowing for sensorimotor translations of dynamic sensory input (Kilner,
2011). There are also reciprocal connections from premotor to superior
temporal sulcus allowing for a predictive coding model of action involving
sensory representations (Kilner et al., 2007). This type of predictive coding
may be particularly important in shaping auditory judgments of action on
the basis of visual input alone or in situations where auditory input is
ambiguous for some reason (e.g., an individual with severe hearing loss or
an individual with normal hearing listening in low signal-to-noise
conditions).
The mechanism proposed for visual perception and audio-visual
integration of melodic pitch information involves feedforward and feedback
connections along the dorsal stream. Feedforward connections provide
multisensory input to motor planning areas. Feedback connections provide
predictive coding of movement informed by priors that can be compared
with incoming sensory information (Kilner et al., 2007; Maes, Leman,
Palmer, & Wanderley, 2014). In the case of individuals with severe hearing
loss, there may also be an additional contribution owed to visual activation
of the auditory cortex in belt areas (Finney, 2001; Röder, Stock, Bien,
Neville, & Rösler, 2002). Research in animal models suggests that belt
areas undergo profound plastic changes following a period of auditory
deprivation, which leads in some cases to enhanced visual processing.
Lomber, Meredith, and Kral (2010) showed that deactivation of posterior
belt areas selectively eliminates enhancements to visual localization,
whereas deactivation of the dorsal belt areas eliminates enhancement of
visual motion detection.
Transcranial magnetic stimulation (TMS) has been used as one means of
investigating the assumed involvement of motor areas in processing sung
melodic intervals (Royal, Lidji, Théoret, Russo, & Peretz, 2015). Non-
musicians were given brief training that enabled them to apply a label to
intervals of different size (e.g., unison, octave, etc.). Following training,
facilitative TMS was applied over motor cortex, while participants observed
a pitch interval label that was immediately followed by the audio-visual
presentation of a sung interval. Participants were required to make a forced-
choice judgment regarding whether the pitch interval label matched the
pitch interval contained in the two-note vocal melody. Motor-evoked
potentials recorded from the mouth muscles contralateral to the hemisphere
receiving stimulation were found to increase relative to baseline for large
pitch intervals and decrease for small pitch intervals, suggesting that some
type of motor simulation was taking place.
Another line of evidence in support of motor involvement in perception
of song may be found in EEG research investigating the sensorimotor (or
mu) wave. The oscillatory generators of the sensorimotor wave can be
found in the inferior frontal gyrus, and to a lesser extent in the inferior
parietal lobe. The sensorimotor wave becomes desynchronized when an
individual moves intentionally or when they observe others moving
intentionally, and the extent of desynchronization is enhanced under
multisensory presentation conditions (Kaplan & Iacoboni, 2007; McGarry,
Russo, Schalles, & Pineda, 2012). These data have been interpreted as
evidence of an internal simulation involving motor planning and
proprioception. While some controversy exists regarding the putative mirror
system responsible for the sensorimotor wave (Hickok, 2009), its
responsiveness to observation of intentional action is less equivocal. A
meta-analysis by Fox et al. (2016), involving eighty-five studies, found
significant event-related desynchronization during observation of
intentional action (Cohen’s d = 0.31, N = 1,508). With regard to music
stimuli, evidence has been found for sensorimotor desynchronization in
response to audio-only presentations of isolated sung notes (Lévêque &
Schön, 2013) and audio-visual presentations of sung melodic intervals
(McGarry et al., 2015). Although it seems likely, it remains to be
determined whether sensorimotor desynchronization in response to song is
greater in multisensory compared to unisensory presentation conditions.
Somatosensory Influences
Because all sound arises from a source of mechanical vibration, it should be
no surprise that evidence exists for perception of pitch and other musical
dimensions on the basis of vibrotactile input (i.e., mechanical vibration of
the skin). Detection thresholds for vibrotactile stimuli show peak sensitivity
around 250 Hz, and a sharp decline in sensitivity (i.e., larger thresholds)
below 100 Hz (Hopkins, Maté-Cid, Fulford, Seiffert, & Ginsborg, 2016;
Morioka & Griffin, 2005; Verrillo, 1992). Thresholds are also smaller in
smooth (vs. hairy) skin due to increased mechanoreceptor density (Verrillo
& Bolanowski, 1986), and with large (vs. small) contactor areas due to
effects of spatial summation (Morioka & Griffin, 2005). Pitch
discrimination thresholds obtained with vibrotactile stimuli tend to be about
five times greater than those obtained with auditory stimuli (Branje,
Maksimowski, Karam, Fels, & Russo, 2010; Verrillo, 1992). In addition to
this relatively poor pitch discrimination ability, there is no convincing
psychophysical evidence for vibrotactile pitch discriminations beyond about
1,000 Hz.
Single cell recording in macaques has revealed that low-frequency
vibrotactile stimuli can activate belt areas of auditory cortex (Schroeder et
al., 2001). Convergent evidence has been found in imaging studies
involving adults with normal hearing. Low-frequency vibrotactile stimuli
has been shown to activate auditory cortex bilaterally (Levänen, Jousmäki,
& Hari, 1998), particularly in posterior belt areas (Schürmann, Caetano,
Hlushchuk, Jousmäki, & Hari, 2006). The extent of auditory activations
observed in deaf participants is more widespread than that observed in
normal hearing participants (Auer, Bernstein, Sungkarat, & Singh, 2007),
likely due to neuroplastic changes following sensory deprivation. One
question resulting from this work is whether activation of auditory areas by
vibrotactile stimuli is direct or whether it is the result of projections from
somatosensory areas.
Using MEG, Caetano and Jousmäki (2006) were able to track the time
course of vibrotactile activations. They presented normal hearing
participants with 200 Hz vibrotactile stimuli delivered to the fingertips. An
initial response was observed in somatosensory cortex, peaking around 60
ms, followed by transient auditory responses in auditory and secondary
somatosensory cortices between 100 and 200 ms. Finally, a sustained
response was observed in auditory cortex between 200 and 700 ms.
Although these studies all present unisensory stimuli, taken together, these
findings suggest a likely mechanism for audio-tactile integration that is
hierarchical involving a progressive convergence of auditory and
somatosensory pathways. One of the main areas of sensory convergence in
the cortex appears to be the posterior subdivisions of the auditory parabelt
and the superior temporal sulcus (see Fig. 2).
FIGURE 2. Schematic sagittal view of the human brain featuring modules and pathways that are
involved in the multisensory perception of music.
M P T
Visuomotor Influences
Saldaña and Rosenblum (1993) presented participants with audio-visual
presentations of cello tones where bowing and plucking was crossed across
the senses. So, for example, observers were presented with a multisensory
stimulus in which the audio channel consisted of an unequivocal plucking
sound and the visual channel presented an unequivocal bowing movement.
Much like the “McGurk effect” upon which this study is based (McGurk &
Macdonald, 1976), auditory judgments were influenced by visual
information. For instance, plucking sounds were more likely to be heard as
bowing when accompanied by bowing visual movement. The authors
interpreted their results with regard to an automatic internal motor
simulation that is driven by auditory and visual information. Much like the
explanation for sung melodic pitch, an internal motor simulation may have
provided a predictive coding model of action involving sensory
representations. The output of the predictive coding model may have been
integrated with direct auditory input at the level of the superior temporal
sulcus. Consistent with this interpretation, fMRI work involving
multisensory speech has consistently implicated the superior temporal
sulcus and superior temporal gyrus (Callan et al., 2003, 2004; Jones &
Callan, 2003). Similar evidence has been found with multisensory tool use
and the extent of activation in the superior temporal sulcus appears to
adhere to the law of inverse effectiveness (Stevenson & James, 2009).
Somatosensory Influences
Several studies have investigated the ability to discriminate timbre using
vibrotactile stimuli. Russo, Ammirante, and Fels (2012) found that deaf and
hearing observers were able to accurately distinguish instrument timbres on
the basis of vibrotactile information alone. Deaf and hearing participants
were also able to distinguish timbre on the basis of synthetic tones that
differed only with regard to spectral envelope (dull vs. bright). This ability
persisted even though numerous controls were put in place to ensure that
participants received no trace of residual auditory input. Russo et al. (2012)
proposed that vibrotactile discrimination involves the cortical integration of
spectral information filtered through frequency-tuned mechanoreceptors.
There are four known channels that respond to touch (Bolanowski,
Gescheider, Verrillo, & Checkosky, 1988), and each is sensitive to a unique
range of the frequency spectrum. This allows the mechanoreceptors to
collectively code for spectral shape in the same way that has been proposed
for critical bands in the auditory system (Makous, Friedman, & Vierck,
1995). It would only take two such channels to allow for the coding of
spectral tilt. A follow-up study revealed that deaf participants are able to
discriminate sung vowels and that extent of difference in spectral tilt
between pairs strongly predicted their discriminability (Ammirante, Russo,
Good, & Fels, 2013).
In addition to the influence of vibrotactile stimulation on passive
reception of timbre, it seems likely that such stimulation provides
performers with valuable timbre information during active performance
(Marshall & Wanderley, 2011). As an example, the string vibrations of a
piano are detectable at the level of the key press. Vibration detection
thresholds are reduced under natural playing conditions involving active
touch (Papetti, Jarvelainen, Giordano, Schiesser, & Frohlich, 2017) and the
co-occurrence of sound at the same frequency (Ro, Hsu, Yasar, Caitlin
Elmore, & Beauchamp, 2009). Perhaps not surprisingly, the perception of
sound quality as evaluated by the performer has been shown to be
influenced by vibration that is felt through the keys (Fontana, Papetti,
Järveläinen, & Avanzini, 2017).
To date, there have been no neural studies investigating auditory-tactile
perception of timbre. However, it seems likely that this ability would
depend on direct projections from somatosensory cortex to posterior belt
areas of auditory cortex (see top panel of Fig. 1). These direct projections
are likely to be right lateralized because of thinner myelin sheathing in the
right auditory cortex (Anderson et al., 1999), which may better support
communication across frequency channels, thus enabling spectral analysis.
M P R
Visuomotor Influences
Rhythm involves the metrical patterning and grouping of tones that is
shaped by intensity and duration. Visual influences have been found to
affect the ability to track rhythm as well as the low-level dimensions that
contribute to rhythm (e.g., loudness and duration). Because percussionists
do not have the ability to independently control the intensity and duration of
the notes that they produce, the use of gestures may be particularly
important in shaping these dimensions (Schutz, 2008). Rosenblum and
Fowler (1991) recorded handclaps of varying intensity. They presented
participants with audio-visual pairings of the handclaps that were either
congruent or incongruent. Although participants were asked to base
loudness judgments only on what they heard, the visual information
presented had a systematic influence on loudness judgments.
Schutz and colleagues have shown that expressive gestures are also able
to influence the duration of a performed note. Their initial study utilized
recordings of notes performed on a marimba with “long” and “short”
gestures (Schutz & Lipscomb, 2007). Audio and visual channels were
recombined to form congruent and incongruent audio-visual pairings. These
pairings were presented to listeners and they were asked to make duration
estimations on the basis of sound alone. Although the auditory content of
the recordings had no effect on estimations of duration, the visual
presentation influenced perceived duration such that long gestures
lengthened notes and short gestures shortened notes. This effect persisted
even when visual content was substituted with a point-light display,
suggesting that the effect was based on the dynamics of visual movement
(Schutz & Kubovy, 2009).
The ability to synchronize to metrical structures created by discrete
visual flashes has been found to be inferior to synchronization with discrete
auditory tones that have the same temporal characteristics (Patel, Iversen,
Chen, & Repp, 2005). However, the auditory advantage is almost
eliminated if visual rhythms are presented using continuous stimuli such as
a bouncing ball (Grahn, 2012; Hove, Fairhurst, Kotz, & Keller, 2013;
Iversen, Patel, Nicodemus, & Emmorey, 2015). Imaging results have shown
that activation in the putamen, a key timing area involved in motor planning
and beat perception (Grahn & Brett, 2007), parallels results obtained with
sensorimotor synchronization tasks. In particular, continuous visual stimuli
led to greater activation of the putamen than did visual flashes, approaching
activation levels obtained with auditory beeps. This finding suggests that
the ability to synchronize to metrical structure is not simply contingent on
the channel of sensory input but also on the nature of stimulus presentation
(Grahn, 2012; Hove et al., 2013; Ross, Iversen, & Balasubramaniam, 2016).
While discrete events are optimal with auditory stimuli, continuous events
lead to better outcomes with visual stimuli. Some evidence suggests that the
deaf possess some advantage in tracking visual rhythms (Iversen et al.,
2015). The latter finding may be owed to neuroplastic changes resulting
from sensory deprivation and life-long experience with signing (Bavelier et
al., 2000, 2001). Referring back to Fig. 1, the strength of direct visual input
to auditory-motor pathways is likely enhanced in deaf individuals.
Many studies have used EEG to assess neural entrainment to the beat.
When the frequency of the beat is within the range of human movement
(e.g., 1 to 4 Hz), large swathes of cortex entrain to that frequency. These
neural oscillations will persist even after a rhythmic stimulus has been
temporarily paused. Depending on when the rhythmic stimulus is resumed,
the entrained neural oscillations will either increase or decrease in power
(Simon & Wallace, 2017). Power decreases when the rhythmic stimulus
anticipates the beat (too early) and it increases when the rhythmic stimulus
is resumed on the beat (on time). However, if the beat is resumed as an
audio-visual event, there is no modulation of power in the entrained neural
oscillations. These findings reveal that multisensory inputs are not
equivalent to auditory inputs with respect to entrainment. One interpretation
is that multisensory input is “highly reliable or salient” and that resources
should be allocated to processing it independently from the oscillations
manifesting from the original auditory-only beat. This pattern of neural
findings may also help to explain results from sensorimotor synchronization
studies revealing superior synchronization using multisensory rhythms
compared with auditory-only rhythms (Elliott, Wing, & Welchman, 2010;
Varlet, Marin, Issartel, Schmidt, & Bardy, 2012).
Although visual influences on the perception of rhythm can be powerful,
it is important to acknowledge that many listeners will choose to listen with
their eyes closed under challenging conditions. One interpretation of this
phenomenon is that the visual information is somehow distracting. In a task
involving temporal order judgments of varying complexity, researchers
found progressively greater deactivation of visual cortical areas as temporal
asynchronies approached discrimination thresholds (Hairston et al., 2008).
This finding is perhaps best understood from the perspective of the inverse
effectiveness rule (Stein & Meredith, 1993), whereby deactivation of the
visual cortex protects against integration of potentially aberrant timing
information from the visual system in a task that is well handled by
audition.
Somatosensory Influences
Some evidence exists for the somatosensory system contributing to the
perception of rhythm. Tranchant et al. (2017) asked deaf and hearing
participants to synchronize movements to a vibrotactile beat delivered
through a vibrating platform. Hearing participants were also asked to
synchronize movements to the same beat delivered through audition and
without vibrotactile stimulation. Results revealed that most participants
were able to synchronize to the vibrotactile beat with no differences
between groups. However, for hearing participants, synchronization
performance was better in the auditory condition than in the vibrotactile
condition.
Other studies have demonstrated that sensorimotor synchronization to a
beat is possible using vibrotactile stimulation applied to the fingertip
(Brochard, Touzalin, Després, & Dufour, 2008; Elliott et al., 2010), toe
(Müller et al., 2008), or to the back (Ammirante, Patel, & Russo, 2016).
Findings have revealed that synchronization to a simple (metronomic)
vibrotactile beat can be as accurate as synchronization to an auditory beat
but only under certain conditions. For example, Müller et al. (2008) found
equivalence on the fingertip but not the toe and Ammirante et al. (2016)
found equivalence on the back, but only when a large portion of the back
was stimulated. Presumably, spatial summation (involving integration of
information across receptors), improved the somatosensory response to
rhythmic information (Gescheider, Bolanowski, Pope, & Verrillo, 2002).
Ammirante et al. (2016) also included an audio-tactile condition to
investigate multisensory integration. Results indicated that sensorimotor
synchronization to audio was consistently equivalent to auditory-tactile,
regardless of contactors size. These results may be interpreted with respect
to the maximum likelihood estimation model (Ernst & Banks, 2002), where
auditory information represents a highly reliable cue that is resistant to
integration with information from a somewhat less reliable channel of
sensory input (vibrotactile).
The results of Ammirante et al. (2016) may also be considered with
respect to sensorimotor models of perception (Fig. 1, Panel 3). The Action
Simulation for Auditory Perception (ASAP) model suggests that our ability
to find the beat in rhythm is based on an internal simulation of periodic
motor activity (Patel & Iversen, 2014). A secondary hypothesis posited in
the model is that beat perception evolved from mechanisms required for
verbal communication, as both involve periodic timing and the integration
of motor and auditory information. This hypothesis is supported in part by
the observation that beat synchronization exists robustly in vocal-learning
species that are only distally related to humans (e.g., parrots and elephants)
and not at all in non-human primates (Merchant, Grahn, Trainor,
Rohrmeier, & Fitch, 2015).
As vocal communication is primarily based in the auditory modality it
follows that cognitive and neurological timing mechanisms would show a
preference for auditory stimuli. Again, this prediction is confirmed by
evidence demonstrating that sensorimotor synchronization to auditory
stimuli tends to be superior to sensorimotor synchronization to visual or
vibrotactile stimuli. Current research in my lab led by Sean Gilmore is
using EEG and source analysis to investigate the extent to which neural
entrainment to the beat is possible under audio-only, vibrotactile-only, and
audio-vibrotactile stimuli. On the basis of the behavioral results of
Ammirante et al. (2016), we expect to find that neural entrainment in motor
planning areas will be weakest for vibrotactile stimuli and that no
differences will exist between audio and audio-tactile conditions.
Movement-Based Influences
Both passive and active head movements are capable of stimulating the
vestibular system (Cullen & Roy, 2004). Given that people actively move
their heads while listening to music it would seem that vestibular
stimulation is commonplace in music listening. Moreover, given that
vestibular cortex is extensively connected with other sensory systems it
stands to reason that there are ample opportunities for multisensory
integration in music that involve the vestibular system. Phillips-Silver &
Trainor (2005) assessed the contribution of the vestibular system to
multisensory rhythm using an ambiguous auditory rhythm. These rhythms
can be encoded in duple form (a march) or in triple form (a waltz). The
rhythms were presented to infants while they were bounced on every
second or every third beat. On the basis of a head-turn preference
procedure, researchers were able to conclude that when infants were
bounced on every second beat, they were coding the ambiguous rhythm in
duple form, and when they were bounced on every third beat they coded the
rhythm in triple form. A follow-up experiment in the same study showed
that blindfolding infants mitigated but did not eliminate the effect, which
confirms that this example of multisensory integration in rhythm does not
depend on visual perception.
Two other studies by Trainor and colleagues have confirmed that these
effects of auditory-vestibular integration in music persist into adulthood. In
one study, adults were trained to bounce in duple or triple time while
listening to an ambiguous rhythm. A subsequent listening test showed that
adults identified an auditory version of the rhythm pattern with accented
beats that matched their bouncing experience as more similar than a version
whose accents did not match (Phillips-Silver & Trainor, 2007). Because this
study involved self-motion it was not able to separate out the contributions
of vestibular and proprioceptive cues. However, a follow-up study
involving direct galvanic stimulation of the vestibular system was able to
provide evidence that auditory and vestibular information are integrated in
rhythm perception in adults even in the absence of movement. In single cell
recordings involving animal models, the posterior parietal cortex appears to
be a likely locus of multisensory integration involving vestibular input
(Bremmer, Schlack, Duhamel, Graf, & Fink, 2001). This area happens to be
proximal to other cortical areas that have been implicated as contributing to
multisensory processing (i.e., posterior superior temporal gyrus, auditory
parabelt, and medial temporal areas).
Other researchers have considered the consequences of multisensory
integration resulting from moving to the beat. Manning & Schutz (2013)
had participants move or simply listen to an isochronous beat. A final tone
was presented following a brief pause and participants were asked whether
it was consistent with the timing of the preceding sequence. Accuracy in
this timing task was superior in the movement condition. In a follow-up
study, it was found that the accuracy gains in this timing task are greater in
percussionists than in non-percussionists, suggesting a role for experience
with moving to the beat (Manning & Schutz, 2016). It seems likely that the
multisensory timing cues resulting from moving to the beat would lead to
stronger neural entrainment to the beat. Indeed, EEG research involving an
ambiguous rhythm has shown that entrainment is stronger after participants
have been trained to move to the rhythm in a way that suggests a binary or
ternary form (Chemin, Mouraux, & Nozaradan, 2014). In addition, the
entrainment gains were detectable at frequencies related to the meter of
movement.
S C
R
Abel, M. K., Li, H. C., Russo, F. A., Schlaug, G., & Loui, P. (2016). Audiovisual interval size
estimation is associated with early musical training. PLoS ONE 11(10), 1–12.
Alais, D., & Burr, D. (2004). Ventriloquist effect results from near-optimal bimodal integration.
Current Biology 14(3), 257–262.
Ammirante, P., Patel, A. D., & Russo, F. A. (2016). Synchronizing to auditory and tactile
metronomes: A test of the auditory-motor enhancement hypothesis. Psychonomic Bulletin &
Review 23(6), 1882–1890.
Ammirante, P., Russo, F. A., Good, A., & Fels, D. I. (2013). Feeling voices. PloS ONE 8(1), 1–5.
Anderson, B., Southern, B. D., & Powers, R. E. (1999). Anatomic asymmetries of the posterior
superior temporal lobes: A postmortem study. Neuropsychiatry Neuropsychology, and Behavioral
Neurology 12(4), 247–254.
Angelaki, D. E., Gu, Y., & DeAngelis, G. C. (2009). Multisensory integration: Psychophysics,
neurophysiology, and computation. Current Opinion in Neurobiology 19(4), 452–458.
Arieh, Y., & Marks, L. E. (2008). Cross-modal interaction between vision and hearing: A speed-
accuracy analysis. Perception & Psychophysics 70(3), 412–421.
Auer, E. T., Bernstein, L. E., Sungkarat, W., & Singh, M. (2007). Vibrotactile activation of the
auditory cortices in deaf versus hearing adults. Neuroreport 18(7), 645–648.
Bavelier, D., Brozinsky, C., Tomann, A., Mitchell, T., Neville, H., & Liu, G. (2001). Impact of early
deafness and early exposure to sign language on the cerebral organization for motion processing.
Journal of Neuroscience 21(22), 8931–8942.
Bavelier, D., Tomann, A., Hutton, C., Mitchell, T., Corina, D., Liu, G., & Neville, H. (2000). Visual
attention to the periphery is enhanced in congenitally deaf individuals. Journal of Neuroscience
20(17), RC93.
Beauchamp, M. S., Yasar, N. E., Frye, R. E., & Ro, T. (2008). Touch, sound and vision in human
superior temporal sulcus. NeuroImage 41(3), 1011–1020.
Bolanowski, S. J., Gescheider, G. A., Verrillo, R. T., & Checkosky, C. M. (1988). Four channels
mediate the mechanical aspects of touch. Journal of the Acoustical Society of America 84(5),
1680–1694.
Branje, C., Maksimowski, M., Karam, M., Fels, D. I., & Russo, F. A. (2010). Vibrotactile display of
music on the human back. Proceedings of the 3rd International Conference on Advances in
Computer–Human Interactions, ACHI 2010 (pp. 154–159). Retrieved from
https://doi.org/10.1109/ACHI.2010.40
Bremmer, F., Schlack, A., Duhamel, J. R., Graf, W., & Fink, G. R. (2001). Space coding in primate
posterior parietal cortex. NeuroImage 14(1), S46–S51.
Brochard, R., Touzalin, P., Després, O., & Dufour, A. (2008). Evidence of beat perception via purely
tactile stimulation. Brain Research 1223, 59–64.
Caetano, G., & Jousmäki, V. (2006). Evidence of vibrotactile input to human auditory cortex.
NeuroImage 29(1), 15–28.
Callan, D. E., Jones, J. A., Munhall, K., Callan, A. M., Kroos, C., & Vatikiotis-Bateson, E. (2003).
Neural processes underlying perceptual enhancement by visual speech gestures. Neuroreport
14(17), 2213–2218.
Callan, D. E., Jones, J. A., Munhall, K., Kroos, C., Callan, A. M., & Vatikiotis-Bateson, E. (2004).
Multisensory integration sites identified by perception of spatial wavelet filtered visual speech
gesture information. Journal of Cognitive Neuroscience 16(5), 805–816.
Calvert, G. A., Bullmore, E. T., Brammer, M. J., Campbell, R., Williams, S. C. R., McGuire, P. K., …
David, A. S. (1997). Activation of auditory cortex during silent lipreading. Science 276(5312),
593–596.
Chemin, B., Mouraux, A., & Nozaradan, S. (2014). Body movement selectively shapes the neural
representation of musical rhythms. Psychological Science 25(12), 2147–2159.
Cullen, K. E., & Roy, J. E. (2004). Signal processing in the vestibular system during active versus
passive head movements. Journal of Neurophysiology 91(5), 1919–1933.
Elliott, M. T., Wing, A. M., & Welchman, A. E. (2010). Multisensory cues improve sensorimotor
synchronisation. European Journal of Neuroscience 31(10), 1828–1835.
Ernst, M. O., & Banks, M. S. (2002). Humans integrate visual and haptic information in a
statistically optimal fashion. Nature 415(6870), 429–433.
Ernst, M. O., & Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends in Cognitive
Sciences 8(4), 162–169.
Finney, E. M. F. (2001). Visual stimuli activate auditory cortex in the deaf. Nature Neuroscience
4(12), 1171–1173.
Fontana, F., Papetti, S., Järveläinen, H., & Avanzini, F. (2017). Detection of keyboard vibrations and
effects on perceived piano quality. Journal of the Acoustical Society of America 142(5), 2953–
2967.
Formisano, E., Kim, D. S., Di Salle, F., Van De Moortele, P. F., Ugurbil, K., & Goebel, R. (2003).
Mirror-symmetric tonotopic maps in human primary auditory cortex. Neuron 40(4), 859–869.
Fox, N. A., Yoo, K. H., Bowman, L. C., Cannon, E. N., Ferrari, P. F., Bakermans-Kranenburg, M. J.,
… Van IJzendoorn, M. H. (2016). Assessing human mirror activity with EEG mu rhythm: A meta-
analysis. Psychological Bulletin 142(3), 291–313.
Foxe, J. J., & Schroeder, C. E. (2005). The case for feedforward multisensory convergence during
early cortical processing. Neuroreport 16(5), 419–423.
Foxe, J. J., Wylie, G. R., Martinez, A., Schroeder, C. E., Javitt, D. C., Guilfoyle, D., … Murray, M.
M. (2002). Auditory-somatosensory multisensory processing in auditory association cortex: An
fMRI study. Journal of Neurophysiology 88(1), 540–543.
Frith, C. D., & Hasson, U. (2016). Mirroring and beyond: Coupled dynamics as a generalized
framework for modelling social interactions. Philosophical Transactions of the Royal Society B:
Biological Sciences 371(1693), 20150366. Retrieved from https://doi.org/10.1098/rstb.2015.0366
Gescheider, G. A., Bolanowski, S. J., Pope, J. V., & Verrillo, R. T. (2002). A four-channel analysis of
the tactile sensitivity of the fingertip: Frequency selectivity, spatial summation, and temporal
summation. Somatosensory and Motor Research 19(2), 114–124.
Goodale, M. A., & Milner, A. D. (1992). Separate visual pathways for perception and action. Trends
in Neurosciences 15(1), 20–25.
Grahn, J. A. (2012). See what I hear? Beat perception in auditory and visual rhythms. Experimental
Brain Research 220(1), 51–61.
Grahn, J. A., & Brett, M. (2007). Rhythm and beat perception in motor areas of the brain. Journal of
Cognitive Neuroscience 19(5), 893–906.
Gu, Y., Angelaki, D. E., & DeAngelis, G. C. (2008). Neural correlates of multisensory cue
integration in macaque MSTd. Nature Neuroscience 11(10), 1201–1210.
Hairston, W. D., Hodges, D. A., Casanova, R., Hayasaka, S., Kraft, R., Maldjian, J. A., & Burdette, J.
H. (2008). Closing the mind’s eye: Deactivation of visual cortex related to auditory task difficulty.
Neuroreport 19(2), 151–154.
Hickok, G. (2009). Eight problems for the mirror neuron theory of action understanding in monkeys
and humans. Journal of Cognitive Neuroscience 21(7), 1229–1243.
Hopkins, C., Maté-Cid, S., Fulford, R., Seiffert, G., & Ginsborg, J. (2016). Vibrotactile presentation
of musical notes to the glabrous skin for adults with normal hearing or a hearing impairment:
Thresholds, dynamic range and high-frequency perception. PLoS ONE 11(5), e0155807. Retrieved
from https://doi.org/10.1371/journal.pone.0155807
Hove, M. J., Fairhurst, M. T., Kotz, S. A., & Keller, P. E. (2013). Synchronizing with auditory and
visual rhythms: An fMRI assessment of modality differences and modality appropriateness.
NeuroImage 67, 313–321.
Hyde, K. L., Peretz, I., & Zatorre, R. J. (2008). Evidence for the role of the right auditory cortex in
fine pitch resolution. Neuropsychologia 46(2), 632–639.
Iversen, J. R., Patel, A. D., Nicodemus, B., & Emmorey, K. (2015). Synchronization to auditory and
visual rhythms in hearing and deaf individuals. Cognition 134, 232–244.
Johnsrude, I. S., Penhune, V. B., & Zatorre, R. J. (2000). Functional specificity in the right human
auditory cortex for perceiving pitch direction. Brain 123(1), 155–163.
Jones, J. A., & Callan, D. E. (2003). Brain activity during audiovisual speech perception: An fMRI
study of the McGurk effect. Neuroreport 14(8), 1129–1133.
Kaas, J. H., & Hackett, T. A. (2000). Subdivisions of auditory cortex and processing streams in
primates. Proceedings of the National Academy of Sciences 97(22), 11793–11799.
Kaplan, J. T., & Iacoboni, M. (2007). Multimodal action representation in human left ventral
premotor cortex. Cognitive Processing 8(2), 103–113.
Keil, J., Müller, N., Ihssen, N., & Weisz, N. (2012). On the variability of the McGurk effect:
Audiovisual integration depends on prestimulus brain states. Cerebral Cortex 22(1), 221–231.
Kilner, J. M. (2011). More than one pathway to action understanding. Trends in Cognitive Sciences
15(8), 352–357.
Kilner, J. M., Friston, K. J., & Frith, C. D. (2007). Predictive coding: An account of the mirror
neuron system. Cognitive Processing 8(3), 159–166.
Lahav, A., Saltzman, E., & Schlaug, G. (2007). Action representation of sound: Audiomotor
recognition network while listening to newly acquired actions. Journal of Neuroscience?27(2),
308–314.
Levänen, S., Jousmäki, V., & Hari, R. (1998). Vibration-induced auditory-cortex activation in a
congenitally deaf adult. Current Biology 8(15), 869–872.
Lévêque, Y., & Schön, D. (2013). Listening to the human voice alters sensorimotor brain rhythms.
PLoS ONE 8(11), 1–10.
Liégeois-Chauvel, C., Giraud, K., Badier, J. M., Marquis, P., & Chauvel, P. (2012). Intracerebral
evoked potentials in pitch perception reveal a functional asymmetry of human auditory cortex.
Annals of the New York Academy of Sciences 930, 117–132.
Lomber, S. G., Meredith, M. A., & Kral, A. (2010). Cross-modal plasticity in specific auditory
cortices underlies visual compensations in the deaf. Nature Neuroscience 13(11), 1421–1427.
Loui, P. (2015). A dual-stream neuroanatomy of singing. Music Perception: An Interdisciplinary
Journal 32(3), 232–241.
Luo, H., Liu, Z., & Poeppel, D. (2010). Auditory cortex tracks both auditory and visual stimulus
dynamics using low-frequency neuronal phase modulation. PLoS Biology 8(8), e1000445.
Retrieved from http://dx.plos.org/10.1371/journal.pbio.1000445.g007
Luo, H., & Poeppel, D. (2007). Phase patterns of neuronal responses reliably discriminate speech in
human auditory cortex. Neuron 54(6), 1001–1010.
McGarry, L. M., Pineda, J. A., & Russo, F. A. (2015). The role of the extended MNS in emotional
and nonemotional judgments of human song. Cognitive, Affective, & Behavioral Neuroscience
15(1), 32–44. https://doi.org/10.3758/s13415-014-0311-x
McGarry, L. M., Russo, F. A., Schalles, M. D., & Pineda, J. A. (2012). Audio-visual facilitation of
the mu rhythm. Experimental Brain Research 218(4), 527–538.
McGurk, H., & Macdonald, J. (1976). Hearing lips and seeing voices. Nature 264(5588), 746–748.
Maes, P.-J., Leman, M., Palmer, C., & Wanderley, M. M. (2014). Action-based effects on music
perception. Frontiers in Psychology 4. Retrieved from https://doi.org/10.3389/fpsyg.2013.01008
Makous, J. C., Friedman, R. M., & Vierck, C. J. (1995). A critical band filter in touch. Journal of
Neuroscience 15(4), 2808–2818.
Manning, F. C., & Schutz, M. (2013). “Moving to the beat” improves timing perception.
Psychonomic Bulletin and Review 20(6), 1133–1139.
Manning, F. C., & Schutz, M. (2016). Trained to keep a beat: Movement-related enhancements to
timing perception in percussionists and non-percussionists. Psychological Research 80(4), 532–
542.
Marshall, M. T., & Wanderley, M. M. (2011). Examining the effects of embedded vibrotactile
feedback on the feel of a digital musical instrument. New Interfaces for Musical Expression (June),
399–404.
Merabet, L. B., Hamilton, R., Schlaug, G., Swisher, J. D., Kiriakopoulos, E. T., Pitskel, N. B., …
Pascual-Leone, A. (2008). Rapid and reversible recruitment of early visual cortex for touch. PLoS
ONE 3(8). Retrieved from https://doi.org/10.1371/journal.pone.0003046
Merchant, H., Grahn, J., Trainor, L., Rohrmeier, M., & Fitch, W. T. (2015). Finding the beat: A
neural perspective across humans and non-human primates. Philosophical Transactions of the
Royal Society B: Biological Sciences 370(1664), 20140093. Retrieved from
https://doi.org/10.1098/rstb.2014.0093
Meredith, M. A., & Stein, B. E. (1986). Visual, auditory, and somatosensory convergence on cells in
superior colliculus results in multisensory integration. Journal of Neurophysiology 56(3), 640–662.
Milner, B. (1962). Laterality effects in audition. In V. B. Mountcastle (Ed.), Interhemispheric
relations and cerebral dominance (pp. 177–195). Baltimore, MD: Johns Hopkins University Press.
Morioka, M., & Griffin, M. J. (2005). Thresholds for the perception of hand-transmitted vibration:
Dependence on contact area and contact location. Somatosensory and Motor Research 22(4), 281–
297.
Müller, K., Aschersleben, G., Schmitz, F., Schnitzler, A., Freund, H. J., & Prinz, W. (2008). Inter-
versus intramodal integration in sensorimotor synchronization: A combined behavioral and
magnetoencephalographic study. Experimental Brain Research 185(2), 309–318.
Nattiez, J.-J. (1990). Music and discourse: Toward a semiology of music. Princeton, NJ: Princeton
University Press.
North, A. C. (2012). The effect of background music on the taste of wine. British Journal of
Psychology,103(3), 293–301.
North, A. C., Hargreaves, D. J., & McKendrick, J. (1999). The influence of in-store music on wine
selections. Journal of Applied Psychology 84(2), 271–276.
Papetti, S., Jarvelainen, H., Giordano, B. L., Schiesser, S., & Frohlich, M. (2017). Vibrotactile
sensitivity in active touch: Effect of pressing force. IEEE Transactions on Haptics 10(1), 113–122.
Patel, A. D., & Iversen, J. R. (2014). The evolutionary neuroscience of musical beat perception: The
Action Simulation for Auditory Prediction (ASAP) hypothesis. Frontiers in Systems Neuroscience
8, 57. Retrieved from https://doi.org/10.3389/fnsys.2014.00057
Patel, A. D., Iversen, J. R., Chen, Y., & Repp, B. H. (2005). The influence of metricality and
modality on synchronization with a beat. Experimental Brain Research 163(2), 226–238.
Peretz, I. (1990). Processing of local and global musical information by unilateral brain-damaged
patients. Brain 113(4), 1185–1205.
Phillips-Silver, J., & Trainor, L. J. (2005). Feeling the beat: Movement influences infant rhythm
perception. Science 308(5727), 1430.
Phillips-Silver, J., & Trainor, L. J. (2007). Hearing what the body feels: Auditory encoding of
rhythmic movement. Cognition 105(3), 533–546.
Platz, F., & Kopiez, R. (2012). When the eye listens: A meta-analysis of how audio-visual
presentation enhances the appreciation of music performance. Music Perception 30(1), 71–83.
Poeppel, D. (2001). Pure word deafness and the bilateral processing of the speech code. Cognitive
Science 25(5), 679–693.
Quinto, L., Thompson, W. F., Russo, F. A., & Trehub, S. E. (2010). A comparison of the McGurk
effect for spoken and sung syllables. Attention, Perception, & Psychophysics 72(6), 1450–1454.
Rauschecker, J. P., Tian, B., & Hauser, M. (1995). Processing of complex sounds in the macaque
nonprimary auditory cortex. Science 268(5207), 111–114.
Rauschecker, J. P., Tian, B., Pons, T., & Mishkin, M. (1997). Serial and parallel processing in rhesus
monkey auditory cortex. Journal of Comparative Neurology 382(1), 89–103.
Ro, T., Hsu, J., Yasar, N. E., Caitlin Elmore, L., & Beauchamp, M. S. (2009). Sound enhances touch
perception. Experimental Brain Research 195(1), 135–143.
Röder, B., Stock, O., Bien, S., Neville, H., & Rösler, F. (2002). Speech processing activates visual
cortex in congenitally blind humans. European Journal of Neuroscience 16(5), 930–936.
Rohe, T., & Noppeney, U. (2015). Cortical hierarchies perform Bayesian causal inference in
multisensory perception. PLoS Biology 13(2). Retrieved from
https://doi.org/10.1371/journal.pbio.1002073
Rosenblum, L. D., & Fowler, C. A. (1991). Audiovisual investigation of the loudness-effort effect for
speech and nonspeech events. Journal of Experimental Psychology: Human Perception and
Performance 17(4), 976–985.
Ross, J. M., Iversen, J. R., & Balasubramaniam, R. (2016). Motor simulation theories of musical beat
perception. Neurocase 22(6), 558–565.
Royal, I., Lidji, P., Théoret, H., Russo, F. A., & Peretz, I. (2015). Excitability of the motor system: A
transcranial magnetic stimulation study on singing and speaking. Neuropsychologia 75, 525–532.
Russo, F. A., Ammirante, P., & Fels, D. I. (2012). Vibrotactile discrimination of musical timbre.
Journal of Experimental Psychology: Human Perception and Performance 38(4), 822–826.
Russo, F. A., Sandstrom, G. M., & Maksimowski, M. (2011). Mouth versus eyes: Gaze fixation
during perception of sung interval size. Psychomusicology: Music, Mind, and Brain 21(1–2), 98–
107.
Saldaña, H. M., & Rosenblum, L. D. (1993). Visual influences on auditory pluck and bow judgments.
Perception & Psychophysics 54(3), 406–416.
Schalles, M. D., & Pineda, J. A. (2015). Musical sequence learning and EEG correlates of
audiomotor processing. Behavioural Neurology, 2015. Retrieved from
https://doi.org/10.1155/2015/638202
Schroeder, C. E., Lindsley, R. W., Specht, C., Marcovici, A., Smiley, J. F., & Javitt, D. C. (2001).
Somatosensory input to auditory association cortex in the macaque monkey. Journal of
Neurophysiology 85(3), 1322–1327.
Schürmann, M., Caetano, G., Hlushchuk, Y., Jousmäki, V., & Hari, R. (2006). Touch activates human
auditory cortex. NeuroImage 30(4), 1325–1331.
Schutz, M. (2008). Seeing music? What musicians need to know about vision. Empirical Musicology
Review 3(3), 83–108.
Schutz, M., & Kubovy, M. (2009). Deconstructing a musical illusion: Point-light representations
capture salient properties of impact motions. Canadian Acoustics 37(1) 23–28.
Schutz, M., & Lipscomb, S. (2007). Hearing gestures, seeing music: Vision influences perceived tone
duration. Perception 36(6), 888–897.
Senkowski, D., Schneider, T. R., Foxe, J. J., & Engel, A. K. (2008). Crossmodal binding through
neural coherence: Implications for multisensory processing. Trends in Neurosciences 31(8), 401–
409.
Simon, D. M., & Wallace, M. T. (2017). Rhythmic modulation of entrained auditory oscillations by
visual inputs. Brain Topography 30(5), 565–578.
Spence, C. (2011). Crossmodal correspondences: A tutorial review. Attention, Perception, &
Psychophysics 73(4), 971–995.
Stein, B. E., & Meredith, M. A. (1993). The merging of the senses. Cambridge, MA: MIT Press.
Stevenson, R. A., & James, T. W. (2009). Audiovisual integration in human superior temporal sulcus:
Inverse effectiveness and the neural processing of speech and object recognition. NeuroImage
44(3), 1210–1223.
Thomas, C. (1983). Music as heard: A study in applied phenomenology. New Haven, CT: Yale
University Press.
Thompson, W. F., Graham, P., & Russo, F. A. (2005). Seeing music performance: Visual influences
on perception and experience. Semiotica 156(1/4), 203–227.
Thompson, W. F., & Russo, F. A. (2007). Facing the music. Psychological Science 18(9), 756–757.
Thompson, W., Russo, F., & Livingstone, S. (2010). Facial expressions of singers influence perceived
pitch relations. Psychonomic Bulletin & Review 17(3), 317–322.
Thompson, W. F., Russo, F. A., & Quinto, L. (2008). Audio-visual integration of emotional cues in
song. Cognition & Emotion 22(8), 1457–1470.
Tranchant, P., Shiell, M. M., Giordano, M., Nadeau, A., Peretz, I., & Zatorre, R. J. (2017). Feeling
the beat: Bouncing synchronization to vibrotactile music in hearing and early deaf people.
Frontiers in Neuroscience 11. Retrieved from https://doi.org/10.3389/fnins.2017.00507
Varlet, M., Marin, L., Issartel, J., Schmidt, R. C., & Bardy, B. G. (2012). Continuity of visual and
auditory rhythms influences sensorimotor coordination. PLoS ONE 7(9). Retrieved from
https://doi.org/10.1371/journal.pone.0044082
Verrillo, R. T. (1992). Vibration sensation in humans. Music Perception: An Interdisciplinary Journal
9(3), 281–302.
Verrillo, R. T., & Bolanowski, S. J. (1986). The effects of skin temperature on the psychophysical
responses to vibration on glabrous and hairy skin. Journal of the Acoustical Society of America
80(2), 528–532.
Vines, B. W., Krumhansl, C. L., Wanderley, M. M., Dalca, I. M., & Levitin, D. J. (2011). Music to
my eyes: Cross-modal interactions in the perception of emotions in musical performance.
Cognition 118(2), 157–170.
Vuoskoski, J. K., Thompson, M. R., Clarke, E. F., & Spence, C. (2014). Crossmodal interactions in
the perception of expressivity in musical performance. Attention, Perception, & Psychophysics,
76(2), 591–604.
Wapnick, J., Mazza, J. K., & Darrow, A. A. (2000). Effects of performer attractiveness, stage
behaviour, and dress on evaluation of children’s piano performances. Journal of Research in Music
Education 323(4), 323–335.
Zatorre, R. J. (1988). Pitch perception of complex tones and human temporal-lobe function. Journal
of the Acoustical Society of America 84(2), 566–572.
Zatorre, R. J., & Belin, P. (2001). Spectral and temporal processing in human auditory cortex.
Cerebral Cortex 11(10), 946–953.
Zatorre, R. J., Chen, J. L., & Penhune, V. B. (2007). When the brain plays music: Auditory–motor
interactions in music perception and production. Nature Reviews Neuroscience 8, 547–558.
SECTION IV
N E U R A L R E S P ON S E S
TO MU S IC :
C OGN IT ION , A F F E C T,
L A N GU A GE
CHAPT E R 11
M U S I C A N D M E M O RY
L U T Z JÄ N C K E
H M G
Human memory comprises several parts: (1) sensory memory, (2) short-
term memory, (3) working memory, and (4) long-term memory. The
sensory memory stores sensory information for a very short period. This
memory system is strongly associated with neural networks processing
sensory information. Thus, this information is not processed, interpreted,
and encoded. The working memory system is a central system not only for
memory processes; it is rather pivotal for many if not all cognitions. The
main functions controlled by working memory are often coined as
“maintenance and manipulation” to express the fact that working memory
not only holds but also manipulates information. To hold information for a
short period of time without any cognitive manipulation is a matter of short-
term memory. Manipulation, which is a main pillar of the working memory
system, is strongly related to executive functions, pattern recognition, long-
term memory, encoding for long-term memory, language and music
comprehension, problem solving, and even creativity. This is all
accomplished under participation of the working memory system. Thus, this
system is pivotal for nearly all music functions and particularly for music
memory. The neural networks, which are involved in working memory
process, are not focal but are distributed over many brain areas due to the
many functions associated with working memory. In long-term memory,
encoded material is stored for longer time periods, sometimes even
extremely long—up to many decades. Long-term memory is divided into an
explicit and implicit memory system. The explicit memory system contains
consciously available information and comprises the semantic and episodic
memory. The semantic memory contains conscious memory of facts while
the episodic memory is a system for holding events, memory traces
associated with places, times, emotions, and other concept-based
knowledge of an experience. This explicit memory (sometimes also called
declarative memory) is not a simple store; it is rather a mechanism
constructing the past on the basis of stored and new information using
specific strategies (e.g., retrieval schemas, which will be described later).
The neural underpinnings of the explicit memory system are relatively
complex and contain so-called “bottleneck structures” in mesiotemporal
brain areas (including the hippocampus) and networks in temporal, parietal,
as well as frontal brain areas. Thus, the explicit memory system is based on
a distributed network with a mesiotemporal core system. The implicit
memory system contains information that is not easy to verbalize but can be
used without consciously thinking about it. The networks controlling this
implicit memory system do not overlap with the neural networks for the
explicit memory system. The neural networks for the implicit memory
mainly comprise premotor, cerebellar, and basal ganglia structures.
M P M
L
The sound sequences are woven into a melodic contour of pitch and
rhythm. These melodic contours do not appear to be due to bottom-up
processes since the listener is not a passive listener or receiver, but is
actively engaged in processing the music. In this context, the listener uses
acoustic memories, aesthetic judgments, and expectations and combines
them to understand and interpret the particular piece of music (the schema
concept is discussed later in the chapter). Thus, the listener stores many
aspects of the auditory stimuli—such as pitch, pitch interval, timbre, and
rhythm—in memory. Based on this stored information, the listener
constructs an integrated memory of the particular melody. In the following,
I will describe some memory processes associated with tone, tone interval,
and melody processing in more detail.
Tone Memory
Even non-musicians are relatively good at remembering and recognizing
single tones or the pitch of a melody. For example, in an experiment
conducted by Gaab and colleagues (Gaab, Gaser, Zaehle, Jancke, &
Schlaug, 2003), non-musicians performed well in pitch memory tasks
during which they were asked to make a decision on whether the last or
second to last tone of a tone sequence was the same or different as the first
tone. The recognition rate for the tones was astonishingly high, with an
accuracy of about 66 percent. The authors also conducted fMRI
measurements during pitch memory learning. When relating the pitch
memory performance to the task-related hemodynamic responses, they
revealed that bilaterally the supramarginal gyrus and the dorsolateral
cerebellum were significantly correlated with good task performance. The
authors suggest that (besides the auditory cortex), the supramarginal gyrus
and the dorsolateral cerebellum may play a critical role in short-term
storage of pitch information.
Absolute pitch listeners are much better in memorizing tones and
chords. Absolute pitch (AP) is defined as the ability to identify a note
without relying on a reference tone (Levitin & Rogers, 2005; Takeuchi &
Hulse, 1993). It is a rare ability with an incidence of 1 percent in the general
population, although Asian people speaking tonal languages have a higher
rate (Deutsch, Henthorn, Marvin, & Xu, 2006). Absolute pitch is supposed
to originate from an intertwining of genetic factors (Gregersen, Kowalsky,
Kohn, & Marvin, 1999), early exposure to music (Gregersen, Kowalsky,
Kohn, & Marvin, 2001), and intensity of musical training (Gregersen et al.,
2001). Currently, a two-component model is discussed explaining this
extraordinary ability. In the context of this model, it is suggested that AP is
constituted by one perceptual (i.e., “categorical perception”) and two
cognitive—“pitch memory” (i.e., explicit memory) and “pitch labeling”
(i.e., implicit associative memory)—mechanisms, whereby the latter
mechanism has been suggested as constituting the load-bearing skeleton of
AP. Several neurophysiological and neuroanatomical studies support this
suggestion. One main finding in this context is that frontotemporal areas are
strongly activated during tone listening and tone memory tasks in AP
listeners and that these regions are specifically and strongly functionally
and anatomically interconnected (Elmer, Rogenmoser, Kühnis, & Jäncke,
2015; Rogenmoser, Elmer, & Jäncke, 2015; Zatorre, Perry, Beckett,
Westbury, & Evans, 1998). Although these findings are interesting and
important for understanding the neural underpinnings of tone perception
and tone memory, listening to single tones and remembering them are not
adequate tasks in understanding musical listening and the associated
memory processes in their entirety.
M M E
Can music be used as memory enhancer? When asking this question one
has to distinguish which aspect of memory should benefit from music. In
fact there are different influences of music on memory performance. First,
we have to discuss whether musicians or non-professional but musically
trained subjects benefit from musical training in terms of improved memory
performance (e.g., improved verbal working memory or improved long-
term memory). Second, does background music exert beneficial or even
detrimental effects on cognitive functions? Third, can music be used to
enhance memory functions? And fourth, is music beneficial for clinical
samples? In the following I will summarize some of the important findings
in this field.
A M M M
Principally, music memory is not that different from the “classical” memory
system. However, there are some fundamental and important differences
when it comes to natural music and how it is processed, stored, and
retrieved. Music is a dynamic stimulus evolving over time, so when
listeners listen to music they have to integrate the incoming sequential
auditory information and apply specific memory-based mechanisms
(Gestalt perception, chunking, etc.) to form this sequence to a musical
piece. Thus, music listening is not only a matter of simple auditory
information processing, it is much more, since several psychological
functions are involved, from working memory to several aspects of long-
term memory. What is, however, special for music perception and music
memory is the fact that widely distributed neural networks are involved in
perceiving and recognizing musical pieces. Figure 2 demonstrates
schematically the specific nature of the musical memory, which is partly
derived from the SAM model proposed by Raaijmakers and Shiffrin (1981)
and from Kalveram’s model of inverse processing (Kalveram & Seyfarth,
2009).
FIGURE 2. Schematic description of the memory system associating music information with many
non-music aspects.
As one can see from Fig. 2, auditory information is fed to the memory
storage, which can be conceived of as a kind of correlation storage where
auditory information (at) is associated with lots of different information
resulting in a set of efferences (et). In this sense, music information can be
associated with motor programs, which is particularly important for
musicians who have learned to generate music by manipulating
instruments. For that they need specific and highly specialized motor
programs, which they can use to operate their instrument. But non-
musicians do also have (maybe even an innate) audio-motor coupling,
which becomes obvious when we listen to rhythmic music. In these
situations, we tend to move according to the music rhythm. Besides motor
programs, episodic, autobiographical, semantic, and implicit memory
information are also associated with music information. Even emotion and
motivation can be related to the incoming auditory information. These
associations can be conceived of as correlations of varying strength, with
the correlation strength depending on the frequency of repetition and the
salience of the associated information. Some of these correlations give rise
to conscious perceptions (explicit memory) while others remain
unconscious (implicit memory). Executive functions can enhance the
processing of incoming information by directing attention to particular
information at the end enhancing the neural activation of those areas that
are involved in processing this information. We can also apply executive
functions to direct our attention to particular correlations, thus enhancing
their likelihood to result in an appropriate efference. This can also lead to a
kind of suppression and/or inhibition of other correlations. In this context,
perceptual and memory schemas can be applied according to which we
select or enhance incoming information or the pattern of correlations within
the storage. This model can also be operated in an inverse fashion. For
example, when a person wants to change his current mood he might
“activate” the correlations within the storage associated with a particular
emotion. This will “activate” images of those musical pieces that activate
the wished emotion. Thus, the goal (to evoke a particular emotion) is now
fed into the storage, which activates those efferences yielding to the
emotion in question.
No generation has listened to music as often as our generation.
According to a 2016 US survey estimate, more than 90 percent of the
population reported listening to music for an average of 25 hours a week
(Nielsen, 2017). Thus, it is obvious that music is a frequently applied cue
for autobiographical memories because music is associated with lots of
everyday information. Thus music can serve as an efficient strategy to
assess and stimulate our biographical memory.
Music is a complex stimulus carrying much information and which
evolves over time. This could be the reason why music processing is
associated with distributed neural network activations. Obviously, it is
relatively easy to link music information to multimodal information. Music
can carry meaning, emotion, and episodic non-music information; it can
also trigger and control motor behavior. This multiplicity and versatility of
the music network offer many possibilities to insert new information. This
could be the reason why several music-related learning strategies improve
memory functions. However, future studies are necessary to show whether
music interventions can be used to improve memory functions in both
healthy and neurologically impaired subjects.
R
Alonso, I., Dellacherie, D., & Samson, S. (2015). Emotional memory for musical excerpts in young
and older adults. Frontiers in Aging Neuroscience 23. Retrieved from
https://doi.org/10.3389/fnagi.2015.00023
Altenmüller, E., Siggel, S., Mohammadi, B., Samii, A., & Münte, T. F. (2014). Play it again, Sam:
Brain correlates of emotional music recognition. Frontiers in Psychology 114. Retrieved from
https://doi.org/10.3389/fpsyg.2014.00114
Anvari, S. H., Trainor, L. J., Woodside, J., & Levy, B. A. (2002). Relations among musical skills,
phonological processing, and early reading ability in preschool children. Journal of Experimental
Child Psychology 83(2), 111–130.
Baddeley, A. (2010). Working memory. Current Biology 20(4), R136–R140.
Baddeley, A. D., & Hitch, G. (1974). Working memory. In G. H. Bower (Ed.), Psychology of
Learning and Motivation (Vol. 8, pp. 47–89). New York: Academic Press.
Bangert, M., & Altenmüller, E. O. (2003). Mapping perception to action in piano practice: A
longitudinal DC-EEG study. BMC Neuroscience 4, 26. Retrieved from
https://doi.org/10.1186/1471-2202-4-26
Bhattacharya, J., & Petsche, H. (2001). Enhanced phase synchrony in the electroencephalograph
gamma band for musicians while listening to music. Physical Review E: Statistical, Nonlinear, and
Soft Matter Physics 64(1 Pt. 1), 012902.
Bhattacharya, J., Petsche, H., Feldmann, U., & Rescher, B. (2001). EEG gamma-band phase
synchronization between posterior and frontal cortex during mental rotation in humans.
Neuroscience Letters 311(1), 29–32.
Bhattacharya, J., Petsche, H., & Pereda, E. (2001). Long-range synchrony in the gamma band: Role
in music perception. Journal of Neuroscience 21(16), 6329–6337.
Bower, G. H. (1981). Mood and memory. The American Psychologist 36, 129–148.
Calvert, S. L., & Tart, M. (1993). Song versus verbal forms for very-long-term, long-term, and short-
term verbatim recall. Journal of Applied Developmental Psychology 14(2), 245–260.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences. New York: Lawrence
Erlbaum Associates.
Cowan, N. (2011). The focus of attention as observed in visual working memory tasks: Making sense
of competing claims. Neuropsychologia 49, 1401–1406.
Deutsch, D. (1970). Tones and numbers: Specificity of interference in immediate memory. Science
168(3939), 1604–1605.
Deutsch, D., Henthorn, T., Marvin, E., & Xu, H. (2006). Absolute pitch among American and
Chinese conservatory students: Prevalence differences, and evidence for a speech-related critical
period. Journal of the Acoustical Society of America 119(2), 719–722.
Elmer, S., Hänggi, J., Meyer, M., & Jäncke, L. (2013). Increased cortical surface area of the left
planum temporale in musicians facilitates the categorization of phonetic and temporal speech
sounds. Cortex 49(10), 2812–2821.
Elmer, S., Rogenmoser, L., Kühnis, J., & Jäncke, L. (2015). Bridging the gap between perceptual and
cognitive perspectives on absolute pitch. Journal of Neuroscience 35(1), 366–371.
Eschrich, S., Münte, T. F., & Altenmüller, E. O. (2005). Remember Bach: An investigation in
episodic memory for music. Annals of the New York Academy of Sciences 1060, 438–442.
Eschrich, S., Münte, T. F., & Altenmüller, E. O. (2008). Unforgettable film music: The role of
emotion in episodic long-term memory for music. BMC Neuroscience 9, 48. Retrieved from
https://doi.org/10.1186/1471-2202-9-48
Evers, S., & Suhr, B. (2000). Changes of the neurotransmitter serotonin but not of hormones during
short time music perception. European Archives of Psychiatry and Clinical Neuroscience 250(3),
144–147.
Ferreri, L., Bigand, E., Perrey, S., Muthalib, M., Bard, P., & Bugaiska, A. (2014). Less effort, better
results: How does music act on prefrontal cortex in older adults during verbal encoding? An fNIRS
study. Frontiers in Human Neuroscience 8, 301. Retrieved from
https://doi.org/10.3389/fnhum.2014.00301
Ferreri, L., & Rodriguez-Fornells, A. (2017). Music-related reward responses predict episodic
memory performance. Experimental Brain Research 235(12), 3721–3731.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schon, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A randomized
control trial. PLoS ONE 10(9), e0138715.
Ford, J. H., Addis, D. R., & Giovanello, K. S. (2011). Differential neural activity during search of
specific and general autobiographical memories elicited by musical cues. Neuropsychologia 49(9),
2514–2526.
Frieler, K., Fischinger, T., Schlemmer, K., Lothwesen, K., Jakubowski, K., & Müllensiefen, D.
(2013). Absolute memory for pitch: A comparative replication of Levitin’s 1994 study in six
European labs. Musicae Scientiae: The Journal of the European Society for the Cognitive Sciences
of Music 17(3), 334–349.
Gaab, N., Gaser, C., Zaehle, T., Jancke, L., & Schlaug, G. (2003). Functional anatomy of pitch
memory: An fMRI study with sparse temporal sampling. NeuroImage 19(4), 1417–1426.
Gagnepain, P., Fauvel, B., Desgranges, B., Gaubert, M., Viader, F., Eustache, F., … Platel, H. (2017).
Musical expertise increases top-down modulation over hippocampal activation during familiarity
decisions. Frontiers in Human Neuroscience 11, 472. Retrieved from
https://doi.org/10.3389/fnhum.2017.00472
Gobet, F. (1998). Expert memory: A comparison of four theories. Cognition 66(2), 115–152.
Green, A. C., Bærentsen, K. B., Stødkilde-Jørgensen, H., Roepstorff, A., & Vuust, P. (2012). Listen,
learn, like! Dorsolateral prefrontal cortex involved in the mere exposure effect in music. Neurology
Research International 2012, 846270. Retrieved from http://dx.doi.org/10.1155/2012/846270
Gregersen, P. K., Kowalsky, E., Kohn, N., & Marvin, E. W. (1999). Absolute pitch: Prevalence,
ethnic variation, and estimation of the genetic component. American Journal of Human Genetics
65(3), 911–913.
Gregersen, P. K., Kowalsky, E., Kohn, N., & Marvin, E. W. (2001). Early childhood music education
and predisposition to absolute pitch: Teasing apart genes and environment. American Journal of
Medical Genetics 98(3), 280–282.
Groussard, M., La Joie, R., Rauchs, G., Landeau, B., Chetelat, G., Viader, F., … Platel, H. (2010).
When music and long-term memory interact: Effects of musical expertise on functional and
structural plasticity in the hippocampus. PLoS ONE 5(10), e13225.
Halpern, A. R. (1989). Memory for the absolute pitch of familiar songs. Memory & Cognition 17(5),
572–581.
Halpern, A. R., & Müllensiefen, D. (2008). Effects of timbre and tempo change on memory for
music. Quarterly Journal of Experimental Psychology 61(9), 1371–1384.
Halpern, A. R., & O’Connor, M. G. (2000). Implicit memory for music in Alzheimer’s disease.
Neuropsychology 3(14), 391–397.
Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews
Neuroscience 8(5), 393–402.
Honing, H., & Ladinig, O. (2009). Exposure influences expressive timing judgments in music.
Journal of Experimental Psychology: Human Perception and Performance 35(1), 281–288.
Jaencke, L., & Sandmann, P. (2010). Music listening while you learn: No influence of background
music on verbal learning. Behavioral and Brain Functions 6, 3. Retrieved from
https://doi.org/10.1186/1744-9081-6-3
Janata, P. (2009). The neural architecture of music-evoked autobiographical memories. Cerebral
Cortex 19(11), 2579–2594.
Jäncke, L. (2012). The dynamic audio-motor system in pianists. Annals of the New York Academy of
Sciences 1252, 246–252.
Jäncke, L., & Alahmadi, N. (2016). Detection of independent functional networks during music
listening using electroencephalogram and sLORETA-ICA. Neuroreport 27(6), 455–461.
Jäncke, L., Mirzazade, S., & Shah, N. J. (1999). Attention modulates activity in the primary and the
secondary auditory cortex: A functional magnetic resonance imaging study in human subjects.
Neuroscience Letters 266(2), 125–128.
Johnson, M. K., Kim, J. K., & Risse, G. (1985). Do alcoholic Korsakoff’s syndrome patients acquire
affective reactions? Journal of Experimental Psychology: Learning, Memory, and Cognition 11(1),
22–36.
Judde, S., & Rickard, N. (2010). The effect of post-learning presentation of music on long-term
word-list retention. Neurobiology of Learning and Memory 94(1), 13–20.
Kalveram, K. T., & Seyfarth, A. (2009). Inverse biomimetics: How robots can help to verify concepts
concerning sensorimotor control of human arm and leg movements. Journal of Physiology 103(3–
5), 232–243.
Kämpfe, J., Sedlmeier, P., & Renkewitz, F. (2010). The impact of background music on adult
listeners: A meta-analysis. Psychology of Music 39(4), 424–448.
Keitz, M., Martin-Soelch, C., & Leenders, K. L. (2003). Reward processing in the brain: A
prerequisite for movement preparation? Neural Plasticity 10(1–2), 121–128.
Kilgour, A. R., Jakobson, L. S., & Cuddy, L. L. (2000). Music training and rate of presentation as
mediators of text and song recall. Memory & Cognition 28(5), 700–710.
Klimesch, W. (1999). EEG alpha and theta oscillations reflect cognitive and memory performance: A
review and analysis. Brain Research Reviews 29(2), 169–195.
Klimesch, W., Sauseng, P., & Gerloff, C. (2003). Enhancing cognitive performance with repetitive
transcranial magnetic stimulation at human individual alpha frequency. European Journal of
Neuroscience 17(5), 1129–1133.
Koelsch, S., Kasper, E., Sammler, D., Schulze, K., Gunter, T., & Friederici, A. D. (2004). Music,
language and meaning: Brain signatures of semantic processing. Nature Neuroscience 7(3), 302–
307.
Kühn, S., Romanowski, A., Schilling, C., Lorenz, R., Mörsen, C., Seiferth, N., … IMAGEN
Consortium (2011). The neural basis of video gaming. Translational Psychiatry 1(11), e53.
Kühnis, J., Elmer, S., & Jäncke, L. (2014). Auditory evoked responses in musicians during passive
vowel listening are modulated by functional connectivity between bilateral auditory-related brain
regions. Journal of Cognitive Neuroscience 26(12), 2750–2761.
Kussner, M. B., de Groot, A. M., Hofman, W. F., & Hillen, M. A. (2016). EEG beta power but not
background music predicts the recall scores in a foreign-vocabulary learning task. PLoS ONE
11(8), e0161387.
Lee, Y. S., Janata, P., Frost, C., Martinez, Z., & Granger, R. (2015). Melody recognition revisited:
Influence of melodic Gestalt on the encoding of relational pitch information. Psychonomic Bulletin
& Review 22(1), 163–169.
Levitin, D. J. (1994). Absolute memory for musical pitch: Evidence from the production of learned
melodies. Perception & Psychophysics 56, 414–423.
Levitin, D. J., & Rogers, S. E. (2005). Absolute pitch: Perception, coding, and controversies. Trends
in Cognitive Sciences 9(1), 26–33.
Logothetis, N. K. (2008). What we can do and what we cannot do with fMRI. Nature 453(7197),
869–878.
McElhinney, M., & Annett, J. M. (1996). Pattern of efficacy of a musical mnemonic on recall of
familiar words over several presentations. Perceptual and Motor Skills 82(2), 395–400.
McGaugh, J. L. (2000). Memory: A century of consolidation. Science 287(5451), 248–251.
Margulis, E. H., Mlsna, L. M., Uppunda, A. K., Parrish, T. B., & Wong, P. C. M. (2009). Selective
neurophysiologic responses to music in instrumentalists with different listening biographies.
Human Brain Mapping 30(1), 267–275.
Marie, C., Magne, C., & Besson, M. (2011). Musicians and the metric structure of words. Journal of
Cognitive Neuroscience 23(2), 294–305.
Meneses, A., & Liy-Salmeron, G. (2012). Serotonin and emotion, learning and memory. Reviews in
the Neurosciences 23(5–6), 543–553.
Moussard, A., Bigand, E., Belleville, S., & Peretz, I. (2012). Music as an aid to learn new verbal
information in Alzheimer’s disease. Music Perception: An Interdisciplinary Journal 29(5), 521–
531.
Moussard, A., Bigand, E., Belleville, S., & Peretz, I. (2014). Learning sung lyrics aids retention in
normal ageing and Alzheimer’s disease. Neuropsychological Rehabilitation 24(6), 894–917.
Nielsen (2017). Nielsen Music year-end report 2016. Retrieved from
http://www.nielsen.com/us/en/press-room/2017/nielsen-releases-2016-us-year-end-music-
report.html
Oberauer, K., & Lewandowsky, S. (2011). Modeling working memory: A computational
implementation of the Time-Based Resource-Sharing theory. Psychonomic Bulletin & Review
18(1), 10–45.
Okhrei, A., Kutsenko, T., & Makarchuk, M. (2017). Performance of working memory of musicians
and non-musicians in tests with letters, digits, and geometrical shapes. Biologija 62(4), 207–215.
Palisson, J., Roussel-Baclet, C., Maillet, D., Belin, C., Ankri, J., & Narme, P. (2015). Music enhances
verbal episodic memory in Alzheimer’s disease. Journal of Clinical and Experimental
Neuropsychology 37(5), 503–517.
Parks, S. L., & Clancy Dollinger, S. (2014). The positivity effect and auditory recognition memory
for musical excerpts in young, middle-aged, and older adults. Psychomusicology: Music, Mind,
and Brain 24(4), 298–308.
Peretz, I., Gosselin, N., Belin, P., Zatorre, R. J., Plailly, J., & Tillmann, B. (2009). Music lexical
networks. Annals of the New York Academy of Sciences 1169, 256–265.
Peterson, D. A., & Thaut, M. H. (2007). Music increases frontal EEG coherence during verbal
learning. Neuroscience Letters 412(3), 217–221.
Piaget, J. (1923). La langage et la pensée chez l’enfant: Études sur la logique de l’enfant. Retrieved
from
http://pubman.mpdl.mpg.de/pubman/item/escidoc:2375486/component/escidoc:2375485/Piaget_1
923_language_pensee_enfant.pdf
Plailly, J., Tillmann, B., & Royet, J.-P. (2007). The feeling of familiarity of music and odors: The
same neural signature? Cerebral Cortex 17(11), 2650–2658.
Plantinga, J., & Trainor, L. J. (2005). Memory for melody: Infants use a relative pitch code.
Cognition 98(1), 1–11.
Platel, H. (2005). Functional neuroimaging of semantic and episodic musical memory. Annals of the
New York Academy of Sciences 1060, 136–147.
Raaijmakers, J. G., & Shiffrin, R. M. (1981). Search of associative memory. Psychological Review
88(2), 93–134.
Rawson, K. A., & Van Overschelde, J. P. (2008). How does knowledge promote memory? The
distinctiveness theory of skilled memory. Journal of Memory and Language 58(3), 646–668.
Rickard, N. S., Wong, W. W., & Velik, L. (2012). Relaxing music counters heightened consolidation
of emotional memory. Neurobiology of Learning and Memory 97(2), 220–228.
Rogenmoser, L., Elmer, S., & Jäncke, L. (2015). Absolute pitch: Evidence for early cognitive
facilitation during passive listening as revealed by reduced P3a amplitudes. Journal of Cognitive
Neuroscience 27(3), 623–637.
Salamé, P., & Baddeley, A. (1989). Effects of background music on phonological short-term memory.
Quarterly Journal of Experimental Psychology Section A 41(1), 107–122.
Salimpoor, V. N., Zald, D. H., Zatorre, R. J., Dagher, A., & McIntosh, A. R. (2015). Predictions and
the brain: How musical sounds become rewarding. Trends in Cognitive Sciences 19(2), 86–91.
Samson, S., & Peretz, I. (2005). Effects of prior exposure on music liking and recognition in patients
with temporal lobe lesions. Annals of the New York Academy of Sciences 1060, 419–428.
Särkämö, T., Tervaniemi, M., Laitinen, S., Forsblom, A., Soinila, S., Mikkonen, M., … Hietanen, M.
(2008). Music listening enhances cognitive recovery and mood after middle cerebral artery stroke.
Brain: A Journal of Neurology 131, 866–876.
Schellenberg, E. G. (2001). Music and nonmusical abilities. Annals of the New York Academy of
Sciences 930, 355–371. Reprinted in G. E. McPherson (Ed.), The child as musician: A handbook
of musical development (2nd ed., pp. 149–176). Oxford: Oxford University Press, 2016.
Schellenberg, E. G., & Weiss, M. W. (2013). Music and cognitive abilities. In D. Deutsch (Ed.), The
Psychology of Music (3rd ed., pp. 499–550). London: Academic Press.
Schendel, Z. A., & Palmer, C. (2007). Suppression effects on musical and verbal memory. Memory &
Cognition 35(4), 640–650.
Schulze, K., & Koelsch, S. (2012). Working memory for speech and music. Annals of the New York
Academy of Sciences 1252, 229–236.
Schulze, K., Mueller, K., & Koelsch, S. (2011). Neural correlates of strategy use during auditory
working memory in musicians and non-musicians. European Journal of Neuroscience 33(1), 189–
196.
Schulze, K., Mueller, K., & Koelsch, S. (2013). Auditory stroop and absolute pitch: An fMRI study.
Human Brain Mapping 34(7), 1579–1590.
Schulze, K., Zysset, S., Mueller, K., Friederici, A. D., & Koelsch, S. (2011). Neuroarchitecture of
verbal and tonal working memory in nonmusicians and musicians. Human Brain Mapping 32,
771–783.
Semal, C., Demany, L., Ueda, K., & Hallé, P. A. (1996). Speech versus nonspeech in pitch memory.
Journal of the Acoustical Society of America 100(2 Pt. 1), 1132–1140.
Simmons-Stern, N. R., Budson, A. E., & Ally, B. A. (2010). Music as a memory enhancer in patients
with Alzheimer’s disease. Neuropsychologia 48(10), 3164–3167.
Simmons-Stern, N. R., Deason, R. G., Brandler, B. J., Frustace, B. S., O’Connor, M. K., Ally, B. A.,
& Budson, A. E. (2012). Music-based memory enhancement in Alzheimer’s disease: Promise and
limitations. Neuropsychologia 50(14), 3295–3303.
Takeuchi, A. H., & Hulse, S. H. (1993). Absolute pitch. Psychological Bulletin 113(2), 345–361.
Talamini, F., Altoè, G., Carretti, B., & Grassi, M. (2017). Musicians have better memory than
nonmusicians: A meta-analysis. PLoS ONE 12(10), e0186773.
Talamini, F., Carretti, B., & Grassi, M. (2016). The working memory of musicians and nonmusicians.
Music Perception: An Interdisciplinary Journal 34(2), 183–191.
Tamminen, J., Rastle, K., Darby, J., Lucas, R., & Williamson, V. J. (2017). The impact of music on
learning and consolidation of novel words. Memory 25(1), 107–121.
Thaut, M. H., Peterson, D. A., McIntosh, G. C., & Hoemberg, V. (2014). Music mnemonics aid
verbal memory and induce learning-related brain plasticity in multiple sclerosis. Frontiers in
Human Neuroscience 8, 395. Retrieved from https://doi.org/10.3389/fnhum.2014.00395
Trainor, L. J., McDonald, K. L., & Alain, C. (2002). Automatic and controlled processing of melodic
contour and interval information measured by electrical brain activity. Journal of Cognitive
Neuroscience 14(3), 430–442.
Vieillard, S., & Gilet, A.-L. (2013). Age-related differences in affective responses to and memory for
emotions conveyed by music: A cross-sectional study. Frontiers in Psychology 4, 711. Retrieved
from https://doi.org/10.3389/fpsyg.2013.00711
Wallace, W. T. (1994). Memory for music: Effect of melody on recall of text. Journal of
Experimental Psychology: Learning, Memory, and Cognition 20(6), 1471–1485.
Watanabe, T., Yagishita, S., & Kikyo, H. (2008). Memory of music: Roles of right hippocampus and
left inferior frontal gyrus. NeuroImage 39(1), 483–491.
Werbik, H. (1971). Informationsgehalt und emotionale Wirkung von Musik. Mainz: B. Schott.
Wildschut, T., Sedikides, C., Arndt, J., & Routledge, C. (2006). Nostalgia: Content, triggers,
functions. Journal of Personality and Social Psychology 91(5), 975–993.
Williamson, V. J., Baddeley, A. D., & Hitch, G. J. (2010). Musicians’ and nonmusicians’ short-term
memory for verbal and musical sequences: Comparing phonological similarity and pitch
proximity. Memory & Cognition 38(2), 163–175.
Zajonc, R. B. (1968). Attitudinal effects of mere exposure. Journal of Personality and Social
Psychology 9(2 pt. 2), 1–27.
Zatorre, R. J., Perry, D. W., Beckett, C. A., Westbury, C. F., & Evans, A. C. (1998). Functional
anatomy of musical processing in listeners with absolute pitch and relative pitch. Proceedings of
the National Academy of Sciences 95(6), 3172–3177.
CHAPT E R 12
M U S I C A N D AT T E N T I O N ,
EXECUTIVE FUNCTION,
A N D C R E AT I V I T Y
P S Y C H E L O U I A N D R A C H E L E . G U E T TA
M A
Theories of Attention
Patel’s OPERA hypothesis (Patel, 2011b) posits that one of several reasons
why music training benefits the neural encoding of speech is through
attention: by engaging shared brain networks between music and speech
that are associated with focused attention. Attention has been thought of in
terms of early versus late selection theories, and in terms of its operation
over space and time. Early selection theories focus on sensory processing
and more exogenous (reflexive) sources of information, whereas late
selection theories focus more on feature selection and more cognitive,
endogenous operations.
Theories of early versus late attention differ in their posited effects of
perceptual selection, enhancement, and cognitive focus along various points
in the perceptual-cognitive pathway, or along the gradient of primary to
association areas in the human cortex. Such early versus late selection
theories of attention pertain to when, temporally, along the classic sensory-
cognitive pathway attentional processes might operate. Early selection
theories generally focus on sensory processing (closer to the sensory
periphery) whereas late selection theories focus more on cognitive
processing. Evidence for early selection comes from findings from the
dichotic listening paradigm in which event-related brain potentials were
recorded. The amplitude of the N1, an event-related brain response
generated in response to sounds, is enhanced in response to sounds in the
attended ear relative to the unattended ear (Woldorff & Hillyard, 1991).
Magnetoencephalography work subsequently pointed out the source of this
attentional enhancement to the auditory cortex (Woldorff et al., 1993). Since
the auditory cortex is part of the primary sensory cortices, the finding that
this early cortical way station acts on attentional processing as early as 100
ms after sound presentation provides convincing evidence for early
selection.
Theories that posit relatively late selection conceptualize attention as a
feature-based or object-based operation. In particular, the feature integration
theory (Treisman & Gelade, 1980) posits that attention operates by
combining pre-attentively selected features within a busy scene. Support for
this comes from illusory conjunctions, in which unattended features of
visual objects, such as color and shape, are sometimes combined to give
rise to an illusory percept of a nonexistent object. While this theory has
received lots of interest, the definitions of features in vision may not so
readily transfer to audition. In the auditory modality, stimulus
representation has been described as hierarchical, as shown by
psychophysical and modeling studies. At the lowest rung of the hierarchy of
stimulus representation there are “primitive features” such as acoustic
frequency, whereas at higher levels there are more complex, emergent
features such as virtual pitch, which combine with other features to form
objects. Attention can be enhanced by cueing at the appropriate level, thus
reducing uncertainty (Hafter & Saberi, 2001). Object-based attention offers
a direct comparison between visual and auditory processing. Much like the
visual system combines features to form objects, the auditory system forms
objects by grouping together sound elements that share features such as
frequency and harmonic structure (Shinn-Cunningham, 2008). The
temporal evolution of these features is especially relevant for object
formation in the auditory system. At a low-level timescale, the auditory
system may group together sounds based on similar fine-grained temporal
features such as attack time, while at a higher-level timescale, distinct tones
may be grouped together based on temporal proximity to give rise to beat
perception. Beat perception has been proposed as an attentional mechanism,
through which different temporal objects such as tones are combined
together to form larger units such as rhythms and phrases (De Freitas,
Liverence, & Scholl, 2014; Grahn, 2012). The rhythmic effects of attention
over time will be revisited later in this section.
Evidence for late selection in auditory neurophysiology comes from
findings of later attention-related enhancements in event-related brain
responses such as the P300 (Purves et al., 2008). In addition, cases of late
selection are supported by the neuropsychological literature, in which
patients with lesions in the right parietal cortex present with lack of
awareness of their contralesional (usually left) visual field. For these cases,
sometimes the successful perception of one feature can reduce the
detectability of another, simultaneously present feature, a condition known
as extinction. In the auditory/musical modality, interesting evidence comes
from the use of an auditory illusion in a case of auditory extinction
(Deouell, Deutsch, Scabini, Soroker, & Knight, 2007). This study took
advantage of Deutsch’s scale illusion, in which presenting subjects with
alternating high-pitched and low-pitched tones to the left and right ear
paradoxically leads to the percept of a stream of high tones in the right ear
and low tones in the left (Deutsch, 1974). When the patient with auditory
neglect was presented with octave illusion stimuli, he reported only hearing
the high-pitched stream. The fact that he only heard the right-lateralized
stream, rather than the right-ear stimulus, does suggest that some forms of
perceptual analysis, such as the formation of auditory streams, are intact
before attention and its disruption in hemispatial neglect, thus providing
support for late selection.
A third line of literature supports a combination of early and late
selection theories in showing attention-related enhancements of mid-latency
brain responses to sound. For example, the mismatch negativity, an event-
related potential generated around 200 ms after the onset of unexpected
sounds, is both pre-attentively generated and modulated by attention
(Woldorff, Hillyard, Gallen, Hampson, & Bloom, 1998). More specific to
the music literature, the Early Right Anterior Negativity (ERAN), an event-
related potential in response to unexpected musical chords such as the
Neapolitan chord (Koelsch, Gunter, Friederici, & Schröger, 2000), is also
pre-attentively generated but modulated by attention: When subjects were
directing attention away from auditory stimuli in a visual task, they
nevertheless elicited an ERAN in response to the Neapolitan chord;
however its amplitude is larger in the attended condition (Loui, Grent-’T-
Jong, Torpey, & Woldorff, 2005). Taken together, the best available
resolution for the debate on early versus late selection holds that attention
acts on multiple levels of the perceptual-cognitive or primary-association
continuum, by selecting relevant features and processing them more fully at
more sensory stages, and also by combining selected features to form
coherent objects, streams, or scenes at later association stages.
M E F
Far Transfer
The hypothesis that music training enhances EFs assumes that far transfer
of cognitive skills takes place as a result of training; however, far transfer
has not been reliable across studies (Sala & Gobet, 2017b). On one hand,
cross-sectional studies comparing musicians and non-musicians have shown
positive effects of EF: Adult musicians perform better on measures of
cognitive flexibility, working memory, and verbal fluency, and musically
trained children also perform better on behavioral and fMRI indices of
verbal fluency, rule representation, and task switching (Zuk, Benjamin,
Kenyon, & Gaab, 2014). On the other hand, cross-sectional studies are still
limited by the fundamental possibility that results may be due to similar
confounds as the association studies, such as differences in parental
education, socio-economic status (although these were mostly controlled for
in the previous study), or some aspect of exposure in the home environment
that is outside of the experimenter’s control, as well as pre-existing
differences before initiating training. Long-term differences in EF
performance, only after controlling for these potential confounds, would
provide a convincing basis for the possibility of far transfer.
Negative Findings
Studies reviewed thus far have reported positive transfer effects for near
transfer, and more limited but nevertheless successful results on far transfer.
However, not all reports have been positive, and the effect sizes of far
transfer have been small, as shown by a recent meta-analysis of the far
transfer effects of musical training (Sala & Gobet, 2017a, b). Mehr and
colleagues found no reliable evidence for non-musical cognitive benefits
from brief preschool music lessons (Mehr, Schachner, Katz, & Spelke,
2013). Preschool children were either given music classes, arts instruction,
or no lessons. After six weeks, the participants were assessed in four
distinct cognitive areas in which older arts students have been reported to
excel: spatial-navigational reasoning, visual form analysis, numerical
discrimination, and receptive vocabulary. At first, music class participants
showed greater spatial-navigational ability than those in the visual arts
class, while children from the visual arts class showed greater visual form
analysis ability than children from the music class. However, the
researchers were unable to replicate this trend. In the end, the children who
were provided with music classes performed no better overall than those
with visual arts or no classes. These findings demand caution in interpreting
other positive findings for enhanced executive functioning as a result of
music instruction. It may be important to note, however, that the brief
training sessions from this study do not readily compare to long-term
musical training. Furthermore, the selection of transfer tasks needs to take
into account the underlying mechanism that could lead to transfer.
R
Amabile, T. M. (1982). Social psychology of creativity: A consensual assessment technique. Journal
of Personality and Social Psychology 43(5), 997–1013.
Arkin, C., Przysinda, E., Pfeifer, C., Zeng, T., & Loui, P. (2017). Information content predicts
creativity in musical improvisation: A behavioral and voxel-based morphometry study. Under
review.
Arkin, C., Przysinda, E., Pfeifer, C., Zeng, T., & Loui, P. (2019). Grey matter correlates of creativity
in musical improvisation. Under review.
Beaty, R. E. (2015). The neuroscience of musical improvisation. Neuroscience & Biobehavioral
Reviews 51, 108–117.
Benedek, M., Borovnjak, B., Neubauer, A. C., & Kruse-Weber, S. (2014). Creativity and personality
in classical, jazz and folk musicians. Personality and Individual Differences 63, 117–121.
Bengtsson, S. L., Csikszentmihalyi, M., & Ullén, F. (2007). Cortical regions involved in the
generation of musical structures during improvisation in pianists. Journal of Cognitive
Neuroscience 19, 830–842.
Berkowitz, A. L., & Ansari, D. (2008). Generation of novel motor sequences: The neural correlates
of musical improvisation. NeuroImage 41(2), 535–543.
Bharucha, J. J., & Stoeckig, K. (1986). Reaction time and musical expectancy: Priming of chords.
Journal of Experimental Psychology: Human Perception and Performance 12(4), 403–410.
Biasutti, M. (2015). Pedagogical applications of the cognitive research on music improvisation.
Frontiers in Psychology 6. Retrieved from https://doi.org/10.3389/fpsyg.2015.00614
Bigand, E., Poulin, B., Tillmann, B., Madurell, F., & D’Adamo, D. A. (2003). Sensory versus
cognitive components in harmonic priming. Journal of Experimental Psychology: Human
Perception and Performance 29(1), 159–171.
Boccia, M., Piccardi, L., Palermo, L., Nori, R., & Palmiero, M. (2015). Where do bright ideas occur
in our brain? Meta-analytic evidence from neuroimaging studies of domain-specific creativity.
Frontiers in Psychology 6, 1195. Retrieved from https://doi.org/10.3389/fpsyg.2015.01195
Bregman, A. S. (1994). Auditory scene analysis: The perceptual organization of sound. Cambridge,
MA: MIT Press.
Broadbent, D. E. (1982). Task combination and selective intake of information. Acta Psychologica
50(3), 253–290.
Cherry, E. C. (1953). Some experiments on the recognition of speech, with one and with two ears.
Journal of the Acoustical Society of America 25(5), 975–979.
Christoff, K., Irving, Z. C., Fox, K. C., Spreng, R. N., & Andrews-Hanna, J. R. (2016). Mind-
wandering as spontaneous thought: A dynamic framework. Nature Reviews Neuroscience 17(11),
718–731.
Collins, D. (2016). The act of musical composition: Studies in the creative process. New York:
Routledge.
Csikszentmihalyi, M. (1996). Creativity: Flow and the psychology of discovery and invention. New
York: HarperCollins.
De Freitas, J., Liverence, B. M., & Scholl, B. J. (2014). Attentional rhythm: A temporal analogue of
object-based attention. Journal of Experimental Psychology: General 143(1), 71–76.
Dennett, D. C. (2008). Kinds of minds: Toward an understanding of consciousness. New York: Basic
Books.
Deouell, L. Y., Deutsch, D., Scabini, D., Soroker, N., & Knight, R. T. (2007). No disillusions in
auditory extinction: Perceiving a melody comprised of unperceived notes. Frontiers in Human
Neuroscience 1, 15. Retrieved from https://doi.org/10.3389/neuro.09.015.2007
Deutsch, D. (1974). An illusion with musical scales. Journal of the Acoustical Society of America
56(S1). Retrieved from https://doi.org/10.1121/1.1914084.
Diamond, A. (2013). Executive functions. Annual Review of Psychology 64, 135–168.
Ding, N., Patel, A. D., Chen, L., Butler, H., Luo, C., & Poeppel, D. (2017). Temporal modulations in
speech and music. Neuroscience & Biobehavioral Reviews 81(Part B), 181–187.
Donnay, G. F., Rankin, S. K., Lopez-Gonzalez, M., Jiradejvong, P., & Limb, C. J. (2014). Neural
substrates of interactive musical improvisation: An fMRI study of “trading fours” in jazz. PLoS
ONE 9, e88665.
Ellis, R. J., Bruijn, B., Norton, A. C., Winner, E., & Schlaug, G. (2013). Training-mediated leftward
asymmetries during music processing: A cross-sectional and longitudinal fMRI analysis.
NeuroImage 75, 97–107.
Escoffier, N., & Tillmann, B. (2008). The tonal function of a task-irrelevant chord modulates speed
of visual processing. Cognition 107(3), 1070–1083.
Fox, M. D., & Raichle, M. E. (2007). Spontaneous fluctuations in brain activity observed with
functional magnetic resonance imaging. Nature Reviews Neuroscience 8(9), 700–711.
Friston, K. (2010). The free-energy principle: A unified brain theory? Nature Reviews Neuroscience
11, 127–138.
Fujioka, T., Ross, B., & Trainor, L. (2015). Beta-band oscillations represent auditory beat and its
metrical hierarchy in perception and imagery. Journal of Neuroscience 35(45), 15187–15198.
Goldenberg, J., Mazursky, D., & Solomon, S. (1999). Creative sparks. Science 285(5433), 1495–
1496.
Grahn, J. A. (2012). See what I hear? Beat perception in auditory and visual rhythms. Experimental
Brain Research 220(1), 51–61.
Grahn, J. A., & Brett, M. (2007). Rhythm and beat perception in motor areas of the brain. Journal of
Cognitive Neuroscience 19(5), 893–906.
Gromko, J. E., & Poorman, A. S. (1998). The effect of music training on preschoolers’ spatial-
temporal task performance. Journal of Research in Music Education 46(2), 173–181.
Habibi, A., Cahn, B. R., Damasio, A., & Damasio, H. (2016). Neural correlates of accelerated
auditory processing in children engaged in music training. Developmental Cognitive Neuroscience
21, 1–14.
Hafter, E. R., & Saberi, K. (2001). A level of stimulus representation model for auditory detection
and attention. Journal of the Acoustical Society of America 110, 1489. Retrieved from
https://doi.org/10.1121/1.1394220
Hafter, E. R., Sarampalis, A., & Loui, P. (2008). Auditory attention and filters. In W. Yost (Ed.),
Auditory perception of sound sources (pp. 115–142). Dordrecht: Springer.
Hafter, E. R., Schlauch, R. S., & Tang, J. (1993). Attending to auditory filters that were not
stimulated directly. Journal of the Acoustical Society of America 94, 743–747. Retrieved from
https://doi.org/10.1121/1.408203
Hansen, N. C., & Pearce, M. T. (2014). Predictive uncertainty in auditory sequence processing.
Frontiers in Psychology 5, 1052. Retrieved from https://doi.org/10.3389/fpsyg.2014.01052
Henry, M. J., Herrmann, B., Kunke, D., & Obleser, J. (2017). Aging affects the balance of neural
entrainment and top-down neural modulation in the listening brain. Nature Communications 8,
15801. doi:10.1038/ncomms15801
Henry, M. J., Herrmann, B., & Obleser, J. (2015). Selective attention to temporal features on nested
time scales. Cerebral Cortex 25(2), 450–459.
Hollinger, A., Steele, C., Penhune, V., Zatorre, R., & Wanderley, M. (2007). fMRI-compatible
electronic controllers. In Proceedings of the 7th international conference on New Interfaces for
Musical Expression (pp. 246–249). New York: ACM. doi:10.1145/1279740.1279790
James, T., Przysinda, E., Sampaio, G., Woods, K. J. P., Hewett, A., Morillon, B., & Loui, P. (2017).
Acoustic effects on oscillatory markers of sustained attention. Presentation at the International
Conference on Auditory Cortex. Banff, Canada.
James, W. (1890). The principles of psychology. New York: Henry Holt.
Jones, M. R. (1976). Time, our lost dimension: Toward a new theory of perception, attention, and
memory. Psychological Review 83(5), 323–355.
Jones, M. R., & Boltz, M. (1989). Dynamic attending and responses to time. Psychological Review
96(3), 459–491.
Jones, M. R., Moynihan, H., MacKenzie, N., & Puente, J. (2002). Temporal aspects of stimulus-
driven attending in dynamic arrays. Psychological Science 13(4), 313–319.
Justus, T., & List, A. (2005). Auditory attention to frequency and time: An analogy to visual local-
global stimuli. Cognition 98(1), 31–51.
Kleinmintz, O. M., Goldstein, P., Mayseless, N., Abecasis, D., & Shamay-Tsoory, S. G. (2014).
Expertise in musical improvisation and creativity: The mediation of idea evaluation. PLoS ONE 9,
e101568.
Koelsch, S., Gunter, T. C., Friederici, A. D., & Schröger, E. (2000). Brain indices of music
processing: Nonmusicians are musical. Journal of Cognitive Neuroscience 12(3), 520–541.
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11, 599–605.
Lamb, S. J., & Gregory, A. H. (1993). The relationship between music and reading in beginning
readers. Educational Psychology 13(1), 19–27.
Lartillot, O., & Toiviainen, P. (2007). A Matlab toolbox for musical feature extraction from audio. In
Proceedings of the 10th International Conference on Digital Audio Effects (pp. 237–244).
Bordeaux, France. Retrieved from http://dafx.labri.fr/main/papers/p237.pdf
Limb, C. J., & Braun, A. R. (2008). Neural substrates of spontaneous musical performance: An fMRI
study of jazz improvisation. PLoS ONE 3, e1679.
List, A., Justus, T., Robertson, L. C., & Bentin, S. (2007). A mismatch negativity study of local–
global auditory processing. Brain Research 1153, 122–133.
Liu, S., Chow, H. M., Xu, Y., Erkkinen, M. G., Swett, K. E., Eagle, M. W., … Braun, A. R. (2012).
Neural correlates of lyrical improvisation: An fMRI study of freestyle rap. Scientific Reports 2,
834. doi:10.1038/srep00834
Lolli, S., Lewenstein, A. D., Basurto, J., Winnik, S., & Loui, P. (2015). Sound frequency affects
speech emotion perception: Results from congenital amusia. Frontiers in Psychology 6. Retrieved
from https://doi.org/10.3389/fpsyg.2015.01340
Longuet-Higgins, H. C., & Lee, C. S. (1982). The perception of musical rhythms. Perception 11(2),
115–128.
Loui, P., Grent-’T-Jong, T., Torpey, D., & Woldorff, M. (2005). Effects of attention on the neural
processing of harmonic syntax in Western music. Cognitive Brain Research 25(3), 678–687.
Loui, P., Kroog, K., Zuk, J., Winner, E., & Schlaug, G. (2011). Relating pitch awareness to phonemic
awareness in children: Implications for tone-deafness and dyslexia. Frontiers in Psychology 2, 111.
Retrieved from https://doi.org/10.3389/fpsyg.2011.00111
Loui, P., & Wessel, D. (2007). Harmonic expectation and affect in Western music: Effects of attention
and training. Perception & Psychophysics 69(7), 1084–1092.
McCrae, R. R. (1987). Creativity, divergent thinking, and openness to experience. Journal of
Personality and Social Psychology 52(6), 1258–1265.
McPherson, M. J., Barrett, F. S., Lopez-Gonzalez, M., Jiradejvong, P., & Limb, C. J. (2016).
Emotional intent modulates the neural substrates of creativity: An fMRI study of emotionally
targeted improvisation in jazz musicians. Scientific Reports 6, 18460. doi:10.1038/srep18460
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in
Broca’s area: An MEG study. Nature Neuroscience 4, 540–545.
Marmel, F., Tillmann, B., & Dowling, W. J. (2008). Tonal expectations influence pitch perception.
Perception & Psychophysics 70(5), 841–852.
Mehr, S. A., Schachner, A., Katz, R. C., & Spelke, E. S. (2013). Two randomized trials provide no
consistent evidence for nonmusical cognitive benefits of brief preschool music enrichment. PloS
ONE 8(12), e82007.
Millett, D. (2001). Hans Berger: From psychic energy to the EEG. Perspectives in Biology and
Medicine 44(4), 522–542.
Moreno, S., Bialystok, E., Barac, R., Schellenberg, E. G., Cepeda, N. J., & Chau, T. (2011). Short-
term music training enhances verbal intelligence and executive function. Psychological Science
22(11), 1425–1433.
Morillon, B., & Baillet, S. (2017). Motor origin of temporal predictions in auditory attention.
Proceedings of the National Academy of Sciences 114(42), E8913–E8921.
Norgaard, M. (2017). Developing musical creativity through improvisation in the large performance
classroom. Music Educators Journal 103(3), 34–39.
Norgaard, M., Spencer, J., & Montiel, M. (2013). Testing cognitive theories by creating a pattern-
based probabilistic algorithm for melody and rhythm in jazz improvisation. Psychomusicology:
Music, Mind, and Brain 23(4), 243–254.
Nozaradan, S., Zerouali, Y., Peretz, I., & Mouraux, A. (2013). Capturing with EEG the neural
entrainment and coupling underlying sensorimotor synchronization to the beat. Cerebral Cortex
25(3), 736–747.
Patel, A. D. (2003). Language, music, syntax and the brain. Nature Neuroscience 6, 674–681.
Patel, A. D. (2011a). Why would musical training benefit the neural encoding of speech? The
OPERA hypothesis. Frontiers in Psychology 2, 142. Retrieved from
https://doi.org/10.3389/fpsyg.2011.00142
Patel, A. D. (2011b). Why does musical training benefit the neural encoding of speech? A new
hypothesis. Journal of the Acoustical Society of America 130, 2398. Retrieved from
https://doi.org/10.1121/1.3654612
Pinho, A. L., De Manzano, O., Fransson, P., Eriksson, H., & Ullén, F. (2014). Connecting to create:
Expertise in musical improvisation is associated with increased functional connectivity between
premotor and prefrontal areas. Journal of Neuroscience 34(18), 6156–6163.
Povel, D.-J., & Essens, P. (1985). Perception of temporal patterns. Music Perception: An
Interdisciplinary Journal 2(4), 411–440.
Przysinda, E., Zeng, T., Maves, K., Arkin, C., & Loui, P. (2017). Jazz musicians reveal role of
expectancy in human creativity. Brain and Cognition 119, 45–53.
Purves, D., Cabeza, R., Huettel, S. A., Labar, K. S., Platt, M. L., Woldorff, M. G., & Brannon, E. M.
(2008). Cognitive neuroscience. Sunderland: Sinauer Associates.
Rauscher, F. H., Shaw, G. L., & Ky, C. N. (1993). Music and spatial task performance. Nature
365(6447), 611.
Rentfrow, P. J., & Gosling, S. D. (2003). The do re mi’s of everyday life: The structure and
personality correlates of music preferences. Journal of Personality and Social Psychology 84(6),
1236–1256.
Runco, M. A., & Jaeger, G. J. (2012). The standard definition of creativity. Creativity Research
Journal 24(1), 92–96.
Sala, G., & Gobet, F. (2017a). Does far transfer exist? Negative evidence from chess, music, and
working memory training. Current Directions in Psychological Science 26(6), 515–520.
Sala, G., & Gobet, F. (2017b). When the music’s over: Does music skill transfer to children’s and
young adolescents’ cognitive and academic skills? A meta-analysis. Educational Research Review
20, 55–67.
Sawyer, K. (2011). The cognitive neuroscience of creativity: A critical review. Creativity Research
Journal 23(2), 137–154.
Schellenberg, E. G. (2004). Music lessons enhance IQ. Psychological Science 15(8), 511–514.
Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal
27(3), 379–423.
Shinn-Cunningham, B. G. (2008). Object-based auditory and visual attention. Trends in Cognitive
Sciences 12(5), 182–186.
Shirer, W. R., Ryali, S., Rykhlevskaia, E., Menon, V., & Greicius, M. D. (2012). Decoding subject-
driven cognitive states with whole-brain connectivity patterns. Cerebral Cortex 22(1), 158–165.
Silvia, P. J., Beaty, R. E., & Nusbaum, E. C. (2013). Verbal fluency and creativity: General and
specific contributions of broad retrieval ability (Gr) factors to divergent thinking. Intelligence
41(5), 328–340.
Slevc, L. R., Davey, N. S., Buschkuehl, M., & Jaeggi, S. M. (2016). Tuning the mind: Exploring the
connections between musical ability and executive functions. Cognition 152, 199–211.
Slevc, L. R., & Okada, B. M. (2015). Processing structure in language and music: A case for shared
reliance on cognitive control. Psychonomic Bulletin & Review 22(3), 637–652.
Slevc, L. R., Rosenberg, J. C., & Patel, A. D. (2009). Making psycholinguistics musical: Self-paced
reading time evidence for shared processing of linguistic and musical syntax. Psychonomic
Bulletin & Review 16(2), 374–381.
Sternberg, R. J., & Lubart, T. (1999). The concept of creativity: Prospects and paradigms. In R.
Sternberg (Ed.), Handbook of creativity 1 (pp. 3–15). Cambridge: Cambridge University Press.
Sternberg, R. J., Lubart, T. I., Kaufman, J. C., & Pretz, J. E. (2005). Creativity. In K. Holyoak & R.
G. Morrison (Eds.), The Cambridge handbook of thinking and reasoning (pp. 351–370).
Cambridge: Cambridge University Press.
Treisman, A. M., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive
Psychology 12(1), 97–136.
Varèse, E., & Wen-Chung, C. (1966). The liberation of sound. Perspectives of New Music 5(1), 11–
19.
White-Schwoch, T., Carr, K. W., Anderson, S., Strait, D. L., & Kraus, N. (2013). Older adults benefit
from music training early in life: Biological evidence for long-term training-driven plasticity.
Journal of Neuroscience 33(45), 17667–17674.
Woldorff, M. G., Gallen, C. C., Hampson, S. A., Hillyard, S. A., Pantev, C., Sobel, D., & Bloom, F.
E. (1993). Modulation of early sensory processing in human auditory cortex during auditory
selective attention. Proceedings of the National Academy of Sciences 90(18), 8722–8726.
Woldorff, M. G., & Hillyard, S. A. (1991). Modulation of early auditory processing during selective
listening to rapidly presented tones. Electroencephalography and Clinical Neurophysiology 79(3),
170–191.
Woldorff, M. G., Hillyard, S. A., Gallen, C. C., Hampson, S. R., & Bloom, F. E. (1998).
Magnetoencephalographic recordings demonstrate attentional modulation of mismatch-related
neural activity in human auditory cortex. Psychophysiology 35(3), 283–292.
Zeng, T., Przysinda, E., Pfeifer, C., Arkin, C., & Loui, P. (2017). Structural connectivity predicts
success in musical improvisation. Under review.
Zeng, T., Przysinda, E., Pfeifer, C., Arkin, C., & Loui, P. (2018). White matter connectivity reflects
success in musical improvisation. bioRxiv.
Zuk, J., Benjamin, C., Kenyon, A., & Gaab, N. (2014). Behavioral and neural correlates of executive
functioning in musicians and non-musicians. PLoS ONE 9, e99868.
CHAPT E R 13
N E U R A L C O R R E L AT E S O F
MUSIC AND EMOTION
PAT R I K N . JU S L I N A N D L A U R A S . S A K K A
M A : D
D
P M : A
T F
To explain emotional responses to music, we need to uncover the
psychological mechanisms that produce perceived or induced emotion.
Broadly speaking, the mechanism refers to those causal processes through
which an outcome is brought into being. In the present context, this
involves a functional (i.e., psychological) description of what the brain is
“doing” in principle (e.g., retrieving a memory). Such a process description
at the psychological level must not be confused with the separate question
of where in the brain the process is implemented, or with the
phenomenological experience it seeks to explain (Dennett, 1987).
Several authors have proposed possible mechanisms underlying
perception and induction of emotions in music, typically involving one or a
few possibilities (see Berlyne, 1971; Clynes, 1977; Juslin, 2001; Langer,
1957; Meyer, 1956; Scherer & Zentner, 2001; Sloboda & Juslin, 2001).
Space limitations prevent us from reviewing previous work here, but a
parsimonious way to organize current theory is provided by the ICINAS-
BRECVEMAC framework, fully described in Juslin (2019) and briefly
summarized below.
Emotion Perception
The first part of the acronym ICINAS-BRECVEMAC stands for Iconic-
Intrinsic-Associative, and refers to three ways in which music carries
emotional meaning. Although the case can be made that emotion perception
is a more straightforward process than emotion induction, even perceived
emotions may need to be decomposed into different subprocesses.
Accordingly, based on the seminal distinction made by Dowling and
Harwood (1986), Juslin (2013b) proposes that there are three distinct
“layers” of musical expression of emotion. Each layer corresponds to a
specific type of coding of emotional meaning (see Fig. 1).
FIGURE 1. Multiple-layer conceptualization of musical expression of emotions.
Reproduced from Patrik N. Juslin, What does music express? Basic emotions and beyond,
Frontiers in Psychology: Emotion Science 4(596), Figure 2, doi: 10.3389/fpsyg.2013.00596
© 2013 Juslin. This work is licensed under the Creative Commons Attribution License (CC
BY 3.0). It is attributed to the author Patrik N. Juslin.
The core layer is based on iconically coded basic emotions. Icon refers
to how music carries emotional meaning based on a formal resemblance
between the music and other events that have an emotional tone (such as
emotional speech and gesture). This core layer may explain findings of
cross-modal parallels (Juslin & Laukka, 2003) and universal recognition of
basic emotions (i.e., sadness, happiness, anger, fear, and love/tenderness) in
both speech (Bryan & Barrett, 2008) and music (Fritz et al., 2009).
The core layer may be extended, qualified, and even modified by two
additional layers based on intrinsic and associative coding, respectively,
which enable listeners to perceive also more complex or ambiguous
emotions. The two additional layers are less cross-culturally invariant and
depend more on the context and the listener’s individual learning (Juslin,
2019). Intrinsic coding refers to how music carries meaning based on
syntactic relationships within the music itself, how one part of the music
may “refer” to another part of the music, thus contributing to shifting levels
of stability, tension, or arousal (“affective trajectories”; e.g., Spitzer, 2013).
Associative coding, finally, refers to how music carries emotional
meaning based on a more arbitrary association (e.g., temporal or spatial
contiguity); a piece of music can be perceived as expressive of an emotion
just because something in the music (e.g., a melodic theme) has been
repeatedly linked with other emotionally meaningful events in the past—
either through chance or by design (e.g., Wagner’s “Leitmotif” strategy; see
Dowling & Harwood, 1986).
To illustrate this further in a musical piece, the overall emotion category
or broad “emotional tone” (e.g., sadness) might be specified by iconically
coded features (e.g., slow tempo, minor mode, low and often falling pitch
contour, legato articulation); this basic emotion category is given
“expressive shape” by intrinsically coded features (e.g., local structural
features such as syncopations, dissonant intervals, and melodic
appoggiaturas), creating “tension” and “release,” which contribute to more
time-dependent and complex nuances of the same emotion category (e.g.,
sadness vs. hopelessness); to this we add the final and more personal layer
of expression (e.g., that the listener associates the piece with a particular
person, event, or physical location).
It appears plausible that the three sources of perceived emotions—which
might occur alone or in combination—involve partly different neural
correlates (Juslin, 2019).
Emotion Induction
Our main focus in this chapter will be on induced emotion, which appears
to be more complex in terms of its neural substrates. Here, a multi-
mechanism framework is clearly called for. The second part of the ICINAS-
BRECVEMAC acronym refers to nine psychological mechanisms for
induction of emotions (listed below), which may be activated by music (and
other stimuli).
An evolutionary perspective on human perception of sounds suggests
that the survival of our ancient ancestors depended on their ability to detect
patterns in sounds, derive meaning from them, and adjust their behavior
accordingly (Juslin, 2013a; cf. Hodges & Sebald, 2011). This behavioral
function can be achieved in a multitude of ways, reflecting the phylogenetic
origin of our emotions.
The human brain did not develop from scratch. It is the result of a long
evolutionary process, during which newer brain structures were gradually
imposed on older structures (Gärdenfors, 2003). Brain circuits are laid out
like the concentric layers of an onion, functional layer upon functional
layer. One consequence of this arrangement, which is the result of natural
selection rather than design, is that emotion can be evoked at multiple levels
of the brain (Juslin, 2019).
Hence, the first author of this chapter has postulated a set of induction
mechanisms involving (more or less) distinct brain networks, which have
developed gradually and in a specific order during evolution—from simple
reflexes to complex judgments. Different mechanisms rely on different
kinds of mental representation (e.g., associative, analogical,
sensorimotoric), which serve to guide future action. All mechanisms have
in common that they can be triggered by a “musical event” (broadly defined
as music, listener, and context). The mechanisms are:
General Overview
In this section, we summarize seventy-eight neuropsychological studies,
published between 1982 and 2016 (see Appendix table). Studies have been
grouped with regard to methodology: PET/fMRI (38 studies, 49 percent),
EEG (22 studies, 28 percent), lesions (16 studies, 20 percent), and dichotic
listening (2 studies, 3 percent). They are described in terms of listeners,
musical stimuli, contrast/design, method, main findings, and type of affect
(e.g., measuring induced vs. perceived emotions; categories, dimensions,
preferences). The categorization concerning induced vs. perceived emotion
is not entirely straightforward, because brain studies do not always
distinguish the processes in the design. (Previous reviews of the field have
tended to inter-mix studies that focus on different aspects, induced vs.
perceived emotion.)
Sample size varies depending on method—PET/fMRI (M = 16.31), EEG
(M = 32.00), lesions (M = 14.44), and dichotic listening (M = 18.00)—but
tends to be relatively small overall. Note that PET/fMRI and EEG studies
have focused mostly on induced emotion, whereas lesion and dichotic
listening studies have focused mostly on perceived emotion. Blood flow
studies have used mostly fMRI (as opposed to PET) and “real” (as opposed
to synthesized) music, and have mostly adopted dimensional (66 percent) as
opposed to discrete (34 percent) approaches to emotion. EEG studies have
also used mostly “real” music—but have adopted dimensional (34 percent)
and discrete (31 percent) approaches to a roughly equal degree. Lesion
studies have (in contrast to other studies) used mainly synthesized music,
and have mostly studied discrete emotions (75 percent), rather than
dimensions (38 percent). Such differences between studies that use different
methods should clearly be kept in mind when interpreting the overall
results.
Empirical Approaches
The “contrast” and “emotion” columns in the appendix table are suggestive
of the kind of empirical approach adopted in the study. Some early studies
tended to use an open-ended exploratory approach, which simply presents
listeners with supposedly “emotional” music, to see which regions might be
affected. Although such an approach was defensible in the early stages, it
makes it difficult to interpret the results (e.g., “It is not possible to
disentangle the different subcomponents of the activation due to limitations
of this experimental design,” Alfredson, Risberg, Hagberg, & Gustafson,
2004, p. 165). Thus, for instance, it may not be clear whether the study has
measured perceived or induced emotion, in the absence of control
conditions or converging measures.
We identify at least five possible approaches in the neuropsychological
study of emotions, which can serve different aims. These have been
adopted, implicitly or explicitly, in music studies to highly varying degrees.
We briefly summarize these approaches, before looking closer at the actual
data.
To the best of our knowledge, no music study thus far has directly
contrasted perception and induction of emotion, but attempts to interpret
data along those lines have been made (Juslin & Sloboda, 2001, p. 456). We
review further evidence below. The preliminary conclusion of this approach
is that “perception and induction of emotions may involve different patterns
of brain activation.”
Not all of the studies seem to follow this pattern, however—at least on first
view. A problem is that in some cases, it is difficult to know for sure
whether a study has measured perceived or evoked emotion, since multi-
component indices were not used. For instance, it is an open question
whether ratings of pleasantness of music in some studies are just that
(ratings of the stimuli) or whether they index feelings of pleasure. If there is
insufficient control over which process is actually elicited in studies, this
can explain the mixed findings. We submit that the results suggest some
degree of specificity in terms of valence, but the nature of these patterns
and their interpretation remain contested. Yet, a preliminary conclusion of
this approach is that “neural correlates can distinguish the valence of
musically aroused affect.”
Over time, one may discern a change from basic lateralization studies (e.g.,
dichotic listening) and a search for individual brain structures to a
consideration of more complex and distributed networks. But we are not
aware of any study of neural correlates so far that contrasts different
psychological mechanisms. (We will consider such an approach later in the
chapter.) Thus, the increasing awareness of the role of mechanisms has not
yet translated into concrete designs. This becomes clear when we take a
closer look at the findings.
Summary of Brain Imaging Data
At the current stage, the data that are potentially most informative when it
comes to pinpointing neural correlates of musical emotion come from the
(38) brain imaging studies conducted to date. Tables 1 and 2 summarize the
main findings for perceived and induced emotion, respectively, in terms of
broad brain areas for which blood-flow changes have been reported. Ideally,
the interpretation should be made in terms of “networks” (Bressler &
Menon, 2010), rather than “isolated” regions, but current results do not yet
enable such interpretations.
Some broad conclusions can be drawn based on the findings. First,
music listening can cause changes in blood flow in “core” regions for
emotional processing. Second, as noted by Peretz (2010, p. 119), “there is
not a single, unitary emotional system underlying all emotional responses to
music.” On the contrary, a fairly broad range of cortical and subcortical
brain regions seem to be linked to musical emotions. Most of these belong
to the (extended) limbic system and include the amygdala, the
hippocampus, the striatum (including nucleus accumbens), the cingulate
cortex, the insula, the prefrontal and orbitofrontal cortex, the cerebellum,
the frontal gyrus, the parahippocampal gyrus, and various brainstem
structures.
The data in Tables 1 and 2 also enable us to compare induced and
perceived emotions. As may be seen, there is some overlap between the
brain regions reported. This could reflect two things: (a) that there is some
extent of overlap in the neural correlates of these processes or (b) that
studies have not sufficiently distinguished between the processes—such that
some studies that ostensibly focus on induced emotion have measured
perceived emotion and vice versa; or that some studies measure both
processes at the same time—leading to “noisy” data. Few studies have
measured multiple components of emotion so as to enhance the validity of
conclusions about induced emotions (discussed at the beginning of this
chapter).
In any case, note that there are certain differences in the findings for the
two processes: Only for induction of emotion have several studies reported
changes in the amygdala, the striatum (including nucleus accumbens), and
the hippocampus. At least some of these areas may thus distinguish induced
emotions from mere perception, though studies that directly contrast the
two processes under controlled conditions are clearly required to confirm
this hypothesis.1
Beyond these simple and relatively trivial conclusions, interpretations of
the findings tend to become more difficult and “impressionistic” in nature.
Given a general lack of “process-pure” manipulations of mechanisms,
researchers have to rely on “informed speculations” about the possible role
of different brain structures and networks. These are typically based on
general knowledge of the brain, but tend to be relatively vague. This is
because the analyses involve very broad brain areas which have been
proposed to be involved in a wide range of different psychological
processes; that is, they have poor “selectivity” (Poldrack, 2006) when it
comes to “revealing” specific psychological processes.
Koelsch (2014, p. 172) submits that observed changes in the amygdala
“could be because music is perceived as a stimulus with social significance
owing to its communicative properties.” This is, indeed, one possibility—
but we really do not know. And even if this notion is correct, it does not
offer very precise information about the functional role of the amygdala. An
additional problem is that this form of “reverse inference” about cognitive
process is not deductively valid (Poldrack, 2006). Normally, we would infer
from brain imaging data that “when cognitive process X is active, then
brain area Y is active”—not the other way around.
Let us be clear: This is not a matter of competence. Informed
speculations and interpretations by distinguished neuroscientists like Stefan
Koelsch or Isabelle Peretz are as good as they get. The problem is rather
that in the absence of process-specific experimental manipulation in the
field as a whole, theoretical interpretations are rendered difficult for a
number of reasons.
A first problem is that brain imaging “cannot disentangle correlation
from causation” (Peretz, 2010, p. 114); a related problem is that results
from imaging studies tend to be “overinclusive” (Peretz, 2010, p. 114);
therefore, “it is not always easy to determine if the activity is related to
emotional or non-emotional processing of the musical structure” (Peretz,
2010, p. 112). Indeed, the same brain structure can serve different roles both
within and across domains (Kreutz & Lotze, 2007). In addition, as implied
by the ICINAS-BRECVEMAC framework, cognition and emotion are not
neatly separated in the brain: specific cognitive processes may be involved
depending on the mechanism responsible for the perceived or induced
emotion.
The specific listener task (self-report of felt affect, ratings of melodies,
or mere listening) may also affect the patterns of brain
activation/deactivation, and so may differences with respect to the music
stimuli (“real” vs. “synthesized” music, “familiar” vs. “unfamiliar,” “self-
selected” vs. “experimenter-selected”). All of these issues conspire to make
interpretations of findings from brain imaging studies problematic. This has
not prevented researchers from suggesting how to organize the findings
with regard to the processes of perception and induction, respectively.
Perception of Emotions
Double brain dissociations between emotional judgments and melody
recognition (Peretz & Gagnon, 1999), and between emotional judgments
and basic music perception (Peretz, Gagnon, & Bouchard, 1998), initially
lead Peretz (2001) to postulate an “emotion module,” dedicated to
perception of emotion in music. Subsequently, she proposed that a more
distributed network, originally evolved to process vocal emotions, has been
“invaded” by music, such that emotional speech and emotional music will
share neural resources (Peretz, 2010). This idea has received some support
(Escoffier, Zhong, Schirmer, & Qui, 2013) and is in line with documented
parallels in emotions between speech and music (Juslin & Laukka, 2003).
Studies on emotions in speech suggest a network of areas primarily in the
(right) frontal and parietal lobes, including the inferior frontal gyrus
(Schirmer & Kotz, 2006).
The possibility of cross-modal parallels can be explored in the context of
the present results (Table 1). For perceived emotions, the most frequently
reported regions are frontal areas (73 percent of studies) and the frontal
gyrus (45 percent). Note that Escoffier et al. (2013) found that tracing of
emotions in both speech and music was related to activity in the medial
SFG. Moreover, Nair, Large, Steinberg, and Kelso (2002) discovered that
listening to expressive (as compared to “mechanical”) music performances
increased activity in the right inferior frontal gyrus. These findings seem
consistent with the “shared-resources hypothesis” (further evidence of a
shared neural code was recently reported by Paquette, Takerkart, Saget,
Peretz, & Belin, 2018).
There are some additional brain regions implicated in emotion-
perception studies. Curiously, there are three studies (27 percent) that report
changes in the cerebellum during perceived emotion, and five studies (45
percent) that report changes in the anterior cingulate cortex (which occurs
also in evoked emotion; cf. Table 2). We return to these findings later.
It has further been argued that the perception of dissonance is linked to
the parahippocampal gyrus (Blood et al., 1999). This notion receives
support from lesion studies showing that this basic ability suffers after
damage to the parahippocampal gyrus (Gosselin et al., 2006). Only two of
eleven studies (18 percent) report changes in the amygdala (see Table 1)—
though it has been found that recognition of “scary” music suffers after
damage to the amygdala (Gosselin et al., 2005). It cannot be completely
ruled out that the two studies really measured evoked emotion rather than
just perceived (since they featured unpleasant stimuli that may have evoked
some negative emotion). In summary, the most consistent results are that
perception of emotions in music involves the frontal cortex and the frontal
gyrus—and, perhaps, some right hemisphere lateralization (Bryden, Ley, &
Sugarman, 1982).
Induction of Emotions
For induction of emotions, a larger number of brain regions have been
reported (Table 2). The most frequently reported areas include the amygdala
(63 percent of studies), the frontal cortex (70 percent; Pfc 37 percent), the
ventral striatum/NAc (44 percent), the hippocampus (52 percent), the insula
(48 percent), and the anterior cingulate cortex (41 percent). However, note
that the results vary a lot from study to study, in ways that are not easy to
explain. For example, it may be seen in Table 2 that there are numerous
additional regions that were reported in only one or a few studies. These
include the parahippocampus, the thalamus, the basal ganglia, the
cerebellum, motor regions, and the brainstem.
One approach to this problem is to look for areas that are consistently
activated across studies in the hope that this will reveal an emotion network
that is invariably involved in the process. Thus, for instance, Koelsch and
colleagues (Koelsch, Siebel, & Fritz, 2010) argue that a network consisting
of the amygdala, the hippocampus, the parahippocampus, the temporal
poles, and the pregenual cingulate cortex may play a consistent role in
emotional processing of music. But is there support for the idea of a set of
brain regions that are consistently activated?
Close inspection of Table 2 reveals that few brain regions are reported in
more than about half of the studies which purported to measure induced
emotions. If areas are not consistently found to be influenced, how is this to
be interpreted? Some of this variability is surely due to methodological
problems and consequent measurement error. This could include differences
in how regions of interest (ROI) are defined, or in the assumptions made in
the analysis. But assuming that limbic regions were “prime suspects” in the
analyses, the variability is still too large to be accounted for by (only) this
factor.
In principle, one may argue that if these studies have tried to measure
emotion and the listed regions are not consistently activated across studies,
then either these areas are not related to emotions, or these studies have not
consistently managed to induce any emotion. However, a different
interpretation suggested by the BRECVEMAC framework (and supported
by meta-analyses of “general” emotion findings; cf. Lindquist et al., 2012)
is that the variability is due to different psychological mechanisms being
activated in different investigations (depending on the musical stimuli, the
listeners, and the situation, as well as the experimental procedure). This
possibility is elaborated in the following section.
T M P A
C R : A F N
A ?
1
The dopaminergic mesolimbic reward pathway involving nucleus accumbens is a prime suspect
when it comes to positive emotions. The hippocampus and the amygdala are both considered key
areas for emotional memories, which will presumably not be involved as much during comparatively
“neutral” perception of emotions.
CHAPT E R 14
NEUROCHEMICAL
R E S P O N S E S TO M U S I C
Y U K O K O S H I MO R I
D S
Dopamine (DA) is synthesized in the cytosol of catecholaminergic neurons
in the ventral tegmental area (VTA) and in the substantia nigra pars
compacta (SNpc) of the brain. The VTA DA sends projections to the ventral
striatum/nucleus accumbens (NAc), amygdala, and hippocampus as well as
medial prefrontal cortex such as orbitofrontal and anterior cingulate cortices
whereas the SN DA sends projections to the dorsal striatum (i.e., putamen
and caudate nucleus). These DA pathways are commonly referred to as
mesocorticolimbic and nigro-striatal dopamine pathways, respectively. The
former pathway is associated with emotional/motivational functions
whereas the latter pathway is more involved in the executive/cognitive and
sensorimotor functions (Solís & Moratalla, 2018). There are five dopamine
receptor subtypes, which are classified based on their functional properties
and subdivided into D1-like and D2-like families. The D1-like family
consists of D1 and D5 receptors and the D2-like family consists of D2, D3,
and D4 receptors. DA is also synthesized in the adrenal medulla and acts as
a hormone along with other catecholamines (see section “Norepinephrine
Systems”).
Numerous functional neuroimaging studies have demonstrated that
music alters brain activity in the DA pathways associated with
reward/motivation (Menon & Levitin, 2005; Salimpoor, van den Bosch,
Kovacevic, McIntosh, & Dagher, 2013), emotion/pleasure (Koelsch, 2014;
Mueller et al., 2015; Salimpoor, Benovoy, Larcher, Dagher, & Zatorre,
2011), as well as motor functions (Grahn & Rowe, 2009). However, to date
there has been only one study that investigated dopaminergic transmission
in the ventral and dorsal striatum associated with musical pleasure
(Salimpoor et al., 2011). This study employed PET with [11C]raclopride
that is a D2/D3 receptor antagonist and found that greater DA release
occurred in bilateral dorsal and ventral striatum, but notably in the right
caudate nucleus and the right NAc during listening to pleasurable music
selected by the participants compared to neutral music. Furthermore, the
greater DA release in the right caudate nucleus was associated with the
greater number of peak pleasure or “chills” whereas greater DA release in
the right NAc was associated with more intense chills experienced. These
anatomically distinct roles of the subregions in music listening were further
elucidated by analyzing their temporal brain activation using functional
magnetic resonance imaging (fMRI). The increased brain activity in the
right caudate nucleus occurred several seconds prior to experiencing the
pleasurable peak while the enhanced activity in the right NAc occurred
during the pleasurable moments. The authors interpreted these findings as
indicating that the former structure is involved in the anticipation and
prediction of pleasure and the latter structure, in experiencing pleasure. This
study demonstrated that musical pleasure is associated with DA release in
the ventral striatum, particularly in the NAc. However, as the individuals
who regularly experience “chills” during music listening were selected to
participate in the study, further research is needed to investigate whether
DA is also released during listening to pleasurable music in those who have
never experienced “chills.”
In addition to musical pleasure, the role of DA in music perception and
auditory-motor entrainment was investigated in Parkinson’s disease (PD)
that is primarily characterized by loss of dopaminergic neurons in the SNpc,
resulting in depletion of dopaminergic input to the dorsal striatum. The
involvement of DA in these functions was investigated when people with
PD were on and off dopaminergic medication. One study showed that
dopaminergic medication improved music perception (Cameron, Pickett,
Earhart, & Grahn, 2016). However, it is unknown whether the improvement
was due to the practice effect that was also observed in the healthy
participants or the effect of the medication. Another study using PET with
[11C]-DTBZ that binds dopamine transporters (VMAT2) did not find strong
association between dopaminergic denervation and auditory-motor task
performance (Miller et al., 2013). However, when the people with PD were
grouped based on the similarity of dopaminergic denervation, the auditory-
motor synchronization accuracy paralleled the pattern of denervation. On
the other hand, two fMRI studies did not implicate the role of DA during
auditory-motor entrainment (Elsinger et al., 2003; Jahanshahi et al., 2010).
In addition to these central dopaminergic functions, dopamine-
associated gene expression and peripheral dopaminergic levels were also
investigated. The expression of alpha-synuclein (SNCA) that maintains DA
neuronal homeostasis (Murphy, Rueter, Trojanowski, & Lee, 2000;
Oczkowska, Kozubski, Lianeri, & Dorszewska, 2014) was upregulated in
professional musicians after a two-hour concert performance compared to
after a music-free control condition (Kanduri, Kuusi, et al., 2015), as well
as in listeners with musical experiences longer than 10 years and those with
high musical aptitude after listening to 20-minutes of classical music
(Kanduri, Raijas, et al., 2015). The latter study also reported that the
upregulation was absent after a music-free control condition or in listeners
with no significant musical experience (Kanduri, Kuusi, et al., 2015;
Kanduri, Raijas, et al., 2015).
A few studies investigated the dopaminergic levels in the peripheral
samples as well as associated psychological measures using an auditory
stimulus and music. Two studies reported decreased dopaminergic levels:
one in the urine sample following daily listening to binaural beats for 60
days in healthy adults, who also reported decreased trait anxiety (Wahbeh,
Calabrese, & Zwickey, 2007) and the other study in the plasma sample
following a 12-week dance movement therapy (DMT) that combines music,
light exercise, and sensory stimulation in female adolescents with mild
depression whose psychological distress was also reduced (Jeong et al.,
2005). One study reported no change in the plasma dopaminergic level
following listening to music (high-uplifting music or low-uplifting) in
healthy adults who had performed a stress-induced task (Hirokawa & Ohira,
2003). However, it is unknown whether the stress-induced task affected the
plasma dopaminergic level before listening to music.
To summarize, there is some evidence that music enhances
dopaminergic function. Listening to pleasurable music induces dopamine
release, and music upregulates the SNCA expression that may facilitate
dopaminergic neurotransmission. However, these responses may occur only
in specific people (i.e., those with extensive musical training or those who
regularly experience “chills” by listening to pleasurable music). Further
studies are needed to investigate these effects in individuals with varying
music education/training levels and listening habits and experiences, as well
as using different genres of music. Music may also be able to reduce the
peripheral DA levels and psychological disturbances. Future studies
including both clinical and healthy control participants are needed to clarify
these effects. In addition, some PD studies are suggestive of the role of DA
in music perception and auditory cuing. Further pharmacological studies in
people with PD are needed to address the limitations of previous studies for
clarification.
E O S
Central endogenous opioid systems (EOS) include three opioid peptides
including β-endorphin, enkephalins, and dynorphin and their three
receptors, Mu (μ), Delta (δ), and Kappa (κ) (Benarroch, 2012). Neurons
containing β-endorphin are localized in two main areas, the arcuate nucleus
of the hypothalamus and nucleus tractus solitarius tract of the brainstem,
which send widespread projections to the rest of the brain. Enkephalin and
dynorphin are primarily located in local neurons. Opioid receptors are
widely but differentially expressed throughout the central nervous system
(CNS): the μ receptor is the most abundant receptor in the cerebral cortex,
amygdala, thalamus, brainstem, dorsal horn and dorsal root ganglion (DRG)
neurons; the δ receptor is mostly expressed in the olfactory system,
striatum, limbic cortex, dorsal horn and DRG neurons; and the κ receptor in
the claustrum, striatum, hippocampus, hypothalamus, brainstem, and dorsal
horn. The EOS involves various functions such as reward, pain modulation,
stress responses, and autonomic control.
In the previous section, the study by Salimpoor et al. (2011) showed that
musical pleasure was associated with DA release in the ventral striatum.
However, the central EOS likely plays a primary role in the positive affect
or pleasure induced by music, which acts synergistically with DA (Chanda
& Levitin, 2013). Both endogenous and exogenous opioids can activate DA
neurons in the VTA (Hjelmstad, Xia, Margolis, & Fields, 2013) that
innervates the NAc. Both EOS and dopamine systems are involved in
reward mechanisms, which consist of liking/pleasure (core reactions to
hedonic experience of receiving a reward), wanting (motivational aspect),
and learning (association and cognitive representations) (Berridge &
Kringelbach, 2015), but animal research favors the notion that the EOS but
not the DA system generates pleasure. Specifically, stimulation of μ opioid
receptors in the rostrodorsal part of the medial shell in the NAc, the
posterior ventral pallidum, as well as anterior orbitofrontal cortex and
posterior insula (Castro & Berridge, 2017) enhances pleasure. In fact, two
studies reported that blocking opioid receptors attenuated musical thrills
(Goldstein, 1980) and pleasure (Mallik, Chanda, & Levitin, 2017) in
response to participant-selected music.
In addition to the role of the central EOS in musical pleasure, several
studies investigated plasma levels of β-endorphin, which is released from
the anterior pituitary gland (see also the section “Neuroendocrine Systems
II”), acts as a hormone associated with stress (Kreutz, Murcia, & Bongard,
2012), and therefore functions differently from the β-endorphin in the CNS
(Veening, Gerrits, & Barendregt, 2012).
Listening to techno-music increased the plasma β-endorphin level,
accompanied with increases in other psychophysiological measures as well
as changes in emotional states whereas listening to classical music did not
affect them (Gerra et al., 1998). Interestingly, this study also showed the
association between the β-endorphin responses to music and personal traits.
Higher β-endorphin responses were associated with less novel seeking.
In contrast to enhanced EOS responses, a decrease in the plasma
concentration of β-endorphin was reported in response to experimenter-
selected relaxation music, accompanied with a reduction in worries, fear,
and blood pressure in coronary patients (Vollert, Störk, Rose, & Möckel,
2003) as well as after a single one-hour singing session in choirs affected by
cancer including carers, bereaved carers, and patients (Fancourt, Williamon,
et al., 2016). In this latter study, the β-endorphin level showed negative
correlations with the levels of immune biomarkers and a positive
correlation with another stress biomarker. Another study reported a
decrease in response to experimenter-selected classical music and imagery,
but not to music only or imagery only (McKinney, Tims, Kumar, & Kumar,
1997).
In summary, musical pleasure is associated with the central EOS and
music induces changes in the plasma concentration of β-endorphin. As the
EOS plays an important role in various functions and has two different
functional systems (i.e., central and peripheral systems), more research is
needed to replicate and extend existing literature. Suggested future studies
include PET studies that investigate both EOS and dopamine systems
associated with musical pleasure/reward, studies that investigate the effects
of different genres of music and personal traits on the release of β-
endorphin, as well as studies that investigate the effects of music on EOS
associated with pain modulation (i.e., μ opioid receptors in the central pain
network; Benarroch, 2012) and stress regulation including both healthy
participants and those with pain and stress. When the cerebrospinal fluid
(CSF) or peripheral β-endorphin levels are assessed, the diurnal fluctuations
should be taken into account. Moreover, it should be noted that responses of
the plasma β-endorphin little reflect those in the CSF although they are not
entirely independent (Veening et al., 2012).
S S
N S I (P
P )
Two neuropeptides released from the posterior pituitary are oxytocin (OT)
and arginine vasopressin (AV). They are highly conserved across species
(Johnson & Young, 2017) and modulate social behaviors (Bachner-Melman
& Ebstein, 2014), including social cognition (Donaldson & Young, 2008)
and social affiliation (Insel, 2010), as well as reproductive behaviors. They
are also implicated in psychiatric disorders such as autism spectrum
disorder (Bachner-Melman & Ebstein, 2014; Donaldson & Young, 2008).
OT and AV are collectively called nonapeptides because they are composed
of nine amino acid residues (Acher & Chauvet, 1995). They are
predominantly synthesized in the magnocellular neurons in the
hypothalamic supraoptic and paraventricular nuclei and released centrally
and peripherally into the circulation through the posterior pituitary (Johnson
& Young, 2017) and thereby act as neuromodulators or neurohormones
(Bachner-Melman & Ebstein, 2014; Donaldson & Young, 2008). Although
several nonapeptide receptors are identified in the brain, OT receptor
(OTR), vasopressin receptor 1a (V1aR), and vasopressin receptor 1b
(V1bR) have been a major focus of investigation. These nonapeptide
receptors are expressed throughout auditory and mesolimbic pathways
(Johnson & Young, 2017).
Oxytocin
Several studies investigated peripheral OT responses to musical activities.
A single 30-minute singing lesson increased the serum OT level compared
to the baseline level in both professional and amateur singers (Grape,
Sandgren, Hansson, Ericson, & Theorell, 2003). Compared to a chatting
group, a singing group showed increase in the salivary OT level and
improvement of psychological well-being (Kreutz et al., 2012). In another
study, the plasma OT level increased in a small sample of four singers after
improvised singing, but it did not change after pre-composed singing
(Keeler et al., 2015). Furthermore, a group of boys with mild emotional
disturbance aged between 8 and 12 years showed an increased level of
salivary OT in the free session of group drumming compared to the practice
session, which was not observed in the same age group of girls or an older
aged group of boys (Yuhi et al., 2017).
In contrast to these findings, two studies reported that group singing
reduced the OT levels. One study found a decrease in the salivary OT level
after choir singing (Schladt et al., 2017). However, this change was not
observed after solo singing in the same participants. Instead, the OT level
increased after solo singing. Another study reported that singing in a single
70-minute choir rehearsal was associated with a decrease in the salivary OT
level across three populations affected by cancer (Fancourt, Williamon, et
al., 2016).
In addition to these studies with musical activities, the effect of passive
listening was investigated in two studies. Elevated plasma OT level was
reported in cardiac surgery patients who listened passively to experimenter-
selected “soothing” music (soft, relaxing of 60–80 beats per minute with a
volume of 50–60 dB) for 30 minutes one day after the surgery, but not in
those who rested without listening to music (Nilsson, 2009). Elevated
plasma OT level was also observed in participants with Williams Syndrome
(WS) who listened to their favorite music that elicited positive emotions
(Dai et al., 2012).
Arginine Vasopressin
Arginine vasopressin receptor 1A (AVPR1A) is one of the main genes that
have been associated with musical activities and related behaviors in
genome-wide linkage and association studies (Bachner-Melman et al.,
2005; Granot et al., 2007; Mariath et al., 2017; Ukkola et al., 2009; Ukkola-
Vuoti et al., 2011). The AVPR1A microsatellites have been associated with
musical working memory (Granot et al., 2007; but also see Granot,
Uzefovsky, Bogopolsky, & Ebstein, 2013), musical aptitude (Ukkola et al.,
2009), active music listening (Ukkola-Vuoti et al., 2011), and a wide range
of musical abilities (e.g., musical abilities associated with tempo, rhythm,
dynamics, vocality, and pitch, as well as creativity and development of
musical ideas and accompaniment) (Mariath et al., 2017). Except for these
genetic studies, the relationships between AV and music have been little
explored. The only study that has measured the AV level was the study by
Dai et al. (2012) mentioned above, which also found an increase in the AV
level in participants with WS.
In summary, music induces the peripheral OT responses and there is
some genetic association between AVR and music. However, the directional
changes are not consistent among those OT studies. The elevated OT levels
are generally implicated in positive social experiences (Chanda & Levitin,
2013). However, OT is also released in response to various kinds of stress
(Brown, Cardoso, & Ellenbogen, 2016; de Jong et al., 2015; Pierrehumbert
et al., 2010). The reduction in the OT level may reflect lower arousal and
stress during choir singing (Schladt et al., 2017). Taking blood samples may
cause stress and increase the OT level in some participants at the baseline
measurement, confounding the findings. Alternatively, the inconsistent
findings may also be partially derived from different sampling methods.
Some studies measured the OT levels in plasma and others measured the
salivary OT. These peripheral levels are used as a proxy for the central OT
levels. However, there are no strong correlations between central (CSF) and
peripheral measures as well as between the peripheral measures (Carson et
al., 2015; Hoffman, Brownley, Hamer, & Bulik, 2012; Javor et al., 2014;
Lefevre et al., 2017; Valstad et al., 2017). Other possible factors explaining
the inconsistencies include the influence by gonadal steroid (Insel, 2010)
and subjective experiences of the musical activities employed in the study
(Yuhi et al., 2017). The baseline measurement in a healthy control group is
also needed to understand the directional changes and interactions when
clinical populations are studied.
To date, there has been only one study that investigated the effects of
music on both neuropeptides in a clinical population and no study in
healthy participants. For future studies measuring both neuropeptides, it
should be noted that OT and AV show similar directional changes for some
social behaviors such as pair bonding (Caldwell, 2017), whereas they show
different effects in some cases, and opposite effects for aggression (Ferris,
1992; MacLean et al., 2017), anxiety and stress (Bachner-Melman &
Ebstein, 2014; Heinrichs, von Dawans, & Domes, 2009), and social
approach (Thompson & Walton, 2004). In fact, several electrophysiological
experiments revealed their differential regulations of excitatory projections
in the limbic system (Campbell-Smith, Holmes, Lingawi, Panayi, &
Westbrook, 2015; Huber, Veinante, & Stoop, 2005; Lubin, Elliot, Black, &
Johns, 2003; Numan et al., 2010). Furthermore, animal research suggests
that neuroanatomical distribution of their receptors may be critical for
determining function.
N S II (A
P )
A neuroendocrine system, commonly referred to as the hypothalamic-
pituitary-adrenal (HPA) axis releases cortisol as its main effector hormone
(Spencer, Chun, Hartsock, & Woodruff, 2018). Cortisol plays an important
role in circadian and stress regulation. The basal cortisol levels fluctuate in
a circadian fashion in the absence of stressors and the levels rise in response
to acute physical or psychological stressors as well as a circadian
entrainment. Circadian and stress-induced cortisol secretion is determined
by the neurohormone, corticotropin releasing factor (CRF) produced in and
secreted from the medial paraventricular nucleus of the hypothalamus. In
response to CRF, the anterior pituitary produces and secretes
adrenocorticotropin hormone (ACTH) and ß-endorphin (see also the section
“Endogenous Opioid Systems”). Triggered by the ACTH, the cortisol is
synthesized in the adrenal cortex. It passively diffuses into the adrenal vein
and is carried throughout the circulatory system. In addition to CRF,
vasopressin (AVP) is also involved in the process of the secretion.
The measured cortisol levels discussed in this section are primarily
salivary cortisol unless mentioned otherwise. Salivary cortisol is a valid and
reliable measure for unbound hormone in blood (Kirschbaum &
Hellhammer, 1994). Cortisol has been most studied as a stress biomarker in
response to music (Chanda & Levitin, 2013; Fancourt, Ockelford, & Belai,
2014; Hodges, 2010). There is a general consensus that relaxing music
regardless whether it is experimenter- or participant-selected reduces
cortisol levels (Beaulieu-Boire et al., 2013; Chanda & Levitin, 2013; Chen,
Sung, Lee, & Chang, 2015; Fancourt et al., 2014; Hodges, 2010; Jayamala,
Lakshmanagowda, Pradeep, & Goturu, 2015; Kreutz et al., 2012; Mejía-
Rubalcava, Alanís-Tavira, Mendieta-Zerón, & Sánchez-Pérez, 2015; but
also see null findings by Chen et al., 2015; Chlan, Engeland, & Savik,
2013; Good et al., 2013; Tan, McPherson, Peretz, Berkovic, & Wilson,
2014). However, when experimenter-selected relaxing music and
participant-selected music from a choice of genres were compared,
participant-selected music was more effective in reducing the cortisol level
by showing the prolonged effect post-surgery (Leardi et al., 2007). Another
study suggests that the sound of rippling water may be more effective than
relaxing music in reducing the cortisol level (Thoma et al., 2013). However,
neither was significantly different from the control condition without
acoustic stimulation.
In addition to the effect of relaxing music, several studies investigated
the effect of stimulating music on the cortisol levels. Five studies reported
that stimulating music also reduced the cortisol levels in female adolescents
with chronic depression (Field et al., 1998), in surgical patients (Koelsch et
al., 2011), in participants with hypertensives (Möckel et al., 1995), in
dancers (Quiroga Murcia, Kreutz, Clift, & Bongard, 2010), in participants
with lung infection (le Roux, Bouic, & Bester, 2007), and in healthy males
(Ooishi, Mukai, Watanabe, Kawato, & Kashino, 2017), whereas other
studies found an increase in healthy participants (Brownley, McMurray, &
Hackney, 1995; Gerra et al., 1998; Hébert, Béland, Dionne-Fournelle,
Crête, & Lupien, 2005; Karageorghis et al., 2017). These results suggest
that stimulating music can either attenuate or enhance the cortisol level,
which may depend on participant characteristics and/or their preference of
music.
Furthermore, there are some studies suggesting that music in general
may reduce the cortisol levels. For example, both participant-selected chill-
inducing music and music they disliked significantly reduced the cortisol
levels in both male and female participants (Fukui & Toyoshima, 2013).
Additionally, listening to music regardless of genre (Mozart, Strauss, and
ABBA) led to a significant reduction in the serum cortisol concentrations,
which was also significantly lower compared to those in the silence
condition (Trappe & Voit, 2016). Furthermore, another study reported that
both repetitive drumming and instrumental meditation music decreased the
cortisol levels (Gingras, Pohler, & Fitch, 2014). Taken together, cortisol
appears to be responsive to music in general.
Cortisol responses to music were also investigated in surgical patients
pre- and post- as well as during surgery. Listening to participant-selected
and experimenter-selected music during and post-surgery prevented the
cortisol level from increasing, and/or decreased the cortisol level post-
surgery (Graversen & Sommer, 2013; Nilsson, 2009; Schneider,
Schedlowski, Schürmeyer, & Becker, 2001; Tabrizi, Sahraei, & Rad, 2012;
but also see Lin, Lin, Huang, Hsu, & Lin, 2011). Comparing different
periods of time of music listening, one study reported that listening to
experimenter-selected relaxing music following the surgery was most
effective in reducing the level of serum cortisol relative to listening to
music in the pre- and peri-operative periods (Nilsson, Unosson, & Rawal,
2005). Altogether, these studies suggest that listening to music is beneficial
to surgical patients by reducing cortisol levels. Schneider et al. (2001)
reported that more than a majority of patients in the music group thought
that the beneficial effect of music was a distraction. Thus, how music exerts
the beneficial effect on the cortisol level and other behavioral measures in
surgical patients needs further investigation.
The counteracting effect of music on the elevated cortisol levels induced
by acute stressors was also studied in younger healthy participants.
Experimenter-selected relaxing music helped to reduce the cortisol level
immediately following a psychological stressor whereas the silence
condition led to an increase during the same recovery period (Khalfa, Dalla
Bella, Roy, Peretz, & Lupien, 2003), suggesting that relaxing music
facilitates faster recovery from the stressor. On the other hand,
experimenter-selected relaxing music did not lower the cortisol levels after
the exposure to a psychological stressor whereas it prevented stress-induced
increases in heart rate, systolic blood pressure, and anxiety compared to the
silence condition (Knight & Rickard, 2001). Another study used an acute
physiological stressor and demonstrated that tapping to the experimenter-
selected positive music post-stressor was associated with more positive
mood and stronger cortisol responses (i.e., increase) compared to tapping to
the neural music (Koelsch et al., 2016). The positive mood was also
associated with the greater cortisol response to the acute stressor in the
music group. Authors interpreted these findings as indicating that the
stronger cortisol response may reflect an early sign of immuno-enhancing
response to the acute stressor, but not a higher stress level because the
music group overall had a more positive mood (Koelsch et al., 2016). There
was no effect of music on the level of ACTH in this study. The inconsistent
findings of these studies may be partly due to different types of stressor and
how music was applied in these studies.
Moreover, the effects of group musical activities on endocrine responses
have been studied. Singing was associated with a reduction in endocrine
responses (Fancourt, Aufegger, & Williamon, 2015; Fancourt, Williamon,
et al., 2016; Schladt et al., 2017). Cortisol reduction was greater for choir
singing than solo singing, accompanied with a reduction in the salivary OT
(Schladt et al., 2017). In addition, the effect of group singing on the
endocrine responses was modulated by the conditions of performance
(Fancourt et al., 2015). More specifically, the reduced levels in cortisol and
cortisone were only observed in the low-stress condition (singing without
an audience) compared to the high-stress condition (singing in a live
concert). On the other hand, no endocrine changes were found following a
single session of group drumming (Bittman et al., 2001) or multiple
sessions of group drumming (Fancourt, Perkins, et al., 2016) in healthy
participants.
Music therapy, in which musical and other activities are led by a
therapist, showed mixed results. Following guided imaginary and music
(GIM) therapy that combines relaxation techniques and listening to classical
music, the cortisol level was reduced to be lower compared to the silence
condition in healthy participants (McKinney et al., 1997) and in individuals
on sick-leave (Beck, Hansen, & Gold, 2015). On the other hand, there were
no endocrine responses to an individualized music therapy in older healthy
adults (Suzuki, Kanamori, Nagasawa, Tokiko, & Takayuki, 2007); to an
individualized music therapy or to a multisensory stimulation environment
including auditory stimulation in older adults with severe dementia
(Valdiglesias et al., 2017); or to movement music therapy in older healthy
adults (Shimizu et al., 2013).
One study investigated the social effect of music listening on the cortisol
level (Linnemann, Strahler, & Nater, 2016). Listening to music in the
presence of others (mostly friends), but not listening alone, attenuated the
secretion of cortisol. However, the presence of others alone significantly
explained the variance in the cortisol level. In addition, the findings of this
study should be interpreted with caution since the time intervals between
music listening and the cortisol measurement were unknown.
Interestingly, listening to music for relaxation was associated with
significant reductions in the subjective stress level and in the cortisol
concentration in healthy participants (Linnemann, Ditzen, Strahler, Doerr,
& Nater, 2015). Moreover, the reduction in the cortisol level was not
associated with the perception of music as relaxing. The authors
emphasized the importance of non-musical, contextual factors such as
reasons for music listening. It would be interesting to compare the cortisol
response to a non-musical control activity for relaxation. This study also
showed that listening to music for distraction increased the stress level,
which contrasted the findings in surgical patients (Schneider et al., 2001).
This may be due to differences in participants’ characteristics and/or
circumstances.
The effects of music on endocrine measures may be mediated by sex.
For example, testosterone showed opposite responses between men and
women. Music decreased testosterone in men but it increased in women
(Fukui & Toyoshima, 2013; Fukui & Yamashita, 2003). In addition, music
may have differential effects on men and women. In one study, after
strenuous exercise men and women showed different trajectory of the
cortisol levels during a recovery period with music. This was observed
regardless of musical tempo (Karageorghis et al., 2017). In another study,
the cortisol level decreased more steeply in men relative to women in both
choir and solo singing (Schladt et al., 2017). In contrast, other studies did
not find any sex effect on the level of cortisol associated with music
listening (Fukui & Yamashita, 2003; Nater, Abbruzzese, Krebs, & Ehlert,
2006).
The studies discussed above included adult participants. Several studies
investigated the cortisol responses to music in younger age groups. In
schoolchildren, extra two-hour musical activities including singing,
moving, dancing, or playing instruments during a school year resulted in a
reduction of the cortisol level measured in the afternoon at the end of school
year. However, this result reached a statistical significance only when a one-
tail t-test was used (Lindblad, Hogmark, & Theorell, 2007). In preterm
infants, exposure to live instrumental music reduced the cortisol level along
with improvement of other measures for oxygen desaturations, apneas, and
pain (Schwilling et al., 2015). On the other hand, recorded lullabies did not
affect the cortisol level or sleep–awake behavior (Dorn et al., 2014).
Another study also did not find any effect of recorded lullaby combined
with touch on the cortisol level (Qiu et al., 2017). This study, however,
showed that following the intervention, blood ß-endorphin was significantly
increased, accompanied with decreased pain responses.
To summarize, it is relatively conclusive that music reduces the cortisol
level. The beneficial effects of music may be associated with distraction
from aversive states (Chanda & Levitin, 2013) in the context of acute
stressors (Linnemann, Kappert, et al., 2015) and/or listener’s intention of
music listening (Linnemann, Ditzen, et al., 2015). Further studies are
needed to clarify how music exerts beneficial effects on stress biomarkers.
The endocrine responses are primarily studied associated with stress in
which multiple factors can affect the findings, for example depending on
whether the stressor is either acute or chronic (Koelsch et al., 2016) or
whether it is psychological, physiological, or physical. In addition,
appropriate stress response differs depending on the circadian phase
(Spencer et al., 2018). Therefore, more studies warrant further elucidation
of the effect of music on stress responses.
N S
P I S
The immune system functions to protect and defend the body against
infection and damage from foreign organisms and toxins, while maintaining
checks and balances to prevent self-reactivity. It has two branches: innate
and adaptive immunity. The innate immune responses occur immediately
following an insult and are the first component of the immune system to be
activated against invasion (Turvey & Broide, 2010), and include activation
of immune cells such as granulocyte and monocytes/macrophages, and
secretion of pro-inflammatory cytokines such as interleukin IL-1β, IL-6,
tumor necrosis factor alpha TNF-α, and interferon-gamma IFN-γ to
upregulate the acute inflammatory response. In contrast, the adaptive
immune system, consisting of B cells and T cells, is slower acting with its
responses occurring days to weeks after exposure. Unlike the innate
immune system, the adaptive immune system is capable of memory and is
able to adjust in response to pathogens. Anti-inflammatory cytokines
include IL-1Ra, IL-4, IL-6, IL-10, IL-11 and IL-13, and TNF-β that
modulate the inflammatory immune response to prevent the harmful effects
of prolonged immune system activation. It should be noted that the immune
cells such as natural killer (NK) cells and dendritic cells cannot be clearly
defined as innate or adaptive and that some cytokines have both pro- and
anti-inflammatory properties depending on the amount of the cytokines
expressed, the length of time they are expressed, or which form of the
receptors the cytokines activate (Rainville, Tsyglakova, & Hodes, 2018).
Immune Cells
Group drumming led to an increase in the levels of NK cell in healthy
adults (Bittman et al., 2001) and increased CD4+ T cell, and memory T cell
counts only in older adults, but not in younger adults (Koyama et al., 2009).
In contrast, listening to experimenter-selected relaxing music during surgery
decreased the levels of NK cell, which was not observed in patients who
chose their music from preselected music (Leardi et al., 2007).
Cytokines
Among cytokines, IL-6 (that presents both pro- and anti-inflammatory
properties) has been most researched associated with music. Music therapy
sessions using relaxing music reduced the IL-6 levels, which was
accompanied with reductions in SNS biomarkers in surgical patients
(Conrad et al., 2007), and in the elderly with cerebrovascular disease and
dementia (Okada et al., 2009). Experimenter-selected classical music also
decreased the IL-6 level among older adults who liked the genre of music,
which was accompanied with an increase in the expression of μ opioid
receptors (Stefano, Zhu, Cadet, Salamon, & Mantione, 2004) whereas it did
not change the levels of other cytokines. On the other hand, group
drumming exercises led to an increase in the IL-6 level, along with
increased levels of pro-inflammatory IFN-γ in older adults, which was not
observed in younger adults (Koyama et al., 2009). Because Koyama et al.
(2009) also reported increased CD4+ T cell and memory T cell counts only
in older adults, the increased IL-6 level may be anti-inflammatory.
Although it appeared that IL-6 showed “the greatest levels of
responsiveness” (Fancourt et al., 2014, p. 18), more recent studies showed
otherwise (Beaulieu-Boire et al., 2013; Fancourt, Perkins, et al., 2016;
Fancourt, Williamon, et al., 2016; Koelsch et al., 2016). Further research is
needed to determine whether or not IL-6 is a sensitive immune biomarker in
response to music.
Other cytokines also showed responses to music. The level of anti-
inflammatory IL-1 was increased, along with the cortisol reduction in
response to participant-selected music compared to the control conditions
(Bartlett, Kaufman, & Smeltekop, 1993). In another study, anti-
inflammatory IL-4 was increased, accompanied with a reduction in a pro-
inflammatory marker, monocyte chemoattractant protein (MCP) in response
to multiple group drumming sessions (Fancourt, Perkins, et al., 2016).
Another study reported increased inflammatory markers including pro-
inflammatory IL-2 and soluble IL2 receptor α; anti-inflammatory IL-4; and
IL-17 that displays both pro- and anti-inflammatory profiles, along with
improved affects and reductions in the cortisol, ß-endorphin, and OT levels
in response to a single session of singing in choirs affected by cancer
(Fancourt, Williamon, et al., 2016). One study found that Mozart, but not
Beethoven or Schubert, downregulated the levels of anti-inflammatory IL-
4, 10 and 13 and upregulated the levels of pro-inflammatory cytokine such
as IFN-γ and IL-12, which was also associated with alleviated allergic skin
responses (Kimata, 2003). The findings reported by Kimata (2003) may
reflect the enhancement of pro-inflammatory responses induced by music,
which was similar to the increased cortisol responses following the acute
physiological stressor (Koelsch et al., 2016).
Immunoglobulin A
Along with these peripheral immune biomarkers, immunoglobulin A (IgA)
is one of the most commonly studied immune biomarkers associated with
music. Immunoglobulin A is a major serum immunoglobulin that is
predominantly produced in the bone marrow and mediates various
protective functions through interaction with specific receptors and immune
mediators (Woof & Ken, 2006). Immunoglobulin A is also a principal
antibody class in the external secretions that bathe vast mucosal surfaces of
the gastrointestinal, respiratory, and genitourinary tracts and plays an
important role in first line immune protection. Secretory and serum IgA are
different biochemical and immunochemical properties produced by cells
with different organ distributions. Therefore, different methods of
immunization can induce either secretory or serum IgA responses or a
combination of both.
In general, research has yielded consistent results: music increases the
concentrations or secretion rate of secretory IgA (S-IgA) (Chanda &
Levitin, 2013; Fancourt et al., 2014; Hodges, 2010), suggesting that music
enhances immunity in healthy individuals. Furthermore, the S-IgA increase
was greater when engaging in group singing compared to passive listening
(Beck, Cesario, Yousefi, & Enamoto, 2000; Kreutz, Bongard, Rohrmann,
Hodapp, & Grebe, 2004; Kuhn, 2002). Another study showed that S-IgA
was increased only in repose to “designer music” that brings positive
feelings, but not to relaxing (new age) or rock music (McCraty, Atkinson, &
Rein, 1996).
However, there are also a few studies reporting no changes in the levels
of IgA. In two studies, the serum levels of IgA did not change in patients
who listened to experimenter-selected calming music post-surgery (Nilsson
et al., 2005) or joyful music (which was described to the patients as
“relaxing” acoustic stimulation to reduce noise) before, during, or after
surgery (Koelsch et al., 2011). No music effect on the plasma IgA
concentrations may be due to the effects of local anesthetic infiltration
(Nilsson et al., 2005) or due to differences in response to music between S-
IgA and serum IgA (Woof & Ken, 2006). Furthermore, two studies reported
no changes following stressors such as eating adverse/allergic food (Kejr et
al., 2010) or a stressful cognitive task (Hirokawa & Ohira, 2003). The
immunoenhancement effect of music may be limited in healthy individuals
without exposure to stressors.
To summarize, there is some evidence that music induces changes in
immune biomarkers. S-IgA appears to respond most consistently and
robustly to music in healthy individuals. Music-induced increase of S-IgA
is interpreted as immunoenhancement. Future studies can investigate how
long this effect lasts and whether music experiences and habits modulate
the effect. Although music induces responses of other immune biomarkers,
the interpretations can be challenging due to the inconsistency in the
directional changes of cytokines with different inflammatory properties.
An interesting observation from animal research is that individual
differences in the peripheral immune system influence the development of
stress susceptibility, demonstrated by higher circulating levels of IL-6 and
leukocytes in susceptible mice compared to resilient and control mice
(Hodes et al., 2014; Rainville et al., 2018). Therefore, it may be useful to
separate participants depending on the baseline level of immune
biomarkers. Furthermore, as immune biomarkers are closely connected with
hormones (Yovel, Shakhar, & Ben-Eliyahu, 2001), sex may need to be
accounted for in the study design.
C S
D F D
R
Acher, R., & Chauvet, J. (1995). The neurohypophysial endocrine regulatory cascade: Precursors
mediators, receptors, and effectors. Frontiers in Neuroendocrinology 16(3), 237–289.
Albers, H. E. (2015). Species, sex and individual differences in the vasotocin/vasopressin system:
Relationship to neurochemical signaling in the social behavior neural network. Frontiers in
Neuroendocrinology 36, 49–71.
Bachner-Melman, R., Dina, C., Zohar, A. H., Constantini, N., Lerer, E., Hoch, S., … Ebstein, R. P.
(2005). AVPR1a and SLC6A4 gene polymorphisms are associated with creative dance
performance. PLoS Genetics 1(3), 394–403. Retrieved from
https://doi.org/10.1371/journal.pgen.0010042
Bachner-Melman, R., & Ebstein, R. P. (2014). The role of oxytocin and vasopressin in emotional and
social behaviors. Handbook of Clinical Neurology 124, 53–68.
Bandelow, B., Baldwin, D., Abelli, M., Bolea-Alamanac, B., Bourin, M., Chamberlain, S. R., …
Riederer, P. (2017). Biological markers for anxiety disorders, OCD and PTSD: A consensus
statement. Part II: Neurochemistry, neurophysiology and neurocognition. World Journal of
Biological Psychiatry 18(3), 162–214.
Barrett, F. S., Preller, K. H., Herdener, M., Janata, P., & Vollenweider, F. X. (2017). Serotonin 2A
receptor signaling underlies LSD-induced alteration of the neural response to dynamic changes in
music. Cerebral Cortex (December), 1–12. Retrieved from https://doi.org/10.1093/cercor/bhx257
Bartlett, D., Kaufman, D., & Smeltekop, R. (1993). The effects of music listening and perceived
sensory experience on the immune system as measured by interleukin-1 and cortisol. Journal of
Music Therapy 30(4), 194–209.
Beaulieu-Boire, G., Bourque, S., Chagnon, F., Chouinard, L., Gallo-Payet, N., & Lesur, O. (2013).
Music and biological stress dampening in mechanically-ventilated patients at the intensive care
unit ward: A prospective interventional randomized crossover trial. Journal of Critical Care 28(4),
442–450.
Beck, B. D., Hansen, A. M., & Gold, C. (2015). Coping with work-related stress through guided
imagery and music (GIM): Randomized controlled trial. Journal of Music Therapy 52(3), 323–
352.
Beck, R. J., Cesario, T. C., Yousefi, A., & Enamoto, H. (2000). Choral singing, performance
perception, and immune system changes in salivary immunoglobulin A and cortisol. Music
Perception: An Interdisciplinary Journal 18(1), 87–106.
Benarroch, E. E. (2012). Endogenous opioid systems: Current concepts and clinical correlations.
Neurology 79, 807–814.
Berridge, K. C., & Kringelbach, M. L. (2015). Pleasure systems in the brain. Neuron 86(3), 646–664.
Bittman, B., Berk, L., Felten, D., Westengard, J., Simonton, O., Pappas, J., & Ninehouser, M. (2001).
Composite effects of group drumming music therapy on modulatin of neuroendocrine-immune
parameters in normal subjects. Alternative Therapies 7(1), 38–47.
Brown, C. A., Cardoso, C., & Ellenbogen, M. A. (2016). A meta-analytic review of the correlation
between peripheral oxytocin and cortisol concentrations. Frontiers in Neuroendocrinology 43, 19–
27.
Brownley, K. A., McMurray, R. G., & Hackney, A. C. (1995). Effects of music on physiological and
affective responses to graded treadmill exercise in trained and untrained runners. International
Journal of Psychophysiology 19(3), 193–201.
Caldwell, H. K. (2017). Oxytocin and vasopressin: Powerful regulators of social behavior.
Neuroscientist 23(5), 517–528.
Cameron, D. J., Pickett, K. A., Earhart, G. M., & Grahn, J. A. (2016). The effect of dopaminergic
medication on beat-based auditory timing in Parkinson’s disease. Frontiers in Neurology 7, 1–8.
Retrieved from https://doi.org/10.3389/fneur.2016.00019
Campbell-Smith, E. J., Holmes, N. M., Lingawi, N. W., Panayi, M. C., & Westbrook, R. F. (2015).
Oxytocin signaling in basolateral and central amygdala nuclei differentially regulates the
acquisition, expression, and extinction of context-conditioned fear in rats. Learning & Memory
22(5), 247–257.
Carson, D. S., Berquist, S. W., Trujillo, T. H., Garner, J. P., Hannah, S. L., Hyde, S. A., … Parker, K.
J. (2015). Cerebrospinal fluid and plasma oxytocin concentrations are positively correlated and
negatively predict anxiety in children. Molecular Psychiatry 20(9), 1085–1090.
Castro, D. C., & Berridge, K. C. (2017). Opioid and orexin hedonic hotspots in rat orbitofrontal
cortex and insula. Proceedings of the National Academy of Sciences 114(43), E9125–E9134.
Chanda, M. L., & Levitin, D. J. (2013). The neurochemistry of music. Trends in Cognitive Sciences
17(4), 179–191.
Chen, C. J., Sung, H. C., Lee, M. S., & Chang, C. Y. (2015). The effects of Chinese five-element
music therapy on nursing students with depressed mood. International Journal of Nursing Practice
21(2), 192–199.
Chen, J., Nakamura, M., Kawamura, T., Takahashi, T., & Nakahara, D. (2006). Roles of
pedunculopontine tegmental cholinergic receptors in brain stimulation reward in the rat.
Psychopharmacology 184(3–4), 514–522.
Chlan, B. L. L., Engeland, W. C., & Anthony, A. (2007). Influence of music on the stress response in
patients receiving mechanical ventilatory support: A pilot study. American Journal of Critical
Care 16(2), 141–146.
Chlan, L. L., Engeland, W. C., & Savik, K. (2013). Does music influence stress in mechanically
ventilated patients? Intensive and Critical Care Nursing 29(3), 121–127.
Ciurleo, R., Di Lorenzo, G., Bramanti, P., & Marino, S. (2014). Magnetic resonance spectroscopy:
An in vivo molecular imaging biomarker for Parkinson’s disease? BioMed Research International
2014, 519816. Retrieved from https://doi.org/10.1155/2014/519816
Conrad, C., Niess, H., Jauch, K., Bruns, C., Hartl, W., & Welker, L. (2007). Overture for growth
hormone: Requiem for interleukin-6? Critical Care Medicine 35(12), 2709–2713.
Dai, L., Carter, C. S., Ying, J., Bellugi, U., Pournajafi-Nazarloo, H., & Korenberg, J. R. (2012).
Oxytocin and vasopressin are dysregulated in Williams syndrome, a genetic disorder affecting
social behavior. PLoS ONE 7(6), e38513. Retrieved from
https://doi.org/10.1371/journal.pone.0038513
de Jong, T. R., Menon, R., Bludau, A., Grund, T., Biermeier, V., Klampfl, S. M., … Neumann, I. D.
(2015). Salivary oxytocin concentrations in response to running, sexual self-stimulation,
breastfeeding and the TSST: The Regensburg Oxytocin Challenge (ROC) study.
Psychoneuroendocrinology 62, 381–388.
Djurovic, S., Le Hellard, S., Kähler, A. K., Jönsson, E. G., Agartz, I., Steen, V. M., … Andreassen,
O. A. (2009). Association of MCTP2 gene variants with schizophrenia in three independent
samples of Scandinavian origin (SCOPE). Psychiatry Research 168(3), 256–258.
Dölen, G., Darvishzadeh, A., Huang, K. W., & Malenka, R. C. (2013). Social reward requires
coordinated activity of nucleus accumbens oxytocin and serotonin. Nature 501(7466), 179–184.
Donaldson, Z. R., & Young, L. J. (2008). Oxytocin, vasopressin, and the neurogenetics of sociality.
Science 322(5903), 900–904. Correction (2009): Science 323(5920), 1429.
Dorn, F., Wirth, L., Gorbey, S., Wege, M., Zemlin, M., Maier, R. F., & Lemmer, B. (2014). Influence
of acoustic stimulation on the circadian and ultradian rhythm of premature infants. Chronobiology
International 31(9), 1062–1074.
Elsinger, C. L., Rao, S. M., Zimbelman, J. L., Reynolds, N. C., Blindauer, K. A., & Hoffmann, R. G.
(2003). Neural basis for impaired time reproduction in Parkinson’s disease: An fMRI study.
Journal of the International Neuropsychological Society 9(7), 1088–1098.
Evers, S., & Suhr, B. (2000). Changes of the neurotransmitter serotonin but not of hormones during
short time music perception. European Archives of Psychiatry and Clinical Neuroscience 250(3),
144–147.
Fancourt, D., Aufegger, L., & Williamon, A. (2015). Low-stress and high-stress singing have
contrasting effects on glucocorticoid response. Frontiers in Psychology 6, 1–5. Retrieved from
https://doi.org/10.3389/fpsyg.2015.01242
Fancourt, D., Ockelford, A., & Belai, A. (2014). The psychoneuroimmunological effects of music: A
systematic review and a new model. Brain, Behavior, and Immunity 36, 15–26.
Fancourt, D., Perkins, R., Ascenso, S., Carvalho, L. A., Steptoe, A., & Williamon, A. (2016). Effects
of group drumming interventions on anxiety, depression, social resilience and inflammatory
immune response among mental health service users. PLoS ONE 11(3), 1–16. Retrieved from
https://doi.org/10.1371/journal.pone.0151136
Fancourt, D., Williamon, A., Carvalho, L. A., Steptoe, A., Dow, R., & Lewis, I. (2016). Singing
modulates mood, stress, cortisol, cytokine and neuropeptide activity in cancer patients and carers.
Ecancermedicalscience 10, 1–13. Retrieved from https://doi.org/10.3332/ecancer.2016.631
Ferris, C. (1992). Role of vasopressin in aggressive and dominant/subordinate behaviors. Annals of
the New York Academy of Sciences 652, 212–226.
Field, T., Martinez, A., Nawrocki, T., Pickens, J., Fox, N., & Schanberg, S. (1998). Music shifts
frontal EEG in depressed adolescents. Adolescence 33(129), 109–116.
Frost, J. J., Douglass, K. H., Mayberg, H. S., Dannals, R. F., Links, J. M., Wilson, A. A., … Wagner,
H. N. (1989). Multicompartmental analysis of [11C]-Carfentanil binding to opiate receptors in
humans measured by positron emission tomography. Journal of Cerebral Blood Flow &
Metabolism 9(3), 398–409.
Fukui, H., & Toyoshima, K. (2013). Influence of music on steroid hormones and the relationship
between receptor polymorphisms and musical ability: A pilot study. Frontiers in Psychology 4, 1–
8. Retrieved from https://doi.org/10.3389/fpsyg.2013.00910
Fukui, H., & Yamashita, M. (2003). The effects of music and visual stress on testosterone and
cortisol in men and women. Neuro Endocrinology Letters 24(3–4), 173–180.
Gerra, G., Zaimovic, A., Franchini, D., Palladino, M., Giucastro, G., Reali, N., … Brambilla, F.
(1998). Neuroendocrine responses of healthy volunteers to “techno-music”: Relationships with
personality traits and emotional state. International Journal of Psychophysiology 28, 99–111.
Gingras, B., Pohler, G., & Fitch, W. T. (2014). Exploring shamanic journeying: Repetitive drumming
with shamanic instructions induces specific subjective experiences but no larger cortisol decrease
than instrumental meditation music. Plos ONE 9(7). Retrieved from
https://doi.org/10.1371/journal.pone.0102103
Goldstein, A. (1980). Thrills in response to music and other stimuli. Physiological Psychology 8(1),
126–129.
Good, M., Albert, J. M., Arafah, B., Anderson, G. C., Wotman, S., Cong, X., … Ahn, S. (2013).
Effects on postoperative salivary cortisol of relaxation/music and patient teaching about pain
management. Biological Research for Nursing 15(3), 318–329.
Graff-Guerrero, A., Willeit, M., Ginovart, N., Mamo, D., Mizrahi, R., Rusjan, P., … Kapur, S.
(2008). Brain region binding of the D2/3 agonist [11C]-(+)- PHNO and the D2/3 antagonist
[11C]raclopride in healthy humans. Human Brain Mapping 29(4), 400–410.
Grahn, J. A., & Rowe, J. B. (2009). Feeling the beat: Premotor and striatal interactions in musicians
and nonmusicians during beat perception. Journal of Neuroscience 29(23), 7540–7548.
Granot, R. Y., Frankel, Y., Gritsenko, V., Lerer, E., Gritsenko, I., Bachner-Melman, R., … Ebstein, R.
P. (2007). Provisional evidence that the arginine vasopressin 1a receptor gene is associated with
musical memory. Evolution and Human Behavior 28(5), 313–318.
Granot, R. Y., Uzefovsky, F., Bogopolsky, H., & Ebstein, R. P. (2013). Effects of arginine vasopressin
on musical working memory. Frontiers in Psychology 4, 1–12. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00712
Grape, C., Sandgren, M., Hansson, L., Ericson, M., & Theorell, T. (2003). Does singing promote
well-being? An empirical study of professional and amateur singers during a singing lesson.
Integrative Physiological & Behavioral Science 38(1), 65–74.
Graversen, M., & Sommer, T. (2013). Perioperative music may reduce pain and fatigue in patients
undergoing laparoscopic cholecystectomy. Acta Anaesthesiologica Scandinavica 57(8), 1010–
1016.
Griffiths, T. D., Uppenkamp, S., Johnsrude, I., Josephs, O., & Patterson, R. D. (2001). Encoding of
the temporal regularity of sound in the human brainstem. Nature Neuroscience 4(6), 633–637.
Gunn, R. N., Slifstein, M., Searle, G. E., & Price, J. C. (2015). Quantitative imaging of protein
targets in the human brain with PET. Physics in Medicine and Biology 60(22), R363–R411.
Hébert, S., Béland, R., Dionne-Fournelle, O., Crête, M., & Lupien, S. J. (2005). Physiological stress
response to video-game playing: The contribution of built-in music. Life Sciences 76(20), 2371–
2380.
Heinrichs, M., von Dawans, B., & Domes, G. (2009). Oxytocin, vasopressin, and human social
behavior. Frontiers in Neuroendocrinology 30(4), 548–557.
Hirokawa, E., & Ohira, H. (2003). The effects of music listening after a stressful task on immune
functions, neuroendocrine responses, and emotional states in college students. Journal of Music
Therapy 40(3), 189–211.
Hjelmstad, G. O., Xia, Y., Margolis, E. B., & Fields, H. L. (2013). Opioid modulation of ventral
pallidal afferents to ventral tegmental area neurons. Journal of Neuroscience 33(15), 6454–6459.
Hodes, G. E., Pfaua, M. L., Marylene Leboeufb, C., Goldena, S. A., Christoffela, D. J., Bregmana,
D., … Russo, S. J. (2014). Individual differences in the peripheral immune system promote
resilience versus susceptibility to social stress. Proceedings of the National Academy of Sciences
111(45), 16136–16141.
Hodges, D. (2010). Pyschophysiological measures. In P. Juslin & J. Sloboda (Eds.), Handbook of
music and emotion: Theory, research, applications (pp. 279–312). Oxford: Oxford University
Press.
Hoffman, E. R., Brownley, K. A., Hamer, R. M., & Bulik, C. M. (2012). Plasma, salivary, and
urinary oxytocin in anorexia nervosa: A pilot study. Eating Behaviors 13(3), 256–259.
Huber, D., Veinante, P., & Stoop, R. (2005). Vasopressin and oxytocin excite distinct neuronal
populations in the central amygdala. Science 308(5719), 245–248.
Hurley, L. M., & Sullivan, M. R. (2012). From behavioral context to receptors: Serotonergic
modulatory pathways in the IC. Frontiers in Neural Circuits 6, 1–17. Retrieved from
https://doi.org/10.3389/fncir.2012.00058
Hurley, R. A, Flashman, L. A., Chow, T. W., & Taber, K. H. (2010). The brainstem: Anatomy,
assessment, and clinical syndromes. Journal of Neuropsychiatry and Clinical Neuroscience 22(1),
2–6. Retrieved from https://doi.org/10.1176/appi.neuropsych.23.2.121
Insel, T. R. (2010). The challenge of translation in social neuroscience: A review of oxytocin,
vasopressin, and affiliative behavior. Neuron 65(6), 768–779.
Jahanshahi, M., Jones, C. R. G., Zijlmans, J., Katzenschlager, R., Lee, L., Quinn, N., … Lees, A. J.
(2010). Dopaminergic modulation of striato-frontal connectivity during motor timing in
Parkinson’s disease. Brain 133(3), 727–745.
Javor, A., Riedl, R., Kindermann, H., Brandstatter, W., Ransmayr, G., & Gabriel, M. (2014).
Correlation of plasma and salivary oxytocin in healthy young men: Experimental evidence. Neuro
Endocrinology Letters 35(6), 470–473.
Jayamala, A. K., Lakshmanagowda, P. B., Pradeep, G. C. M., & Goturu, J. (2015). Impact of music
therapy on breast milk secretion in mothers of premature newborns. Journal of Clinical and
Diagnostic Research 9(4), CC04–CC06.
Jeong, Y. J., Hong, S. C., Myeong, S. L., Park, M. C., Kim, Y. K., & Suh, C. M. (2005). Dance
movement therapy improves emotional responses and modulates neurohormones in adolescents
with mild depression. International Journal of Neuroscience 115(12), 1711–1720.
Johnson, Z. V., & Young, L. J. (2017). Oxytocin and vasopressin neural networks: Implications for
social behavioral diversity and translational neuroscience. Neuroscience & Biobehavioral Reviews
76, 87–98.
Jones, A. K. P., Cunningham, V. J., Ha-Kawa, S. K., Fujiwara, T., Liyii, Q., Luthra, S. K., … Jones,
T. (1994). Quantitation of [11C]diprenorphine cerebral kinetics in man acquired by PET using
presaturation, pulse-chase and tracer-only protocols. Journal of Neuroscience Methods 51(2), 123–
134.
Juslin, P. N., & Västfjäll, D. (2008). Emotional responses to music: The need to consider underlying
mechanisms. Behavioral and Brain Sciences 31(5), 559–575; discussion 575–621.
Kaelen, M., Barrett, F. S., Roseman, L., Lorenz, R., Family, N., Bolstridge, M., … Carhart-Harris, R.
L. (2015). LSD enhances the emotional response to music. Psychopharmacology 232(19), 3607–
3614.
Kaelen, M., Roseman, L., Kahan, J., Santos-Ribeiro, A., Orban, C., Lorenz, R., … Carhart-Harris, R.
(2016). LSD modulates music-induced imagery via changes in parahippocampal connectivity.
European Neuropsychopharmacology 26(7), 1099–1109.
Kanduri, C., Kuusi, T., Ahvenainen, M., Philips, A. K., Lähdesmäki, H., & Järvelä, I. (2015). The
effect of music performance on the transcriptome of professional musicians. Scientific Reports 5,
1–7. Retrieved from https://doi.org/10.1038/srep09506
Kanduri, C., Raijas, P., Ahvenainen, M., Philips, A. K., Ukkola-Vuoti, L., Lähdesmäki, H., & Järvelä,
I. (2015). The effect of listening to music on human transcriptome. PeerJ 3, e830. Retrieved from
https://doi.org/10.7717/peerj.830
Karageorghis, C. I., Bruce, A. C., Pottratz, S. T., Stevens, R. C., Bigliassi, M., & Hamer, M. (2017).
Psychological and psychophysiological effects of recuperative music postexercise. Medicine &
Science in Sports & Exercise 50(4), 739–746.
Katori, S., Hamada, S., Noguchi, Y., Fukuda, E., Yamamoto, T., Yamamoto, H., … Yagi, T. (2009).
Protocadherin-α family is required for serotonergic projections to appropriately innervate target
brain areas. Journal of Neuroscience 29(29), 9137–9147.
Keeler, J. R., Roth, E. A., Neuser, B. L., Spitsbergen, J. M., Waters, D. J. M., & Vianney, J.-M.
(2015). The neurochemistry and social flow of singing: Bonding and oxytocin. Frontiers in Human
Neuroscience 9, 1–10. Retrieved from https://doi.org/10.3389/fnhum.2015.00518
Kejr, A., Gigante, C., Hames, V., Krieg, C., Mages, J., König, N., … Diel, F. (2010). Receptive music
therapy and salivary histamine secretion. Inflammation Research 59(Suppl. 2), 217–218.
Khalfa, S., Dalla Bella, S., Roy, M., Peretz, I., & Lupien, S. J. (2003). Effects of relaxing music on
salivary cortisol level after psychological stress. Annals of the New York Academy of Sciences 999,
374–376.
Kimata, H. (2003). Listening to Mozart reduces allergic skin wheal responses and in vitro allergen-
specific IGE production in atopic dermatitis patients with latex allergy. Behavioral Medicine 29(1),
15–19.
Kirschbaum, C., & Hellhammer, D. H. (1994). Salivary cortisol in psychoneuroendocrine research:
Recent developments and applications. Psychoneuroendocrinology, 19(4), 313–333.
Knight, W. E. J., & Rickard, N. S. (2001). Relaxing music prevents stress-induced increase in
subjective anxiety, systolic blood pressure, and heart rate in healthy males and female. Journal of
Music Therapy 34(4), 254–272.
Koelsch, S. (2014). Brain correlates of music-evoked emotions. Nature Reviews Neuroscience 15(3),
170–180.
Koelsch, S., Boehlig, A., Hohenadel, M., Nitsche, I., Bauer, K., & Sack, U. (2016). The impact of
acute stress on hormones and cytokines, and how their recovery is affected by music-evoked
positive mood. Scientific Reports 6, 1–11. Retrieved from https://doi.org/10.1038/srep23008
Koelsch, S., Fuermetz, J., Sack, U., Bauer, K., Hohenadel, M., Wiegel, M., … Heinke, W. (2011).
Effects of music listening on cortisol levels and propofol consumption during spinal anesthesia.
Frontiers in Psychology 2, 1–9. Retrieved from https://doi.org/10.3389/fpsyg.2011.00058
Koyama, M., Wachi, M., Utsuyama, M., Bittman, B., Hirokawa, K., & Kitagawa, M. (2009).
Recreational music-making modulates immunological responses and mood states in older adults.
Journal of Medical and Dental Sciences 56, 79–90.
Koyama, Y., Jodo, E., & Kayama, Y. (1994). Sensory responsiveness of “broad-spike” neurons in the
laterodorsal tegmental nucleus, locus coeruleus and dorsal raphe of awake rats: Implications for
cholinergic and monoaminergic neuron-specific responses. Neuroscience 63(4), 1021–1031.
Kreutz, G., Bongard, S., Rohrmann, S., Hodapp, V., & Grebe, D. (2004). Effects of choir singing or
listening on secretory immunoglobulin A, cortisol, and emotional state. Journal of Behavioral
Medicine 27(6), 623–635.
Kreutz, G., Murcia, C., & Bongard, S. (2012). Psychoneuroendocrine research on music and health:
An overview. In R. MacDonald, G. Kreutz, & L. Mitchell (Eds.), Music, health, and wellbeing (pp.
457–476). Oxford: Oxford University Press.
Kuhn, D. (2002). The effects of active and passive participation in musical activity on the immune
system as measured by salivary immunoglobulin A (SIgA). Journal of Music Therapy 39(1), 30–
39.
Kumar, A., Tims, F., Cruess, D., Mintzer, M., Ironson, G., Loewenstein, D., … Kumar, M. (1999).
Music therapy increases serum melatonin levels in patients with Alzheimer’s disease. Alternative
Therapies in Health Medicine 5(9), 49–57.
Leardi, S., Pietroletti, R., Angeloni, G., Necozione, S., Ranalletta, G., & Del Gusto, B. (2007).
Randomized clinical trial examining the effect of music therapy in stress response to day surgery.
British Journal of Surgery 94(8), 943–947.
le Roux, F., Bouic, P., & Bester, M. (2007). The effect of Bach’s Magnificat on emotions, immune,
and endocrine parameters during physiotherapy treatment of patients with infectious lung
conditions. Journal of Music Therapy 44(2), 156–168.
Lefevre, A., Mottolese, R., Dirheimer, M., Mottolese, C., Duhamel, J. R., & Sirigu, A. (2017). A
comparison of methods to measure central and peripheral oxytocin concentrations in human and
non-human primates. Scientific Reports 7(1), 17222. Retrieved from
https://doi.org/10.1038/s41598-017-17674-7
Levitt, P., & Moore, R. Y. (1979). Origin and organization of brainstem catecholamine innervation in
the rat. Journal of Comparative Neurology 186(4), 505–528.
Lin, P. C., Lin, M. L., Huang, L. C., Hsu, H. C., & Lin, C. C. (2011). Music therapy for patients
receiving spine surgery. Journal of Clinical Nursing 20(7–8), 960–968.
Lindblad, F., Hogmark, Å., & Theorell, T. (2007). Music intervention for 5th and 6th graders: Effects
on development and cortisol secretion. Stress and Health 23(1), 9–14.
Linnemann, A., Ditzen, B., Strahler, J., Doerr, J. M., & Nater, U. M. (2015). Music listening as a
means of stress reduction in daily life. Psychoneuroendocrinology 60, 82–90.
Linnemann, A., Kappert, M. B., Fischer, S., Doerr, J. M., Strahler, J., & Nater, U. M. (2015). The
effects of music listening on pain and stress in the daily life of patients with fibromyalgia
syndrome. Frontiers in Human Neuroscience 9, 1–10. Retrieved from
https://doi.org/10.3389/fnhum.2015.00434
Linnemann, A., Strahler, J., & Nater, U. M. (2016). The stress-reducing effect of music listening
varies depending on the social context. Psychoneuroendocrinology 72, 97–105.
Lubin, D., Elliot, J., Black, M., & Johns, J. (2003). An oxytocin antagonist infused into the central
nucleus of the amygdala increases maternal aggressive behavior. Behavioral Neuroscience 117(2),
195–201.
McCraty, R., Atkinson, M., & Rein, G. (1996). Music enhances the effect of positive emotional states
on salivary IgA. Stress Medicine 12, 167–175.
McKinney, C. H., Tims, F. C., Kumar, A. M., & Kumar, M. (1997). The effect of selected classical
music and spontaneous imagery on plasma β-endorphin. Journal of Behavioral Medicine 20(1),
85–99.
MacLean, E. L., Gesquiere, L. R., Gruen, M. E., Sherman, B. L., Martin, W. L., & Carter, C. S.
(2017). Endogenous oxytocin, vasopressin, and aggression in domestic dogs. Frontiers in
Psychology 8. Retrieved from https://doi.org/10.3389/fpsyg.2017.01613
Mallik, A., Chanda, M. L., & Levitin, D. J. (2017). Anhedonia to music and mu-opioids: Evidence
from the administration of naltrexone. Scientific Reports 7, 1–8. Retrieved from
https://doi.org/10.1038/srep41952
Mariath, L. M., da Silva, A. M., Kowalski, T. W., Gattino, G. S., De Araujo, G. A., Figueiredo, F. G.,
… Schuch, J. B. (2017). Music genetics research: Association with musicality of a polymorphism
in the AVPR1A gene. Genetics and Molecular Biology 40(2), 421–429.
Maulina, T., Djustiana, N., & Shahib, M. N. (2017). The effect of music intervention on dental
anxiety during dental extraction procedure. The Open Dentistry Journal 11(1), 565–572.
Mejía-Rubalcava, C., Alanís-Tavira, J., Mendieta-Zerón, H., & Sánchez-Pérez, L. (2015). Changes
induced by music therapy to physiologic parameters in patients with dental anxiety.
Complementary Therapies in Clinical Practice 21(4), 282–286.
Menon, V., & Levitin, D. J. (2005). The rewards of music listening: Response and physiological
connectivity of the mesolimbic system. NeuroImage 28(1), 175–184.
Metherate, R. (2011). Functional connectivity and cholinergic modulation in auditory cortex.
Neuroscience & Biobehavioral Reviews 35(10), 2058–2063.
Migneault, B., Girard, F., Albert, C., Chouinard, P., Boudreault, D., Provencher, D., … Girard, D. C.
(2004). The effect of music on the neurohormonal stress response to surgery under general
anesthesia. Anesthesia & Analgesia 98(2), 527–532.
Miller, N. S., Kwak, Y., Bohnen, N. I., Müller, M. L. T. M., Dayalu, P., & Seidler, R. D. (2013). The
pattern of striatal dopaminergic denervation explains sensorimotor synchronization accuracy in
Parkinson’s disease. Behavioural Brain Research 257, 100–110.
Möckel, M., Störk, T., Vollert, J., Röcker, L., Danne, O., Hochrein, H., … Frei, U. (1995). Stress
reduction through listening to music: Effects on stress hormones, hemodynamics and mental state
in patients with arterial hypertension and in healthy persons. Deutsche Medizinische Wochenschrift
120(21), 745–752.
Moriizumi, T., & Hattori, T. (1992). Choline acetyltransferase-immunoreactive neurons in the rat
entopeduncular nucleus. Neuroscience 46(3), 721–728.
Morley, A. P., Narayanan, M., Mines, R., Molokhia, A., Baxter, S., Craig, G., … Craig, I. (2012).
AVPR1A and SLC6A4 polymorphisms in choral singers and non-musicians: A gene association
study. PLoS ONE 7(2), 2–8. Retrieved from https://doi.org/10.1371/journal.pone.0031763
Morley, B. J., & Happe, H. K. (2000). Cholinergic receptors: Dual roles in transduction and
plasticity. Hearing Research 147(1–2), 104–112.
Motts, S. D., & Schofield, B. R. (2010). Cholinergic and non-cholinergic projections from the
pedunculopontine and laterodorsal tegmental nuclei to the medial geniculate body in guinea pigs.
Frontiers in Neuroanatomy 4, 1–8. Retrieved from https://doi.org/10.3389/fnana.2010.00137
Mueller, K., Fritz, T., Mildner, T., Richter, M., Schulze, K., Lepsien, J., … Möller, H. E. (2015).
Investigating the dynamics of the brain response to music: A central role of the ventral
striatum/nucleus accumbens. NeuroImage 116, 68–79.
Murphy, D. D., Rueter, S. M., Trojanowski, J. Q., & Lee, V. M. (2000). Synucleins are
developmentally expressed, and alpha-synuclein regulates the size of the presynaptic vesicular
pool in primary hippocampal neurons. Journal of Neuroscience 20(9), 3214–3220.
Naganawa, M., Zheng, M.-Q., Henry, S., Nabulsi, N., Lin, S.-F., Ropchan, J., … Huang, Y. (2015).
Test-retest reproducibility of binding parameters in humans with 11C-LY2795050, an antagonist
PET radiotracer for the opioid receptor. Journal of Nuclear Medicine 56(2), 243–248.
Narendran, R., Mason, N. S., Laymon, C. M., Lopresti, B. J., Velasquez, N. D., May, M. A., …
Frankle, W. G. (2010). A comparative evaluation of the dopamine D(2/3) agonist radiotracer [11C]
(-)-N-propyl-norapomorphine and antagonist [11C]raclopride to measure amphetamine-induced
dopamine release in the human striatum. Journal of Pharmacology and Experimental Therapeutics
333(2), 533–539.
Narendran, R., Slifstein, M., Guillin, O., Hwang, Y., Hwang, D. R., Scher, E., … Laruelle, M. (2006).
Dopamine (D2/3) receptor agonist Positron Emission Tomography radiotracer [11C]-(+)-PHNO is
a D3 receptor preferring agonist in vivo. Synapse 60(7), 485–495.
Nater, U. M., Abbruzzese, E., Krebs, M., & Ehlert, U. (2006). Sex differences in emotional and
psychophysiological responses to musical stimuli. International Journal of Psychophysiology
62(2), 300–308.
Nilsson, U. (2009). Soothing music can increase oxytocin levels during bed rest after open-heart
surgery: A randomised control trial. Journal of Clinical Nursing 18(15), 2153–2161.
Nilsson, U., Unosson, M., & Rawal, N. (2005). Stress reduction and analgesia in patients exposed to
calming music postoperatively: A randomized controlled trial. European Journal of
Anaesthesiology 22(2), 96–102.
Numan, M., Bress, J. A., Ranker, L. R., Gary, A. J., DeNicola, A. L., Bettis, J. K., & Knapp, S. E.
(2010). The importance of the basolateral/basomedial amygdala for goal-directed maternal
responses in postpartum rats. Behavioural Brain Research 214(2), 368–376.
Oczkowska, A., Kozubski, W., Lianeri, M., & Dorszewska, J. (2014). Mutations in PRKN and SNCA
genes important for the progress of Parkinson’s disease. Current Genomics 14(8), 502–517.
Okada, K., Kurita, A., Takase, B., Otsuka, T., Kodani, E., Kusama, Y., … Mizuno, K. (2009). Effects
of music therapy on autonomic nervous system activity, incidence of heart failure events, and
plasma cytokine and catecholamine levels in elderly patients with cerebrovascular disease and
dementia. International Heart Journal 50(1), 95–110.
Ooishi, Y., Mukai, H., Watanabe, K., Kawato, S., & Kashino, M. (2017). Increase in salivary
oxytocin and decrease in salivary cortisol after listening to relaxing slow-tempo and exciting fast-
tempo music. PLoS ONE 12(12), 1–16. Retrieved from
https://doi.org/10.1371/journal.pone.0189075
Owen, D. R. J., Gunn, R. N., Rabiner, E. A., Bennacef, I., Fujita, M., Kreisl, W. C., … Parker, C. A.
(2011). Mixed-affinity binding in humans with 18-kDa translocator protein ligands. Journal of
Nuclear Medicine 52(1), 24–32.
Pan, W. X., & Hyland, B. I. (2005). Pedunculopontine tegmental nucleus controls conditioned
responses of midbrain dopamine neurons in behaving rats. Journal of Neuroscience 25(19), 4725–
4732.
Pierrehumbert, B., Torrisi, R., Laufer, D., Halfon, O., Ansermet, F., & Beck Popovic, M. (2010).
Oxytocin response to an experimental psychosocial challenge in adults exposed to traumatic
experiences during childhood or adolescence. Neuroscience 166(1), 168–177.
Qiu, J., Jiang, Y.-F., Li, F., Tong, Q.-H., Rong, H., & Cheng, R. (2017). Effect of combined music
and touch intervention on pain response and β-endorphin and cortisol concentrations in late
preterm infants. BMC Pediatrics 17(1), 1–7. Retrieved from https://doi.org/10.1186/s12887-016-
0755-y
Quiroga Murcia, C., Kreutz, G., Clift, S., & Bongard, S. (2010). Shall we dance? An exploration of
the perceived benefits of dancing on well-being. Arts & Health 2(2), 149–163.
Rabiner, E. A., & Laruelle, M. (2010). Imaging the D3 receptor in humans in vivo using [11C](+)-
PHNO positron emission tomography (PET). International Journal of Neuropsychopharmacology
13(3), 289–290.
Rainville, J. R., Tsyglakova, M., & Hodes, G. E. (2018). Deciphering sex differences in the immune
system and depression. Frontiers in Neuroendocrinology (August). Retrieved from
https://doi.org/10.1016/j.yfrne.2017.12.004
Reese, N. B., Garciarill, E., & Skinner, R. D. (1995a). Auditory input to the pedunculopontine
nucleus: I. Evoked potentials. Brain Research Bulletin 37(3), 257–264.
Reese, N. B., Garciarill, E., & Skinner, R. D. (1995b). Auditory input to the pedunculopontine
nucleus: II. Unite responses. Brain Research Bulletin 37(3), 265–273.
Salimpoor, V. N., Benovoy, M., Larcher, K., Dagher, A., & Zatorre, R. J. (2011). Anatomically
distinct dopamine release during anticipation and experience of peak emotion to music. Nature
Neuroscience 14(2), 257–262.
Salimpoor, V. N., van den Bosch, I., Kovacevic, N., McIntosh, A. R., & Dagher, A. Z. R. (2013).
Interactions between the nucleus accumbens and auditory cortices predict music reward value.
Science 340(6129), 216–219.
Schladt, T. M., Nordmann, G. C., Emilius, R., Kudielka, B. M., de Jong, T. R., & Neumann, I. D.
(2017). Choir versus solo singing: Effects on mood, and salivary oxytocin and cortisol
concentrations. Frontiers in Human Neuroscience 11, 1–9. Retrieved from
https://doi.org/10.3389/fnhum.2017.00430
Schneider, N., Schedlowski, M., Schürmeyer, T. H., & Becker, H. (2001). Stress reduction through
music in patients undergoing cerebral angiography. Neuroradiology 43(6), 472–476.
Schofield, B. R. (2010). Projections from auditory cortex to midbrain cholinergic neurons that project
to the inferior colliculus. Neuroscience 166(1), 231–240.
Schwilling, D., Vogeser, M., Kirchhoff, F., Schwaiblmair, F., Boulesteix, A. L., Schulze, A., &
Flemmer, A. W. (2015). Live music reduces stress levels in very low-birthweight infants. Acta
Paediatrica (Oslo, Norway), 104(4), 360–367.
Shimizu, N., Umemura, T., Hirai, T., Tamura, T., Sato, K., & Kusaka, Y. (2013). Effects of movement
music therapy with the naruko clapper on psychological, physical and physiological indices among
elderly females: A randomized controlled trial. Gerontology 59(4), 355–367.
Shotbolt, P., Tziortzi, A. C., Searle, G. E., Colasanti, A., Van Der Aart, J., Abanades, S., … Rabiner,
E. A. (2012). Within-subject comparison of [11C]-(+)-PHNO and [11C]raclopride sensitivity to
acute amphetamine challenge in healthy humans. Journal of Cerebral Blood Flow and Metabolism
32(1), 127–136.
Solís, O., & Moratalla, R. (2018). Dopamine receptors: Homomeric and heteromeric complexes in
l‑DOPA‑induced dyskinesia. Journal of Neural Transmission 1, 1–8. Retrieved from
https://doi.org/10.1007/s00702-018-1852-x
Spencer, R. L., Chun, L. E., Hartsock, M. J., & Woodruff, E. R. (2018). Glucocorticoid hormones are
both a major circadian signal and major stress signal: How this shared signal contributes to a
dynamic relationship between the circadian and stress systems. Frontiers in Neuroendocrinology
49, 52–71.
Stefano, G. B., Zhu, W., Cadet, P., Salamon, E., & Mantione, K. J. (2004). Music alters constitutively
expressed opiate and cytokine processes in listeners. Medical Science Monitor: International
Medical Journal of Experimental and Clinical Research 10(6), MS18–MS27.
Suzuki, M., Kanamori, M., Nagasawa, S., Tokiko, I., & Takayuki, S. (2007). Music therapy-induced
changes in behavioral evaluations, and saliva chromogranin A and immunoglobulin A
concentrations in elderly patients with senile dementia. Geriatrics & Gerontology International
7(1), 61–71.
Tabrizi, E. M., Sahraei, H., & Rad, S. M. (2012). The effect of music on the level of cortisol, blood
glucose and physiological variables. EXCLI Journal 11, 556–565. Retrieved from
https://doi.org/10.3389/fpsyg.2011.00058
Tan, Y. T., McPherson, G. E., Peretz, I., Berkovic, S. F., & Wilson, S. J. (2014). The genetic basis of
music ability. Frontiers in Psychology 5, 1–19. Retrieved from
https://doi.org/10.3389/fpsyg.2014.00658
Thoma, M. V., La Marca, R., Brönnimann, R., Finkel, L., Ehlert, U., & Nater, U. M. (2013). The
effect of music on the human stress response. PLoS ONE 8(8), 1–12. Retrieved from
https://doi.org/10.1371/journal.pone.0070156
Thompson, A. M. (2003). Pontine sources of norepinephrine in the cat cochlear nucleus. Journal of
Comparative Neurology 457(4), 374–383.
Thompson, R. R., & Walton, J. C. (2004). Peptide effects on social behavior: Effects of vasotocin and
isotocin on social approach behavior in male goldfish (Carassius auratus). Behavioral
Neuroscience 118(3), 620–626.
Trappe, H.-J., & Voit, G. (2016). The cardiovascular effect of musical genres. Deutzsches Ärzteblatt
International 113(20), 347–352.
Turvey, S. E., & Broide, D. H. (2010). Innate immunity. Journal of Allergy and Clinical Immunology
125(2 Suppl. 2), S24–S32.
Ukkola-Vuoti, L., Kanduri, C., Oikkonen, J., Buck, G., Blancher, C., Raijas, P., … Järvelä, I. (2013).
Genome-wide copy number variation analysis in extended families and unrelated individuals
characterized for musical aptitude and creativity in music. PLoS ONE 8(2). Retrieved from
https://doi.org/10.1371/journal.pone.0056356
Ukkola-Vuoti, L., Oikkonen, J., Onkamo, P., Karma, K., Raijas, P., & Järvelä, I. (2011). Association
of the arginine vasopressin receptor 1A (AVPR1A) haplotypes with listening to music. Journal of
Human Genetics 56(4), 324–329.
Ukkola, L., Onkamo, P., Raijas, P., Karma, K., & Järvelä, I. (2009). Musical aptitude is associated
with AVPR1A-Halotypes. PLoS ONE 4(5), e5534. Retrieved from
https://doi.org/10.1371/journal.pone.0005534
Valdiglesias, V., Maseda, A., Lorenzo-López, L., Pásaro, E., Millán-Calenti, J. C., & Laffon, B.
(2017). Is salivary chromogranin A a valid psychological stress biomarker during sensory
stimulation in people with advanced dementia? Journal of Alzheimer’s Disease 55(4), 1509–1517.
Valstad, M., Alvares, G. A., Egknud, M., Matziorinis, A. M., Andreassen, O. A., Westlye, L. T., &
Quintana, D. S. (2017). The correlation between central and peripheral oxytocin concentrations: A
systematic review and meta-analysis. Neuroscience & Biobehavioral Reviews 78, 117–124.
Veening, J. G., Gerrits, P. O., & Barendregt, H. P. (2012). Volume transmission of beta-endorphin via
the cerebrospinal fluid: A review. Fluids and Barriers of the CNS 9(1), 1. Retrieved from
https://doi.org/10.1186/2045-8118-9-16
Venneti, S., Lopresti, B. J., & Wiley, C. A. (2013). Molecular imaging of microglia/macrophages in
the brain. Glia 61(1), 10–23. Retrieved from https://doi.org/10.1002/glia.22357
Vollert, J., Störk, T., Rose, M., & Möckel, M. (2003). Musik als begleitende therapie bei koronarer
herzkrankheit. Deutsche Medizinische Wochenschrift, 128, 2712–2716.
Wahbeh, H., Calabrese, C., & Zwickey, H. (2007). Binaural beat technology in humans: A pilot study
to assess psychologic and physiologic effects. Journal of Alternative and Complementary
Medicine 13(1), 25–32.
Wang, S., Kulkarni, L., Dolev, J., & Kain, Z. (2002). Music and preoperative anxiety: A randomized,
controlled study. Anesthesia & Analgesia 94(6), 1489–1494.
Willeit, M., Ginovart, N., Graff, A., Rusjan, P., Vitcu, I., Houle, S., … Kapur, S. (2008). First human
evidence of d-amphetamine induced displacement of a D2/3agonist radioligand: A [11C]-(+)-
PHNO positron emission tomography study. Neuropsychopharmacology 33(2), 279–289.
Willeit, M., Ginovart, N., Kapur, S., Houle, S., Hussey, D., Seeman, P., & Wilson, A. A. (2006).
High-affinity states of human brain dopamine D2/3 receptors imaged by the agonist [11C]-(+)-
PHNO. Biological Psychiatry 59(5), 389–394.
Woof, J. M., & Ken, M. A. (2006). The function of immunoglobulin A in immunity. Journal of
Pathology 208(2), 270–282.
Yamamoto, T., Ohkuwa, T., Itoh, H., Kitoh, M., Terasawa, J., Tsuda, T., … Sato, Y. (2003). Effects of
pre-exercise listening to slow and fast rhythm music on supramaximal cycle performance and
selected metabolic variables. Archives of Physiology and Biochemistry 111(3), 211–214.
Yovel, G., Shakhar, K., & Ben-Eliyahu, S. (2001). The effects of sex, menstrual cycle, and oral
contraceptives on the number and activity of natural killer cells. Gynecologic Oncology 81(2),
254–262.
Yuhi, T., Kyuta, H., Mori, H.-A., Murakami, C., Furuhara, K., Okuno, M., … Higashida, H. (2017).
Salivary oxytocin concentration changes during a group drumming intervention for maltreated
school children. Brain Sciences 7, 152. Retrieved from https://doi.org/10.3390/brainsci7110152
CHAPT E R 15
THE NEUROAESTHETICS
OF MUSIC: A RESEARCH
AGENDA COMING OF AGE
E LV I R A B R AT T I C O
T N S N
D A
The term neuroaesthetics was first coined by Semir Zeki almost two
decades ago (Zeki, 1999) to indicate a multidisciplinary field of research,
focused at first on visual art, merging a long history of philosophical and
empirical aesthetics with the methodology of cognitive and affective
neurosciences (Chatterjee, 2011; Chatterjee & Vartanian, 2014, 2016;
Conway & Rehding, 2013; Nadal & Pearce, 2011; Pearce et al., 2016).
Neuroaesthetics seeks to understand the neural principles underlying the
different processes that compose a human aesthetic experience with an
artistic object (Livingstone & Hubel, 2002).
An aesthetic experience has been defined as a psychological state
determined by interaction with an object to which we intend to attribute
(evaluate/appraise) positive or negative qualities according to perceptual,
cognitive, affective, or cultural criteria. It is intrinsically different from
other affective experiences due to a special attitude (also referred to as
focus, stance, or pre-classification) toward the object. According to a
Kantian notion, this aesthetic stance is often characterized by being
disinterested, distanced from the primary emotional needs of the organism
(Leder, Gerger, Brieber, & Schwarz, 2014). According to a somewhat
tautological definition, an aesthetic experience is “an experience of a
particular kind of object that has been designed to produce such an
experience” (Bundgaard, 2015, p. 788). According to this
conceptualization, an aesthetic experience arises when, through perceptual-
representational processes, we attribute to the stimulus a meaning based on
aesthetic evaluation. While there exist some universal laws of preference
for some stimulus configurations (e.g., according to Gestalt laws humans
tend to like symmetry, equilibrium, and order due to organizational function
of the organism; Cupchik, 2007; Eysenck, 1942), the stimulus alone is not
by itself the source of an aesthetic experience. Rather, it is the intentional
relation and attitude that the subject has with the stimulus. Because of this,
subjectivity is intrinsic in aesthetic responses. A stimulus that is
aesthetically appealing to one person can be repulsive to another. These
variations derive from both the internal state, including the personal
experience of previous encounters with the stimulus, and the attitudes
toward the stimulus, the current mood, and the innate biological
predispositions for processing the stimulus and for having an aesthetic
experience as a whole (Pelowski, Markey, Forster, Gerger, & Leder, 2017).
Along this conceptualization, the research field of neuroaesthetics is
dedicated to studying how the brain facilitates the human capacity for
experiencing phenomena as “aesthetic” and for creating objects that evoke
such experience. To delve into these aims, one can choose two possible
directions of investigation, as also conceptualized by Brattico (2015),
Cupchik and colleagues (Cupchik, Vartanian, Crawley, & Mikulis, 2009),
Jacobsen and Beudt (2017), and Pelowski et al. (2017): on one hand, the
bottom-up perceptual facilitation of aesthetic responses based on the
physical properties of an artwork, and, on the other hand, the feedback and
feedforward relationship between top-down, intentional orientation of
attention and the artwork. Following Redies (Redies, 2015), this dualism in
how aesthetic phenomena are studied can be represented as a dichotomy
between formalist and contextual theories. Formalist theories propose that
the aesthetic experience relies on formal properties of the stimulus (e.g.,
symmetry, sensual beauty), which are considered to be universal and based
on human brain physiology. Often in these theories, aesthetic responses to
art are described as automatic and independent from conscious control
(Zeki, 2013). In turn, in contextual theories the aesthetic experience
depends on the intention of the artist and the circumstances under which the
artwork has been created and is displayed. Some of these theories focus on
contemporary abstract art, characterized by a lesser role given to sensory
features (Jacobsen, 2014; Leder, Belke, Oeberst, & Augustin, 2004;
Pelowski, Markey, Lauring, & Leder, 2016). Some proposals also attempt a
reconciliation between the two opposite stands, modeling the impact of top-
down and bottom-up factors depending on the type of artistic stimulus that
is at hand. For instance, in the model by Redies (Redies, 2015) external
information, meaning the stimulus features and context in which it is
displayed, is distinct from internal representation, meaning the subjective
representation and reaction to the stimulus by the beholder. In this particular
model, aesthetic experience is reached only with favorable encoding and
cognitive mastering of the stimulus. In most proposals, mainly focused on
visual art (Pearce et al., 2016; Pelowski et al., 2016), the aesthetic
experience seems to emerge from the interaction of cognitive, affective, and
evaluative processes, involving at least three different brain processes: (a)
an enhancement of low-level sensory processing; (b) high-level top-down
processing and activation of cortical areas involved in evaluative judgment;
(c) an engagement of the reward circuit, including cortical and subcortical
regions.
The initial efforts within neuroaesthetics of visual art involved
measurements of subjects’ brain activity while they evaluated the beauty or
preference of artistic versus natural pictures (e.g., Vartanian & Goel, 2004),
while they rated the beauty or correctness of abstract visual patterns (e.g.,
Jacobsen & Höfel, 2003), or while they viewed abstract, still life,
landscape, or portrait pictures classified as beautiful, ugly, or neutral prior
to the brain scanning session (e.g., Kawabata & Zeki, 2004). After these
inspiring works, a great number of publications using neuroimaging and
neurophysiological techniques have followed. Current neuroaesthetic
research has fractionated human responses to art into the main outcomes of
aesthetic emotions (e.g., pleasure, being moved, interest), preference (e.g.,
conscious liking), and judgment (e.g., beauty), associating to each of them a
replicable and reliable pattern of neural and physiological activity (Brattico
et al., 2016; Brattico, Bogert, & Jacobsen, 2013; Brattico & Pearce, 2013;
Chatterjee & Vartanian, 2014, 2016; Istok, Brattico, Jacobsen, Ritter, &
Tervaniemi, 2013; Jacobsen, 2014; Leder, Markey, & Pelowski, 2015;
Nieminen, Istok, Brattico, Tervaniemi, & Huotilainen, 2011; Pearce et al.,
2016; Pelowski et al., 2016; Reybrouck & Brattico, 2015). In these
proposals, aesthetic emotions are the subjective feelings elicited by an
artistic object whereas aesthetic judgments are defined as subjective
evaluations based on an individual set of criteria. Moreover, several factors
affecting the aesthetic experience have been targeted by neuroscientific
investigation: environment, intentions, familiarity, expertise, and attitudes.
In the latest overarching proposal called the Vienna Integrated Model of Art
Perception or VIMAP (Pelowski et al., 2017), bottom-up processing of low-
level artwork derived features, listing perceptual analysis, implicit memory
integration, and explicit classification, is conjoined with top-down factors.
Among those latter factors, cognitive mastery, namely the matching of all
information collected in previous processing stages to existing predictions
and schemata, plays a central role and leads to the creation of meaning and
associations. Brain substrates of the difference stages of the visual aesthetic
experiences have also been identified particularly in visual cortices for
feature analysis, dorsolateral prefrontal cortex for cognitive mastery
default-mode network regions, error-monitoring regions of the anterior
cingulate cortex, limbic regions (particularly, insula and amygdala) for
controlling emotions, and orbitofrontal cortex for integrating signals from
cognitive and emotional brain regions and issuing aesthetic judgments.
Lately, while the initial and majority of efforts have concentrated on
visual art (paintings), researchers keen on the neuroaesthetics approach
have expanded their interest from visual art toward several other artistic
domains, such as sculpture (Di Dio, Macaluso, & Rizzolatti, 2007),
architecture (Coburn, Vartanian, & Chatterjee, 2017), dance (Calvo-Merino,
Glaser, Grezes, Passingham, & Haggard, 2005; Calvo-Merino, Jola, Glaser,
& Haggard, 2008), and poetry (Wassiliwizky, Koelsch, Wagner, Jacobsen,
& Menninghaus, 2017). In the past few years, the field has seen a fast
growth with several special issues of journals and books (e.g., Huston,
Nadal, Mora, Agnati, & Cela Conde, 2015; Martindale, Locher, & Petrov,
2007), reviews (Chatterjee, 2011; Chatterjee and Vartanian, 2014, 2016;
Leder & Nadal, 2014; Nadal et al., 2008; Pearce et al., 2016; Pelowski et
al., 2016, 2017), and conferences (e.g., Nadal & Pearce, 2011). While
critiques do exist (Tallis, 2008, 2011), and are indeed welcome for a healthy
scientific debate, in the past two years the status of neuroaesthetics,
especially for visual arts, has changed from that of contingent or trendy to
that of a mature discipline (Chatterjee, 2011; Leder & Nadal, 2014; Pearce
et al., 2016).
N : A R A
M
E M M
A E
Even if the past few years have witnessed several studies on aesthetic-
related phenomena during music listening, the scientific questions asked
have often been addressed without any explicit reference to overarching
aesthetic frameworks, differently from what happens in visual research
(Brattico & Pearce, 2013; Hodges, 2016). In a critical integrative analysis
of thirty-one empirical aesthetic studies conducted between the years 1990
and 2015 (out of the initial 1,450 references first obtained) (Tiihonen,
Brattico, Maksimainen, Wikgren, & Saarikallio, 2017), it was noted that
scientific investigations of pleasure, one of the main subjective aesthetic
responses to any artwork, have been contextualized within aesthetic
frameworks and concepts for the visual modality (studies using stimuli
from figurative arts, such as painting or sculptures) whereas they were
linked to basic neuroscientific literature on primary pleasure (or the absence
of it) for music modality. This analysis confirms that visual empirical and
neuroaesthetics are active fields counting a number of established and well-
recognized frameworks, whereas research on music is dominated by
sensory and basic emotion models.
The current situation can be attributed to the scarcity of brain-based
models of aesthetic processes in music, leading to limited efforts of
overarching interpretations of the individual neuroscientific findings
obtained. One of these models (illustrated in Fig. 2) is characterized by a
chronometric distinction of the information processing stages leading to
aesthetic responses. This and further developments by the same authors
establish a distinction between pre-attentive, low-level perceptual and
emotional stages, and reflective processes involving cognitive control
(Brattico, 2015; Brattico et al., 2013; Brattico & Pearce, 2013; Nieminen et
al., 2011; Reybrouck & Brattico, 2015). These stages lead to the three main
outcomes of an aesthetic experience, namely emotion, preference, and
judgment (Brattico, 2015; Brattico & Pearce, 2013). These previous
accounts include a locationist view combined with a temporal information
processing description of the brain mechanisms involved in the aesthetic
experience of music: each temporally evolving stage depends on a distinct
set of specific brain structures. The final outcomes of the aesthetic
experience require the succession of all previous stages in order to
materialize. For instance, according to Brattico et al. (2013), conscious
liking judgments can be issued after the brainstem, thalamus, and limbic
regions have quickly reacted to salient features of the sound, and after the
frontotemporal cortex has encoded and integrated those sound features with
learned cognitive schemata, using parietal and action observation neural
resources for attributing emotional connotations to the sounds (see Fig. 2).
If all these stages are successfully completed, and if limbic, prefrontal, and
mentalizing brain regions are conspicuously activated, then a liking
judgment, possibly accompanied also by a beauty verdict, is issued.
FIGURE 2. A schematic representation of a previous framework concerning the timing,
localization, and effects of neural processes contributing to aesthetic experience (modified from
Brattico et al., 2013). The lower block shows how the various processes evolve as a function of time,
beginning from the first sensory analyses to the main outcomes of aesthetic emotions, preference,
and judgments. The upper block illustrates their rough anatomic locations and connections in the
human brain. ABR = auditory brainstem response; LPP = late positive potential.
M B S R
A R M
Recent years have seen a change in the way brain physiology is described,
from a locationist view where each structure subserves one or a few main
functions, to a distributed view where the brain is described as a complex
dynamic system and where the interactions between its components govern
cognitive functions (Bassett & Sporns, 2017; Medaglia, Lynall, & Bassett,
2015). This novel view derives from the technological and scientific
progress of network neuroscience, namely the marriage between network
science and cognitive neuroscience (Bassett & Sporns, 2017). Network
techniques are mathematical tools to describe complex systems organized in
networks that change over time (dynamics) (Medaglia et al., 2015;
Newman, 2010).
In previous overviews of the music neuroaesthetic field (Brattico &
Pearce, 2013; Hodges, 2016), studies from network neuroscience have not
been much mentioned. Indeed, most studies on functional connectivity have
been published in the past two years. For instance, it has been recently
found that functional connectivity between the superior temporal gyrus
(where the auditory cortex is located), the inferofrontal cortex (where
hierarchical predictions for sounds are computed), and reward regions
determines the pleasurable rewarding responses to music, or the absence of
them (Martínez-Molina, Mas-Herrero, Rodriguez-Fornells, Zatorre, &
Marco-Pallares, 2016; Sachs, Ellis, Schlaug, & Loui, 2016; Salimpoor et
al., 2013; Wilkins, Hodges, Laurienti, Steen, & Burdette, 2014). For
instance, in a study where subjects had to decide how much money they
would use to buy songs, it was found that the connections between the
nucleus accumbens and its surrounding regions (the amygdala and the
hippocampus) predicted how much a participant would spend on each song
(Salimpoor et al., 2013).
The importance of the neural interactions between the nucleus
accumbens and the auditory cortex for determining aesthetic pleasure to
music has been remarked also by studies aiming at identifying the neural
sources of individual differences in pleasurable reactions to musical sounds
(Keller et al., 2013; Martínez-Molina et al., 2016; Sachs et al., 2016). These
studies originate from the recently empirically proven observation that
music is not universally liked and appreciated but rather individuals vary
greatly in their sensitivity to musical reward, ranging from musicophilics
characterized by acute craving for music and increased responsiveness and
interest for musical sounds (Sacks, 2007) to musical anhedonics, with a
total indifference to music (Mas-Herrero, Marco-Pallares, Lorenzo-Seva,
Zatorre, & Rodriguez-Fornells, 2013; Mas-Herrero, Zatorre, Rodriguez-
Fornells, & Marco-Pallares, 2014). A recent study using diffusion tensor
imaging (DTI) has evidenced that the white-matter tracts between the
posterior portion of the superior temporal lobe and emotion- and reward-
processing regions such as the anterior insula and the medial prefrontal
cortex explain the individual differences in reward sensitivity to music
(Sachs et al., 2016). In that study, reward sensitivity was quantified with the
amount of chills experienced by each individual combined with the degree
of physiological changes (heart rate and skin conductance response) during
listening to music inducing chills versus neutral music. Another study
(Martínez-Molina et al., 2016) used the newly developed BMRQ
questionnaire to identify music-specific anhedonic, hedonic, and
hyperhedonic subjects. They were measured with fMRI during a music
listening task where they rated the pleasantness of the music excerpts, and a
gambling task, where they either won or lost a symbolic amount of money.
Results evidenced decreased regional activity in the ventral striatum
(including the nucleus accumbens) in anhedonics and increased regional
activity in hyperhedonics as well as downregulated functional connectivity
between this area and the right superior temporal gyrus in anhedonics.
These results were obtained only in relation with pleasantness responses to
the music and not with the gambling task.
These findings are not confined to receptive pleasure during listening
but also relate to the desire to move to rhythmic aspects of the music. A
study by Witek et al. (forthcoming) found local changes in directed
effective connectivity between motor (dorsomedial prefrontal) and reward
(striatal) networks during maximal rhythm-induced pleasurable urge to
move. In addition, they showed that maximal pleasurable desire to move to
sound was predicted by a meta-stable brain network organization, namely a
neural organization lying between an ordered and a disordered state
(computed as whole-brain shuffling speed of effective connectivity
matrices) (Deco, Kringelbach, Jirsa, & Ritter, 2017).
These and other studies compellingly demonstrate that functional
connectivity between the superior temporal gyrus (where the auditory
cortex is located), the inferofrontal cortex (where hierarchical predictions
for sounds are computed), and reward regions of the brain are linked with
pleasurable rewarding responses to music, or the absence of them
(Martínez-Molina et al., 2016; Sachs et al., 2016; Salimpoor et al., 2013;
Wilkins et al., 2014). Notably, the neural transmission between these brain
areas is regulated by the monoamine neurotransmitter dopamine that has
been linked to incentive salience and motivation for acting, namely to the
“wanting” phase of the reward cycle (Kringelbach & Berridge, 2017). A
very recent investigation has discovered a molecular link between affective
sensitivity to (musical) sounds and dopamine functionality (Quarto et al.,
2017): a functional variation in a dopamine receptor gene modulates the
impact of sounds on mood states and emotion-related prefrontal and striatal
brain activity.
The studies reviewed above, while having the important merit to reveal
the complex architecture subserving the rewarding experience of music
listening, have not examined whether this experience can be consumed
spontaneously, even with casual listening, or whether it requires focused
attention and a particular attitude (that is sometimes referred to as aesthetic
stance). A fresh study (Liu et al., in press) contrasted conditions varying in
the type of focused attentional involvement toward the music requested
from subjects. Similarly to previous findings (Bogert et al., 2016; Brattico
et al., 2016; Liu, Abu-Jamous, Brattico, & Nandi, 2017), the study observed
a co-activation in a network of mesiotemporal limbic structures, including
the nucleus accumbens, in response to the liked musical stimuli,
irrespectively of whether subjects were focusing on making a conscious
liking evaluation or not. Functional connectivity within prefrontal and
parieto-occipital regions was instead obtained for the liking judgments.
F C P
Until now, the musical experience has been analyzed from the point of view
of the subject. Yet, music (like other arts) can represent a means of
communication between the judgmental intentions of the perceiver and the
meaning-making intentions of the composer/artist. The act of meaning
attribution, which is essential to an aesthetic experience, as argued, for
example, by Chatterjee and Vartanian (2014), Pearce et al. (2016), Leder et
al. (2004), and Menninghaus et al. (2017), cannot exist without the
assignment of an intention to the agent producing the artistic object
(Acquadro, Congedo, & De Riddeer, 2016). Modern neuroscience offers
unprecedented opportunities to capture the essence of such aesthetic
processes, thanks to the hyperscanning approach, namely the synchronized
brain recordings of two or more persons doing an experimental task
together (Hari, Henriksson, Malinen, & Parkkonen, 2015; Konvalinka &
Roepstorff, 2012; Zhdanov et al., 2015; Zhou, Bourguignon, Parkkonen, &
Hari, 2016).
Even if presently, “core” neuroaesthetics of music does not much
account for motor production, the mirror neuron or action observation
system (a set of neurons in the fronto-parietal regions of the brain that
responds when watching others doing a motor action; Freedberg & Gallese,
2007; Gallese & Freedberg, 2007; Rizzolatti, Fadiga, Gallese, & Fogassi,
1996) has been proposed as a key mechanism allowing aesthetic responses
to music in an interactive situation (Molnar-Szakacs & Overy, 2006).
According to one model (Molnar-Szakacs & Overy, 2006), music is
described as a sequence of hierarchically organized sequences of motor acts
synchronous with auditory information and activating both the auditory
cortex and motor regions of the action observation network in the posterior
inferior frontal gyrus (BA 44) and adjacent premotor cortex. In this model,
the anterior insula serves to evaluate the internal visceral changes derived
from music and relay these changes to activity in the limbic system, which
ultimately is responsible for the complex affective experiences originating
from music listening. The co-activation of the same motor systems in
musician and perceiver is supposed to allow the co-representation and
sharing of emotions during an aesthetic musical experience. Hence, future
studies using hyperscanning techniques might measure the aesthetic value
of a musical interaction and determine the responsible neural mechanisms.
Initial investigations measuring the inter-subject coupling of
electroencephalographic signals (especially in beta frequency range) from
guitarists playing in a duet prove the feasibility of this approach
(Lindenberger, Li, Gruber, & Müller, 2009; Müller, Sanger, &
Lindenberger, 2013).
To conclude, the agenda of the neuroaesthetics of music, by addressing
questions related to intra- and inter-subjectivity during a musical activity,
comes close to the essence of music and of what we are as humans. While
there still is the risk of “biologism,” researchers working under the music
neuroaesthetics umbrella reach out to the “humanistic” approach since they
strive to explain how “musical appreciation is dependent on culture,
memory, mood and many other factors such as personal taste” (Tallis, 2011,
p. 54).
R
Acquadro, M. A., Congedo, M., & De Riddeer, D. (2016). Music performance as an experimental
approach to hyperscanning studies. Frontiers in Human Neuroscience 10, 242. Retrieved from
https://doi.org/10.3389/fnhum.2016.00242
Alluri, V., & Toiviainen, P. (2015). Musical expertise modulates functional connectivity of limbic
regions during continuous music listening. Psychomusicology: Music, Mind, and Brain 25(4),
443–454.
Altenmüller, E., Demorest, S. M., Fujioka, T., Halpern, A. R., Hannon, E. E., Loui, P., … Zatorre, R.
J. (2012). Introduction to the neurosciences and music IV: Learning and memory. Annals of the
New York Academy of Sciences 1252, 1–16.
Aubert, M., Brumm, A., Ramli, M., Sutikna, T., Saptomo, E. W., Hakim, B., … Dosseto, A. (2014).
Pleistocene cave art from Sulawesi, Indonesia. Nature 514(7521), 223–227.
Bassett, D. S., & Sporns, O. (2017). Network neuroscience. Nature Neuroscience 20(3), 353–364.
Berns, G. S., Capra, C. M., Moore, S., & Noussair, C. (2010). Neural mechanisms of the influence of
popularity on adolescent ratings of music. NeuroImage 49(3), 2687–2696.
Berridge, K. C., & Kringelbach, M. L. (2015). Pleasure systems in the brain. Neuron 86(3), 646–664.
Bigand, E., & Tillmann, B. (2015). Introduction to the neurosciences and music V: Cognitive
stimulation and rehabilitation. Annals of the New York Academy of Sciences 1337, vii–ix.
Blood, A. J., & Zatorre, R. J. (2001). Intensely pleasurable responses to music correlate with activity
in brain regions implicated in reward and emotion. Proceedings of the National Academy of
Sciences 98(20) 11818–11823.
Blood, A. J., Zatorre, R. J., Bermudez, P., & Evans, A. C. (1999). Emotional responses to pleasant
and unpleasant music correlate with activity in paralimbic brain regions. Nature Neuroscience
2(4), 382–387.
Blum, K., Chen, T. J., Chen, A. L., Madigan, M., Downs, B. W., Waite, R. L., … Gold, M. S. (2010).
Do dopaminergic gene polymorphisms affect mesolimbic reward activation of music listening
response? Therapeutic impact on Reward Deficiency Syndrome (RDS). Medical Hypotheses 74(3),
513–520.
Bogert, B., Numminen-Kontti, T., Gold, B., Sams, M., Numminen, J., Burunat, I., … Brattico, E.
(2016). Hidden sources of joy, fear, and sadness: Explicit versus implicit neural processing of
musical emotions. Neuropsychologia 89, 393–402.
Brattico, E. (2015). From pleasure to liking and back: Bottom-up and top-down neural routes to the
aesthetic enjoyment of music. In M. Nadal, J. P. Houston, L. Agnati, F. Mora, & C. J. Cela Conde
(Eds.), Art, aesthetics, and the brain (pp. 303–318). Oxford: Oxford University Press.
Brattico, E., Bogert, B., Alluri, V., Tervaniemi, M., Eerola, T., & Jacobsen, T. (2016). It’s sad but I
like it: The neural dissociation between musical emotions and liking in experts and laypersons.
Frontiers in Human Neuroscience 9, 676. Retrieved from
https://doi.org/10.3389/fnhum.2015.00676
Brattico, E., Bogert, B., & Jacobsen, T. (2013). Toward a neural chronometry for the aesthetic
experience of music. Frontiers in Psychology 4, 206. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00206
Brattico, E., Brattico, P., & Jacobsen, T. (2009). The origins of the aesthetic enjoyment of music: A
review of the literature. Musicae Scientiae 13(2), 15–39.
Brattico, E., Brusa, A., Fernandes, H. M., Jacobsen, T., Gaggero, G., Toiviainen, P., Vuust, P., &
Proverbio, A. M. (submitted). The beauty and the brain: Investigating the neural correlates of
musical beauty during a realistic listening experience.
Brattico, E., Jacobsen, T., De Baene, W., Glerean, E., & Tervaniemi, M. (2010). Cognitive vs.
affective listening modes and judgments of music: An ERP study. Biological Psychology 85(3),
393–409.
Brattico, E., & Pearce, M. T. (2013). The neuroaesthetics of music. Psychology of Aesthetics,
Creativity, and the Arts 7, 48–61.
Brattico, P., Brattico, E., & Vuust, P. (2017). Global sensory qualities and aesthetic experience of
music. Frontiers in Neuroscience 11. Retrieved from https://doi.org/10.3389/fnins.2017.00159
Brielmann, A. A., & Pelli, D. G. (2017). Beauty requires thought. Current Biology 27(10), 1506–
1513 e3.
Brown, S., Gao, X., Tisdelle, L., Eickhoff, S. B., & Liotti, M. (2011). Naturalizing aesthetics: Brain
areas for aesthetic appraisal across sensory modalities. NeuroImage 58(1), 250–258.
Bundgaard, H. (2015). Feeling, meaning, and intentionality: A critique of the neuroaesthetics of
beauty. Phenomenology and the Cognitive Sciences 14(4), 781–801.
Calvo-Merino, B., Glaser, D. E., Grezes, J., Passingham, R. E., & Haggard, P. (2005). Action
observation and acquired motor skills: An FMRI study with expert dancers. Cerebral Cortex 15(8),
1243–1249.
Calvo-Merino, B., Jola, C., Glaser, D. E., & Haggard, P. (2008). Towards a sensorimotor aesthetics of
performing art. Consciousness and Cognition 17(3), 911–922.
Chanda, M. L., & Levitin, D. J. (2013). The neurochemistry of music. Trends in Cognitive Sciences
17(4), 179–193.
Chapin, H., Jantzen, K., Kelso, J. A., Steinberg, F., & Large, E. (2010). Dynamic emotional and
neural responses to music depend on performance expression and listener experience. PloS ONE
5(12), e13812.
Chatterjee, A. (2011). Neuroaesthetics: A coming of age story. Journal of Cognitive Neuroscience
23(1), 53–62.
Chatterjee, A., & Vartanian, O. (2014). Neuroaesthetics. Trends in Cognitive Sciences 18(7), 370–
375.
Chatterjee, A., & Vartanian, O. (2016). Neuroscience of aesthetics. Annals of the New York Academy
of Sciences 1369, 172–194.
Coburn, A., Vartanian, O., & Chatterjee, A. (2017). Buildings, beauty, and the brain: A neuroscience
of architectural experience. Journal of Cognitive Neuroscience 29(9), 1521–1531.
Conway, B. R., & Rehding, A. (2013). Neuroaesthetics and the trouble with beauty. PLoS Biology 11,
e1001504.
Cupchik, G. C. (2007). A critical reflection on Arnheim’s Gestalt theory of aesthetics. Psychology of
Aesthetics, Creativity, and the Arts 1(1), 16–24.
Cupchik, G. C., Vartanian, O., Crawley, A., & Mikulis, D. J. (2009). Viewing artworks: Contributions
of cognitive control and perceptual facilitation to aesthetic experience. Brain and Cognition 70(1),
84–91.
Curtis, G. (2006). The cave painters. New York: Anchor Books.
Deco, G., Kringelbach, M. L., Jirsa, V. K., & Ritter, P. (2017). The dynamics of resting fluctuations
in the brain: Metastability and its dynamical cortical core. Scientific Reports 7, 3095.
doi:10.1038/s41598-017-03073-5
Di Dio, C., Macaluso, E., & Rizzolatti, G. (2007). The golden beauty: Brain response to classical and
renaissance sculptures. PLoS ONE 11, 1–9.
Doelling, K. B., & Poeppel, D. (2015). Cortical entrainment to music and its modulation by
expertise. Proceedings of the National Academy of Sciences 112(45), E6233–E6242.
Eysenck, H. J. (1942). The experimental study of the “good Gestalt”: A new approach. Psychological
Review 49(4), 344–364.
Fishman, Y. I., Volkov, I. O., Noh, M. D., Garell, P. C., Bakken, H., Arezzo, J. C., … Steinschneider,
M. (2001). Consonance and dissonance of musical chords: Neural correlates in auditory cortex of
monkeys and humans. Journal of Neurophysiology 86(6), 2761–2788.
Freedberg, D., & Gallese, V. (2007). Motion, emotion and empathy in esthetic experience. Trends in
Cognitive Sciences 11(5), 197–203.
Friston, K. (2005). A theory of cortical responses. Philosophical Transactions of the Royal Society B:
Biological Sciences 360(1456), 815–836.
Früholz, S., Trost, W., & Kotz, S. A. (2016). The sound of emotions: Towards a unifying neural
network perspective of affective sound processing. Neuroscience & Biobehioral Reviews 68, 96–
110.
Gallese, V., & Freedberg, D. (2007). Mirror and canonical neurons are crucial elements in esthetic
response. Trends in Cognitive Sciences 11(10), 411.
Golden, H. L., Clark, C. N., Nicholas, J. M., Cohen, M. H., Slattery, C. F., Paterson, R. W., …
Warren, J. D. (2017). Music perception in dementia. Journal of Alzheimer’s Disease 55(3), 933–
949.
Gosselin, N., Samson, S., Adolphs, R., Noulhiane, M., Roy, M., Hasboun, D., … Peretz, I. (2006).
Emotional responses to unpleasant music correlates with damage to the parahippocampal cortex.
Brain 129(10), 2585–2592.
Hanslick, E. (1954). On the musically beautiful. Indianapolis: Hackett (English translation from the
8th ed. 1891).
Hargreaves, D. J., & North, A. C. (2010). Experimental aesthetics and liking for music. In P. N.
Juslin & J. A. Sloboda (Eds.), Handbook of music and emotion: Theory, research, applications
(pp. 515–46). Oxford: Oxford University Press.
Hari, R., Henriksson, L., Malinen, S., & Parkkonen, L. (2015). Centrality of social interaction in
human brain function. Neuron 88(1), 181–193.
Hodges, D. A. (2016). The neuroaesthetics of music. In S. Hallam, I. Cross, & M. Thaut (Eds.), The
Oxford handbook of music psychology (2nd ed., pp. 247–262). Oxford: Oxford University Press.
Huron, D. (2006). Sweet anticipation: Music and the psychology of expectation. Cambridge, MA:
MIT Press.
Huron, D. (2009). Aesthetics. In S. Hallam, I. Cross, & M. Thaut (Eds.), The Oxford handbook of
music psychology (pp. 151–159). Oxford: Oxford University Press.
Huston, J. P., Nadal, M., Agnati, L., Mora, L., & Cela Conde, C. J. (Eds.). (2015). Art, aesthetics and
the brain. Oxford: Oxford University Press.
Ishizu, T., & Zeki, S. (2011). Toward a brain-based theory of beauty. PLoS ONE 6, e21852.
Istok, E., Brattico, E., Jacobsen, T., Krohn, K., Mueller, M., & Tervaniemi, M. (2009). Aesthetic
responses to music: A questionnaire study. Musicae Scientiae 13, 183–206.
Istok, E., Brattico, E., Jacobsen, T., Ritter, A., & Tervaniemi, M. (2013). “I love rock ’n’ roll”: Music
genre preference modulates brain responses to music. Biological Psychology 92(2), 142–151.
Jacobsen, J. H., Stelzer, J., Fritz, T. H., Chetelat, G., La Joie, R., & Turner, R. (2015). Why musical
memory can be preserved in advanced Alzheimer’s disease. Brain 138(8), 2438–2450.
Jacobsen, T. (2014). Domain specificity and mental chronometry in empirical aesthetics. British
Journal of Psychology 105(4), 471–473.
Jacobsen, T., & Beudt, S. (2017). Domain generality and domain specificity in aesthetic appreciation.
New Ideas in Psychology 47, 97–102.
Jacobsen, T., & Höfel, L. (2003). Descriptive and evaluative judgment processes: Behavioral and
electrophysiological indices of processing symmetry and aesthetics. Cognitive, Affective, &
Behavioral Neuroscience 3(4), 289–299.
Juslin, P. N. (2013). From everyday emotions to aesthetic emotions: Towards a unified theory of
musical emotions. Physics of Life Reviews 10(3), 235–266.
Juslin, P. N., & Laukka, P. (2004). Expression, perception, and induction of musical emotions: A
review and a questionnaire study of everyday listening. Journal of New Music Research 33(3),
217–238.
Juslin, P. N., & Västfjäll, D. (2008). Emotional responses to music: The need to consider underlying
mechanisms. Behavioral and Brain Sciences 31(5), 559–575.
Kawabata, H., & Zeki, S. (2004). Neural correlates of beauty. Journal of Neurophysiology 91(4),
1699–1705.
Keller, J., Young, C. B., Kelley, E., Prater, K., Levitin, D. J., & Menon, V. (2013). Trait anhedonia is
associated with reduced reactivity and connectivity of mesolimbic and paralimbic reward
pathways. Journal of Psychiatric Research 47(10), 1319–1328.
Koelsch, S. (2011). Toward a neural basis of music perception: A review and updated model.
Frontiers in Psychology 2, 110. Retrieved from https://doi.org/10.3389/fpsyg.2011.00110
Koelsch, S. (2014). Brain correlates of music-evoked emotions. Nature Reviews Neuroscience 15,
170–180.
Koelsch, S., & Siebel, W. A. (2005). Towards a neural basis of music perception. Trends in Cognitive
Sciences 9(12), 578–584.
Konvalinka, I., & Roepstorff, A. (2012). The two-brain approach: How can mutually interacting
brains teach us something about social interaction? Frontiers in Human Neuroscience 6, 215.
Retrieved from https://doi.org/10.3389/fnhum.2012.00215
Kornysheva, K., von Cramon, D. Y., Jacobsen, T., & Schubotz, R. I., Tuning-in the beat: Aesthetic
appreciation of musical rhythms correlates with a premotor activity boost. Human Brain Mapping
31(1), 48–64.
Kringelbach, M. L., & Berridge, K. C. (2017). The affective core of emotion: Linking pleasure,
subjective well-being, and optimal metastability in the brain. Emotion Review 9(3), 191–199.
Kühn, S., & Gallinat, J. (2012). The neural correlates of subjective pleasantness. NeuroImage 61(1),
289–294.
Large, E. W., & Snyder, J. S. (2009). Pulse and meter as neural resonance. Annals of the New York
Academy of Sciences 1169, 46–57.
Laukka, P. (2007). Uses of music and psychological well-being among the elderly. Journal of
Happiness Studies 8(2), 215–241.
Leder, H., Belke, B., Oeberst, A., & Augustin, D. (2004). A model of aesthetic appreciation and
aesthetic judgements. British Journal of Psychology 95(4), 489–508.
Leder, H., Gerger, G., Brieber, D., & Schwarz, N. (2014). What makes an art expert? Emotion and
evaluation in art appreciation. Cognition and Emotion 28, 1137–1147.
Leder, H., Markey, P. S., & Pelowski, M. (2015). Aesthetic emotions to art: What they are and what
makes them special. Comment on “The quartet theory of human emotions: An integrative and
neurofunctional model” by S. Koelsch et al. Physics of Life Reviews 13, 67–70.
Leder, H., & Nadal, M. (2014). Ten years of a model of aesthetic appreciation and aesthetic
judgments: The aesthetic episode—developments and challenges in empirical aesthetics. British
Journal of Psychology 105(4), 443–464.
Lehne, M., & Koelsch, S. (2015). Tension-resolution patterns as a key element of aesthetic
experience: Psychological principles and underlying brain mechanisms. In J. P. Huston, M. Nadal,
F. Mora, L. Agnati, & C. J. Cela Conde (Eds.), Art, aesthetics, and the brain (pp. 285–302).
Oxford: Oxford University Press.
Levitin, D. J., & Tirovolas, A. K. (2009). Current advances in the cognitive neuroscience of music.
Annals of the New York Academy of Sciences 1156, 211–231.
Lindenberger, U., Li, S. C., Gruber, W., & Müller, V. (2009). Brains swinging in concert: Cortical
phase synchronization while playing guitar. BMC Neuroscience 10, 22. Retrieved from
https://doi.org/10.1186/1471-2202-10-22
Liu, C., Abu-Jamous, B., Brattico, E., & Nandi, A. K. (2017). Towards tunable consensus clustering
for studying functional brain connectivity during affective processing. International Journal of
Neural Systems 27(2), doi:10.1142/S0129065716500428
Liu, C., Brattico, E., Abu-Jamous, B., Pereira, C. S., Jacobsen, T., & Nandi, A. K. (in press). Effect
of explicit evaluation on the neural connectivity related to listening to unfamiliar music. Frontiers
in Human Neuroscience. Retrieved from https://doi.org/10.3389/fnhum.2017.00611
Livingstone, M., & Hubel, D. H. (2002). Vision and art: The biology of seeing. New York: Harry N.
Abrams.
McDonald, C., & Stewart, L. (2008). Uses and functions of music in congenital amusia. Music
Perception 25(4), 345–355.
Martindale, C., Locher, P., & Petrov, V. M. (2007). Evolutionary and neurocognitive approaches to
aesthetics, creativity and the arts. Amityville, NY: Baywood Publishing.
Martínez-Molina, N., Mas-Herrero, E., Rodriguez-Fornells, A., Zatorre, R. J., & Marco-Pallares, J.
(2016). Neural correlates of specific musical anhedonia. Proceedings of the National Academy of
Sciences 113, E7337–E7345.
Mas-Herrero, E., Dagher, A., & Zatorre, R. J. (2018). Modulating musical reward sensitivity up and
down with transcranial magnetic stimulation. Nature Human Behaviour 2, 27–32.
Mas-Herrero, E., Marco-Pallares, J., Lorenzo-Seva, U., Zatorre, R. J., & Rodriguez-Fornells, A.
(2013). Individual differences in music reward experiences. Music Perception 31(2), 118–138.
Mas-Herrero, E., Zatorre, R. J., Rodriguez-Fornells, A., & Marco-Pallares, J. (2014). Dissociation
between musical and monetary reward responses in specific musical anhedonia. Current Biology
24(6), 699–704.
Matrone, C., & Brattico, E. (2015). The power of music on Alzheimer’s disease and the need to
understand the underlying molecular mechanisms. Journal of Alzheimer’s Disease and
Parkinsonism 5. doi:10.4172/2161-0460.1000196
Medaglia, J. D., Lynall, M. E., & Bassett, D. S. (2015). Cognitive network neuroscience. Journal of
Cognitive Neuroscience 27(8), 1471–1491.
Menninghaus, W., Wagner, V., Hanich, J., Wassiliwizky, E., Jacobsen, T., & Koelsch, S. (2017). The
distancing-embracing model of the enjoyment of negative emotions in art reception. Behavioral
and Brain Sciences 40, 1–58.
Menon, V., & Levitin, D. J. (2005). The rewards of music listening: Response and physiological
connectivity of the mesolimbic system. NeuroImage 28(1), 175–184.
Molnar-Szakacs, I., & Overy, K. (2006). Music and mirror neurons: From motion to “e”motion.
Social Cognitive and Affective Neuroscience 1(3), 235–241.
Montag, C., Reuter, M., & Axmacher, N. (2011). How one’s favorite song activates the reward
circuitry of the brain: Personality matters! Behavioural Brain Research 225(2), 511–514.
Müller, V., Höfel, L., Brattico, E., & Jacobsen, T. (2010). Aesthetic judgments of music in experts
and laypersons: An ERP study. International Journal of Psychophysiology 76(1), 40–51.
Müller, V., Sanger, J., & Lindenberger, U. (2013). Intra- and inter-brain synchronization during
musical improvisation on the guitar. PLoS ONE 8, e73852.
Nadal, M., Munar, E., Capo, M. A., Rossello, J., & Cela-Conde, C. J. (2008). Towards a framework
for the study of the neural correlates of aesthetic preference. Spatial Vision 21(3–5), 379–396.
Nadal, M., & Pearce, M. T. (2011). The Copenhagen neuroaesthetics conference: Prospects and
pitfalls for an emerging field. Brain and Cognition 76(1), 172–183.
Newman, M. E. J. (2010). Networks: An introduction. Oxford: Oxford University Press.
Nieminen, S., Istok, E., Brattico, E., Tervaniemi, M., & Huotilainen, M. (2011). The development of
aesthetic responses to music and their underlying neural and psychological mechanisms. Cortex
47(9), 1138–1146.
Pallesen, K. J., Bailey, C. J., Brattico, E., Gjedde, A., Palva, J. M., & Palva, S. (2015). Experience
drives synchronization: The phase and amplitude dynamics of neural oscillations to musical chords
are differentially modulated by musical expertise. PLoS ONE 10, e0134211.
Pallesen, K. J., Brattico, E., Bailey, C., Korvenoja, A., Koivisto, J., Gjedde, A., & Carlson, S. (2005).
Emotion processing of major, minor, and dissonant chords: a functional magnetic resonance
imaging study. Annals of the New York Academy of Sciences 1060, 450–453.
Patel, A. (2008). Music, language, and the brain. Oxford: Oxford University Press.
Pearce, M. T. (2015). Effects of expertise on the cognitive and neural processes involved in musical
appreciation. In J. P. Huston, M. Nadal, F. Mora, L. Agnati, & C. J. Cela Conde (Eds.), Art,
aesthetics, and the brain (pp. 319–338). Oxford: Oxford University Press.
Pearce, M. T., Ruiz, M. H., Kapasi, S., Wiggins, G. A., & Bhattacharya, J. (2010). Unsupervised
statistical learning underpins computational, behavioural, and neural manifestations of musical
expectation. NeuroImage 50(1), 302–313.
Pearce, M. T., Zaidel, D. W., Vartanian, O., Skov, M., Leder, H., Chatterjee, A., & Nadal, M. (2016).
Neuroaesthetics: The cognitive neuroscience of aesthetic experience. Perspectives on
Psychological Science 11(2), 265–279.
Pelowski, M., Markey, P. S., Forster, M., Gerger, G., & Leder, H. (2017). Move me, astonish me …
delight my eyes and brain: The Vienna Integrated Model of top-down and bottom-up processes in
Art Perception (VIMAP) and corresponding affective, evaluative, and neurophysiological
correlates. Physics of Life Reviews 21, 80–125.
Pelowski, M., Markey, P. S., Lauring, J. O., & Leder, H. (2016). Visualizing the impact of art: An
update and comparison of current psychological models of art experience. Frontiers in Human
Neuroscience 10, 160. doi:10.3389/fnhum.2016.00160
Pereira, C. S., Teixeira, J., Figueiredo, P., Xavier, J., Castro, S. L., & Brattico, E. (2011). Music and
emotions in the brain: Familiarity matters. PloS ONE 6(11), e27241.
Peretz, I., & Coltheart, M. (2003). Modularity of music processing. Nature Neuroscience 6, 688–691.
Peretz, I., & Zatorre, R. (Eds.). (2003). The cognitive neuroscience of music. Oxford: Oxford
University Press.
Peretz, I., & Zatorre, R. J. (2005). Brain organization for music processing. Annual Review of
Psychology 56, 89–114
Quarto, T., Fasano, M. C., Taurisano, P., Fazio, L., Antonucci, L. A., Gelao, B., … Brattico, E.
(2017). Interaction between DRD2 variation and sound environment on mood and emotion-related
brain activity. Neuroscience 341, 9–17.
Redies, C. (2015). Combining universal beauty and cultural context in a unifying model of visual
aesthetic experience. Frontiers in Human Neuroscience 9, 218. Retrieved from
https://doi.org/10.3389/fnhum.2015.00218
Reybrouck, M., & Brattico, E. (2015). Neuroplasticity beyond sounds: Neural adaptations following
long-term musical aesthetic experiences. Brain Sciences 5(1), 69–91.
Rizzolatti, G., Fadiga, L., Gallese, V., & Fogassi, L. (1996). Premotor cortex and the recognition of
motor actions. Cognitive Brain Research 3(2), 131–141.
Sachs, M. E., Ellis, R. J., Schlaug, G., & Loui, P. (2016). Brain connectivity reflects human aesthetic
responses to music. Social Cognitive and Affective Neuroscience 11(6), 884–891.
Sacks, O. (2007). Musicophilia: Tales of music and the brain. New York: Vintage.
Salimpoor, V. N., Benovoy, M., Larcher, K., Dagher, A., & Zatorre, R. J. (2011). Anatomically
distinct dopamine release during anticipation and experience of peak emotion to music. Nature
Neuroscience 14, 257–262.
Salimpoor, V. N., Van Den Bosch, I., Kovacevic, N., McIntosh, A. R., Dagher, A., & Zatorre, R. J.
(2013). Interactions between the nucleus accumbens and auditory cortices predict music reward
value. Science 340(6129), 216–219.
Salimpoor, V. N., & Zattore, R. J. (2013). Neural interactions that give rise to musical pleasure.
Psychology of Aesthetics, Creativity, and the Arts 7, 62–75.
Samson, S., Dellacherie, D., & Platel, H. (2009). Emotional power of music in patients with memory
disorders: Clinical implications of cognitive neuroscience. Annals of the New York Academy of
Sciences 1169, 245–255.
Savage, P. E., Brown, S., Sakai, E., & Currie, T. E. (2015). Statistical universals reveal the structures
and functions of human music. Proceedings of the National Academy of Sciences 112, 8987–8992.
Schafer, T., Sedlmeier, P., Stadtler, C., & Huron, D. (2013). The psychological functions of music
listening. Frontiers in Psychology 4, 511. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00511
Scruton, R. (1999). The aesthetics of music. Oxford: Oxford University Press.
Seghdi, N., & Brattico, E. (in press). The phylogenetic roots of music. Biokulturelle Menneske.
Sloboda, J. A. (1985). The musical mind. Oxford: Oxford University Press.
Sloboda, J. A. (1992). Empirical studies of emotional response to music. In M. R. Jones & S.
Holleran (Eds.), Cognitive Bases of Musical Communication (pp. 33–46). Washington, DC:
American Psychological Association.
Smith, C. U. (2005). Evolutionary neurobiology and aesthetics. Perspectives in Biology and
Medicine 48(1), 17–30.
Steinbeis, N., & Koelsch, S. (2009). Understanding the intentions behind man-made products elicits
neural activity in areas dedicated to mental state attribution. Cerebral Cortex 19(3), 619–623.
Suzuki, M., Okamura, N., Kawachi, Y., Tashiro, M., Arao, H., Hoshishiba, T., … Yanai, K. (2008).
Discrete cortical regions associated with the musical beauty of major and minor chords. Cognitive,
Affective, & Behavioral Neuroscience 8(2), 126–31.
Tallis, R. (2008). The limitations of a neurological approach to art: Review of Neuroarthistory: From
Aristotle and Pliny to Baxandall and Zeki by John Onians (Yale University Press, 2008). Lancet
372, 19–20.
Tallis, R. (2011). Reflections of a metaphysical flaneur. London and New York: Routledge.
Tiihonen, M., Brattico, E., Maksimainen, J., Wikgren, J., & Saarikallio, S. (2017). Constituents of
music and visual-art related pleasure: A critical integrative literature review. Frontiers in
Psychology 8, 1218. Retrieved from https://doi.org/10.3389/fpsyg.2017.01218
Trost, W., Ethofer, T., Zentner, M., & Vuilleumier, P. (2012). Mapping aesthetic musical emotions in
the brain. Cerebral Cortex 22(12), 2769–2783.
Trost, W., Frühholz, S., Cochrane, T., Cojan, Y., & Vuilleumier, P. (2015). Temporal dynamics of
musical emotions examined through intersubject synchrony of brain activity. Social Cognitive and
Affective Neuroscience 10(12), 1705–1721.
Trost, W., Frühholz, S., Schön, D., Labbé, C., Pichon, S., Grandjean, D., & Vuilleumier, P. (2014).
Getting the beat: Entrainment of brain activity by musical rhythm and pleasantness. NeuroImage
103, 55–64.
Vartanian, O., & Goel, V. (2004). Neuroanatomical correlates of aesthetic preference for paintings.
Neuroreport 15(5), 893–897.
Vuust, P., & Kringelbach, M. L. (2010). The pleasure of making sense of music. Interdisciplinary
Science Reviews 35(2), 166–182.
Vuust, P., Ostergaard, L., Pallesen, K. J., Bailey, C., & Roepstorff, A. (2009). Predictive coding of
music: Brain responses to rhythmic incongruity. Cortex 45(1), 80–92.
Vuust, P., & Witek, M. A. (2014). Rhythmic complexity and predictive coding: A novel approach to
modeling rhythm and meter perception in music. Frontiers in Psychology 5, 1111. Retrieved from
https://doi.org/10.3389/fpsyg.2014.01111
Wassiliwizky, E., Koelsch, S., Wagner, V., Jacobsen, T., & Menninghaus, W. (2017). The emotional
power of poetry: Neural circuitry, psychophysiology and compositional principles. Social
Cognitive and Affective Neuroscience 12(8), 1229–1240.
Wilkins, R. W., Hodges, D. A., Laurienti, P. J., Steen, M., & Burdette, J. H. (2014). Network science
and the effects of music preference on functional brain connectivity: From Beethoven to Eminem.
Scientific Reports 4, 6130. doi:10.1038/srep06130
Witek, M. A., Clarke, E. F., Wallentin, M., Kringelbach, M. L., & Vuust, P. (2014). Syncopation,
body-movement and pleasure in groove music. PLoS ONE 9, e94446.
Witek, M. A., Gilson, M., Clarke, E. F., Wallentin, M., Deco, G., Kringelbach, M. L., & Vuust, P.
(forthcoming). The brain dynamics of musical groove: Whole-brain modelling of effective
connectivity reveals increased metastability of reward and motor networks. Nature
Communication.
Witek, M. A., Kringelbach, M. L., & Vuust, P. (2015). Musical rhythm and affect: Comment on “The
quartet theory of human emotions: An integrative and neurofunctional model” by S. Koelsch et al.
Physics of Life Reviews 13, 92–94.
Zatorre, R. J. (2015). Musical pleasure and reward: Mechanisms and dysfunction. Annals of the New
York Academy of Sciences 1337, 202–211.
Zeki, S. (1999). Inner vision: An exploration of art and the brain. Oxford: Oxford University Press.
Zeki, S. (2013). Clive Bell’s “Significant Form” and the neurobiology of aesthetics. Frontiers in
Human Neuroscience 7, 730. Retrieved from https://doi.org/10.3389/fnhum.2013.00730
Zeki, S. (2014). Neurobiology and the humanities. Neuron 84(1), 12–14.
Zhdanov, A., Nurminen, J., Baess, P., Hirvenkari, L., Jousmaki, V., Makela, J. P., … Parkkonen, L.
(2015). An internet-based real-time audiovisual link for dual MEG recordings. PLoS ONE 10,
e0128485.
Zhou, G., Bourguignon, M., Parkkonen, L., & Hari, R. (2016). Neural signatures of hand kinematics
in leaders vs. followers: A dual-MEG study. NeuroImage 125, 731–738.
CHAPT E R 16
D A N I E L E S C H Ö N A N D B E N JA MI N MO R I L L O N
O M M
L
C F O
M L
O R S
The problem that is raised here is the scale problem of human anatomy.
Historically, there has been a very rough distinction of music and language
in terms of hemispheric dominance and this led many people to believe that
language is processed by the left hemisphere and music by the right
hemisphere. We now clearly know that this is not the case (Lindell, 2006;
Vigneau et al., 2011). Then, there have been more specific claims that the
left Broca’s area would be language specific, but this has also been
falsified, by showing for instance that musical harmony (Koelsch et al.,
2002; Maess, Koelsch, Gunter, & Friederici, 2001) and rhythm processing
(Herdener et al., 2012) are mediated by the same regions processing
language syntax (Friederici & Kotz, 2003). Further work based on
multivariate pattern analysis has shown that within overlapping regions,
distinct brain patterns of responses can be found to linguistic and musical
sounds (Abrams et al., 2010; Fedorenko, McDermott, Norman-Haignere, &
Kanwisher, 2012). However, these differences could be accounted for in
terms of differences in the stimuli manipulation or in the task. For instance,
Abrams et al. (2010) compared scrambled versions of music and speech to
normal music and speech and used a fixed scrambling window of 350 ms.
As the authors acknowledge, it could be that music and speech have
inherently different acoustical regularities and structures, rendering one
material more “scrambled” than the other. Also, different patterns of
activation in common brain areas may result from the same neural
population reacting differently to music and language (Kunert & Slevc,
2015).
The argument raised by Peretz advocates for the possibility of music
dedicated neurons, adjacent to language dedicated neurons. While this is of
course a non-falsifiable hypothesis for the moment, one should not think of
music or language as a whole, but in terms of precisely defined elementary
operations. If these operations are required with both language and music
material, then there would be no reason for the brain to produce two
extremely intermingled networks computing the same algorithm. On the
other side it is clear that the rules determining gender agreement or those
affecting tonality modulations are necessarily represented in different neural
networks. Thus, claiming that differences may always subsist at a smaller
scale is a recursive argument that does not really add much to the debate
(besides the fact that at a quantum level, music and language can be
described by the same equations). In our view, the major advances will not
come from single unit recordings showing specific neurons to the last chord
of a precise Haydn piano sonata, but rather from neurocomputational
models precisely describing what particular operations are subtended by a
given neural network when listening to speech and to music.
A more promising approach seems to us to study whether two different
levels of music and language processing interact or not. Indeed, the
interaction is a measure of the extent to which two processes influence each
other and as such it can be used to infer that one process is not independent
of the other. Several studies have tackled this issue by using interference
paradigms. For instance Slevc and colleagues (Slevc, Rosenberg, & Patel,
2009) measured the reading time of garden path sentences and found that it
was influenced by simultaneous presentation of irrelevant harmonically
unexpected chords while it was not affected by timbrally unexpected chords
(e.g., a different instrument). These results have been interpreted as
evidence for shared music–language resources processing structural
(syntactic) relations. The task-irrelevant music being processed
automatically, it uses some resources resulting in a suboptimal processing
of the language syntactic relations. Other studies have used this approach to
show an interaction between melodic and syntactic processing (Fedorenko,
Patel, Casasanto, Winawer, & Gibson, 2009), harmonic and syntactic
processing but not semantic processing (Hoch, Poulin-Charronnat, &
Tillmann, 2011), and harmonic processing and word recall (Fiveash &
Pammer, 2014). This has also been coupled to electrophysiological
measures, confirming that melodic or harmonic unexpected events affect
the syntax-related left anterior negativity (Carrus, Pearce, & Bhattacharya,
2013; Koelsch, Gunter, Wittfoth, & Sammler, 2005). Interestingly Sammler
et al. (2013) showed a co-localization of early components elicited by
musical and linguistic syntactic deviations using intracranial recordings.
Surprisingly, few neuroimaging studies have exploited the possibly most
natural setting to compare music and language which is a stimulus that
combines both speech and music: song. The use of songs has the clear
advantage of solving the problem of using different stimuli in the language
and musical task. Schön et al. (2010) used an interference paradigm based
on sung sentences and showed that the processing demands of melodic and
lexical/phonological processing interact in a large network including
bilateral temporal cortex and left inferior frontal cortex. Importantly, most
voxels sensitive to the lexical/phonological manipulation are also sensitive
to the interaction between the lexical/phonological and the melodic
dimensions. In other words there seem to be very few voxels that are
involved in lexical/phonological and are not influenced by melodic
structure (see Fig. 1).
FIGURE 1. Number of surviving voxels for the main effect of lexical/phonological dimension as a
function of the threshold of the interaction between phonological and melodic dimensions. The
dotted vertical line indicates the p-value of 0.05 for the mask. The right edge corresponds to a very
conservative p-value (adapted from Schön et al., 2010).
M T L S
We have seen that the approach of studying music and language brain
correlates is limited by a number of methodological problems that render
the interpretation of the results in terms of sharing or not of the resources
rather complex. Another way to address the sharing resources hypothesis is
to investigate whether music training affects the way the brain processes
language, and vice versa. The reasoning is the following. Musical expertise
requires an intense training often starting at an early age. As a result of
learning, all the operations required by music perception and production
will be affected by this training and become more efficient. If some of these
operations are also required by language perception and production, then
one should be able to observe a more efficient processing whenever the
appropriate language processing levels are investigated. By contrast with
the approach described above, the validation of this hypothesis does not
necessarily require brain imaging data, insofar as behavioral differences can
be taken as evidence that resource sharing exists. Psychologists and some
neuroscientists often use the term “transfer of learning.” This term is,
however, rather vague as it seems to point to some sort of magic transfer of
learning from one domain to another or from one function to another
function without specifying how this transfer would actually take place.
However, an alternative explanation is to hypothesize that these so-called
transfer effects are simply due to an elementary function that is shared by
both music and language processing. According to this view there is no
transfer taking place, but only sharing of functions and resources.
Importantly, while there is no clear way of showing how transfer could be
possibly implemented, shared elementary operations can be defined via
careful experimental manipulations.
Considering the early steps of sound analysis helps to clarify this point.
The group of Nina Kraus has studied for many years the effect of music
training on sound perception in general, including speech. Using EEG and
focusing on high frequency (>200 Hz) neural responses, possibly
principally occurring at the subcortical level, this group of researchers has
shown that, compared to non-musicians, musicians have a stronger
representation of several features of speech sounds, including the
fundamental frequency (Wong, Skoe, Russo, Dees, & Kraus, 2007), the
harmonics (Kraus & Chandrasekaran, 2010), and rapid transients that may
be important in distinguishing consonants (Parbery-Clark, Tierney, Strait, &
Kraus, 2012). Overall, the correlation between the neural response and the
stimulus is greater in musicians than in non-musicians and this
independently of whether the stimulus is a music or a speech sound
(Musacchia, Sams, Skoe, & Kraus, 2007). Most importantly this correlation
is more resistant to acoustic noise in musicians. In other words, musicians
seem to be able to filter out the noise better than non-musicians (Parbery-
Clark, Skoe, & Kraus, 2009). Interestingly some of these differences can be
observed in adults that had a few years of music training during childhood,
thus showing that these changes last in time and do not necessarily require a
long-lasting and intense training (Skoe & Kraus, 2012).
Moreover, these differences induced by music training are not simply
due to a better processing of any sound feature. Indeed, results of a recent
experiment show that music training can facilitate the selective processing
of certain relevant features of speech. In this study, Intartaglia and
colleagues (Intartaglia, White-Schwoch, Kraus, & Schön, 2017) compared
French and American participants listening to an American phoneme, not
existing in French. The comparison of the neural signatures showed that
American participants had a more robust representation compared to French
participants. The differences concerned the high formant frequencies that
are necessary to encode the specific features of consonants and vowels.
They then tested French musicians and the differences with the Americans
disappeared. In other words, music training seems to allow a better
encoding of the relevant features of speech sounds, even when these sounds
are not familiar.
When interpreting these overall results one should keep in mind that two
possible non-exclusive explanations co-exist. First, the subcortical relays
may be more efficient in sound processing due to massive bottom-up
processing. In this case one can clearly see that there is no need to advocate
for a transfer effect. There is a dedicated auditory subcortical network that
processes both musical and linguistic sounds. If this network becomes more
efficient via intensive musical training, then speech processing will also
benefit from the enhanced efficiency. Second, the cortical regions are
known to send efferent signals to the subcortical relays and these
modulatory top-down signals may play a role in enhancing the
representation of certain features of sounds or in reducing the noise (Strait,
Kraus, Parbery-Clark, & Ashley, 2010; Tenenbaum, Kemp, Griffiths, &
Goodman, 2011). In this perspective, the changes are possibly due to an
enhanced connectivity that allows a finer modulatory activity of cortical
over subcortical activity.
Independently of whether these enhanced subcortical representations
reflect a bottom-up or a top-down modulation, these results are important in
interpreting the differences that may be observed at a more integrated level.
Indeed, differences observed at a phonological, syntactic, or prosodic level
may result from a cascade effect of early auditory processing differences.
The studies on prosody and phoneme perception in musicians are
particularly sensitive to this issue. Indeed, pitch is important in speech at
the supra-segmental level, by signaling the emotional content of an
utterance (Kotz et al., 2003), the linguistic structure (Steinhauer, Alter, &
Friederici, 1999), and certain syntactic features such as to determine
whether the utterance is a question or not (Astésano, Besson, & Alter,
2004). Pitch contour also plays a role at the segmental level in tone
languages: it plays a linguistically contrastive function.
Musicians are more accurate in detecting subtle pitch variations in both
music and speech prosody. These variations in speech prosody are detected
earlier by musicians’ brains and elicit more distinguishable event-related
potentials compared to normal speech (Schön et al., 2004). This has been
replicated with 8-year-old musician children (Magne, Schön, & Besson,
2006). Music lessons also seem to promote sensitivity to emotions
conveyed by speech prosody. Indeed, musically trained adults perform
better than untrained adults in discrimination and identification of
emotional prosody (Thompson, Schellenberg, & Husain, 2004). Finally,
musicians are more accurate at identifying, reproducing, or discriminating
Mandarin tones (Gottfried & Riester, 2000; Gottfried, Staby, & Ziemer,
2004; Marie, Delogu, Lampis, Belardinelli, & Besson, 2011). However, as
previously stated, it is difficult to know to what extent these differences are
due to cortical or subcortical plasticity. Considering that anatomical
differences have been observed at the cortical level in the auditory cortex
(Benner et al., 2017; Kleber et al., 2016; Schlaug, Jäncke, Huang, Staiger,
& Steinmetz, 1995; Shahin, Bosnyak, Trainor, & Roberts, 2003), it seems
reasonable to believe that the whole auditory network is modified by music
training, thus affecting speech processing at multiple levels.
Interestingly, previous studies provided evidence for a positive
relationship between the function or the anatomy of the planum temporale
and performance during syllable categorization (Elmer, Hänggi, Meyer, &
Jäncke, 2013). Recently, Elmer and colleagues (Elmer, Hänggi, & Jäncke,
2016) provided evidence for a relationship between planum temporale
connectivity, musicianship, and phonetic categorization. They found an
increased connectivity between the left and right plana temporalia in
musicians compared to non-musicians. This increased connectivity
positively correlated with the performance in a phonetic categorization task
as well as with musical aptitudes. Indeed, music training seems to affect the
sensitivity to some acoustic features that are important to categorization of
syllables, in particular temporal features such as voice-onset time (Chobert,
Marie, François, Schön, & Besson, 2011; Zuk et al., 2013).
Very few studies have examined whether musical expertise influences
the processing of the speech temporal structures. While isochrony is absent
in speech, several nested temporal hierarchies are present in speech
(Cummins & Port, 1998; Ghitza, 2011; Giraud & Poeppel, 2012).
Musicians outperform non-musicians when asked to judge the lengthening
of a syllable in a sentence (Marie, Magne, & Besson, 2011). Also,
independently of whether musicians direct attention to the temporal or
semantic content, they are more sensitive to subtle changes in the temporal
structure of speech than non-musicians (Magne et al., 2006; Marie, Delogu,
et al., 2011). Milovanov et al. (2009) reported a positive correlation
between musical aptitudes and sensitivity to syllable discrimination in
children. In artificial language learning, speech segmentation results from
the capacity to parse a continuous stream of syllables and to build and
maintain probabilistic relationship of the different elements (syllables) that
compose words. François and Schön (2011) showed that musicians have
improved segmentation skills compared to non-musicians. Indeed, when
listening to a new stream of an artificial language, they are faster and more
accurate at segmenting the continuous stream. Children, after only one year
of music training already show an improvement in speech segmentation
(François, Chobert, Besson, & Schön, 2012). This ability, namely
discovering word boundaries in the continuous stream of natural speech, is
of utmost importance during language learning in the first years of life
(Saffran et al., 1996).
The evidence concerning an effect of music training on language
semantic and syntactic levels is rather scarce. One study showed that music
training seems to influence semantic aspects of language processing
(Dittinger et al., 2016). However, in this study, French participants had to
learn new words that were in Thai language. Thus, differences may be due
to the difficulty of the task at the perceptual level in terms of discriminating
Thai tokens that differed in terms of pitch or vowel length. At the neural
level, results indicate an increased functional connectivity in the ventral and
dorsal streams of the left hemisphere during retrieval of novel words in
musicians compared to non-musicians (Dittinger, Valizadeh, Jäncke,
Besson, & Elmer, 2018). An effect of musical expertise on syntactic
processing was shown by Jentschke and Koelsch (2009) with earlier and
larger evoked responses to syntactic errors in children with musical
training. However, others described that differences are absent at the
behavioral level and that musical expertise does not modulate the amplitude
of responses evoked by syntactic violations but only the topographical
distribution (Fitzroy & Sanders, 2013). Thus, the evidence that music
training affects language semantic and syntactic processing is not yet
compelling and further studies are awaited.
Overall, while the theoretical framework of transfer of learning remains
uncertain, there is a rather massive amount of data pointing to an
improvement induced by music training at different levels of speech and
language processing. Patel (2014) has tried to formalize the conditions
under which music training may be beneficial to speech processing. In the
OPERA hypothesis (Overlap, Precision, Emotion, Repetition, and
Attention) he suggests that, in order for music training to enhance speech
processing, music and speech need to share sensory or cognitive processing
mechanisms and music must place higher demands on these mechanisms
compared to speech. These mechanisms are tightly bound to the music
emotional rewards system (Salimpoor et al., 2013). The last ingredients of
music-induced and speech-related neural plasticity would be the fact that
music training requires a repetition of sound patterns and gestures for an
enormous amount of time under conditions of highly focused attention.
B M L
A T F M
L
R
Abrams, D. A., Bhatara, A., Ryali, S., Balaban, E., Levitin, D. J., & Menon, V. (2010). Decoding
temporal structure in music and speech relies on shared brain resources but elicits different fine-
scale spatial patterns. Cerebral Cortex 21(7), 1507–1518.
Albert, M. L., Sparks, R. W., & Helm, N. A. (1973). Melodic intonation therapy for aphasia. Archives
of Neurology 29, 130–131.
Arnal, L. (2012). Predicting “when” using the motor system’s beta-band oscillations. Frontiers in
Human Neuroscience 6, 225. Retrieved from https://doi.org/10.3389/fnhum.2012.00225
Arnal, L., & Giraud, A. L. (2012). Cortical oscillations and sensory predictions. Trends in Cognitive
Sciences 16(7), 390–398.
Astésano, C., Besson, M., & Alter, K. (2004). Brain potentials during semantic and prosodic
processing in French. Cognitive Brain Research 18(2), 172–184.
Basso, A., & Capitani, E. (1985). Spared musical abilities in a conductor with global aphasia and
ideomotor apraxia. Journal of Neurology, Neurosurgery & Psychiatry 48(5), 407–412.
Bedoin, N., Brisseau, L., Molinier, P., Roch, D., & Tillmann, B. (2016). Temporally regular musical
primes facilitate subsequent syntax processing in children with specific language impairment.
Frontiers in Neuroscience 10. Retrieved from https://doi.org/10.3389/fnins.2016.00245
Belin, P., Zatorre, R. J., Lafaille, P., Ahad, P., & Pike, B. (2000). Voice-selective areas in human
auditory cortex. Nature 403(6767), 309–312.
Benner, J., Wengenroth, M., Reinhardt, J., Stippich, C., Schneider, P., & Blatow, M. (2017).
Prevalence and function of Heschl’s gyrus morphotypes in musicians. Brain Structure and
Function 222(8), 1–17.
Besle, J., Schevon, C. A., Mehta, A. D., Lakatos, P., Goodman, R. R., McKhann, G. M., …
Schroeder, C. E. (2011). Tuning of the human neocortex to the temporal dynamics of attended
events. Journal of Neuroscience 31(9), 3176–3185.
Bhide, A., Power, A., & Goswami, U. (2013). A rhythmic musical intervention for poor readers: A
comparison of efficacy with a letter-based intervention. Mind, Brain, and Education 7(2), 113–
123.
Carrus, E., Pearce, M. T., & Bhattacharya, J. (2013). Melodic pitch expectation interacts with neural
responses to syntactic but not semantic violations. Cortex 49(8), 2186–2200.
Cason, N., Astésano, C., & Schön, D. (2015). Bridging music and speech rhythm: Rhythmic priming
and audio-motor training affect speech perception. Acta Psychologica 155, 43–50.
Cason, N., Hidalgo, C., Isoard, F., Roman, S., & Schön, D. (2015). Rhythmic priming enhances
speech production abilities: Evidence from prelingually deaf children. Neuropsychology 29(1),
102.
Cason, N., & Schön, D. (2012). Rhythmic priming enhances the phonological processing of speech.
Neuropsychologia 50(11), 2652–2658.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008). Listening to musical rhythms recruits motor
regions of the brain. Cerebral Cortex 18(12), 2844–2854.
Chern, A., Tillmann, B., Vaughan, C., & Gordon, R. L. (2018). New evidence of a rhythmic priming
effect that enhances grammaticality judgments in children. Journal of Experimental Child
Psychology 173, 371–379.
Cheung, C., Hamilton, L. S., Johnson, K., & Chang, E. F. (2016). The auditory representation of
speech sounds in human motor cortex. eLife 5, e12577.
Chobert, J., François, C., Velay, J. L., & Besson, M. (2012). Twelve months of active musical
training in 8- to 10-year-old children enhances the preattentive processing of syllabic duration and
voice onset time. Cerebral Cortex 24(4), 956–967.
Chobert, J., Marie, C., François, C., Schön, D., & Besson, M. (2011). Enhanced passive and active
processing of syllables in musician children. Journal of Cognitive Neuroscience 23(12), 3874–
3887.
Chomsky, N. (1959). A review of B. F. Skinner’s Verbal Behavior. Language 35(1), 26–58.
Cogo-Moreira, H., de Avila, C. R. B., Ploubidis, G. B., & de Jesus Mari, J. (2013). Effectiveness of
music education for the improvement of reading skills and academic achievement in young poor
readers: A pragmatic cluster-randomized, controlled clinical trial. PloS ONE 8(3), e59984.
Coull, J. T. (2011). Discrete neuroanatomical substrates for generating and updating temporal
expectations. In S. Dehaene & E. Brannon (Eds.), Space, time and number in the brain: Searching
for the foundations of mathematical thought (pp. 87–101). Amsterdam: Elsevier.
Cravo, A. M., Rohenkohl, G., Wyart, V., & Nobre, A. C. (2013). Temporal expectation enhances
contrast sensitivity by phase entrainment of low-frequency oscillations in visual cortex. Journal of
Neuroscience 33(9), 4002–4010.
Cummins, F., & Port, R. (1998). Rhythmic constraints on stress timing in English. Journal of
Phonetics 26(2), 145–171.
DeWitt, L. A., & Samuel, A. G. (1990). The role of knowledge-based expectations in music
perception: Evidence from musical restoration. Journal of Experimental Psychology: General
119(2), 123–144.
Ding, N., Patel, A. D., Chen, L., Butler, H., Luo, C., & Poeppel, D. (2017). Temporal modulations in
speech and music. Neuroscience & Biobehavioral Reviews 81(B), 181–187.
Dittinger, E., Barbaroux, M., D’Imperio, M., Jäncke, L., Elmer, S., & Besson, M. (2016).
Professional music training and novel word learning: From faster semantic encoding to longer-
lasting word representations. Journal of Cognitive Neuroscience 28(10), 1584–1602.
Dittinger, E., Valizadeh, S. A., Jäncke, L., Besson, M., & Elmer, S. (2018). Increased functional
connectivity in the ventral and dorsal streams during retrieval of novel words in professional
musicians. Human Brain Mapping 39(2), 722–734.
Du, Y., & Zatorre, R. J. (2017). Musical training sharpens and bonds ears and tongue to hear speech
better. Proceedings of the National Academy of Sciences 5, 201712223. Retrieved from
https://doi.org/10.1073/pnas.1712223114
Elhilali, M., & Shamma, S. A. (2008). A cocktail party with a cortical twist: How cortical
mechanisms contribute to sound segregation. Journal of the Acoustical Society of America 124(6),
3751–3771.
Elmer, S., Hänggi, J., & Jäncke, L. (2016). Interhemispheric transcallosal connectivity between the
left and right planum temporale predicts musicianship, performance in temporal speech
processing, and functional specialization. Brain Structure and Function 221(1), 331–344.
Elmer, S., Hänggi, J., Meyer, M., & Jäncke, L. (2013). Increased cortical surface area of the left
planum temporale in musicians facilitates the categorization of phonetic and temporal speech
sounds. Cortex 49(10), 2812–2821.
Falk, S., Lanzilotti, C., & Schön, D. (2017). Tuning neural phase entrainment to speech. Journal of
Cognitive Neuroscience 29(8), 1378–1389.
Farbood, M. M., Marcus, G., & Poeppel, D. (2013). Temporal dynamics and the identification of
musical key. Journal of Experimental Psychology Human Perception & Performance 39(4), 911–
918.
Farbood, M. M., Rowland, J., Marcus, G., Ghitza, O., & Poeppel, D. (2015). Decoding time for the
identification of musical key. Attention, Perception, & Psychophysics 77(1), 28–35.
Fedorenko, E., McDermott, J. H., Norman-Haignere, S., & Kanwisher, N. (2012). Sensitivity to
musical structure in the human brain. Journal of Neurophysiology 108(12), 3289–3300.
Fedorenko, E., Patel, A., Casasanto, D., Winawer, J., & Gibson, E. (2009). Structural integration in
language and music: Evidence for a shared system. Memory & Cognition 37(1), 1–9.
Fitzroy, A. B., & Sanders, L. D. (2013). Musical expertise modulates early processing of syntactic
violations in language. Frontiers in Psychology 3, 603. Retrieved from
https://doi.org/10.3389/fpsyg.2012.00603
Fiveash, A., & Pammer, K. (2014). Music and language: Do they draw on similar syntactic working
memory resources? Psychology of Music 42(2), 190–209.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schön, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A randomized
control trial. PLoS ONE 10(9), e0138715.
Fodor, J. A. (1983). The modularity of mind: An essay on faculty psychology. Cambridge, MA: MIT
Press.
François, C., Chobert, J., Besson, M., & Schön, D. (2012). Music training for the development of
speech segmentation. Cerebral Cortex 23(9), 2038–2043.
François, C., & Schön, D. (2011). Musical expertise boosts implicit learning of both musical and
linguistic structures. Cerebral Cortex 21(10), 2357–2365.
Friederici, A. D., & Kotz, S. A. (2003). The brain basis of syntactic processes: Functional imaging
and lesion studies. NeuroImage 20, S8–S17.
Ghitza, O. (2011). Linking speech perception and neurophysiology: Speech decoding guided by
cascaded oscillators locked to the input rhythm. Frontiers in Psychology 2. Retrieved from
https://doi.org/10.3389/fpsyg.2011.00130
Ghitza, O., & Greenberg, S. (2009). On the possible role of brain rhythms in speech perception:
Intelligibility of time-compressed speech with periodic and aperiodic insertions of silence.
Phonetica 66, 113–126.
Giraud, A. L., Kleinschmidt, A., Poeppel, D., Lund, T. E., Frackowiak, R. S., & Laufs, H. (2007).
Endogenous cortical rhythms determine cerebral specialization for speech perception and
production. Neuron 56(6), 1127–1134.
Giraud, A. L., & Poeppel, D. (2012). Cortical oscillations and speech processing: Emerging
computational principles and operations. Nature Neuroscience 15(4), 511–517.
Goswami, U. (2011). A temporal sampling framework for developmental dyslexia. Trends in
Cognitive Sciences 15(1), 3–10.
Gottfried, T. L., & Riester, D. (2000). Relation of pitch glide perception and Mandarin tone
identification. Journal of the Acoustical Society of America 108(5), 2604.
Gottfried, T. L., Staby, A. M., & Ziemer, C. J. (2004). Musical experience and Mandarin tone
discrimination and imitation. Journal of the Acoustical Society of America 115(5), 2545.
Grahn, J. A., & Rowe, J. B. (2012). Finding and feeling the musical beat: Striatal dissociations
between detection and prediction of regularity. Cerebral Cortex 23(4), 913–921.
Haegens, S., & Golumbic, E. Z. (2018). Rhythmic facilitation of sensory processing: A critical
review. Neuroscience & Biobehavioral Reviews 86, 150–165.
Herdener, M., Humbel, T., Esposito, F., Habermeyer, B., Cattapan-Ludewig, K., & Seifritz, E.
(2012). Jazz drummers recruit language-specific areas for the processing of rhythmic structure.
Cerebral Cortex 24(3), 836–843.
Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews
Neuroscience 8, 393–402.
Hidalgo, C., Falk, S., & Schön, D. (2017). Speak on time! Effects of a musical rhythmic training on
children with hearing loss. Hearing Research 351, 11–18.
Hillis, A. E., & Caramazza, A. (1995). Representation of grammatical categories of words in the
brain. Journal of Cognitive Neuroscience 7(3), 396–407.
Hoch, L., Poulin-Charronnat, B., & Tillmann, B. (2011). The influence of task-irrelevant music on
language processing: Syntactic and semantic structures. Frontiers in Psychology 2. Retrieved from
https://doi.org/10.3389/fpsyg.2011.00112
Intartaglia, B., White-Schwoch, T., Kraus, N., & Schön, D. (2017). Music training enhances the
automatic neural processing of foreign speech sounds. Scientific Reports 7(1), 12631.
Jaramillo, S., & Zador, A. M. (2011). The auditory cortex mediates the perceptual effects of acoustic
temporal expectation. Nature Neuroscience 14, 246–251.
Jentschke, S., & Koelsch, S. (2009). Musical training modulates the development of syntax
processing in children. NeuroImage 47(2), 735–744.
Jones, M. R. (1976). Time, our lost dimension: Toward a new theory of perception, attention, and
memory. Psychological Review 83(5), 323–355.
Kleber, B., Veit, R., Moll, C. V., Gaser, C., Birbaumer, N., & Lotze, M. (2016). Voxel-based
morphometry in opera singers: Increased gray-matter volume in right somatosensory and auditory
cortices. NeuroImage 133, 477–483.
Koelsch, S., Gunter, T. C., Cramon, D. Y. V., Zysset, S., Lohmann, G., & Friederici, A. D. (2002).
Bach speaks: A cortical “language-network” serves the processing of music. NeuroImage 17(2),
956–966.
Koelsch, S., Gunter, T. C., Wittfoth, M., & Sammler, D. (2005). Interaction between syntax
processing in language and in music: An ERP study. Journal of Cognitive Neuroscience 17(10),
1565–1577.
Kotz, S. A., Gunter, T. C., & Wonneberger, S. (2005). The basal ganglia are receptive to rhythmic
compensation during auditory syntactic processing: ERP patient data. Brain and Language 95(1),
70–71.
Kotz, S. A., Meyer, M., Alter, K., Besson, M., von Cramon, D. Y., & Friederici, A. D. (2003). On the
lateralization of emotional prosody: An event-related functional MR investigation. Brain &
Language 86(3), 366–376.
Kotz, S. A., & Schwartze, M. (2010). Cortical speech processing unplugged: A timely subcortico-
cortical framework. Trends in Cognitive Sciences 14(9), 392–399.
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11(8), 599–605.
Kunert, R., & Slevc, L. R. (2015). A commentary on: “Neural overlap in processing music and
speech.” Frontiers in Human Neuroscience 9. Retrieved from
https://doi.org/10.3389/fnhum.2015.00330
Kunert, R., Willems, R. M., Casasanto, D., Patel, A. D., & Hagoort, P. (2015). Music and language
syntax interact in Broca’s area: An fMRI study. PloS One 10(11), e0141069.
Lakatos, P., Musacchia, G., O’Connel, M. N., Falchier, A. Y., Javitt, D. C., & Schroeder, C. E.
(2013). the spectrotemporal filter mechanism of auditory selective attention. Neuron 77, 750–761.
Large, E. W., & Jones, M. R. (1999). The dynamics of attending: How people track time-varying
events. Psychological Review 106(1), 119–159.
Liberman, A. M., Cooper, F. S., Shankweiler, D. P., & Studdert-Kennedy, M. (1967). Perception of
the speech code. Psychological Review 74, 431–461.
Liebenthal, E., Binder, J. R., Spitzer, S. M., Possing, E. T., & Medler, D. A. (2005). Neural substrates
of phonemic perception. Cerebral Cortex 15(10), 1621–1631.
Lindell, A. K. (2006). In your right mind: Right hemisphere contributions to language processing and
production. Neuropsychology Review 16(3), 131–148.
Luria, A. R., Tsevetkova, L. S., & Futer, D. S. (1965). Aphasia in a composer. Journal of the
Neurological Sciences 2(3), 288–292.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in
Broca’s area: an MEG study. Nature neuroscience, 4(5), 540–545.
Magne, C., Schön, D., & Besson, M. (2006). Musician children detect pitch violations in both music
and language better than nonmusician children: Behavioral and electrophysiological approaches.
Journal of Cognitive Neuroscience 18(2), 199–211.
Marie, C., Delogu, F., Lampis, G., Belardinelli, M. O., & Besson, M. (2011). Influence of musical
expertise on segmental and tonal processing in Mandarin Chinese. Journal of Cognitive
Neuroscience 23(10), 2701–2715.
Marie, C., Magne, C., & Besson, M. (2011). Musicians and the metric structure of words. Journal of
Cognitive Neuroscience 23(2), 294–305.
Marr, D. (1982). Vision. San Francisco, CA: Freeman.
Merchant, H., Grahn, J., Trainor, L., Rohrmeier, M., & Fitch, W.T. (2015). Finding the beat: A neural
perspective across humans and non-human primates. Philosophical Transactions of the Royal
Society B: Biological Sciences 370(1664), 20140093. doi:10.1098/rstb.2014.0093
Milovanov, R., Huotilainen, M., Esquef, P. A., Alku, P., Välimäki, V., & Tervaniemi, M. (2009). The
role of musical aptitude and language skills in preattentive duration processing in school-aged
children. Neuroscience Letters 460(2), 161–165.
Moore, E., Branigan, H. & Overy, K. (2017). Exploring the role of auditory-motor synchronisation in
the transfer of music to language skills in dyslexia. Outstanding Poster Award talk at
Neurosciences and Music VI conference.
Morillon, B., & Baillet, S. (2017). Motor origin of temporal predictions in auditory attention.
Proceedings of the National Academy of Sciences 114(42), E8913–E8921.
Morillon, B., Hackett, T. A., Kajikawa, Y., & Schroeder, C. E. (2015). Predictive motor control of
sensory dynamics in auditory active sensing. Current Opinion in Neurobiology 31, 230–238.
Morillon, B., Schroeder, C. E., & Wyart, V. (2014). Motor contributions to the temporal precision of
auditory attention. Nature Communications 5, 5255.
Morillon, B., Schroeder, C. E., Wyart, V., & Arnal, L. H. (2016). Temporal prediction in lieu of
periodic stimulation. Journal of Neuroscience 36(8), 2342–2347.
Musacchia, G., Sams, M., Skoe, E., & Kraus, N. (2007). Musicians have enhanced subcortical
auditory and audiovisual processing of speech and music. Proceedings of the National Academy of
Sciences 104(40), 15894–15898.
Nobre, A. C., & van Ede, F. (2018). Anticipated moments: Temporal structure in attention. Nature
Reviews Neuroscience 19, 34–38.
Overy, K. (2000). Dyslexia, temporal processing and music: The potential of music as an early
learning aid for dyslexic children. Psychology of Music 28(2), 218–229.
Parbery-Clark, A., Skoe, E., & Kraus, N. (2009). Musical experience limits the degradative effects of
background noise on the neural processing of sound. Journal of Neuroscience 29(45), 14100–
14107
Parbery-Clark, A., Tierney, A., Strait, D. L., & Kraus, N. (2012). Musicians have fine-tuned neural
distinction of speech syllables. Neuroscience 219, 111–119.
Park, H., Ince, R. A. A., Schyns, P. G., Thut, G., & Gross, J. (2015). Frontal top-down signals
increase coupling of auditory low-frequency oscillations to continuous speech in human listeners.
Current Biology 25(12), 1649–1653.
Patel, A. D. (2003). Language, music, syntax and the brain. Nature Neuroscience 6(7), 674–681.
Patel, A. D. (2011). Why would musical training benefit the neural encoding of speech? The OPERA
hypothesis. Frontiers in Psychology 2, 142. doi:10.3389/fpsyg.2011.00142
Patel, A. D. (2014). Can nonlinguistic musical training change the way the brain processes speech?
The expanded OPERA hypothesis. Hearing Research 308, 98–108.
Patel, A. D., & Iversen, J. R. (2014). The evolutionary neuroscience of musical beat perception: The
Action Simulation for Auditory Prediction (ASAP) hypothesis. Frontiers in System Neuroscience
8, 57. Retrieved from https://doi.org/10.3389/fnsys.2014.00057
Patel, A. D., Iversen, J. R., Wassenaar, M., & Hagoort, P. (2008). Musical syntactic processing in
agrammatic Broca’s aphasia. Aphasiology 22(7–8), 776–789.
Peretz, I., & Coltheart, M. (2003). Modularity of music processing. Nature Neuroscience 6(7), 688–
691.
Peretz, I., Vuvan, D., Lagrois, M. É., & Armony, J. L. (2015). Neural overlap in processing music
and speech. Philosophical Transactions of the Royal Society B: Biological Sciences 370(1664),
20140090.
Poeppel, D. (2003). The analysis of speech in different temporal integration windows: Cerebral
lateralization as “asymmetric sampling in time.” Speech Communication 41(1), 245–255.
Poeppel, D. (2012). The maps problem and the mapping problem: Two challenges for a cognitive
neuroscience of speech and language. Cognitive Neuropsychology 29(1–2), 34–55.
Przybylski, L., Bedoin, N., Krifi-Papoz, S., Herbillon, V., Roch, D., Léculier, L., … Tillmann, B.
(2013). Rhythmic auditory stimulation influences syntactic processing in children with
developmental language disorders. Neuropsychology 27(1), 121–131.
Rauschecker, J. P., & Scott, S. K. (2009). Maps and streams in the auditory cortex: Nonhuman
primates illuminate human speech processing. Nature Neuroscience 12, 718–724.
Rauschecker, J. P., & Tian, B. (2000). Mechanisms and streams for processing of “what” and
“where” in auditory cortex. Proceedings of the National Academy of Sciences 97(22), 11800–
11806.
Rimmele, J. M., Morillon, B., Poeppel, D., & Arnal, L. H. (submitted). The proactive and flexible
sense of timing.
Rizzolatti, G., & Craighero, L. (2004). The mirror-neuron system. Annual Review of Neuroscience
27, 169–192.
Rogalsky, C., & Hickok, G. (2011). The role of Broca’s area in sentence comprehension. Journal of
Cognitive Neuroscience 23(7), 1664–1680.
Rohenkohl, G., Cravo, A. M., Wyart, V., & Nobre, A. C. (2012). Temporal expectation improves the
quality of sensory information. Journal of Neuroscience 32(24), 8424–8428.
Saffran, J. R., Aslin, R. N., & Newport, E. L. (1996). Statistical learning by 8-month-old infants.
Science 274(5294), 1926–1928.
Saffran, J. R., Johnson, E. K., Aslin, R. N., & Newport, E. L. (1999). Statistical learning of tone
sequences by human infants and adults. Cognition 70(1), 27–52.
Salimpoor, V. N., van den Bosch, I., Kovacevic, N., McIntosh, A. R., Dagher, A., & Zatorre, R. J.
(2013). Interactions between the nucleus accumbens and auditory cortices predict music reward
value. Science 340(6129), 216–219.
Sammler, D., Baird, A., Valabrègue, R., Clément, S., Dupont, S., Belin, P., & Samson, S. (2010). The
relationship of lyrics and tunes in the processing of unfamiliar songs: A functional magnetic
resonance adaptation study. Journal of Neuroscience 30(10), 3572–3578.
Sammler, D., Koelsch, S., Ball, T., Brandt, A., Grigutsch, M., Huppertz, H. J., … Friederici, A. D.
(2013). Co-localizing linguistic and musical syntax with intracranial EEG. NeuroImage 64, 134–
146.
Sammler, D., Koelsch, S., & Friederici, A. D. (2011). Are left fronto-temporal brain areas a
prerequisite for normal music-syntactic processing? Cortex 47(6), 659–673.
Schlaug, G., Jäncke, L., Huang, Y., Staiger, J. F., & Steinmetz, H. (1995). Increased corpus callosum
size in musicians. Neuropsychologia 33(8), 1047–1055.
Schön, D., Boyer, M., Moreno, S., Besson, M., Peretz, I., & Kolinsky, R. (2008). Songs as an aid for
language acquisition. Cognition 106(2), 975–983.
Schön, D., Gordon, R., Campagne, A., Magne, C., Astésano, C., Anton, J. L., & Besson, M. (2010).
Similar cerebral networks in language, music and song perception. NeuroImage 51(1), 450–461.
Schön, D., Magne, C., & Besson, M. (2004). The music of speech: Music training facilitates pitch
processing in both music and language. Psychophysiology 41(3), 341–349.
Schön, D., & Tillmann, B. (2015). Short- and long-term rhythmic interventions: Perspectives for
language rehabilitation. Annals of the New York Academy of Sciences 1337, 32–39.
Schroeder, C. E., & Lakatos, P. (2009). Low-frequency neuronal oscillations as instruments of
sensory selection. Trends in Neurosciences 32(1), 9–18.
Schroeder, C. E., Wilson, D. A., Radman, T., Scharfman, H., & Lakatos, P. (2010). Dynamics of
active sensing and perceptual selection. Current Opinion in Neurobiology 20, 172–176.
Schubotz, R. I. (2007). Prediction of external events with our motor system: Towards a new
framework. Trends in Cognitive Sciences 11(5), 211–218.
Shahin, A., Bosnyak, D. J., Trainor, L. J., & Roberts, L. E. (2003). Enhancement of neuroplastic P2
and N1c auditory evoked potentials in musicians. Journal of Neuroscience 23(13), 5545–5552.
Skoe, E., & Kraus, N. (2012). A little goes a long way: How the adult brain is shaped by musical
training in childhood. Journal of Neuroscience 32(34), 11507–11510.
Slevc, L. R., Faroqi-Shah, Y., Saxena, S., & Okada, B. M. (2016). Preserved processing of musical
structure in a person with agrammatic aphasia. Neurocase 22(6), 505–511.
Slevc, L. R., Rosenberg, J. C., & Patel, A. D. (2009). Making psycholinguistics musical: Self-paced
reading time evidence for shared processing of linguistic and musical syntax. Psychonomic
Bulletin & Review 16(2), 374–381.
Staeren, N., Renvall, H., De Martino, F., Goebel, R., & Formisano, E. (2009). Sound categories are
represented as distributed patterns in the human auditory cortex. Current Biology 19(6), 498–502.
Stahl, B., Kotz, S. A., Henseler, I., Turner, R., & Geyer, S. (2011). Rhythm in disguise: Why singing
may not hold the key to recovery from aphasia. Brain 134(10), 3083–3093.
Stefanics, G., Hangya, B., Hernadi, I., Winkler, I., Lakatos, P., & Ulbert, I. (2010). Phase entrainment
of human delta oscillations can mediate the effects of expectation on reaction speed. Journal of
Neuroscience 30(41), 13578–13585.
Steinhauer, K., Alter, K., & Friederici, A. D. (1999). Brain potentials indicate immediate use of
prosodic cues in natural speech processing. Nature Neuroscience 2(2), 191–196.
Strait, D. L., Kraus, N., Parbery-Clark, A., & Ashley, R. (2010). Musical experience shapes top-down
auditory mechanisms: Evidence from masking and auditory attention performance. Hearing
Research 261(1), 22–29.
Teki, S., Grube, M., Kumar, S., & Griffiths, T. D. (2011). Distinct neural substrates of duration-based
and beat-based auditory timing. Journal of Neuroscience 31(10), 3805–3812.
Tenenbaum, J. B., Kemp, C., Griffiths, T. L., & Goodman, N. D. (2011). How to grow a mind:
Statistics, structure, and abstraction. Science 331(6022), 1279–1285.
Thompson, W. F., Schellenberg, E. G., & Husain, G. (2004). Decoding speech prosody: Do music
lessons help? Emotion 4(1), 46–64.
Tierney, A., & Kraus, N. (2014). Auditory-motor entrainment and phonological skills: Precise
auditory timing hypothesis (PATH). Frontiers in Human Neuroscience 8. Retrieved from
https://doi.org/10.3389/fnhum.2014.00949
Tinbergen, N. (1963). On aims and methods of ethology. Ethology 20, 410–433.
Vallentin, D., Kosche, G., Lipkind, D., & Long, M. A. (2016). Inhibition protects acquired song
segments during vocal learning in zebra finches, Science 351(6270), 267–271.
Vigneau, M., Beaucousin, V., Hervé, P. Y., Jobard, G., Petit, L., Crivello, F., … Tzourio-Mazoyer, N.
(2011). What is right-hemisphere contribution to phonological, lexico-semantic, and sentence
processing? Insights from a meta-analysis. NeuroImage 54(1), 577–593.
Vuust, P., & Witek, M. A. G. (2014). Rhythmic complexity and predictive coding: A novel approach
to modeling rhythm and meter perception in music. Frontiers in Psychology 5, 1111. Retrieved
from https://doi.org/10.3389/fpsyg.2014.01111
Wong, P. C., Skoe, E., Russo, N. M., Dees, T., & Kraus, N. (2007). Musical experience shapes human
brainstem encoding of linguistic pitch patterns. Nature Neuroscience 10(4), 420–422.
Zatorre, R. J., Belin, P., & Penhune, V. B. (2002). Structure and function of auditory cortex: Music
and speech. Trends in Cognitive Sciences 6(1), 37–46.
Zatorre, R. J., Chen, J. L., & Penhune, V. B. (2007). When the brain plays music: Auditory–motor
interactions in music perception and production. Nature Reviews Neuroscience 8, 547–558.
Zuk, J., Ozernov-Palchik, O., Kim, H., Lakshminarayanan, K., Gabrieli, J. D. E., Tallal, P., & Gaab,
N. (2013). Enhanced syllable discrimination thresholds in musicians. PLoS ONE 8(12), e80546.
Zumbansen, A., Peretz, I., & Hébert, S. (2014). Melodic intonation therapy: Back to basics for future
research. Frontiers in Neurology 5. Retrieved from https://doi.org/10.3389/fneur.2014.00007
SECTION V
MU S IC IA N S HIP A N D
B R A IN F U N C T ION
CHAPT E R 17
M U S I C A L E X P E RT I S E A N D
BRAIN STRUCTURE: THE
CAUSES AND
CONSEQUENCES OF
TRAINING
VIRGINIA B. PENHUNE
O the past twenty years, brain imaging studies have demonstrated that
music training can change brain structure, predominantly in the auditory-
motor network that underlies music performance. These studies have also
shown that brain structural variation is related to performance on a range of
musical tasks, and that even short-term training can result in brain plasticity.
In this chapter, we will argue that the observed differences in brain structure
between experts and novices derive from at least four sources. First, there
may be pre-existing individual differences in structural features supporting
specific skills that predispose people to undertake music training. Second,
lengthy and consistent training likely produces structural change in the
brain networks tapped by performance through repeated cycles of
prediction, feedback, and error-correction that drive learning. Third, the
timing of practice during specific periods of development may result in
brain changes that do not occur at other periods of time, and which may
promote future learning and plasticity. Fourth, both the rewarding nature of
music itself, as well as the reward value of practice and accurate
performance may make music training a particularly effective driver of
brain plasticity.
S B D
A M
D I T -
R P
T I D
T
So, why does training affect the cerebellum differently than the cortex,
and how do these findings challenge our understanding of the effects of
early experience? There are several features of cerebellar anatomy that may
explain this result. First, developmental studies show that peak growth in
the cerebellum occurs much later than in most of the cortex, between the
ages of 12 and 18 (Tiemeier et al., 2010). Thus early experience may have a
different effect on cerebellar plasticity, such that experience leads to greater
efficiency and reduced expansion. Second, the cerebellum is unique in
being structurally homogeneous, with the identical cytoarchitecture and
input–output circuitry throughout (Schmahmann, 1997). In the motor
system, the cerebellar circuits are known to play a role in error-correction
and optimization. Because these circuits are uniform across the structure, it
is hypothesized that they perform the same role in optimizing a wide variety
of functions in the regions to which it is connected (Balsters, Whelan,
Robertson, & Ramnani, 2013; Koziol et al., 2014; Sokolov et al., 2017).
The cerebellar regions that are smaller in ET musicians in our study are
connected to frontal motor and association regions, including M1, PMC,
and prefrontal cortex (Diedrichsen, Balsters, Flavell, Cussans, & Ramnani,
2009; Kelly & Strick, 2003). Based on this information, it is possible that
training-related skills and cortical expansion might be supported by greater
optimization and reduced expansion in the cerebellum. If this is true, then
cortical and cerebellar changes with training should be inversely related.
A S -T T
B I A T
W I M S E
D B P ?
Why does music training produce such robust changes in brain structure?
One very obvious answer is practice—lots of practice. For the studies
reviewed here, the average length of training for musicians was 15–20
years. This is the equivalent of thousands of hours of practice across a large
portion of the person’s life. While the idea that simply practicing long
enough will result in expertise has been largely debunked (for review, see
Mosing et al., 2014), long-term, consistent practice is strongly associated
with expertise in a range of domains (Macnamara, Hambrick, & Oswald,
2014). Further, in the studies reviewed here, the length of training is
typically strongly related to both structural brain differences and task
performance. The impact of practice on brain organization is supported by
studies in animals showing that practice on new motor tasks is associated
with expanded representations in motor areas (Elbert, Pantev, Wienbruch,
Rockstroh, & Taub, 1995; Nudo, Milliken, Jenkins, & Merzenich, 1996),
changes in MR measures of gray- and white-matter (Scholz, Allemang-
Grand, Dazai, & Lerch, 2015; Scholz, Niibori, Frankland, & Lerch, 2015),
and increased numbers of synapses and dendritic spines (Kleim, Barnaby, et
al., 2002; Kleim, Freeman, et al., 2002; Kleim et al., 2004). Neuronal
changes in gray matter that are related to learning include neurogenesis,
synaptogenesis, and changes in neuronal morphology. In white matter,
changes related to learning including increases in the number of axons,
axon diameter, packing density of fibers, and myelination can be found
(Zatorre, Fields, & Johansen-Berg, 2012).
A second reason that music training may be particularly effective in
driving brain plasticity is the highly specific nature of practice. The
majority of musicians are experts on a single instrument; thus they perform
millions of repetitions of the same movements, and listen attentively to an
even larger number of associated sounds. When practicing, a musician
imagines and plans a precise sequence of sounds and the movements
required to produce them. Once the plan is set in motion, they use auditory
and somatosensory information to detect subtle deviations in sound and
movement, implementing adjustments to enhance performance. Practice is
therefore a repeated prediction, feedback, and error-correction cycle.
Auditory-motor prediction is thought to be a central function of the dorsal
stream, particularly of the premotor cortex. Brain imaging studies have
shown increased activity in the PMC when people listen to melodies that
they have learned to play (Chen, Rae, & Watkins, 2012; Lahav, Saltzman,
& Schlaug, 2007), and recent work from our laboratory has shown that
transcranical magnetic stimulation (TMS) over dorsal PMC disrupts
learning of auditory-motor associations (Lega, Stephan, Zatorre, &
Penhune, 2016). Feedback and error-correction are key components of
motor learning (Diedrichsen, Shadmehr, & Ivry, 2010; Sokolov et al., 2017;
Wolpert, Diedrichsen, & Flanagan, 2011), and studies of both motor and
sensory learning show that functional and structural changes in the brain are
driven by decreases in error and improved precision. For example, learning
to juggle (Scholz, Klein, Behrens, & Johansen-Berg, 2009), balance on a
tilting board (Taubert et al., 2010), or to perform a complex visuomotor task
(Lakhani et al., 2016; Landi, Baguear, & Della-Maggiore, 2011) have all
been shown to produce changes in gray- or white-matter architecture that
were related to decreases in error with learning. Thus error-driven learning,
particularly during periods of high developmental plasticity may be an
important contributor to structural brain changes measured in adult
musicians.
Another reason that music training may be so successful in producing
brain plasticity is that it is inherently multisensory. To produce music,
performers must learn to link sounds to actions, but they must also link
visual, somatosensory, and proprioceptive feedback to these sounds and
actions. As described earlier, training is a prediction to feedback to error-
correction cycle in which musicians use all their sensory resources to
produce the perfect sound. Sounds are linked to actions relatively rapidly, as
has been shown by changes in the strength of motor activity during passive
listening to learned melodies after short-term training (Bangert et al., 2006;
D’Ausilio, Altenmüller, Olivetti Belardinelli, & Lotze, 2006; Lega et al.,
2016; Stephan, Brown, Lega, & Penhune, 2016). In particular, it was shown
that learning to play a melody resulted in greater changes in the activity of
auditory cortex than learning to remember the melody by listening alone
(Lappe, Herholz, Trainor, & Pantev, 2008). This may partly be based on
strong intrinsic connections between the auditory and motor systems (Chen
et al., 2012; Poeppel, 2014; Zatorre, Chen, & Penhune, 2007). But it can
also be hypothesized that co-activation of circuits deriving from multiple
senses may drive plasticity even more strongly than input from a single
sense (Lee & Noppeney, 2011, 2014).
A final feature of music training that is likely crucial in promoting
plasticity is the rewarding nature of performance. There are three aspects of
reward that may stimulate plasticity: first, the rewarding nature of music
itself that is experienced through playing; second, the intrinsic reward of
performing, both for the player and through the acclaim it may bring; and
finally, the potentially rewarding nature of practice and the pleasure of
accurate performance. The intrinsic pleasure derived from music appears to
be common to most people (Mas-Herrero, Marco-Pallares, Lorenzo-Seva,
Zatorre, & Rodriguez-Fornells, 2011), and is hypothesized to be based on
the same dopamine-modulated, predictive systems that regulate reward in
other domains with direct biological consequences, including drugs, food,
sex, and money (Salimpoor, Zald, Zatorre, Dagher, & McIntosh, 2015).
Thus learning to produce a rewarding stimulus, such as music, is likely to
be rewarding to the player.
We also know that learning and brain plasticity are strongly affected by
the reward value of what is learned. Animal studies show that brain
plasticity associated with auditory learning is greater when the information
to be learned is rewarded, or behaviorally relevant. For example, the
responses of neurons in the auditory cortex of ferrets were modulated by the
reward value of stimuli (David, Fritz, & Shamma, 2012). Further, pairing a
tone with stimulation of dopamine circuits in the brainstem increased the
selectivity of responding in auditory neurons tuned to the same tone (Bao,
Chan, & Merzenich, 2001). Importantly, dopamine has been shown to
modulate motor learning in both humans and animals (Floel et al., 2005;
Tremblay et al., 2009, 2010); possibly through the reinforcement and habit-
formation circuitry of the striatum (Graybiel & Grafton, 2015; Haith &
Krakauer, 2013). Thus, if the output of practice, a beautiful piece of music,
is rewarding and stimulates dopamine release, then playing such a piece
should promote learning. It is also likely that the social benefits of playing
music add to this type of reward.
Finally, humans seem to have a strong internal motivation to practice
and perfect many skills, even if those skills do not have immediate
physiological, psychological, or social outcomes. In addition to music,
people spend hours perfecting their golf swing, playing video games, or
baking elaborate cakes. All of these skills require practice, and the outcome
of practice is often not immediate. Thus we hypothesize that practice itself
may be rewarding, and that the prediction–feedback–error-correction cycle
that is important for learning, may be motivating across a range of domains.
When musicians are learning a new and challenging piece, or perfecting an
old one, they know exactly what they want it to sound like. This
representation is translated into a motor plan, and both the imagined
outcome and the plan become predictions against which they will measure
their performance. When musicians attempt to play the piece, they will
likely make errors, which lead to corrections and learning; but when they
play the piece as imagined, they experience the reward of accurate
performance. Because error feedback and reward are so important for
learning, these mechanisms seem like strong candidates for promoting brain
plasticity, but have been little explored.
W D W G F H ?
Bringing together the data from this review, we suggest three directions for
future research.
R
Abdul-Kareem, I. A., Stancak, A., Parkes, L. M., Al-Ameen, M., Alghamdi, J., Aldhafeeri, F. M., …
Sluming, V. (2011). Plasticity of the superior and middle cerebellar peduncles in musicians
revealed by quantitative analysis of volume and number of streamlines based on diffusion tensor
tractography. Cerebellum 10(3), 611–623.
Abdul-Kareem, I. A., Stancak, A., Parkes, L. M., & Sluming, V. (2011). Increased gray matter
volume of left pars opercularis in male orchestral musicians correlate positively with years of
musical performance. Journal of Magnetic Resonance Imaging 33(1), 24–32.
Altenmüller, E., & Furuya, S. (2016). Brain plasticity and the concept of metaplasticity in skilled
musicians. Advances in Experimental Medicine and Biology 957, 197–208.
Amunts, K., Schlaug, G., Jancke, L., Steinmetz, H., Schleicher, A., Dabringhaus, A., & Zilles, K.
(1997). Motor cortex and hand motor skills: Structural compliance in the human brain. Human
Brain Mapping 5(3), 206–215.
Andersen, R. A., & Cui, H. (2009). Intention, action planning, and decision making in parietal-frontal
circuits. Neuron 63(5), 568–583.
Baer, L., Park, M., Bailey, J., Chakravarty, M., Li, K., & Penhune, V. (2015). Regional cerebellar
volumes are related to early musical training and finger tapping performance. NeuroImage 109,
130–139.
Bailey, J. A., & Penhune, V. B. (2010). Rhythm synchronization performance and auditory working
memory in early- and late-trained musicians. Experimental Brain Research 204(1), 91–101.
Bailey, J. A., & Penhune, V. B. (2012). A sensitive period for musical training: Contributions of age
of onset and cognitive abilities. Annals of the New York Academy of Sciences 1252, 163–170.
Bailey, J. A., & Penhune, V. B. (2013). The relationship between the age of onset of musical training
and rhythm synchronization performance: Validation of sensitive period effects. Frontiers in
Neuroscience 7, 227. Retrieved from https://doi.org/10.3389/fnins.2013.00227
Bailey, J. A., Zatorre, R. J., & Penhune, V. B. (2014). Early musical training is linked to gray matter
structure in the ventral premotor cortex and auditory-motor rhythm synchronization performance.
Journal of Cognitive Neuroscience 26(4), 755–767.
Balsters, J. H., Whelan, C. D., Robertson, I. H., & Ramnani, N. (2013). Cerebellum and cognition:
Evidence for the encoding of higher order rules. Cerebral Cortex 23(6), 1433–1443.
Bangert, M., Peschel, T., Schlaug, G., Rotte, M., Drescher, D., Hinrichs, H., … Altenmüller, E.
(2006). Shared networks for auditory and motor processing in professional pianists: Evidence from
fMRI conjunction. NeuroImage 30(3), 917–926.
Bao, S., Chan, V. T., & Merzenich, M. M. (2001). Cortical remodelling induced by activity of ventral
tegmental dopamine neurons. Nature 412(6842), 79–83.
Bengtsson, S., Nagy, Z., Skare, S., Forsman, L., Forssberg, H., & Ullén, F. (2005). Extensive piano
practicing has regionally specific effects on white matter development. Nature Neuroscience 8(9),
1148–1150.
Bermudez, P., Lerch, J. P., Evans, A. C., & Zatorre, R. J. (2009). Neuroanatomical correlates of
musicianship as revealed by cortical thickness and voxel-based morphometry. Cerebral Cortex
19(7), 1583–1596.
Blecher, T., Tal, I., & Ben-Shachar, M. (2016). White matter microstructural properties correlate with
sensorimotor synchronization abilities. NeuroImage 138, 1–12.
Brown, R. M., Zatorre, R. J., & Penhune, V. B. (2015). Expert music performance: Cognitive, neural,
and developmental bases. Progress in Brain Research 217, 57–86.
Butkovic, A., Ullén, F., & Mosing, M. A. (2015). Personality-related traits as predictors of music
practice: Underlying environmental and genetic influences. Personality and Individual Differences
74, 133–138.
Chang, E. F., & Merzenich, M. M. (2003). Environmental noise retards auditory cortical
development. Science 300(5618), 498–502.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008). Moving on time: Brain network for auditory-
motor synchronization is modulated by rhythm complexity and musical training. Journal of
Cognitive Neuroscience 20(2), 226–239.
Chen, J. L., Rae, C., & Watkins, K. E. (2012). Learning to play a melody: An fMRI study examining
the formation of auditory-motor associations. NeuroImage 59(2), 1200–1208.
Chiang, M. C., Barysheva, M., Shattuck, D. W., Lee, A. D., Madsen, S. K., Avedissian, C., …
Thompson, P. M. (2009). Genetics of brain fiber architecture and intellectual performance. Journal
of Neuroscience 29(7), 2212–2224.
D’Ausilio, A., Altenmüller, E., Olivetti Belardinelli, M., & Lotze, M. (2006). Cross-modal plasticity
of the motor cortex while listening to a rehearsed musical piece. European Journal of
Neuroscience 24(3), 955–958.
Dalla Bella, S., Farrugia, N., Benoit, C. E., Begel, V., Verga, L., Harding, E., & Kotz, S. A. (2017).
BAASTA: Battery for the Assessment of Auditory Sensorimotor and Timing Abilities. Behavior
Research Methods 49(3), 1128–1145.
David, S. V., Fritz, J. B., & Shamma, S. A. (2012). Task reward structure shapes rapid receptive field
plasticity in auditory cortex. Proceedings of the National Academy of Sciences 109(6), 2144–2149.
de Manzano, O., & Ullén, F. (2018). Same genes, different brains: Neuroanatomical differences
between monozygotic twins discordant for musical training. Cerebral Cortex 28(1), 387–394.
de Villers-Sidani, E., Chang, E. F., Bao, S., & Merzenich, M. M. (2007). Critical period window for
spectral tuning defined in the primary auditory cortex (A1) in the rat. Journal of Neuroscience
27(1), 180–189.
Diedrichsen, J., Balsters, J. H., Flavell, J., Cussans, E., & Ramnani, N. (2009). A probabilistic MR
atlas of the human cerebellum. NeuroImage 46(1), 39–46.
Diedrichsen, J., Shadmehr, R., & Ivry, R. B. (2010). The coordination of movement: Optimal
feedback control and beyond. Trends in Cognitive Sciences 14(1), 31–39.
E, K. H., Chen, S. H., Ho, M. H., & Desmond, J. E. (2014). A meta-analysis of cerebellar
contributions to higher cognition from PET and fMRI studies. Human Brain Mapping 35(2), 593–
615.
Elbert, T., Pantev, C., Wienbruch, C., Rockstroh, B., & Taub, E. (1995). Increased cortical
representation of the fingers of the left hand in string players. Science 270(5234), 305–307.
Engel, A., Hijmans, B. S., Cerliani, L., Bangert, M., Nanetti, L., Keller, P. E., & Keysers, C. (2014).
Inter-individual differences in audio-motor learning of piano melodies and white matter fiber tract
architecture. Human Brain Mapping 35(5), 2483–2497.
Floel, A., Breitenstein, C., Hummel, F., Celnik, P., Gingert, C., Sawaki, L., … Cohen, L. G. (2005).
Dopaminergic influences on formation of a motor memory. Annals of Neurology 58(1), 121–130.
Foster, N. E., & Zatorre, R. J. (2010). Cortical structure predicts success in performing musical
transformation judgments. NeuroImage 53(1), 26–36.
Gaser, C., & Schlaug, G. (2003). Brain structure differences between musicians and non-musicians.
Journal of Neuroscience 23(27), 9240–9245.
Giedd, J., Blumenthal, J., Jeffries, N., Castellanos, F., Liu, H., Zijdenbos, A., … Rapoport, J. (1999).
Brain development during childhood and adolescence: A longitudinal MRI study. Nature
Neuroscience 2(10), 861–863.
Gogos, A., Gavrilescu, M., Davison, S., Searle, K., Adams, J., Rossell, S. L., … Egan, G. F. (2010).
Greater superior than inferior parietal lobule activation with increasing rotation angle during
mental rotation: An fMRI study. Neuropsychologia 48(2), 529–535.
Golestani, N., Molko, N., Dehaene, S., LeBihan, D., & Pallier, C. (2007). Brain structure predicts the
learning of foreign speech sounds. Cerebral Cortex 17(3), 575–582.
Golestani, N., Paus, T., & Zatorre, R. (2002). Anatomical correlates of learning novel speech sounds.
Neuron 35, 997–1010.
Graybiel, A. M., & Grafton, S. T. (2015). The striatum: Where skills and habits meet. Cold Spring
Harbor Perspectives in Biology 7(8), a021691. doi:10.1101/cshperspect.a021691
Habibi, A., Damasio, A., Ilari, B., Veiga, R., Joshi, A. A., Leahy, R. M., … Damasio, H. (2017).
Childhood music training induces change in micro and macroscopic brain structure: Results from a
longitudinal study. Cerebral Cortex 1–12. doi:10.1093/cercor/bhx286
Haith, A. M., & Krakauer, J. W. (2013). Model-based and model-free mechanisms of human motor
learning. Advances in Experimental Medicine and Biology 782, 1–21.
Halwani, G. F., Loui, P., Ruber, T., & Schlaug, G. (2011). Effects of practice and experience on the
arcuate fasciculus: Comparing singers, instrumentalists, and non-musicians. Frontiers in
Psychology 2, 156. Retrieved from https://doi.org/10.3389/fpsyg.2011.00156
Herholz, S. C., Boh, B., & Pantev, C. (2011). Musical training modulates encoding of higher-order
regularities in the auditory cortex. European Journal of Neuroscience 34(3), 524–529.
Herholz, S. C., & Zatorre, R. (2012). Musical training as a framework for brain plasticity: Behavior,
function, and structure. Neuron 76(3), 486–502.
Hund-Georgiadis, M., & von Cramon, D. (1999). Motor-learning-related changes in piano players
and non-musicians revealed by functional magnetic-resonance signals. Experimental Brain
Research 125(4), 417–425.
Hutchinson, S., Lee, L. H., Gaab, N., & Schlaug, G. (2003). Cerebellar volume of musicians.
Cerebral Cortex 13(9), 943–949.
Hyde, K. L., Lerch, J., Norton, A., Forgeard, M., Winner, E., Evans, A. C., & Schlaug, G. (2009).
Musical training shapes structural brain development. Journal of Neuroscience 29(10), 3019–3025.
Ireland, K., Parker, A., Foster, N., & Penhune, V. (in press). Rhythm and melody tasks for school-
aged children with and without musical training: Age-equivalent scores and reliability. Frontiers in
Auditory Cognitive Neuroscience.
Ivry, R. B., Spencer, R. M., Zelaznik, H. N., & Diedrichsen, J. (2002). The cerebellum and event
timing. Annals of the New York Academy of Sciences 978, 302–317.
James, C. E., Oechslin, M. S., Van De Ville, D., Hauert, C. A., Descloux, C., & Lazeyras, F. (2014).
Musical training intensity yields opposite effects on grey matter density in cognitive versus
sensorimotor networks. Brain Structure & Function 219(1), 353–366.
Jorgensen, H. (2011). Instrumental learning: Is an early start a key to success? British Journal of
Music Education 18(3), 227–239.
Karpati, F. J., Giacosa, C., Foster, N. E. V., Penhune, V. B., & Hyde, K. L. (2017). Dance and music
share gray matter structural correlates. Brain Research 1657, 62–73.
Kelly, R., & Strick, P. (2003). Cerebellar loops with motor cortex and prefrontal cortex of a non-
human primate. Journal of Neuroscience 23(23), 8432–8444.
Kleim, J. A., Barnaby, S., Cooper, N., Hogg, T., Reidel, C., Remple, M., & Nudo, R. (2002). Motor
learning-dependent synaptogenesis is localized to functionally reorganized motor cortex.
Neurobiology of Learning and Memory 77(1), 63–77.
Kleim, J. A., Freeman, J. H., Jr., Bruneau, R., Nolan, B. C., Cooper, N. R., Zook, A., & Walters, D.
(2002). Synapse formation is associated with memory storage in the cerebellum. Proceedings of
the National Academy of Sciences 99(20), 13228–13231.
Kleim, J. A., Hogg, T., VandenBerg, P., Cooper, N., Bruneau, R., & Remple, M. (2004). Cortical
synaptogenesis and motor map reorganziation occur during late, but not early, phase of motor skill
learning. Journal of Neuroscience 24(3), 628–633.
Kolb, B., Mychasiuk, R., Muhammad, A., Li, Y., Frost, D. O., & Gibb, R. (2012). Experience and the
developing prefrontal cortex. Proceedings of the National Academy of Sciences 109(Suppl. 2),
17186–17193.
Koziol, L. F., Budding, D., Andreasen, N., D’Arrigo, S., Bulgheroni, S., Imamizu, H., … Yamazaki,
T. (2014). Consensus paper: The cerebellum’s role in movement and cognition. Cerebellum 13(1),
151–177.
Lahav, A., Saltzman, E., & Schlaug, G. (2007). Action representation of sound: Audiomotor
recognition network while listening to newly acquired actions. Journal of Neuroscience 27(2),
308–314.
Lakhani, B., Borich, M. R., Jackson, J. N., Wadden, K. P., Peters, S., Villamayor, A., … Boyd, L. A.
(2016). Motor skill acquisition promotes human brain myelin plasticity. Neural Plasticity 2016,
7526135. doi:10.1155/2016/7526135
Landi, S. M., Baguear, F., & Della-Maggiore, V. (2011). One week of motor adaptation induces
structural changes in primary motor cortex that predict long-term memory one year later. Journal
of Neuroscience 31(33), 11808–11813.
Lappe, C., Herholz, S. C., Trainor, L. J., & Pantev, C. (2008). Cortical plasticity induced by short-
term unimodal and multimodal musical training. Journal of Neuroscience 28(39), 9632–9639.
Lee, H., & Noppeney, U. (2011). Long-term music training tunes how the brain temporally binds
signals from multiple senses. Proceedings of the National Academy of Sciences 108(51), E1441–
E1450.
Lee, H., & Noppeney, U. (2014). Music expertise shapes audiovisual temporal integration windows
for speech, sinewave speech, and music. Frontiers in Psychology 5, 868. Retrieved from
https://doi.org/10.3389/fpsyg.2014.00868
Lega, C., Stephan, M. A., Zatorre, R. J., & Penhune, V. (2016). Testing the role of dorsal premotor
cortex in auditory-motor association learning using transcranical magnetic stimulation (TMS).
PLoS ONE 11(9), e0163380.
Li, X., De Beuckelaer, A., Guo, J., Ma, F., Xu, M., & Liu, J. (2014). The gray matter volume of the
amygdala is correlated with the perception of melodic intervals: a voxel-based morphometry study.
PLoS ONE 9(6), e99889.
Macnamara, B. N., Hambrick, D. Z., & Oswald, F. L. (2014). Deliberate practice and performance in
music, games, sports, education, and professions: A meta-analysis. Psychological Science 25(8),
1608–1618.
Mas-Herrero, E., Marco-Pallares, J., Lorenzo-Seva, U., Zatorre, R. J., & Rodriguez-Fornells, A.
(2011). Individual differences in music reward experiences. Music Perception 31(2), 118–138.
Mosing, M. A., Madison, G., Pedersen, N. L., Kuja-Halkola, R., & Ullén, F. (2014). Practice does not
make perfect: No causal effect of music practice on music ability. Psychological Science 25(9),
1795–1803.
Mullensiefen, D., Gingras, B., Musil, J., & Stewart, L. (2014). The musicality of non-musicians: An
index for assessing musical sophistication in the general population. PLoS ONE 9(2), e89642.
Novembre, G., & Keller, P. E. (2014). A conceptual review on action-perception coupling in the
musicians’ brain: What is it good for? Frontiers in Human Neuroscience 8, 603. Retrieved from
https://doi.org/10.3389/fnhum.2014.00603
Nudo, R., Milliken, G., Jenkins, W., & Merzenich, M. (1996). Use-dependent alterations of
movement representations in primary motor cortex of adult squirrel monkeys. Journal of
Neuroscience 16(2), 785–807.
Paquette, S., Fujii, S., Li, H. C., & Schlaug, G. (2017). The cerebellum’s contribution to beat interval
discrimination. NeuroImage 163, 177–182.
Peretz, I., Gosselin, N., Nan, Y., Caron-Caplette, E., Trehub, S. E., & Beland, R. (2013). A novel tool
for evaluating children’s musical abilities across age and culture. Frontiers in Systems
Neuroscience 7, 30. Retrieved from https://doi.org/10.3389/fnsys.2013.00030
Poeppel, D. (2014). The neuroanatomic and neurophysiological infrastructure for speech and
language. Current Opinion in Neurobiology 28, 142–149.
Putkinen, V., Tervaniemi, M., Saarikivi, K., Ojala, P., & Huotilainen, M. (2014). Enhanced
development of auditory change detection in musically trained school-aged children: A
longitudinal event-related potential study. Developmental Science 17(2), 282–297.
Ragert, P., Schmidt, A., Altenmüller, E., & Dinse, H. (2004). Superior tactile performance and
learning in professional pianists: Evidence for meta-plasticity in musicians. European Journal of
Neuroscience 19(2), 473–478.
Rauschecker, J. (2011). An expanded role for the dorsal auditory pathway in sensorimotor control
and integration. Hearing Research 271, 16–25.
Repp, B. (2005). Sensorimotor synchronization: A review of the tapping literature. Psychonomic
Bulletin and Review 12(6), 969–992.
Rosenkranz, K., Williamon, A., & Rothwell, J. C. (2007). Motorcortical excitability and synaptic
plasticity is enhanced in professional musicians. Journal of Neuroscience 27(19), 5200–5206.
Salimpoor, V. N., Zald, D. H., Zatorre, R. J., Dagher, A., & McIntosh, A. R. (2015). Predictions and
the brain: How musical sounds become rewarding. Trends in Cognitive Sciences 19(2), 86–91.
Schlaug, G., Jancke, L., Huang, Y., Staiger, J. F., & Steinmetz, H. (1995). Increased corpus callosum
size in musicians. Neuropsychologia 33(8), 1047–1055.
Schmahmann, J. (1997). The cerebrocerebellar system. In J. Schmahmann (Ed.), The Cerebellum and
Cognition (Vol. 41, pp. 31–55). San Diego, CA: Academic Press.
Schneider, P., Scherg, M., Dosch, H., Specht, H., Gutschalk, A., & Rupp, A. (2002). Morphology of
Heschl’s gyrus reflects enhanced activation in the auditory cortex of musicians. Nature
Neuroscience 5(7), 688–694.
Schneider, P., Sluming, V., Roberts, N., Scherg, M., Goebel, R., Specht, H. J., … Rupp, A. (2005).
Structural and functional asymmetry of lateral Heschl’s gyrus reflects pitch perception preference.
Nature Neuroscience 8(9), 1241–1247.
Scholz, J., Allemang-Grand, R., Dazai, J., & Lerch, J. P. (2015). Environmental enrichment is
associated with rapid volumetric brain changes in adult mice. NeuroImage 109, 190–198.
Scholz, J., Klein, M. C., Behrens, T. E., & Johansen-Berg, H. (2009). Training induces changes in
white-matter architecture. Nature Neuroscience 12(11), 1370–1371.
Scholz, J., Niibori, Y., Frankland, P. W., & Lerch, J. P. (2015). Rotarod training in mice is associated
with changes in brain structure observable with multimodal MRI. NeuroImage 107, 182–189.
Seither-Preisler, A., Parncutt, R., & Schneider, P. (2014). Size and synchronization of auditory cortex
promotes musical, literacy, and attentional skills in children. Journal of Neuroscience 34(33),
10937–10949.
Sluming, V., Barrick, T., Howard, M., Cezayirli, E., Mayes, A., & Roberts, N. (2002). Voxel-based
morphometry reveals increased gray matter density in Broca’s area in male symphony orchestra
musicians. NeuroImage 17(3), 1613–1622.
Sokolov, A. A., Miall, R. C., & Ivry, R. B. (2017). The cerebellum: Adaptive prediction for
movement and cognition. Trends in Cognitive Sciences 21(5), 313–332.
Steele, C. J., Bailey, J. A., Zatorre, R. J., & Penhune, V. B. (2013). Early musical training and white-
matter plasticity in the corpus callosum: Evidence for a sensitive period. Journal of Neuroscience
33(3), 1282–1290.
Stephan, M. A., Brown, R., Lega, C., & Penhune, V. (2016). Melodic priming of motor sequence
performance: The role of the dorsal premotor cortex. Frontiers in Neuroscience 10, 210. Retrieved
from https://www.frontiersin.org/articles/10.3389/fnins.2016.00210
Taubert, M., Draganski, B., Anwander, A., Muller, K., Horstmann, A., Villringer, A., & Ragert, P.
(2010). Dynamic properties of human brain structure: Learning-related changes in cortical areas
and associated fiber connections. Journal of Neuroscience 30(35), 11670–11677.
Tiemeier, H., Lenroot, R. K., Greenstein, D. K., Tran, L., Pierson, R., & Giedd, J. N. (2010).
Cerebellum development during childhood and adolescence: A longitudinal morphometric MRI
study. NeuroImage 49(1), 63–70.
Tremblay, P. L., Bedard, M. A., Langlois, D., Blanchet, P. J., Lemay, M., & Parent, M. (2010).
Movement chunking during sequence learning is a dopamine-dependent process: A study
conducted in Parkinson’s disease. Experimental Brain Research 205(3), 375–385.
Tremblay, P. L., Bedard, M. A., Levesque, M., Chebli, M., Parent, M., Courtemanche, R., &
Blanchet, P. J. (2009). Motor sequence learning in primate: Role of the D2 receptor in movement
chunking during consolidation. Behavioural Brain Research 198(1), 231–239.
Ullén, F., Hambrick, D. Z., & Mosing, M. A. (2016). Rethinking expertise: A multifactorial gene–
environment interaction model of expert performance. Psychological Bulletin 142(4), 427–446.
Vaquero, L., Hartmann, K., Ripolles, P., Rojo, N., Sierpowska, J., Francois, C., … Altenmüller, E.
(2016). Structural neuroplasticity in expert pianists depends on the age of musical training onset.
NeuroImage 126, 106–119.
Vaquero, L., Rodriguez-Fornells, A., & Reiterer, S. M. (2017). The left, the better: White-matter
brain integrity predicts foreign language imitation ability. Cerebral Cortex 27(8), 3906–3917.
Westerhausen, R., Luders, E., Specht, K., Ofte, S. H., Toga, A. W., Thompson, P. M., … Hugdahl, K.
(2011). Structural and functional reorganization of the corpus callosum between the age of 6 and 8
years. Cerebral Cortex 21(5), 1012–1017.
Wolpert, D. M., Diedrichsen, J., & Flanagan, J. R. (2011). Principles of sensorimotor learning.
Nature Reviews Neuroscience 12(12), 739–751.
Wong, P. C., Warrier, C. M., Penhune, V. B., Roy, A. K., Sadehh, A., Parrish, T. B., & Zatorre, R. J.
(2008). Volume of left Heschl’s gyrus and linguistic pitch learning. Cerebral Cortex 18(4), 828–
836.
Zatorre, R. J., Chen, J., & Penhune, V. (2007). When the brain plays music: Sensory-motor
interactions in music perception and production. Nature Reviews Neuroscience 8, 547–558.
Zatorre, R. J., Fields, R. D., & Johansen-Berg, H. (2012). Plasticity in gray and white: Neuroimaging
changes in brain structure during learning. Nature Neuroscience 15(4), 528–536.
CHAPT E R 18
GENOMICS APPROACHES
FOR STUDYING MUSICAL
A P T I T U D E A N D R E L AT E D
TRAITS
I R MA JÄ RV E L Ä
G A S H
T
M A B
T
FIGURE 2. Parental music education is related to children’s music education. High music
education is common among parents of professional musicians (n = 100).
Reproduced from Irma Järvelä, Genomics studies on musical aptitude, music perception, and
practice, Annals of the New York Academy of Sciences, Special Issue: The Neurosciences and
Music 6, p. 2, Figure 1, doi:10.1111/nyas.13620, Copyright © 2018, New York Academy of
Sciences.
E M A
G -W L A
A M T
For assigning genetic markers associated with a trait such as musical
aptitude the definition of the phenotype is required. As musical aptitude is a
complex cognitive trait, it is likely that its individual components have
distinct molecular backgrounds. Each of these components (subphenotype)
can be analyzed separately and they can also be combined.
In a genome-wide study of musical aptitude nearly 800 family members
were defined for auditory structuring ability (Karma Music Test, KMT)
(Karma, 1994) and perception of pitch and time (Seashore et al., 1960) in
music and a combined test score of all the three aforementioned test scores
(COMB). When the family material was analyzed for 660,000 genetic
markers several genetic loci were found in the human genome (Oikkonen et
al., 2015). The identified loci contained candidate genes that affect inner ear
development and neurocognitive processes, which are necessary traits for
music perception. The highest probability of linkage was obtained at 4q22
(Oikkonen et al., 2015). Earlier, chromosome 4q22 was found in a smaller
family material using microsatellite marker scan (Pulli et al., 2008). The
strongest association (in unrelated subjects) was found upstream of GATA
binding protein 2 (GATA2) at chromosome 3q21.3. GATA2 is a relevant
candidate gene as it regulates the development of cochlear hair cells
(Haugas, Lilleväli, Hakanen, & Salminen, 2010) and the inferior colliculus
(IC) (Lahti, Achim, & Partanen, 2013) important in tonotopic mapping, that
is, the processing of sounds of different frequency in the brain.
Interestingly, GATA2 is abundantly expressed in dopaminergic neurons
(Scherzer et al., 2008) that release dopamine during emotional arousal to
music (Salimpoor et al., 2011). Several plausible candidate genes were
located at 4p14 with the highest probability of linkage in the family study
(Oikkonen et al., 2015). The pitch perception accuracy (SP) was linked next
to the protocadherin 7 gene (PCHD7), expressed in the cochlear (Lin et al.,
2012) and amygdaloid (Hertel, Redies, & Medina, 2012) complexes.
PCHD7 is a relevant candidate gene for pitch perception functioning in the
hair cells of the cochlea that recognize pitches (Gosselin, Peretz, Johnsen,
& Adolphs, 2007). The amygdala is the emotional center of the human
brain affected by music (Koelsch, 2010). Interestingly, the homologous
gene PCDH15 also affects hair cell sensory transduction and together with
cadherin type 23 (CDH23) (another candidate gene at chromosome 16)
form a tip-link with each other in sensory hair cells (Sotomayor, Weihofen,
Gaudet, & Corey, 2012). Moreover, the Pcdha–gene cluster was found in
the CNV-study of musical aptitude (Ukkola-Vuoti et al., 2013). Platelet-
derived growth factor receptor alpha polypeptide (PDGFRA) is expressed
in the hippocampus (Di Pasquale et al., 2003), associated with learning and
memory. The potassium channel tetramerisation domain containing 8
(KCTD8) is expressed in the spiral ganglion of the cochlea (Metz,
Gassmann, Fakler, Schaeren-Wiemers, & Bettler, 2011). KCTD8 also
interacts with the GABA receptors GABRB1 and GABRB2; of these,
GABRb1 protein is reduced in schizophrenia, bipolar disorder, and major
depression, diseases that severely affect human cognition and mood
regulation (Fatemi, Folsom, Rooney, & Thuras, 2013). Cholinergic
receptor, nicotinic alpha 9 (neuronal) (CHRNA9) (Katz et al., 2004) and the
paired-like homeobox 2b (PHOX2B) (Ousdal et al., 2012) on chromosome 4
also affect inner ear development. In addition, PHOX2B increases amygdala
activity and autonomic functions (blood pressure, heart rate, and
respiration) that are reported to be affected by music (Blood & Zatorre,
2001). The genome-wide analyses performed on Mongolian families using
the pitch perception accuracy (PPA) test identified a partly shared genetic
region on chromosome 4q (Park et al., 2012). The statistically most
significant locus found in a genome-wide linkage study of absolute pitch
(AP) is located at 8q24.21 (Theusch, Basu, & Gitschier, 2009). The results
suggest that musical aptitude is an innate ability that is affected by several
predisposing genetic variants (Fig. 1).
Genome-wide copy number variation (CNV) analysis revealed regions
that contain candidate genes for neuropsychiatric disorders were associated
with musical aptitude (Ukkola-Vuoti et al., 2013). A deletion covering the
protocadherin-a gene cluster 1–9 (PCDHA 1–9) was associated with low
music test scores (COMB) both in familial and sporadic cases. PCDHAs
affect synaptogenesis and maturation of the serotonergic projections in the
brain and Pcdha mutant mice show abnormalities in learning and memory
(Katori et al., 2009).
T E M P
P H T
Music acts as an environmental trigger. Numerous studies have shown that
listening to and performing classical music have an effect on the human
body (Blood & Zatorre, 2001; Salimpoor et al., 2011). When comparing
genome-wide RNA expression profiles before and after listening to
classical music and after a “music-free” control session, the activity of
genes involved in dopamine secretion and transport (SNCA, RTN4, and
SLC6A8), and learning and memory (SNCA, NRGN, NPTN, RTN4) were
enhanced (Kanduri, Kuusi, et al., 2015). Of these genes, SNCA (George,
Jin, Woods, & Clayton, 1995), NRGN (Wood, Olson, Lovell, & Mello,
2008), and RGS2 affect song learning and singing in songbirds (Clayton,
2000) suggesting a shared evolutionary background of sound perception
between vocalizing birds and humans. It is noteworthy that the effect of
music was only detectable in musically-experienced listeners. The lack of
the effect of music in novices could be explained by differences in the
amount of exposure to music that is known to affect brain structure and
function (Elbert, Pantev, Wienbruch, Rockstroh, & Taub, 1995; Gaser &
Schlaug, 2003), unfamiliarity with the music (Salimpoor, Benovoy, Longo,
Cooperstock, & Zatorre, 2009), or musical anhedonia (Martínez-Molina,
Mas-Herrero, Rodríguez-Fornells, Zatorre, & Marco-Pallarés, 2016). In
addition, listening to music increased the expression of the target genes of
the dopaminoceptive neuronal glucocorticoid receptor (NR3C1), which
increases the synaptic concentration of dopamine linked to rewarding and
reinforcing properties (Ambroggi et al., 2009). It is of note that NR3C1 is
also a key molecule in the regulation of addictive behavior.
Music performance by professional musicians involves a wide spectrum
of cognitive and multisensory motor skills, whose molecular basis is largely
unknown. The effect of music performance on the genome-wide peripheral
blood transcriptome of professional musicians was analyzed by collecting
RNA-samples before and after a two-hour concert performance and after a
“music-free” control session. The upregulated genes were found to affect
dopaminergic neurotransmission, motor behavior, neuronal plasticity, and
neurocognitive functions including learning and memory. Specifically,
performance of music by professional musicians increased the expression of
FOS, DUSP1, SNCA, catecholamine biosynthesis, and dopamine
metabolism (Kanduri, Raijas, et al., 2015). Interestingly, SNCA, FOS, and
DUSP1 are involved in song perception and production in songbirds. Thus,
both listening to and performing music shared partially the same genes as
those affected in songbird singing. It is noteworthy that although the brains
of songbirds are small they have a double density neuron structure
compared to primate brains of the same mass. Thus, the large number of
neurons can contribute to the neural basis of cognitive capacity (Enard,
2016).
In both listening to and performing music (Kanduri, Kuusi, et al., 2015;
Kanduri, Raijas, et al., 2015), one of the strongest activations was detected
in the alfa-synuclein gene (SNCA), which has a physiological role in the
development of nerve cells and releases neurotransmitters, especially
dopamine from the presynaptic cells. Dopamine is responsible for motor
functions and genes known to affect growth and the plasticity of nerve cells
and the inactivation of genes affecting neurodegeneration (Kanduri, Raijas,
et al., 2015). SNCA is located in the best linkage region of musical aptitude
on chromosome 4q22.1 and is regulated by GATA2 residing at 3q21, the
region with the most significant association with musical aptitude thus
linking the results of the GWA study and transcriptional profiling studies to
the same locus (Kanduri, Kuusi, et al., 2015; Kanduri, Raijas, et al., 2015;
Oikkonen et al., 2015) (Fig. 3). GATA2 is abundantly expressed in
dopaminergic neurons and binds to intron-1 of endogenous neuronal SNCA
to regulate its expression. The results are in agreement with
neurophysiological studies where increases in endogenous dopamine have
been detected in the striatum when listening to music (Blood & Zatorre,
2001). Interestingly, SNCA is a causative gene for Parkinson’s disease (with
disturbed dopamine metabolism) (Petrucci, Ginevrino, & Valente, 2016)
and variations in SNCA predispose to Lewy-body dementia (Peuralinna et
al., 2008).
Listening to music and music performance had partially different effects
on gene expression. Some genes such as ZNF223 and PPP2R3A were
downregulated after music listening but upregulated after music
performance (Kanduri, Kuusi, et al., 2015; Kanduri, Raijas, et al., 2015).
ZNF223 is a zinc-finger transcription regulator and similar to an immediate
early response gene (IEG) ZNF225 (also known as ZENK, EGR1) that
regulates the song control system of songbirds (Dong & Clayton, 2008).
PPP2R3A, abundantly expressed in the striatum, is known to integrate the
effects of dopamine and other neurotransmitters (Ahn et al., 2007). Other
IEGs such as FOS and DUSP1 that are known to be responsible for the
song control nuclei of songbirds were upregulated only after music
performance (Kanduri, Kuusi, et al., 2015), but not music listening
(Kanduri, Raijas, et al., 2015). Many other song perception-related genes in
songbirds like RGS2 were found to be differentially regulated after listening
to music, but not after music performance (Kanduri, Kuusi, et al., 2015;
Kanduri, Raijas, et al., 2015). The reasons for the differences are plausibly
due, for example, to different types of musical activity and different study
subjects.
FIGURE 3. The results of DNA- and RNA-studies of music-related traits converge at chromosome
4q22. The alpha-synuclein gene (SNCA) upregulated by listening to music (Kanduri, Raijas, et al.,
2015) and music performance by professional musicians (Kanduri, Kuusi, et al., 2015) is located at
the most significant region of musical aptitude (Oikkonen et al., 2015; Park et al., 2012; Pulli et al.,
2008) and regulated by GATA2, associated with musical aptitude (Oikkonen et al., 2015).
Reproduced from Irma Järvelä, Genomics studies on musical aptitude, music perception, and
practice, Annals of the New York Academy of Sciences, Special Issue: The Neurosciences and
Music 6, p. 4, Fig. 2, doi:10.1111/nyas.13620, Copyright © 2018, New York Academy of
Sciences.
C A
B B C
A M
R
Ahn, J. H., Sung, J. Y., McAvoy, T., Nishi, A., Janssens, V., Goris, J., … Nairn, A. C. (2007). The
B"/PR72 subunit mediates Ca2+-dependent dephosphorylation of DARPP-32 by protein
phosphatase 2A. Proceedings of the National Academy of Sciences 104(23), 9876–9881.
Alcock, K. J., Passingham, R. E., Watkins, K., & Vargha-Khadem, F. (2000). Pitch and timing
abilities in inherited speech and language impairment. Brain & Language 75(1), 34–46.
Ambroggi, F., Turiault, M., Milet, A., Deroche-Gamonet, V., Parnaudeau, S., Balado, E., … Tronche,
F. (2009). Stress and addiction: Glucocorticoid receptor in dopaminoceptive neurons facilitates
cocaine seeking. Nature Neuroscience 12(3), 247–249.
Araki, M., Bandi, M. M., & Yazaki-Sukiyama, Y. (2016). Mind the gap: Neural coding of species
identity in birdsong prosody. Science 354(6317), 1282–1287.
Atik, T., Bademci, G., Diaz-Horta, O., Blanton, S. H., & Tekin, M. (2017). Whole-exome sequencing
and its impact in hereditary hearing loss. Genetics Research 97, e4.
doi:10.1017/S001667231500004X
Avey, M. T., Kanyo, R. A., Irwin, E. L., & Sturdy, C. B. (2008). Differential effects of vocalization
type, singer and listener on ZENK immediate early gene response in black-capped chickadees
(Poecile atricapillus). Behavioural Brain Research 188(1), 201–208.
Ayub, Q., Yngvadottir, B., Chen, Y., Xue, Y., Hu, M., Vernes, S. C., … Tyler-Smith, C. (2013).
FOXP2 targets show evidence of positive selection in European populations. American Journal of
Human Genetics 92(5), 696–706.
Baharloo, S., Service, S. K., Risch, N., Gitschier, J., & Freimer, N. B. (2000). Familial aggregation of
absolute pitch. American Journal of Human Genetics 67(3), 755–758.
Blood, A. J., & Zatorre, R. J. (2001). Intensely pleasurable responses to music correlate with activity
in brain regions implicate in reward and emotion. Proceedings of the National Academy of
Sciences 98(20), 11818–11823.
Brown, S., Martinez, M. J., & Parsons, L. M. (2006). Music and language side by side in the brain: A
PET study of the generation of melodies and sentences. European Journal of Neuroscience 23(10),
2791–2803.
Cáceres, M., Lachuer, J., Zapala, M. A., Redmond, J. C., Kudo, L., Geschwind, D. H., … Barlow, C.
(2003). Elevated gene expression levels distinguish human from non-human primate brains.
Proceedings of the National Academy of Sciences 100(22), 13030–13035.
Clayton, D. F. (2000). The genomic action potential. Neurobiology of Learning and Memory 74(3),
185–216.
Clayton, D. F. (2013). The genomics of memory and learning in songbirds. Annual Review of
Genomics and Human Genetics 14, 45–65.
Di Pasquale, G., Davidson, B. L., Stein, C. S., Martins, I., Scudiero, D., Monks, A., & Chiorini, J. A.
(2003). Identification of PDGFR as a receptor for AAV-5 transduction. Nature Medicine 9, 1306–
1312.
Dietrich, A., & Kanso, R. (2010). A review of EEG, ERP, and neuroimaging studies of creativity and
insight. Psychological Bulletin 136(5), 822–848.
Dixon-Salazar, T. J., & Gleeson, J. G. (2010). Genetic regulation of human brain development:
Lessons from Mendelian diseases. Annals of the New York Academy of Sciences 1214, 156–167.
Dong, S., & Clayton, D. F. (2008) Partial dissociation of molecular and behavioral measures of song
habituation in adult zebra finches. Genes, Brain and Behavior 7(7), 802–809.
Drayna, D., Manichaikul, A., de Lange, M., & Snieder, H. (2001). Genetic correlates of musical pitch
recognition in humans. Science 291(5510), 1969–1972.
Drnevich, J., Replogle, K. L., Lovell, P., Hahn, T. P., Johnson, F., Mast, T. G., … Clayton, D. F.
(2012). Impact of experience-dependent and -independent factors on gene expression in songbird
brain. Proceedings of the National Academy of Sciences 109(Suppl. 2), 17245–17252.
Duclot, F., & Kabbaj, M. (2017). The role of Early Growth Response 1 (EGR1) in brain plasticity
and neuropsychiatric disorders. Frontiers in Behavioral Neuroscience 11, 35. Retrieved from
https://doi.org/10.3389/fnbeh.2017.00035
Elbert, T., Pantev, C., Wienbruch, C., Rockstroh, B., & Taub, E. (1995). Increased cortical
representation of the fingers of the left hand in string players. Science 270(5234), 305–307.
Enard, W. (2016). The molecular basis of human brain evolution. Current Biology 26(20), R1109–
R1117.
Enard, W., Przeworski, M., Fisher, S. E., Lai, C. S., Wiebe, V., Kitano, T., … Pääbo, S. (2002).
Molecular evolution of FOXP2, a gene involved in speech and language. Nature 418(6900), 869–
872.
Fatemi, S. H., Folsom, T. D., Rooney, R. J., & Thuras, P. D. (2013). Expression of GABAA a2-, b1-
and e-receptors are altered significantly in the lateral cerebellum of subjects with schizophrenia,
major depression and bipolar disorder. Translational Psychiatry 3, e303. doi:10.1038/tp.2013.64
Gaser, C., & Schlaug, G. (2003). Brain structures differ between musicians and non-musicians.
Journal of Neuroscience 23(27), 9240–9245.
George, J. M., Jin, H., Woods, W. S., & Clayton, D. F. (1995). Characterization of a novel protein
regulated during the critical period for song learning in the zebra finch. Neuron 15, 361–372.
Gosselin, N., Peretz, I., Johnsen, E., & Adolphs, R. (2007). Amygdala damage impairs emotion
recognition from music. Neuropsychologia 45(2), 236–244.
Guo, Y. P., Sun, X., Li, C., Wang, N. Q., Chan, Y. S., & He, J. (2007). Corticothalamic
synchronization leads to c-fos expression in the auditory thalamus. Proceedings of the National
Academy of Sciences 104(28), 11802–11807.
Haesler, S., Wada, K., Nshdejan, A., Morrisey, E. E., Lints, T., Jarvis, E. D., & Scharff, C. (2004).
FoxP2 expression in avian vocal learners and non-learners. Journal of Neuroscience 24(13), 3164–
3175.
Hambrick, D. Z., Oswald, F. L., Altmann, E. M., Meinz, E. J., Gobet, F., & Campitelli, G. (2014).
Deliberate practice: Is that all it takes to become an expert? Intelligence 45, 34–45.
Haugas, M., Lilleväli, K., Hakanen, J., & Salminen, M. (2010). Gata2 is required for the
development of inner ear semicircular ducts and the surrounding perilymphatic space.
Developmental Dynamics 239(9), 2452–2469.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Hertel, N., Redies, C., & Medina, L. (2012). Cadherin expression delineates the divisions of the
postnatal and adult mouse amygdala. Journal of Comparative Neurology 520(17), 3982–4012.
Hilliard, A. T., Miller, J. E., Fraley, E. R., Horvath, S., & White, S. A. (2012). Molecular
microcircuitry underlies functional specification in a basal ganglia circuit dedicated to vocal
learning. Neuron 73(3), 537–552.
Honing, H., Ten Cate, C., Peretz, I., & Trehub, S. E. (2015). Without it no music: Cognition, biology
and evolution of musicality. Philosophical Transactions of the Royal Society of London B:
Biological Sciences 370(1664): 20140088. doi:10.1098/rstb.2014.0088
Horita, H., Kobayashi, M., Liu, W.-C., Oka, K., Jarvis, E. D., & Wada, K. (2012). Specialized motor-
driven dusp1 expression in the song systems of multiple lineages of vocal learning birds. PLoS
ONE 7, e42173.
Hu, C., Chen, W., Myers, S. J., Yuan, H., & Traynelis, S. F. (2016). Human GRIN2B variants in
neurodevelopmental disorders. Journal of Pharmacological Sciences 132(2), 115–121.
Justus, T., & Hutsler, J. J. (2005). Fundamental issues in the evolutionary psychology of music:
Assessing innateness and domain specificity. Music Perception 23(1), 1–27.
Kanduri, C., Kuusi, T., Ahvenainen, M., Philips, A. K., Lähdesmäki, H., & Järvelä, I. (2015).The
effect of music performance on the transcriptome of professional musicians. Scientific Reports 5,
9506. doi:10.1038/srep09506
Kanduri, C., Raijas, P., Ahvenainen, M., Philips, A. K., Ukkola-Vuoti, L., Lähdesmäki, H., & Järvelä,
I (2015). The effect of listening to music on human transcriptome. PeerJ 3, e830. Retrieved from
https://doi.org/10.7717/peerj.830
Karma, K. (1994). Auditory and visual temporal structuring: How important is sound to musical
thinking? Psychology of Music 22(1), 20–30.
Katori, S., Hamada, S., Noguchi, Y., Fukuda, E., Yamamoto, T., Yamamoto, H., … Yagi, T. (2009).
Protocadherin-alpha family is required for serotonergic projections to appropriately innervate
target brain areas. Journal of Neuroscience 29(29), 9137–9147.
Katz, E., Elgoyhen, A. B., Gómez-Casati, M. E., Knipper, M., Vetter, D. E., Fuchs, P. A., &
Glowatski, E. (2004). Developmental regulation of nicotinic synapses on cochlear inner hair cells.
Journal of Neuroscience 24(36), 7814–7820.
Kauppi, K., Nilsson, L.-G., Adolfsson, R., Eriksson, E. & Nyberg, L. (2011). KIBRA polymorphism
is related to enhanced memory and elevated hippocampal processing. Journal of Neuroscience 31,
14218–14222.
Kim, H. S., & Sasaki, J. Y. (2014). Cultural neuroscience: Biology of the mind in cultural contexts.
Annual Review of Psychology 65, 487–514.
Kitayama, S., King, A., Hsu, M., Liberzon, I., & Yoon, C. (2016). Dopamine-system genes and
cultural acquisition: The norm sensitivity hypothesis. Current Opinion in Psychology 8, 167–174.
Koelsch, S. (2010). Towards a neural basis of music-evoked emotions. Trends in Cognitive Sciences
14(3), 131–137.
Kyaga, S., Landén, M., Boman, M., Hultman, C. M., Långström, N., & Lichtenstein, P. (2013).
Mental illness, suicide and creativity: 40-year prospective total population study. Journal of
Psychiatric Research 47(1), 83–90.
Kyaga, S., Lichtenstein, P., Boman, M., Hultman, C., Långström, N., & Landén, M. (2011).
Creativity and mental disorder: Family study of 300,000 people with severe mental disorder.
British Journal of Psychiatry 199(5), 373–379.
Lahti, L., Achim, K., & Partanen, J. (2013). Molecular regulation of GABAergic neuron
differentiation and diversity in the developing midbrain. Acta Physiologica (Oxford) 207(4), 616–
627.
Lai, C. S., Fisher, S. E., Hurst, J. A., Vargha-Khadem, F., & Monaco, A. P. (2001). A forkhead-
domain gene is mutated in a severe speech and language disorder. Nature 413 (6855), 519–523.
Lander, E. S. (2011). Initial impact of the sequencing of the human genome. Nature 470, 187–197.
Lee, S. H., Ripke, S., Neale, B. M., & Cross-Disorder Group of the Psychiatric Genomics
Consortium (2013). Genetic relationship between five psychiatric disorders estimated from
genome-wide SNPs. Nature Genetics 45(9), 984–994.
Limb, C. J., & Braun, A. R. (2008). Neural substrates of spontaneous musical performance: An
FMRI study of jazz improvisation. PLoS ONE 3, e1679.
Lin, J., Yan, X., Wang, C., Guo, Z., Rolfs, A., & Luo, J. (2012). Anatomical expression patterns of
delta-protocadherins in developing chicken cochlea. Journal of Anatomy 221(6), 598–608.
Lindor, N. M., Thibodeau, S., & Burke, W. (2017). Whole-genome sequencing in healthy people.
Mayo Clinic Proceedings 92(1), 159–172.
Lipkind, D., & Tchernichovski, O. (2011). Colloquium paper: Quantification of developmental
birdsong learning from the subsyllabic scale to cultural evolution. Proceedings of the National
Academy of Sciences 108(Suppl. 3), 15572–15579.
Liu, S., Chow, H. M., Xu, Y., Erkkinen, M. G., Swett, K. E., Eagle, M. W., … Braun, A. R. (2012).
Neural correlates of lyrical improvisation: An fMRI study of freestyle rap. Scientific Reports 2,
834. doi:10.1038/srep00834
Liu, X., Kanduri, C., Oikkonen, J., Karma, K., Raijas, P., Ukkola-Vuoti, L., … Järvelä, I. (2016).
Detecting signatures of positive selection associated with musical aptitude in the human genome.
Scientific Reports 6, 21198. doi:10.1038/srep21198
London, S. E., & Clayton, D. F. (2008). Functional identification of sensory mechanisms required for
developmental song learning. Nature Neuroscience 11(5), 579–586.
Martínez-Molina, N., Mas-Herrero, E., Rodríguez-Fornells, A., Zatorre, R. J., & Marco-Pallarés, J.
(2016). Neural correlates of specific musical anhedonia. Proceedings of the National Academy of
Sciences 113(46), E7337–E7345.
Metz, M., Gassmann, M., Fakler, B., Schaeren-Wiemers, N., & Bettler, B. (2011). Distribution of the
auxiliary GABAB receptor subunits KCTD8, 12, 12b, and 16 in the mouse brain. Journal of
Comparative Neurology 519(8), 1435–1454.
Montealegre-Z, F., Jonsson, T., Robson-Brown, K. A., Postles, M., & Robert, D. (2012). Convergent
evolution between insect and mammalian audition. Science 338(6109), 968–971.
Mosing, M. A., Madison, G., Pedersen, N. L., Kuja-Halkola, R., & Ullén, F. (2014). Practice does not
make perfect: No causal effect of music practice on music ability. Psychological Science 25(9),
1795–1803.
Mosing, M. A., Madison, G., Pedersen, N. L., & Ullén, F. (2016). Investigating cognitive transfer
within the framework of music practice: Genetic pleiotropy rather than causality. Developmental
Science 19(3), 504–512.
Murugan, M., Harward, S., Scharff, C., & Mooney, R. (2013). Diminished FoxP2 levels affect
dopaminergic modulation of corticostriatal signaling important to song variability. Neuron 80(6),
1464–1476.
Nakahara, H., Masuko, T., Kinoshita, H., Francis, P. R., & Furuya, S. (2011). Performing music can
induce greater modulation of emotion-related psychophysiological responses than listening to
music. International Journal of Psychophysiology 81(3), 152–158.
Nordeen, E. J., Holtzman, D. A., & Nordeen, K. W. (2009). Increased Fos expression among
midbrain dopaminergic cell groups during birdsong tutoring. European Journal of Neuroscience
30(4), 662–670.
Oikkonen, J., Huang, Y., Onkamo, P., Ukkola-Vuoti, L., Raijas, P., Karma, K., … Järvelä, I. (2015).
A genome-wide linkage and association study of musical aptitude identifies loci containing genes
related to inner ear development and neurocognitive functions. Molecular Psychiatry 20(2), 275–
282.
Oikkonen, J., & Järvelä, I. (2014). Genomics approaches to study musical aptitude. Bioessays 36(11),
1102–1108.
Oikkonen, J., Onkamo, P., Järvelä, I., & Kanduri, C. (2016). Convergent evidence for the molecular
basis of musical traits. Scientific Reports 6, 39707. doi:10.1038/srep39707
Ousdal, O. T., Anand Brown, A., Jensen, J., Nakstad, P. H., Melle, I., Agartz, I., … Andreassen, O.
A. (2012). Associations between variants near a monoaminergic pathways gene (PHOX2B) and
amygdala reactivity: A genome-wide functional imaging study. Twin Research and Human
Genetics 15(3), 273–285.
Park, H., Lee, S., Kim, H. J., & Ju, Y. S. (2012). Comprehensive genomic analyses associate UGT8
variants with musical ability in a Mongolian population. Journal of Medical Genetics 49(12), 747–
752.
Parker, J., Tsagkogeorga, G., Cotton, J. A., Liu, Y., Provero, P., Stupka, E., & Rossiter, S. J. (2013).
Genome-wide signatures of convergent evolution in echolocating mammals. Nature 502(7470),
228–231.
Penhune, V., & de Villers-Sidani, E. (2014). Time for new thinking about sensitive periods. Frontiers
in Systems Neuroscience 8, 55. Retrieved from https://doi.org/10.3389/fnsys.2014.00055
Perani, D., Saccuman, M. C., Scifo, P., Spada, D., Andreolli, G., Rovelli, R., … Koelsch, S. (2010).
Functional specializations for music processing in the human newborn brain. Proceedings of the
National Academy of Sciences 107(10), 4758–4763.
Peretz, I., Cummings, S., & Dube, M. P. (2007). The genetics of congenital amusia (tone deafness):
A family-aggregation study. American Journal of Human Genetics 81(3), 582–588.
Petrucci, S., Ginevrino, M., & Valente, E. M. (2016). Phenotypic spectrum of alpha-synuclein
mutations: New insights from patients and cellular models. Parkinsonism & Related Disorders
22(Suppl. 1), S16–S20.
Peuralinna, T., Oinas, M., Polvikoski, T., Paetau, A., Sulkava, R., Niinistö, L., … Myllykangas, L.
(2008). Neurofibrillary tau pathology modulated by genetic variation of alpha-synuclein. Annals of
Neurology 64(3), 348–352.
Pfenning, A. R., Hara, E., Whitney, O., Rivas, M. V., Wang, R., Roulhac, P. L., … Jarvis, E. D.
(2014). Convergent transcriptional specializations in the brains of humans and song-learning birds.
Science 346(6215), 1256846–1256846.
Power, R. A, Steinberg, S., Bjornsdottir, G., Rietveld, C. A., Abdellaoui, A., Nivard, M. M., …
Steffanson, K. (2015). Polygenic risk scores for schizophrenia and bipolar disorder predict
creativity. Nature Neuroscience 18(7), 953–955.
Pulli, K., Karma, K., Norio, R., Sistonen, P., Göring, H. H., & Järvelä, I. (2008). Genome-wide
linkage scan for loci of musical aptitude in Finnish families: Evidence for a major locus at 4q22.
Journal of Medical Genetics 45(7), 451–456.
Qian, W., Deng, L., Lu, D., & Xu, S. (2013). Genome-wide landscapes of human local adaptation in
Asia. PLoS ONE 8, e54224.
Rothenberg, D., Roeske, T. C., Voss, H. U., Naguib, M., & Tchernichovski, O. (2014). Investigation
of musicality in birdsong. Hearing Research 308, 71–83.
Sabeti, P. C., Schaffner, S. F., Fry, B., Lohmueller, J., Varilly, P., Shamovsky, O., … Lander, E. S.
(2006). Positive natural selection in the human lineage. Science 312(5780), 1614–1620.
Salimpoor, V. N., Benovoy, M., Larcher, K., Dagher, A., & Zatorre, R. J. (2011). Anatomically
distinct dopamine release during anticipation and experience of peak emotion to music. Nature
Neuroscience 14(2), 257–262.
Salimpoor, V. N., Benovoy, M., Longo, G., Cooperstock, J. R., & Zatorre, R. J. (2009). The
rewarding aspects of music listening are related to degree of emotional arousal. PLoS ONE 4,
e7487.
Salimpoor, V. N., Zald, D. H., Zatorre, R. J., Dagher, A., & McIntosh, A. R. (2015). Predictions and
the brain: How musical sounds become rewarding. Trends in Cognitive Sciences 19(2), 86–91.
Scherzer, C. R., Grass, J. A., Liao, Z., Pepivani, I., Zheng, B., Eklund, A. C., … Schlossmacher, M.
G. (2008). GATA transcription factors directly regulate the Parkinson’s disease-linked gene alpha-
synuclein. Proceedings of the National Academy of Sciences 105(31), 10907–10912.
Schizophrenia Working Group of the Psychiatric Genomics Consortium (2014). Biological insights
from 108 schizophrenia-associated genetic loci. Nature 511(7510), 421–427.
Schneider, P., Scherg, M., Dosch, H. G., Specht, H. J., Gutschalk, A., & Rupp, A. (2002).
Morphology of Heschl’s gyrus reflects enhanced activation in the auditory cortex of musicians.
Nature Neuroscience 5(7), 688–694.
Seashore, C., Lewis, D., & Saetveit, J. (1960). Seashore measures of musical talents. New York:
Psychological Corporation.
Sotomayor, M., Weihofen, W. A., Gaudet, R., & Corey, D. P. (2012). Structure of a force-conveying
cadherin bond essential for inner-ear mechanotransduction. Nature 492(7427), 128–132.
Theusch, E., Basu, A., & Gitschier, J. (2009). Genome-wide study of families with absolute pitch
reveals linkage to 8q24.21 and locus heterogeneity. American Journal of Human Genetics 85(1),
112–119.
Ukkola-Vuoti, L., Kanduri, C., Oikkonen, J., Buck, G., Blancher, C., Raijas, P., … Järvelä, I. (2013).
Genome-wide copy number variation analysis in extended families and unrelated individuals
characterized for musical aptitude and creativity in music. PLoS ONE 8, e56356.
Vernes, S. C., Spiteri, E., Nicod, J., Groszer, M., Taylor, J. M., Davies, K. E., … Fisher, S. E. (2007).
High-throughput analysis of promoter occupancy reveals direct neural targets of FOXP2, a gene
mutated in speech and language disorders. American Journal of Human Genetics 81(6), 1232–
1250.
White, E. J., Hutka, S. A., Williams, L. J., & Moreno, S. (2013). Learning, neural plasticity and
sensitive periods: Implications for language acquisition, music training and transfer across the
lifespan. Frontiers in Systems Neuroscience 7, 90. Retrieved from
https://doi.org/10.3389/fnsys.2013.00090
Whitney, O., Pfenning, A. R., Howard, J. T., Blatti, C. A., Liu, F., Ward, J. M., … Jarvis, E. D.
(2014). Core and region-enriched networks of behaviorally regulated genes and the singing
genome. Science 346(6215), 1256780.
Wood, W. E., Olson, C. R., Lovell, P. V., & Mello, C. V. (2008). Dietary retinoic acid affects song
maturation and gene expression in the song system of the zebra finch. Developmental
Neurobiology 68(10), 1213–1224.
Zhang, G., Li, C., Li, Q., Li, B., Larkin, D. M., Lee, C., … Wang, J. (2014). Comparative genomics
reveals insights into avian genome evolution and adaptation. Science 346(6215), 1311–1320.
CHAPT E R 19
BRAIN RESEARCH IN
MUSIC PERFORMANCE
E C K A RT A LT E N MÜ L L E R, S H I N I C H I F U R U YA, D A N I E L S .
S C H O L Z, A N D C H R I S TO S I . I O A N N O U
P M D
B P
T E M T
B F
T E M T
B S
FIGURE 2. Summary of the results of the study on pianists by Vaquero et al. (2016). Explanations
are given in the figure and text.
Reprinted from NeuroImage, 126, Lucía Vaquero, Karl Hartmann, Pablo Ripollés, Nuria
Rojo, Joanna Sierpowska, Clément François, Estela Càmara, Floris Tijmen van Vugt,
Bahram Mohammadi, Amir Samii, Thomas F. Münte, Antoni Rodríguez-Fornells, and Eckart
Altenmüller, Structural neuroplasticity in expert pianists depends on the age of musical
training onset, pp. 106–119, Copyright © 2015 Elsevier Ltd. All rights reserved.
D -E : M ’ D
S M P
B C A L
S C
B P P
R E P
M
A
This work was in part supported by the DYSTRACT grant, awarded from
the BMBF to authors EA and DS.
R
Altenmüller, E., Baur, V., Hofmann, A., Lim, V. K., & Jabusch, H. C. (2012). Musician’s cramp as
manifestation of maladaptive brain plasticity: Arguments from instrumental differences. Annals of
the New York Academy of Sciences 1252, 259–265.
Altenmüller, E., & Furuya, S. (2015). Planning and performance. In S. Hallam, I. Cross, & M. Thaut
(Eds.), The Oxford handbook of music psychology (2nd ed., pp. 529–546). Oxford: Oxford
University Press.
Altenmüller, E., Ioannou, C. I., & Lee, A. (2015). Apollo’s curse: Neurological causes of motor
impairments in musicians. Progress in Brain Research 217, 89–106.
Altenmüller, E., Münte, T. H., & Gerloff, C. (2004). Neurocognitive functions and the EEG. In E.
Niedermeyer & F. Lopes da Silva (eds.), Electroencephalography (5th ed., pp. 661–682).
Baltimore, MD: Lippincott Williams.
Amunts, K., Schlaug, G., Jäncke, L., Steinmetz, H., Schleicher, A., Dabringhaus, A., & Zilles, K.
(1997). Motor cortex and hand motor skills: Structural compliance in the human brain. Human
Brain Mapping 5(3), 206–215.
Amunts, K., & Zilles, K. (2015). Architectonic mapping of the human brain beyond Brodmann.
Neuron 88(6), 1086–1107.
Auerbach, S. (1906–1913). Zur Lokalisation des musicalischen Talentes im Gehirn und am Schädel.
Archives of Anatomy and Physiology (1906, 197–230; 1908, 31–38; 1911, 1–10; 1913 (Suppl.),
89–96).
Baharloo, S., Johnston, P. A., Service, S. K., Gitschier, J., & Freimer, N. B. (1998). Absolute pitch:
An approach for identification of genetic and nongenetic components. American Journal of
Human Genetics 62(2), 224–231.
Baharloo, S., Service, S. K., Risch, N., Gitschier, J., & Freimer, N. B. (2000). Familial aggregation of
absolute pitch. American Journal of Human Genetics 67(3), 755–758.
Bandettini, P. A. (2009). What is new in neuroimaging methods? Annals of the New York Academy of
Sciences 1156, 260–293.
Bangert, M., & Altenmüller, E. (2003). Mapping perception to action in piano practice: A
longitudinal DC-EEG-study. BMC Neuroscience 4, 26–36.
Bangert, M., Peschel, T., Rotte, M., Drescher, D., Hinrichs, H., Schlaug, G., … Altenmüller, E.
(2006). Shared networks for auditory and motor processing in professional pianists: Evidence from
fMRI conjunction. NeuroImage 30(3), 917–926.
Bangert, M., & Schlaug, G. (2006). Specialization of the specialized in features of external brain
morphology. European Journal of Neuroscience 24(6), 1832–1834.
Battistella, G., Termsarasab, P., Ramdhani, R. A., Fuertinger, S., & Simonyan, K. (2017). Isolated
focal dystonia as a disorder of large-scale functional networks. Cerebral Cortex 27(2), 1203–1215.
Baur, V., Jabusch, H. C., & Altenmüller, E. (2011). Behavioral factors influence the phenotype of
musician’s dystonia. Movement Disorders 26(9), 1780–1781.
Bengtsson, S. L., Nagy, Z., Skare, S., Forsman, L., Forssberg, H., & Ullén, F. (2005). Extensive piano
practicing has regionally specific effects on white matter development. Nature Neuroscience 8(9),
1148–1150.
Bermudez, P., Lerch, J. P., Evans, A. C., & Zatorre, R. J. (2008). Neuroanatomical correlates of
musicianship as revealed by cortical thickness and voxel-based morphometry. Cerebral Cortex
19(7), 1583–1596.
Brown, R. M., Chen, J. L., Hollinger, A., Palmer, C., Penhune, V., & Zatorre, R. J. (2013). Repetition
suppression in auditory-motor regions to pitch and temporal structure in music. Journal of
Cognitive Neuroscience 25(2), 313–328.
Brown, R. M., Penhune, V. B., & Zatorre, R. (2015). Expert music performance: Cognitive, neural,
and developmental bases. Progress in Brain Research 217, 57–86.
Byl, N. N., Merzenich, M. M., & Jenkins, W. M. (1996). A primate genesis model of focal dystonia
and repetitive strain injury: I. Learning-induced dedifferentiation of the representation of the hand
in the primary somatosensory cortex in adult monkeys. Neurology 47(2), 508–520.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008a). Moving on time: Brain network for auditory-
motor synchronization is modulated by rhythm complexity and musical training. Journal of
Cognitive Neuroscience 20(2), 226–239.
Chen, J. L., Penhune, V. B., & Zatorre, R. J. (2008b). Listening to musical rhythms recruits motor
regions of the brain. Cerebral Cortex 18(12), 2844–2854.
Corrigall, K. A., Schellenberg, E. G., & Misura, N. M. (2013). Music training, cognition, and
personality. Frontiers in Psychology 4, 222. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00222
De Manzano, Ö., & Ullén, F. (2012). Activation and connectivity patterns of the presupplementary
and dorsal premotor areas during free improvisation of melodies and rhythms. NeuroImage 63(1),
272–280.
Elbert, T., Candia, V., Altenmüller, E., Rau, H., Rockstroh, B., Pantev, C., & Taub, E. (1998).
Alteration of digital representations in somatosensory cortex in focal hand dystonia. Neuroreport
9(16), 3571–3575.
Elbert, T., Pantev, C., Wienbruch, C., Rockstroh, B., & Taub, E. (1995). Increased cortical
representation of the fingers of the left hand in string players. Science 270(5234), 305–307.
Ellis, R. J., Norton, A., Overy, K., Winner, E., Alsop, D., & Schlaug, G. (2012). Differentiating
maturational and training influences on fMRI activation during music processing. NeuroImage
60(3), 1902–1912.
Ericsson, K. A., Krampe, R. T., & Tesch-Römer, C. (1993). The role of deliberate practice in the
acquisition of expert performance. Psychological Review 100(3), 363–406.
Ericsson, K. A., & Lehmann, A. C. (1996). Expert and exceptional performance: Evidence of
maximal adaptation to task constraints. Annual Review of Psychology 47, 273–305.
Foster, N. E., & Zatorre, R. J. (2010). A role for the intraparietal sulcus in transforming musical pitch
information. Cerebral Cortex 20(6), 1350–1359.
Furukawa, Y., Uehara, K., & Furuya, S. (2017). Expertise-dependent motor somatotopy of music
perception. Neuroscience Letters 650, 97–102.
Furuya, S., & Altenmüller, E. (2013). Finger-specific loss of independent control of movements in
musicians with focal dystonia. Neuroscience 247, 152–163.
Furuya, S., Klaus, M., Nitsche, M. A., Paulus, W., & Altenmüller, E. (2014). Ceiling effects prevent
further improvement of transcranial stimulation in skilled musicians. Journal of Neuroscience
34(41), 13834–13839.
Furuya, S., Nitsche, M. A., Paulus, W., & Altenmüller, E. (2014). Surmounting retraining limits in
musicians’ dystonia by transcranial stimulation. Annals of Neurology 75(5), 700–707.
Furuya, S., Oku, T., Miyazaki, F., & Kinoshita, H. (2015). Secrets of virtuoso: Neuromuscular
attributes of motor virtuosity in expert musicians. Scientific Reports 5, 15750.
Gaab, N., Gaser, C., & Schlaug, G. (2006). Improvement-related functional plasticity following pitch
memory training. NeuroImage 31(1), 255–263.
Gärtner, H., Minnerop, M., Pieperhoff, P., Schleicher, A., Zilles, K., Altenmüller, E., & Amunts, K.
(2013). Brain morphometry shows effects of long-term musical practice in middle-aged keyboard
players. Frontiers in Psychology 4, 636. Retrieved from https://doi.org/10.3389/fpsyg.2013.00636
Gaser, C., & Schlaug, G. (2003). Brain structures differ between musicians and non-musicians.
Journal of Neuroscience 23(27), 9240–9245.
Gentner, R., & Classen, J. (2006). Modular organization of finger movements by the human central
nervous system. Neuron 52(4), 731–742.
Gentner, R., Gorges, S., Weise, D., aufm Kampe, K., Buttmann, M., & Classen, J. (2010). Encoding
of motor skill in the corticomuscular system of musicians. Current Biology 20(20), 1869–1874.
Gingras, B., Honing, H., Peretz, I., Trainor, L. J., & Fisher, S. E. (2015). Defining the biological
bases of individual differences in musicality. Philososophical Transactions of the Royal Society of
London B: Biological Sciences 370(1664), 20140092.
Granert, O., Peller, M., Jabusch, H. C., Altenmüller, E., & Siebner, H. R. (2011). Sensorimotor skills
and focal dystonia are linked to putaminal grey-matter volume in pianists. Journal of Neurology,
Neurosurgery, and Psychiatry 82(11), 1225–1231.
Gregersen, P. K., Kowalsky, E., Kohn, N., & Marvin, E. W. (2001). Early childhood music education
and predisposition to absolute pitch: Teasing apart genes and environment. American Journal of
Medical Genetics 98(3), 280–282.
Haber, J. (2003). The primate basal ganglia: Parallel and integrative networks. Journal of Chemical
Neuroanatomy 26(4), 317–330.
Halwani, G. F., Loui, P., Rüber, T., & Schlaug, G. (2011). Effects of practice and experience on the
arcuate fasciculus: Comparing singers, instrumentalists, and non-musicians. Frontiers in
Psychology 2, 156. Retrieved from https://doi.org/10.3389/fpsyg.2011.00156
Haslinger, B., Altenmüller, E., Castrop, F., Zimmer, C., & Dresel, C. (2010). Sensorimotor
overactivity as a pathophysiologic trait of embouchure dystonia. Neurology 74(22), 1790–1797.
Haslinger, B., Erhard, P., Altenmüller, E., Schroeder, U., Boecker, H., & Ceballos-Baumann, A. O.
(2005). Transmodal sensorimotor networks during action observation in professional pianists.
Journal of Cognitive Neuroscience 17(2), 282–293.
Herholz, S. C., Coffey, E. B., Pantev, C., & Zatorre, R. J. (2016). Dissociation of neural networks for
predisposition and for training-related plasticity in auditory-motor learning. Cerebral Cortex 26(7),
3125–3134.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Herrojo-Ruiz, M., Jabusch, H. C., & Altenmüller, E. (2009). Detecting wrong notes in advance:
Neuronal correlates of error monitoring in pianists. Cerebral Cortex 19(11), 2625–2639.
Herrojo-Ruiz, M., Senghaas, P., Grossbach, M., Jabusch, H. C., Bangert, M., Hummel, F., …
Altenmüller E. (2009). Defective inhibition and inter-regional phase synchronization in pianists
with musician’s dystonia (MD): An EEG study. Human Brain Mapping 30(8), 2689–2700.
Hikosaka, O., & Nakamura, K. (2002). Central mechanisms of motor skill learning. Current Opinion
in Neurobiology 12(2), 217–222.
Hirata, Y., Kuriki, S., & Pantev, C. (1999). Musicians with absolute pitch show distinct neural
activities in the auditory cortex. Neuroreport 10(5), 999–1002.
Hosoda, M., & Furuya, S. (2016). Shared somatosensory and motor functions in musicians. Scientific
Reports 6, 37632.
Hyde, K. L., Lerch, J., Norton, A., Forgeard, M., Winner, E., Evans, A. C., & Schlaug, G. (2009).
Musical training shapes structural brain development. Journal of Neuroscience 29(10), 3019–3025.
Imfeld, A., Oechslin, M. S., Meyer, M., Loenneker, T., & Jäncke, L. (2009). White matter plasticity
in the corticospinal tract of musicians: A diffusion tensor imaging study. NeuroImage 46(3), 600–
607.
Ioannou, C. I., & Altenmüller, E. (2014). Psychological characteristics in musician’s dystonia: A new
diagnostic classification. Neuropsychology 61, 80–88.
Ioannou, C. I., Furuya, S., & Altenmüller, E. (2016). The impact of stress on motor performance in
skilled musicians suffering from focal dystonia: Physiological and psychological characteristics.
Neuropsychologia 85, 226–236.
Jabusch, H. C., & Altenmüller, E. (2004). Anxiety as an aggravating factor during onset of focal
dystonia in musicians. Medical Problems of Performing Artists 19(2), 75–81.
Jabusch, H. C., Müller, S. V., & Altenmüller, E. (2004). High levels of perfectionism and anxiety in
musicians with focal dystonia. Movement Disorders 19(10), 990–991.
James, C. E., Oechslin, M. S., Van De Ville, D., Hauert, C. A., Descloux, C., & Lazeyras, F. (2013).
Musical training intensity yields opposite effects on grey matter density in cognitive versus
sensorimotor networks. Brain Structure and Function 219(1), 53–66.
Kleber, B., Veit, R., Birbaumer, N., Gruzelier, J., & Lotze, M. (2010). The brain of opera singers:
Experience-dependent changes in functional activation. Cerebral Cortex 20(5), 1144–1152.
Koelsch S. (2011). Toward a neural basis of music perception: A review and updated model.
Frontiers in Psychology 2, 110. Retrieved from https://doi.org/10.3389/fpsyg.2011.00110
Krampe, R., & Ericsson, K. (1996). Maintaining excellence: Deliberate practice and elite
performance in young and older pianists. Journal of Experimental Psychology: General 125(4),
331–359.
Kraus, N., McGee, T. J., & Koch, D. B. (1998). Speech sound representation, perception and
plasticity: A neurophysiologic perspective. Audiology & Neurotology 3, 168–182.
Kuhtz-Buschbeck, J. P., Mahnkopf, C., Holzknecht, C., Siebner, H., Ulmer, S., & Jansen, O. (2003).
Effector-independent representations of simple and complex imagined finger movements: A
combined fMRI and TMS study. European Journal of Neuroscience 18(12), 3375–3387.
Lahav, A., Saltzman, E., & Schlaug, G. (2007). Action representation of sound: Audiomotor
recognition network while listening to newly acquired actions. Journal of Neuroscience 27(2),
308–314.
Lee, A., Heiß, P., Eich, C., Ioannou, I. C., & Altenmüller, E. (2018). Phenomenology, risk-factors and
treatment outcome in 369 musicians with focal dystonia. Submitted to Journal of Clinical
Movement Disorders.
Lin, P. T., & Hallett, M. (2009). The pathophysiology of focal hand dystonia. Journal of Hand
Therapy 22(2), 109–114.
Loui, P., Li, H. C., Hohmann, A., & Schlaug, G. (2011). Enhanced cortical connectivity in absolute
pitch musicians: A model for local hyperconnectivity. Journal of Cognitive Neuroscience 23(4),
1015–1026.
Meinz, E. J. (2000). Experience-based attenuation of age-related differences in music cognition tasks.
Psychology and Aging 15(2), 297–312.
Miyazaki, K. (1988). Musical pitch identification by absolute pitch possessors. Perception &
Psychophysics 44(6), 501–512.
Mosing, M. A., Madison, G., Pedersen, N. L., Kuja-Halkola, R., & Ullén, F. (2014). Practice does not
make perfect: No causal effect of music practice on music ability Psychological Science 25(9),
17965–1803.
Münte, T. F., Kohlmetz, C., Nager, W., & Altenmüller, E. (2001). Neuroperception: Superior auditory
spatial tuning in professional conductors. Nature 409(6820), 580.
Münte, T. F., Nager, W., Beiss, T., Schroeder, C., & Altenmüller, E. (2003). Specialization of the
specialized: Electrophysiological investigations in professional musicians. Annals of the New York
Academy of Sciences 999, 131–139.
Oechslin, M. S., Imfeld, A., Loenneker, T., Meyer, M., & Jäncke, L. (2010). The plasticity of the
superior longitudinal fasciculus as a function of musical expertise: A diffusion tensor imaging
study. Frontiers in Human Neuroscience 3, 76. Retrieved from
https://doi.org/10.3389/neuro.09.076.2009
Oztürk, A. H., Tascioglu, B., Aktekin, M., Kurtoglu, Z., & Erden, I. (2002). Morphometric
comparison of the human corpus callosum in professional musicians and non-musicians by using
in vivo magnetic resonance imaging. Journal of Neuroradiology 29(1), 29–34.
Pantev, C., Oostenveld, R., Engelien, A., Ross, B., Roberts, L. E., & Hoke, M. (1998). Increased
auditory cortical representation in musicians. Nature 392, 811–814.
Pascual-Leone, A. (2001). The brain that plays music and is changed by it. Annals of the New York
Academy of Sciences 930, 315–329.
Pascual-Leone, A., Grafman, J., & Hallett, M. (1994). Modulation of cortical motor output maps
during development of implicit and explicit knowledge. Science 263(5151), 1287–1289.
Paulig, J., Jabusch, H. C., Großbach, M., Boullet, L., & Altenmüller, E. (2014). Sensory trick
phenomenon improves motor control in pianists with dystonia: Prognostic value of glove-effect.
Frontiers in Psychology 5, 1012. Retrieved from https://doi.org/10.3389/fpsyg.2014.01012
Ragert, P., Schmidt, A., Altenmüller, E., & Dinse, H. R. (2003). Superior tactile performance and
learning in professional pianists: Evidence for meta-plasticity in musicians. European Journal of
Neuroscience 19(2), 473–478.
Ramnani, N. (2014). Automatic and controlled processing in the corticocerebellar system. Progress
in Brain Research 210, 255–285.
Ridding, M. C., Brouwer, B., & Nordstrom, M. A. (2000). Reduced interhemispheric inhibition in
musicians. Experimental Brain Research 133(2), 249–253.
Rizzolatti, G., Fadiga, L., Gallese, V., & Fogassi, L. (1996). Premotor cortex and the recognition of
motor actions. Brain Research: Cognitive Brain Research 3(2), 131–141.
Roland, P. E., & Zilles, J. (1996). Functions and structures of the motor cortices in humans. Current
Opinion in Neurobiology 6(6), 773–781.
Rosenkranz, K., Altenmüller, E., Siggelkow, S., & Dengler, R. (2000). Alteration of sensorimotor
integration in musician’s cramp: Impaired focussing of proprioception. Clinical Neurophysiology
111(11), 2036–2041.
Rosenkranz, K., Williamon, A., Butler, K., Cordivari, C., Lees, A. J., & Rothwell, J. C. (2005).
Pathophysiological differences between musician’s dystonia and writer’s cramp. Brain 128(4),
918–931.
Salimpoor, V. N., van den Bosch, I., Kovacevic, N., McIntosh, A. R., Dagher, A., & Zatorre, R. J.
(2013). Interactions between the nucleus accumbens and auditory cortices predict music reward
value. Science 340(6129), 216–219.
Schlaug, G. (2001). The brain of musicians: A model for functional and structural plasticity. Annals
of the New York Academy of Sciences 930, 281–299.
Schlaug, G. (2015). Musicians and music making as a model for the study of brain plasticity.
Progress in Brain Research 217, 37–55.
Schlaug, G., Jäncke, L., Huang, Y., Staiger, J. F., & Steinmetz, H. (1995). Increased corpus callusum
size in musicians. Neuropsychologia 33(8), 1047–1055.
Schlaug, G., Jäncke, L., Huang, Y., & Steinmetz, H. (1995). In vivo evidence of structural brain
asymmetry in musicians. Science 267(5198), 699–701.
Schmidt, A., Jabusch, H. C., Altenmüller, E., Hagenah, J., Brüggemann, N., Lohmann, K., … Klein,
C. (2009). Etiology of musicians dystonia: Familial or environmental? Neurology 72(14), 1248–
1254.
Schneider, P., Scherg, M., Dosch, H. G., Specht, H. J., Gutschalk, A., & Rupp, A. (2002).
Morphology of Heschl’s gyrus reflects enhanced activation in the auditory cortex of musicians.
Nature Neuroscience 5(7), 688–694.
Schneider, P., Sluming, V., Roberts, N., Scherg, M., Goebel, R., Specht, H. J., … Rupp, A. (2005).
Structural and functional asymmetry of lateral Heschl’s gyrus reflects pitch perception preference.
Nature Neuroscience 8(9), 1241–1247.
Seger, C. A. (2006). The basal ganglia in human learning Neuroscientist 12(4), 285–290.
Sergeant, D. (1968). Experimental investigation of absolute pitch. Journal of Research in Music
Education 17(1), 135–143.
Skoe, E., & Kraus, N. (2013). Musical training heightens auditory brainstem function during
sensitive periods in development. Frontiers in Psychology 4, 622. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00622
Sommer, M., Ruge, D., Tergau, F., Beuche, W., Altenmüller, E., & Paulus, W. (2002). Spatial
distribution of intracortical inhibition and facilitation in focal dystonia Movement Disorders 17,
1017–1025.
Steele, C. J., Bailey, J. A., Zatorre, R. J., & Penhune, V. B. (2013). Early musical training and white-
matter plasticity in the corpus callosum: Evidence for a sensitive period. Journal of Neuroscience
33(3), 1282–1290.
Stetson, C., & Anderson, C. A. (2015). Early planning in frontal and parietal cortex in a simplified
task. Journal of Neurophysiology 113(10), 3915–3922.
Stewart, L., Henson, R., Kampe, K., Walsch, V., Turner, R., & Frith, U. (2003). Brain changes after
learning to read and play music. NeuroImage 20(1), 71–83.
Strübing, F., Ruiz, M. H., Jabusch, H. C., & Altenmüller, E. (2012). Error monitoring is altered in
musician’s dystonia: Evidence from ERP-based studies. Annals of the New York Academy of
Sciences 1252, 192–199.
Taubert, M., Villringer, A., & Ragert, P. (2012). Learning-related gray and white matter changes in
humans: An update. Neuroscientist 18(4), 320–325.
Termsarasab, P., Ramdhani, R. A., Battistella, G., Rubien-Thomas, E., Choy, M., Farwell, I. M., &
Simonyan, K. (2015). Neural correlates of abnormal sensory discrimination in laryngeal dystonia.
Neuroimage Clinical 10, 18–26.
van Vugt, F. T., Boullet, L., Jabusch, H. C., & Altenmüller, E. (2014). Musician’s dystonia in
pianists: Long-term evaluation of retraining and other therapies. Parkinsonism & Related
Disorders 20(1), 8–12.
Vaquero, L., Hartmann, K., Ripolles, P., Rojo, N., Sierpowska, J., François, C., … Altenmüller, E.
(2016). Structural neuroplasticity in expert pianists depends on the age of musical training onset.
NeuroImage 126, 106–119.
Vollmann, H., Ragert, P., Conde, V., Villringer, A., Classen, J., Witte, O. W., & Steel, J. (2014).
Instrument specific use-dependent plasticity shapes the anatomical properties of the corpus
callosum: A comparison between musicians and non-musicians. Frontiers in Behavioral
Neuroscience 8, 245. Retrieved from https://doi.org/10.3389/fnbeh.2014.00245
Warren, J. E., Wise, R. J., & Warren, J. D. (2005). Sounds do-able: Auditory-motor transformations
and the posterior temporal plane. Trends in Neuroscience 28(12), 636–643.
Wong, P. C. M., Skoe, E., Russo, N. M., Dees, T., & Kraus, N. (2007). Musical experience shapes
human brainstem encoding of linguistic pitch patterns. Nature Neuroscience 10(4), 420–422.
Zatorre, R. J. (2001). Neural specializations for tonal processing. Annals of the New York Academy of
Sciences 930, 193–210.
Zatorre, R. J., Chen, J. L., & Penhune, V. B. (2007). When the brain plays music: Auditory–motor
interactions in music perception and production. Nature Reviews Neuroscience 8(7), 547–558.
Zatorre, R. J., & Salimpoor, V. N. (2013). From perception to pleasure: Music and its neural
substrates Proceedings of the National Academy of Sciences 110(2), 10430–10437.
CHAPT E R 20
BRAIN RESEARCH IN
M U S I C I M P R O V I S AT I O N
MI C H A E L G . E R K K I N E N A N D A A R O N L . B E R K O WI T Z
F M R
I ( MRI)
Results
Behavioral: Subjects accurately reproduced their improvisations during the
“Reproduce” condition, and a comparison between “Improv” and
“Reproduce” showed no differences in musical or motor output. In the
comparison of “FreeImprov” versus “Improv,” the subjects played more
notes, for a longer period, and made more musical modifications during the
former condition.
fMRI: The comparison of “Improv” and “Reproduce” conditions
demonstrated increased signal in the right dorsolateral prefrontal cortex
(DLPFC), right pre-supplementary motor area (pre-SMA), bilateral dorsal
premotor cortex (PMd), left posterior superior temporal gyrus (STG) near
the temporoparietal junction (TPJ), right fusiform gyrus, and bilateral
middle occipital gyri. Improvisational complexity correlated with increased
activity within the pre-SMA. There were no significant differences in brain
activity in the “FreeImprov” versus “Improv” contrast.
Conclusions/Highlighted Discussion
The authors hypothesized that the DLPFC may “maintain and execute an
overall plan for the improvisation through top-down influences on the
activity in subordinate premotor areas,” thus guiding the selection of actions
best suited for a particular improvisation to be executed by the motor
system. The authors suggest the pre-SMA may be involved in the temporal
aspects of response selection during improvisation, including “decisions
about timing and rhythmic patterning.” The rostral PMd may be involved in
guiding the selection of responses based on visual cues (in this case musical
notation), and transforming those cues into motor sequences. The authors
speculate that the medially-located pre-SMA may be related to rhythm and
the lateral rostral PMd to melody, given these regions’ known roles in
temporal and visual motor integration, respectively. The posterior STG
activation was felt to represent increased processing in auditory-motor
feedback loops required to guide the improvisation based on the ongoing
performance, increased auditory monitoring and working memory demands,
or the online retrieval or musical motifs from long-term memory. The
bilateral fusiform and middle occipital activations were felt to reflect
increased visual processing of the musical scores when using them to guide
improvisation. In sum, these authors conclude that improvisation requires a
brain network relying heavily on higher-order cognitive and motor centers
in conjunction with auditory memory/feedback, and, when score-guided,
higher-order visual regions.
Results
Behavioral: Subjects were largely adherent to the rhythmic and melodic
constraints, and demonstrated a higher percentage of unique melodic
sequences when improvising.
fMRI: Melodic improvisation was associated with increased activity in
the anterior cingulate cortex (ACC), left ventral premotor cortex (PMv) and
inferior frontal gyrus (IFG), bilateral dorsal premotor cortex (PMd), and left
cerebellum.
Rhythmic improvisation was associated with increased activity in the
left ACC, PMd, IFG/PMv, sensorimotor cortex, superior parietal lobule
(SPL), and inferior parietal lobule.
The interaction of melodic and rhythmic freedom—the condition of total
improvisational freedom—did not reveal any unique areas of activity
beyond that of melodic or rhythmic freedom considered separately.
However, a conjunction analysis evaluating which areas were involved with
both rhythmic and melodic freedom demonstrated a three-region network:
the left IFG, PMd, and ACC.
Conclusions/Highlighted Discussion
The authors suggest the areas of common activation in both rhythmic and
melodic freedom—IFG/PMv, ACC, and PMd—to be involved in “the
generation of possible musical phrases, selection among them, and
execution of the decided-upon motor output,” respectively.
Conclusions/Highlighted Discussion
The authors highlight the dissociation of activity within the prefrontal
cortex seen during musical improvisation—widespread deactivation of
lateral regions and focal activation with the medial regions—and interpret
this pattern to suggest that “musical creativity vis-à-vis improvisation may
be a result of the combination of intentional, internally generated self-
expression (MPFC-mediated) with the suspension of self-monitoring and
related processes (lateral OFC and DLPFC-mediated) that typically regulate
conscious control of goal-directed, predictable, or planned actions.” The
authors speculate that the increased neocortical activity in multiple sensory
and motor regions may reflect task-related demands (e.g., temporal activity
reflects processing of highly-structured acoustic information, motor region
activity with the generation/selection of novel motor programs) and raise
the possibility that these changes may be triggered in a top-down manner
driven by the prefrontal areas. The authors speculate that deactivations
within parts of the limbic system may reflect the positive emotional valence
associated with creative act, as this is also seen when listening to
pleasurable music. The authors suggest the medial PFC serves as “an index
of internally motivated behaviors” and performs “a broad-based integrative
function, [that] combines multiple cognitive operations in the pursuit of
higher behavioral goals.” The lateral PFC is conceived of as the site of
conscious monitoring, where goal-directed behaviors are evaluated and
corrected, and is “responsible for planning, stepwise implementation, and
on-line adjustment of behavioral sequences that require retention of
preceding steps in working memory.” This state of medial activation and
lateral deactivation within the PFC may allow for highly-creative and
unconstrained actions, which when directed toward musical improvisation,
may allow for creative expression in this highly-trained group of musicians.
Results
Behavioral: There was no difference in the level of novelty between experts
and novices during the improvisational tasks.
The fMRI data demonstrates that experts had significantly reduced
activity in one region—the right TPJ—when compared with non-experts
during improvisational tasks. This was true under conditions of melodic
improvisation but not rhythmic improvisation.
Conclusions/Highlighted Discussion
The finding that TPJ deactivation was associated with expertise while
improvising was unexpected, as the authors had hypothesized that musical
expertise may be associated with increased activity in the network of
regions described in their prior study (Berkowitz & Ansari, 2008). TPJ
deactivation was felt to reflect a training-related utilization of a “top-down,”
goal-directed cognitive strategy during improvisation. They review
evidence that the TPJ, as part of the broader ventral attention network, is
normally involved in “bottom-up, stimulus-driven” attentional processing,
and is important for reorienting attention to behaviorally relevant sensory
stimuli. The authors suggest that during improvisational tasks, experts may
more effectively filter out potentially task-irrelevant stimuli by inhibiting
the TPJ, reflecting a top-down, goal-driven approach to the task.
Results
Behavioral: The performance data showed that the pianists accurately
performed the presented melody during the “Notes” condition, and that
larger intervals were played during the “Notes” and “Random” sections
compared with “Improvise.”
fMRI: Internally-generated actions (conjunction of “Improvise” +
“Random” versus “Notes”) were associated with increased activity in
prefrontal regions (right ACC, left DLPFC, left dorsal medial PFC),
bilateral motor regions (bilateral pre-SMA and cerebellum), bilateral IFG,
and bilateral insula. The contrast of “Improvise”—“Random” did not reveal
any unique areas of neural activity. The contrast of
“Random”—“Improvise,” by comparison, showed a widespread pattern of
bilateral activations, including frontal, temporal, parietal, occipital, insular,
and cerebellar regions. This pattern was maintained when controlling for
the larger interval sizes seen in “Random.” The largest clusters were within
TPJ, medial and lateral PMC, DLPFC, and cerebellum. A parametric
analysis using the number of keys (2, 6, 12) as a modulator of brain activity
did not identify regions with significant correlations during “Improvisation”
contrasted with either “Random” or “Notes.”
Conclusions/Highlighted Discussion
The authors suggest that the observed pattern of activation associated with
internally-generated tasks, namely the bilateral IFG extending into the
DLPFC, right ACC, pre-SMA, bilateral insula, and cerebellum, is “likely
involved in general functions important for a wide-range of free generation
tasks, e.g., attention to action, monitoring of responses in working memory,
response inhibition, and selection.” Several areas reported previously in
Bengtsson et al. (2007) were not seen, including the bilateral rostral PMd,
TPJ, and occipital regions, and the authors suggest this to reflect differences
in the baseline condition, as sight-reading may require additional visual
processing and visually-guided selection of motor sequences. Bilateral IFG
activations, which were not observed in Bengtsson et al. (2007), were felt to
result from comparatively increased musical freedom, and are thought to
contribute to cognitive processes important for the generation of novel
musical phrases, including “retrieval and selection of semantic information
from long-term memory, rule maintenance, and sequential control in
general including syntax processing.”
The widespread increased activity seen with random keyboard presses
(compared with musical improvisation) was felt to reflect task demands:
“far from being a simple model of free choice, [pseudo-random generation]
is a complex and novel task with high demands on attention, planning,
working memory, and executive functions.”
This study suggests musical improvisation utilizes prefrontal, high-order
motor, inferior frontal, and insular regions, but that these regions are not
specific for musical improvisation, as this network is also active during
performance of internally-generated actions that are not musical in nature.
Results
Behavioral: There were more reproduction errors during constrained
improvisational (e.g., “Melody”) than non-improvisational (“Notes”) tasks,
although the absolute number of errors was small.
fMRI: The ROI analysis looking specifically at the pre-SMA and PMd
showed melodic improvisation to be associated with increased activity in
both the left PMd and bilateral SMA, whereas rhythmic improvisation
demonstrated only bilateral pre-SMA activity. Functional connectivity
studies demonstrated increased connectivity of pre-SMA with the
cerebellum (right > left, lobules VI/VII) during rhythmic improvisation
when compared with melodic improvisation or “Notes.” Melodic
improvisation was not associated with increased pre-SMA functional
connectivity with other brain regions. The left PMd was not associated with
changes in function connectivity during either task. The GLM (general
linear model) contrast of free improvisation (“Free”—“Notes”) was
associated with increased activity in prefrontal, motor, and lateral frontal
network, including the pre-SMA, left PMd, left DLPFC extending into the
left IFG (pars triangularis), and decreased activity within the bilateral
inferior occipital gyrus, right precentral gyrus extending across the midline
into the right superior parietal cortex, bilateral medial frontal gyrus, left
superior parietal lobe, and right inferior parietal lobe.
Conclusions/Highlighted Discussion
This study demonstrated differential activation of high-order motor regions
during musical tasks that isolate melodic and improvisational freedom, and
functional connectivity of these regions to other brain regions varies
depending on the task. The authors argue that the pre-SMA plays a critical
role in motor timing and the hierarchical control and sequencing of
movements, which is important in both melody and rhythm generation. The
authors suggest that the PMd may be important for melodic and spatial
processing based on evidence showing the region to be important for the
cognitive aspects of visuomotor integration and spatial targeting of
movement sequences.
The regions associated with free improvisation (“Free”—“Notes”)—the
pre-SMA, dorsal PMC, and DLPFC—are known to be associated with
“explicit processing of novel motor sequences.” The authors contrast their
results with that of Limb and Braun (2008), which showed task-related
deactivations in these areas in musicians with expertise specifically in
improvisation, and suggest that expertise may make the tasks less
cognitively demanding. Expert improvisers may utilize “implicit, routine,
and automated behavior” strategies, which is reflected neurologically in “a
more caudal distribution of activity in the SMA and PMd.”
The observed increase in functional connectivity of the pre-SMA with
the cerebellum during rhythmic improvisation was anticipated given the
cerebellum’s role in motor timing, and demonstrates the region’s ability to
modulate its interactions with other areas based on task demands.
Neural Correlates of Lyrical Improvisation: An
fMRI Study of Freestyle Rap (Liu et al., 2012)
Design
This study sought to understand musical improvisation using a different
creative modality than the keyboard studies described earlier: freestyle rap.
Twelve freestyle rap artists with more than five years of experience rapped
over an eight-measure instrumental track at 85 beats per minute, and the
experimental conditions included “Conventional,” where the subjects were
asked to rap a memorized lyric they had been given prior to the scanning
session; and “Improvise,” where the subjects were asked to improvise
lyrics. Behavioral measurements included blinded ratings of the creative
use of language and rhythm in the improvised compositions (by an expert
panel), and the number of syllables per minute. The subjects also completed
standard beside neurological tests of generative verbal fluency, with both
phonological and semantic constraints.
Results
Behavioral: The subjects generated the same number of syllables in both
the “Improvise” and “Conventional” conditions. On tests of verbal fluency,
the subjects scored above the 80th percentile when compared with age and
education matched controls in both semantic and phonemic tests,
suggesting superior abilities.
fMRI: The main neuroimaging contrast designed to isolate the unique
aspects of improvisation—“Improvise” versus “Conventional”—revealed
increased activity within the several functional networks, including
prefrontal (left medial PFC extending from the frontopolar cortex to the
pre-SMA), language/perisylvian (left IFG, left MTG, left STS, and left
fusiform), and motor (left cingulate motor area [CMA], left pre-SMA, left
dorsal PMC, left caudate, left globus pallidus, and right posterior
cerebellum and vermis) regions. There was decreased activity within the
right DLPFC extending from orbital to superior regions.
A parametric analysis using the expert-rated creativity scores as
predictors of regional activity showed higher creativity scores during
“Improvise” to be associated with higher activity in the left posterior and
middle temporal gyrus, left medial PFC near the superior frontal sulcus, and
the left PCC.
A functional connectivity analysis showed that during “Improvise”
(versus “Conventional”), activity in the left medial PFC (seed selected by
the parametric results) had reduced connectivity with the right DLPFC and
increased connections with the anterior perisylvian (e.g., left IFG) and
cortical motor areas (e.g., CMA, ACC, pre-SMA, and PMd). To define the
extent of this improvisation-associated medial PFC network further,
functional connectivity studies were repeated using the areas of increased
medial PFC connectivity as seeds (e.g., left IFG, pre-SMA, CMA); these
regions all showed increased connectivity with the left amygdala. When the
left amygdala was used as a seed, there was a widespread, bilaterally-
distributed neural network of increased connectivity, which included the
right IFG, right inferior parietal lobule, and bilateral insula.
Conclusions/Highlighted Discussion
The study showed similar results to Limb and Braun (2008), with a
dissociation of medial activation and lateral deactivation within the PFC
during musical improvisation. The authors suggest that this pattern reflects
“a state in which internally-motivated, stimulus-independent behaviors are
allowed to unfold in the absence of conscious volitional control.” The
medial PFC, which regulates motivational drive and guides self-generated
behaviors, is normally regulated by the DLPFC, where executive control
occurs and ongoing adjustments are made “to ensure that actions conform
to explicit goals.” The authors speculate that information from the medial
PFC could bypass the DLPFC via its rich connections to the CMA, an area
known to integrate affective and cognitive representations to guide
behavior. The deactivations of the right DLPFC and other elements of the
dorsal attention network (e.g., intraparietal sulcus) are explained: “top-
down attentional processes mediated by this network may be attenuated
during improvisation, consistent with the notion that a state of defocused
attention enables the generation of novel, unexpected associations that
underlie spontaneous creative activity.”
The areas of activation during improvisational tasks tended to be
lateralized to the left, whereas the deactivations were more right-lateralized
(e.g., DLPFC). The authors suggest the dominant hemisphere activations
are consistent given the unique demands of freestyle rap, an inherently
language-based musical form, and may reflect “spontaneous phonetic
encoding and articulation of rapidly selected words during improvisation …
and spontaneous incorporation into established rhythmic patterns … which
may place additional demands on these regions.”
The widespread, bilaterally-distributed network identified during the
functional connectivity analyses using the medial PFC as the initial seed
may underlie “multi-modal sensory processing and the representation of
subjective experience, and that as a whole, this entire network is more
effectively coupled during spontaneous behavior—perhaps facilitating what
has been described as a psychological ‘flow’ state.”
The correlation of creativity scores with neural activity in the MTG and
STS may reflect superior verbal fluency, as these areas are important for
accessing the mental lexicon, and the medial PFC—also associated with
higher scores of creativity—may suggest a role for motivation/drive in
innovative compositions.
The authors argue that the DLPFC deactivations, which were not
reported in other studies of musical improvisation, may be the result of
fewer secondary cognitive demands associated with other studies’ tasks.
Results
Behavioral: There was a wide but bimodal distribution in both fluidity and
flexibility scores, and the parameters were strongly correlated. This allowed
for a relatively clear separation into the HCG and LCG groups.
fMRI: Comparisons of HCG versus LCG (when “Create” and “Repeat”
tasks were combined) demonstrated increased signal in the left pre- and
postcentral gyrus and left DLPFC. The main improvisational contrast,
“Create” versus “Repeat,” revealed increased signal in the SMA, DLPFC,
and right ventral lateral PFC, when both groups were collapsed into one
(HCG+LCG). When examining the groups separately for this contrast, the
HCG showed increased activity in the left DLPFC, right insula, and right
ventral lateral PFC, whereas the LCG showed increased signal in the left
precentral gyrus and SMA. The contrast of these maps—“Create” versus
“Repeat” for the HCG only compared with the LCG—revealed only an
uncorrected activation within the left DLPFC and right insula.
Parametric analysis showed that flexibility scores covaried with signal in
the left DLPFC, right ventral lateral PFC, and right insula.
Conclusions/Highlighted Discussion
The authors argue that the left DLPFC and right insular activations seen in
the HCG during improvisation, combined with these regions’ positive
correlation with creativity scores in the parametric analysis represent
“widespread integration of networks associated with cognitive,
motivational, and emotional processes,” which may be important for
novelty idea generation. The DLPFC activity observed in the HCG, similar
to that suggested in Bengtsson et al. (2007), reflects “a greater focus of
attention, greater reliance on working memory to retain diverse musical
images in their mind while other images were being processed, greater
inhibition of interfering stimuli to avoid adhering to the original rhythmical
patterns, and greater amount of manipulating to organize their products into
unique and recognizable original combinations.” The tasks employed are
similar to those used by Bengtsson et al. (2007), and may help explain the
discrepant DLPFC findings with Limb and Braun (2008) and Liu et al
(2012). The insular activity, via its interactions with other regions, “serves
to develop subjective emotional and motivational states and to translate
these states into specific action plans. … and the correlation between
anterior insula activation and creativity … likely reflects a positive
association between the capacity to integrate information and creativity
level.”
The LCG group showed only SMA activity, and this was attributed to
the SMA’s role in a network that includes cortical (SMA, IPL), basal
ganglia, and cerebellar structures to integrate sensory and motor
information during performances involving rhythmic movements in
response to auditory stimuli. The authors suggest SMA may have been
more prominent in the LCG because their compositions did not differ from
the originally-presented rhythmic stimuli.
Results
Behavioral: There was significant variability in the amount of experience
amongst the pianists, both overall and specifically for improvisation. The
amount of improvisational training did not correlate with measures of
musical complexity.
fMRI: There were strong correlations of the BOLD signal across all four
experimental conditions, and as such, the conditions were collapsed into
one (“Improvisation”) for the purposes of imaging analysis. Improvisational
experience was negatively correlated with neural activity during
“Improvisation” (contrasted with rest) in several right hemisphere regions,
namely the DLPFC, IFG, anterior insula, and angular gyrus. In a functional
connectivity analysis, improvisational experience was associated with
higher connectivity between prefrontal, premotor, and motor regions of the
frontal lobe during improvisation (contrasted with rest). This was shown
using six different seed regions within the bilateral DLPFC, pre-SMA, and
PMd, although the most extensive connectivity was seen using the right
PMd. Additional areas of increased functional connectivity outside the
frontal lobe were also observed using these seeds, including the parietal,
posterior temporal, primary sensorimotor, and cerebellar regions. The
regions in each of these studies affected by improvisational experience—
right hemisphere regions for neural activity, and bilateral frontal regions in
the functional connectivity analysis—were non-overlapping in their
anatomical distribution. All of these effects were independent of the amount
of classical piano experience or the age of the pianists.
Conclusions/Highlighted Discussion
This study demonstrates a link between the type of training and the
functional neuroanatomy underlying improvised musical performance. The
authors suggest that “greater functional connectivity of the frontal brain
regions seen in the most experienced participants may reflect a more
efficient integration of representations of musical structures at different
levels of abstraction. A higher functional connectivity with the seed regions
was observed with premotor regions and parietal and prefrontal association
cortex and the cerebellum, suggesting the training-related functional
reorganizations may affect both cognitive and sensorimotor aspects of
improvisation.” The authors suggest that the reduced activations of the right
DLPFC and parietal regions observed in those with more extensive
improvisational experience may indicate “automation and reduced top-
down cognitive control,” similar to what was reported by Limb and Braun
(2008) and Liu et al. (2012). The authors explain the finding of training
being associated with reduced brain activity but increased connectivity
between regions during the task of musical improvisation as signifying that
“skilled improvisational performance may thus be characterized by both
lower demands on executive control and a more efficient interaction within
the network of involved brain areas.”
Results
Behavioral: Performers used more keystrokes and showed higher musical
complexity during the emotional conditions (“Happy” and “Fearful”)
compared with the pitch-set conditions (“Tonal” and “Atonal”).
fMRI: In the contrast of the pitch-set conditions versus the emotional
conditions, there was increased activity within the bilateral DLPFC
(extended on the right throughout the middle frontal gyrus into the PMd),
inferior parietal lobes, inferior temporal gyri, left inferior occipital gyri, and
left cerebellum. During the opposite contrast—emotional versus pitch-set
conditions—there was increased signal within the left dorsal medial PFC
(in the superior medial gyrus), left medial orbital gyrus, bilateral IFG,
bilateral insula (extending into the amygdala), left STG, left mid-cingulate,
right precentral gyrus, left central sulcus, right Rolandic operculum, and
bilateral occipital gyri.
The right DLPFC seed ROI used in the functional connectivity analyses
was chosen as the region of overlap in the DLPFC between the GLM
contrast of pitch-set versus emotion and a previously reported DLPFC area
whose activity is related to improvisation practice (Pinho et al., 2014).
Functional connectivity during pitch-set condition (compared with
emotional sets) showed increased connectivity of the right DLPFC with
motor areas (bilateral PMd, left PMv, left SMA), auditory areas (bilateral
STG), left primary sensorimotor cortex, left parietal lobe, and right
cerebellum. During emotional conditions, the right DLPFC showed more
connectivity with parts of the default-mode network (medial PFC and
medial parietal regions). The left DLPFC showed a similar connectivity
pattern as the right DLPFC during emotional conditions, but not during
pitch-set conditions.
Conclusions/Highlighted Discussion
This study demonstrates that improvisation-associated neural activity and
connectivity are modulated by emotional and musical constraints. The
pitch-set task, the authors suggest, requires “an explicit approach to creative
thinking,” and consequently the DLPFC is more active and functionally
connected to premotor, sensorimotor, and cerebellum, which are important
for “integrating goal-oriented information, that is, internal (musical) and
external (response set) constraints, for attentional selection, that is,
cognitive control of action sequencing and motor execution.” The authors
argue that “top-down executive control extends to the level of motor
execution,” and the DLPFC, PMd, and parietal cortex “constitute an
‘intentional framework’ for sensorimotor processing.”
Emotional improvisation, in contrast, may rely on a more “implicit”
strategy, which is reflected in reduced DLPFC activity and its increased
connections with parts of the default mode. The increased activity of the
medial PFC during emotional improvisation is notable given its role in
“representing the affective meaning of stimuli,” and its “functional
interconnections with cortical, striatal, and limbic regions … [that] allow
convergence of sensorimotor integration and visceromotor control in the
processing of emotionally salient information and regulation of behavior.”
The authors point to evidence of tonal representations in the medial PFC,
which “may enable associative processes between music, emotion, and
memories.” During emotional conditions, the IFG controls response
selection “based on retrieval and sequencing processes that … utilized
internalized musical syntactic rules and semantic associations.”
The authors argue these two modes of musical improvisation represent
two neurological “meta-systems,” one an executive system “where the
DLPFC drives integration of sensory, autonomic, and goal-related
information to implement adaptive control,” and another an integrative
system “constituted primarily by the default mode network, where largely
automated processes in specialized brain systems are organized under the
influence of the MPFC for the flexible integration of exogenous and
endogenous information.” An individual may shift between these two
cognitive modes depending on their training and the improvisational
context.
Neural Substrates of Interactive Musical
Improvisation: An fMRI Study of “Trading Fours”
in Jazz (Donnay, Rankin, Lopez-Gonzalez,
Jiradejvong, & Limb, 2014)
Design
This study sought to examine musical improvisation as it occurs with an
interlocutor, as in “trading fours” in jazz. Eleven professional musicians
proficient in jazz piano performance interacted musically with an
interlocutor by alternating four-bar phrases with each other. The constraints
on these interactions between the musician pair characterized the
experimental conditions, which included “Scale-Control,” where only
quarter notes and repeated playing of the D Dorian scale was permitted;
“Scale-Improv,” where melodies were improvised using the D Dorian scale,
but only quarter notes were allowed; “Jazz-Control,” where subjects played
a memorized composition with background accompaniment; and “Jazz-
Improv,” where melodies and rhythms were unrestricted and played with
background accompaniment.
Behavioral assessments included measurement of the performance and
quantification of the musical interactions between the subject and
interlocutor, including note density, pitch class distribution, pitch class
transition, duration distribution, duration transition, interval distribution,
interval transitions, and melodic complexity.
Results
Behavioral: There were more notes played in the “Jazz-Improv” condition
compared with the “Jazz-Control,” with the comparable “Scale” conditions
showing no difference. Melodic complexity was highest and most variable
for “Jazz-Improv.” The melodies traded in phrase pairings were related to
each other in terms of duration, pitch, interval, and melodic complexity.
fMRI: The main contrast in the MRI data, improvised melodies versus
controls, revealed a widespread pattern of activation and deactivation in
both “Scale” and “Jazz” conditions. Areas of increased activity included
language (bilateral IFG pars opercularis and triangularis, bilateral [right
more so than left] posterior STG within Wernicke’s area), prefrontal
(bilateral DLPFC), motor (bilateral SMA), parietal areas (bilateral IPL,
bilateral SPL), bilateral SMG, and bilateral middle occipital gyrus. Areas of
decreased signal included prefrontal areas (bilateral dorsal prefrontal cortex
over the superior frontal gyrus and middle frontal gyrus), default mode
areas (bilateral angular gyrus, bilateral precuneus), and motor areas
(bilateral precentral gyrus). This pattern of BOLD signal change was
similar in both “Jazz” and “Scale” paradigms.
Functional connectivity measured during improvisational exchanges
revealed increased connectivity between the left and right IFG, and anti-
correlations between the bilateral IFG and STG, and the left IFG and
bilateral angular gyri.
Conclusions/Highlighted Discussion
The study demonstrates that improvised musical exchanges are associated
with increased activity in a network that includes traditional perisylvian
language areas and their right-sided homologues (e.g., IFG, posterior STG),
prefrontal and attentional regions (bilateral DLPFC, IPL), premotor/motor
areas (e.g., bilateral SMA, precentral gyrus), and parietal regions (e.g., SPL,
IPL). The right IFG was felt to be important for the “detection of task-
relevant cues, such as those involved in the identification of salient
harmonic and rhythmic elements,” and the right STG important for auditory
short-term memory, as would be required to keep track of the interlocutor’s
ongoing improvisations. The bilateral IFG is important in syntactic
processing of music and speech, and the STG has been implicated in
harmonic processing. The authors suggest a link between linguistic and
musical discourse, and point to shared regions of activity in this study and
those using a speech interlocutor, as well as similarities in their hierarchical
structures, and propose that both utilize a “common neural network for
syntactic operations.”
Increased activity within the DLPFC during improvisation was felt to
represent increased conscious self-monitoring of musical behavior in the
social musical setting, and possibly also increased working demands
associated with trading fours. The authors speculate that increased activity
in the sensorimotor areas represents a “primed” state “as the musician
prepares to execute unplanned ideas in a spontaneous context.”
The authors suggest that the functional deactivations within the bilateral
angular gyrus, and its reduced connectivity with left IFG, may be
“indicative of the lesser role semantic processing has in moment-to-moment
recall and improvisatory musical generation whereby only musical syntactic
information is exchanged and explicit meaning is intangible and possibly
superfluous.”
This study suggests that social paired musical improvisations may utilize
inferior frontal systems important for hierarchical structuring of musical
and linguistic discourse (e.g., musical syntax), and require increased
working memory demands and harmonic processing. Areas important for
the communication of explicit semantic ideas (and their functional
connections) are less active during these exchanges, suggesting a de-
emphasis of these features in musical conversation.
Results
Behavioral: The emotional valence of the facial expressions was associated
with differences in performance, as “positive” improvisations were most apt
to be performed in a major key (71% of the time, compared with 31% for
negative and 46% for ambiguous), had higher note maxima (whereas
“negative” conditions had lower note minima), highest note density
(followed by ambiguous, then negative), and significantly more notes of
shorter duration.
fMRI: When combining all groups, improvisation (versus chromatic)
was associated with increased signal in the left IFG, and decreased signal in
the bilateral medial and lateral frontopolar cortex, DLPFC, angular gyrus,
precuneus, and bilateral mid-cingulate.
Emotional valence was associated with different regions of brain activity
during improvisation. Positive improvisations were associated with
decreased signal within left hippocampus, and more extensive deactivation
in the DLPFC, angular gyrus, and precuneus compared with
negative/ambiguous. Both negative and ambiguous improvisations were
associated with increased activity in the bilateral SMA, and negative
improvisations were associated with decreased signal within the bilateral
hippocampi.
During improvisational blocks, the contrast of positive versus
ambiguous showed increased activity within limbic areas (left
hippocampus, left amygdala, right parahippocampal gyrus). The contrast of
negative versus ambiguous revealed increased signal in dorsal medial
prefrontal (right ACC [BA9]), posterior default mode (left angular gyrus
[BA39]), high-order sensory (SMG [BA40]), and limbic regions (right
hippocampus), and decreased signal within motor (right cerebellum, left
primary motor [BA4]), and auditory areas (bilateral Heschel’s gyrus).
Negative and ambiguous versus positive revealed increased signal in
prefrontal (bilateral frontopolar cortex [BA10]), right ACC (BA 32), right
insula (BA13 and 47), and perisylvian areas (right SMG [BA40], bilateral
middle temporal [BA22]). The contrast of positive versus negative revealed
increased signal only within the right cerebellum. Viewing the emotional
expression itself was not associated with any significant differences in brain
activity, which suggests the observed difference during improvisation did
not simply reflect viewing the emotional stimulus itself.
Functional connectivity analyses using seeds within the left amygdala
and left insula revealed changes in connectivity associated with the
emotional valence of the stimulus. During positive improvisations (versus
chromatic), the left amygdala had reduced connectivity with the left
cerebellum, and the left insula had lower connectivity with areas important
for attention and executive functioning (left superior frontal gyrus, bilateral
middle frontal gyrus), high-order sensory processing (left SMG), and
primary sensorimotor functions (precentral and postcentral gyri), increased
connectivity with visual areas (middle occipital gyrus). During negative
improvisations, the left amygdala had lower connectivity with the right IFG
and left postcentral gyrus. When contrasting positive versus negative
emotions during improvisational trials (not versus chromatic), the left
amygdala had greater connectivity with left-sided attention/executive areas
(superior medial and superior frontal gyri, IPS), ACC, and high-order
sensory areas (SMG). Using the same contrast, the left insula showed
increased connectivity with the Rolandic operculum and reduced
connectivity with midbrain (including substantia nigra).
Conclusions/Highlighted Discussion
The study reveals a network of brain regions important for musical
improvisation (e.g., deactivations within the angular gyrus, precuneus,
medial PFC; activations of the IFG) and demonstrates that activity in these
and other regions is altered by the intended emotional valence of the
compositions. Positive improvisations were associated with robust
deactivations of the DLPFC, which, in association with a lack of increased
activity in the SMA (which is active during tasks requiring continuous
monitoring of motor output), may “indicate that positive improvisation
induces a deeper state of flow than negative or ambiguous improvisation.”
Negative improvisations are associated with increased insular connectivity
with the substantia nigra, a midbrain nucleus containing neurons with
dopaminergic projections to subcortical reward centers. The insula is known
to represent afferent information about internal body states, and the authors
suggest that negative improvisations may be associated with “binding of
visceral awareness” within the insula without any “real-life” negative
consequences, creating a potentially rewarding situation. This may depend
on maintaining “cognitive distance” from the performance, which they
argue is substantiated by the finding of increased activity within the SMA
and frontopolar cortex during negative improvisations, which are regions
known to be involved in cognitive control and self-monitoring.
The authors suggest that positive and negative musical improvisation
may be pleasurable by different mechanisms: “While positive emotional
targets enable more widespread hypofrontality and deeper flow states
during spontaneous creativity, negative emotional targets may be more
closely linked to a stronger visceral experience and greater activity in
reward processing areas of the brain during improvisation.” This study
demonstrates that emotional intent activates different neural networks
during musical improvisation, and that positive and negative emotions
utilize different aspects of attentional, limbic, and sensory processing
during the generation of novel melodies.
P E T (PET)
Results
The authors found that melody generation (relative to rest) was associated
with increased PET signal in the SMA, pre-SMA, primary motor, lateral
premotor, frontal operculum, anterior insula, primary auditory, secondary
auditory, and superior temporal polar cortices. The SMA, primary motor,
and frontal opercular signals were bilateral, and the auditory cortices were
lateralized to smaller foci in the right hemisphere, and more widespread on
the left. An extensive subcortical network was also identified, including
thalamus, putamen, globus pallidus, caudate, midbrain, pons, and
cerebellum. The bilateral parieto-occipital cortices were deactivated. A
similar, overlapping network was identified for sentence generation. The
network of shared activation included the bilateral SMA, left primary
motor, bilateral premotor, left IFG (pars triangularis), left primary auditory,
bilateral secondary auditory, anterior insular, and left anterior cingulate
cortices; the subcortical areas were nearly identical between the two tasks.
Regions specific for melody generation were the dorsal right temporal pole
and right frontal operculum.
Conclusions/Highlighted Discussion
The authors identified a wide regional network that is associated with the
generation of new melodies that includes motor, language, auditory, limbic,
and subcortical areas. The authors suggest that these regions support
processes integral to improvisation, including “(i) accessing rules of
harmony, and (ii) re-ordering, rhythmically altering, re-harmonizing, or
concatenating the stimulus or recalled musical associations to generate
musically-appropriate phrases.” The opercular and planum polare
activations are hypothesized to subserve the “use of implicit knowledge for
harmonic and melodic rules,” the premotor, basal ganglia, and cerebellar
activity to subserve the “representation of rhythmic musical features” such
as meter, and the insula to reflect “kinaesthetically based musical
expressivity.” The other areas—the SMA, ACC, premotor areas, basal
ganglia—are “likely to be involved in the improvised manipulation of
musical structures or perceived in the stimulus,” and also to aid in the
response selection of “generated possibilities to determine the next note in a
phrase.” The authors suggest that music and language shared many
resources, including for audition and vocalization, and use parallel
resources for phonological generativity of different semantic units.
T D C
S ( DCS)
Results
The individual components of the expert ratings of performance were all
positively correlated with one another, and were collapsed into a single
“quality” score. The musical scores improved with increasing prior
improvisational experience.
When all subjects were considered together, tDCS (i.e., sham, anodal,
cathodal) did not affect musical quality. However, there was a quality-by-
expert interaction: right DLPFC stimulation increased the musical quality in
less experienced subjects (anodal more than cathodal) and decreased quality
in experts (anodal only).
Conclusion/Highlighted Discussion
Based on data from prior reports, the authors hypothesized that anodal
stimulation to the right DLPFC would improve the improvisational
performance in less experienced individuals by enhancing top-down
conscious control mechanisms (“Type 2 processes”), and that cathodal
stimulation would improve performance in more experienced improvisers
by enhancing implicit, automatic performance (“Type 1 processes”) that is
hypothesized to occur in hypofrontal states with expertise. The finding that
anodal (i.e., activating) stimulation led to increased quality of performance
in novices and decreased quality in experts is consistent with this idea. The
authors suggest that right DLPFC stimulation may enhance cognitive
processes that are important for creativity more generally, such as working
memory, attention, inhibitory control, and visuospatial memory. The authors
argue that right DLPFC anodal stimulation may also activate and strengthen
a functionally-connected network of brain regions including prefrontal,
premotor, and motor areas, which may “appear similar to more experienced
musicians,” or it may increase theta coherence, which is believed to
integrate “widely distributed neural networks that underlie creativity.”
Experts do not benefit from this stimulation, they argue, because it
disrupted their highly-trained neural networks by recruiting explicit, top-
down processing, “similar to what happens when one attends to the
components of a well-learned skill, causing performance decrements.”
The authors argue that cathodal stimulation did not have the expected,
opposite effects to that of anodal stimulation due to the unclear inhibitory
effects of cathodal stimulation or compensation from other cognitive
domains. Novices may have benefited from cathodal stimulation for similar
reasons as experts, by allowing them to “perform using a more bottom-up
approach.”
E (EEG)
Results
Subjects’ performance was generally accurate: they performed the correct
order of tones on 88 percent of prelearned trials, and only slightly
anticipated the metronomic beat.
During play conditions, improvisation (versus prelearned) was
associated with higher peak amplitudes over left frontal, left central,
bilateral parietal, and bilateral occipital nodes. The same contrast during
imagined performance showed a similar area of peak amplitude difference
over the left frontal region, and novel regions within right lateral temporal
areas. The anatomical sources of the EEG signal during these tasks were
calculated to correspond to the left superior frontal gyrus (SFG), SMA, left
IPL, DLPFC, and right superior temporal gyrus (STG).
There was globally increased alpha power in all leads (most pronounced
in the parieto-occipital regions) during prelearned versus improvised
performances during the play tasks, and slightly increased beta power in the
frontal and parietal regions. There were no power differences in any
frequency range when comparing the “Imagine-Improvise” and “Imagine-
Prelearned” conditions.
The Granger causality analysis revealed dynamic intra-network
interactions. During overt musical performance, improvisation was
associated with decreased causal influences from the SFG to SMA, SMA to
IPL, and IPL to SFG. The strength of the connectivity between these
regions was also negatively correlated with the originality of the
compositions.
Conclusions/Highlighted Discussion
The authors propose that the finding of increased alpha power during the
overt prelearned tasks reflects top-down inhibition or suppression of
potentially interfering alternative responses (e.g., the other three prelearned
melodies.) The increase in beta power during the prelearned task may
reflect “improvement in cerebral integrative and motor functions,” and
“planning and execution of motor movements.”
The authors suggest the SMA may be involved in motor readiness,
motor imagery involving covert vocalizations, and “monitoring of current
and planned motor movements.” The left IPL and right STG are believed to
be involved in a feedback loop involving somatosensory and auditory
perception. The frontal areas (SFG, right DLPFC) may be involved in
cognitive control of musical improvisation. The causality analyses showing
reduced influence of the SFG to SMA to IPL to SFG loop during
improvisation aligns with the hypofrontality hypothesis, whereby “top-
down control may inhibit a creative process driven by bottom-up
processes.” The authors argue that during more complex improvisations, the
information flow is reversed through parts of the network (e.g., the SFG
receives information from the SMA), resulting in bottom-up processing
during more creative output.
The authors conclude that “creative performance in a real-time musical
improvisational task involves regions that may function outside of the top-
down control networks usually seen in traditional decision-making tasks.”
This may be driven by the time constraints related to the task, wherein
deliberate decision making about individual note choices is not possible,
resulting in reliance on “bottom-up processes to control note choices using
aesthetic rules that our advanced musician participants have internalized
during a lifetime of music engagement.”
Creativity as a Distinct Trainable Mental State: An
EEG Study of Musical Improvisation (Lopata,
Nowicki, & Joanisse, 2017)
Design
This study used EEG to evaluate three questions about the neural substrates
underlying improvisation. The first was whether there is a difference in
frontal alpha activity between musical improvisation, rote playback, and
passive listening, since synchronous frontal alpha oscillations are
hypothesized to serve as a marker of implicit, bottom-up “Type 1” creative
processes (see above). The second was to look for changes in alpha
synchronization associated with improvisational expertise/training, and the
third was to assess whether changes in alpha frequency correlate with the
quality of improvised compositions, as rated by experts. Twenty-two
musicians with a wide variety of musical experience (range: 4–48 years;
mean 18.5; SD 11.7) were split into two groups, one with formal
institutional training in improvisation (“FITI”) and the other without (“Non-
FITI”). Prior to testing, the musicians were shown three charts of 16 bars of
chord progressions and given the diatonic structures for each progression
(e.g., C-blues, G-major), but without overlying melodies. The experimental
tasks were performed in the same order each time: “Listen,” where a
melody was played and the subjects passively listened; “Learn,” where
subjects actively learned to play the melody on a keyboard; “Imagine
Playback,” where subjects imagined playing the prior melody; “Actual
Playback,” where subjects overtly played the learned melody; “Imagine
Improvisation,” where the subjects imagined improvising melodies over the
chord progressions; and “Actual Improvisation,” where they improvised
over the chords.
Behavioral measures included an expert assessment of the creativity of
the improvisational musical creativity via the use of a questionnaire.
EEG data included measurements of upper alpha range power (10–12
Hz), which were calculated for each condition (versus a pre-stimulus
reference interval) at each electrode, and a measurement of synchronization
across electrodes.
Results
In both the FITI and non-FITI groups, there was increased frontal alpha
synchronization in the right hemisphere during “Listen,” “Playback,” and
“Improvisation” tasks. During “Improvisation,” however, the FITI group
showed increased right hemisphere alpha power compared with the non-
FITI group. In the FITI group, alpha synchronization was higher during
improvisation compared with both “Listen” and “Playback,” suggesting a
unique interaction in alpha synchrony with improvisation and expertise.
Also in the FITI group there were positive correlations between the left and
right hemisphere frontal alpha synchrony for all tasks.
In the non-FITI group, there was a strong negative correlation of right
hemisphere alpha synchrony with both musical and improvisational
experience during “Improvisation”; this was true to a lesser extent during
“Listen” and “Playback.” In the FITI-group, there were positive correlations
of left hemisphere alpha synchrony and age, musical experience, and
improvisational experience during all tasks.
There were no significant differences in creativity scores between FITI
and non-FITI groups, but in the FITI group only, there was a positive
correlation between creativity scores and right hemisphere alpha
synchronization.
Conclusions/Highlighted Discussion
The study demonstrates that frontal alpha synchronization is associated with
musical improvisation, which is enhanced by formal training experience,
and is associated with more creative performances in the most experienced
improvisers. The authors interpreted the increased frontal alpha
synchronization during improvisation to be “evidence of an underlying
creative mental state characterized by immersion in a Type 1 spontaneous
processing mode,” that suggests “top-down processing and internal focus of
attention,” and not merely “a suppression of executive functions and
logical-rational thought processes.” The authors speculate that increased
right hemisphere alpha synchronization in the FITI group during
“Improvisation” to “support the view of a special role of right frontal brain
areas in the generation of original ideas, and as benefitting from expertise
and development through training.”
In the non-FITI group, the finding of negative correlations of right
hemisphere alpha synchronization with musical and improvisational
experience during improvisation is interpreted to reflect this group’s “lack
of immersion in Type 1 spontaneous processing,” and suggests an
engagement with music that is more deliberate than spontaneous. The
authors argue that the correlation between right frontal alpha synchrony and
creativity scores in those with FITI suggest that “Type 1 spontaneous
processing tends to yield higher quality improvised performance” in these
experts.
S D
Motor Regions
Motor areas control the body’s movements and include primary motor
cortex (e.g., precentral gyrus) and high-order regions important for
planning, sequencing, initiation, and monitoring of movement (e.g., PMd,
PMv, SMA, pre-SMA), emotionally-guided movement (e.g., CMA),
patterning and sequencing of movements (basal ganglia), and coordination
of movements (cerebellum). Given that musical improvisation can only be
externalized through movement, it is not surprising that all of these regions
are involved in musical improvisation.
The SMA and pre-SMA are important in the selection, initiation, timing,
and monitoring of motor movements, and are thought to play a role in the
rhythmic patterning during improvisational tasks (Bengtsson et al., 2007;
Brown et al., 2006; de Manzano & Ullén, 2012a; Donnay et al., 2014; Liu
et al., 2012; Villarreal et al., 2013) and hierarchical control of motor
sequencing (de Manzano & Ullén, 2012a). The pre-SMA is more strongly
connected with the cerebellum during tasks of rhythmic improvisation (de
Manzano & Ullén, 2012a), which highlights its role in timing, and also with
the limbic areas during freestyle rap (Liu et al., 2012), suggesting an
interaction beyond that of other motor areas. The connection between
higher-order motor regions may be modulated by training, as more
experienced improvisers show increased connections of the SMA and PMd
with a widespread network involving prefrontal, premotor, motor, parietal,
and auditory regions (Pinho et al., 2014).
The PMd is reported in many studies (Bengtsson et al., 2007; Berkowitz
& Ansari, 2008; Brown et al., 2006; de Manzano & Ullén, 2012a; Limb &
Braun, 2008; Liu et al., 2012). It is suggested to play a role in sensorimotor
integration, whereby sensory information (often visual) is used to guide the
sequencing and planning of motor movements; it is also important for
internally-generated actions, and is connected with prefrontal areas. The
region may be important in reading musical notation (Bengtsson et al.,
2007), melodic performance (Bengtsson et al., 2007), and more broadly in
top-down, explicit processing of novel motor sequencing (de Manzano &
Ullén, 2012a).
The CMA is thought to guide the selection of voluntary movements
based on expected rewards, and is known to integrate limbic information to
guide motor behaviors. It is associated with increased activity during
musical improvisation (Liu et al., 2012). During freestyle rap, the CMA
shows increased functional connectivity with the amygdala as part of a
broader network integrating affective, motor, and perceptual processes (Liu
et al., 2012). The authors speculate it may represent an alternative pathway
of behavioral expression occurring outside DLPFC-mediated, explicit
motor selection.
Limbic/Affective Processing
Given the emotional nature of musical improvisation, it is not surprising
that limbic areas are involved. These regions help to represent emotion,
motivation, and memory, and are connected richly with the autonomic
nervous system, which provides information about internal body states.
There are reports of reduced activity, including the hypothalamus,
amygdala, hippocampus, parahippocampal gyrus, temporopolar cortex, and
ventral striatum, which may be indicative of the positive emotional valence
associated with improvising (Limb & Braun, 2008). During freestyle rap,
the medial PFC shows increased functional connectivity with the amygdala
via the IFG, CMA, and pre-SMA (Liu et al., 2012).
Insula: The insula is important in representing subjective emotional and
motivational states as it receives interoceptive inputs from the body via the
autonomic nervous system and integrates this information with sensory,
limbic, hedonic, and cognitive inputs. It is also important in salience
detection, and serves an important role in switching between different large-
scale networks (e.g., central executive and default mode networks).
Successful integration of highly-integrated emotional representations into
specific motor plans may be important for creative expression under certain
conditions, a process that may be enhanced when guided by the medial
PFC. Activity in the insular cortices has been reported in several studies of
musical improvisation (de Manzano & Ullén, 2012b; Limb & Braun, 2008;
Pinho et al., 2016; Villarreal et al., 2013), with reports of both increased
(Brown et al., 2006; de Manzano & Ullén, 2012b; Pinho et al., 2016;
Villarreal et al., 2013) and decreased (Limb & Braun, 2008) activity. When
comparing groups with high versus low creativity scores on a task of
rhythmic improvisation, the right insula was associated with increased
creativity scores, and its activity was positively correlated with higher
scores (Villarreal et al., 2013). Improvisation with emotional intent is
associated with bilateral insular activation (Pinho et al., 2016), and the
expression of negatively-valenced emotional improvisations is associated
with increased insular connectivity with the midbrain substantia nigra
(McPherson et al., 2016).
Language Areas
The left IFG is critical for expressive language and syntax/grammar,
functioning as part of Broca’s area. The IFG may be involved in other
functions, including response inhibition, mirroring of external motor
movements, generative verbal fluency, and hierarchical motor sequencing.
The IFG has been implicated in several studies of improvisation (Berkowitz
& Ansari, 2008; de Manzano & Ullén, 2012a, 2012b; Donnay et al., 2014;
Limb & Braun, 2008; Liu et al., 2012; McPherson et al., 2016; Pinho et al.,
2016), and is thought to play a role in the generation and selection of motor
sequences (Berkowitz & Ansari, 2008), novel musical phrases (de Manzano
& Ullén, 2012b), detection of salient harmonic and rhythmic elements
(Donnay et al., 2014), and hierarchical structuring of musical phrases
(Donnay et al., 2014). During freestyle rap, the medial PFC has enhanced
functional connections with the left IFG as part of a broader, integrative
network. Donnay et al. (2014) demonstrate the IFG to be functionally
disconnected from areas important for communication of explicit semantic
information (e.g., angular gyrus) during improvisational tasks that utilize a
musical interlocutor, and the authors suggest that in musical
communication, explicit semantic knowledge is superfluous, and that
“acoustic-phonologic-analysis” areas are paramount.
Sensory Processing
Sensory information is represented in the cortex in a hierarchical manner.
Incoming sensory information is initially processed as simple unimodal
representations (primary sensory areas), and progressively organized into
complex unimodal representations (secondary sensory cortex), then later
combined with other sensory modalities in heteromodal regions. These
regions are subject to both bottom-up and top-down regulation. A number
of studies report increased activity within primary and unimodal sensory
areas (Bengtsson et al., 2007; Brown et al., 2006; Limb & Braun, 2008).
This may be related to task demands, suggesting a role for increased
sensory processing during improvisation, or possibly a release phenomenon
with reduced top-down inhibition (Limb & Braun, 2008).
Auditory: Auditory sensory streams are located within the superior and
lateral temporal areas, and musical improvisation is associated with
activation of these areas, including the STG (Bengtsson et al., 2007; Brown
et al., 2006; Donnay et al., 2014; Limb & Braun, 2008; Liu et al., 2012;
Pinho et al., 2016), MTG (Limb & Braun, 2008; Liu et al., 2012), and ITG
(Limb & Braun, 2008). Activity within the posterior superior temporal
areas (Bengtsson et al., 2007; Limb & Braun, 2008; Liu et al., 2012), a
region known to be important in highly-structured auditory processing, is
involved in auditory working memory (Bengtsson et al., 2007: Donnay et
al., 2014). It may be part of an auditory-motor feedback loop where
auditory information is utilized online to guide the next musical idea via
higher-order motor planning through instrumental performance (Bengtsson
et al., 2007), and may aid the retrieval of stored musical motifs (Bengtsson
et al., 2007). The posterior temporal regions may also be important in
harmonic processing (Donnay et al., 2014). Activity in this region is
associated with higher scores of creativity during improvised rap (Liu et al.,
2012).
Somatosensory: The SPL—a secondary somatosensory region—has
been reported to have increased activity in several studies (Berkowitz &
Ansari, 2008; Donnay et al., 2014; Limb & Braun, 2008), but not all (Liu et
al., 2012). The primary sensory areas are also reported. This may reflect
task demands, although in these studies the motor output (and thus
somatosensory feedback) in both the improvisational tasks and controls was
similar, and as such was suggested to represent a “generalized
intensification of activity in all sensory modalities” associated with musical
spontaneity (Limb & Braun, 2008).
Visual: The occipital cortex, which is the site of hierarchical visual
processing, including the fusiform and lingual gyri, was activated in a
number of studies (Bengtsson et al., 2007; Donnay et al., 2014; Limb &
Braun 2008; Liu et al., 2012; Pinho et al., 2016), and has been reported to
reflect visual demands associated with using a musical score to guide
improvisation (Bengtsson et al., 2007).
R
Adhikari, B. M., Norgaard, M., Quinn, K. M., Ampudia, J., Squirek, J., & Dhamala, M. (2016). The
brain network underpinning novel melody creation. Brain Connectivity 6(10), 772–785.
Beaty, R. E. (2015). The neuroscience of musical improvisation. Neuroscience & Biobehavioral
Reviews 51, 108–117.
Beaty, R. E., Benedek, M., Silvia, P. J., & Schacter, D. L. (2016). Creative cognition and brain
network dynamics. Trends in Cognitive Sciences 20(2), 87–95.
Bengtsson, S. L., Csikszentmihalyi, M., & Ullén, F. (2007). Cortical regions involved in the
generation of musical structures during improvisation in pianists. Journal of Cognitive
Neuroscience 19(5), 830–842.
Berkowitz, A. L., & Ansari, D. (2008). Generation of novel motor sequences: The neural correlates
of musical improvisation. NeuroImage 41(2), 535–543.
Berkowitz, A. L., & Ansari, D. (2010). Expertise-related deactivation of the right temporoparietal
junction during musical improvisation. NeuroImage 49(1), 712–719.
Brown, S., Martinez, M. J., & Parsons, L. M. (2006). Music and language side by side in the brain: A
PET study of the generation of melodies and sentences. European Journal of Neuroscience 23(10),
2791–2803.
de Manzano, O., & Ullén, F. (2012a). Activation and connectivity patterns of the presupplementary
and dorsal premotor areas during free improvisation of melodies and rhythms. NeuroImage 63(1),
272–280.
de Manzano, O., & Ullén, F. (2012b). Goal-independent mechanisms for free response generation:
Creative and pseudo-random performance share neural substrates. NeuroImage 59(1), 772–780.
Donnay, G. F., Rankin, S. K., Lopez-Gonzalez, M., Jiradejvong, P., & Limb, C. J. (2014). Neural
substrates of interactive musical improvisation: An FMRI study of “trading fours” in jazz. PloS
ONE 9(2), e88665.
Limb, C. J., & Braun, A. R. (2008). Neural substrates of spontaneous musical performance: An fMRI
study of jazz improvisation. PloS ONE 3(2), e1679.
Liu, S., Chow, H. M., Xu, Y., Erkkinen, M. G., Swett, K. E., Eagle, M. W., … Braun, A. R. (2012).
Neural correlates of lyrical improvisation: An fMRI study of freestyle rap. Scientific Reports 2,
834. doi:10.1038/srep00834
Lopata, J. A., Nowicki, E. A., & Joanisse, M. F. (2017). Creativity as a distinct trainable mental state:
An EEG study of musical improvisation. Neuropsychologia 99, 246–258.
Lu, J., Yang, H., Zhang, X., He, H., Luo, C., & Yao, D. (2015). The brain functional state of music
creation: An fMRI study of composers. Scientific Reports 5, 12277. doi:10.1038/srep12277
McPherson, M. J., Barrett, F. S., Lopez-Gonzalez, M., Jiradejvong, P., & Limb, C. J. (2016).
Emotional intent modulates the neural substrates of creativity: An fMRI study of emotionally
targeted improvisation in jazz musicians. Scientific Reports 6, 18460. doi:10.1038/srep18460
Pinho, A. L., de Manzano, O., Fransson, P., Eriksson, H., & Ullén, F. (2014). Connecting to create:
Expertise in musical improvisation is associated with increased functional connectivity between
premotor and prefrontal areas. Journal of Neuroscience 34(18), 6156–6163.
Pinho, A. L., Ullén, F., Castelo-Branco, M., Fransson, P., & de Manzano, O. (2016). Addressing a
paradox: Dual strategies for creative performance in introspective and extrospective networks.
Cerebral Cortex 26(7), 3052–3063.
Rosen, D. S., Erickson, B., Kim, Y. E., Mirman, D., Hamilton, R. H., & Kounios, J. (2016). Anodal
tDCS to right dorsolateral prefrontal cortex facilitates performance for novice jazz improvisers but
hinders experts. Frontiers in Human Neuroscience 10, 579. Retrieved from
https://doi.org/10.3389/fnhum.2016.00579
Villarreal, M. F., Cerquetti, D., Caruso, S., Schwarcz López Aranguren, V., Gerschcovich, E. R.,
Frega, A. L., & Leiguarda, R. C. (2013). Neural correlates of musical creativity: Differences
between high and low creative subjects. PloS ONE 8(9), e75427.
CHAPT E R 21
NEURAL MECHANISMS OF
M U S I C A L I M A G E RY
T I MO T H Y L . H U B B A R D
I P M
Brain Damage
Although there have been numerous studies of the effects of brain damage
on music perception, cognition, and production (for reviews, see Marin &
Perry, 1999; Peretz & Zatorre, 2005; Stewart, von Kriegstein, Warren, &
Griffiths, 2006), there have been relatively few studies of musical imagery
in patients with brain damage. The studies that have been reported typically
compared performance involving imagery in brain-damaged patients with
performance on the same task in a control group. Patients with damage to
the right temporal lobe performed worse on pitch comparisons in imagery
and in perception than did patients with damage to the left temporal lobe or
control participants (Zatorre & Halpern 1993). Halpern (2003) suggested
these lesion data and subsequent imaging data (from Zatorre et al., 1996)
demonstrated the right superior temporal gyrus is involved in comparisons
of pitch in imagery (see also Samsom & Zatorre, 1991). Patients with right
temporal lobe damage to the area including Heschl’s gyrus do not perceive
a missing fundamental (Zatorre, 1988), and this is consistent with a role for
this area in top-down representation of pitch. Patients with right hemisphere
damage have difficulty in processing information regarding musical interval
and musical contour (Liégois-Chauvel, Peretz, Babaï, Laguitton, &
Chauvel, 1998; Peretz, 1990) and in identification of sad music (Khalfa,
Schon, Anton, & Liégeois-Chauvel, 2005), and this predicts such patients
would have similar difficulties in musical imagery. More positively, music
influences brain plasticity, and so it could be predicted that musical imagery
might be useful in the treatment of some neurological damage or disorders
(e.g., melodic intonation therapy, Peretz, 2013; also Bringas et al., 2015;
Sabaté, Llanos, & Rodriguez, 2008; Särkämö, Altenmüller, Rodriguez-
Fornells, & Peretz, 2016).
Clinical studies of individuals with trauma-induced amusia (e.g., Marin
& Perry, 1999; Satoh, 2014) or congenital amusia (e.g., Peretz, 2013) have
shed light on neural mechanisms of music processing, but musical imagery
has typically not been studied in such individuals. Amusias might have a
basis in perception or memory (Peretz, 2002); to the extent an amusia
involves dysfunction of memory, imagery might be impacted (e.g., Satoh,
2014, explicitly identifies memory as internal imagery), but to the extent an
amusia involves dysfunction in perception, imagery might be relatively
spared. Also, parallels between types of amusia and types of aphasia (e.g.,
receptive, production) suggest there may be some overlap in neural
mechanisms that process music and neural mechanisms that process
language (cf. Besson & Schön, 2003; Marin & Perry, 1999; Patel, 2008).
Additionally, findings that patients with amusia have difficulty in spatial
tasks such as mental rotation (Douglas & Bilkey, 2007; but see Tillmann et
al., 2010), coupled with findings that some types of musical imagery
manipulation involve cortical areas implicated in mental rotation (Zatorre et
al., 2010), suggest such patients might have impaired musical imagery.
Studies of patients with amusia suggest music functions are not as strongly
lateralized as language functions (Alossa & Castelli, 2009), and this has
been confirmed in non-patient studies as well (e.g., Parsons, 2003; Platel et
al., 1997). Also, presence of amusia predicts deficits in auditory emotion
recognition in schizophrenia, and this might reflect development of music
and language from the same musical protolanguage (Kantrowitz et al.,
2014).
Physiological Measures
Many studies recorded physiological measures in an attempt to understand
neural mechanisms of musical imagery. These studies typically involved
electrophysiology such as EEG and event-related potential (ERP) or brain
imaging such as PET and fMRI (for review, see Koelsch, 2012).
Electrophysiology
Imaging a melody results in more high-band synchronized alpha than does
perceiving a melody (Schaefer, Vlek, & Desain, 2011; Villena-González,
López, & Rodríguez, 2016), and alpha is increased during imagery of more
complex tones (van Dijk, Nieuwenhuis, & Jensen, 2010). Emitted potentials
occur when a musical note is expected but not presented (Cebrian & Janata,
2010b; Janata, 2001), and these are similar to evoked potentials elicited by
presentation of a musical note. There are differences in size of the N1 in
response to a perceived tone as a function of image accuracy and whether
preceding tones were imaged or perceived (Cebrian & Janata, 2010a). If a
participant deliberately generates an auditory image appropriate to a
stimulus seen in a visual picture, P2 and LPC are increased (Wu, Mai,
Chan, Zheng, & Luo, 2006). A larger mismatch in loudness or pitch
between imaged tones and subsequent perceived tones elicits a larger N2
(Wu, Mai, Yu, Qin, & Luo, 2010) and lower-pitched or louder images and
percepts evoke a larger N1 and LPC (Wu, Yu, Mai, Wei, & Luo, 2011).
Accented beats in a sequence of imaged or perceived beats result in a larger
positive amplitude after 180–250 milliseconds and a larger negative
amplitude after 350 milliseconds (Vlek, Schaefer, Gielen, Farquhar, &
Desain, 2011). Relatedly, rhythmic aspects of melody are more easily
isolated in EEG than are pitch or melody-driven aspects (Schaefer, Desain,
& Suppes, 2009). Mismatch negativity is evoked in musicians for perceived
and for imaged musical stimuli (Herholz, Lappe, Knief, & Pantev, 2008;
Yumoto et al., 2005). Continuation of a lyric in imagery during an
unexpected silent gap in familiar music results in several changes in
perceptual, attentional, and cognitive components of ERPs (Gabriel et al.,
2016). In highly trained musicians, ERPs while reading a visual musical
score are indistinguishable from ERPs while listening to auditory notes
(Simoens & Tervaniemi, 2013). In general, imagery of a musical stimulus
results in generation of ERP or EEG patterns similar to those generated by
perception of a musical stimulus.
Brain Imaging
There have been numerous studies involving brain imaging during
processing of musical stimuli (for reviews, see Koelsch, 2010, 2012; also
Peretz & Zatorre, 2003) and changes in the brain related to musical training
(Wan & Schlaug, 2013). There is substantial overlap of cortical areas
activated in musical imagery and activated in music perception, especially
in Wernicke’s area and its right hemisphere homologue (Zhang, Chen, Wen,
Lu, & Liu, 2017) and auditory association areas (e.g., Daselaar, Porat,
Huijbers, & Pennartz, 2010; Herholz, Halpern, & Zatorre, 2012; Zatorre et
al., 1996). Spontaneous imagery during an unexpected gap in a well-known
musical piece (Kraemer, Macrae, Green, & Kelley, 2005) or during a silent
gap prior to the start of an expected music track on a familiar CD (Leaver,
van Lare, Zielinski, Halpern, & Rauschecker, 2009) involves activation of
auditory association areas as well as prefrontal and motor areas. Auditory
imagery may activate frequency-specific regions in primary auditory cortex
(Oh, Kwon, Yang, & Jeong, 2013). When participants listen to four-part
harmony, there is greater activation in bilateral temporal lobes, cingulate
gyrus, and medial cerebellum when participants focus on the harmony as a
whole, but greater activation of superior parietal, bilateral precuneus, and
bilateral orbital frontal cortices if participants focus on a particular (e.g.,
alto) line (Satoh, Takeda, Nagata, Hatazawa, & Kuzuhara, 2001). Judgment
of similarities of perceived timbres and of imaged timbres results in similar
cortical activation (Halpern, Zatorre, Bouffard, & Johnson, 2004):
secondary auditory cortex and supplementary motor cortex are activated in
both imagery and perception, but primary auditory cortex is activated only
in perception (see also Zhang et al., 2017). Indeed, passive listening to
music by musicians (Haueisen & Knösche, 2001) and non-musicians
(Perrone-Capano, Volpicelli, & di Porzio, 2017) who remain motionless
results in activation of cortical motor areas.
Participants who self-report more vivid musical imagery exhibit greater
activation in right superior temporal gyrus and prefrontal cortex (Herholz et
al., 2012) and in right parietal cortex (Zatorre et al., 2010). Higher self-
reported vividness of auditory imagery correlates with gray matter volume
in left inferior parietal lobe, medial superior frontal gyrus, middle frontal
gyrus, and left supplementary motor area (Lima et al., 2015). Application of
TMS over the right hemisphere (to disrupt cortical activation) disrupts pitch
discrimination (Halpern, 2003). Imagery reversal of a musical stimulus (i.e.,
scanning backward through a melody; Zatorre et al., 2010) activates
intraparietal sulcus and ventrolateral and dorsolateral frontal cortex (areas
involved in manipulating sensory information). Musicians who read a
musical score initially exhibit activation in occipital areas that spreads to
midline parietal and then to left temporal auditory association areas and
right premotor areas, and this pattern could reflect emergence of notational
audiation, that is, auditory imagery of a piece of music that is evoked by
reading the musical score of that piece (Schürmann, Raij, Fujiki, & Hari,
2002). Participants instructed to image a single note exhibit activation of
bilateral superior temporal gyri, medial and inferior frontal gyri, and
precuneus (Yoo, Lee, & Choi, 2001). Overall, brain imaging studies
generally support the idea that neural mechanisms are shared between
imagery and perception and between imagery and production (see later
subsection on “Mental Practice and Performance”), although there are
exceptions (e.g., primary auditory cortex is less likely to be activated during
imagery than during perception).
I M I
The majority of laboratory studies of musical imagery involve images
created in response to a stimulus or task demand, and as noted above, these
studies suggest such imagery generally recruits neural mechanisms similar
to those used in music perception, cognition, and performance. However,
musical imagery can be involuntary and occur spontaneously and without
conscious control. Five types of involuntary musical imagery are
considered here, namely (a) anticipatory musical imagery, (b) musical
hallucinations, (c) musical imagery in schizophrenia, (d) earworms, and (e)
synesthesia.
Schizophrenia
The most commonly investigated psychopathology within auditory imagery
literature is schizophrenia. Although the majority of investigations of
schizophrenia focused on auditory hallucinations involving verbal stimuli
(e.g., Cho & Wu, 2013; Evans, McGuire, & David, 2000; Johns et al., 2001;
McGuire et al., 1996; Shergill, Bullmore, Simmons, Murray, & McGuire,
2000), cases of musical hallucinations have been documented. Saba and
Keshavan (1997) documented sixteen patients with schizophrenia who
reported musical hallucinations. Musical imagery in schizophrenia is
typically hallucinatory (not under voluntary control), and Baba and
colleagues (Baba, Hamada, & Koca, 2003) suggested a model of musical
hallucination in schizophrenia in which musical imagery becomes more
obsessive in quality, is perceived as originating outside the individual, and
is ultimately accepted as part of the self. The content of musical
hallucinations in schizophrenia is often described as religious, and this is
consistent with observations that delusions in schizophrenia often contain
religious themes (e.g., Galant-Swafford & Bota, 2015). Brain imaging
acquired during a schizophrenic patient’s musical hallucinations revealed
increased activity in right orbitofrontal cortex (Bleich-Cohen, Hendler,
Pashinian, Faragian, & Poyurovsky, 2011). Relatedly, differences in brain
activation patterns of patients with schizophrenia and controls when spoken
sentences were imaged in another person’s voice, but not when sentences
were imaged in the participant’s own voice (McGuire et al., 1995), suggest
articulatory information relative to the inner voice (discussed later in the
chapter) might be overly represented in schizophrenia.
Earworms
Perhaps the fastest growing area of research on musical imagery during the
past several years involves earworms (also referred to as involuntary
musical imagery, stuck-song syndrome, brain worms, sticky music,
intrusive musical imagery, and perpetual music track; see Williams, 2015).
Earworms are a fragment of a song or melody that repeatedly and
involuntarily occupies an individual’s awareness. Unlike musical
hallucinations, earworms are generally not considered to reflect
psychopathology and are usually not considered distressing by those who
experience them (Beaty et al., 2013; Halpern & Bartlett, 2011; Hemming &
Merrill, 2015). Research on earworms has focused on descriptive
phenomenology and behavioral correlates (for summary, see Hubbard,
forthcoming), and there has been little consideration of neural mechanisms
of earworms. Levitin (2007) suggested earworms occur when neural areas
representing a specific piece of music get stuck in “playback mode.”
Farrugia and colleagues (Farrugia, Jakubowski, Cusack, & Stewart, 2015)
found that frequency of occurrence of earworms was related to cortical
thickness in the right frontal and temporal cortices and anterior cingulate,
whereas affective aspects of involuntary musical imagery were related to
gray matter volume in right temporopolar and parahippocampal cortices. It
could be predicted that neural mechanisms previously shown to be involved
in voluntary musical imagery might be activated during earworms, and
there might be additional (or lack of) activation in other areas or differences
in time course of activation that reflect the involuntary nature of earworms
(e.g., differences in voluntary voice imagery and involuntary voice
hallucinations; Linden et al., 2011).
Synesthesia
Synesthesia occurs if a stimulus in one dimension or modality induces
systematic and idiosyncratic perceptual experience of a specific stimulus in
a different dimension or modality (e.g., hearing a specific sound induces
visual experience of a specific color, e.g., Baron-Cohen & Harrison, 1997;
Cytowic, 2002; Robertson & Sagiv, 2005). Reports of synesthesia in which
a non-musical stimulus elicits musical imagery are rare (e.g., see listings in
Cytowic & Eagleman, 2011; Day, 2016), and perhaps the most well-known
is that of composer Jean Sibelius, who experienced different musical chords
when viewing different colors (Pearce, 2007).1 It might be tempting to
consider musical hallucinations or earworms as forms of synesthesia, but
neither musical hallucinations nor earworms match the typical
phenomenology of synesthesia (e.g., specific synesthetic experiences are
evoked by specific stimuli and consistent over long periods of time). Most
research on neural mechanisms of synesthesia focused on color-grapheme
synesthesia (in which perception of letters or numerals induced experience
of color; e.g., Rouw & Scholte, 2010), and there has been little research
involving neural mechanisms of synesthesia involving musical imagery.
Possible neural mechanisms of synesthesia involve activation
(Ramachandran & Hubbard, 2001) or disinhibition (e.g., Grossenbacher &
Lovelace, 2001) of cross-connections between sensory areas, and so one
speculative possibility is that lack of evoked musical imagery in synesthesia
might be related to the general lack of activation in primary auditory cortex
during musical imagery. Another speculative possibility is that musical
perception already involves non-auditory (e.g., kinesthetic) elements, and
so activation of other non-auditory information in musical imagery is not
experienced as synesthesia per se.
E M I
Mimicry
Listening to or recalling music has been suggested to involve motor
mimicry (Cox, 2016). Many musical sounds provide information regarding
the human motor action that produced those sounds (including information
involving spatial and force metaphors), and musical imagery is often
accompanied by visual or motor images related to the sound source (Godøy,
2001). Cox suggested an important component of music comprehension is
imitating, either overtly or covertly, the sound-producing actions of
performers. Such imitative movements might involve movements
appropriate to playing an instrument or subvocal imitation of musical
sounds, and musical features (e.g., pitch, duration, strength, etc.) might be
represented mimetically; indeed, even simple tapping along with the beat
might be considered mimicry. Western popular music has been dominated
by music that is easily singable or danceable (Cox, 2016), and this is
consistent with the importance of embodiment and mimicry in music
processing. Such mimicry might involve overt physical action or covert
firing of mirror neurons. Mirror neurons can be activated by sounds
associated with a given action (Kohler et al., 2002), and so might be
involved in neural activity relevant to singing or playing an instrument. One
consequence of such mimicry is that musical imagery involves kinesthetic
and proprioceptive information (Hubbard, 2013b), and the importance of
kinesthetic and proprioceptive information in auditory and musical imagery
is seen in the distinction between the inner ear and inner voice and in
separate roles of auditory imagery and kinesthetic imagery in mental
practice and performance.
Musical Affect
The evolutionary origins of music have been linked to communication of
emotional information (e.g., Bryant, 2013; Snowdon, Zimmerman, &
Altenmüller, 2015), and this might account for the common observation of
a link between music and affect (for review, see Juslin & Sloboda, 2001).
Indeed, music perception increases activation in mesocorticolimbic areas,
especially in the amygdala and hippocampus (e.g., Blood & Zatorre, 2001;
for review, Koelsch, 2010). Listening to music is linked with release of
dopamine in dorsal and ventral striatum, and the amount of dopamine
released appears related to the amount of pleasure experienced (Salimpoor,
Benovoy, Larcher, Dagher, & Zatorre, 2011). Furthermore, perception of
music is linked with an increase of oxytocin (Chanda & Levitin, 2013),
which is linked with social bonding. To the extent that musical imagery
involves activation of the same neural mechanisms as music perception,
cognition, and production, then musical imagery would presumably be
linked with affect. Indeed, as noted earlier, the majority of cases of
earworms are generally pleasant. Also, if images function as anticipatory
predictive processes (e.g., Neisser, 1976; Tian & Poepple, 2012), then
matching of musical imagery to subsequent music perception might result
in positive affect resulting from a successful prediction (cf. expectancy as a
contributor to emotion; Huron, 2006; Juslin & Västfjäll, 2008). Given that
perceived music might activate different cortical areas as a function of
whether that music is perceived as happy or sad (Khalfa et al., 2005;
Mitterschiffthaler, Fu, Dalton, Andrew, & Williams, 2007), analogous
patterns of cortical activation could be predicted during music imagery.
S C
R
Aleman, A., & van’t Wout, M. (2004). Subvocalization in auditory-verbal imagery: Just a form of
motor imagery? Cognitive Processing 5(4), 228–231.
Alossa, N., & Castelli, L. (2009). Amusia and musical functioning. European Neurology 61(5), 269–
277.
Baba, A., Hamada, H., & Koca, H. (2003). Musical hallucinations in schizophrenia. 2. Relations with
verbal hallucinations. Psychopathology 36(2), 104–110.
Baddeley, A. D. (1986). Working memory. New York: Oxford University Press.
Baddeley, A. D. (2000). The episodic buffer: A new component of working memory? Trends in
Cognitive Sciences 4(11), 417–423.
Baddeley, A. D., & Logie, R. H. (1992). Auditory imagery and working memory. In D. Reisberg
(Ed.), Auditory imagery (pp. 179–197). Hillsdale, NJ: Lawrence Erlbaum Associates.
Bangert, M., Peschel, T., Schlaug, G., Rotte, M., Drescher, D., Hinrichs, H., … Altenmüller, E.
(2006). Shared networks for auditory and motor processing in professional pianists: Evidence from
fMRI conjunction. NeuroImage 30(3), 917–926.
Baron-Cohen, S., & Harrison, J. E. (Eds.). (1997). Synaesthesia: Classic and contemporary readings.
Cambridge, MA: MIT Press/Blackwell.
Barsalou, L. W. (2008). Grounded cognition. Annual Review of Psychology 59, 617–645.
Baumann, S., Koeneke, S., Schmidt, C. F., Meyer, M., Lutz, K., & Jänke, L. A. (2007). A network for
audio-motor coordination in skilled pianists and non-musicians. Brain Research 1161, 65–78.
Beaty, R. E., Burgin, C. J., Nusbaum, E. C., Kwapil, T. R., Hodges, D. A., & Silvia, P. J. (2013).
Music to the inner ears: Exploring individual differences in musical imagery. Consciousness and
Cognition 22(4), 1163–1173.
Bernardini, F., Attademo, L., Blackmon, K., & Devinsky, O. (2017). Musical hallucinations: A brief
review of functional neuroimaging findings. CNS Spectrums 22(5), 397–403.
Berrios, G. E. (1991). Musical hallucinosis: A statistical analysis of 46 cases. Psychopathology 24(6),
356–360.
Besson, M., & Schön, D. (2003). Comparison between language and music. In I. Peretz & R. J.
Zatorre (Eds.), The cognitive neuroscience of music (pp. 269–293). New York: Oxford University
Press.
Bleich-Cohen, M., Hendler, T., Pashinian, A., Faragian, S., & Poyurovsky, M. (2011). Obsessive
musical hallucinations in a schizophrenic patient: Psychopathological and fMRI characteristics.
CNS Spectrums 16(7), 153–156.
Blood, A. J., & Zatorre, R. J. (2001). Intensely pleasurable responses to music correlate with activity
in brain regions implicated in reward and emotion. Proceedings of the National Academy of
Sciences 98(20), 11818–11823.
Bringas, M. L., Zaldivar, M., Rojas, P. A., Martinez-Montes, K., Chongo, D. M., Ortega, M. A., …
Valdes-Sosa, P. A. (2015). Effectiveness of music therapy as an aid to neurorestoration of children
with severe neurological disorders. Frontiers in Neuroscience 9, 427. Retrieved from
https://doi.org/10.3389/fnins.2015.00427
Brodsky, W., Henik, A., Rubenstein, B. S., & Zorman, M. (2003). Auditory imagery from musical
notation in expert musicians. Perception & Psychophysics 65(4), 602–612.
Brodsky, W., Kessler, Y., Rubenstein, B. S., Ginsborg, J., & Henik, A. (2008). The mental
representation of music notation: Notational audiation. Journal of Experimental Psychology:
Human Perception and Performance 34(2), 427–445.
Bryant, G. A. (2013). Animal signals and emotion in music: Coordinating affect across groups.
Frontiers in Psychology 4, 990. Retrieved from https://doi.org/10.3389/fpsyg.2013.00990
Cahn, D. (2008). The effects of varying ratios of physical and mental practice, and task difficulty on
performance of a tonal pattern. Psychology of Music 36, 179–191.
Cebrian, A. N., & Janata, P. (2010a). Electrophysiological correlates of accurate mental image
formation in auditory perception and imagery tasks. Brain Research 1342, 39–54.
Cebrian, A. N., & Janata, P. (2010b). Influences of multiple memory systems on auditory mental
image acuity. Journal of the Acoustical Society of America 127, 3189–3202.
Chanda, M. L., & Levitin, D. J. (2013). The neurochemistry of music. Trends in Cognitive Sciences
17(4), 179–193.
Cho, R., & Wu, W. (2013). Mechanisms of auditory verbal hallucination in schizophrenia. Frontiers
in Psychiatry 4, 155. Retrieved from https://doi.org/10.3389/fpsyt.2013.00155
Coebergh, J. A. F., Lauw, R. F., Bots, R., Sommer, I. E. C., & Blom, J. D. (2015). Musical
hallucinations: Review of treatment effects. Frontiers in Psychology 6, 814. Retrieved from
https://doi.org/10.3389/fpsyg.2015.00814
Cox, A. (2016). Music and embodied cognition. Bloomington, IN: Indiana University Press.
Crowder, R. G. (1989). Imagery for musical timbre. Journal of Experimental Psychology: Human
Perception and Performance 15(3), 472–478.
Cytowic, R. E. (2002). Synesthesia: A union of the senses (2nd ed.). Cambridge, MA: MIT Press.
Cytowic, R. E., & Eagleman, D. M. (2011). Wednesday is indigo blue: Discovering the brain of
synesthesia. Cambridge, MA: MIT Press.
Daselaar, S. M., Porat, Y., Huijbers, W., & Pennartz, C. M. (2010). Modality-specific and modality-
independent components of the human imagery system. NeuroImage 52(2), 677–685.
Davidson-Kelly, K., Schaeffer, R. S., Moran, N., & Overy, K. (2015). “Total Inner Memory”:
Deliberate uses of multimodal musical imagery during performance preparation.
Psychomusicology: Music, Mind and Brain 25(1), 83–92.
Day, S. A. (2016). Synesthetes: A handbook. CreateSpace Independent Publishing Platform.
Deroy, O., Fernandez-Prieto, I., Navarra, J., & Spence, C. (2018). Unraveling the paradox of spatial
pitch. In T. L. Hubbard (Ed.), Spatial biases in perception and cognition (pp. 77–93). New York:
Cambridge University Press.
Douglas, K. M., & Bilkey, D. K. (2007). Amusia is associated with deficits in spatial processing.
Nature Neuroscience 10(7), 915–921.
Driskell, J. E., Copper, C., & Moran, A. (1994). Does mental practice enhance music performance?
Journal of Applied Psychology 79(4), 481–492.
Eitan, Z., & Granot, R. Y. (2006). How music moves: Musical parameters and listeners’ images of
motion. Music Perception 23(3), 221–247.
Eitan, Z., & Timmers, R. (2010). Beethoven’s last piano sonata and those who follow crocodiles:
Cross-domain mappings of auditory pitch in a musical context. Cognition 114(3), 405–422.
Elkin, J., & Leuthold, H. (2011). The representation of pitch in auditory imagery: Evidence from S-R
compatibility and distance effects. Journal of Cognitive Psychology 23(1), 76–91.
Evans, C. L., McGuire, P. K., & David, A. S. (2000). Is auditory imagery defective in patients with
auditory hallucinations? Psychological Medicine 30(1), 137–148.
Evers, S. (2006). Musical hallucinations. Current Psychiatry Reports 8(3), 205–210.
Evers, S., & Ellger, T. (2004). The clinical spectrum of musical hallucinations. Journal of the
Neurological Sciences 227(1), 55–65.
Farrugia, N., Jakubowski, K., Cusack, R., & Stewart, L. (2015). Tunes stuck in your brain: The
frequency and affective evaluation of involuntary musical imagery correlate with cortical structure.
Consciousness and Cognition 35, 66–77.
Fraisse, P. (1982). Rhythm and tempo. In D. Deutsch (Ed.), The psychology of music (pp. 149–181).
New York: Academic Press.
Friberg, A., & Sundberg, J. (1999). Does music performance allude to locomotion? A model of final
ritardandi derived from measurements of stopping runners. Journal of the Acoustical Society of
America 105(3), 1469–1484.
Gabriel, D., Wong, T. C., Nicolier, M., Giustiniani, J., Mignot, C., Noiret, N., … Vandel, P. (2016).
Don’t forget the lyrics! Spatiotemporal dynamics of neural mechanisms spontaneously evoked by
gaps of silence in familiar and newly learned songs. Neurobiology of Learning and Memory 132,
18–28.
Galant-Swafford, J., & Bota, R. (2015). Musical hallucinations in schizophrenia. Mental Illness 7(1),
6065.
Gibbs, R. W. (2005). Embodiment and cognitive science. New York: Cambridge University Press.
Godøy, R. I. (2001). Imagined action, excitation, and resonance. In R. I. Godøy & H. Jørgensen
(Eds.), Musical imagery (pp. 237–250). New York: Taylor & Francis.
Greenwood, D. D. (1961). Critical bandwidth and the frequency coordinates of the basilar membrane.
Journal of the Acoustical Society of America 33, 1344–1356.
Griffiths, T. D. (2000). Musical hallucinosis in acquired deafness: Phenomenology and brain
substrate. Brain 123(10), 2065–2076.
Griffiths, T. D., Jackson, M. C., Spillane, J. A., Friston, K. J., & Frackowiak, R. S. J. (1997). A
neural substrate for musical hallucinosis. Neurocase 3(3), 167–172.
Grossenbacher, P. G., & Lovelace, C. T. (2001). Mechanisms of synesthesia: Cognitive and
physiological constraints. Trends in Cognitive Sciences 5(1), 36–41.
Halpern, A. R. (1988a). Mental scanning in auditory imagery for songs. Journal of Experimental
Psychology: Learning, Memory, and Cognition 14, 434–443.
Halpern, A. R. (1988b). Perceived and imaged tempos of familiar songs. Music Perception 6(2),
193–202.
Halpern, A. R. (1989). Memory for the absolute pitch of familiar songs. Memory & Cognition 17(5),
572–581.
Halpern, A. R. (2003). Cerebral substrates of musical imagery. In I. Peretz & R. Zatorre (Eds.), The
cognitive neuroscience of music (pp. 217–230). New York: Oxford University Press.
Halpern, A. R., & Bartlett, J. C. (2011). The persistence of musical memories: A descriptive study of
earworms. Music Perception 28(4), 425–432.
Halpern, A. R., & Zatorre, R. J. (1999). When that tune runs through your head: A PET investigation
of auditory imagery for familiar melodies. Cerebral Cortex 9(7), 697–704.
Halpern, A. R., Zatorre, R. J., Bouffard, M., & Johnson, J. A. (2004). Behavioral and neural
correlates of perceived and imagined musical timbre. Neuropsychologia 42(9), 1281–1292.
Hammeke, T. A., McQuillen, M. P., & Cohen, B. A. (1983). Musical hallucinations associated with
acquired deafness. Journal of Neurology, Neurosurgery, and Psychiatry 46(6), 570–572.
Haslinger, B., Erhard, P., Altenmüller, E., Schroeder, U., Boecker, H., & Ceballos-Baumann, A. O.
(2005). Transmodal sensorimotor networks during action observation in professional pianists.
Journal of Cognitive Neuroscience 17(2), 282–293.
Haueisen, J., & Knösche, T. (2001). Involuntary motor activity in patients evoked by music
perception. Journal of Cognitive Neuroscience 13(6), 786–792.
Hemming, J., & Merrill, J. (2015). On the distinction between involuntary musical imagery, musical
hallucinosis, and musical hallucinations. Psychomusicology: Music, Mind, and Brain 25(4), 435–
442.
Herholz, S. C., Halpern, A. R., & Zatorre, R. J. (2012). Neuronal correlates of perception, imagery,
and memory for familiar tunes. Journal of Cognitive Neuroscience 24(6), 1382–1397.
Herholz, S. C., Lappe, C., Knief, A., & Pantev, C. (2008). Neural basis of music imagery and the
effect of musical expertise. European Journal of Neuroscience 28(11), 2352–2360.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Highben, Z., & Palmer, C. (2004). Effects of auditory and motor mental practice in memorized piano
performance. Bulletin of the Council for Research in Music Education 159, 58–65.
Hodges, D. A. (2009). Bodily responses to music. In S. Hallam, I. Cross, & M. Thaut (Eds.), The
Oxford handbook of music psychology (2nd ed., pp. 183–196). New York: Oxford University
Press.
Hubbard, T. L. (2010). Auditory imagery: Empirical findings. Psychological Bulletin 136(2), 302–
329.
Hubbard, T. L. (2013a). Auditory aspects of auditory imagery. In S. Lacey & R. Lawson (Eds.),
Multisensory imagery (pp. 51–76). New York: Springer.
Hubbard, T. L. (2013b). Auditory imagery contains more than audition. In S. Lacey & R. Lawson
(Eds.), Multisensory imagery (pp. 221–247). New York: Springer.
Hubbard, T. L. (2017). Momentum in music: Musical succession as physical motion.
Psychomusicology: Music, Mind, and Brain 27(1), 14–30.
Hubbard, T. L. (2018). Some methodological and conceptual considerations in studies of auditory
imagery. Auditory Perception and Cognition 1, 6–41.
Hubbard, T. L. (forthcoming). Some anticipatory, kinesthetic, and dynamic aspects of auditory
imagery. In M. Grimshaw, M. Walther-Hansen, & M. Knakkergaard (Eds.), The Oxford handbook
of sound and imagination. New York: Oxford University Press.
Hubbard, T. L., & Stoeckig, K. (1988). Musical imagery: Generation of tones and chords. Journal of
Experimental Psychology: Learning, Memory, and Cognition 14, 656–667.
Huron, D. (2006). Sweet anticipation: Music and the psychology of expectation. Cambridge, MA:
MIT Press.
Iacoboni, M. (2009). Imitation, empathy, and mirror neurons. Annual Review of Psychology 60, 653–
670.
Jakubowski, K., Farrugia, N., & Stewart, L. (2016). Probing imagined tempo for music: Effects of
motor engagement and musical experience. Psychology of Music 44(6), 1274–1288.
Janata, P. (2001). Brain electrical activity evoked by mental formation of auditory expectations and
images. Brain Topography 13(3), 169–193.
Janata, P., & Paroo, K. (2006). Acuity of auditory images in pitch and time. Perception &
Psychophysics 68(5), 829–844.
Johns, L. C., Rossell, S., Frith, C., Ahmad, F., Hemsley, D., Kuipers, E., & McGuire, P. K. (2001).
Verbal self-monitoring and auditory verbal hallucinations in patients with schizophrenia.
Psychological Medicine 31(4), 705–715.
Johnson, M. L., & Larson, S. (2003). “Something in the way she moves” - Metaphors of musical
motion. Metaphor and Symbol 18(2), 63–84.
Juslin, P. N., & Sloboda, J. A. (Eds.). (2001). Music and emotion: Theory and research. New York:
Oxford University Press.
Juslin, P. N., & Västfjäll, D. (2008). Emotional responses to music: The need to consider underlying
mechanisms. Behavioral and Brain Sciences 31(5), 559–621.
Kantrowitz, J. T., Scaramello, N., Jakubovitz, A., Lehrfeld, J. M., Laukka, P., Elfenbein, H. A., …
Javitt, D. C. (2014). Amusia and protolanguage impairments in schizophrenia. Psychological
Medicine 44(13), 2739–2748.
Kasai, K., Asada, T., Yumoto, M., Takeya, J., & Matsuda, H. (1999). Evidence of functional
abnormality in the right auditory cortex during musical hallucinations. Lancet 354, 1703–1704.
Keller, P. E. (2012). Mental imagery in musical performance: Underlying mechanisms and potential
benefits. Annals of the New York Academy of Sciences 1252, 206–213.
Keller, P. E., Dalla Bella, S., & Koch, I. (2010). Auditory imagery shapes movement timing and
kinematics: Evidence from a musical task. Journal of Experimental Psychology: Human
Perception and Performance 36(2), 508–513.
Keshavan, M. S., Davis, A. S., Steingard, S., & Lishman, W. A. (1992). Musical hallucinosis: A
review and synthesis. Neuropsychiatry, Neuropsychology, & Behavioral Neurology 5(3), 211–223.
Khalfa, S., Schon, D., Anton, J.-L., & Liégeois-Chauvel, C. (2005). Brain regions involved in the
recognition of happiness and sadness in music. Neuroreport 16(18), 1981–1984.
Kleber, B., & Birbaumer, N., Veit, R., Trevorrow, T., & Lotze, M. (2007). Overt and imagined
singing of an Italian aria. NeuroImage 36(3), 889–900.
Koelsch, S. (2010). Towards a neural basis of music-evoked emotions. Trends in Cognitive Sciences
14(3), 131–137.
Koelsch, S. (2012). Brain and music. Cambridge, MA: Wiley-Blackwell.
Kohler, E., Keysers, C., Umiltà, M. A., Fogassi, L., Gallese, V., & Rizzolatti, G. (2002). Hearing
sounds, understanding actions: Action representation in mirror neurons. Science 297(5582), 846–
848.
Kraemer, D. J. M., Macrae, C. N., Green, A. E., & Kelley, W. M. (2005). Musical imagery: Sound of
silence activates auditory cortex. Nature 434, 158.
Kristeva, R., Chakarov, V., Schulte-Mönting, J., & Spreer, J. (2003). Activation of cortical areas in
music execution and imagining: A high-resolution EEG study. NeuroImage 20(3), 1872–1883.
Krumhansl, C. L., & Schenck, D. L. (1997). Can dance reflect the structural and expressive qualities
of music? A perceptual experiment on Balanchine’s choreography of Mozart’s Divertimento No.
15. Musicae Scientiae 1, 63–85.
Kumar, S., Sedley, W., Barnes, G. R., Teki, S., Friston, K. J., & Griffiths, T. D. (2014). A brain basis
for musical hallucinations. Cortex 52(100), 86–97.
Lakoff, G., & Johnson, M. (1980). Metaphors we live by. Chicago, IL: University of Chicago Press.
Lakoff, G., & Johnson, M. (1999). Philosophy in the flesh: The embodied mind and its challenge to
western thought. New York: Basic Books.
Langheim, F. J., Callicott, J. H., Mattay, V. S., Duyn, J. H., & Weinberger, D. R. (2002). Cortical
systems associated with covert music rehearsal. Neuroimage 16(4), 901–908.
Larson, S. (2012). Musical forces: Motion, metaphor and meaning in music. Bloomington, IN:
Indiana University Press.
Leaver, A. M., van Lare, J., Zielinski, B., Halpern, A. R., & Rauschecker, J. P. (2009). Brain
activation during anticipation of sound sequences. Journal of Neuroscience 29(8), 2477–2485.
Levitin, D. J. (2007). This is your brain on music: The science of a human obsession. New York:
Penguin Group.
Levitin, D. J., & Tirovolas, A. K. (2009). Current advances in the cognitive neuroscience of music.
Annals of the New York Academy of Sciences 1156, 211–231.
Liberman, A. M., & Mattingly, I. G. (1985). The motor theory of speech perception revised.
Cognition 21, 1–36.
Liégois-Chauvel, C., Peretz, I., Babaï, M., Laguitton, V., & Chauvel, P. (1998). Contribution of
different cortical areas in the temporal lobes in music processing. Brain 121(10), 1853–1867.
Lima, C. F., Krishnan, S., & Scott, S. K. (2016). Roles of supplementary motor areas in auditory
processing and auditory imagery. Trends in Neurosciences 39(8), 527–542.
Lima, C. F., Lavan, N., Evans, S., Agnew, Z., Halpern, A. R., Shanmugalingam, P., … Scott, S. K.
(2015). Feel the noise: Relating individual differences in auditory imagery to the structure and
function of sensorimotor systems. Cerebral Cortex 25(11), 4638–4650.
Linden, D. E. J., Thornton, K., Kuswanto, C. N., Johnston, S. J., van de Ven, V., & Jackson, M. C.
(2011). The brain’s voices: Comparing nonclinical auditory hallucinations and imagery. Cerebral
Cortex 21(2), 330–337.
Linke, A. C., & Cusack, R. (2015). Flexible information coding in human auditory cortex during
perception, imagery, and STM of complex sounds. Journal of Cognitive Neuroscience 27(7),
1322–1333.
Lotze, M. (2013). Kinesthetic imagery of musical performance. Frontiers in Human Neuroscience 7,
280. Retrieved from https://doi.org/10.3389/fnhum.2013.00280
Lotze, M., Scheler, G., Tan, H. R., Braun, C., & Birbaumer, N. (2003). The musician’s brain:
Functional imaging of amateurs and professionals during performance and imagery. Neuroimage
20(3), 1817–1829.
McGuire, P. K., Silbersweig, D. A., Wright, I., Murray, R. M., David, A. S., Frackowiak, R. S. J., &
Frith, C. D. (1995). Abnormal monitoring of inner speech: A physiological basis for auditory
hallucinations. Lancet 346, 596–600.
McGuire, P. K., Silbersweig, D. A., Murray, R. M., David, A. S., Frackowiak, R. S., & Frith, C. D.
(1996). Functional anatomy of inner speech and auditory verbal imagery. Psychological Medicine
26(1), 29–38.
Marin, O. S. M., & Perry, D. W. (1999). Neurological aspects of music perception and performance.
In D. Deutsch (Ed.), The psychology of music (2nd ed., pp. 653–724). New York: Academic Press.
Meister, I. G., Krings, T., Foltys, H., Boroojerdi, B., Müller, M., Töpper, R., & Thron, A. (2004).
Playing piano in the mind: An fMRI study on music imagery and performance in pianists.
Cognitive Brain Research 19(3), 219–228.
Mikumo, M. (1994). Motor encoding strategy for pitches of melodies. Music Perception 12(2), 175–
197.
Mitchell, R. W., & Gallaher, M. C. (2001). Embodying music: Matching music and dance in memory.
Music Perception 19(1), 65–85.
Mitterschiffthaler, M. T., Fu, C. H. Y., Dalton, J. A., Andrew, C. M., & Williams, S. C. R. (2007). A
functional MRI study of happy and sad affective states induced by classical music. Human Brain
Mapping 28(11), 1150–1162.
Neisser, U. (1976). Cognition and reality: Principles and implications of cognitive psychology. New
York: W. H. Freeman.
Nirkko, A. C., Baader, A. P., Loevblad, K.-O., Milani, P., & Wiesendanger, M. (2000). Cortical
representation of music production in violin players: Behavioral assessment and functional
imaging of finger sequencing, bimanual coordination and music specific brain activation.
NeuroImage 11(5), S106.
Oh, J., Kwon, J. H., Yang, P. S., & Jeong, J. (2013). Auditory imagery modulates frequency-specific
areas in the human auditory cortex. Journal of Cognitive Neuroscience 25(2), 175–187.
Oztop, E., Kawato, M., & Arbib, M. (2006). Mirror neurons and imitation: A computationally guided
review. Neural Networks 19(3), 254–271.
Parsons, L. M. (2003). Exploring the functional neuroanatomy of music performance, perception, and
comprehension. In I. Peretz & R. Zatorre (Eds.), The cognitive neuroscience of music (pp. 247–
268). New York: Oxford University Press.
Patel, A. D. (2008). Music, language, and the brain. New York: Oxford University Press.
Pearce, J. M. S. (2007). Synaesthesia. European Neurology 57(2), 120–124.
Pecenka, N., & Keller, P. E. (2009). Auditory pitch imagery and its relationship to musical
synchronization. Annals of the New York Academy of Sciences 1169, 282–286.
Penfield, W., & Perot, P. (1963). The brain’s record of auditory and visual experience. Brain 86(4),
595–696.
Peretz, I. (1990). Processing of local and global musical information by unilateral brain-damaged
patients. Brain 113(4), 1185–1205.
Peretz, I. (2002). Brain specialization for music. Neuroscientist 8(4), 374–382.
Peretz, I. (2013). The biological foundations of music: Insight from congenital amusia. In D. Deutsch
(Ed.). The psychology of music (3rd ed., pp. 551–564). New York: Academic Press.
Peretz, I., & Zatorre, R. J. (Eds.). (2003). The cognitive neuroscience of music. New York: Oxford
University Press.
Peretz, I., & Zatorre, R. J. (2005). Brain organization for music processing. Annual Review of
Psychology 56, 89–114.
Perrone-Capano, C., Volpicelli, F., & di Porzio, U. (2017). Biological bases of human musicality.
Review of Neuroscience 28(3), 235–245.
Petsche, H., von Stein, A., & Filz, O. (1996). EEG aspects of mentally playing an instrument.
Cognitive Brain Research 3(2), 115–123.
Platel, H., Price, C., Baron, J.-C., Wise, R., Lambert, J., Frackowiak, R. S. J., … Eustache, F. (1997).
The structural components of music perception: A functional anatomical study. Brain: A Journal
of Neurology 120(2), 229–243.
Prete, G., Marzoli, D., Brancucci, A., & Tommasi, L. (2016). Hearing it right: Evidence of
hemispheric lateralization in auditory imagery. Hearing Research 332, 80–86.
Ramachandran, V. S., & Hubbard, E. M. (2001). Synaesthesia: A window into perception, thought
and language. Journal of Consciousness Studies 8(12), 3–34.
Reisberg, D., Smith, J. D., Baxter, D. A., & Sonenshine, M. (1989). “Enacted” auditory images are
ambiguous; “pure” auditory images are not. Quarterly Journal of Experimental Psychology A:
Human Experimental Psychology 41(3), 619–641.
Reybrouck, M. (2001). Musical imagery between sensory processing and ideomotor simulation. In R.
I. Godøy & H. Jørgensen (Eds.), Musical imagery (pp. 117–135). New York: Taylor & Francis.
Rizzonelli, M., Kim, J. H., Gladow, T., & Mainka, S. (2017). Musical stimulation with feedback in
gait training for Parkinson’s disease. Psychomusicology: Music, Mind, and Brain 27, 213–218.
Robertson, L. C., & Sagiv, N. (Eds.). (2005). Synesthesia: Perspectives from cognitive neuroscience.
New York: Oxford University Press.
Rouw, R., & Scholte, H. S. (2010). Neural basis of individual differences in synesthetic experiences.
Journal of Neuroscience 30(18), 6205–6213.
Saba, P. R., & Keshavan, M. S. (1997). Musical hallucinations and musical imagery: Prevalence and
phenomenology in schizophrenic inpatients. Psychopathology 30, 185–190.
Sabaté, M., Llanos, C., & Rodriguez, M. (2008). Integration of auditory and kinesthetic information
in motion: Alterations in Parkinson’s disease. Neuropsychology 22(4), 462–468.
Salimpoor, V. N., Benovoy, M., Larcher, K., Dagher, A., &. Zatorre, R. J. (2011). Anatomically
distant dopamine release during anticipation and experience of peak emotion to music. Nature
Neuroscience 14, 257–262.
Samson, S., & Zatorre, R. J. (1991). Recognition memory for text and melody of songs after
unilateral temporal lobe lesion: Evidence for dual encoding. Journal of Experimental Psychology:
Learning, Memory, and Cognition 17(4), 793–804.
Särkämö, T., Altenmüller, E., Rodriguez-Fornells, A., & Peretz, I. (2016). Editorial. Music, brain,
and rehabilitation: Emerging therapeutic applications and potential neural mechanisms. Frontiers
in Human Neuroscience 10, 103. Retrieved from https://doi.org/10.3389/fnhum.2016.00103
Satoh, M. (2014). Musical processing in the brain: A neuropsychological approach through cases
with amusia. Austin Journal of Clinical Neurology 1(2), 1009.
Satoh, M., Takeda, K., Nagata, N., Hatazawa, J., & Kuzuhara, S. (2001). Activated brain regions in
musicians during an ensemble: A PET study. Cognitive Brain Research 12(1), 101–108.
Schaefer, R. S., Desain, P., & Suppes, P. (2009). Structural decomposition of EEG signatures of
melodic processing. Biological Psychology 82(3), 253–259.
Schaefer, R. S., Vlek, R. J., & Desain, P. (2011). Music perception and imagery in EEG: Alpha band
effects of task and stimulus. International Journal of Psychophysiology 82(3), 254–259.
Schellenberg, E. G., & Trehub, S. E. (2003). Good pitch memory is widespread. Psychological
Science 14(3), 262–266.
Schlaug, G. (2015). Musicians and music making as a model for the study of brain plasticity.
Progress in Brain Research 127, 37–55.
Schürmann, M., Raij, T., Fujiki, N., & Hari, R. (2002). Mind’s ear in a musician: Where and when in
the brain. NeuroImage 16(2), 434–440.
Senior, C., Barnes, J., Giampietroc, V., Simmons, A., Bullmore, E. T., Brammer, M., & David, A. S.
(2000). The functional neuroanatomy of implicit-motion perception or “representational
momentum.” Current Biology 10(1), 16–22.
Senior, C., Ward, J., & David, A. S. (2002). Representational momentum and the brain: An
investigation of the functional necessity of V5/MT. Visual Cognition 9(1), 81–92.
Shapiro, L. (2010). Embodied cognition. New York: Routledge.
Shergill, S. S., Bullmore, E., Simmons, A., Murray, R., & McGuire, P. (2000). Functional anatomy of
auditory verbal imagery in schizophrenic patients with auditory hallucinations. American Journal
of Psychiatry 157(10), 1691–1693.
Simoens, V. L., & Tervaniemi, M. (2013). Auditory short-term memory activation during score
reading. PLoS ONE 8(1), e53691.
Smith, J. D., Wilson, M., & Reisberg, D. (1995). The role of subvocalization in auditory imagery.
Neuropsychologia 33(11), 1433–1454.
Snowdon, C. T., Zimmerman, E., & Altenmüller, E. (2015). Music evolution and neuroscience.
Progress in Brain Research 217, 17–34.
Spiller, M. J., Jonas, C. N., Simner, J., & Jansari, A. (2015). Beyond visual imagery: How modality-
specific is enhanced mental imagery in synesthesia? Consciousness and Cognition 31, 73–85.
Stewart, L., von Kriegstein, K., Warren, J. D., & Griffiths, T. D. (2006). Music and the brain:
Disorders of musical listening. Brain 129(10), 2533–2553.
Thaut, M. H., McIntosh, G. C., Rice, R. R., Miller, R. A., Rathbun, J., & Brault, J. M. (1996).
Rhythmic auditory stimulation in gait training with Parkinson’s disease patients. Movement
Disorders 11(2), 193–200.
Theiler, A. M., & Lippman, L. G. (1995). Effects of mental practice and modeling on guitar and
vocal performance. Journal of General Psychology 122(4), 329–343.
Tian, X., & Poeppel, D. (2012). Mental imagery of speech: Linking motor and perceptual systems
through internal simulation and estimation. Frontiers in Human Neuroscience 6, 314. Retrieved
from https://doi.org/10.3389/fnhum.2012.00314
Tillmann, B., Jolicoeur, P., Ishihara, M., Gosselin, N., Bertrand, O., Rossetti, Y., & Peretz, I. (2010).
The amusic brain: Lost in music, but not in space. PLoS ONE 5(4), e10173.
van Dijk, H., Nieuwenhuis, I. L., & Jensen, O. (2010). Left temporal alpha band activity increases
during working memory retention of pitches. European Journal of Neuroscience 31(9), 1701–
1707.
Villena-González, M., López, V., & Rodríguez, E. (2016). Data of ERPs and spectral alpha power
when attention is engaged on visual or verbal/auditory imagery. Data in Brief 7, 882–888.
Vines, B. W., Krumhansl, C. L., Wanderley, M. M., & Levitin, D. J. (2006). Cross-modal interactions
in the perception of musical performance. Cognition 101(1), 80–113.
Vlek, R. J., Schaefer, R. S., Gielen, C. C. A. M., Farquhar, J. D. R., & Desain, P. (2011). Shared
mechanisms in perception and imagery of auditory accents. Clinical Neurophysiology 122(8),
1526–1532.
Vogt, S., Buccino, G., Wohlschlager, A. M., Canessa, N., Shah, N. J., Zilles, K., … Fink, G. R.
(2007). Prefrontal involvement in imitation learning of hand actions: Effects of practice and
expertise. NeuroImage 37(4), 1371–1383.
Vuvan, D. T., & Schmuckler, M. A. (2011). Tonal hierarchy representations in auditory imagery.
Memory & Cognition 39(3), 477–490.
Wan, C. Y., & Schlaug, G. (2013). Brain plasticity induced by musical training. In D. Deutsch (Ed.).
The psychology of music (3rd ed., pp. 565–581). New York: Academic Press.
Williams, T. I. (2015). The classification of involuntary musical imagery: The case for earworms.
Psychomusicology: Music, Mind, and Brain 25(1), 5–13.
Wilson, M. (2002). Six views of embodied cognition. Psychonomic Bulletin & Review 9(4), 625–636.
Wu, J., Mai, X., Chan, C. C. H., Zheng, Y., & Luo, Y. (2006). Event related potentials during mental
imagery of animal sounds. Psychophysiology 43(6), 592–597.
Wu, J., Mai, X., Yu, Z., Qin, S., & Luo, Y. (2010). Effects of discrepancy between imagined and
perceived sounds on the N2 component of the event-related potential. Psychophysiology 47(2),
289–298.
Wu, J., Yu, Z., Mai, X., Wei, J., & Luo, Y. (2011). Pitch and loudness information encoded in
auditory imagery as revealed by event-related potentials. Psychophysiology 48(3), 415–419.
Yoo, S. S., Lee, C. U., & Choi, B. G. (2001). Human brain mapping of auditory imagery: Event-
related functional MRI study. Neuroreport 12(14), 3045–3049.
Yumoto, M., Matsuda, M., Itoh, K., Uno, A., Karino, S., Siatoh, O., … Kaga, K. (2005). Auditory
imagery mismatch negativity elicited in musicians. Neuroreport 16(11), 1175–1178.
Zatorre, R. J. (1988). Pitch perception of complex tones and human temporal-lobe function. Journal
of the Acoustical Society of America 84(2), 566–572.
Zatorre, R. J., & Halpern, A. R. (1993). Effect of unilateral temporal-lobe excision on perception and
imagery of songs. Neuropsychologia 31(3), 221–232.
Zatorre, R. J., & Halpern, A. R. (2005). Mental concerts: Musical imagery and the auditory cortex.
Neuron 47(1), 9–12.
Zatorre, R. J., Halpern, A. R., & Bouffard, M. (2010). Mental reversal of imagined melodies: A role
for the posterior parietal cortex. Journal of Cognitive Neuroscience 22(4), 775–789.
Zatorre, R. J., Halpern, A. R., Perry, D. W., Meyer, E., & Evans, A. C. (1996). Hearing in the mind’s
ear: A PET investigation of musical imagery and perception. Journal of Cognitive Neuroscience
8(1), 29–46.
Zhang, Y., Chen, G., Wen, H., Lu, K. H., & Liu, Z. (2017). Musical imagery involves Wernicke’s
area in bilateral and anti-correlated network interactions in musicians. Scientific Reports 7(1),
17068.
1
Interestingly, a number of prominent composers are suspected of or have admitted to
experiencing synesthesia in which musical stimuli evoked different colors or other visual qualities
(e.g., Leonard Bernstein, Duke Ellington, Billy Joel, Franz Liszt, Oliver Messiaen, Nikolai Rimsky-
Korsakov, Alexander Scriabin). The greater prevalence of non-musical imagery (e.g., visual color)
triggered by a musical stimulus (e.g., pitch), coupled with the relative lack of musical imagery (e.g.,
pitch) triggered by a non-musical stimulus (e.g., visual color), is consistent with findings that
auditory stimuli evoke non-auditory qualities in a large percentage of synesthetes, but non-auditory
stimuli evoke auditory qualities in a very small percentage of synesthetes (Spiller, Jonas, Simner, &
Jansari, 2015). Also, it is not clear if the apparent lack of auditory imagery induced in synesthesia is
due to a limitation of synesthesia or to a bias in reporting.
CHAPT E R 22
NEUROPLASTICITY IN
MUSIC LEARNING
V E S A P U T K I N E N A N D MA R I T E RVA N I E MI
E M
E C S P
M
S E A
P
The auditory brainstem response (ABR) is another electrophysiological
measure that has been widely used to compare sound encoding between
musicians and non-musicians (Kraus & Chandrasekaran, 2010). The ABR
elicited by complex sounds like phonemes or musical interval is typically
characterized by transient series of peaks within the first few milliseconds
followed by sustained response that closely mimics the period features of
the sound. As the nomenclature suggests, the ABR is thought to originate
mainly from brainstem nuclei (Chandrasekaran & Kraus, 2010) although
there is emerging evidence for a cortical contribution to the response
(Coffey, Herholz, Chepesiuk, Baillet, & Zatorre, 2016; Coffey, Musacchia,
& Zatorre, 2017). The sustained portion of the ABR, termed the frequency
following response (FFR), preserves the spectro-temporal features of the
stimulus with high fidelity. Therefore, the FFR lends itself to the study of
encoding of sound features that are important for pitch and timbre
processing in music but also essential for differentiating speech sound.
Indeed, the majority of the studies employing the FFR in musicians and
non-musicians have been conducted using speech sound and indicate that
relative to musically untrained controls musicians show more robust coding
of the spectrum of speech stimuli or faster or stronger neural responses at
the very early stages of sound encoding (Kraus & Chandrasekaran, 2010;
Strait & Kraus, 2014). Furthermore, there is evidence to suggest that these
group differences are particularly pronounced in challenging listening
conditions such as in the presence of background noise (Coffey, Mogilever,
& Zatorre, 2017; Strait & Kraus, 2014). Such findings have raised hopes
that musical training could be used to alleviate problems of speech-in-noise
perception and other auditory processing deficits that can occur in language
and other neurological disorders as well as in normal aging (Alain, Zendel,
Hutka, & Bidelman, 2014; Skoe & Kraus, 2010).
The first of these studies used the FFR to investigate the encoding of
pitch contours in spoken Mandarin Chinese syllables (Wong, Skoe, Russo,
Dees, & Kraus, 2007). In Mandarin Chinese, the meaning of syllables is
dependent on the pitch contour and a previous study had shown that native
speakers of this language display enhanced pitch tracking as indexed by the
ABR (Krishnan, Gandour, Ananthakrishnan, & Vijayaraghavan, 2015). The
study by Wong et al. in turn found evidence for a more robust pitch tracking
in musicians than in non-musicians. As none of the subjects had previous
exposure to Mandarin, the results suggest a generalization of the enhanced
sound encoding in musicians to foreign speech sounds. A later study found
that musicians had enhanced ABRs to cello and spoken syllables (/da/) both
in the early transient portion of the ABR as well as in the FFR time window
(Musacchia, Sams, Skoe, & Kraus, 2007). Since these seminal studies, the
enhanced encoding of linguistic and non-linguistic sounds in musicians as
indexed by the ABR has been replicated numerous times (Strait & Kraus,
2014) also in children and adolescents (discussed below) as well as in aging
participants (Alain et al., 2014). Results from laboratory training in sound
identification indicate that pitch tracking of the ABR is boosted even by
short-term experience (Song, Skoe, Wong, & Kraus, 2008) and thereby
support (but obviously do not prove) the notion that the higher quality
sound representation in musicians may be attributable to experience.
In sum, ABR studies indicate that frequent engagement with musical
sounds might tune sound processing in the nuclei along the auditory
pathway to fine-grained acoustic information that is important for sound
processing in both music and speech. The mechanism underlying the ABR
enhancement in musicians is unclear, but has been speculated to be driven
by top-down influence through descending (cortico-fugal) pathways from
the cortex to the auditory brainstem (Kraus & Chandrasekaran, 2010). The
enhanced auditory skills of musicians indeed tend to be accompanied by
above-average performance in non-auditory tasks that tap into higher-order
cognitive processes such as executive functions (discussed below) which, in
some studies, have been found to correlate with the degree of enhancement
in ABR indices of sound encoding (e.g., Strait, Kraus, Parbery-Clark, &
Ashley, 2010).
M T D
B
T S P :
D M I E
F ?
C F D
R
Alain, C., Zendel, B. R., Hutka, S., & Bidelman, G. M. (2014). Turning down the noise: The benefit
of musical training on the aging auditory brain. Hearing Research 308, 162–173.
Bengtsson, S. L., Nagy, Z., Skare, S., Forsman, L., Forssberg, H., & Ullén, F. (2005). Extensive piano
practicing has regionally specific effects on white matter development. Nature Neuroscience 8(9),
1148–1150.
Bermudez, P., Lerch, J. P., Evans, A. C., & Zatorre, R. J. (2009). Neuroanatomical correlates of
musicianship as revealed by cortical thickness and voxel-based morphometry. Cerebral Cortex
19(7), 1583–1596.
Besson, M., & Faïta, F. (1995). An event-related potential (ERP) study of musical expectancy:
Comparison of musicians with nonmusicians. Journal of Experimental Psychology: Human
Perception and Performance 21(6), 1278–1296.
Besson, M., Faïta, F., & Requin, J. (1994). Brain waves associated with musical incongruities differ
for musicians and non-musicians. Neuroscience Letters 168(1), 101–105.
Bialystok, E., & DePape, A.-M. (2009). Musical expertise, bilingualism, and executive functioning.
Journal of Experimental Psychology: Human Perception and Performance 53(2), 565–574.
Brattico, E., Näätänen, R., & Tervaniemi, M. (2001). Context effects on pitch perception in musicians
and nonmusicians: Evidence from event-related-potential recordings. Music Perception: An
Interdisciplinary Journal 19(2), 199–222.
Brattico, E., Pallesen, K. J., Varyagina, O., Bailey, C., Anourova, I., Järvenpää, M., … Tervaniemi,
M. (2008). Neural discrimination of nonprototypical chords in music experts and laymen: An
MEG study. Journal of Cognitive Neuroscience 21(11), 2230–2244.
Buonomano, D. V., & Merzenich, M. M. (1998). Cortical plasticity: From synapses to maps. Annual
Review of Neuroscience 21, 149–186.
Chandrasekaran, B., & Kraus, N. (2010). The scalp-recorded brainstem response to speech: Neural
origins and plasticity. Psychophysiology 47(2), 236–246.
Chobert, J., François, C., Velay, J.-L., & Besson, M. (2012). Twelve months of active musical
training in 8- to 10-year-old children enhances the preattentive processing of syllabic duration and
voice onset time. Cerebral Cortex 24(4), 956–967.
Coffey, E. B. J., Herholz, S. C., Chepesiuk, A. M. P., Baillet, S., & Zatorre, R. J. (2016). Cortical
contributions to the auditory frequency-following response revealed by MEG. Nature
Communications 7, 11070. doi:10.1038/ncomms11070
Coffey, E. B., Mogilever, N., & Zatorre, R. J. (2017). Speech-in-noise perception in musicians: A
review. Hearing Research 352, 49–69.
Coffey, E. B., Musacchia, G., & Zatorre, R. J. (2017). Cortical correlates of the auditory frequency-
following and onset responses: EEG and fMRI evidence. Journal of Neuroscience 37(4), 830–838.
Costa-Giomi, E. (1999). The effects of three years of piano instruction on children’s cognitive
development. Journal of Research in Music Education 47(3), 198–212.
de Manzano, Ö., & Ullén, F. (2018). Same genes, different brains: Neuroanatomical differences
between monozygotic twins discordant for musical training. Cerebral Cortex 28(1), 387–394.
Diamond, A. (2013). Executive functions. Annual Review of Psychology 64, 135–168.
Drayna, D., Manichaikul, A., de Lange, M., Snieder, H., & Spector, T. (2001). Genetic correlates of
musical pitch recognition in humans. Science 291(5510), 1969–1972.
Elbert, T., Pantev, C., Wienbruch, C., Rockstroh, B., & Taub, E. (1995). Increased cortical
representation of the fingers of the left hand in string players. Science 270(5234), 305–307.
François, C., Chobert, J., Besson, M., & Schön, D. (2012). Music training for the development of
speech segmentation. Cerebral Cortex 23(9), 2038–2043.
Friedman, N. P., & Miyake, A. (2017). Unity and diversity of executive functions: Individual
differences as a window on cognitive structure. Cortex 86, 186–204.
Fujioka, T., Ross, B., Kakigi, R., Pantev, C., & Trainor, L. J. (2006). One year of musical training
affects development of auditory cortical-evoked fields in young children. Brain 129(10), 2593–
2608.
Fujioka, T., Trainor, L. J., Ross, B., Kakigi, R., & Pantev, C. (2004). Musical training enhances
automatic encoding of melodic contour and interval structure. Journal of Cognitive Neuroscience
16(6), 1010–1021.
Fujioka, T., Trainor, L. J., Ross, B., Kakigi, R., & Pantev, C. (2005). Automatic encoding of
polyphonic melodies in musicians and nonmusicians. Journal of Cognitive Neuroscience 17(10),
1578–1592.
Gaser, C., & Schlaug, G. (2003). Brain structures differ between musicians and non-musicians.
Journal of Neuroscience 23(27), 9240–9245.
George, E. M., & Coch, D. (2011). Music training and working memory: An ERP study.
Neuropsychologia 49(5), 1083–1094.
Habibi, A., Cahn, B. R., Damasio, A., & Damasio, H. (2016). Neural correlates of accelerated
auditory processing in children engaged in music training. Developmental Cognitive Neuroscience
21, 1–14.
Habibi, A., Damasio, A., Ilari, B., Veiga, R., Joshi, A. A., Leahy, R. M., … Damasio, H. (2017).
Childhood music training induces change in micro and macroscopic brain structure: Results from a
longitudinal study. Cerebral Cortex, 1–12. Retrieved from https://doi.org/10.1093/cercor/bhx286
Halwani, G. F., Loui, P., Rüber, T., & Schlaug, G. (2011). Effects of practice and experience on the
arcuate fasciculus: Comparing singers, instrumentalists, and non-musicians. Frontiers in
Psychology 2. Retrieved from https://doi.org/10.3389/fpsyg.2011.00156
Hannon, E. E., & Trainor, L. J. (2007). Music acquisition: Effects of enculturation and formal
training on development. Trends in Cognitive Sciences 11(11), 466–472.
Hansen, M., Wallentin, M., & Vuust, P. (2013). Working memory and musical competence of
musicians and non-musicians. Psychology of Music 41(6), 779–793.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
behavior, function, and structure. Neuron 76(3), 486–502.
Ho, Y. C., Cheung, M. C., & Chan, A. S. (2003). Music training improves verbal but not visual
memory: Cross-sectional and longitudinal explorations in children. Neuropsychology 17(3), 439–
450.
Hyde, K. L., Lerch, J., Norton, A., Forgeard, M., Winner, E., Evans, A. C., & Schlaug, G. (2009).
Musical training shapes structural brain development. Journal of Neuroscience 29(10), 3019–3025.
Janus, M., Lee, Y., Moreno, S., & Bialystok, E. (2016). Effects of short-term music and second-
language training on executive control. Journal of Experimental Child Psychology 144, 84–97.
Jaschke, A. C., Honing, H., & Scherder, E. J. (2018). Longitudinal analysis of music education on
executive functions in primary school children. Frontiers in Neuroscience 12, 103. Retrieved from
https://www.frontiersin.org/articles/10.3389/fnins.2018.00103
Jentschke, S., & Koelsch, S. (2009). Musical training modulates the development of syntax
processing in children. NeuroImage 47(2), 735–744.
Koelsch, S. (2009). Music-syntactic processing and auditory memory: Similarities and differences
between ERAN and MMN. Psychophysiology 46(1), 179–190.
Koelsch, S., Grossmann, T., Gunter, T. C., Hahne, A., Schröger, E., & Friederici, A. D. (2003).
Children processing music: Electric brain responses reveal musical competence and gender
differences. Journal of Cognitive Neuroscience 15(5), 683–693.
Koelsch, S., Gunter, T., Friederici, A. D., & Schröger, E. (2000). Brain indices of music processing:
“Nonmusicians” are musical. Journal of Cognitive Neuroscience 12(3), 520–541.
Koelsch, S., Schmidt, B.-H., & Kansok, J. (2002). Effects of musical expertise on the early right
anterior negativity: An event-related brain potential study. Psychophysiology 39(5), 657–663.
Koelsch, S., Schröger, E., & Tervaniemi, M. (1999). Superior pre-attentive auditory processing in
musicians. Neuroreport 10(6), 1309–1313.
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11(8), 599–605.
Kraus, N., Slater, J., Thompson, E. C., Hornickel, J., Strait, D. L., Nicol, T., & White-Schwoch, T.
(2014). Music enrichment programs improve the neural encoding of speech in at-risk children.
Journal of Neuroscience 34(36), 11913–11918.
Krishnan, A., Gandour, J. T., Ananthakrishnan, S., & Vijayaraghavan, V. (2015). Language
experience enhances early cortical pitch-dependent responses. Journal of Neurolinguistics 33,
128–148.
Kühnis, J., Elmer, S., Meyer, M., & Jäncke, L. (2013). The encoding of vowels and temporal speech
cues in the auditory cortex of professional musicians: An EEG study. Neuropsychologia 51(8),
1608–1618.
Lappe, C., Herholz, S. C., Trainor, L. J., & Pantev, C. (2008). Cortical plasticity induced by short-
term unimodal and multimodal musical training. Journal of Neuroscience 28(39), 9632–9639.
Linnavalli, T., Putkinen, V., Lipsanen, J., Huotilainen, M., & Tervaniemi, M. (2018). Music
playschool enhances children’s linguistic skills. Scientific Reports 8(1), 8767.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001). Musical syntax is processed in
Broca’s area: An MEG study. Nature Neuroscience 4(5), 540–545.
Magne, C., Schön, D., & Besson, M. (2006). Musician children detect pitch violations in both music
and language better than nonmusician children: Behavioral and electrophysiological approaches.
Journal of Cognitive Neuroscience 18(2), 199–211.
Marie, C., Kujala, T., & Besson, M. (2012). Musical and linguistic expertise influence pre-attentive
and attentive processing of non-speech sounds. Cortex 48(4), 447–457.
Menning, H., Roberts, L. E., & Pantev, C. (2000). Plastic changes in the auditory cortex induced by
intensive frequency discrimination training. Neuroreport 11(4), 817–822.
Meyer, M., Elmer, S., Ringli, M., Oechslin, M. S., Baumann, S., & Jäncke, L. (2011). Long-term
exposure to music enhances the sensitivity of the auditory system in children. European Journal of
Neuroscience 34(5), 755–765.
Moradzadeh, L., Blumenthal, G., & Wiseheart, M. (2015). Musical training, bilingualism, and
executive function: A closer look at task switching and dual-task performance. Cognitive Science
39(5), 992–1020.
Moreno, S., Bialystok, E., Barac, R., Schellenberg, E. G., Cepeda, N. J., & Chau, T. (2011). Short-
term music training enhances verbal intelligence and executive function. Psychological Science
22(11), 1425–1433.
Moreno, S., & Bidelman, G. M. (2014). Examining neural plasticity and cognitive benefit through
the unique lens of musical training. Hearing Research 308, 84–97.
Moreno, S., Lee, Y., Janus, M., & Bialystok, E. (2015). Short-term second language and music
training induces lasting functional brain changes in early childhood. Child Development 86(2),
394–406.
Moreno, S., Marques, C., Santos, A., Santos, M., & Besson, M. (2009). Musical training influences
linguistic abilities in 8-year-old children: More evidence for brain plasticity. Cerebral Cortex
19(3), 712–723.
Mosing, M. A., Madison, G., Pedersen, N. L., Kuja-Halkola, R., & Ullén, F. (2014). Practice does not
make perfect: No causal effect of music practice on music ability. Psychological Science 25(9),
1795–1803.
Musacchia, G., Sams, M., Skoe, E., & Kraus, N. (2007). Musicians have enhanced subcortical
auditory and audiovisual processing of speech and music. Proceedings of the National Academy of
Sciences 104(40), 15894–15898.
Näätänen, R., Paavilainen, P., Rinne, T., & Alho, K. (2007). The mismatch negativity (MMN) in
basic research of central auditory processing: A review. Clinical Neurophysiology 118(12), 2544–
2590.
Näätänen, R., Tervaniemi, M., Sussman, E., Paavilainen, P., & Winkler, I. (2001). “Primitive
intelligence” in the auditory cortex. Trends in Neurosciences 24(5), 283–288.
Nan, Y., Liu, L., Geiser, E., Shu, H., Gong, C. C., Dong, Q., … Desimone, R. (2018). Piano training
enhances the neural processing of pitch and improves speech perception in Mandarin-speaking
children. Proceedings of the National Academy of Sciences 115(28), E6630–E6639.
Paavilainen, P. (2013). The mismatch-negativity (MMN) component of the auditory event-related
potential to violations of abstract regularities: A review. International Journal of Psychophysiology
88(2), 109–123.
Pallesen, K. J., Brattico, E., Bailey, C. J., Korvenoja, A., Koivisto, J., Gjedde, A., & Carlson, S.
(2010). Cognitive control in auditory working memory is enhanced in musicians. PloS ONE 5(6),
e11120.
Pantev, C., Oostenveld, R., Engelien, A., Ross, B., Roberts, L. E., & Hoke, M. (1998). Increased
auditory cortical representation in musicians. Nature 392(6678), 811–814.
Pantev, C., Roberts, L. E., Schulz, M., Engelien, A., & Ross, B. (2001). Timbre-specific
enhancement of auditory cortical representations in musicians. Neuroreport 12(1), 169–174.
Paraskevopoulos, E., Kraneburg, A., Herholz, S. C., Bamidis, P. D., & Pantev, C. (2015). Musical
expertise is related to altered functional connectivity during audiovisual integration. Proceedings
of the National Academy of Sciences 112(40), 12522–12527.
Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., & Pantev, C. (2012a). Evidence for training-
induced plasticity in multisensory brain structures: An MEG study. PloS ONE 7(5), e36534.
Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., & Pantev, C. (2012b). Musical expertise
induces audiovisual integration of abstract congruency rules. Journal of Neuroscience 32(50),
18196–18203.
Patel, A. D. (2011). Why would musical training benefit the neural encoding of speech? The OPERA
hypothesis. Frontiers in Psychology 2, 142. doi:10.3389/fpsyg.2011.00142
Peper, J. S., Brouwer, R. M., Boomsma, D. I., Kahn, R. S., & Hulshoff Pol, H. E. (2007). Genetic
influences on human brain structure: A review of brain imaging studies in twins. Human Brain
Mapping 28(6), 464–473.
Praag, H. van, Kempermann, G., & Gage, F. H. (2000). Neural consequences of environmental
enrichment. Nature Reviews Neuroscience 1(3), 191–198.
Putkinen, V., Tervaniemi, M., Saarikivi, K., de Vent, N., & Huotilainen, M. (2014). Investigating the
effects of musical training on functional brain development with a novel melodic MMN paradigm.
Neurobiology of Learning and Memory 110, 8–15.
Putkinen, V., Tervaniemi, M., Saarikivi, K., Ojala, P., & Huotilainen, M. (2014). Enhanced
development of auditory change detection in musically trained school-aged children: A
longitudinal event-related potential study. Developmental Science 17(2), 282–297.
Rinne, T., Alho, K., Ilmoniemi, R. J., Virtanen, J., & Näätänen, R. (2000). Separate time behaviors of
the temporal and frontal mismatch negativity sources. NeuroImage 12(1), 14–19.
Rüsseler, J., Altenmüller, E., Nager, W., Kohlmetz, C., & Münte, T. F. (2001). Event-related brain
potentials to sound omissions differ in musicians and non-musicians. Neuroscience Letters 308(1),
33–36.
Saarikivi, K., Putkinen, V., Tervaniemi, M., & Huotilainen, M. (2016). Cognitive flexibility
modulates maturation and music-training-related changes in neural sound discrimination.
European Journal of Neuroscience 44(2), 1815–1825.
Sachs, M., Kaplan, J., Der Sarkissian, A., & Habibi, A. (2017). Increased engagement of the
cognitive control network associated with music training in children during an fMRI Stroop task.
PloS ONE 12(10), e0187254.
Sala, G., & Gobet, F. (2017a). Does far transfer exist? Negative evidence from chess, music, and
working memory training. Current Directions in Psychological Science 26(6), 515–520.
Sala, G., & Gobet, F. (2017b). When the music’s over: Does music skill transfer to children’s and
young adolescents’ cognitive and academic skills? A meta-analysis. Educational Research Review
20, 55–67.
Schlaug, G., Jäncke, L., Huang, Y., Staiger, J. F., & Steinmetz, H. (1995). Increased corpus callosum
size in musicians. Neuropsychologia 33(8), 1047–1055.
Schneider, P., Scherg, M., Dosch, H. G., Specht, H. J., Gutschalk, A., & Rupp, A. (2002).
Morphology of Heschl’s gyrus reflects enhanced activation in the auditory cortex of musicians.
Nature Neuroscience 5(7), 688–694.
Schönwiesner, M., Novitski, N., Pakarinen, S., Carlson, S., Tervaniemi, M., & Näätänen, R. (2007).
Heschl’s gyrus, posterior superior temporal gyrus, and mid-ventrolateral prefrontal cortex have
different roles in the detection of acoustic changes. Journal of Neurophysiology 97(3), 2075–2082.
Schulze, K., Mueller, K., & Koelsch, S. (2011). Neural correlates of strategy use during auditory
working memory in musicians and non-musicians. European Journal of Neuroscience 33(1), 189–
196.
Schulze, K., Zysset, S., Mueller, K., Friederici, A. D., & Koelsch, S. (2011). Neuroarchitecture of
verbal and tonal working memory in nonmusicians and musicians. Human Brain Mapping 32(5),
771–783.
Seesjärvi, E., Särkämö, T., Vuoksimaa, E., Tervaniemi, M., Peretz, I., & Kaprio, J. (2016). The nature
and nurture of melody: A twin study of musical pitch and rhythm perception. Behavior Genetics
46(4), 506–515.
Shahin, A., Roberts, L. E., & Trainor, L. J. (2004). Enhancement of auditory cortical development by
musical experience in children. Neuroreport 15(12), 1917–1921.
Skoe, E., & Kraus, N. (2010). Auditory brainstem response to complex sounds: A tutorial. Ear and
Hearing 31(3), 302–324.
Sluming, V., Barrick, T., Howard, M., Cezayirli, E., Mayes, A., & Roberts, N. (2002). Voxel-based
morphometry reveals increased gray matter density in Broca’s area in male symphony orchestra
musicians. NeuroImage 17(3), 1613–1622.
Song, J. H., Skoe, E., Wong, P. C., & Kraus, N. (2008). Plasticity in the adult human auditory
brainstem following short-term linguistic training. Journal of Cognitive Neuroscience 20(10),
1892–1902.
Strait, D. L., & Kraus, N. (2014). Biological impact of auditory expertise across the life span:
Musicians as a model of auditory learning. Hearing Research 308(Suppl. C), 109–121.
Strait, D. L., Kraus, N., Parbery-Clark, A., & Ashley, R. (2010). Musical experience shapes top-down
auditory mechanisms: Evidence from masking and auditory attention performance. Hearing
Research 261(1), 22–29.
Tervaniemi, M., Castaneda, A., Knoll, M., & Uther, M. (2006). Sound processing in amateur
musicians and nonmusicians: Event-related potential and behavioral indices. Neuroreport 17(11),
1225–1228.
Tervaniemi, M., Janhunen, L., Kruck, S., Putkinen, V., & Huotilainen, M. (2015). Auditory profiles
of classical, jazz, and rock musicians: Genre-specific sensitivity to musical sound features.
Frontiers in Psychology 6. Retrieved from
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4703758/
Tervaniemi, M., Just, V., Koelsch, S., Widmann, A., & Schröger, E. (2005). Pitch discrimination
accuracy in musicians vs. nonmusicians: An event-related potential and behavioral study.
Experimental Brain Research 161(1), 1–10.
Tervaniemi, M., Kruck, S., De Baene, W., Schröger, E., Alter, K., & Friederici, A. D. (2009). Top-
down modulation of auditory processing: Effects of sound context, musical expertise and
attentional focus. European Journal of Neuroscience 30(8), 1636–1642.
Tervaniemi, M., Rytkönen, M., Schröger, E., Ilmoniemi, R. J., & Näätänen, R. (2001). Superior
formation of cortical memory traces for melodic patterns in musicians. Learning & Memory 8(5),
295–300.
Tierney, A. T., Krizman, J., & Kraus, N. (2015). Music training alters the course of adolescent
auditory development. Proceedings of the National Academy of Sciences 112(32), 10062–10067.
Tierney, A. T., Krizman, J., Skoe, E., Johnston, K., & Kraus, N. (2013). High school music classes
enhance the neural processing of speech. Frontiers in Psychology 4. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00855
Titz, C., & Karbach, J. (2014). Working memory and executive functions: Effects of training on
academic achievement. Psychological Research 78(6), 852–868.
Ullén, F., Hambrick, D. Z., & Mosing, M. A. (2016). Rethinking expertise: A multifactorial gene–
environment interaction model of expert performance. Psychological Bulletin 142(4), 427–446.
Ullén, F., Mosing, M. A., & Madison, G. (2015). Associations between motor timing, music practice,
and intelligence studied in a large sample of twins. Annals of the New York Academy of Sciences
1337, 125–129.
Van Beijsterveldt, C. E. M., & Van Baal, G. C. M. (2002). Twin and family studies of the human
electroencephalogram: A review and a meta-analysis. Biological Psychology 61(1), 111–138.
van Zuijen, T. L., Sussman, E., Winkler, I., Näätänen, R., & Tervaniemi, M. (2004). Grouping of
sequential sounds: An event-related potential study comparing musicians and nonmusicians.
Journal of Cognitive Neuroscience 16(2), 331–338.
van Zuijen, T. L., Sussman, E., Winkler, I., Näätänen, R., & Tervaniemi, M. (2005). Auditory
organization of sound sequences by a temporal or numerical regularity: A mismatch negativity
study comparing musicians and non-musicians. Brain Research: Cognitive Brain Research 23(2–
3), 270–276.
Vinkhuyzen, A. A., Van der Sluis, S., Posthuma, D., & Boomsma, D. I. (2009). The heritability of
aptitude and exceptional talent across different domains in adolescents and young adults. Behavior
Genetics 39(4), 380–392.
Virtala, P., Huotilainen, M., Putkinen, V., Makkonen, T., & Tervaniemi, M. (2012). Musical training
facilitates the neural discrimination of major versus minor chords in 13-year-old children.
Psychophysiology 49(8), 1125–1132.
Vuust, P., Brattico, E., Seppänen, M., Näätänen, R., & Tervaniemi, M. (2012). The sound of music:
Differentiating musicians using a fast, musical multi-feature mismatch negativity paradigm.
Neuropsychologia 50(7), 1432–1443.
Vuust, P., Pallesen, K. J., Bailey, C., van Zuijen, T. L., Gjedde, A., Roepstorff, A., & Østergaard, L.
(2005). To musicians, the message is in the meter: Pre-attentive neuronal responses to incongruent
rhythm are left-lateralized in musicians. NeuroImage 24(2), 560–564.
Winkler, I., Denham, S. L., & Nelken, I. (2009). Modeling the auditory scene: Predictive regularity
representations and perceptual objects. Trends in Cognitive Sciences 13(12), 532–540.
Wong, P. C. M., Skoe, E., Russo, N. M., Dees, T., & Kraus, N. (2007). Musical experience shapes
human brainstem encoding of linguistic pitch patterns. Nature Neuroscience 10(4), 420–422.
Zuk, J., Benjamin, C., Kenyon, A., & Gaab, N. (2014). Behavioral and neural correlates of executive
functioning in musicians and non-musicians. PloS ONE 9(6), e99868.
SECTION VI
D E V E L OP ME N TA L
IS S U E S IN MU S IC A N D
T HE B R A IN
CHAPT E R 23
A N T H O N Y B R A N D T, MO L LY G E B R I A N, A N D L . R O B E RT
SLEVC
M L P
A B
In societies across the globe, getting children conversant with their native
tongue as fast as possible is a universal goal. Thus, children’s ability to
learn language is an essential constraint on its structure. As Deacon (1997,
p. 110) writes, “The structure of a language is under intense selection
pressure because in its reproduction from generation to generation, it must
pass through a narrow bottleneck: children’s minds.” Do language’s musical
features help to facilitate this process?
Despite their inability to speak and understand language, newborns
display an impressive sensitivity to a variety of linguistic contrasts. This
sensitivity has been often cited as evidence that the ability to learn language
is innate (e.g., Eimas, Siqueland, Jusczyk, & Vigorito, 1971; Vouloumanos
& Werker, 2007). However, infants are most responsive to the sounds of
words, not to their meaning: they are drawn first to the musical aspects of
language.
For instance, as they encounter the world for the first time, newborns are
famously able to discriminate the phonemes of all languages (Dehaene-
Lambertz & Dehaene, 1994; Eimas et al., 1971; Werker & Tees, 1984). As
discussed above, this reflects a sensitivity to (vocal) timbre. Although less
research has been done to probe musical timbre perception in newborns, 3-
to 4-day-old infants can organize auditory streams on the basis of timbre,
similar to adults (McAdams & Bertoncini, 1997). Older infants remain
highly sensitive to timbre: 6-month-olds have long-term memory for the
timbre of folk songs, and 7- to 8.5-month-old infants can differentiate tones
that differ only in their spectral structure (Trainor, Wu, & Tsang, 2004;
Trehub, Endman, & Thorpe, 1990). Timbre appears to be so salient to
infants that it can actually affect their ability to recognize and discriminate
other basic features of music and speech (for summaries, see Costa-Giomi
& Davila, 2014; Creel, 2016). For instance, infants take longer to learn
words spoken by different speakers than when they are spoken by just one
person (Jusczyl, Pisoni, & Mullennix, 1992). Trainor and colleagues (2004)
familiarized infants with a melody played on a single instrument and found
that the infants could not recognize this same melody when played on a
different instrument. The salience of timbre extends through preschool:
when asked to associate sound to visual stimuli, preschoolers do so more
readily with timbral contrasts than with pitch contours (Creel, 2016). Even
adults show memory facilitation dependent on timbre (e.g., Halpern &
Müllensiefen, 2008; Radvansky & Potter, 2000).
Newborns are also sensitive to the rhythmic features of language: they
distinguish between languages of different rhythmic classes—stress-timed
or syllable-timed—whether or not the contrast includes their native
language (Nazzi, Bertoncini, & Mehler, 1998). Although newborns prefer
their native language (Moon, Cooper, & Fifer, 1993), this seems to reflect a
preference for the rhythmic class (stress patterns) of their native language: it
is not until 4 months of age that infants can reliably tell the difference
between languages of the same rhythmic class (Bosch & Sebastian-Galles,
1997; Gervain & Mehler, 2010; Nazzi et al., 1998).
In addition to being sensitive to linguistic timbre and rhythm, infants can
also discriminate the characteristic prosody (or melody) of their native
language (Friederici, 2006) and even show evidence of discriminating
affective prosody in the first two days of life (Cheng, Lee, Chen, Wang, &
Decety, 2012). Research on infant cries also sheds light on the importance
of melodic abilities in linguistic development: over the first few months of
life, melodic complexity of crying increases (Wermke & Mende, 2009) and
infants who do not show this increasing melodic complexity show poorer
language performance two years later (Wermke, Leising, & Stellzig-
Eisenhauer, 2007). The “melody” of infants’ cries reflects the prosody of
their native language, further support for the idea that infants are sensitive
to the musical aspects of the language to which they are most exposed
(Mampe, Friederici, Christophe, & Wermke, 2009; Prochnow, Erlandsson,
Hesse, & Wermke, 2017).
Additional evidence that the musical features of language are salient to
infants comes from the way we talk to them: in “baby talk” or motherese,
melodic contours are exaggerated, speech is slower and rhythmic stresses
are emphasized. Researchers have debated why we talk to babies that way.
Some have argued that the function of motherese is limited to emotional
communication (Trainor, Austin, & Desjardins, 2000). Others propose that,
in addition to communicating emotion, baby talk also engages infants’
attention (Fernald, 1989): for instance, a parent’s pitch contours vary
depending on whether the child is smiling and/or maintaining eye contact or
not (Stern, Spieker, & MacKain, 1982). Across a variety of languages,
mothers modulate their vocal timbre in similar ways when speaking to
infants versus adults, perhaps as a way to draw their infants’ attention—
again highlighting the salience timbre has for infants (Piazza, Iordan, &
Lew-Williams, 2017). Other researchers emphasize the didactic role of
motherese, observing that mothers lengthen the vowels in content words
and exaggerate word and sentence boundaries (Kuhl et al., 1997; Saint-
Georges et al., 2013).
The functions of motherese may certainly change in the course of
development: as Saint-Georges and colleagues write (Saint-Georges et al.,
2013, p. 9), “Mothers adjust their infant-directed speech to infants’ age,
cognitive abilities and linguistic level.” The universality of motherese (Falk,
2004; Fernald, 1992) highlights how crucial it is to the caregiver–child
relationship. Infants are drawn to the musical features of speech—and that
attraction helps them to engage socially and to learn.
Various other evidence illuminates the extensive sensitivity infants have
to the musical sounds of language. For instance, newborns can use different
patterns of lexical stress to discriminate individual words (Sansavini,
Bertoncini, & Giovanelli, 1997); can use acoustic cues that signal word
boundaries (Christophe, Dupoux, Bertoncini, & Mehler, 1994); can
distinguish content words from function words based on their different
acoustic characteristics (Shi, Werker, & Morgan, 1999); and appear to be
sensitive to the prosodic boundaries in sentences (Pannekamp, Weber, &
Friederici, 2006). A long-held debate in language acquisition research is
how infants solve the so-called “bootstrapping problem”—how infants
connect sounds to meaning. It may be that infants use the musical aspects of
language (timbre, rhythm, melody) as the scaffolding on which they hang
their later developments in semantic and syntactic comprehension. Infants
are listening for how their language is composed and using this musical
information to support later linguistic developments.
Still, early life is the hardest period to study and the evidence is at times
contradictory, especially when it comes to timbre perception: for instance,
infants born ten weeks premature can discriminate a ba/da contrast, but
have more difficulty discriminating two different speakers
(Mahmoudzadeh, Wallois, Kongolo, Goudjil, & Dehaene-Lambertz, 2016),
and infants can recognize different phones before they can recognize
different voices (Dehaene-Lambertz, 2017). This discrepancy highlights the
importance of continuing to probe the aural abilities of newborns.
If the precocious discrimination abilities we see in newborns are
domain-general and not limited to language, we would expect to find
similarly precocious music perception abilities. Indeed, young infants have
shown that they have very fine-grained pitch discrimination abilities, being
able to distinguish pitch contrasts as small as a third of a half-step (Olsho,
Schoon, Sakai, Turpin, & Sperduto, 1982). Newborns can also detect a
deviant pitch, even when the timbre of the pitches is also changing (Háden
et al., 2009), and can extract pitch patterns and use them in a predictive
manner (Háden, Németh, Török, & Winkler, 2015)—further evidence of
their advanced pitch-processing abilities.
In the rhythmic domain, newborns can detect the beat in music and can
detect when an important rhythmic event is removed from a repeating drum
pattern (Winkler, Háden, Ladinig, Sziller, & Honing, 2009). Newborns can
also distinguish between small changes (60–100 ms) in the length of two
tones (Čėponiené et al., 2002; Cheour et al., 2002). At two months of age,
they can discriminate an isochronous sequence of tones from a non-
isochronous sequence (Demany, McKenzie, & Vurpillot, 1977). Far more
research has been done on the speech perception abilities than the music
perception abilities of newborns and very young infants, but existing
evidence suggests that infants’ music perception abilities are every bit as
sensitive as their language perception abilities at birth. This further supports
the idea that these are domain-general sound processing capabilities, not
unique to either music or language.
C - M
L P 6 12
M A
C - M
L C
Parallels between music and language development do not stop after the
first year of life. Indeed, they continue throughout childhood, until
children’s musical and linguistic sensitivity reaches adult levels (see Fig. 2).
One challenge in comparing music and language development after infancy
is that, whereas linguistic ability is often measured against the general
population, musical ability is often (implicitly) measured against the
expertise of professional musicians. This has contributed to the idea that
language ability is an innate skill all typically developing humans possess,
whereas musical skill is due to “talent” or a “gift” that is slower to mature.
Although it certainly takes a tremendous amount of hard work and
dedication to master the viola or the trumpet, acquiring the musical
conventions of your native culture is no more difficult or slow than learning
your native language.
FIGURE 2. Parallel development in music and language milestones from 2 to 12 years. Regular
text denotes parallel development. Italics denote related, but not analogous development. See main
text for references. (1) Two-year-olds can repeat brief, sung phrases with identifiable rhythm and
contour. (2) Eighteen-month-olds produce two-word utterances; 2-year-olds tend to eliminate
function words, but not content words. (3) Two-year-olds show basic knowledge of word order
constraints. (4) Three-year-olds have some knowledge of key membership and harmony and sing
“outline songs.” (5) Four- to six-year olds show knowledge of scale and key membership and detect
changes more easily in diatonic melodies than in non-diatonic ones. Five-year-olds show a typical
electrophysiological response to unexpected chords (the early right anterior negativity, or ERAN),
but do not detect a melodic change that implies a change in harmony. (6) At 5 years, processing of
function words depends on semantic context and brain activation is not function-specific for
semantic vs. syntactic processing (unlike adults). (7) Six-year-olds are able to speak in complete,
well-formed sentences. (8) Seven-year-olds have a knowledge of Western tonal structure comparable
to adults’ and can detect melodic changes that imply a change in harmony. (9) Only after 10 years of
age do children show adult-like electrophysiological responses to syntactic errors (Hahne, Eckstein,
& Friederici, 2006).
Adapted from Frontiers in Psychology 3, p. 327, Figure 2, Anthony Brandt, Molly Gebrian,
and L. Robert Slevc, Music and early language acquisition, doi: 10.3389/fpsyg.2012.00327,
© 2012 Brandt, Gebrian, and Slevc. Reprinted under the terms of the Creative Commons
Attribution License, which permits unrestricted use, distribution, and reproduction in any
medium, provided the original author and source are credited.
I N B S
L P ?
Despite the evidence above, many have claimed that music and language
are separate systems, and are processed as such by the brain (e.g., Peretz &
Coltheart, 2003). Some researchers claim that this separation is innate based
on evidence that infants show left hemisphere lateralization for speech at
birth (e.g., Dehaene-Lambertz, Dehaene, & Hertz-Pannier, 2002). However,
a close look at this research reveals a more complicated picture. Bilateral
activation for both music and language has been found in a number of
studies, ranging from those that include full sentences and/or musical
phrases (Fava, Hull, Baumbauer, & Bortfeld, 2014; Perani, 2012), to those
that use tightly controlled speech-like sounds that test specific parameters
of the linguistic acoustic signal (Dehaene-Lambertz, 2000; Kotilahti et al.,
2010; Minagawa-Kawai, Cristià, Vendelin, Cabrol, & Dupoux, 2011a).
Fava and colleagues (2014) used full sentences and musical phrases and
found no differences in activation. Perani and colleagues (2011) found right
lateralization for speech that closely parallels the activation they observed
to music in an earlier study (Perani et al., 2010).
Researchers who argue that music and language processing are separate
in the brain often point to rapid temporal processing as evidence for this
separation (e.g., Boemio, Fromm, Braun, & Poeppel, 2005; Zatorre &
Belin, 2001). The argument is that rapid temporal processing is required for
language comprehension, but not for music perception. However, as
described earlier, the perception of musical timbre requires processing at the
same rate as phonemes: for instance, in a percussion solo, rapid temporal
processing is at least as salient as slower melodic processing. Nevertheless,
even taken at face value, this research has produced conflicting results.
While the right hemisphere is believed to respond more strongly to
spectral information, the left hemisphere shows greater sensitivity to rapid
temporal contrasts, which underlies the supposed innate predisposition for
language perception. As noted earlier, Mahmoudzadeh and colleagues
(2016) found that infants born 10 weeks preterm could distinguish
consonants differing along a temporal continuum (specifically, a ba/da
contrast varying in voice onset time and so requiring rapid temporal
processing), but not between different speakers (whose voices presumably
differ mostly in spectral, not temporal, dimensions). Because the infants
were born so prematurely, the authors argued that this is evidence for an
early advantage for the processing of fast changes in the acoustic signal that
is likely genetic. However, an earlier study on premature infants by this
same research group (Mahmoudzadeh et al., 2013) found right hemisphere
activation increases to both a change in phoneme and a change in speaker (a
left hemisphere increase was also observed, but only for the phoneme
change). Minagawa-Kawai and colleagues (2011a) tested sounds that
differed in only their temporal composition or only their spectral
composition and found no evidence for hemispheric asymmetries; in fact,
both hemispheres were equally activated in the temporal condition.
Dehaene-Lambertz, who has long argued that music and language are
separate and that language perception is innate and prioritized in the infant
brain, has recently acknowledged that the brain activation to language in
young infants is complex and the leftward preference sometimes seen for
linguistic stimuli may have more to do with the fact that temporally
responsive neurons on the left mature earlier, rather than reflecting anything
special about language per se (Dehaene-Lambertz, 2017).
What is more fully supported by the research, at least to date, is the idea
that functional specialization emerges as a result of exposure and learning.
In 2011, Minagawa-Kawai and colleagues proposed such a developmental
scenario, arguing that
language acquisition recruits several specialized (but not necessarily domain-specific)
learning subsystems … The establishment of feature based, categorical phonetic units in the
extraction of words and rules on the basis of hierarchical and adjacent regularities requires
specific learning algorithms that are especially efficient in the left hemisphere and, as a
result, speech perception comes to be left-lateralized as a function of experience.
(Minagawa-Kawai, Cristià, & Dupoux, 2011b, p. 219)
This hypothesis predicts that second language learning in adults will be less
left-lateralized as a function of proficiency, which is exactly what is found
(Dehaene et al., 1997; Perani et al., 1996, 1998). In addition, sound pairs
only elicit asymmetrical activation when they constitute a contrast in the
speaker’s native language (Dehaene-Lambertz, 1997; Minagawa-Kawa,
Mori, & Sato, 2005; Näätänen et al., 1997), but after extensive training with
a non-native contrast, there is a shift to left-hemisphere dominance (Best &
Avery, 1999; Zhang et al., 2009). Gervain (2015, p. 16) echoes this idea:
“Features of the native language are processed in an increasingly lateralized
fashion in a network of focal brain areas, as processing turns from
acoustic/auditory to linguistic in nature, whereas non-native sound patterns
are handled in a more distributed and more bilateral way.”
L D D
Other evidence for the inseparability of music and language during early
life comes from linked deficits in individuals with abnormal linguistic or
musical development. For example, the speech perception and reading
deficits associated with developmental dyslexia have often been linked to
underlying problems with auditory processing, including processing of
rapid temporal changes in speech (for a review, see Hämäläinen, Salminen,
& Leppänen, 2013). Given the discussion above, it is unsurprising that
these deficits impact music as well. In particular, children with dyslexia
tend to also show deficits with musical rhythm (e.g., synchronizing with a
metronome) and (while not as often studied) musical timbre (e.g.,
Goswami, Huss, Mead, Fosker, & Verney, 2013; Huss, Verney, Fosker,
Mead, & Goswami, 2011; Overy, Nicolson, Fawcett, & Clarke, 2003).
Assuming a crucial link between musical and linguistic development, one
might imagine that musical experience could help treat dyslexia. In fact,
musical (especially rhythm) training can help remediate linguistic deficits in
dyslexia (e.g., Flaugnacco et al., 2015), although music training is clearly
no panacea given that dyslexic musicians show good musical abilities
despite pronounced reading deficits (Weiss, Granot, & Ahissar, 2014).
Note, however, that dyslexic musicians do show other types of perceptual
and auditory deficits including discrimination of amplitude envelope cues
(which relies on rapid temporal processing) and auditory working memory
(Weiss et al., 2014; Zuk et al., 2017). Thus developmental dyslexia appears
to be a deficit not only of language but also of music.
Similar parallels emerge in other purportedly language-specific deficits
such as specific language impairment (SLI). Individuals with SLI, whose
primary deficit involves syntactic processing, also show deficits in the
processing of musical (harmonic) structure (Jentschke, Koelsch, Sallat, &
Friederici, 2008), deficits in rhythmic processing (e.g., Cumming, Wilson,
Leong, Colling, & Goswami, 2015), and show more accurate grammatical
processing following rhythmic stimulation in a priming paradigm (Bedoin,
Brisseau, Molinier, Roch, & Tillmann, 2016; Przybylski et al., 2013).
Production deficits in SLI are also associated with productive deficits in
music—specifically, in pitch-matching and melody reproduction (Clément,
Planchou, Béland, Motte, & Samson, 2015).
Even deficits in language production such as developmental stuttering
have been linked to deficits with musical rhythm (Wieland, McAuley,
Dilley, & Chang, 2015), and suggestive relationships between language and
music deficits appear in other disorders not linked specifically to language.
For example, autistic children often have age-appropriate responses to
music but suffer from language disabilities. This has been viewed as
evidence that music and language must involve innately distinct neural
networks, with the music network healthy and the language one impaired.
However, in an fMRI study of autistic children, Lai and colleagues (Lai,
Pantazatos, Schneider, & Hirsch, 2012, p. 961) found that, “paradoxically,
brain regions associated with these functions typically overlap”: while song
and speech activated the same networks, the response to song was vigorous
whereas that to speech was subdued (Lai et al., 2012). And while the causes
are not yet fully understood, there is growing evidence for the efficacy of
musical engagement and therapy on language (and other) outcomes in
autism spectrum disorders (e.g., Geretsegger, Elefant, Mössler, & Gold,
2014).
Of course, the prediction is not just that linguistic deficits show
concomitant musical processing problems, but also that developmental
deficits in music should have consequences for language development.
Congenital amusia, the most well-studied deficit of musical development, is
primarily associated with deficits in pitch perception and/or pitch memory,
although can also include deficits in the processing of temporal (rhythmic)
aspects of music (reviews: Peretz & Hyde, 2003; Tillman, Albouy, &
Caclin, 2015). While early conceptions of amusia suggested a deficit
specific to music, more recent work has found related deficits in processing
of linguistic pitch, both for lexical tones (in tone-language speakers; e.g.,
Liu, Patel, Fourcin, & Stewart, 2010; Wang & Peng, 2014), and for the
recognition of emotional prosody in non-tone languages (Thompson, Marin,
& Stewart, 2012). Speech perception deficits in congenital amusia appear to
extend to non-pitch-based aspects of language as well (e.g., Liu, Jiang,
Wang, Xu, & Patel, 2015), showing subtle, but widespread, linguistic
consequences of congenital amusia.
One deficit that severely impacts both music and speech processing is
deafness. Given that deaf individuals can successfully learn full-fledged
sign languages, deafness could be seen as a notable counterexample for
many of the claims made here. Note, however, that music is not just an
auditory stimulus, but also a kinesthetic and visual one, and the rhythmic
“babbling” characteristic of sign-exposed infants (e.g., Petitto, Holowka,
Sergio, & Ostry, 2001) may lay the foundation for the temporal processing
abilities underlying later linguistic acquisition in signers.
E M L
A
It can be difficult for adults to conceive of music and language being treated
as one and the same in the infant brain because they are so obviously
different once we reach maturity. However, this entanglement between
music and language persists throughout our lives. Research on the
perception of tunes and lyrics shows evidence of integrated processing at
the pre-lexical, phonemic processing stage in adults (Sammler et al., 2010).
Sine-wave speech, which initially sounds like meaningless whistles, can not
only be perceived as speech after training, but it also activates speech areas
(specifically the left posterior superior temporal sulcus) once it is perceived
as speech (Mottonen et al., 2006). Silbo Gomero, a whistled speech used on
La Gomera in the Canary Islands, activates areas of the brain normally
associated with spoken language perception in proficient whistlers, but not
in those who do not speak the language (Carreiras, Lopez, Rivero, &
Corina, 2005).
Something as abstract as syntax may also be rooted in the music of
language. Kreiner and Eviatar (2014) argue that syntactic comprehension is
rooted in prosody: syntactic and prosodic boundaries largely correspond in
spoken language and this helps aid syntactic comprehension (see also
Heffner & Slevc, 2015). They note that prosody helps disambiguate unclear
syntactic structures (such as garden-path sentences), which can be
misleading in print, but rarely in spoken conversation. As argued above, this
melody of language is what infants first use, but even as adults, it continues
to underlie our syntactic understanding. There is also evidence for a
correlation between rhythm processing and word stress processing in
normal adults, another entanglement of music and language (Hausen,
Torppa, Salmela, Vainio, & Särkämö, 2013).
Patel (2012) has proposed a framework for the adult brain in which music
and language share neural resources for those perceptual and cognitive
tasks that they share in common. Given the evidence for infants’
sensitivities to the musical features of speech, the co-development of
musical and linguistic abilities, and shared developmental disorders, it
seems plausible that music and language are even more deeply entangled in
the newborn brain and that modularity emerges in the course of
development. Speech may be privileged in the infant brain (Shultz,
Vouloumanos, Bennett, & Pelphrey, 2014), but it is first experienced as a
vocal performance whose musical features are what engage the newborn’s
attention. It is only later that the child begins to apprehend the referential
function of words, and the music of words begins to sink into the
background. As Deutsch’s “looped speech” experiment shows, the music of
speech is never really absent, even in a non-tonal language—it is just that
we pay less attention to it. Music—and by extension, poetry—may give
human culture ways to creatively engage in the features of our aural
imaginations that conversational speech does not prioritize.
A burning question of human cognition is whether language is innate.
Are we language animals, with a universal grammar encoded in our genes
(Chomsky, 1980)? Or do we have a strong biological need to communicate
—and an aptitude for learning how to do so? The jury is still out, but there
has been a gradual shift toward viewing language as a cultural inheritance
rather than a genetic one. Iterated learning—in which initially random data
becomes coherent over time as generations of subjects instruct one another
—is a plausible way of describing the emergence of both language and
music (Kirby, Griffiths, & Smith, 2014; Ravignani, Delgado, & Kirby,
2016; Smith, Kirby, & Brighton, 2003). The study of music’s role in early
language acquisition may have important implications: when we learn
language, the music of speech comes first, thereby providing a key
mechanism by which language is transmitted from generation to generation.
In a newborn’s first months, speech is akin to bebop: musical attention to
how languages are composed—through their unique phonemic inventory
and prosody—helps an infant born into any community learn its native
tongue. The same acuities used in these early developmental stages—
sensitivities to timbre, pitch, and rhythm, and the ability to recognize their
consistencies—embody musical aptitude later in life. Early language
acquisition thus lies at the crossroads of music and language and provides
tantalizing glimpses into what it means to be human.
R
Albin, D. D., & Echols, C. H. (1996). Stressed and word-final syllables in infant-directed speech.
Infant Behavior and Development 19(4), 401–418.
Batuman, E. (2012). Talking drums. The New Yorker, July 9. Retrieved from
https://www.newyorker.com/culture/culture-desk/talking-drums
Bedoin, N., Brisseau, L., Molinier, P., Roch, D., & Tillmann, B. (2016). Temporally regular musical
primes facilitate subsequent syntax processing in children with specific language impairment.
Frontiers in Neuroscience 10. Retrieved from https://doi.org/10.3389/fnins.2016.00245
Bergeson, T. R., & Trehub, S. E. (2006). Infants’ perception of rhythmic patterns. Music Perception
23(4), 345–360.
Best, C. T., & Avery, R. A. (1999). Left-hemisphere advantage for click consonants is determined by
linguistic significance and experience. Psychological Science 10(1), 65–70.
Boemio, A., Fromm, S., Braun, A., & Poeppel, D. (2005). Hierarchical and asymmetric temporal
sensitivity in human auditory cortices. Nature Neuroscience 8(3), 389–395.
Bosch, L., & Sebastián-Gallés, N. (1997). Native-language recognition abilities in 4-month-old
infants from monolingual and bilingual environments. Cognition 65(1), 33–69.
Brandt, A., Gebrian, M., & Slevc, L. R. (2012). Music and early language acquisition. Frontiers in
Psychology 3. Retrieved from https://doi.org/10.3389/fpsyg.2012.00327
Brauer, J., & Friederici, A. D. (2007). Functional neural networks of semantic and syntactic
processes in the developing brain. Journal of Cognitive Neuroscience 19(10), 1609–1623.
Carreiras, M., Lopez, J., Rivero, F., & Corina, D. (2005). Linguistic perception: Neural processing of
a whistled language. Nature 433(7021), 31–32.
Čėponiené, R., Kushnerenko, E., Fellman, V., Renlund, M., Suominen, K., & Näätänen, R. (2002).
Event-related potential features indexing central auditory discrimination by newborns. Cognitive
Brain Research 13(1), 101–113.
Cheng, Y., Lee, S. Y., Chen, H. Y., Wang, P. Y., & Decety, J. (2012). Voice and emotion processing in
the human neonatal brain. Journal of Cognitive Neuroscience 24(6), 1411–1419.
Cheour, M., Čėponiené, R., Lehtokoski, A., Luuk, A., Allik, J., Alho, K., & Näätänen, R. (1998).
Development of language-specific phoneme representations in the infant brain. Nature
Neuroscience 1, 351–353.
Cheour, M., Čėponiené, R., Leppänen, P., Alho, K., Kujala, T., Renlund, M., & Näätänen, R. (2002).
The auditory sensory memory trace decays rapidly in newborns. Scandinavian Journal of
Psychology 43(1), 33–39.
Chomsky, N. (1980). Rules and representations. New York: Columbia University Press.
Christophe, A., Dupoux, E., Bertoncini, J., & Mehler, J. (1994). Do infants perceive word
boundaries? An empirical study of the bootstrapping of lexical acquisition. Journal of the
Acoustical Society of America 95(3), 1570–1580.
Clément, S., Planchou, C., Béland, R., Motte, J., & Samson, S. (2015). Singing abilities in children
with Specific Language Impairment (SLI). Frontiers in Psychology 6. Retrieved from
https://doi.org/10.3389/fpsyg.2015.00420
Corrigall, K. A., & Trainor, L. J. (2009). Effects of musical training on key and harmony perception.
Annals of the New York Academy of Sciences 1169, 164–168.
Costa-Giomi, E. (2003). Young children’s harmonic perception. Annals of the New York Academy of
Sciences 999, 477–484.
Costa-Giomi, E., & Davila, Y. (2014). Infants’ discrimination of female singing voices. International
Journal of Music Education 32(3), 324–332.
Creel, S. C. (2016). Ups and downs in auditory development: Preschoolers’ sensitivity to pitch
contour and timbre. Cognitive Science 40(2), 373–403.
Cross, I. (2001). Music, cognition, culture, and evolution. Annals of the New York Academy of
Sciences 930, 28–42.
Cumming, R., Wilson, A., Leong, V., Colling, L. J., & Goswami, U. (2015). Awareness of rhythm
patterns in speech and music in children with specific language impairments. Frontiers in Human
Neuroscience 9. Retrieved from https://doi.org/10.3389/fnhum.2015.00672
Davidson, L. (1994). Song singing by young and old: A developmental approach to music. In R.
Aiello with J. Sloboda (Eds.), Musical perceptions (pp. 99–130). New York: Oxford University
Press.
Deacon, T. W. (1997). The symbolic species: The coevolution of language and the brain. New York:
W. W. Norton.
Dehaene, S., Dupoux, E., Mehler, J., Cohen, L., Paulesu, E., Perani, D., & Le Bihan, D. (1997).
Anatomical variability in the cortical representation of first and second language. NeuroReport
8(17), 3809–3815.
Dehaene-Lambertz, G. (1997). Electrophysiological correlates of categorical phoneme perception in
adults. NeuroReport 8(4), 919–924.
Dehaene-Lambertz, G. (2000). Cerebral specialization for speech and non-speech stimuli in infants.
Journal of Cognitive Neuroscience 12(3), 449–460.
Dehaene-Lambertz, G. (2017). The human infant brain: A neural architecture able to learn language.
Psychonomic Bulletin and Review 24(1), 48–55.
Dehaene-Lambertz, G., & Dehaene, S. (1994). Speed and cerebral correlates of syllable
discrimination in infants. Nature 370, 1–4.
Dehaene-Lambertz, G., Dehaene, S., & Hertz-Pannier, L. (2002). Functional neuroimaging of speech
perception in infants. Science 298(5600), 2013–2015.
Demany, L., McKenzie, B., & Vurpillot, E. (1977). Rhythm perception in early infancy. Nature
266(5604), 718–719.
Deutsch, D., Henthorn, T., & Lapidis, R. (2011). Illusory transformation from speech to song.
Journal of the Acoustical Society of America 129(4), 2245–2252
Dowling, W. J. (1999). The development of music perception and cognition. In D. Deutsch (Ed.), The
Psychology of Music (2nd ed.; pp. 603–625). London: Academic Press.
Eimas, P., Siqueland, E., Jusczyk, P., & Vigorito, J. (1971). Speech perception in infants. Science
171(3968), 202–206.
Falk, D. (2004). Prelinguistic evolution in early hominins: Whence motherese? Behavioral and Brain
Sciences 27(4), 491–503; discussion: 503–583.
Fava, E., Hull, R., Baumbauer, K., & Bortfeld, H. (2014). Hemodynamic responses to speech and
music in preverbal infants. Child Neuropsychology 20(4), 430–448.
Fernald, A. (1989). Intonation and communicative intent in mothers’ speech to infants: Is the melody
the message? Child Development 60(6), 1497–1510.
Fernald, A. (1992). Human maternal vocalizations to infants as biologically relevant signals: An
evolutionary perspective. In J. H. Barkow, L. Cosmides, & J. Tooby (Eds.), The adapted mind:
Evolutionary psychology and the generation of culture (pp. 391–428). Oxford: Oxford University
Press.
Fernald, A., & Mazzie, C. (1991). Prosody and focus in speech to infants and adults. Developmental
Psychology 27(2), 209–221.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schön, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A randomized
control trial. PLoS ONE 10(9), e0138715.
Fleming, K. (2013). Five surprising facts about Richie’s classic “All Night Long.” New York Post,
September 21. Retrieved from http://nypost.com/2013/09/21/five-surprising-facts-about-richies-
classic-all-night-long/
Forsyth, M. (2012). The etymologicon: A circular stroll through the hidden connections of the
English language. New York: Berkley Books.
Fox, C. (1990). Steve Reich’s “Different Trains.” Tempo, New Series, No. 172 (March), 2–8.
Friederici, A. D. (1983). Children’s sensitivity to function words during sentence comprehension.
Linguistics 21, 717–739.
Friederici, A. D. (2006). The neural basis of language development and its impairment. Neuron
52(6), 941–952.
Geretsegger, M., Elefant, C., Mössler, K. A., & Gold, C. (2014). Music therapy for people with
autism spectrum disorder. Cochrane Database of Systematic Reviews 17(6), CD004381.
doi:10.1002/14651858.CD004381.pub3
Gerken, L., Landau, B., & Remez, R. (1990). Function morphemes in young children’s speech
perception and production. Developmental Psychology 26(2), 204–216.
Gervain, J. (2015). Plasticity in early language acquisition: The effects of prenatal and early
childhood experience. Current Opinion in Neurobiology 35, 13–20.
Gervain, J., & Mehler, J. (2010). Speech perception and language acquisition in the first year of life.
Annual Review of Psychology 61, 191–218.
Gervain, J., Nespor, M., Mazuka, R., Horie, R., & Mehler, J. (2008). Bootstrapping word order in
prelexical infants: A Japanese–Italian cross-linguistic study. Cognitive Psychology 57(1), 56–74.
Goswami, U., Huss, M., Mead, N., Fosker, T., & Verney, J. P. (2013). Perception of patterns of
musical beat distribution in phonological developmental dyslexia: Significant longitudinal
relations with word reading and reading comprehension. Cortex 49(5), 1363–1376.
Háden, G. P., Németh, R., Török, M., & Winkler, I. (2015). Predictive processing of pitch trends in
newborn infants. Brain Research 1626, 14–20.
Háden, G. P., Stefanics, G., Vestergaard, M. D., Denham, S. L., Sziller, I., & Winkler, I. (2009).
Timbre-independent extraction of pitch in newborn infants. Psychophysiology 46(1), 69–74.
Hahne, A., Eckstein, K., & Friederici, A. D. (2006). Brain signatures of syntactic and semantic
processes during children’s language development. Brain 16(7), 1302–1318.
Halpern, A. R., & Müllensiefen, D. (2008). Effects of timbre and tempo change on memory for
music. Quarterly Journal of Experimental Psychology 61(9), 1371–1384.
Hämäläinen, J. A., Salminen, H. K., & Leppänen, P. H. (2013). Basic auditory processing deficits in
dyslexia: Systematic review of the behavioral and event-related potential/field evidence. Journal of
Learning Disabilities 46(5), 413–427.
Hannon, E. E., & Trehub, S. E. (2005a). Metrical categories in infancy and adulthood. Psychological
Science 16(1), 48–55.
Hannon, E. E., & Trehub, S. E. (2005b). Tuning in to musical rhythms: Infants learn more readily
than adults. Proceedings of the National Academy of Sciences 102(35), 12639–12643.
Hargreaves, D. J. (1996). The development of artistic and musical competence. In I. Deliege & J.
Sloboda (Eds.), Musical beginnings (pp. 145–170). Oxford: Oxford University Press.
Harich-Schneider, E. (1954). The rhythmical patterns in Gagaku and Bugaku. Leiden: E. J. Brill.
Hausen, M., Torppa, R., Salmela, V. R., Vainio, M., & Särkämö, T. (2013). Music and speech
prosody: A common rhythm. Frontiers in Psychology 4. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00566
Heffner, C. C., & Slevc, L. R. (2015). Prosodic structure as a parallel to musical structure. Frontiers
in Psychology 6. Retrieved from https://doi.org/10.3389/fpsyg.2015.01962
Hochmann, J. R., Endress, A. D., & Mehler, J. (2010). Word frequency as a cue for identifying
function words in infancy. Cognition 115(3), 444–457.
Höhle, B., Weissenborn, J., Schmitz, M., & Ischebeck, A. (2001). Discovering word order
regularities: The role of prosodic information for early parameter setting. In J. Weissenborn & B.
Höhle (Eds.), Approaches to bootstrapping: Phonological, lexical, syntactic and
neurophysiological aspects of early language acquisition (pp. 249–265). Amsterdam: John
Benjamins.
Howard, D. M., Angus, J. A., & Welch, G. F. (1994). Singing pitching accuracy from years 3 to 6 in
a primary school. Proceedings of the Institute of Acoustics 16(5), 223–230.
Hugo, V. (1864). William Shakespeare. Paris: A. Lacroix, Verboeckhoeven et Cie.
Hukin, R. W., & Darwin, C. J. (1995). Comparison of the effect of onset asynchrony on auditory
grouping in pitch matching and vowel identification. Perception and Psychophysics 57(2), 191–
196.
Huss, M., Verney, J. P., Fosker, T., Mead, N., & Goswami, U. (2011). Music, rhythm, rise time
perception and developmental dyslexia: Perception of musical meter predicts reading and
phonology. Cortex 47(6), 674–689.
Jackendoff, R. (2009). Parallels and nonparallels between language and music. Music Perception
26(3), 195–204.
Jentschke, S., Koelsch, S., Sallat, S., & Friederici, A. D. (2008). Children with specific language
impairment also show impairment of music-syntactic processing. Journal of Cognitive
Neuroscience 20(11), 1940–1951.
Jusczyk, P. W. (2000). The discovery of spoken language. Cambridge, MA: MIT Press.
Jusczyk, P. W., Hohne, E. A., & Bauman, A. (1999). Infants’ sensitivity to allophonic cues to word
segmentation. Perception and Psychophysiology 61(8), 1465–1476.
Jusczyl, P. W., Pisoni, D. B., & Mullennix, J. (1992). Some consequences of stimulus variability on
speech processing by 2-month-old infants. Cognition 43(3), 253–291.
Kirby, S., Griffiths, T., & Smith, K. (2014). Iterated learning and the evolution of language. Current
Opinion in Neurobiology 28, 108–114.
Koelsch, S., Grossmann, T., Gunter, T. C., Hahne, A., Schröger, E., & Friederici, A. D. (2003).
Children processing music: Electric brain responses reveal musical competence and gender
differences. Journal of Cognitive Neuroscience 15(5), 683–693.
Kostelanetz, R. (1987). Notes on Milton Babbitt as text-sound artist. Perspectives of New Music
25(1/2), 280–284.
Kotilahti, K., Nissilä, I., Näsi, T., Lipiäinen, L., Noponen, T., Meriläinen, P., … Fellman, V. (2010).
Hemodynamic responses to speech and music in newborn infants. Human Brain Mapping 31(4),
595–603.
Kreiner, H., & Eviatar, Z. (2014). The missing link in the embodiment of syntax: Prosody. Brain &
Language 137, 91–102.
Kreutzer, N. (2001). Song acquisition among from rural Shona-speaking Zimbabwean children from
birth to 7 years. Journal of Research in Music Education 49(3), 198–211.
Kuhl, P. K. (2010). Brain mechanisms in early language acquisition. Neuron 67(5), 713–727.
Kuhl, P. K., Andruski, J. E., Chistovich, I., Chistovich, L. A., Kozhevnikova, E. V., Ryskina, V. L., …
Lacerda, F. (1997). Cross-language analysis of phonetic units in language addressed to infants.
Science 277(5326), 684–686.
Kuhl, P. K., Williams, K., Lacerda, F., Stevens, K., & Lindblom, B. (1992). Linguistic experience
alters phonetic perception in infants by 6 months of age. Science 255(5044), 606–608.
Lai, G., Pantazatos, S. P., Schneider, H., & Hirsch, J. (2012). Neural systems for speech and song in
autism. Brain 135(3), 961–975.
Liu, F., Jiang, C., Wang, B., Xu, Y., & Patel, A. D. (2015). A music perception disorder (congenital
amusia) influences speech comprehension. Neuropsychologia 66, 111–118.
Liu, F., Patel, A. D., Fourcin, A., & Stewart, L. (2010). Intonation processing in congenital amusia:
Discrimination, identification and imitation. Brain 133(6), 1682–1693.
Lynch, M. P., & Eilers, R. E. (1992). A study of perceptual development for musical tuning.
Perception and Psychophysics 52(6), 599–608.
Lynch, M. P., Eilers, R. E., Kimbrough Oller, D., & Urbano, R. C. (1990). Innateness, experience,
and music perception. Psychological Science 1(4), 272–276.
McAdams, S., & Bertoncini, J. (1997). Organization and discrimination of repeating sound sequences
by newborn infants. Journal of the Acoustical Society of America 102(5), 2945–2953.
McMullen, E., & Saffran, J. R. (2004). Music and language: A developmental comparison. Music
Perception 21(3), 289–311.
McWhorter, J. (2015). The world’s most musical languages. The Atlantic, November 13. Retrieved
from https://www.theatlantic.com/international/archive/2015/11/tonal-languages-linguistics-
mandarin/415701/
Mahmoudzadeh, M., Dehaene-Lambertz, G., Fournier, M., Kongolo, G., Goudjil, S., Dubois, J., &
Wallois, F. (2013). Syllabic discrimination in premature human infants prior to complete formation
of cortical layers. Proceedings of the National Academy of Sciences 110(12), 4846–4851.
Mahmoudzadeh, M., Wallois, F., Kongolo, G., Goudjil, S., & Dehaene-Lambertz, G. (2016).
Functional maps at the onset of auditory inputs in very early preterm human neonates. Cerebral
Cortex 27(4), 2500–2512.
Mampe, B., Friederici, A. D., Christophe, A., & Wermke, K. (2009). Newborns’ cry melody is
shaped by their native language. Current Biology 19, 1994–1997.
Mann, B. (1961). Who Put the Bomp. ABC-Paramount 10237.
Minagawa-Kawai, Y., Cristià, A., & Dupoux, E. (2011b). Cerebral lateralization and early speech
acquisition: A developmental scenario. Developmental Cognitive Neuroscience 1(3), 217–232.
Minagawa-Kawai, Y., Cristià, A., Vendelin, I., Cabrol, D., & Dupoux, E. (2011a). Assessing signal-
driven mechanisms in neonates: Brain responses to temporally and spectrally different sounds.
Frontiers in Psychology 2. Retrieved from https://doi.org/10.3389/fpsyg.2011.00135
Minagawa-Kawai, Y., Mori, K., & Sato, Y. (2005). Different brain strategies underlie the categorical
perception of foreign and native phonemes. Journal of Cognitive Neuroscience 17(9), 1376–1385.
Miranda, R. A., & Ullman, M. T. (2007). Double dissociation between rules and memory in music:
An event-related potential study. NeuroImage 38(2), 331–345.
Moog, H. (1976). The musical experience of the pre-school child. Trans. C. Clarke. London: Schott.
Moon, C., Cooper, R. P., & Fifer, W. P. (1993).Two-day-olds prefer their native language. Infant
Behavior and Development 16(4), 495–500.
Moran, J. (2006). Artist-in-residence. New York: Blue Note Records.
Möttönen, R., Calvert, G. A., Jääskeläinen, I. P., Matthews, P. M., Thesen, T., Tuomainen, J., &
Sams, M. (2006). Perceiving identical sounds as speech or non-speech modulates activity in the
left posterior superior temporal sulcus. NeuroImage 30(2), 563–569.
Näätänen, R., Lehtokoski, A., Lennes, M., Cheour, M., Huotilainen, M., Iivonen, A., & Allik, J.
(1997). Language-specific phoneme representations revealed by electric and magnetic brain
responses. Nature 385(6615), 432–434.
Nazzi, T., Bertoncini, J., & Mehler, J. (1998). Language discrimination by newborns: Toward an
understanding of the role of rhythm. Journal of Experimental Psychology: Human Perception and
Performance 24(3), 756–766.
Nespor, M., Shukla, M., van de Vijver, R., Avesani, C., Schraudolf, H., & Donati, C. (2008).
Different phrasal prominence realization in VO and OV languages. Lingue e Linguaggio 7(2), 1–
28.
Norman-Haignere, S., Kanwisher, N. G., & McDermott, J. H. (2015). Distinct cortical pathways for
music and speech revealed by hypothesis-free voxel decomposition. Neuron 88(6), 1281–1296.
Nuñez, S. C., Dapretto, M., Katzir, T., Starr, A., Bramen, J., Kan, E., … Sowell, E. R. (2011). fMRI
of syntactic processing in typically developing children: Structural correlates in the inferior frontal
gyrus. Developmental Cognitive Neuroscience 1(3), 313–323.
Olsho, L. W., Schoon, C., Sakai, R., Turpin, R., & Sperduto, V. (1982). Auditory frequency
discrimination in infancy. Developmental Psychology 18(5), 721–726.
Overy, K., Nicolson, R. I., Fawcett, A. J., & Clarke, E. F. (2003). Dyslexia and music: Measuring
musical timing skills. Dyslexia 9(1), 18–36.
Pannekamp, A., Weber, C., & Friederici, A. D. (2006). Prosodic processing at the sentence level in
infants. NeuroReport 17(6), 675–678.
Patel, A. D. (2012). Language, music, and the brain: A resource-sharing framework. In P. Rebuschat,
M. Rohrmeier, J. A. Hawkins, & I. Cross (Eds.), Language and music as cognitive systems (pp.
204–223). Oxford: Oxford University Press.
Perani, D. (2012). Functional and structural connectivity for language and music processing at birth.
Rendiconti Lincei 23(3), 305–314.
Perani, D., Dehaene, S., Grassi, F., Cohen, L., Cappa, S. F., Dupoux, E., … Mehler, J. (1996). Brain
processing of native and foreign languages. NeuroReport 7(15–17), 2439–2444.
Perani, D., Paulesu, E., Galles, N. S., Dupoux, E., Dehaene, S., Bettinardi, V., & Mehler, J. (1998).
The bilingual brain: Proficiency and age of acquisition of the second language. Brain 121(10),
1841–1852.
Perani, D., Saccuman, M. C., Scifo, P., Anwander, A., Spada, D., Baldoli, C., … Friederici, A. D.
(2011). Neural language networks at birth. Proceedings of the National Academy of Sciences
108(38), 16056–16061.
Perani, D., Saccumann, M. C., Scifo, P., Spada, D., Andreolli, G., Rovelli, R., … Koelsch, S. (2010).
Functional specializations for music processing in the human newborn brain. Proceedings of the
National Academy of Sciences 107(10), 4758–4763.
Peretz, I. (2001). The biological foundations of music. In E. Dupoux (Ed.), Language, Brain, and
Cognitive Development: Essays in Honor of Jacques Mehler (pp. 435–466). Cambridge, MA: MIT
Press.
Peretz, I., & Coltheart, M. (2003). Modularity of music processing. Nature Neuroscience 6(7), 688–
691.
Peretz, I., & Hyde, K. L. (2003). What is specific to music processing? Insights from congenital
amusia. Trends in Cognitive Sciences 7(8), 362–367.
Perruchet, P., & Pacton, S. (2006). Implicit learning and statistical learning: One phenomenon, two
approaches. Trends in Cognitive Sciences 10(5), 233–238.
Petitto, L. A., Holowka, S., Sergio, L. E., & Ostry, D. (2001). Language rhythms in baby hand
movements. Nature 413(6851), 35–36.
Piazza, E. A., Iordan, M. C., & Lew-Williams, C. (2017). Mothers consistently alter their unique
vocal fingerprints when communicating with infants. Current Biology 27(20), 3162–3167.
Polka, L., & Werker, J. F. (1994). Developmental changes in perception of nonnative vowel contrasts.
Journal of Experimental Psychology: Human Perception and Performance 20(2), 421–435.
Prochnow, A., Erlandsson, S., Hesse, V., & Wermke, K. (2017). Does a “musical” mother tongue
influence cry melodies? A comparative study of Swedish and German newborns. Musicae
Scientiae (October). doi:1029864917733035
Przybylski, L., Bedoin, N., Krifi-Papoz, S., Herbillon, V., Roch, D., Léculier, L., & Tillmann, B.
(2013). Rhythmic auditory stimulation influences syntactic processing in children with
developmental language disorders. Neuropsychology 27(1), 121–131.
Radvansky, G. A., & Potter, J. K. (2000). Source cuing: Memory for melodies. Memory and
Cognition 28(5), 693–699.
Ramus, F., & Mehler, J. (1999). Language identification with suprasegmental cues: Study based on
speech resynthesis. Journal of the Acoustical Society of America 105, 512–521.
Ravignani, A., Delgado, T., & Kirby, S. (2016). Musical evolution in the lab exhibits rhythmic
universals. Nature Human Behaviour. doi:10.1038/s41562-016-0007
Reich, S. (1989). Different Trains/Electric Counterpoint. Nonesuch 979176-2.
Rivera-Gaxiola, M., Klarman, L., Garcia-Sierra, A., & Kuhl, P. K. (2005). Neural patterns to speech
and vocabulary growth in American infants. NeuroReport 16, 495–498.
Robinson, K., & Patterson, R. D. (1995). The duration required to identify the instrument, the octave,
or the pitch chroma of a musical note. Music Perception 13, 1–14.
Rousseau, B. (2016). Which language uses the most sounds? Click 5 times for the answer. The New
York Times, November 25. Retrieved from https://www.nytimes.com/2016/11/25/world/what-in-
the-world/click-languages-taa-xoon-xoo-botswana.html
Saffran, J., Aslin, R., & Newport, E. (1996). Statistical learning by eight-month-old infants. Science
274(5294), 926–928.
Saint-Georges, C., Chetouani, M., Cassel, R., Apicella, F., Mahdhaoui, A., Muratori, F., & Cohen, D.
(2013). Motherese in interaction: At the cross-road of emotion and cognition? (A systematic
review). PLoS ONE, 8(10). Retrieved from https://doi.org/10.1371/journal.pone.0078103
Sammler, D., Baird, A., Valabrègue, R., Clément, S., Dupont, S., Belin, P., & Samson, S. (2010). The
relationship of lyrics and tunes in the processing of unfamiliar songs: A functional magnetic
resonance adaptation study. Journal of Neuroscience 30(10), 3572–3578.
Sansavini, A., Bertoncini, J., & Giovanelli, G. (1997). Newborns discriminate the rhythm of
multisyllabic stressed words. Developmental Psychology 33(1), 3–11.
Schellenberg, E. G., & Trainor, L. J. (1996). Sensory consonance and the perceptual similarity of
complex-tone harmonic intervals: Tests of adult and infant listeners. Journal of the Acoustical
Society of America 100(5), 3321–3328.
Scott, C. (2004). Syntactic ability in children and adolescents with language and learning disabilities.
In R. A. Berman (Ed.), Language Development Across Childhood and Adolescence (pp. 111–134).
Amsterdam: John Benjamins.
Shepard, R. N. (1980). Multidimensional scaling, tree-fitting, and clustering. Science 210(4468),
390–398.
Shi, R., Werker, J. F., & Morgan, J. L. (1999). Newborn infants’ sensitivity to perceptual cues to
lexical and grammatical words. Cognition 72(2), B11–B21.
Shultz, S., Vouloumanos, A., Bennett, R. H., & Pelphrey, K. (2014). Neural specialization for speech
in the first months of life. Developmental Science 17(5), 766–774.
Smith, K., Kirby, S., & Brighton, H. (2003). Iterated learning: A framework for the emergence of
language. Artificial Life 9, 371–386.
Speer, J. R., & Meeks, P. U. (1985). School children’s perception of pitch in music.
Psychomusicology 5, 49–56.
Spiller, H. (2004). The traditional sounds of Indonesia. Santa Barbara, CA: ABC-CLIO.
Stern, D. N., Spieker, S., & MacKain, K. (1982). Intonation contours as signals in maternal speech to
prelinguistic infants. Developmental Psychology 18(5), 727–735.
Thompson, W. F., Marin, M. M., & Stewart, L. (2012). Reduced sensitivity to emotional prosody in
congenital amusia rekindles the musical protolanguage hypothesis. Proceedings of the National
Academy of Sciences 109(46), 19027–19032.
Tillmann, B., Albouy, P., & Caclin, A. (2015). Congenital amusias. In G. G. Celesia & G. S. Hickok
(Eds.), The human auditory system: Fundamental organization and clinical disorder (3rd ed.; pp.
589–605). Amsterdam: Elsevier.
Trainor, L. J., Austin, C. M., & Desjardins, N. (2000). Is infant-directed speech a result of the vocal
expression of emotion? Psychological Science 11(3), 188–195.
Trainor, L. J., & Corrigall, K. A. (2010). Music acquisition and effects of musical experience. In M.
Riess Jones, R. R. Fay, & A. N. Popper (Eds.), Music Perception (Vol. 36; pp. 89–127). New York:
Springer.
Trainor, L. J., & Trehub, S. E. (1994). Key membership and implied harmony in Western tonal
music: Developmental perspectives. Attention, Perception, and Psychophysics 56(2), 125–132.
Trainor, L. J., Wu, L., & Tsang, C. D. (2004). Long-term memory for music: Infants remember tempo
and timbre. Developmental Science 7(3), 289–296.
Trehub, S., Cohen, A., Thorpe, L., & Morrongiello, B. (1986). Development of the perception of
musical relations: Semitone and diatonic structure. Journal of Experimental Psychology: Human
Perception and Performance 12, 295–301.
Trehub, S. E., Endman, M. W., & Thorpe, L. A. (1990). Infants’ perception of timbre: Classification
of complex tones by spectral structure. Journal of Experimental Child Psychology 49(2), 300–313.
Trehub, S. E., & Thorpe, L. A. (1989). Infants’ perception of rhythm: Categorization of auditory
sequences by temporal structure. Canadian Journal of Psychology/Revue canadienne de
psychologie 43(2), 217–229.
Vouloumanos, A., & Werker, J. F. (2007). Listening to language at birth: Evidence for a bias for
speech in neonates. Developmental Science 10(2), 159–164.
Wang, X., & Peng, G. (2014). Phonological processing in Mandarin speakers with congenital amusia.
Journal of the Acoustical Society of America 136(6), 3360–3370.
Weiss, A. H., Granot, R. Y., & Ahissar, M. (2014). The enigma of dyslexic musicians.
Neuropsychologia 54, 28–40.
Welch, G. F. (2002). Early childhood musical development. In L. Bresler & C. Thompson (Eds.), The
arts in children’s lives: Context, culture and curriculum (pp. 113–128). Dordrecht: Kluwer.
Welch, G. F. (2009). Evidence of the development of vocal pitch matching ability in children.
Japanese Journal of Music Education Research 21, 1–13.
Werker, J. F., & Tees, R. (1984). Cross-language speech perception: Evidence for perceptual
reorganization during the first year of life. Infant Behavior and Development 7, 49–63.
Wermke, K., Leising, D., & Stellzig-Eisenhauer, A. (2007). Relation of melody complexity in
infants’ cries to language outcome in the second year of life: A longitudinal study. Clinical
Linguistics and Phonetics 21(11–12), 961–973.
Wermke, K., & Mende, W. (2009). Musical elements in human infants’ cries: In the beginning is the
melody. Musicae Scientiae 13(2), 151–175.
Werner, L. A., & Marean, G. C. (1996). Human auditory development. Madison, WI: Brown
Benchmark.
Wieland, E. A., McAuley, J. D., Dilley, L. C., & Chang, S. E. (2015). Evidence for a rhythm
perception deficit in children who stutter. Brain & Language 144, 26–34.
Winkler, I., Háden, G. P., Ladinig, O., Sziller, I., & Honing, H. (2009). Newborn infants detect the
beat music. Proceedings of the National Academy of Sciences 106(7), 2468–2471.
Zatorre, R. J., & Belin, P. (2001). Spectral and temporal processing in human auditory cortex.
Cerebral Cortex 11(10), 946–953.
Zhang, Y., Kuhl, P. K., Imada, T., Iverson, P., Pruitt, J., Stevens, E. B., & Nemoto, I. (2009). Neural
signatures of phonetic learning in adulthood: A magnetoencephalography study. NeuroImage
46(1), 226–240.
Zuk, J., Bishop-Liebler, P., Ozernov-Palchik, O., Moore, E., Overy, K., Welch, G., & Gaab, N.
(2017). Revisiting the “enigma” of musicians with dyslexia: Auditory sequencing and speech
abilities. Journal of Experimental Psychology: General 146(4), 495–511.
CHAPT E R 24
L A U R E L J. T R A I N O R A N D S U S A N MA R S H - R O L L O
E P B , M ,
R
From at least as young as two months of age, infants detect tempo changes
(Baruch & Drake, 1997) and can discriminate different rhythm patterns
composed of the same interval durations but in different orders, such as
100–600–300 ms versus 600–300–100 ms (Chang & Trehub, 1977;
Demany, McKenzie, & Vurpillot, 1977; Lewkowicz, 2003). Impressively,
they can make such discriminations across changes in the tempo and pitch
level of the comparison patterns (Trehub & Thorpe, 1989). Young infants
also show sensitivity to phrase and grouping structures in music. For
example, they are more sensitive to small timing perturbations inserted in
the middle of a phrase than at phrase boundaries, where an elongation is
likely to occur in any case (Jusczyk & Krumhansl, 1993; Thorpe & Trehub,
1989; Trainor & Adams, 2000). However, the temporal cues infants use to
determine grouping boundaries of short sequences may be influenced by the
language they are exposed to in their environment. Trainor and Adams
(2000) found that English-learning infants marked the ends of short
perceptual groups with longer tones, whereas Yoshida et al. (2010) found
that Japanese-learning infants marked the beginnings of perceptual groups
with longer tones, consistent with linguistic accent structures in the
respective languages.
Infants also show precocious sensitivity to beat and metrical structure.
For example, newborns exposed to a rhythmic pattern with a 4/4 time
structure show a larger event-related potential (ERP) response in
electroencephalographic (EEG) recordings to omissions on strong
compared to weak beats (Winkler, Háden, Ladinig, Sziller, & Honing,
2009), suggesting sensitivity to metrical structure. However, this study
needs to be replicated, as whether a beat was strong or weak was
confounded with the number of sounds that were omitted. By 7 months of
age, however, there is clear evidence of metrical processing. After
habituation to repeated trials containing rhythms in either duple or triple
meter, infants showed renewed interest for a rhythm presented with the
other (novel) meter (Hannon & Johnson, 2005). In another study, infants
bounced on either every second or third beat of a repeating six-beat
ambiguous rhythm pattern subsequently preferred to listen to a version of
the rhythm pattern with accents added on either every second or every third
beat that matched the meter set up during the bouncing experience,
indicating both discrimination between meters and early involvement of the
motor system in the perception of meter (Phillips-Silver & Trainor, 2005).
In adults, listening to a rhythm pattern entrains neural activity at the
perceived beat and meter frequencies (i.e., tempos) (Fujioka et al., 2012;
Fujioka, Ross, & Trainor, 2015; Nozaradan, 2014; Nozaradan, Peretz, &
Mouraux, 2012; Tal et al., 2017). Evidence from EEG studies also indicates
that at least as young as 7 months, neural oscillations in auditory cortex
entrain to both beat and metrical levels of rhythm patterns (Cirelli, Spinelli,
Nozaradan, & Trainor, 2016). Seven-month-old and 15-month-old infants
listened to a repeating six-beat rhythm pattern that could be interpreted as
either in duple or triple meter, as in Phillips-Silver and Trainor (2005). EEG
recordings were subjected to Fourier analysis following the frequency-
tagging methods of Nozaradan and colleagues (Nozaradan, Peretz, Missal,
& Mouraux, 2011). Peaks in the frequency spectrum were found that
corresponded to the beat frequency as well as to both duple and triple
metrical interpretations, indicating neural entraining at both beat and meter
frequencies in both age groups. Interestingly, at 7 months, the amplitude at
the duple meter frequency was enhanced in infants engaged in infant–parent
music classes compared to those not enrolled in music classes. At 15
months, beat and both meter frequencies were all enhanced in infants whose
parents were musically trained compared to those whose parents were not
musically trained. Thus, early in development neural circuits are sensitive
to the temporal structure of incoming auditory rhythmic patterns.
Just as tonality, pitch, and harmonic structures vary across the musical
systems used in different cultures, so do metrical structures (Hannon et al.,
2018; Hannon & Trainor, 2007; Trainor & Hannon, 2013). In Western
music, durations most commonly stand in simple 2:1 ratios (e.g., a march
meter), with 3:1 as the next most common ratio (e.g., a waltz meter).
However, in many parts of the world (e.g., Africa, the Balkans, South Asia,
South America), more complex metrical patterns that create a non-
isochronous beat at one or more levels of the metrical hierarchy are
common (Hannon, Soley, & Ullal, 2012; London, 2004). For example, an
isochronous beat at a basic level of the hierarchy might be grouped into
alternating groups of three and two beats at the next level of the hierarchy,
creating a five-beat pattern. The alternating groups of two and three beats
create a more complex duration ratio of 3:2. Western adults without
exposure to music with such metrical patterns are much better at
discriminating, remembering, reproducing, and tapping to rhythm patterns
with simple compared to complex meters (e.g., Essens, 1986; Essens &
Povel, 1985; Fraisse, 1982; Hannon & Trehub, 2005a; Repp, London, &
Keller, 2005; Snyder, Hannon, Large, & Christiansen, 2006). However,
adults who grew up in cultures employing complex meter do not show
processing differences between familiar rhythms with simple and complex
meters (e.g., Hannon & Trehub, 2005a; Hannon, Soley, & Ullal, 2012).
Just as infants learn the particular language(s) in their environment,
becoming more sensitive to the phonemic structure of that language, and
less sensitive to alternative phonemic structures by their first birthday (Kuhl
et al., 2006; Werker & Tees, 2005), a similar process of perceptual
narrowing occurs in music acquisition, such that infants become specialized
at processing both the tonal (Gerry, Unrau, & Trainor, 2012; Lynch &
Eilers, 1992; Lynch, Eilers, Oller, & Urbano, 1990; Trainor, Marie, Gerry,
Whiskin, & Unrau, 2012; Trainor & Trehub, 1992, 1994) and metrical
structures (Gerry, Faux, & Trainor, 2010; Hannon & Trehub, 2005a, 2005b)
in the music they experience in their environment (Hannon & Trainor,
2007; Trainor & Corrigall, 2010; Trainor & Hannon, 2013; Trainor &
Unrau, 2012). With respect to metrical processing, at 4 to 6 months of age,
Western infants notice if an extra beat is added to a 7/4 meter as well as if a
beat is dropped from an 8/4 meter (Hannon & Trehub, 2005a). However,
performance on the 7/4 meter declines between 7 and 12 months, such that
12-month-old Western infants, like Western adults, perform very poorly on
this task (Hannon & Trehub, 2005b). That these declines for non-native
meters are driven by experience is reinforced by findings that listening
experience can speed up or slow down (or even reverse) the perceptual
narrowing. As far as slowing down or reversing perceptual narrowing,
providing daily listening experience with non-Western non-isochronous
meters reinstates the loss of sensitivity for the non-isochronous meters in
12-month-olds (Hannon & Trehub, 2005a). Interestingly, there appears to
be a window of sensitivity for reversing perceptual narrowing for meter, as
Western 5- to 7-year-old children show some, but not full, reinstatement
after a similar listening experience with non-isochronous meters, whereas
adults show no evidence of reinstatement after such experience (Hannon,
Vandenbosch der Nederlanden, & Tichko, 2012). As far as speeding up
perceptual narrowing, Gerry et al. (2010) found that 7-month-old infants
enrolled in Kindermusik classes showed a listening preference for a rhythm
with accents on every second beat compared to the same rhythm with
accents on every third beat, the former meter being more common in
Western music, whereas infants not enrolled in music classes did not show
this preference. In general, a preference for native meters may be evident
prior to perceptual narrowing. Soley and Hannon (2010) found that a
preference for isochronous over non-isochronous meters increases between
4 and 8 months in Western infants whereas no listening preferences are
evident during this age period in Turkish infants.
The bias for culture-specific meters continues into childhood. Einarson
and Trainor (2015, 2016) developed a child-friendly version of the Beat
Alignment Task (BAT) (Iversen & Patel, 2008) that included music with
both simple and complex meters (cBAT). In this task, children watch pairs
of short video excerpts of puppets drumming to musical excerpts. One
puppet drums on the beat and the other puppet’s drumming is either at the
wrong tempo or misaligned in phase with the music. The children decide
which puppet is the best drummer for a band. Western 5-year-old children
were at change levels on music with complex meters, but performed
significantly better (and above change levels) on music with simple meters.
Together, the infant and child studies indicate that perceptual narrowing for
the metrical structures common in the music of one’s culture develop early
and are maintained in childhood, raising the interesting question of whether
best pedagogical practice might be to expose infants and young children to
complex meters if the goal is to provide them with the perceptual tools to
understand the rhythms of music from around the world.
T R I -D
S
For most infants, their first experience of music is likely hearing their
mother sing, and a diary study of North American mothers, where
opportunities to listen to recorded music abound, indicates that most
mothers still sing to their infants many times during the day, such as when
bathing them, playing, feeding, during diaper changes, in the car, and at
sleep time (Trehub et al., 1997). Infants often experience their parents’
singing while being held and rocked or walked rhythmically, or while
feeling their parent tap their back or touch other body parts rhythmically
during the song, so that from early ages, infants experience musical rhythms
in a multisensory context involving hearing, movement, and vision. Singing
to infants appears to be a spontaneous intuitive response to the presence of
an infant, and universal across human cultures. Furthermore, across
cultures, Western adults were able to discern lullabies that were intended for
infants from other songs matched in tempo and general style (Trehub,
Unyk, & Trainor, 1993), suggesting that infant-directed singing might have
been an evolutionary adaptation that helped infants to survive. The one
song category that Western adults found difficult to distinguish from
lullabies was love songs (Trehub & Trainor, 1998), suggesting that lullabies
may express and communicate the deep emotional bonds between parents
and their infants. Music appears to be particularly effective at controlling
infants’ states—when left alone, hearing their mothers’ infant-directed
singing was found to keep infants happier for considerably longer than
hearing their mothers’ infant-directed speech (Corbeil, Trehub, & Peretz,
2016), and cortisol levels were found to decrease in infants when their
mothers sang to them (Shenfield, Trehub, & Nakata, 2003).
Just as adults use a different speaking style when talking to infants
compared to adults, termed infant-directed or musical speech (Fernald,
1991; Papoušek, Papoušek, & Symmes, 1991), they sing differently to
infants than they sing in other circumstances (Trehub & Trainor, 1998).
Trainor (1996) recorded mothers singing the same song when their infant
was present and when their infant was absent and found that adults were
highly accurate at identifying the infant-directed versions. Furthermore,
using a preferential looking paradigm, she found that infants preferred to
listen to the infant-directed versions. In addition to being sung at a higher
pitch and in a more loving tone of voice, in comparison to non-infant-
directed singing, infant-directed singing also differs in timing and rhythmic
features. It is generally slower in tempo and has exaggerated structural
features, such as enhanced phrase boundaries, rhythm, and grouping
(Longhi, 2009; Trainor, Clark, Huntley, & Adams, 1997). For example,
infant-directed singing contains longer pauses between phrases (Trainor et
al., 1997). There is also evidence that mothers exaggerate the hierarchical
beat structure of songs when singing to infants, using both acoustic accents
and body movements to do so (Longhi, 2009). They particularly emphasize
upbeats, which is interesting in that upbeats provide anticipatory
information that a downbeat is expected to follow. And while infants are not
yet able to synchronize movements precisely to the beats of music, infants
in this study made more synchronous movements to beats at the beginnings
and ends of phrases than in the middle, suggesting some understanding of
the temporal structure of phrases in infant-directed singing with
exaggerated rhythmic cues. There is some evidence that depressed mothers
do not employ the full repertoire of infant-directed singing features,
generally singing faster and with less expression than non-depressed
mothers (de l’Etoile & Leider, 2011), and possibly compromising
communication with their infants.
Two basic categories of infant-directed singing have been identified:
lullabies, where the intention is to help a fussy infant to fall asleep, and
playsongs, where the intention is to rouse the infant, interact with them in
play, and direct their attention to interesting people and things in the
environment (Rock, Trainor, & Addison, 1999; Trainor, 1996). These two
categories arise more from the style or manner in which the caregiver sings
than the structural content of the music. Indeed, Rock et al. (1999) recorded
mothers singing a song of their choice to their infant, once in a lullaby style
and once in a playsong style. Adult raters were 100 percent accurate at
identifying which were lullabies and which playsongs, indicating that these
styles are highly distinct. Furthermore they rated playsongs as sounding
more rhythmic, clipped, and accented compared to lullabies, which were
rated as sounding smoother. Importantly, infants show differential behaviors
in response to lullaby and playsong renditions of the same song (Rock et al.,
1999) and prefer faster tempos for playsongs but not for lullabies (Conrad,
Walsh, Allen, & Tsang, 2011). That such timing and rhythmic differences
are likely universal across human cultures and musical systems suggests
that the perceptual, emotional, and social consequences of these temporal
features may have evolutionary origins.
Singing to infants is a social interaction that requires temporal
coordination between the caregiver and the infant. Such temporal
coordination during the first months after birth appears to promote
communication, the development of successful social interactions, and
emotion regulation (Ilari, 2016; Malloch & Trevarthen, 2009). These social
consequences of rhythmic interactions are discussed later in the chapter.
R , P , N
O , D
R
The regularity of rhythms enables prediction of when the next beat is likely
to occur (Large & Jones, 1999; Trainor & Zatorre, 2015), which can aid in
preparing for incoming information and focusing attention at information-
rich points in time. Predictive timing is critical for the perception of stimuli
such as speech and music that unfold rapidly over time and are fleeting in
that once each note or phoneme ends, the next begins, and it is not possible
to hear the input again. Indeed there is considerable evidence that the adult
brain is continually predicting the future and comparing its predictions with
what actually occurs (Fujioka et al., 2012, 2015; Herrmann, Henry,
Haegens, & Obleser, 2016; Morillon & Schroeder, 2015). In the case of
incorrect predictions, an error signal is generated which can engage
attention and lead to additional processing and learning (Arnal & Giraud,
2012; Chang, Bosnyak, & Trainor, 2018; Ding et al., 2017; Haegens & Zion
Golumbic, 2018; Nobre et al., 2007; Nobre & van Ede, 2018; Schroeder &
Lakatos, 2009; Schröger, Marzecová, & SanMiguel, 2015). In adults, both
behavioral and neural evidence indicates that the perception of sounds at
beat onsets presented in rhythmic contexts is enhanced (Arnal & Giraud,
2012; Chang et al., under review b; Haegens & Zion-Golumbic, 2018;
Henry & Obleser, 2012; Herrmann et al., 2016; Jones et al., 2002; Morillon,
Schroeder, Wyart, & Arnal, 2016; Nobre & van Ede, 2018). Predictive
processes are evident very early in infancy in that occasional unexpected
changes (deviants) in isochronous sound sequences lead to ERP mismatch
responses (MMRs) in EEG recordings (Basirat, Dehaene, & Dehaene-
Lambertz, 2014; Háden, Németh, Török, & Winkler, 2015; He, Hotson, &
Trainor, 2007; Trainor, 2012; Trainor & He, 2013; Trainor & Zatorre,
2015).
The neural mechanisms that underlie predictive timing are beginning to
be understood in the adult brain in terms of neural oscillations. Specifically,
low frequency neural oscillations (delta band, ~1–3 Hz) phase align with
the onsets of beats in an auditory rhythmic stimulus such that predictive
timing is enhanced (Arnal, Poeppel, & Giraud, 2015; Bauer, Bleichner,
Jaeger, Thorne, & Debener, 2018; Calderone, Lakatos, Butler, &
Castellanos, 2014; Henry, Herrmann, & Obleser, 2014; Henry & Obleser,
2012; Schroeder & Lakatos, 2009; Stefanics et al., 2010). The power in
higher frequency oscillations (beta band, ~20 Hz) is also modulated by
auditory rhythmic stimuli such that beta power decreases after a beat onset
and rebounds so as to reach maximum amplitude at the expected time of the
next beat, dependent on the tempo of the rhythmic input (Fujioka et al.,
2012, 2015). This rebound time appears to be a neural signature of timing
prediction in the brain, and beta oscillations are proposed to reflect
attentional processes leading to enhanced perception at particular time
points (Arnal & Giraud, 2012; Chang et al., 2018, under review b; Iversen,
Repp, & Patel, 2009; Snyder & Large, 2005). Beta oscillations also appear
to be associated with capture of attention (Chang et al., 2018).
Very little research has examined the development of neural oscillations
involved in predictive processes in rhythmic contexts. However, one study
compared beta oscillations in 7-year-old children and adults in response to
isochronous beat sequences at different tempos (Cirelli, Bosnyak, et al.,
2014). Beta power entrainment to the tempo of the input was found, but the
responses of children were noisier than those of adults, and were
measurable only over a more narrow range of tempos than in adults. This
suggests that the neural oscillatory responses underlying predictive timing
follow a protracted developmental trajectory. Clearly more research is
needed to understand the brain development underlying rhythmic predictive
processes.
While neural oscillation studies of predictive timing have focused on
stimuli with constant tempos, in real music performances the tempo
typically modulates continuously (James et al., 2012; Palmer, 1989; Rankin
et al., 2009; Repp, 1992; Todd, 1985). These timing perturbations are not
random, but interact with the structure and content of the music. In
particular, expressive timing emphasizes phrase boundaries by lengthening
phrase-final notes or chords, and plays with temporal expectations by, for
example, elongating notes or chords that embody harmonic tension, thus
delaying their resolution (London, 2004; Repp, 2005; Repp & Su, 2013).
Perceptual studies indicate that both musically untrained adults (Clarke &
Krumhansl, 1990; Deliege, 1987; Palmer & Krumhansl, 1987; Peretz, 1989)
and infants (Krumhansl & Jusczyk, 1990; Trainor & Adams, 2000) are
sensitive to phrase boundaries. Even non-musicians produce phrase-final
lengthening in their performances (Kragness & Trainor, 2016), suggesting
that it is not a learned performance technique, but is based on intrinsic
temporal expectations. Specifically, Kragness and Trainor (2016) used a
self-paced tapping paradigm in which non-musician adults pressed a key to
get the next chord in an unfamiliar sequence of chords. In their renditions,
adults tended to speed up in the middle of phrases defined by typical
Western cadences, and to slow down at the ends of phrases, even after
metrical regularity and melodic contour were controlled.
One possible explanation for phrase-final lengthening is that ends of
phrases tend to be points of high entropy in that it is difficult to predict what
will come next, whereas points in the middle of phrases tend to be of low
entropy in that it is relatively easy to predict the next note or chord (Pearce,
Müllensiefen, & Wiggins, 2010; Pearce & Wiggins, 2006). The uncertainty
at phrase boundaries might require more processing time, leading to a
natural slowing. Further evidence for an entropy explanation comes from a
developmental study in which children as young as 3 years were found to
dwell longer at phrase endings, although sophistication in the cues used to
detect phrase boundaries increased between 3 and 7 years of age (Kragness
& Trainor, 2018). That very young children produce phrase-final
lengthening in their musical productions is consistent with the possibility
that it is based on intrinsic processing properties of perception rather than
reflecting learning of a particular musical performance style.
These studies suggest, first, that the brain entrains to the beat and meter
in auditory rhythms early in development, but that this entrainment and its
relation to attention and error monitoring continues to develop for many
years and, second, that although a steady beat is typically experienced when
listening to music, timing perturbations in isochrony that follow the
structure and content of the music are present early in childhood, suggesting
that beat perception and the neural processes underlying it are not strictly
isochronous, but involve an interaction between time and context.
T M E
D A –M
E S E
S M
It is often noted that across cultures past and present music serves social
functions, including action coordination, communication, and social
cohesion (Cirelli, 2018; Cirelli, Trehub, & Trainor, 2018; D’Ausilio,
Novembre, Fadiga, & Keller, 2015; Ilari, 2016; Patel & Iversen, 2014;
Trainor, 2015; Trainor & Cirelli, 2015). Indeed music is present at virtually
all important social occasions including weddings, funerals, religious
rituals, parties, sporting events, and political rallies. A number of
researchers have suggested that the main function of joint musical
experiences among adults is the increased social cohesion that results from
synchronous movement (e.g., Bispham, 2006; Brown & Volgsten, 2006;
Fitch, 2006; Huron, 2001; McNeill, 1995; Merker, 2000). It has also been
suggested that music similarly enhances the social relationships between
infants and their caregivers, with the coordinated interaction that music
engenders increasing attachment, bonding, emotional recognition, and self-
regulation in early development (Cirelli, 2018; Cirelli, Trehub, & Trainor,
2018; Dissanayake, 2012; Ilari, 2016; Malloch & Trevarthen, 2009; Trainor
& Cirelli, 2015).
Many motor movements across species are rhythmic, including
heartbeats, locomotion (e.g., walking, running, skipping, swimming, wing
flapping), pulsating in fireflies, and sound productions from speech in
humans to chirping in crickets (Ackermann, 2008; Bentley & Hoy, 1974;
Buck, 1935, 1937, 1988; Kelso, Saltzman, & Tuller, 1986; Partridge, 1982;
Peelle & Davis, 2012; Weimerskirch, Martin, Clerquin, Alexandre, &
Jiraskova, 2001). However, spontaneous synchronization of movements to
an external auditory beat appears to be relatively rare among non-human
species (Merchant & Honing, 2014; Merchant et al., 2015; Patel, Iversen,
Bregman, & Schulz, 2009; Schachner, Brady, Pepperberg, & Hauser, 2009),
but very common and seemingly effortless in most human adults (Iversen &
Patel, 2008; Patel & Iversen, 2014; Repp, 2005; Repp & Su, 2013; Trainor,
2015).
Despite the ease with which adults achieve auditory–motor rhythmic
coupling, it appears to take a long time to develop (Cirelli, Trehub, &
Trainor, 2018; Drake, 1993; Einarson & Trainor, 2013, 2015, 2016;
Fitzpatrick, Schmidt, & Lockman, 1996; Luck & Toiviainen, 2006; Phillips-
Silver & Trainor, 2005; Provasi & Bobin-Bègue, 2003; Trainor & Cirelli,
2015; Van Noorden & De Bruyn, 2009; Zentner & Eerola, 2010), so
interactions in early development involving synchronous movement likely
rely on the caregiver to achieve the synchrony. Zentner and Eerola (2010)
analyzed the movements of a European sample of infants while they
listened to music and found no evidence for precise synchronization with
the tempo of the music. However, infants moved more in response to music
than to speech, and moved differently to music with faster compared to
slower tempos. Such early responses are likely influenced by early
experiences; in a follow-up study, Ilari (2015) found increased spontaneous
rhythmic movements to music in a Brazilian sample of infants compared to
those in the European sample of Zentner and Eerola (2010).
At least among children growing up in Western cultures, children
younger than 4 years of age generally have difficulty entraining to a beat.
One study found that 2.5-year-old children only succeeded at tapping to a
beat when the tempo was around their spontaneous tapping rate of about
400 ms onset-to-onset (Provasi & Bobin-Bègue, 2003). Another study
reported that 3-year-olds performed poorly in general at clapping to a beat
(Fitzpatrick et al., 1996). And although children younger than 4 years of age
readily engage in whole body movements in response to music, their
hopping, swaying, and circling are not generally entrained to the tempo of
the music (Eerola, Luck, & Toiviainen, 2006). However, by 4 years of age,
clear motoric entrainment to a beat emerges (Drake, Jones, & Baruch, 2000;
Eerola et al., 2006; Endedijk et al., 2015; Fitzpatrick et al., 1996; McAuley,
Jones, Holub, Johnston, & Miller, 2006; Provasi & Bobin-Bègue, 2003),
although school-aged children still perform worse than adults (Einarson &
Trainor, 2013; Van Noorden & De Bruyn, 2009).
Interestingly, evidence for auditory–motor entrainment can be seen at
younger ages when the task is embedded in a social situation. While not
synchronization, coordination between mothers and infants aged 3 to 9
months is evident in infancy, and correlates with self-regulation, future IQ,
and development of empathy (Feldman, 2007). Furthermore, during infant-
directed singing, infants’ head, body, hand, and leg movements coordinate
most with the music at the beginnings and ends of phrases (Longhi, 2009).
At somewhat older ages, precursors of entrainment can be seen in the
spontaneous drumming of pairs of children in social situations; for instance,
at ages 2 and 3 years, children will stop and start drumming when a partner
does so, but only 4-year-olds appear to adapt the tempo of their drumming
to that of their partner, even though all children showed tempo stability (i.e.,
the ability to produce relatively isochronous drumming sequences)
(Endedijk et al., 2015).
Interestingly, one study indicates that children as young as 2.5 years of
age will entrain their drumming to that of an adult social partner (Kirschner
& Tomasello, 2009). That this ability does not rely simply on visual cues is
evident in that the children performed better when drumming with a human
social partner than with a machine that hit the drum. Unlike when
entraining to a predetermined stimulus as in most laboratory studies, in real
musical interactions between people, all participants can adaptively adjust
their timing in response to the other musicians (D’Ausilio et al., 2015;
Keller, Novembre, & Hove, 2014; Nakata & Trainor, 2015). Indeed, the
information flow among members of a musical ensemble can be measured
through correlational and directional causal analyses of movement or EEG
(Chang, Livingstone, Bosnyak, & Trainor, 2017; Lindenberger, Li, Gruber,
& Müller, 2009; Sänger, Müller, & Lindenberger, 2012). With such
approaches, it is possible that socially adaptive musical interactions could
be measured in young children.
Cultural experience also has an effect. Parental reports suggest that
children in Brazil engage to a greater extent in such social music making
than do children from Germany, and children from Brazil show greater
propensity to spontaneously synchronize their drumming with another
person than children from Germany (Kirschner & Ilari, 2014).
Several studies indicate that when adults move in synchrony with each
other, they subsequently cooperate more, like and trust each other more,
remember more about each other, and engage in more altruistic acts (e.g.,
Anshel & Kipper, 1988; Hove & Risen, 2009; Launay, Dean, & Bailes,
2013; Macrae, Duffy, Miles, & Lawrence, 2008; Tarr, Launay, & Dunbar,
2014; Valdesolo & DeSteno, 2011; Valdesolo, Ouyang, & DeSteno, 2010;
Wiltermuth & Heath, 2009; Woolhouse, Tidhar, Demorest, Morrison, &
Campbell, 2010). Furthermore, synchronized drumming can increase
activation in the caudate, a brain region linked to expectation and reward
(Kokal, Engel, Kirschner, & Keysers, 2011). Music provides an ideal
context for facilitating synchronous movement between people. Rhythmic
regularity in music enables prediction of when the next beat is expected,
and therefore advanced planning of the motor movements necessary to
synchronize with the beat. When people hear the same music and
synchronize their movements with the beat of that music, they necessarily
become synchronized with each other.
Although infants cannot precisely entrain their movements to an
auditory beat, they often experience such synchronization when they are
held and walked or rocked to singing and other music. The social effects of
synchronization with others begin to appear around the end of the first year
after birth, a time during which infants’ social understanding is undergoing
rapid development (Dunfield, Kuhlmeier, O’Connell, & Kelley, 2011). For
example, at 12 months, but not at 9 months, infants preferred a toy bear that
rocked in sync with them over one that rocked out-of-sync with them
(Tunçgenç, Cohen, & Fawcett, 2015). By 14 months, infants will engage in
overt helping behaviors, for example, picking up and handing back a
marker “accidentally” dropped by an experimenter engaged in drawing a
picture, or picking up and handing back a clothespin “accidentally” dropped
by an experimenter hanging clothes on a line (Warneken & Tomasello,
2006, 2007, 2009). At 14 months, infants are not yet able to move in sync to
music, but they can be bounced in sync to music when held by an assistant
in an infant carrier. Cirelli, Einarson, and Trainor (2014) had an assistant
bounce infants to “Twist and Shout” by the Beatles while the infant faced
forward in the carrier across from the experimenter, who either bounced in
sync with them (i.e., at the same tempo) or out of sync (at a different
tempo), by having the experimenter bounce according to a click track
delivered to her over headphones. They found that after less than 3 minutes
of such bouncing, infants were more likely to help the experimenter if they
experienced synchronous compared to asynchronous bouncing in a series of
helping tasks as just described. Furthermore, the increased helpfulness was
targeted at the person that the infant experienced the bouncing with—
infants showed no increased helpfulness toward a neutral experimenter who
was present during the bouncing episode, but did not move to the music
(Cirelli, Wan, & Trainor, 2014). However, if infants were shown a skit that
either indicated that a second experimenter was a “friend” of the bouncer, or
was simply an “acquaintance,” infants who bounced in sync with the
experimenter transferred their helpfulness to the friend but not to the
acquaintance. This suggests that infants use movement synchrony as one
cue to identify who is in their social group and who is not.
Infants also form expectations for future behaviors between other people
(third party relationships) by observing how they interact. It appears that
infants begin to use synchrony as a cue for third party relationships around
the same age. Fawcett and Tunçgenç (2017) found that 15-month-old but
not 12-month-old infants who watched bears move either in sync or out of
sync expected those who moved in sync to affiliate socially. Cirelli, Wan,
Johanis, and Trainor (2018) found that 12- and 15-month-old infants who
watched videos of two women bouncing either in sync with each other or
out of sync were surprised when those who bounced asynchronously
subsequently displayed friendly behavior, although no significant
expectations were observed for those who bounced synchronously.
The use of synchronous movement to music as a cue to social affiliation
continues into childhood. For example, after clapping together
synchronously, 4- to 6-year-old children are more likely to help each other
compared to children who experienced asynchronous clapping (Tunçgenç &
Cohen, 2016). Even passive synchronous versus asynchronous movement
(children were pushed on swings) at 4 years of age results in increased
coordination and cooperation between children (Rabinowitch & Meltzoff,
2017).
Interestingly, moving infants synchronously versus asynchronously with
an experimenter in the absence of music has similar prosocial effects on
subsequent helping behavior as when the music is present (Cirelli, Wan, et
al., 2017). This suggests that music may facilitate synchronous movement,
but that experiencing music together per se might not be crucial for
increasing prosocial behavior. However, infants bounced with no music
were much less happy and cooperative than those bounced with music
(Cirelli, Wan, et al., 2017), suggesting that the music does play a role in
infant emotion regulation, which is likely helpful for encouraging prosocial
behavior. Furthermore, synchronous movement between the experimenter
and infant increases helping even when the musical beat is irregular and
movements are therefore synchronized but not isochronous (Cirelli, Wan, &
Trainor, 2014). Interestingly, anti-phase bouncing between an infant and
experimenter (i.e., the experimenter is at maximum height when the infants
is at minimum height and vice versa) appears to be as powerful as in-phase
bouncing for eliciting helping behaviors from infants, suggesting that the
mechanism is not one of simple self-similarity (Cirelli, Wan, & Trainor,
2014). Together, these results suggest that the role of music in encouraging
prosocial behavior is one of promoting synchronous movement. If
synchronous movement can be achieved in the absence of music, it is still a
powerful force for encouraging prosocial behavior. However, music is an
ideal stimulus for promoting synchronous movement because its temporal
regularity enables the prediction necessary for coordinated movement.
A D
D
Young infants are sensitive to timing, rhythm, and meter, which helps them
to organize inputs such as speech and music into hierarchical meaningful
structures, and to enhance processing of auditory streams that unfold over
time by using the regularities of rhythms to predict when important
upcoming information will occur. Tempo and expressive timing are used in
both caregivers’ infant-directed singing and in children’s early musical
productions to convey emotional information. Auditory–motor connections
can be seen early in development in the influence of movement on metrical
interpretation, and in the influence of synchronous movement between an
infant and an adult on infants’ altruistic helping behaviors. At the same
time, it takes considerable development before children become proficient
at entraining their motor movements to musical beats. The critical
importance of timing and rhythm in development is evident in the strong
associations between poor skills in these domains and developmental
disorders including dyslexia, DCD, autism, attention deficits, and stuttering.
Early diagnosis of poor timing and rhythm skills holds promise for early
assessment of risk for developmental disorders and age-appropriate
interventions that can put young children on a better developmental
trajectory.
A
The writing of this chapter was supported by grants from the Natural
Science and Engineering Research Council of Canada, the Canadian
Institute of Health Research, and the Social Sciences and Humanities
Research Council of Canada.
R
Ackermann, H. (2008). Cerebellar contributions to speech production and speech perception:
Psycholinguistic and neurobiological perspectives. Trends in Neurosciences 31(6), 265–272.
Adachi, M., & Trehub, S. E. (1998). Children’s expression of emotion in song. Psychology of Music
26(2), 133–153.
Adachi, M., Trehub, S. E., & Abe, J. I. (2004). Perceiving emotion in children’s songs across age and
culture. Japanese Psychological Research 46(4), 322–336.
American Psychiatric Association (2013). Diagnostic and statistical manual of mental disorders (5th
ed.). Washington, DC: APA.
Anshel, A., & Kipper, D. A. (1988). The influence of group singing on trust and cooperation. Journal
of Music Therapy 25(3), 145–155.
Arnal, L. H., & Giraud, A. L. (2012). Cortical oscillations and sensory predictions. Trends in
Cognitive Sciences 16(7), 390–398.
Arnal, L. H., Poeppel, D., & Giraud, A. L. (2015). Temporal coding in the auditory cortex. In G.
Celesia & G. Hickok (Eds.), Handbook of clinical neurology, Vol. 129: The human auditory system
(pp. 85–98). Amsterdam: Elsevier.
Balkwill, L. L., & Thompson, W. F. (1999). A cross-cultural investigation of the perception of
emotion in music: Psychophysical and cultural cues. Music Perception: An Interdisciplinary
Journal 17(1), 43–64.
Balkwill, L. L., Thompson, W. F., & Matsunaga, R. I. E. (2004). Recognition of emotion in Japanese,
Western, and Hindustani music by Japanese listeners. Japanese Psychological Research 46(4),
337–349.
Baruch, C., & Drake, C. (1997). Tempo discrimination in infants. Infant Behavior and Development
20(4), 573–577.
Basirat, A., Dehaene, S., & Dehaene-Lambertz, G. (2014). A hierarchy of cortical responses to
sequence violations in three-month-old infants. Cognition 132(2), 137–150.
Bauer, A. K. R., Bleichner, M. G., Jaeger, M., Thorne, J. D., & Debener, S. (2018). Dynamic phase
alignment of ongoing auditory cortex oscillations. NeuroImage 167, 396–407.
Behrens, G. A., & Green, S. B. (1993). The ability to identify emotional content of solo
improvisations performed vocally and on three different instruments. Psychology of Music 21(1),
20–33.
Bentley, D., & Hoy, R. R. (1974). The neurobiology of cricket song. Scientific American 231(2), 34–
45.
Bhat, A. N., & Srinivasan, S. (2013). A review of “music and movement” therapies for children with
autism: Embodied interventions for multisystem development. Frontiers in Integrative
Neuroscience 7, 22. Retrieved from https://doi.org/10.3389/fnint.2013.00022
Bhide, A., Power, A., & Goswami, U. (2013). A rhythmic musical intervention for poor readers: A
comparison of efficacy with a letter-based intervention. Mind, Brain, and Education 7(2), 113–
123.
Bigand, E., Vieillard, S., Madurell, F., Marozeau, J., & Dacquet, A. (2005). Multidimensional scaling
of emotional responses to music: The effect of musical expertise and of the duration of the
excerpts. Cognition & Emotion 19(8), 1113–1139.
Bispham, J. (2006). Rhythm in music: What is it? Who has it? And why? Music Perception: An
Interdisciplinary Journal 24(2), 125–134.
Brochard, R., Abecasis, D., Potter, D., Ragot, R., & Drake, C. (2003). The “ticktock” of our internal
clock: Direct brain evidence of subjective accents in isochronous sequences. Psychological
Science 14(4), 362–366.
Brown, S., & Volgsten, U. (Eds.). (2006). Music and manipulation: On the social uses and social
control of music. New York: Berghahn.
Brown-Lum, M., & Zwicker, J. G. (2015). Brain imaging increases our understanding of
developmental coordination disorder: A review of literature and future directions. Current
Developmental Disorders Reports 2(2), 131–140.
Brown-Lum, M., & Zwicker, J. G. (2017). Neuroimaging and occupational therapy: Bridging the gap
to advance rehabilitation in developmental coordination disorder. Journal of Motor Behavior
49(1), 98–110.
Buck, J. B. (1935). Synchronous flashing of fireflies experimentally induced. Science 81, 339–340.
Buck, J. B. (1937). Studies on the firefly. I. The effects of light and other agents on flashing in
Photinus pyralis, with special reference to periodicity and diurnal rhythm. Physiological Zoology
10(1), 45–58.
Buck, J. B. (1988). Synchronous rhythmic flashing of fireflies. II. Quarterly Review of Biology 63(3),
265–289.
Calderone, D. J., Lakatos, P., Butler, P. D., & Castellanos, F. X. (2014). Entrainment of neural
oscillations as a modifiable substrate of attention. Trends in Cognitive Sciences 18(6), 300–309.
Cason, N., Astésano, C., & Schön, D. (2015). Bridging music and speech rhythm: Rhythmic priming
and audio-motor training affect speech perception. Acta Psychologica 155, 43–50.
Cason, N., & Schön, D. (2012). Rhythmic priming enhances the phonological processing of speech.
Neuropsychologia 50(11), 2652–2658.
Chang, A., Bosnyak, D., & Trainor, L. J. (under review a). Beta oscillatory power modulation reflects
the predictability of pitch change.
Chang, A., Bosnyak, D., & Trainor, L. J. (2018). Beta oscillatory power modulation reflects the
predictability of pitch change. Cortex, 106, 248–260.
Chang, A., Chan, J., Li, Y.-C., Cairney, J., & Trainor, L. J. (2017). Auditory timing deficits in
developmental coordination disorder. Presented at the 1st Conference of the Timing Research
Forum, Strasbourg, France, October 23–25.
Chang, A., Livingstone, S. R., Bosnyak, D. J., & Trainor, L. J. (2017). Body sway reflects leadership
in joint music performance. Proceedings of the National Academy of Sciences 114(21), E4134–
E4141.
Chang, H. W., & Trehub, S. E. (1977). Infants’ perception of temporal grouping in auditory patterns.
Child Development 48(4), 1666–1670.
Cirelli, L. K. (2018). How interpersonal synchrony facilitates early prosocial behavior. Current
Opinion in Psychology 20, 35–39.
Cirelli, L. K., Bosnyak, D., Manning, F. C., Spinelli, C., Marie, C., Fujioka, T., … Trainor, L. J.
(2014). Beat-induced fluctuations in auditory cortical beta-band activity: Using EEG to measure
age-related changes. Frontiers in Psychology 5, 742. Retrieved from
https://doi.org/10.3389/fpsyg.2014.00742
Cirelli, L. K., Einarson, K. M., & Trainor, L. J. (2014). Interpersonal synchrony increases prosocial
behavior in infants. Developmental Science 17(6), 1003–1011.
Cirelli, L. K., Spinelli, C., Nozaradan, S., & Trainor, L. J. (2016). Measuring neural entrainment to
beat and meter in infants: Effects of music background. Frontiers in Neuroscience 10, 229.
Retrieved from https://doi.org/10.3389/fnins.2016.00229
Cirelli, L. K., Trehub, S. E., & Trainor, L. J. (2018). Rhythm and melody as social signals for infants.
Annals of the New York Academy of Sciences. doi:10.1111/nyas.13580
Cirelli, L. K., Wan, S. J., Johanis, T. C., & Trainor, L. J. (2018). Infants’ use of interpersonal
asynchrony as a signal for third-party affiliation. Music & Science 1.
doi:10.1177/2059204317745855
Cirelli, L. K., Wan, S. J., Spinelli, C., & Trainor, L. J. (2017). Effects of interpersonal movement
synchrony on infant helping behaviors: Is music necessary? Music Perception: An
Interdisciplinary Journal 34(3), 319–326.
Cirelli, L. K., Wan, S. J., & Trainor, L. J. (2014). Fourteen-month-old infants use interpersonal
synchrony as a cue to direct helpfulness. Philosophical Transactions of the Royal Society B:
Biological Sciences 369(1658), 20130400.
Clarke, E. F., & Krumhansl, C. L. (1990). Perceiving musical time. Music Perception: An
Interdisciplinary Journal 7(3), 213–251.
Cogo-Moreira, H., de Avila, C. R. B., Ploubidis, G. B., & de Jesus Mari, J. (2013). Effectiveness of
music education for the improvement of reading skills and academic achievement in young poor
readers: A pragmatic cluster-randomized, controlled clinical trial. PloS ONE 8(3), e59984.
Conrad, N. J., Walsh, J., Allen, J. M., & Tsang, C. D. (2011). Examining infants’ preferences for
tempo in lullabies and playsongs. Canadian Journal of Experimental Psychology/Revue
canadienne de psychologie expérimentale 65(3), 168–172.
Corbeil, M., Trehub, S. E., & Peretz, I. (2016). Singing delays the onset of infant distress. Infancy
21(3), 373–391.
Cunningham, J. G., & Sterling, R. S. (1988). Developmental change in the understanding of affective
meaning in music. Motivation and Emotion 12(4), 399–413.
Dalla Bella, S., Peretz, I., Rousseau, L., & Gosselin, N. (2001). A developmental study of the
affective value of tempo and mode in music. Cognition 80(3), B1–B10.
D’Ausilio, A., Novembre, G., Fadiga, L., & Keller, P. E. (2015). What can music tell us about social
interaction? Trends in Cognitive Sciences 19(3), 111–114.
Debrabant, J., Gheysen, F., Caeyenberghs, K., Van Waelvelde, H., & Vingerhoets, G. (2013). Neural
underpinnings of impaired predictive motor timing in children with developmental coordination
disorder. Research in Developmental Disabilities 34(5), 1478–1487.
de l’Etoile, S. K., and Leider, C. N. (2011). Acoustic parameters of infant-directed singing in mothers
with depressive symptoms. Infant Behavior and Development 34(2), 248–256.
Deliege, I. (1987). Grouping conditions in listening to music: An approach to Lerdahl & Jackendoff’s
grouping preference rules. Music Perception: An Interdisciplinary Journal 4(4), 325–359.
Demany, L., McKenzie, B., & Vurpillot, E. (1977). Rhythm perception in early infancy. Nature
266(5604), 718–719.
Ding, N., Melloni, L., Zhang, H., Tian, X., & Poeppel, D. (2016). Cortical tracking of hierarchical
linguistic structures in connected speech. Nature Neuroscience 19(1), 158–164.
Ding, N., Patel, A. D., Chen, L., Butler, H., Luo, C., & Poeppel, D. (2017). Temporal modulations in
speech and music. Neuroscience & Biobehavioral Reviews 81(B), 181–187.
Dissanayake, E. (2012). The earliest narratives were musical. Research Studies in Music Education
34(1), 3–14.
Dolgin, K. G., & Adelson, E. H. (1990). Age changes in the ability to interpret affect in sung and
instrumentally-presented melodies. Psychology of Music 18(1), 87–98.
Dowling, W. J. (1973). Rhythmic groups and subjective chunks in memory for melodies. Perception
& Psychophysics 14(1), 37–40.
Drake, C. (1993). Reproduction of musical rhythms by children, adult musicians, and adult
nonmusicians. Perception & Psychophysics 53(1), 25–33.
Drake, C., Jones, M. R., & Baruch, C. (2000). The development of rhythmic attending in auditory
sequences: Attunement, referent period, focal attending. Cognition 77(3), 251–288.
Dunfield, K., Kuhlmeier, V. A., O’Connell, L., & Kelley, E. (2011). Examining the diversity of
prosocial behavior: Helping, sharing, and comforting in infancy. Infancy 16(3), 227–247.
Eerola, T., Luck, G., & Toiviainen, P. (2006). An investigation of pre-schoolers’ corporeal
synchronization with music. In Proceedings of the 9th International Conference on Music
Perception and Cognition (pp. 472–476). The Society for Music Perception and Cognition and
European Society for the Cognitive Sciences of Music Bologna.
Einarson, K. M., & Trainor, L. J. (2013). Five-year-old children’s beat perception and beat
synchronization abilities. Frontiers in Human Neuroscience. Conference abstract: 14th Rhythm
Production and Perception Workshop, Birmingham, September 11–13.
Einarson, K. M., & Trainor, L. J. (2015). The effect of visual information on young children’s
perceptual sensitivity to musical beat alignment. Timing & Time Perception 3(1–2), 88–101.
Einarson, K. M., & Trainor, L. J. (2016). Hearing the beat: Young children’s perceptual sensitivity to
beat alignment varies according to metric structure. Music Perception: An Interdisciplinary
Journal 34(1), 56–70.
Endedijk, H. M., Ramenzoni, V. C., Cox, R. F., Cillessen, A. H., Bekkering, H., & Hunnius, S.
(2015). Development of interpersonal coordination between peers during a drumming task.
Developmental Psychology 51(5), 714–721.
Esposito, A., and Serio, M. (2007). Children’s perception of musical emotional expressions. In A.
Esposito, M. Faundez-Zanuy, E. Keller, & M. Marinaro (Eds.), Verbal and nonverbal
communication behaviours: COST Action 2102 international workshop, Vietri sul Mare, Italy,
March 29–31, 2007. Revised selected and invited papers (pp. 51–64). Berlin: Springer.
Essens, P. J. (1986). Hierarchical organization of temporal patterns. Perception & Psychophysics
40(2), 69–73.
Essens, P. J., & Povel, D. J. (1985). Metrical and nonmetrical representations of temporal patterns.
Perception & Psychophysics 37(1), 1–7.
Estil, L., Ingvaldsen, R., & Whiting, H. (2002). Spatial and temporal constraints on performance in
children with movement co-ordination problems. Experimental Brain Research 147(2), 153–161.
Falter, C. M., & Noreika, V. (2014). Time processing in developmental disorders: A comparative
view. In V. Arstila & D. Lloyd (Eds.), Subjective time: The philosophy, psychology, and
neuroscience of temporality (pp. 557–598). Cambridge, MA: MIT Press.
Fawcett, C., & Tunçgenç, B. (2017). Infants’ use of movement synchrony to infer social affiliation in
others. Journal of Experimental Child Psychology 160, 127–136.
Feldman, R. (2007). Parent–infant synchrony and the construction of shared timing: Physiological
precursors, developmental outcomes, and risk conditions. Journal of Child Psychology and
Psychiatry 48(3–4), 329–354.
Fernald, A. (1991). Prosody in speech to children: Prelinguistic and linguistic functions. Annals of
Child Development 8, 43–80.
Fitch, W. T. (2006). The biology and evolution of music: A comparative perspective. Cognition
100(1), 173–215.
Fitzpatrick, P., Schmidt, R. C., & Lockman, J. J. (1996). Dynamical patterns in the development of
clapping. Child Development 67(6), 2691–2708.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schön, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A randomized
control trial. PLoS ONE 10(9), e0138715.
Flaugnacco, E., Lopez, L., Terribili, C., Zoia, S., Buda, S., Tilli, S., … Schön, D. (2014). Rhythm
perception and production predict reading abilities in developmental dyslexia. Frontiers in Human
Neuroscience 8, 392. doi:10.3389/fnhum.2014.00392
Flom, R., Gentile, D. A., & Pick, A. D. (2008). Infants’ discrimination of happy and sad music.
Infant Behavior and Development 31(4), 716–728.
Flom, R., & Pick, A. D. (2012). Dynamics of infant habituation: Infants’ discrimination of musical
excerpts. Infant Behavior and Development 35(4), 697–704.
Fraisse, P. (1982). Rhythm and tempo. In D. Deutch (Ed.), The psychology of music (pp. 149–180).
San Diego: Academic Press.
Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., … Koelsch, S. (2009).
Universal recognition of three basic emotions in music. Current Biology 19(7), 573–576.
Fujioka, T., Ross, B., & Trainor, L. J. (2015). Beta-band oscillations represent auditory beat and its
metrical hierarchy in perception and imagery. Journal of Neuroscience 35(45), 15187–15198.
Fujioka, T., Trainor, L. J., Large, E. W., & Ross, B. (2012). Internalized timing of isochronous sounds
is represented in neuromagnetic beta oscillations. Journal of Neuroscience 32(5), 1791–1802.
Gabrielsson, A., & Juslin, P. N. (1996). Emotional expression in music performance: Between the
performer’s intention and the listener’s experience. Psychology of Music 24(1), 68–91.
Gabrielsson, A., & Lindström, E. (2010). The role of structure in the musical expression of emotions.
In P. Juslin and J. A. Sloboda (Eds.), Handbook of music and emotion: Theory, research,
applications (pp. 367–400). Oxford: Oxford University Press.
Gagnon, L., & Peretz, I. (2003). Mode and tempo relative contributions to “happy-sad” judgements
in equitone melodies. Cognition & Emotion 17(1), 25–40.
Gerardi, G. M., & Gerken, L. (1995). The development of affective responses to modality and
melodic contour. Music Perception: An Interdisciplinary Journal 12(3), 279–290.
Gerry, D. W., Faux, A. L., & Trainor, L. J. (2010). Effects of Kindermusik training on infants’
rhythmic enculturation. Developmental Science 13(3), 545–551.
Gerry, D. W., Unrau, A., & Trainor, L. J. (2012). Active music classes in infancy enhance musical,
communicative and social development. Developmental Science 15(3), 398–407.
Giomo, C. J. (1993). An experimental study of children’s sensitivity to mood in music. Psychology of
Music 21(2), 141–162.
Giraud, A. L., Kleinschmidt, A., Poeppel, D., Lund, T. E., Frackowiak, R. S., & Laufs, H. (2007).
Endogenous cortical rhythms determine cerebral specialization for speech perception and
production. Neuron 56(6), 1127–1134.
Giraud, A. L., & Poeppel, D. (2012). Cortical oscillations and speech processing: Emerging
computational principles and operations. Nature Neuroscience 15(4), 511–517.
Gjerdingen, R. O. (1989). Meter as a mode of attending: A network simulation of attentional
rhythmicity in music. Intégral 3, 67–91.
Gomez, A., & Sirigu, A. (2015). Developmental coordination disorder: Core sensori-motor deficits,
neurobiology and etiology. Neuropsychologia 79(B), 272–287.
Goswami, U. (2011). A temporal sampling framework for developmental dyslexia. Trends in
Cognitive Sciences 15(1), 3–10.
Goswami, U., Huss, M., Mead, N., Fosker, T., & Verney, J. P. (2013). Perception of patterns of
musical beat distribution in phonological developmental dyslexia: Significant longitudinal
relations with word reading and reading comprehension. Cortex 49(5), 1363–1376.
Grahn, J. A., & Brett, M. (2007). Rhythm and beat perception in motor areas of the brain. Journal of
Cognitive Neuroscience 19(5), 893–906.
Gregory, A. H., Worrall, L., & Sarge, A. (1996). The development of emotional responses to music in
young children. Motivation and Emotion 20(4), 341–348.
Habib, M., Lardy, C., Desiles, T., Commeiras, C., Chobert, J., & Besson, M. (2016). Music and
dyslexia: A new musical training method to improve reading and related disorders. Frontiers in
Psychology 7, 26. doi:10.3389/fpsyg.2016.00026
Háden, G. P., Németh, R., Török, M., & Winkler, I. (2015). Predictive processing of pitch trends in
newborn infants. Brain Research 1626, 14–20.
Haegens, S., & Zion Golumbic, E. (2018). Rhythmic facilitation of sensory processing: A critical
review. Neuroscience & Biobehavioral Reviews 86, 150–165.
Hannon, E. E., Vandenbosch der Nederlanden, C. M., & Tichko, P. (2012). Effects of perceptual
experience on children’s and adults’ perception of unfamiliar rhythms. Annals of the New York
Academy of Sciences 1252, 92–99.
Hannon, E. E., & Johnson, S. P. (2005). Infants use meter to categorize rhythms and melodies:
Implications for musical structure learning. Cognitive Psychology 50(4), 354–377.
Hannon, E. E., Nave-Blodgett, J. E., & Nave, K. M. (2018). The developmental origins of the
perception and production of musical rhythm. Child Development Perspectives.
doi:10.1111/cdep.12285
Hannon, E. E., Soley, G., & Ullal, S. (2012). Familiarity overrides complexity in rhythm perception:
A cross-cultural comparison of American and Turkish listeners. Journal of Experimental
Psychology: Human Perception and Performance 38(3), 543–548.
Hannon, E. E., & Trainor, L. J. (2007). Music acquisition: Effects of enculturation and formal
training on development. Trends in Cognitive Sciences 11(11), 466–472.
Hannon, E. E., & Trehub, S. E. (2005a). Metrical categories in infancy and adulthood. Psychological
Science 16(1), 48–55.
Hannon, E. E., & Trehub, S. E. (2005b). Tuning in to musical rhythms: Infants learn more readily
than adults. Proceedings of the National Academy of Sciences 102(35), 12639–12643.
Hardy, M. W., and LaGasse, A. B. (2013). Rhythm, movement, and autism: Using rhythmic
rehabilitation research as a model for autism. Frontiers in Integrative Neuroscience 7, 19.
Retrieved from https://doi.org/10.3389/fnint.2013.00019
He, C., Hotson, L., & Trainor, L. J. (2007). Mismatch responses to pitch changes in early infancy.
Journal of Cognitive Neuroscience 19(5), 878–892.
Henry, M. J., Herrmann, B., & Obleser, J. (2014). Entrained neural oscillations in multiple frequency
bands comodulate behavior. Proceedings of the National Academy of Sciences 111(41), 14935–
14940.
Henry, M. J., & Obleser, J. (2012). Frequency modulation entrains slow neural oscillations and
optimizes human listening behavior. Proceedings of the National Academy of Sciences 109(49),
20095–20100.
Herrmann, B., Henry, M. J., Haegens, S., & Obleser, J. (2016). Temporal expectations and neural
amplitude fluctuations in auditory cortex interactively influence perception. NeuroImage 124,
487–497.
Hove, M. J., & Risen, J. L. (2009). It’s all in the timing: Interpersonal synchrony increases affiliation.
Social Cognition 27(6), 949–960.
Huron, D. (2001). Is music an evolutionary adaptation? Annals of the New York Academy of Sciences
930(1), 43–61.
Huron, D. B. (2006). Sweet anticipation: Music and the psychology of expectation. Cambridge, MA:
MIT Press.
Huss, M., Verney, J. P., Fosker, T., Mead, N., & Goswami, U. (2011). Music, rhythm, rise time
perception and developmental dyslexia: Perception of musical meter predicts reading and
phonology. Cortex 47(6), 674–689.
Ilari, B. (2015). Rhythmic engagement with music in early childhood: A replication and extension.
Journal of Research in Music Education 62(4), 332–343.
Ilari, B. (2016). Music in the early years: Pathways into the social world. Research Studies in Music
Education 38(1), 23–39.
Ilie, G., & Thompson, W. F. (2006). A comparison of acoustic cues in music and speech for three
dimensions of affect. Music Perception: An Interdisciplinary Journal 23(4), 319–330.
Isaksson, S., Salomäki, S., Tuominen, J., Arstila, V., Falter-Wagner, C. M., & Noreika, V. (2018). Is
there a generalized timing impairment in autism spectrum disorders across time scales and
paradigms? Journal of Psychiatric Research 99, 111–121.
Iversen, J. R., & Patel, A. D. (2008). The Beat Alignment Test (BAT): Surveying beat processing
abilities in the general population. In K. Miyazaki, M. Adachi, Y. Hiraga, Y. Nakajima, & M.
Tsuzaki (Eds.), Proceedings of the 10th International Conference on Music Perception and
Cognition (ICMPC 10) (pp. 465–468).
Iversen, J. R., Repp, B. H., & Patel, A. D. (2009). Top-down control of rhythm perception modulates
early auditory responses. Annals of the New York Academy of Sciences 1169, 58–73.
Iversen, S., Berg, K., Ellertsen, B., & Tønnessen, F. E. (2005). Motor coordination difficulties in a
municipality group and in a clinical sample of poor readers. Dyslexia 11(3), 217–231.
James, C. E., Michel, C. M., Britz, J., Vuilleumier, P., & Hauert, C. A. (2012). Rhythm evokes action:
Early processing of metric deviances in expressive music by experts and laymen revealed by ERP
source imaging. Human Brain Mapping 33(12), 2751–2767.
Jones, M. R., & Boltz, M. (1989). Dynamic attending and responses to time. Psychological Review
96(3), 459–491.
Jones, M. R., Moynihan, H., MacKenzie, N., & Puente, J. (2002). Temporal aspects of stimulus-
driven attending in dynamic arrays. Psychological Science 13(4), 313–319.
Jusczyk, P. W., & Krumhansl, C. L. (1993). Pitch and rhythmic patterns affecting infants’ sensitivity
to musical phrase structure. Journal of Experimental Psychology: Human Perception and
Performance 19(3), 627–640.
Juslin, P. N. (1997). Emotional communication in music performance: A functionalist perspective
and some data. Music Perception: An Interdisciplinary Journal 14(4), 383–418.
Juslin, P. N. (2000). Cue utilization in communication of emotion in music performance: Relating
performance to perception. Journal of Experimental Psychology: Human Perception and
Performance 26(6), 1797–1813.
Juslin, P. N., & Laukka, P. (2003). Emotional expression in speech and music. Annals of the New
York Academy of Sciences 1000, 279–282.
Juslin, P. N., & Timmers, R. (2010). Expression and communication of emotion in music
performance. In P. Juslin and J. A. Sloboda (Eds.), Handbook of music and emotion: Theory,
research, applications (pp. 453–489). Oxford: Oxford University Press.
Kastner, M. P., & Crowder, R. G. (1990). Perception of the major/minor distinction: IV. Emotional
connotations in young children. Music Perception: An Interdisciplinary Journal 8(2), 189–201.
Keller, P. E., Novembre, G., & Hove, M. J. (2014). Rhythm in joint action: Psychological and
neurophysiological mechanisms for real-time interpersonal coordination. Philosophical
Transactions of the Royal Society B: Biological Sciences 369(1658), 20130394.
doi:10.1098/rstb.2013.0394
Kelso, J. A., Saltzman, E. L., & Tuller, B. (1986). The dynamical perspective on speech production:
Data and theory. Journal of Phonetics 14(1), 29–59.
King-Dowling, S., Missiuna, C., Rodriguez, M. C., Greenway, M., & Cairney, J. (2015). Reprint of
“Co-occurring motor, language and emotional–behavioral problems in children 3–6 years of age.”
Human Movement Science 42, 344–351.
Kirschner, S., & Ilari, B. (2014). Joint drumming in Brazilian and German preschool children:
Cultural differences in rhythmic entrainment, but no prosocial effects. Journal of Cross-Cultural
Psychology 45(1), 137–166.
Kirschner, S., & Tomasello, M. (2009). Joint drumming: Social context facilitates synchronization in
preschool children. Journal of Experimental Child Psychology 102(3), 299–314.
Kokal, I., Engel, A., Kirschner, S., & Keysers, C. (2011). Synchronized drumming enhances activity
in the caudate and facilitates prosocial commitment—if the rhythm comes easily. PLoS ONE 6(11),
e27272.
Kragness, H. E., Baksh, A., Battcock, A., & Trainor, L. J. (2017). Children’s use of expressive cues
in music: A developmental self-pacing study. Presented at the Neurosciences & Music VI
conference, Boston, USA, June 15–18.
Kragness, H. E., & Trainor, L. J. (2016). Listeners lengthen phrase boundaries in self-paced music.
Journal of Experimental Psychology: Human Perception and Performance 42(10), 1676–1686.
Kragness, H. E., & Trainor, L. J. (2018). Young children pause on phrase boundaries in self-paced
music listening: The role of harmonic cues. Developmental Psychology 54(5), 842–856.
Kratus, J. (1993). A developmental study of children’s interpretation of emotion in music.
Psychology of Music 21(1), 3–19.
Krumhansl, C. L., & Jusczyk, P. W. (1990). Infants’ perception of phrase structure in music.
Psychological Science 1(1), 70–73.
Kuhl, P. K., Stevens, E., Hayashi, A., Deguchi, T., Kiritani, S., & Iverson, P. (2006). Infants show a
facilitation effect for native language phonetic perception between 6 and 12 months.
Developmental Science 9(2), F13–F21.
Large, E. W., & Jones, M. R. (1999). The dynamics of attending: How people track time-varying
events. Psychological Review 106(1), 119–159.
Laukka, P., & Gabrielsson, A. (2000). Emotional expression in drumming performance. Psychology
of Music 28(2), 181–189.
Launay, J., Dean, R. T., & Bailes, F. (2013). Synchronization can influence trust following virtual
interaction. Experimental Psychology 60(1), 1–11.
Lerdahl, F., & Jackendoff, R. (1983). A generative theory of tonal music. Cambridge, MA: MIT
Press.
Lewkowicz, D. J. (2003). Learning and discrimination of audiovisual events in human infants: The
hierarchical relation between intersensory temporal synchrony and rhythmic pattern cues.
Developmental Psychology 39(5), 795–804.
Lindenberger, U., Li, S. C., Gruber, W., & Müller, V. (2009). Brains swinging in concert: Cortical
phase synchronization while playing guitar. BMC Neuroscience 10(1), 22.
London, J. (2004). Hearing in time: Psychological aspects of musical meter (2nd ed.). New York:
Oxford University Press.
Longhi, S. (2009). Bloch oscillations in complex crystals with PT symmetry. Physical Review Letters
103(12), 123601.
Luck, G., & Toiviainen, P. (2006). Ensemble musicians’ synchronization with conductors’ gestures:
An automated feature-extraction analysis. Music Perception: An Interdisciplinary Journal 24(2),
189–200.
Lynch, M. P., & Eilers, R. E. (1992). A study of perceptual development for musical tuning.
Perception & Psychophysics 52(6), 599–608.
Lynch, M. P., Eilers, R. E., Oller, D. K., & Urbano, R. C. (1990). Innateness, experience, and music
perception. Psychological Science 1(4), 272–276.
McAuley, J. D., Jones, M. R., Holub, S., Johnston, H. M., & Miller, N. S. (2006). The time of our
lives: Life span development of timing and event tracking. Journal of Experimental Psychology:
General 135(3), 348–367.
McLeod, K. R., Langevin, L. M., Goodyear, B. G., & Dewey, D. (2014). Functional connectivity of
neural motor networks is disrupted in children with developmental coordination disorder and
attention-deficit/hyperactivity disorder. NeuroImage: Clinical 4, 566–575.
McNeill, W. H. (1995). Keeping together in time. Boston, MA: Harvard University Press.
Macrae, C. N., Duffy, O. K., Miles, L. K., & Lawrence, J. (2008). A case of hand waving: Action
synchrony and person perception. Cognition 109(1), 152–156.
Malloch, S. E., & Trevarthen, C. E. (2009). Communicative musicality: Exploring the basis of human
companionship. New York: Oxford University Press.
Merchant, H., Grahn, J., Trainor, L., Rohrmeier, M., & Fitch, W. T. (2015). Finding the beat: A
neural perspective across humans and non-human primates. Philosophical Transactions of the
Royal Society B: Biological Sciences 370(1664), 20140093. doi:10.1098/rstb.2014.0093
Merchant, H., & Honing, H. (2014). Are non-human primates capable of rhythmic entrainment?
Evidence for the gradual audiomotor evolution hypothesis. Frontiers in Neuroscience 7, 274.
Retrieved from https://doi.org/10.3389/fnins.2013.00274
Merker, B. (2000). Synchronous chorusing and human origins. In N. L. Wallin, B. Merker, & S.
Brown (Eds.), The origins of music (pp. 315–327). Cambridge, MA: MIT Press.
Meyer, L. B. (1956). Emotion and meaning in music. Chicago, IL: University of Chicago Press.
Mohn, C., Argstatter, H., & Wilker, F. W. (2011). Perception of six basic emotions in music.
Psychology of Music 39(4), 503–517.
Morillon, B., & Schroeder, C. E. (2015). Neuronal oscillations as a mechanistic substrate of auditory
temporal prediction. Annals of the New York Academy of Sciences 1337, 26–31.
Morillon, B., Schroeder, C. E., Wyart, V., & Arnal, L. H. (2016). Temporal prediction in lieu of
periodic stimulation. Journal of Neuroscience 36(8), 2342–2347.
Mote, J. (2011). The effects of tempo and familiarity on children’s affective interpretation of music.
Emotion 11(3), 618–622.
Nakata, T., & Mitani, C. (2005). Influences of temporal fluctuation on infant attention. Music
Perception: An Interdisciplinary Journal 22(3), 401–409.
Nakata, T., & Trainor, L. J. (2015). Perceptual and cognitive enhancement with an adaptive timing
partner: Electrophysiological responses to pitch change. Psychomusicology: Music, Mind, and
Brain 25(4), 404–415.
Nawrot, E. S. (2003). The perception of emotional expression in music: Evidence from infants,
children and adults. Psychology of Music 31(1), 75–92.
Nobre, A. C., Correa, A., & Coull, J. T. (2007). The hazards of time. Current Opinion in
Neurobiology 17(4), 465–470.
Nobre, A. C., & van Ede, F. (2018). Anticipated moments: Temporal structure in attention. Nature
Reviews Neuroscience 19(1), 34–48.
Nozaradan, S. (2014). Exploring how musical rhythm entrains brain activity with
electroencephalogram frequency-tagging. Philosophical Transactions of the Royal Society B:
Biological Sciences 369(1658), 20130393. doi:10.1098/rstb.2013.0393
Nozaradan, S., Peretz, I., Missal, M., & Mouraux, A. (2011). Tagging the neuronal entrainment to
beat and meter. Journal of Neuroscience 31(28), 10234–10240.
Nozaradan, S., Peretz, I., & Mouraux, A. (2012). Selective neuronal entrainment to the beat and
meter embedded in a musical rhythm. Journal of Neuroscience 32(49), 17572–17581.
Overy, K., Nicolson, R. I., Fawcett, A. J., & Clarke, E. F. (2003). Dyslexia and music: Measuring
musical timing skills. Dyslexia 9(1), 18–36.
Palmer, C. (1989). Mapping musical thought to musical performance. Journal of Experimental
Psychology: Human Perception and Performance 15(2), 331–346.
Palmer, C., & Krumhansl, C. L. (1987). Pitch and temporal contributions to musical phrase
perception: Effects of harmony, performance timing, and familiarity. Perception & Psychophysics
41(6), 505–518.
Papoušek, M., Papoušek, H., & Symmes, D. (1991). The meanings of melodies in motherese in tone
and stress languages. Infant Behavior and Development 14(4), 415–440.
Partridge, B. L. (1982). The structure and function of fish schools. Scientific American 246(6), 114–
123.
Patel, A. D., & Iversen, J. R. (2014). The evolutionary neuroscience of musical beat perception: The
Action Simulation for Auditory Prediction (ASAP) hypothesis. Frontiers in Systems Neuroscience
8, 57. Retrieved from https://doi.org/10.3389/fnsys.2014.00057
Patel, A. D., Iversen, J. R., Bregman, M. R., & Schulz, I. (2009). Experimental evidence for
synchronization to a musical beat in a nonhuman animal. Current Biology 19(10), 827–830.
Pearce, M. T., Müllensiefen, D., & Wiggins, G. A. (2010). The role of expectation and probabilistic
learning in auditory boundary perception: A model comparison. Perception 39(10), 1367–1391.
Pearce, M. T., & Wiggins, G. A. (2006). Expectation in melody: The influence of context and
learning. Music Perception: An Interdisciplinary Journal 23(5), 377–405.
Peelle, J. E., & Davis, M. H. (2012). Neural oscillations carry speech rhythm through to
comprehension. Frontiers in Psychology 3, 320. Retrieved from
https://doi.org/10.3389/fpsyg.2012.00320
Peretz, I. (1989). Clustering in music: An appraisal of task factors. International Journal of
Psychology 24(1–5), 157–178.
Phillips-Silver, J., & Trainor, L. J. (2005). Feeling the beat: Movement influences infant rhythm
perception. Science 308(5727), 1430.
Piek, J. P., & Dyck, M. J. (2004). Sensory-motor deficits in children with developmental coordination
disorder, attention deficit hyperactivity disorder and autistic disorder. Human Movement Science
23(3–4), 475–488.
Provasi, J., & Bobin-Bègue, A. (2003). Spontaneous motor tempo and rhythmical synchronisation in
2½- and 4-year-old children. International Journal of Behavioral Development 27(3), 220–231.
Przybylski, L., Bedoin, N., Krifi-Papoz, S., Herbillon, V., Roch, D., Léculier, L., … Tillmann, B.
(2013). Rhythmic auditory stimulation influences syntactic processing in children with
developmental language disorders. Neuropsychology 27(1), 121–131.
Rabinowitch, T. C., & Meltzoff, A. N. (2017). Synchronized movement experience enhances peer
cooperation in preschool children. Journal of Experimental Child Psychology 160, 21–32.
Rankin, S. K., Large, E. W., & Fink, P. W. (2009). Fractal tempo fluctuation and pulse prediction.
Music Perception: An Interdisciplinary Journal 26(5), 401–413.
Reiersen, A. M., Constantino, J. N., & Todd, R. D. (2008). Co-occurrence of motor problems and
autistic symptoms in attention-deficit/hyperactivity disorder. Journal of the American Academy of
Child & Adolescent Psychiatry 47(6), 662–672.
Repp, B. H. (1992). Diversity and commonality in music performance: An analysis of timing
microstructure in Schumann’s “Träumerei.” Journal of the Acoustical Society of America 92(5),
2546–2568.
Repp, B. H. (2005). Sensorimotor synchronization: A review of the tapping literature. Psychonomic
Bulletin & Review 12(6), 969–992.
Repp, B. H., London, J., & Keller, P. E. (2005). Production and synchronization of uneven rhythms at
fast tempi. Music Perception: An Interdisciplinary Journal 23(1), 61–78.
Repp, B. H., & Su, Y. H. (2013). Sensorimotor synchronization: A review of recent research (2006–
2012). Psychonomic Bulletin & Review 20(3), 403–452.
Rock, A. M., Trainor, L. J., & Addison, T. L. (1999). Distinctive messages in infant-directed lullabies
and play songs. Developmental Psychology 35(2), 527–534.
Rosenblum, S., & Regev, N. (2013). Timing abilities among children with developmental
coordination disorders (DCD) in comparison to children with typical development. Research in
Developmental Disabilities 34(1), 218–227.
Russell, J. A. (1980). A circumplex model of affect. Journal of Personality and Social Psychology
39(6), 1161–1178.
Sänger, J., Müller, V., & Lindenberger, U. (2012). Intra- and interbrain synchronization and network
properties when playing guitar in duets. Frontiers in Human Neuroscience 6, 312. Retrieved from
https://doi.org/10.3389/fnhum.2012.00312
Schachner, A., Brady, T. F., Pepperberg, I. M., & Hauser, M. D. (2009). Spontaneous motor
entrainment to music in multiple vocal mimicking species. Current Biology 19(10), 831–836.
Schön, D., & Tillmann, B. (2015). Short- and long-term rhythmic interventions: Perspectives for
language rehabilitation. Annals of the New York Academy of Sciences 1337, 32–39.
Schroeder, C. E., & Lakatos, P. (2009). Low-frequency neuronal oscillations as instruments of
sensory selection. Trends in Neurosciences 32(1), 9–18.
Schröger, E., Marzecová, A., & SanMiguel, I. (2015). Attention and prediction in human audition: A
lesson from cognitive psychophysiology. European Journal of Neuroscience 41(5), 641–664.
Shenfield, T., Trehub, S. E., & Nakata, T. (2003). Maternal singing modulates infant arousal.
Psychology of Music 31(4), 365–375.
Snyder, J. S., Hannon, E. E., Large, E. W., & Christiansen, M. H. (2006). Synchronization and
continuation tapping to complex meters. Music Perception: An Interdisciplinary Journal 24(2),
135–146.
Snyder, J. S., & Large, E. W. (2005). Gamma-band activity reflects the metric structure of rhythmic
tone sequences. Cognitive Brain Research 24(1), 117–126.
Soley, G., & Hannon, E. E. (2010). Infants prefer the musical meter of their own culture: A cross-
cultural comparison. Developmental Psychology 46(1), 286–292.
Stefanics, G., Hangya, B., Hernádi, I., Winkler, I., Lakatos, P., & Ulbert, I. (2010). Phase entrainment
of human delta oscillations can mediate the effects of expectation on reaction speed. Journal of
Neuroscience 30(41), 13578–13585.
Tal, I., Large, E. W., Rabinovitch, E., Wei, Y., Schroeder, C. E., Poeppel, D., & Golumbic, E. Z.
(2017). Neural entrainment to the beat: The “missing-pulse” phenomenon. Journal of
Neuroscience 37(26), 6331–6341.
Tarr, B., Launay, J., & Dunbar, R. I. (2014). Music and social bonding: “Self-other” merging and
neurohormonal mechanisms. Frontiers in Psychology 5, 1096. Retrieved from
https://doi.org/10.3389/fpsyg.2014.01096
Taub, G. E., McGrew, K. S., & Keith, T. Z. (2007). Improvements in interval time tracking and
effects on reading achievement. Psychology in the Schools 44(8), 849–863.
Terwogt, M. M., & Van Grinsven, F. (1991). Musical expression of moodstates. Psychology of Music
19(2), 99–109.
Thomson, J. M., Fryer, B., Maltby, J., & Goswami, U. (2006). Auditory and motor rhythm awareness
in adults with dyslexia. Journal of Research in Reading 29(3), 334–348.
Thomson, J. M., & Goswami, U. (2008). Rhythmic processing in children with developmental
dyslexia: Auditory and motor rhythms link to reading and spelling. Journal of Physiology-Paris
102(1–3), 120–129.
Thomson, J. M., Leong, V., & Goswami, U. (2013). Auditory processing interventions and
developmental dyslexia: A comparison of phonemic and rhythmic approaches. Reading and
Writing 26(2), 139–161.
Thorpe, L. A., & Trehub, S. E. (1989). Duration illusion and auditory grouping in infancy.
Developmental Psychology 25(1), 122–127.
Todd, N. (1985). A model of expressive timing in tonal music. Music Perception: An
Interdisciplinary Journal 3(1), 33–57.
Toplak, M. E., Dockstader, C., & Tannock, R. (2006). Temporal information processing in ADHD:
Findings to date and new methods. Journal of Neuroscience Methods 151(1), 15–29.
Trainor, L. J. (1996). Infant preferences for infant-directed versus noninfant-directed playsongs and
lullabies. Infant Behavior and Development 19(1), 83–92.
Trainor, L. J. (2012). Musical experience, plasticity, and maturation: Issues in measuring
developmental change using EEG and MEG. Annals of the New York Academy of Sciences 1252,
25–36.
Trainor, L. J. (2015). The origins of music in auditory scene analysis and the roles of evolution and
culture in musical creation. Philosophical Transactions of the Royal Society B: Biological Sciences
370(1664), 20140089. doi: 10.1098/rstb.2014.0089
Trainor, L. J., & Adams, B. (2000). Infants’ and adults’ use of duration and intensity cues in the
segmentation of tone patterns. Perception & Psychophysics 62(2), 333–340.
Trainor, L. J., Chang, A., Cairney, J., Li, Y. C. (2018). Is auditory perceptual timing a core deficit of
developmental coordination disorder? Annals of the New York Academy of Sciences 1423, 30–39
Trainor, L. J., & Cirelli, L. (2015). Rhythm and interpersonal synchrony in early social development.
Annals of the New York Academy of Sciences 1337, 45–52.
Trainor, L. J., Clark, E. D., Huntley, A., & Adams, B. A. (1997). The acoustic basis of preferences for
infant-directed singing. Infant Behavior and Development 20(3), 383–396.
Trainor, L. J., & Corrigall, K. A. (2010). Music acquisition and effects of musical experience. In M.
Riess Jones, R. Fay, & A. Popper (Eds.), Music perception (pp. 89–127). New York: Springer.
Trainor, L. J., & Hannon, E. E. (2013). Musical development. In D. Deutsch (Ed.), The psychology of
music (3rd ed., pp. 423–497). London: Academic Press.
Trainor, L. J., & He, C. (2013). Auditory and musical development. In P. D. Zelazo (Ed.), The Oxford
handbook of developmental psychology, Vol. 1: Body and mind (pp. 310–337). Oxford: Oxford
University Press.
Trainor, L. J., Marie, C., Gerry, D., Whiskin, E., & Unrau, A. (2012). Becoming musically
enculturated: Effects of music classes for infants on brain and behavior. Annals of the New York
Academy of Sciences 1252, 129–138.
Trainor, L. J., & Trehub, S. E. (1992). A comparison of infants’ and adults’ sensitivity to western
musical structure. Journal of Experimental Psychology: Human Perception and Performance
18(2), 394–402.
Trainor, L. J., & Trehub, S. E. (1994). Key membership and implied harmony in Western tonal
music: Developmental perspectives. Perception & Psychophysics 56(2), 125–132.
Trainor, L. J., & Unrau, A. (2012). Development of pitch and music perception. In L. Werner, R. R.
Fay, & A. N. Popper (Eds.), Springer handbook of auditory research: Human auditory
development (pp. 223–254). New York: Springer.
Trainor, L. J., & Zatorre, R. J. (2015). The neurobiology of musical expectations from perception to
emotion. In S. Hallam, I. Cross, & M. Thaut (Eds.), The Oxford handbook of music psychology
(2nd ed., pp. 285–306). Oxford: Oxford University Press.
Trehub, S. E., & Thorpe, L. A. (1989). Infants’ perception of rhythm: Categorization of auditory
sequences by temporal structure. Canadian Journal of Psychology/Revue canadienne de
psychologie 43(2), 217–229.
Trehub, S. E., & Trainor, L. J. (1998). Singing to infants: Lullabies and play songs. Advances in
Infancy Research 12, 43–78.
Trehub, S. E., Unyk, A. M., Kamenetsky, S. B., Hill, D. S., Trainor, L. J., Henderson, J. L., & Saraza,
M. (1997). Mothers’ and fathers’ singing to infants. Developmental Psychology 33(3), 500–507.
Trehub, S. E., Unyk, A. M., & Trainor, L. J. (1993). Adults identify infant-directed music across
cultures. Infant Behavior and Development 16(2), 193–211.
Tunçgenç, B., & Cohen, E. (2016). Movement synchrony forges social bonds across group divides.
Frontiers in Psychology 7, 782. Retrieved from https://doi.org/10.3389/fpsyg.2016.00782
Tunçgenç, B., Cohen, E., & Fawcett, C. (2015). Rock with me: The role of movement synchrony in
infants’ social and nonsocial choices. Child Development 86(3), 976–984.
Valdesolo, P., & DeSteno, D. (2011). Synchrony and the social tuning of compassion. Emotion 11(2),
262–266.
Valdesolo, P., Ouyang, J., & DeSteno, D. (2010). The rhythm of joint action: Synchrony promotes
cooperative ability. Journal of Experimental Social Psychology 46(4), 693–695.
van Ede, F., Niklaus, M., & Nobre, A. C. (2017). Temporal expectations guide dynamic prioritization
in visual working memory through attenuated α oscillations. Journal of Neuroscience 37(2), 437–
445.
Van Noorden, L., & De Bruyn, L. (2009). The development of synchronization skills of children 3 to
11 years old. In Proceedings of ESCOM—7th Triennial Conference of the European Society for the
Cognitive Sciences of Music. Jyväskylä, Finland: University of Jyväskylä.
Volman, M. C. J., & Geuze, R. H. (1998). Relative phase stability of bimanual and visuomanual
rhythmic coordination patterns in children with a developmental coordination disorder. Human
Movement Science 17(4–5), 541–572.
Warneken, F., & Tomasello, M. (2006). Altruistic helping in human infants and young chimpanzees.
Science 311(5765), 1301–1303.
Warneken, F., & Tomasello, M. (2007). Helping and cooperation at 14 months of age. Infancy 11(3),
271–294.
Warneken, F., & Tomasello, M. (2009). Varieties of altruism in children and chimpanzees. Trends in
Cognitive Sciences 13(9), 397–402.
Watemberg, N., Waiserberg, N., Zuk, L., & Lerman-Sagie, T. (2007). Developmental coordination
disorder in children with attention-deficit-hyperactivity disorder and physical therapy intervention.
Developmental Medicine & Child Neurology 49(12), 920–925.
Weimerskirch, H., Martin, J., Clerquin, Y., Alexandre, P., & Jiraskova, S. (2001). Energy saving in
flight formation. Nature 413(6857), 697–698.
Werker, J. F., & Tees, R. C. (2005). Speech perception as a window for understanding plasticity and
commitment in language systems of the brain. Developmental Psychobiology 46(3), 233–251.
Wieland, E. A., McAuley, J. D., Dilley, L. C., & Chang, S. E. (2015). Evidence for a rhythm
perception deficit in children who stutter. Brain & Language 144, 26–34.
Williams, H. G., Woollacott, M. H., & Ivry, R. (1992). Timing and motor control in clumsy children.
Journal of Motor Behavior 24(2), 165–172.
Williams, J., Thomas, P. R., Maruff, P., Butson, M., & Wilson, P. H. (2006). Motor, visual and
egocentric transformations in children with developmental coordination disorder. Child: Care,
Health and Development 32(6), 633–647.
Wilmut, K., & Wann, J. (2008). The use of predictive information is impaired in the actions of
children and young adults with developmental coordination disorder. Experimental Brain Research
191(4), 403–418.
Wilson, P. H., Ruddock, S., Smits-Engelsman, B., Polatajko, H., & Blank, R. (2013). Understanding
performance deficits in developmental coordination disorder: A meta-analysis of recent research.
Developmental Medicine & Child Neurology 55(3), 217–228.
Wiltermuth, S. S., & Heath, C. (2009). Synchrony and cooperation. Psychological Science 20(1), 1–
5.
Winkler, I., Háden, G. P., Ladinig, O., Sziller, I., & Honing, H. (2009). Newborn infants detect the
beat in music. Proceedings of the National Academy of Sciences 106(7), 2468–2471.
Wolff, P. H. (2002). Timing precision and rhythm in developmental dyslexia. Reading and Writing
15(1–2), 179–206.
Woolhouse, M., Tidhar, D., Demorest, S., Morrison, S., & Campbell, P. (2010). Group dancing leads
to increased person-perception. In Proceedings of the 11th International Conference on Music
Perception and Cognition (pp. 605–608). Seattle, WA: University of Washington.
Yoshida, K. A., Iversen, J. R., Patel, A. D., Mazuka, R., Nito, H., Gervain, J., & Werker, J. F. (2010).
The development of perceptual grouping biases in infancy: A Japanese–English cross-linguistic
study. Cognition 115(2), 356–361.
Zatorre, R. J., Chen, J. L., & Penhune, V. B. (2007). When the brain plays music: Auditory–motor
interactions in music perception and production. Nature Reviews Neuroscience 8(7), 547–558.
Zentner, M., & Eerola, T. (2010). Rhythmic engagement with music in infancy. Proceedings of the
National Academy of Sciences 107(13), 5768–5773.
Zwicker, J. G., Missiuna, C., & Boyd, L. A. (2009). Neural correlates of developmental coordination
disorder: A review of hypotheses. Journal of Child Neurology 24(10), 1273–1281.
CHAPT E R 25
L A U R A F E R R E R I, A L I N E MO U S S A R D, E MMA N U E L
B I G A N D, A N D B A R B A R A T I L L MA N N
N A : M C
P W -B
P A : N -
I T R
M
Memory
Because of degeneration in medio-temporal and prefrontal lobes, encoding
and retrieving information is one of the most affected abilities in AD
dementia and of particular interest for music interventions. Several findings
revealed that dementia patients with severe memory deficits can show a
surprisingly robust musical memory (see Baird & Samson, 2015). A
possible neuroanatomical explanation is that brain regions specifically
involved in musical memory, such as the caudal anterior cingulate and the
ventral pre-supplementary motor area, are also regions relatively spared in
the first stages of AD (Jacobsen et al., 2015). The question thus arises as to
whether music can be used to promote the encoding and retrieval of non-
musical material in dementia patients. It has been shown that presenting
verbal information in a sung, rather than spoken version, improves its later
recognition after an immediate (Simmons-Stern, Budson, & Ally, 2010;
Simmons-Stern et al., 2012) or delayed (i.e., four weeks, Moussard, Bigand,
Belleville, & Peretz, 2012, 2014a) recall at a mild stage of AD. This
supports the hypothesis that music may act as a good anchor point for
verbal information, in turn enhancing its retrieval (see Ferreri & Verga,
2016). Different results come from Baird and colleagues (Baird, Samson,
Miller, & Chalmers, 2017a), where no beneficial effects of the sung versus
spoken modality were observed on subsequent immediate (30 min) and
delayed (24 hours) recall. These contrasting findings might be due to
experimental differences. As discussed by Baird et al. (2017a), using only
one session of learning, rather than several learning sessions over weeks
(Moussard et al., 2012, 2014a), may have hindered revealing the potential
benefit of music. Thus, it is possible that music could represent an efficient
mnemonic to learn or relearn information in early stages of AD, but only if
there are sufficient time and learning trials allocated to encoding this
information.
Other findings based on non-verbal memory tasks support music-related
memory improvement in dementia. In another study by Moussard and
colleagues (Moussard, Bigand, Belleville, & Peretz, 2014b), mild AD
participants were asked to learn a sequence of gestures in synchrony (i.e.,
shadowing the experimenter) or not (i.e., observing the experimenter),
performed with music or a metronome beat. Results showed that AD
patients learned better in the music condition when tested after an
immediate (but not delayed, i.e., 10 min) recall. Several studies also showed
that music facilitates the recall of autobiographical (i.e., personal) memories
in people suffering from AD. El Haj and colleagues (El Haj, Fasotti, &
Allain, 2012a; El Haj, Clément, Fasotti, & Allain, 2013) found higher
quality (in terms of speed of recall, content specificity, and grammatical
complexity) of autobiographical memories when recalled after having
listened to music. These studies compared music (selected by the patients or
by the experimenter) to a silent control condition. Critically, Foster and
Valentine (2001) showed that a beneficial effect of background sound on
autobiographical memory for recent (but not for remote) events was also
observed for cafeteria noise (instead of music). Hence, it could be argued
that not the music itself, but rather a more general auditory stimulation
drives the observed positive effect. However, in a further study, El Haj and
colleagues (El Haj, Postal, & Allain, 2012b) found better autobiographical
memories after exposure to patients-selected music compared to another
musical condition (i.e., Vivaldi’s Four Seasons), thus supporting the pivotal
and probably particularly arousing role of personally relevant music
stimulation (see also Lord & Garner, 1993). Remarkably, most of these
studies also reported an emotional component: in AD patients, preferred
music-triggered autobiographical memories were rated higher in emotional
content (El Haj et al., 2012a), and with prevalence of positive over negative
content (El Haj et al., 2012b), than memories retrieved during silence or
with music chosen by the experimenter. Consistent with the hypothesis that
music-driven modulations in emotions can enhance memory (Jäncke,
2008), these findings suggest that music may enhance autobiographical
recall by promoting positive emotional memories (El Haj et al., 2012b; see
Irish et al., 2006, for alternative explanation based on anxiety reduction).
Language
While numerous studies have investigated the role of music on memory in
dementia, few investigations exist in the language domain. Findings from
studies investigating non-degenerative disease related to brain traumas or
vascular problems, such as stroke, suggest that music can be a useful tool
for language rehabilitation. Melodic intonation therapy (MIT), a speech
therapy based on the potential rehabilitative effects of singing and rhythmic
movement, has been described as a valid intervention for improving post-
stroke aphasia in younger and older adults (e.g., Belin et al., 1996; Racette,
Bard, & Peretz, 2006; see Zumbansen, Peretz, & Hébert, 2014 for a
review). Consistently with the hypothesis that music stimulation may
enhance cognitive functions by promoting brain connectivity, Schlaug and
colleagues (Schlaug, Marchina, & Norton, 2009) showed increased number
of fibers and volume of the arcuate fasciculus comparing patients’ brain
pre- and post-MIT treatment. Crucially, the arcuate fasciculus is a white
matter tract connecting several regions involved in language production,
such as superior temporal lobe, premotor regions and posterior inferior
frontal gyrus, and primary motor cortex.
Although to our knowledge no study has explored MIT interventions in
dementia care, positive results of music interventions on language function
have also been found in AD patients. Brotons and Koger (2000) observed
that four sessions of music therapy (i.e., listening to songs related to a
conversation topic) significantly improved both speech content and fluency
in a group of ten AD patients. However, considering the small sample size
of the study and the lack of similar investigations, further research is needed
to better understand the effect of musical interventions on language in
dementia. This becomes particularly relevant when considering the
progressive loss of language functioning usually observed in dementias,
such as AD.
Motor Functions
Although findings on AD patients suggest that music may help motor
functions by enhancing gesture learning (Moussard et al., 2014b), the
investigation of motor rehabilitation through music mainly concerns the
study of patients with post-stroke motor impairment or Parkinson’s disease
(PD). PD is characterized by severe movement impairments (e.g.,
bradykinesia or akinesia, limb rigidity and postural instability) that usually
causes gait disorders, such as small steps, lower cadence and reduced gait
speed, festination and freezing (Dalla Bella et al., 2015). Several studies
clearly revealed that rhythm, and more specifically isochronous stimulation
through metronome or music (i.e., with its underlying beats), can
significantly improve gait (Thaut et al., 1996; see De Dreu, Van Der Wilk,
Poppe, Kwakkel, & Van Wegen, 2012 and Dalla Bella et al., 2015 for meta-
analysis and review). In particular, rhythmic auditory stimulation (i.e., when
PD patients are asked to walk along with the auditory cue) has been shown
to improve gait velocity, cadence, and stride length (e.g., McIntosh, Brown,
Rice, & Thaut, 1997; Thaut et al., 1996) and reduce freezing episodes (e.g.,
Arias & Cudeiro, 2010). Such improvements have been shown to persist
also in the absence of the auditory stimulation following cuing-based
training programs, thus leading to long-lasting effects (see Dalla Bella et al.,
2015). Importantly, rhythmic auditory stimulation has been shown to be an
effective intervention for gait rehabilitation not only in PD patients, but also
in subjects with hemi-paretic stroke (Thaut, McIntosh, & Rice, 1997; Thaut
et al., 2007). One possible explanation for this beneficial effect relies on the
fact that external auditory cues, by promoting neural entrainment, generate
temporal expectations allowing to predict the occurrence of a next event
(Jones, 1976; Large, 2008). Such rhythm-driven expectations can help in
regularizing and stabilizing movement when the sensorimotor network
underlying temporal processing is impaired, as in PD or stroke patients, by
reinforcing compensatory neural networks able to enhance motor behavior
(Benoit et al., 2014; Nombela, Hughes, Owen, & Grahn, 2013).
C F D
S WAT H I S WA MI N AT H A N A N D E . G L E N N
SCHELLENBERG
O W F : M M
A M T
FIGURE 1. Individuals with high cognitive ability and music aptitude have an increased likelihood
of taking music lessons, which could then go on to improve cognitive and musical abilities.
In other correlational research, we used a similar approach to examine
the association between music training and music aptitude (Swaminathan &
Schellenberg, 2018). The simple association between the two variables was
significant, as in previous research, with music training accounting for 24.5
percent of the variance in music aptitude. When socio-economic status,
openness to experience, short-term memory, and general cognitive ability
were considered jointly with music training, the predictive power of the
model increased to 36.7 percent. Music training continued to have the
largest partial association, accounting independently for 6.2 percent of the
variance in music aptitude. Note that the reduction in variance explained
(from 24.5 percent to 6.2 percent) highlights the overlap between music
training and non-musical variables, which are typically overlooked in this
line of research. When the non-musical variables were considered jointly,
they accounted uniquely for 12.2 percent of the variance in music aptitude
(with music training held constant). In other words, music aptitude was
predicted better by the non-musical variables than it was by music training
alone. These findings highlight that music aptitude is more than the simple
consequence of music training. Although music training might be the best
predictor variable, other, non-musical variables play an important role.
In a fourth study (Swaminathan et al., 2018), we used the same approach
to examine the association between music training and reading ability
among adults who were native or non-native speakers of English. As in
previous research, reading ability was positively associated with duration of
music training. We also found that reading ability improved in tandem with
general cognitive ability, and it was better among native than non-native
speakers of English. When these variables were considered jointly, general
cognitive ability and native-language status had significant partial
associations with reading ability, but music training did not. In other words,
associations between music training and reading may be an artifact of other
variables that are typically ignored in this line of research.
R
Alain, C., Zendel, B. R., Hutka, S., & Bidelman, G. M. (2014). Turning down the noise: The benefit
of musical training on the aging auditory brain. Hearing Research 308, 162–173.
Amer, T., Kalender, B., Hasher, L., Trehub, S. E., & Wong, Y. (2013). Do older professional
musicians have cognitive advantages? PLoS ONE 8(8), e71630.
Besson, M., Schön, D., Moreno, S., Santos, A., & Magne, C. (2007). Influence of musical expertise
and musical training on pitch processing in music and language. Restorative Neurology and
Neuroscience 25(3–4), 399–410.
Bhatara, A., Yeung, H. H., & Nazzi, T. (2015). Foreign language learning in French speakers is
associated with rhythm perception, but not with melody perception. Journal of Experimental
Psychology: Human Perception and Performance 41(2), 277–282.
Bialystok, E., & DePape, A.-M. (2009). Musical expertise, bilingualism, and executive functioning.
Journal of Experimental Psychology: Human Perception and Performance 35(2), 565–574.
Bidelman, G. M., & Alain, C. (2015). Musical training orchestrates coordinated neuroplasticity in
auditory brainstem and cortex to counteract age-related declines in categorical vowel perception.
Journal of Neuroscience 35(3), 1240–1249.
Bidelman, G. M., Hutka, S., & Moreno, S. (2013). Tone language speakers and musicians share
enhanced perceptual and cognitive abilities for musical pitch: Evidence for bidirectionality
between the domains of language and music. PLoS ONE 8(4): e60676.
Bidelman, G. M., & Krishnan, A. (2010). Effects of reverberation on brainstem representation of
speech in musicians and nonmusicians. Brain Research 1355, 112–125.
Bilhartz, T. D., Bruhn, R. A., & Olson, J. E. (2000). The effect of early music training on child
cognitive development. Journal of Applied Developmental Psychology 20(4), 615–636.
Boebinger, D., Evans, S., Rosen, S., Lima, C.F., Manly, T., & Scott, S. K. (2015). Musicians and non-
musicians are equally adept at perceiving masked speech. Journal of the Acoustical Society of
America 137(1), 378–387.
Brandler, S., & Rammsayer, T. H. (2003). Differences in mental abilities between musicians and non-
musicians. Psychology of Music 31(2), 123–138.
Brochard, R., Dufour, A., & Deprés, O. (2004). Effect of musical expertise on visuospatial abilities:
Evidence from reaction times and mental imagery. Brain and Cognition 54(2), 103–109.
Brody, N. (1992). Intelligence (2nd ed.). San Diego, CA: Academic Press.
Bugos, J. A., Perlstein, W. M., McCrae, C. S., Brophy, T. S., & Bedenbaugh, P. H. (2007).
Individualized piano instruction enhances executive functioning and working memory in older
adults. Aging and Mental Health 11(4), 464–471.
Butzlaff, R. (2000). Can music be used to teach reading? Journal of Aesthetic Education 34(3–4),
167–178.
Catterall, J., Chapleau, R., & Iwanaga, J. (1999). Involvement in the arts and human development:
General involvement and intensive involvement in music and theatre arts. In E. Fiske (Ed.),
Champions of change: The impact of the arts on learning (pp. 1–18). Washington, DC: The Arts
Education Partnership and The President’s Committee on the Arts and the Humanities.
Chan, A. S., Ho, Y. C., & Cheung, M. C. (1998). Music training improves verbal memory. Nature
396(6707), 128.
Cheek, J. M., & Smith, L. R. (1999). Music training and mathematics achievement. Adolescence
34(136), 759–761.
Chobert, J., François, C., Velay, J. L., & Besson, M. (2014). Twelve months of active musical
training in 8- to 10-year-old children enhances the preattentive processing of syllabic duration and
voice onset time. Cerebral Cortex 24(4), 956–967.
Chobert, J., Marie, C., François, C., Schön, D., & Besson, M. (2011). Enhanced passive and active
processing of syllables in musician children. Journal of Cognitive Neuroscience 23(12), 3874–
3887.
Conway, A. R. A., Getz, S. J., Macnamara, B., & Engel de Abreu, P. M. J. (2011). Working memory
and fluid intelligence. In R. J. Sternberg & S. B. Kaufman (Eds.), The Cambridge handbook of
intelligence (pp. 394–418). Cambridge: Cambridge University Press.
Corenblum, B., & Marshall, E. (1998). The band played on: Predicting students’ intentions to
continue studying music. Journal of Research in Music Education 46(1), 128–140.
Corrigall, K. A., & Schellenberg, E. G. (2015). Predicting who takes music lessons: Parent and child
characteristics. Frontiers in Psychology 6, 282. doi: 10.3389/fpsyg.2015.00282
Corrigall, K. A., Schellenberg, E. G., & Misura, N. M. (2013). Music training, cognition, and
personality. Frontiers in Psychology 4, 222. doi: 10.3389/fpsyg.2013.00222
Corrigall, K. A., & Trainor, L. J. (2011). Associations between length of music training and reading
skills in children. Music Perception 29(2), 147–155.
Costa-Giomi, E. (1999). The effects of three years of piano instruction on children’s cognitive
development. Journal of Research in Music Education 47(3), 198–212.
Costa-Giomi, E. (2004). Effects of three years of piano instruction on children’s academic
achievement, school performance and self-esteem. Psychology of Music 32(2), 139–152.
Dankovičová, J., House, J., Crooks, A., & Jones, K. (2007). The relationship between musical skills,
music training, and intonation analysis skills. Language and Speech 50(2), 177–225.
Deary, I. J., Strand, S., Smith, P., & Fernandes, C. (2007). Intelligence and educational achievement.
Intelligence 35(1), 13–21.
Degé, F., Kubicek, C., & Schwarzer, G. (2011). Music lessons and intelligence: A relation mediated
by executive functions. Music Perception 29(2), 195–201.
Degé, F., & Schwarzer, G. (2011). The effect of a music program on phonological awareness in
preschoolers. Frontiers in Psychology 2, 124. doi: 10.3389/fpsyg.2011.00124.
Degé, F., Wehrum, S., Stark, R., & Schwarzer, G. (2014). Music lessons and academic self-concept in
12- to 14-year-old children. Musicae Scientiae 18(2), 203–215.
Delogu, F., Lampis, G., & Belardinelli, M. O. (2010). From melody to lexical tone: Musical ability
enhances specific aspects of foreign language perception. European Journal of Cognitive
Psychology 22(1), 46–61.
Elbert, T., Pantev, C., Wienbruch, C., Rockstroh, B., & Taub, E. (1995). Increased cortical
representation of the fingers of the left hand in string players. Science 270(5234), 305–307.
Faßhauer, C., Frese, A., & Evers, S. (2015). Musical ability is associated with enhanced auditory and
visual cognitive processing. BMC Neuroscience 16(1), 1. Retrieved from
https://doi.org/10.1186/s12868-015-0200-4
Fauvel, B., Groussard, M., Mutlu, J., Arenaza-Urquijo, E. M., Eustache, F., Desgranges, B., & Platel,
H. (2014). Musical practice and cognitive aging: Two cross-sectional studies point to phonemic
fluency as a potential candidate for a use-dependent adaptation. Frontiers in Aging Neuroscience
6, 227. doi:10.3389/fnagi.2014.00227
Fitzpatrick, K. R. (2006). The effect of instrumental music participation and socioeconomic status on
Ohio fourth-, sixth-, and ninth-grade proficiency test performance. Journal of Research in Music
Education 54(1), 73–84.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schön, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A Randomized
control trial. PloS ONE 10(9), e0138715.
Forgeard, M., Winner, E., Norton, A., & Schlaug, G. (2008). Practicing a musical instrument in
childhood is associated with enhanced verbal ability and nonverbal reasoning. PLoS ONE 3(10),
e3566.
Frakes, L. (1985). Differences in music achievement, academic achievement, and attitude among
participants, dropouts, and nonparticipants in secondary school music. Dissertation Abstracts
International 46, 370A. University Microfilms No. AAC8507938.
François, C., Chobert, J., Besson, M., & Schön, D. (2013). Music training for the development of
speech segmentation. Cerebral Cortex 23(9), 2038–2043.
Franklin, M. S., Moore, K. S., Yip, C., Jonides, J., Rattray, K., & Moher, J. (2008). The effects of
musical training on verbal memory. Psychology of Music 36(3), 353–365.
Gardiner, M., Fox, A., Knowles, F., & Jeffry, D. (1996). Learning improved by arts training. Nature
381(6580), 284.
Gaser, C., & Schlaug, G. (2003). Brain structures differ between musicians and non-musicians.
Journal of Neuroscience 23(27), 9240–9245.
Gibson, C., Folley, B. S., & Park, S. (2009). Enhanced divergent thinking and creativity in musicians:
A behavioral and near-infrared spectroscopy study. Brain and Cognition 69(1), 162–169.
Good, A., Gordon, K. A., Papsin, B. C., Nespoli, G., Hopyan, T., Peretz, I., & Russo, F. A. (2017).
Benefits of music training for perception of emotional speech prosody in deaf children with
cochlear implants. Ear & Hearing 38(4), 455–464.
Gordon, E. E. (1965). Music aptitude profile. Chicago: GIA.
Gouzouasis, P., Guhn, M., & Kishor, N. (2007). The predictive relationship between achievement and
participation in music and achievement in core Grade 12 academic subjects. Music Education
Research 9(1), 81–92.
Grassi, M., Meneghetti, C., Toffalini, E., & Borella, E. (2017). Auditory and cognitive performance
in elderly musicians and nonmusicians. PLoS ONE 12(11), e0187881.
Graziano, A. B., Peterson, M., & Shaw, G. L. (1999). Enhanced learning of proportional math
through music training and spatial-temporal training. Neurological Research 21(2), 139–152.
Gromko, J. E. (2005). The effect of music instruction on phonemic awareness in beginning readers.
Journal of Research in Music Education 53(3), 199–209.
Gromko, J. E., & Poorman, A. S. (1998). Developmental trends and relationships in children’s aural
perception and symbol use. Journal of Research in Music Education 46(1), 16–23.
Gruhn, W., Galley, N., & Kluth, C. (2003). Do mental speed and musical abilities interact? Annals of
the New York Academy of Sciences 999, 485–496.
Guo, X., Ohsawa, C., Suzuki, A., & Sekiyama, K. (2018). Improved digit span in children after a 6-
week intervention of playing a musical instrument: An exploratory randomized controlled trial.
Frontiers in Psychology 8, 2303. doi: 10.3389/fpsyg.2017.02303
Hanna-Pladdy, B., & Gajewski, B. (2012). Recent and past musical activity predicts cognitive aging
variability: Direct comparison with general lifestyle activities. Frontiers in Human Neuroscience
6, 198. doi: 10.3389/fnhum.2012.00198
Hanna-Pladdy, B., & MacKay, A. (2011). The relation between instrumental musical activity and
cognitive aging. Neuropsychology 25(3), 378–386.
Hannon, E. E., & Trainor, L. J. (2007). Music acquisition: Effects of enculturation and formal
training on development. Trends in Cognitive Sciences 11(11), 466–472.
Hansen, M., Wallentin, M., & Vuust, P. (2013). Working memory and musical competence of
musicians and non-musicians. Psychology of Music 41(6), 779–793.
Hassler, M., Birbaumer, N., & Feil, A. (1985). Musical talent and visuo-spatial abilities: A
longitudinal study. Psychology of Music 13(2), 99–113.
Haywood, S., Griggs, J., Lloyd, C., Morris, S., Kiss, Z., & Skipp, A. (2015). Creative futures: Act,
sing, play. Evaluation report and executive summary. London: Educational Endowment
Foundation.
Helmbold, N., Rammsayer, T., & Altenmüller, E. (2005). Differences in primary mental abilities
between musicians and nonmusicians. Journal of Individual Differences 26, 74–85.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Herrero, L., & Carriedo, N. (2018). Differences in updating processes between musicians and non-
musicians from late childhood to adolescence. Learning and Individual Differences 61, 188–195.
Hetland, L. (2000). Learning to make music enhances spatial reasoning. Journal of Aesthetic
Education 34(3–4), 179–238.
Ho, Y., Cheung, M., & Chan, A. S. (2003). Music training improves verbal but not visual memory:
Cross-sectional and longitudinal explorations in children. Neuropsychology 17(3), 439–450.
Jaeggi, S. M., Buschkuehl, M., Jonides, J., & Perrig, W. J. (2008). Improving fluid intelligence with
training on working memory. Proceedings of the National Academy of Sciences 105, 6829–6833.
Jakobson, L. S., Lewycky, S. T., Kilgour, A. R., & Stoesz, B. M. (2008). Memory for verbal and
visual material in highly trained musicians. Music Perception 26(1), 41–55.
Jaschke, A. C., Honing, H., & Scherder, E. J. A. (2018). Longitudinal analysis of music education on
executive functions in primary school children. Frontiers in Neuroscience 12, 103.
doi:10.3389/fnins.2018.00103
Jensen, A. R. (1969). How much can we boost IQ and scholastic achievement? Harvard Educational
Review 39(1), 1–123.
Jensen, A. R. (1998). The g factor: The science of mental ability. Westport, CT: Praeger.
Judge, T. A., Higgins, C. A., Thoresen, C. J., & Barrick, M. R. (1999). The big five personality traits,
general mental ability, and career success across the life span. Personnel Psychology, 52(3), 621–
652.
Kaviani, H., Mirbaha, H., Pournaseh, M., & Sagan, O. (2014). Can music lessons increase the
performance of preschool children in IQ tests? Cognitive Processing 15(1), 77–84.
Kinney, D. W. (2008). Selected demographic variables, school music participation and achievement
test scores of urban middle school students. Journal of Research in Music Education 56(2), 145–
161.
Kinney, D. W. (2010). Selected nonmusic predictors of urban students’ decisions to enroll and persist
in middle school band programs. Journal of Research in Music Education 57(4), 334–350.
Klinedinst, R. E. (1991). Predicting performance achievement and retention of fifth-grade
instrumental students. Journal of Research in Music Education 39(3), 225–238.
Klingberg, T. (2010). Training and plasticity of working memory. Trends in Cognitive Sciences 14(7),
317–324.
Koelsch, S., Schröger, E., & Tervaniemi, M. (1999). Superior pre-attentive auditory processing in
musicians. Neuroreport 10, 1309–1313.
Kolinsky, R., Cuvelier, H., Goetry, V., Peretz, I., & Morais, J. (2009). Music training facilitates
lexical stress processing. Music Perception 26(3), 235–246.
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11, 599–605.
Kraus, N., Slater, J., Thompson, E. C., Hornickel, J., Strait, D. L., Nicol, T., & White-Schwoch, T.
(2014). Music enrichment programs improve the neural encoding of speech in at-risk children.
Journal of Neuroscience 34(36), 11913–11918.
Law, L. N. C., & Zentner, M. (2012). Assessing musical abilities objectively: Construction and
validation of the Profile of Music Perception Skills. PLoS ONE 7(12), e52508.
Leng, X., Shaw, G. L., & Wright, E. L. (1990). Coding of musical structure and the trion model of
cortex. Music Perception 8(1), 49–62.
Love, J. M., Chazan-Cohen, R., Raikes, H., & Brooks-Gunn, J. (2013). What makes a difference:
Early Head Start evaluation findings in a developmental context. Monographs of the Society for
Research in Child Development 78(1), 1–173.
Mackintosh, N. J. (2011). IQ and human intelligence (2nd ed.). Oxford: Oxford University Press.
Magne, C., Schön, D., & Besson, M. (2006). Musician children detect pitch violations in both music
and language better than nonmusician children: Behavioral and electrophysiological approaches.
Journal of Cognitive Neuroscience 18(2), 199–211.
Mansens, D., Deeg, D. J. H., & Comijs, H. C. (2017). The association between singing and/or
playing a musical instrument and cognitive functions in older adults. Aging & Mental Health.
doi:10.1080/13607863.2017.1328481
Marques, C., Moreno, S., Castro, S. L., & Besson, M. (2007). Musicians detect pitch violation in a
foreign language better than non-musicians: Behavioral and electrophysiological evidence. Journal
of Cognitive Neuroscience 19(9), 1453–1463.
Mehr, S. A., Schachner, A., Katz, R. C., & Spelke, E. S. (2013). Two randomized trials provide no
consistent evidence for nonmusical cognitive benefits of brief preschool music enrichment. PLoS
ONE 8(12), e82007.
Melby-Lervåg, M., & Hulme, C. (2013). Is working memory training effective? A meta-analytic
review. Developmental Psychology 49(2), 270–291.
Melby-Lervåg, M., Reddick, T. S., & Hulme, C. (2016). Working memory training does not improve
performance on measures of intelligence or other measures of “far transfer.” Perspectives on
Psychological Science 11(4), 512–534.
Moody, D. E. (2009). Can intelligence be increased by training on a task of working memory?
Intelligence 37, 327–328.
Moreno, S., Bialystok, E., Barac, R., Schellenberg, E. G., Cepeda, N. J., & Chau, T. (2011). Short-
term music training enhances verbal intelligence and executive function. Psychological Science
22(11), 1425–1433.
Moreno, S., Friesen, D., & Bialystok, E. (2011). Effect of music training on promoting preliteracy
skills: Preliminary causal evidence. Music Perception 29(2), 165–172.
Moreno, S., Marques, C., Santos, A., Santos, M., Castro, S. L., & Besson, M. (2009). Musical
training influences linguistic abilities in 8-year-old children: More evidence for brain plasticity.
Cerebral Cortex 19(3), 712–723.
Mosing, M. A., Madison, G., Pedersen, N. L., & Ullén, F. (2016). Investigating cognitive transfer
within the framework of music practice: Genetic pleiotropy rather than causality. Developmental
Science 19(3), 504–512. A
Münte, T. F., Altenmüller, E., & Jäncke, L. (2002). The musician’s brain as a model of
neuroplasticity. Nature Reviews Neuroscience 3(6), 473–478.
Okada, B. M., & Slevc, R. L. (2018). Individual differences in musical training and executive
functions: A latent variable approach. Memory & Cognition. doi:10.3758/s13421-018-0822-8
Overy, K. (2003). Dyslexia and music. Annals of the New York Academy of Sciences 999, 497–505.
Parbery-Clark, A., Skoe, E., Lam, C., & Kraus, N. (2009). Musician enhancement for speech in
noise. Ear and Hearing 30(6), 653–661.
Parbery-Clark, A., Strait, D. L., Anderson, S., Hittner, E., & Kraus, N. (2011). Musical experience
and the aging auditory system: Implications for cognitive abilities and hearing speech in noise.
PLoS ONE 6(5), e18082.
Parbery-Clark, A., Tierney, A., Strait, D. L., & Kraus, N. (2012). Musicians have fine-tuned neural
distinction of speech syllables. Neuroscience 219, 111–119.
Patel, A. D. (2003). Language, music, syntax and the brain. Nature Neuroscience 6, 674–681.
Patel, A. D. (2011). Why would musical training benefit the neural encoding of speech? The OPERA
hypothesis. Frontiers in Psychology 2, 142. doi: 10.3389/fpsyg.2011.00142
Patston, L. L. M., & Tippett, L. J. (2011). The effect of background music on cognitive performance
in musicians and nonmusicians. Music Perception 29(2), 173–183.
Petitto, L. (2008). Arts education, the brain, and language. In B. Rich & C. Asbury (Eds.), Learning,
arts, and the brain: The Dana Consortium report on arts and cognition (pp. 93–104). New
York/Washington, DC: The Dana Foundation.
Piro, J. M., & Oritz, C. (2009). The effect of piano lessons on the vocabulary and verbal sequencing
skills of primary grade students. Psychology of Music 37(3), 325–347.
Portowitz, A., Lichtenstein, O., Egorova, L., & Brand, E. (2009). Underlying mechanisms linking
music education and cognitive modifiability. Research Studies in Music Education 31(2), 107–128.
Posedel, J., Emery, L., Souza, B., & Fountain, C. (2012). Pitch perception, working memory, and
second-language phonological production. Psychology of Music 40(4), 508–517.
Posner, M., Rothbart, M. K., Sheese, B. E., & Kieras, J. (2008). How arts training influences
cognition. In B. Rich & C. Asbury (Eds.), Learning, arts, and the brain: The Dana Consortium
report on arts and cognition (pp. 1–10). New York/Washington, DC: The Dana Foundation.
Rapport, M. D., Orban, S. A., Kofler, M. J., & Friedman, L. M. (2013). Do programs designed to
train working memory, other executive functions, and attention benefit children with ADHD? A
meta-analytic review of cognitive, academic, and behavioral outcomes. Clinical Psychology
Review 33(8), 1237–1252.
Rauscher, F. H., & Hinton, S. C. (2011). Music instruction and its diverse extra-musical benefits.
Music Perception 29(2), 215–226.
Rauscher, F. H., & Shaw, G. L. (1998). Key components of the Mozart effect. Perceptual and Motor
Skills 86(3), 835–841.
Rauscher, F. H., Shaw, G. L., Levine, L. J., Wright, E. L., Dennis, W. R., & Newcomb, R. L. (1997).
Music training causes long-term enhancement of preschool children’s spatial-temporal reasoning.
Neurological Research 19(1), 2–8.
Rauscher, F. H., & Zupan, M. A. (2000). Classroom keyboard instruction improves kindergarten
children’s spatial-temporal performance: A field experiment. Early Childhood Research Quarterly
15(2), 215–228.
Roden, I., Grube, D., Bongard, S., & Kreutz, G. (2014). Does music training enhance working
memory performance? Findings from a quasi-experimental longitudinal study. Psychology of
Music 42(2), 284–298.
Rueda, M. R., Rothbart, M. K., McCandliss, B. D., Saccamanno, L., & Posner, M. I. (2005).
Training, maturation and genetic influences on the development of executive attention.
Proceedings of the National Academy of Sciences 102, 14931–14936.
Ruggles, D. R., Freyman, R. L., & Oxenham, A. J. (2014). Influence of musical training on
understanding voiced and whispered speech in noise. PLoS ONE 9(1), e86980.
Sala, G., & Gobet, F. (2016). Do the benefits of chess instruction transfer to academic and cognitive
skills? A meta-analysis. Educational Research Review 18, 46–57.
Sala, G., & Gobet, F. (2017a). When the music’s over: Does music skill transfer to children’s and
young adolescents’ cognitive and academic skills? A meta-analysis. Educational Research Review
20, 55–67.
Sala, G., & Gobet, F. (2017b). Working memory training in typically developing children: A meta-
analysis of the available evidence. Developmental Psychology 53, 671–685.
Sala, G., Tatlidil, K. S., & Gobet, F. (2018). Video game training does not enhance cognitive ability:
A comprehensive meta-analytic investigation. Psychological Bulletin 144, 111–139.
Salthouse, T. A. (2004). What and when of cognitive aging. Current Directions in Psychological
Science 13(4), 140–144.
Salthouse, T. A. (2005). Relations between cognitive abilities and measures of executive functioning.
Neuropsychology 19(4), 532–545.
Schellenberg, E. G. (2004). Music lessons enhance IQ. Psychological Science 15(8), 511–514.
Schellenberg, E. G. (2006). Long-term positive associations between music lessons and IQ. Journal
of Educational Psychology 98(2), 457–468.
Schellenberg, E. G. (2011a). Examining the association between music lessons and intelligence.
British Journal of Psychology 102(3), 283–302.
Schellenberg, E. G. (2011b). Music lessons, emotional intelligence, and IQ. Music Perception 29(2),
185–194.
Schellenberg, E. G., & Mankarious, M. (2012). Music training and emotion comprehension in
childhood. Emotion 12(5), 887–891.
Schellenberg, E. G., & Moreno, S. (2010). Music lessons, pitch processing and g. Psychology of
Music 38(2), 209–221.
Schellenberg, E. G., & Peretz, I. (2008). Music, language and cognition: Unresolved issues. Trends in
Cognitive Sciences 12(2), 45–46.
Schellenberg, E. G., & Weiss, M. W. (2013). Music and cognitive abilities. In D. Deutsch (Ed.), The
psychology of music (3rd ed.). San Diego, CA: Elsevier.
Seashore, C. E., Lewis, D., & Saetveit, J. G. (1960). The Seashore measures of musical talents. New
York: Psychological Corporation.
Shipstead, Z., Redick, T. S., & Engle, R. W. (2012). Is working memory training effective?
Psychological Bulletin 138, 628–654.
Sirin, S. R. (2005). Socioeconomic status and academic achievement: A meta-analytic review of
research. Review of Educational Research 75(3), 417–453.
Skoe, E., & Kraus, N. (2012). Human subcortical auditory function provides a new conceptual
framework for considering modularity. In P. Rebuschat, M. Rohrmeier, J. A. Hawkins, & I. Cross
(Eds.), Language and music as cognitive systems (pp. 269–282). Oxford: Oxford University Press.
Slater, J., Skoe, E., Strait, D. L., O’Connell, S., Thompson, E., & Kraus, N. (2015). Music training
improves speech-in-noise perception: Longitudinal evidence from a community-based music
program. Behavioural Brain Research 291, 244–252.
Slevc, L. R., Davey, N. S., Buschkuehl, M., & Jaeggi, S. M. (2016). Tuning the mind: Exploring the
connections between musical ability and executive functions. Cognition 152, 199–211.
Sluming, V., Brooks, J., Howard, M., Downes, J. J., & Roberts, N. (2007). Broca’s area supports
enhanced visuospatial cognition in orchestral musicians. Journal of Neuroscience 27(14), 3799–
3806.
Soveri, A., Antfolk, J., Karlsson, L., Salo, B., & Laine, M. (2017). Working memory training
revisited: A multi-level meta-analysis of n-back training studies. Psychonomic Bulletin & Review
24(4), 1077–1096.
Spelke, E. (2008). Effects of music instruction on developing cognitive systems at the foundation of
mathematics and science. In B. Rich & C. Asbury (Eds.), Learning, arts, and the brain: The Dana
Consortium report on arts and cognition (pp. 17–49). New York/Washington, DC: The Dana
Foundation.
Spinath, B., Spinath, F. M., Harlaar, N., & Plomin, R. (2006). Predicting school achievement from
general cognitive ability, self-perceived ability, and intrinsic value. Intelligence 34(4), 363–374.
Standley, J. M. (2008). Does music instruction help children learn to read? Evidence of a meta-
analysis. Update: Applications of Research in Music Education 27(1), 17–32
Stoesz, B., Jakobson, L., Kilgour, A., & Lewycky, S. (2007). Local processing advantage in
musicians: Evidence from disembedding and constructional tasks. Music Perception 25(2), 153–
165.
Strait, D. L., & Kraus, N. (2011). Can you hear me now? Musical training shapes functional brain
networks for selective auditory attention and hearing speech in noise. Frontiers in Psychology 2,
113. doi: 10.3389/fpsyg.2011.00113
Strait, D. L., O’Connell, S., Parbery-Clark, A., & Kraus, N. (2014). Musicians’ enhanced neural
differentiation of speech sounds arises early in life: Developmental evidence from ages 3 to 30.
Cerebral Cortex 24(9), 2512–2521.
Strait, D. L., Parbery-Clark, A., Hittner, E., & Kraus, N. (2012). Musical training during early
childhood enhances the neural encoding of speech in noise. Brain and Language 123(3), 191–201.
Strait, D. L., Parbery-Clark, A., O’Connell, S., & Kraus, N. (2013). Biological impact of preschool
music classes on processing speech in noise. Developmental Cognitive Neuroscience 6, 51–60.
Swaminathan, J., Mason, C. R., Streeter, T. M., Best, V., Kidd Jr, G., & Patel, A. D. (2015). Musical
training, individual differences and the cocktail party problem. Scientific Reports 5, 11628.
doi:10.1038/srep11628
Swaminathan, S., & Gopinath, J. K. (2013). Music training and second-language English
comprehension and vocabulary skills in Indian children. Psychological Studies 58(2), 164–170.
Swaminathan, S., & Schellenberg, E. G. (2016). Music training. In T. Strobach & J. Karbach (Eds.),
Cognitive training: An overview of features and applications (pp. 137–144). New York: Springer.
Swaminathan, S., & Schellenberg, E. G. (2017). Musical competence and phonemic perception in a
foreign language. Psychonomic Bulletin and Review 24(6), 1929–1934.
Swaminathan, S., & Schellenberg, E. G. (2018). Musical competence is predicted by music training,
cognitive abilities, and personality. Manuscript submitted for publication.
Swaminathan, S., Schellenberg, E. G., & Khalil, S. (2017). Revisiting the association between music
lessons and intelligence: Training effects or music aptitude? Intelligence 62, 119–124.
Swaminathan, S., Schellenberg, E. G., & Venkatesan, K. (2018). Explaining the association between
music training and reading in adults. Journal of Experimental Psychology: Learning, Memory, and
Cognition. doi: 10.1037/xlm0000493
Talamini, F., Grassi, M., Toffalini, E., Santoni, R., & Carretti, B. (2018). Learning a second language:
Can music aptitude or music training have a role? Learning and Individual Differences 64, 1–7.
Thompson, W. F., Schellenberg, E. G., & Husain, G. (2004). Decoding speech prosody: Do music
lessons help? Emotion 4(1), 46–64.
Thomson, J. M., Leong, V., & Goswami, U. (2013). Auditory processing interventions and
developmental dyslexia: A comparison of phonemic and rhythmic approaches. Reading and
Writing 26(2), 139–161.
Thorndike, E. L., & Woodworth, R. S. (1901a). The influence of improvement in one mental function
upon the efficiency of other functions (I). Psychological Review 8, 247–261.
Thorndike, E. L., & Woodworth, R. S. (1901b). The influence of improvement in one mental
function upon the efficiency of other functions (II). The estimation of magnitudes. Psychological
Review 8, 384–395.
Tierney, A., Krizman, J., Skoe, E., Johnston, K., & Kraus, N. (2013). High school music classes
enhance the neural processing of speech. Frontiers in Psychology 4, 855. doi:
10.3389/fpsyg.2013.00855.
Trimmer, C. G., & Cuddy, L. L. (2008). Emotional intelligence, not music training, predicts
recognition of emotional speech prosody. Emotion 8(6), 838–849.
Vaughn, K. (2000). Music and mathematics: Modest support for the oft-claimed relationship. Journal
of Aesthetic Education 34(3–4), 149–166.
Vaughn, K., & Winner, E. (2000). SAT scores of students who study the arts: What we can and
cannot conclude about the association. Journal of Aesthetic Education 34(3–4), 77–89.
Wallentin, M., Nielsen, A. H., Friis-Olivarius, M., Vuust, C., & Vuust, P. (2010). The Musical Ear
Test: A new reliable rest for measuring musical competence. Learning and Individual Differences
20(3), 188–196.
Wan, C. Y., & Schlaug, G. (2010). Music making as a tool for promoting brain plasticity across the
life span. The Neuroscientist 16(5), 566–577.
Wandell, B., Dougherty, R. F., Ben-Shachar, M., Deutsch, G. K., & Tsang, J. (2008). Training in the
arts, reading, and brain imaging. In B. Rich & C. Asbury (Eds.), Learning, arts, and the brain: The
Dana Consortium report on arts and cognition (pp. 51–59). New York/Washington, DC: The Dana
Foundation.
Weicker, J., Villringer, A., & Thöne-Otto, A. (2016). Can impaired working memory functioning be
improved by training? A meta-analysis with a special focus on brain injured patients.
Neuropsychology 30(2), 190–212.
Weiss, M. W., & Bidelman, G. M. (2015). Listening to the brainstem: Musicianship enhances
intelligibility of subcortical representations for speech. Journal of Neuroscience 35(4), 1687–1691.
Wetter, O. E., Koerner, F., & Schwaninger, A. (2009). Does musical training improve school
performance? Instructional Science 37(4), 365–374.
White-Schwoch, T., Carr, K. W., Anderson, S., Strait, D. L., & Kraus, N. (2013). Older adults benefit
from music training early in life: Biological evidence for long-term training-driven plasticity.
Journal of Neuroscience 33(45), 17667–17674.
Winner, E., & Cooper, M. (2000). Mute those claims: No evidence (yet) for a causal link between
arts study and academic achievement. Journal of Aesthetic Education 34(3–4), 11–75.
Winner, E., Goldstein, T. R., & Vincent-Lancrin, S. (2012). The impact of arts education: What do
we know? Paris: Organisation for Economic Co-operation and Development.
Wong, P. C. M., Skoe, E., Russo, N. M., Dees, T., & Kraus, N. (2007). Musical experience shapes
human brainstem encoding of linguistic pitch patterns. Nature Neuroscience 10, 420–422.
Wu, H., Ma, X., Zhang, L., Liu, Y., Zhang, Y., & Shu, H. (2015). Musical experience modulates
categorical perception of lexical tones in native Chinese speakers. Frontiers in Psychology 6, 436.
doi:10.3389/fpsyg.2015.00436
Zendel, B. R., & Alain, C. (2009). Concurrent sound segregation is enhanced in musicians. Journal
of Cognitive Neuroscience 21(8), 1488–1498.
Zendel, B. R., & Alain, C. (2012). Musicians experience less age-related decline in central auditory
processing. Psychology and Aging 27(20), 410–417.
Zendel, B. R., & Alain, C. (2013). The influence of lifelong musicianship on neurophysiological
measures of concurrent sound segregation. Journal of Cognitive Neuroscience 25(4), 503–516.
Zuk, J., Benjamin, C., Kenyon, A., & Gaab, N. (2014). Behavioral and neural correlates of executive
functioning in musicians and non-musicians. PLoS ONE 9(6), e99868.
Zuk, J., Ozernov-Palchik, O., Kim, H., Lakshminarayanan, K., Gabrieli, J. D., Tallal, P., & Gaab, N.
(2013). Enhanced syllable discrimination thresholds in musicians. PLoS ONE 8(12), e80546.
CHAPT E R 27
THE NEUROSCIENCE OF
CHILDREN ON THE
AUTISM SPECTRUM WITH
EXCEPTIONAL MUSICAL
ABILITIES
ADAM OCKELFORD
G ’ E T
L A
In his “ecological” approach to understanding auditory perception, Gaver
describes how, in everyday contexts, listeners typically privilege the
function of sounds over their acoustic properties (Gaver, 1993). He gives
the example of a pedestrian walking along an alley when a sound starts to
emerge from behind: that of a car with a large and powerful engine. It is
possible, Gaver contends, that the person concerned may attend to the
sound’s timbre, noticing whether it is rough or smooth, or bright or dull.
Paying attention to these attributes, which have to do with quality of the
sound itself, Gaver terms “musical” listening. It is more likely however, in
the situation described, that the pedestrian will notice that the car is
approaching rather quickly from behind and that the sound of its engine is
echoing off the narrow walls of the alley. There is a need to move quickly to
get out of the vehicle’s way! This Gaver terms “everyday” listening: the
experience of focusing on the significance of an event rather than its
acoustic properties. However, there are people for whom this unthinking
prioritization of function over form does not appear to occur—among them,
many children with so-called “classic” autism—the type first identified by
Leo Kanner in 1943.
Autism is a lifelong, neurological condition that typically manifests
itself within the first two or three years of childhood (see, for example,
Boucher, 2009; Frith, 2003; Wing, 2003). Its effects pervade the whole of a
child’s development. According to the Diagnostic and Statistical Manual of
Mental Disorders, Fifth Edition (the so-called “DSM-5”), published by the
American Psychiatric Association in 2013, Autism Spectrum Disorder is
identified through two criteria (APA, 2013, 299.00 (F84.0)):
Z T
The theory sets out from the reductionist position that music can be
regarded as a system of perceived sonic variables. Some of these, such as
duration, have a single axis of variability, while others, like timbre, are
multidimensional in nature; some, including loudness, gauge qualities,
while others detail its perceived location in time or space; and some, like
pitch, pertain to individual notes, while others, such as tonality, are
characteristic of a group. Despite this diversity, all these variables have a
common characteristic: they each have many potential modes of existence,
or “values,” whose range represents the freedom of choice available to
composers. Conversely, each may be deemed to be constrained or “ordered”
to the extent that its value is reckoned to be subject to restriction. While
some of the causes of perceived sonic constraint may lie beyond a
composer’s immediate control (the selection of timbre will be dictated by
the availability of performers, for example, and a singer may be unable to
reach a particular pitch), and while external influences (such as the cross-
media effects of song-texts, for instance) often have a bearing, most—and
certainly the most important—perceived sonic restrictions in fact function
intra-musically, through the process of repetition. In short, a value may be
thought to be ordered if it is reckoned to exist in imitation of another.
Since the vast majority of listeners are quite unaware of this type of
cognitive activity, clearly it need not operate at a conscious level. Yet it
must be universally present, if only subliminally, otherwise an orderly
sequence of sounds would prove no more effective a means of musical
communication than a random one, which is not the case. While the
acknowledgment of the central role of repetition in musical structure is
widespread in the music-theoretical and music-psychological literatures, the
function of imitation is less well understood. In this respect, zygonic theory
most closely resonates with the thinking of Cone (1987, p. 237), who
asserts, in relation to the derivation of musical material, that “y is derived
from x (y ← x), or, to use the active voice, x generates y (x → y), if y
resembles x and y follows x. By ‘resembles’, I mean ‘sounds like’.”
In zygonic theory, the connections between x and y that Cone identifies,
through which a sense of derivation (or generation) is imagined to exist, are
termed “zygonic relationships” (after the Greek “zygon,” meaning a yoke
connecting two similar things). This single theoretical concept bequeaths a
vast perceptual legacy, with many manifestations: potentially involving any
perceived aspect of sound; existing over different periods of perceived time;
and operating within the same and between different pieces, performances,
and hearings. Zygonic relationships may function in a number of different
ways: reactively, for example, in assessing the relationship between two
extant values, or proactively, in ideating a value as an orderly continuation
from one presented. They may operate between anticipated or remembered
values, or even those that are wholly imagined, only ever existing in the
mind.
In music-structural terms, zygonic relationships function at three
hierarchical levels: between individual events (notes or chords), groups
(motifs, “hooks,” “licks,” or riffs) and frameworks (imaginary matrices of
pitch and time, whose elements have different perceived probabilities of
occurrence according to a listener’s previous exposure to pieces in a
particular style or genre). Recognizing imitation between individual musical
events is thought to take the least mental processing power of all forms of
structure (Ockelford, 2017, p. 187), since it requires at most two or three
items of musical information, in the form of notes or chords or the intervals
between them, to be held in working memory and compared. The temporal
envelope within which such structures occur is constrained, sometimes
extending to little more than Edmund Husserl’s “perceived present.”
Recognizing relationships between motifs is cognitively more demanding:
organization of this kind necessarily involves four events or more, since at
least two are required to create a group (and a minimum of two groups is
required). The timespans of such structures are potentially greater than
those involving events alone, and may even implicate long-term memory.
There is likely to be a greater degree of abstraction from the perceptual
“surface” too. Imitative links between frameworks of pitch and onset times
appear to be the most cognitively demanding of all. They depend on the
existence of long-term “schematic” memories, whereby the details of the
perceptual surface of music and individual connections perceived between
musical events are not encoded in long-term memory discretely or
independently, but are combined with many thousands of other similar data
to create probabilistic networks of relationships between notional
representations of pitch and perceived time. That is, large amounts of
perceptual information are merged to enable the deep level of cognitive
abstraction to occur.
To sum up: the cognitive correlates of musical structure grow in
complexity as one moves from events, to groups and then frameworks,
reflecting an increasing amount of perceptual input, experienced over
longer periods of time, and processed and stored using progressively more
abstract forms of mental representation. Moreover, the cognitive operations
pertaining to higher levels of structure must build on and incorporate those
required to process lower levels, since connections between groups
comprise series of relationships between events, and links between
frameworks are established by acknowledging the correspondences that
exist between groups.
T N D M
A P
To what extent does this hierarchy of music structures tie in with the
development of children’s understanding of music? It seems that very
young children have a built-in propensity to imitate others, and that this
plays a part in early interactive sound-making using individual musical
sounds (Meltzoff & Prinz, 2002). Similarly, in analyzing preverbal
communication in babies from just two to seven months old, Papoušek
(1996) found that up to half of these infants’ vocal sounds are part of
reciprocal matching sequences that the children engage in with their
mothers. These findings complement work by other researchers showing
that babies less than five months of age can replicate individual pitches
(Kessen, Levine, & Wendrich, 1979), copy changes in pitch (Kuhl &
Meltzoff, 1982), and emulate vowel-like sounds made by others. Each of
these forms of interaction involves imitation at the level of events—
showing engagement with musical sounds at the first level of the structural
hierarchy.
Engagement at the next level first appears from seven to eleven months,
when babies repeat and vary groups of sounds, using them as the basic units
of structure, through babbling that, according to Papoušek (1996, p. 106),
involves producing “short musical patterns or phrases that soon become the
core units for a new level of vocal practicing and play.” Gradually, groups
of sounds may be linked through repetition or transposition to form chains,
and the first self-sufficient improvised pieces emerge. Welch (2006) notes
that between the ages of one and two, a typically spontaneous song
comprises repetitions of a brief melodic phrase at different pitch centers.
These are unlike adult singing, however, because “they lack a framework of
stable pitches (a scale) and use a very limited set of contours in one song”
(Dowling, 1982, pp. 416, 417). From the age of two-and-a-half, so-called
“potpourri” melodies may appear (Moog, 1976, p. 115), which borrow and
may transform features and fragments from other, standard songs that have
been assimilated into the child’s own spontaneous singing (Hargreaves,
1986, p. 73). These self-generated melodies, which use materials derived
from a repertoire that is familiar from a child’s musical culture, are termed
“referent-guided improvisation” by Mang (2005).
Finally, from around the age of four (the age can vary considerably), two
advances occur, which pertain to the third level of the music-structural
hierarchy: frameworks. First, children develop the capacity to abstract an
underlying pulse from the surface rhythm of songs and other pieces
(meaning that he or she can perform “in time” to a regular beat that is
provided). Second, children’s singing acquires “tonal stability,” with the
clear projection of a key center across all the phrases of a piece
(Hargreaves, 1986, pp. 76, 77). These abilities imply a cognizance of
repetition at a deeper structural level in the “background” organization of
music.
How does this compare with language development? Children start to
understand a few key words towards the end of their first year, and will
begin to speak using some of these words from around the age of 12
months. Around this time, they develop the capacity to process short
phrases, and from 18–24 months they learn to juxtapose words in pairs
themselves. Over the next two years, these become amalgamated into
longer and more complex sentences, which are generated through an
intuitive understanding of the syntax of the language (or languages) to
which a child is exposed (Saxton, 2010).
So much for the development of music and language processing in the
early years. We can surmise that the third category in the ecological model
of auditory perception—“everyday” sounds—must perceptually be the most
primitive of all, since it requires less cognitive processing than either music
or speech. Hence we can reasonably assume that, early on in human
development, the brain typically treats all sound in the same way and that
music processing starts to emerge as a distinct strand, first, followed by
language. We can speculate that “everyday” sounds form the residue that is
left.
Based on these assumptions, a developmental model of ecological
auditory perception can be constructed along the following lines as in Fig.
1. Note the underlying assumption that, in addition to their shared neural
resources, music and language come to have other, distinct neural correlates
during the first year of life. Since the precise nature of the sounds that
constitute speech and music varies from one culture to another, it is
appropriate to regard the model as indicative rather than prescriptive.
FIGURE 1. A visual representation of how the emerging streams of music and language processing
arise in auditory development.
T D A
P S C
A S
FIGURE 3. Speech may also be processed musically by some children on the autism spectrum.
It is worth noting that echolalia is not only found in the context of
“special” development; it is a feature of “typical” language acquisition in
young children too (Mcglone-Dorrian & Potter, 1984) when, it seems, the
urge to imitate what is heard outstrips semantic understanding. This accords
with a stage in the ecological model of auditory development when the two
strands of communication through sound—language and music—are not
yet cognitively distinct, and supports the notion that musical development
precedes the onset of language.
For some children on the autism spectrum, music itself can become
“super-structured” with additional repetition, as the account of Ben (above)
shows; it is common for children on the autism spectrum to play snippets of
pieces or videos with music over and over again. It is as though the high
proportion of repetition that characterizes music (which is at least 80
percent—see Ockelford, 2005), is insufficient for the mind that craves
structure, and so it makes even more. In conversing with autistic adults who
are able to verbalize why, as children, they would repeat musical excerpts in
this way, it seems that the main reason for obsessively repeating a
particularly fascinating series of sounds (apart from the sheer enjoyment
that the regularity brings) is that they could hear more and more in the
sequence concerned as they listened to it again and again. Bearing in mind
that most music tends to be highly complex, with many events occurring
simultaneously (and given that even individual notes tend to comprise many
pitches in the form of harmonics), to the child with finely tuned auditory
perception, there are many different things to attend to in even a few
seconds of music, and an even greater number of potential relationships
between sounds to fathom. So, for example, while listening to a passage for
orchestra one hundred times may be extremely tedious to the “neurotypical”
ear, which can detect only half a dozen composite events, each fused in
perception, to the mind of the autistic child, which can break down the
sequence into a dozen different melodic lines, the stimulus may be
captivating.
A P
One of the consequences of an early preoccupation with the “musical”
qualities of sounds appears to be the development of “absolute pitch”—or
“AP.” This is the ability to identify or produce pitches in isolation from
others. In the Western population, the capacity is very rare, with an
estimated prevalence of 1 in 10,000 (Takeuchi & Hulse, 1993). However,
among those on the autism spectrum, the position is markedly different;
recent estimates, derived from parental questionnaires, vary between 8
percent, n = 118 (Vamvakari, 2013) and 21 percent, n = 305 (Reese, 2014).
These figures are broadly supported by DePape, Hall, Tillmann, and Trainor
(2012) who found that 11 percent of 27 high-functioning adolescents with
autism had AP. It is unusual to find such high orders of difference in the
incidence of a perceptual ability and, evidently, there is something distinct
in the way that the parts of the brain responsible for pitch memory wire
themselves up in a significant minority of autistic children.
Although AP is a useful skill in “neurotypical” musicians—including
elite performers—it is an indispensable factor in the development of
performance skills in autistic children with learning difficulties—so-called
“musical savants” (Miller, 1989). It seems to be this unusual ability that
both motivates and enables some young children with a very limited general
understanding of the world around, from the age of 24 months or so, to pick
out tunes or chords on instruments that they encounter (sometimes more or
less by chance) at home or elsewhere. Often the instrument concerned will
be an electric keyboard or piano. The children’s early experiments in
producing music may well occur with no adult intervention—or, indeed,
awareness of what they are doing. It is my contention that AP has this
impact since each pitch sounds distinct, and potentially can elicit a powerful
emotional response; hence, being able to reproduce these at will must surely
be an intoxicating experience. But more than this, having AP makes
learning to play by ear manageable, in a way that “relative pitch”—the
capacity to process the differences between pitches (“intervals”)—does not.
To understand why this should be so, consider a typical playground chant,
based on the intervals of a minor third, a perfect fourth, and a major second
(Fig. 4).
FIGURE 4. An archetypal playground chant.
T I AP M
E C Y
P W A A
S
Derek
After many hours of the same dull drone – auditory chewing gum that has long since lost its
flavor or interest – there is a sudden, almost imperceptible change in the humming of the plane’s
engines. I glance outside and see that, at last, we are over the Nevada desert. Only an hour or so
now until we hit Los Angeles.
The young man sitting next to me – noticeably upright in his seat – stiffens slightly as he hears
the tiny deviation in sound.
‘F sharp’, he intones. ‘It’s F sharp, Adam.’
He leans towards me, demanding a response, and the sun bounces off his trademark Prada
sunglasses, but without penetrating the world of darkness beneath.
‘Yes, Derek’, I reply, ‘We’ll soon be landing at LAX.’
‘Landing at LAX’, he echoes, apparently relishing the sound of the words – and their import – in
equal measure.
‘And I will see Dana, and I will play the piano’, he continues.
‘Yes, Derek.’ I offer the same reply again, the sound of my voice as much as the words offering a
reassurance forged in a relationship of many years – as Derek’s teacher, mentor and friend.
‘You’ll play the piano.’
Repetition confers calm, a hint of a smile crosses Derek’s features, and he relaxes back in his
seat.
Derek Paravicini – blind autistic savant, musician extraordinaire, learning disabled genius,
unflagging companion – is on his way to California to perform in a series of concerts: grist to his
globe-trotting mill.
For him, airplanes are one of life’s many mysteries: a series of awkward slopes and steps to be
negotiated; well-meaning helping hands; a waft of warm, stale air; ‘doors to automatic and cross-
check’; the sound of the engines starting up. Soon the seat seems to move and bump about, then
steadiness; a long, vibrating steadiness. Les Mis on the headphones – once, twice, three times?
At last, everything goes into reverse, and abruptly, we’re off the plane. Now there are new
voices, new accents. A new hotel. Oatmeal instead of porridge for breakfast. And … finally …
the piano. At last, something familiar. Every note a close friend. The band plays the same as in
England. The clapping is familiar too, though people seem to clap louder in America.
‘Good job, Derek!’ ‘Awesome!’ ‘Can you smile for the photo?’ Derek wrinkles his nose, and
everyone laughs, infectiously. He catches the humor, and smiles as well. Music has worked its
magic, as it always does.1
This account illustrates how, even as a young adult, Derek’s propensity
for processing everyday sounds and language in a musical way, acquired as
a child, is still evident. For him, the sound of the jet engine has to be
accepted as a feature of air travel, but is not understood; hence it remains
for him at the level of pure auditory input, which he hears as musical notes.
And there are traces too of his childhood tendency to repeat words—
echolalia—as much for their sounds as their meanings, as he copies the
ends of my contributions to our conversation. These are testament to his
idiosyncratic neural circuitry, produced both as a result of hypoxia due to
his extreme prematurity and the consequent exceptional cognitive
environment that his developing brain had to endure, without visual input
and with limited capacity to process language.
And yet, with the necessary support, Derek can function well: meeting
new people, interacting with them in unrehearsed social situations, and
tolerating unfamiliar environments with equanimity. By celebrating the
advantages and ameliorating the disadvantages of Derek’s autism spectrum
condition, he has a quality of life that would be the envy of many,
acknowledged internationally as a “special” musician.
The current cohort of autistic children with whom I work visit me, with
their parents, in a large practice room at the University of Roehampton.
There are two pianos, to avoid potential difficulties over personal space. A
number of the children rarely say a word. Some, like Romy, don’t speak at
all. She converses through her playing, telling me what piece she would like
next, and indicating when she’s had enough. Sometimes, she will tease me
by apparently suggesting one thing when she means another. In this way,
jokes are shared and, sometimes, feelings of sadness too. For Romy, music
replaces words, and truly functions as a proxy language, with the
exceptional neurological correlates that must entail.
Romy
On Sunday mornings, at 10.00am, I steel myself for Romy’s arrival. I know that the next two
hours will be an exacting test of my musical mettle. Yet Romy has severe learning difficulties,
and she doesn’t speak at all. She is musical to the core, though; she lives and breathes music – it
is the very essence of her being. With her passion comes a high degree of particularity; Romy
knows precisely which piece she wants me to play, at what tempo, and in which key. And woe
betide me if I get it wrong.
When we started working together, six years ago, mistakes and misunderstandings occurred all
too frequently since, as it turned out, there were very few pieces that Romy would tolerate: for
example, the theme from Für Elise (never the middle section); the Habanera from Carmen; and
some snippets from ‘Buckaroo Holiday’ (the first movement of Aaron Copland’s Rodeo).
Romy’s acute neophobia meant that even one note of a different piece would evoke shrieks of
fear-cum-anger, and the session could easily grow into an emotional conflagration.
So gradually, gradually, over weeks, then months, and then years, I introduced new pieces –
sometimes, quite literally, at the rate of one note per session. On occasion, if things were
difficult, I would even take a step back before trying to move on again the next time. And,
imperceptibly at first, Romy’s fears started to melt away. The theme from Brahms’s Haydn
Variations became something of an obsession, followed by the slow movement of Beethoven’s
Pathetique sonata. Then it was Joplin’s The Entertainer, and Rocking All Over the World by
Status Quo.
Over the six years, Romy’s jigsaw box of musical pieces – fragments ranging from just a few
seconds to a minute or so in length – has filled up at an ever-increasing rate. Now it’s
overflowing, and it’s difficult to keep up with Romy’s mercurial musical mind; mixing and
matching ideas in our improvised sessions, and even changing melodies and harmonies so they
mesh together, or to ensure that my contributions don’t!
As we play, new pictures in sound emerge and then retreat as a kaleidoscope of ideas whirls
between us. Sometimes a single melody persists for fifteen minutes, even half an hour. For
Romy, no matter how often it is repeated, a fragment of music seems to stay fresh and vibrant.
At other times, it sounds as though she is trying to play several pieces at the same time – she just
can’t get them out quickly enough, and a veritable nest of earworms wriggle their way onto the
piano keyboard. Vainly I attempt to herd them into a common direction of musical travel.
So here I am, sitting at the piano in Roehampton, on a Sunday morning in mid-November,
waiting for Romy to join me (not to be there when she arrives is asking for trouble). I’m
limbering up with a rather sedate rendition of the opening of Chopin’s Etude in C major, Op. 10,
No. 1 when I hear her coming down the corridor, vocalizing with increasing fervor. I feel the
tension rising, and as her father pushes open the door, she breaks away from him, rushes over to
the piano and, with a shriek and an extraordinarily agile sweep of her arm, elbows my right hand
out of the way at the precise moment that I was going to hit the D an octave above middle C. She
usurps this note to her own ends, ushering in her favorite Brahms-Haydn theme. Instantly, Romy
smiles, relaxes and gives me the choice of moving out of the way or having my lap appropriated
as an unwilling cushion on the piano stool. I choose the former, sliding to my left onto a chair
that I’d placed earlier in readiness for the move that I knew I would have to make.
I join in the Brahms, and encourage her to use her left hand to add a bass line. She tolerates this
up to the end of the first section of the theme, but in her mind she’s already moved on, and
without a break in the sound, Romy steps onto the set of A Little Night Music, gently noodling
around the introduction to Send in the Clowns. But it’s in the wrong key – G instead of E flat –
which I know from experience means that she doesn’t really want us to go into the Sondheim
classic, but instead wants me to play the first four bars (and only the first four bars) of
Schumann’s Kleine Studie Op. 68, No. 14. Trying to perform the fifth bar would, in any case, be
futile since Romy’s already started to play … now, is it I am Sailing or O Freedom? The opening
ascent from D through E to G could signal either of those possibilities. Almost tentatively, Romy
presses those three notes down and then looks at me and smiles, waiting, and knowing that
whichever option I choose will be the wrong one. I just shake my head at her and plump for O
Freedom, but sure enough Rod Stewart shoves the Spiritual out of the way before it has time to
draw a second breath.
From there, Romy shifts up a gear to the Canon in D – or is it really Pachelbel’s masterpiece?
With a deft flick of her little finger up to a high A, she seems to suggest that she wants Streets of
London instead (which uses the same harmonies). I opt for Ralph McTell, but another flick, this
time aimed partly at me as well as the keys, shows that Romy actually wants Beethoven’s
Pathetique theme – but again, in the wrong key (D). Obediently I start to play, but Romy takes
us almost immediately to A flat (the tonality that Beethoven originally intended). As soon as I’m
there, though, Romy races back up the keyboard again, returning to Pachelbel’s domain. Before
I’ve had time to catch up, though, she’s transformed the music once more; now we’re hearing the
famous theme from Dvorak’s New World Symphony.
I pause to recover my thoughts, but Romy is impatiently waiting for me to begin the
accompaniment. Two or three minutes into the session, and we’ve already touched on twelve
pieces spanning 300 years of Western music and an emotional range to match. Yet, here is a girl
who in everyday life is supposed to have no ‘theory of mind’ – the capacity to put yourself in
other people’s shoes and think what they are thinking. Here is someone who is supposed to lack
the ability to communicate. Here is someone who functions, apparently, at an 18-month level.
But I say here is a joyous musician who amazes all who hear her. Here is a girl in whom extreme
ability and disability coexist in the most extraordinary way. Here is someone who can reach out
through music and touch one’s emotions in a profound way. If music is important to us all, for
Romy it is truly her lifeblood.2
Freddie
‘Why’s he doing that?’ Freddie’s father, Simon, sounded more than usually puzzled by the antics
of his son.
After months of displacement activity, Freddie was finally sitting next to me at the piano, and
looked as though this time he really were about to play. A final fidget and then his right hand
moved towards the keys. With infinite care, he placed his thumb on middle C as he had watched
me do before – but without pressing it down. Silently, he moved to the next note (D), which he
feathered in a similar way, using his index finger, then with the same precision he touched E, F,
and G, before coming back down the soundless scale to an inaudible C.
I couldn’t help smiling.
‘Fred, we need to hear the notes!’
My comment was rewarded with a deep stare, right into my eyes. Through them, almost. It was
always hard to know what Freddie was thinking, but on this occasion he did seem to understand
and was willing to respond to my request, since his thumb went back to C. Again, the key
remained un-pressed, but this time he sang the note (perfectly in tune), and then the next one,
and the next, until the five-finger exercise was complete.
In most children (assuming that they had the necessary musical skills), such behavior would
probably be regarded as an idiosyncratic attempt at humor or even mild naughtiness. But Freddie
was being absolutely serious and was pleased, I think, to achieve what he’d been asked to do, for
he had indeed enabled me to hear the notes!
He stared at me again, evidently expecting something more, and without thinking I leant
forward.
‘Now on this one, Fred’, I said, touching C sharp.
Freddie gave the tiniest blink and a twitch of his head, and I imagined him, in a fraction of a
second, making the necessary kinesthetic calculations. Without hesitation or error, he produced
the five-finger exercise again, this time using a mixture of black and white notes. Each pressed
silently. All sung flawlessly.
And then, spontaneously, he was off up the keyboard, beginning the same pentatonic pattern on
each of the twelve available keys. At my prompting, Freddie re-ran the sequence with his left
hand – his unbroken voice hoarsely whispering the low notes.
So logical. Why bother to play the notes if you know what they sound like already?
So apparently simple a task, and yet … such a difficult feat to accomplish: the whole
contradiction of autism crystallized in a few moments of music making.3
This chapter sets out a theory of how some children on the autism spectrum
develop prodigious musical talent as a consequence of the way that they
perceive everyday sounds and speech—in musical terms. In a significant
minority of cases, this leads to the development of AP, which, given access
to an appropriate instrument (typically a keyboard), enables such children to
learn to play by ear. This skill is often acquired entirely through their own
efforts and typically first manifests itself in the early years. The neural
correlates of this exceptional development are yet to be explored through
brain imaging, which in the case of children severely affected by autism,
who tend to function successfully only in familiar environments, presents
significant challenges (although one or two passive studies in the field of
music and language have been undertaken; see, for example, Lai,
Pantazatos, Schneider, & Hirsch, 2012; Sharda, Midha, Malik, Mukerji, &
Singh, 2015). It is surely an area worth exploring, however, not only for the
light it would shed on our knowledge of exceptionality, but for the fresh
perspectives that human diversity offers the understanding of our species as
a whole. This is possible because we exist on continua of interests, abilities,
and traits, and it is my contention that, by analyzing the behaviors and their
neural correlates of those who function at the extremes of our tribe’s natural
neurodiversity, we can better understand the ordinary, everyday, musical
engagement that is characteristic of us all. Most importantly, it’s my belief
that, through the prism of the overtly remarkable, we can discover the
uncelebrated exceptionality in each of us, for whether autistic or
neurotypical, we are all musical by design (Ockelford, 2017, p. 9).
R
American Psychiatric Association (2013). Diagnostic and statistical manual of mental disorders (5th
ed.). Washington, DC: APA.
Boucher, J. (2009). The autistic spectrum: Characteristics, causes, and practical issues. London:
Sage Publications.
Brandt, A., Gebrian, M., & Slevc, L. R. (2012). Music and early language acquisition. Frontiers in
Psychology 3. Retrieved from https://doi.org/10.3389/fpsyg.2012.00327
Cone, E. (1987). On derivation: Syntax and rhetoric. Music Analysis 6(3), 237–256.
DePape, A.-M. R., Hall, G. B. C., Tillmann, B., & Trainor, L. J. (2012). Auditory processing in high-
functioning adolescents with autism spectrum disorder. PloS ONE 7(9), e44084.
Dowling, J. (1982). Melodic information processing and its development. In D. Deutsch (Ed.), The
psychology of music (pp. 413–429). New York: Academic Press.
Fay, W. H. (1967). Childhood echolalia. Folia Phoniatrica et Logopaedica 19(4), 297–306.
doi:10.1159/000263153
Fay, W. H. (1973). On the echolalia of the blind and of the autistic child. Journal of Speech and
Hearing Disorders 38(4), 478–489.
Frith, U. (2003). Autism: Explaining the enigma. Oxford: Wiley-Blackwell.
Gaver, W. W. (1993). What in the world do we hear? An ecological approach to auditory event
perception. Ecological Psychology 5(1), 1–29.
Hargreaves, D. (1986). The developmental psychology of music. Cambridge: Cambridge University
Press.
Kessen, W., Levine, J., & Wendrich, K. (1979). The imitation of pitch in infants. Infant Behavior and
Development 2, 93–99.
Kuhl, P., & Meltzoff, A. (1982). The bimodal perception of speech in infancy. Science 218(4577),
1138–1141.
Lai, G., Pantazatos, S., Schneider, H., & Hirsch, J. (2012). Neural systems for speech and song in
autism. Brain 135(3), 961–975.
Lamont, A. (2008). Young children’s musical worlds: Musical engagement in 3.5-year-olds. Journal
of Early Childhood Research 6(3), 247–261.
Lecanuet, J.-P. (1996). Prenatal auditory experience. In I. Deliège & J. Sloboda (Eds.), Musical
beginnings (pp. 3–34). Oxford: Oxford University Press.
McEvoy, R. E., Loveland, K. A., & Landry, S. H. (1988). The functions of immediate echolalia in
autistic children: A developmental perspective. Journal of Autism and Developmental Disorders
18(4), 657–668.
Mcglone-Dorrian, D., & Potter, R. E. (1984). The occurrence of echolalia in three year olds’
responses to various question types. Communication Disorders Quarterly 7(2), 38–47.
McPherson, G. (Ed.). (2016). Musical prodigies: Interpretations from psychology, education,
musicology, and ethnomusicology. New York: Oxford University Press.
Malloch, S., & Trevarthen, C. (Eds.). (2009). Communicative musicality: Exploring the basis of
human companionship. New York: Oxford University Press.
Mang, E. (2005). The referent of early children’s songs. Music Education Research 7(1), 3–20.
Meltzoff, A., & Prinz, W. (2002). The imitative mind: Development, evolution and brain bases.
Cambridge: Cambridge University Press.
Miller, L. (1989). Musical savants: Exceptional skill and mental retardation. Hillsdale, NJ: Lawrence
Erlbaum.
Mills, A. (1993). Visual handicap. In D. Bishop & K. Mogford (Eds.), Language development in
exceptional circumstances (pp. 150–164). Hove: Psychology Press.
Moog, H. (1976). The musical experiences of the pre-school child. Trans. C. Clarke. London: Schott.
Norman-Haignere, S., Kanwisher, N. G., & McDermott, J. H. (2015). Distinct cortical pathways for
music and speech revealed by hypothesis-free voxel decomposition. Neuron 88(6), 1281–1296.
Ockelford, A. (2005). Repetition in music: Theoretical and metatheoretical perspectives. Farnham:
Ashgate.
Ockelford, A. (2009). In the key of genius: The extraordinary life of Derek Paravicini. London:
Random House.
Ockelford, A. (2013). Applied musicology: Using zygonic theory to inform music education, therapy,
and psychology research. New York: Oxford University Press.
Ockelford, A. (2017). Comparing notes: How we make sense of music. London: Profile Books.
Papoušek, M. (1996). Intuitive parenting: A hidden source of musical stimulation in infancy. In I.
Deliège & J. Sloboda (Eds.), Musical beginnings (pp. 88–112). Oxford: Oxford University Press.
Patel, A. D. (2012). Language, music, and the brain: A resource-sharing framework. In P. Rebuschat,
M. Rohmeier, J. A. Hawkins, & I. Cross (Eds.), Language and music as cognitive systems (pp.
204–223). Oxford: Oxford University Press.
Prizant, B. M. (1979). An analysis of the functions of immediate echolalia in autistic children.
Dissertation Abstracts International 39(9-B), 4592–4593.
Prizant, B. M., & Duchan, J. F. (1981). The functions of immediate echolalia in autistic children.
Journal of Speech and Hearing Disorders 46(3), 241–249.
Reese, A. (2014). The effect of exposure to structured musical activities on communication skills and
speech for children and young adults on the autism spectrum (Doctoral dissertation). University of
Roehampton, London.
Saxton, M. (2010). Child language: Acquisition and development. London: Sage Publications.
Sharda, M., Midha, R., Malik, S., Mukerji, S., & Singh, N. C. (2015). Fronto-temporal connectivity
is preserved during sung but not spoken word listening, across the autism spectrum. Autism
Research 8(2), 174–186.
Sloboda, J. A., Davidson, J. W., Howe, M. J., & Moore, D. G. (1996). The role of practice in the
development of performing musicians. British Journal of Psychology 87(2), 287–309.
Stalinski, S. M., & Schellenberg, E. G. (2010). Shifting perceptions: Developmental changes in
judgments of melodic similarity. Developmental Psychology 46(6), 1799–1803.
Sterponi, L., & Shankey, J. (2013). Rethinking echolalia: Repetition as interactional resource in the
communication of a child with autism. Journal of Child Language 41(2), 275–304.
Takeuchi, A. H., & Hulse, S. H. (1993). Absolute pitch. Psychological Bulletin 113(2), 345.
Treffert, D. (2009). The savant syndrome: An extraordinary condition. A synopsis: Past, present,
future. Philosophical Transactions of the Royal Society B: Biological Sciences 364(1522), 1351–
1357.
Vamvakari, T. (2013). My child and music: A survey exploration of the musical abilities and interests
of children and young people diagnosed with autism spectrum conditions (Master’s dissertation).
University of Roehampton, London.
Welch, G. (2006). The musical development and education of young children. In B. Spodek & O.
Saracho (Eds.), Handbook of research on the education of young children (pp. 251–267). Mahwah,
NJ: Lawrence Erlbaum.
Wing, L. (2003). The autistic spectrum: A guide for parents and professionals. London: Robinson.
1
http://www.jkp.com/jkpblog/2013/04/music-language-autism/
2
https://blog.oup.com/2012/12/music-proxy-language-autisic-children/
3
http://www.huffingtonpost.com/adam-ockelford/autism-genius_b_4118805.html.
SECTION VII
MU S IC , T HE B R A IN ,
A N D HE A LT H
CHAPT E R 28
NEUROLOGIC MUSIC
THERAPY IN
S E N S O R I M O TO R
R E H A B I L I TAT I O N
C O R E N E T H A U T A N D K L A U S MA RT I N S T E P H A N
N M T
A M D
D M D
Parkinsonian Syndromes
Parkinson’s disease, the most common movement disorder, is an idiopathic
neurodegenerative disorder associated with progressive loss of
dopaminergic neurons in the basal ganglia, due to the deterioration of the
substantia nigra. Typical symptoms include progressive loss of muscle
control, which leads to bradykinesia (slowing of movements), resting limb
tremor (trembling of the limbs and head while at rest), postural instability,
rigidity (stiffness), and gait instability resulting in impaired balance. Key
treatment goals when working with PD include increasing heel strike in
order to promote longer stride lengths and decreased festinating gait
patterns, increasing step cadence and walking speed, increasing initiation of
movement and functional balance, and decreasing the risk of falls.
Many studies have looked at Parkinson’s disease and the effects of
Rhythmic Auditory Cuing (RAS) as an external time keeper to facilitate
movement sequences that are not receiving the appropriate internal timing
cues from the basal ganglia. Findings have shown that persons with PD, on
and off medication, were able to improve their walking patterns through
better posture; more appropriate step cadence and increased stride length;
and more efficient and symmetric muscle activation patterns (McIntosh et
al., 1997; Miller et al., 1996; Richards, Malouin, Bedard, & Cioni, 1992;
Thaut, McIntosh, et al., 1996). Additionally, McIntosh et al. (1998) looked
at long-term carry-over after a five-week RAS treatment program, finding
that it took an average of 5 weeks for velocity scores to return to baseline.
Falls are among the biggest contributors to loss of independent living,
long-term institutionalization, and increased mortality (Johnell, Melton,
Atkinson, O’Fallon, & Kurland, 1992). The risk of falls in a person with PD
increases substantially from that of healthy elderly, presenting not only a
serious concern over safety, but also over the enormous human and
healthcare cost associated with falling. Wood, Bilclough, Bowron, and
Walker (2002) found that out of 109 subjects with idiopathic PD and a
mean Hoen/Yahr rating of 2, 68 percent experienced falls over a one-year
period. Thaut and colleagues (Thaut, Rice, Braun Janzen, Hurt-Thaut, &
McIntosh, 2018) examined and compared the effects of a continuous 24
week RAS treatment program to an intermittent RAS program with 8 weeks
RAS training, 8 weeks without, for 24 weeks. Changes in ankle
dorsiflexion, cadence, velocity, stride length, the Berg Balance Scale, fear
of falling, the TUG test, and frequency and severity of falls were evaluated.
The findings offered evidence that continuous and intermittent RAS
treatment over time can be effective tools to reduce falls in persons with
Parkinson’s disease; however, continuous RAS treatment resulted in
significantly greater gains in dorsiflexion, cadence, velocity, stride length,
and a reduction in severity level 1 falls and fear of falling, when comparing
treatments. These results suggest that there is only a limited carry-over
effect for RAS in PD patients—presumably due to their pathophysiological
deficit. This encourages the use of ongoing home training programs, also
outside the therapy setting for these patients.
About one third of people with Parkinson’s disease experience freezing
episodes when initiating gait, changing directions, navigating around
obstacles or in small spaces. Numerous studies have shown the
effectiveness of rhythmic auditory cueing on the reduction of freezing
episodes (e.g., Frazzitta, Maestri, Uccellini, Bertotti, & Abelli, 2009;
Freedland et al., 2002; Howe, Lövgreen, Cody, Ashton, & Oldham, 2003;
McIntosh et al., 1997; Morris, Suteerawattananon, Etnyre, Jankovic, &
Protas, 2004; Thaut, McIntosh, et al., 1996; Willems et al., 2006).
Additionally, when looking at kinematic changes due to the immediate
entrainment effects of RAS gait training, Picelli et al. (2010) found
increased hip range of motion and power during pull-off phase of gait and
decreased ankle dorsiplantar flexion with rhythmic cueing. Other studies
which have looked at RAS training programs found a slight increase in
dorsiflexion over 5 weeks (Pau et al., 2016), and a significant increase over
8 week, and in 6 month training programs (Hurt-Thaut, 2014; Thaut et al.,
2018). Until now the sequelae of Parkinson’s have been mainly treated by
RAS approaches. However, it might be that discrete dysfunctions, such as
freezing episodes may also benefit from specifically tailored PSE
approaches.
Huntington’s Disease
Huntington’s disease is a hereditary neurodegenerative disorder which
results in motor disturbances such as hyperkinesia or dystonia, slow
execution of movements, and poor coordination. Perceptual timing is even
more impaired than in Parkinson patients (Cope, Grube, Singh, Burn, &
Griffiths, 2014). The hyperkinetic choreatic movements often coexist with
bradykinesia, and gait can present with a wide base of support, increased
lateral sway, variability in swing and stance phases, difficulty with
frequency modulation, and poor initiation of movement.
Thaut and colleagues (Thaut, Lange, Miltner, Hurt, & Hoemberg, 1996;
Thaut, Miltner, Lange, Hurt, & Hoemberg, 1999) explored velocity
modulation and rhythmic synchronization of gait in persons with
Huntington’s disease, providing the first evidence that rhythmic facilitation
could influence mobility in this population. A high variability in frequency
entrainment was seen across subjects, with exact phase and period matching
highly impaired. Comparisons of self-paced walking, rhythmic metronome
cueing, and music, found that subjects were able to significantly modulate
their gait velocity during both self-paced walking and with metronome
cueing, but not during the music condition. Due to their prominent
difficulties of perception of timing, a “simple” sensory signal may be most
useful for them. Rhythmic facilitation improved locomotor function after a
short training period, although disease progression had a clear impact on
gait parameters.
Parkinsonism
Parkinsonism is a general term used to describe impairments in motor
function presenting with similar characteristics to Parkinson’s disease such
as akinesia, hypokinesia, bradykinesia, motor blocks, rigidity, and problems
with the initiation of cyclical movements. These symptoms are also found
in movement disorders of different etiology such as vascular Parkinsonism,
and drug-induced Parkinsonism and in related disorders, which affect also
additional systems, such as progressive supranuclear palsy (PSP) and
multiple system atrophy (MSA). Generally they do not respond as well to
L-Dopa medications as patients with Parkinson’s. Furthermore, Cope et al.
(2014) could show for patients with MSA, that their timing perception was
more impaired than that in Parkinson’s, similar to that in Huntington’s.
Little is known about the effects of rhythmic auditory cueing on
Parkinsonism; however, given the strong effects seen in some of the related
disorders, further research in this area is warranted.
Multiple Sclerosis
Multiple sclerosis is a prevalent autoimmune disease of the central nervous
system which results in progressive demyelination resulting in scar tissue
which causes widespread neurological sensory, motor, and cognitive
symptoms such as paresthesia, progressive hemiparesis, ataxia, fatigue, and
depression. Gait and postural dysfunctions are common in patients with
multiple sclerosis and can affect static and dynamic stability, motor control,
and coordination, leading to an increased risk of falls and decreased quality
of life. Typical gait characteristics include reduced gait velocity, stride
length, cadence, and increased step width, asymmetric gait, and increased
double limb support time.
Only recently have a number of studies bridged the gap in literature by
looking at the effects of auditory rhythmic cueing on gait in people with
multiple sclerosis (Conklyn et al., 2010; Seebacher, Kuisma, Glynn, &
Berger, 2015, 2016; Shahraki, Sohrabi, Torbati, Nikkhah, & NaeimiKia,
2017). A systematic review of the effects of rhythmic auditory cueing in
gait rehabilitation for multiple sclerosis (Ghai, Ghai, & Effenberg, 2017,
2018; Ghai, Ghai, Schmitz, & Effenberg, 2018), suggested evidence for a
positive impact of rhythmic auditory cueing on reduction in the timed 25-
meter walk test, and spatiotemporal gait parameters: gait velocity, stride
length, and cadence.
The premise for using rhythmic cueing to learn, train, and retrain
movement is built on a feedforward model of rhythm driving the motor and
kinematic changes of movement. Baram and Miller (2007), however,
studied how self-generated auditory feedback through an external apparatus
can serve as a non-imposing reference which can provide a constant
awareness of gait quality and an instantaneous sensory response to changes
in gait for people with multiple sclerosis. Results of this study may provide
evidence for the use of an auditory feedback system to enhance patient
awareness and effort to improve gait quality.
In the last years basic and clinical research has provided more effective
drug treatment for patients with MS. This leads to a change of the
therapeutic goal in more and more patients: the goal is to halt the
progression of MS and not “only” to slow the progression of the disease.
This change of treatment strategy may open the possibility to use TIMP in
these patients more intensively.
Healthy Elderly
Several factors related to the normal process of aging can affect strength,
agility, flexibility, and muscle tone, therefore leading to sensorimotor
changes and safety risk in this population.
Decreased bone density or osteoporosis can not only decrease stability,
but also make bones more vulnerable to breaks. Decreased or lack of
physical activity can result in poor muscle tone, decreased strength, and loss
of bone mass and flexibility, putting someone at higher risk for falls and
injury. Age-related visual impairments such as cataracts and glaucoma can
alter depth perception, visual acuity, and peripheral vision, making it more
difficult to safely maneuver through one’s environment. Medications can
reduce mental alertness, impair balance and gait, and cause drops in systolic
blood pressure while standing. Additionally, environmental hazards such as
poor lighting, loose rugs, lack of grab bars, objects on the floor, or unsturdy
furniture can cause risks for falling.
Listening to music and sometimes dancing to music is often quite
popular with the elderly. We will, however, concentrate in this chapter on
the use of music to prevent falls. In a 2012 Cochrane review, Gillespie and
colleagues assessed the effects of interventions designed to reduce the
incidence of falls in older people living in the community, by examining
159 random control trials with 79,193 participants (Gillespie et al., 2012).
The conclusions of this review were that multifactorial assessment and
intervention programs—such as monitoring medication, treatment of visual
problems, fall prevention education, and non-slip shoes—reduce the rate of
falls, but not the risk of falling. The only interventions which consistently
reduced both the rate and risk of falling were group and home-based
exercise programs and home safety assessments. Hurt-Thaut (2014), found
that healthy elderly achieved a statistically significant increase in degrees of
dorsiflexion, velocity, cadence, stride length, and the Berg Balance Scale
scores when participating in both a continuous and intermittent 6-month
rhythmic based exercise and walking program.
D D
Autism Spectrum Disorder
Autism spectrum disorder (ASD) is a neurodevelopmental disorder that is
often characterized by deficits in social interaction, communication, and
unusual behaviors, such as clumsy uncoordinated or repetitive movement,
poor balance and postural control (Fournier et al., 2010). Only in recent
years has research attention turned to the delays in fundamental motor
development (e.g., oral motor control, coordination, gait) in ASD compared
to typically developing children, and how those delays can directly
influence interpersonal social exchange, cognitive functions such as
attention and executive function, and the acquisition and development of
written and spoken language. Torres et al. (2013) published the most
accessed paper on ASD from a movement perspective, laying out a broad
theoretical framework to research, treat, and track autism. Torres’s research
has been at the forefront of research exploring a complex system for
analyzing micro-movements as a reflection of the layers of multi-directional
internal and external influences on the central and peripheral nervous
systems during goal oriented movement in response to cognitive motor and
social task (Torres & Donnellan, 2015).
Although there is a limited body of research looking specifically at the
influence of elements of music on sensorimotor function in this population,
based on the principles of auditory motor entrainment, rhythm-based
interventions such as RAS, PSE, and TIMP could aid in regulatory control
of proprioceptive movement and provide adaptive mechanisms to decrease
movement variability, smooth movement trajectories, and improve gait
parameters such as symmetry and stability.
Cerebral Palsy
Cerebral palsy (CP) is a chronic disability of the central nervous system
characterized by abnormal control of movement and posture. A person with
CP can present with quadriplegia (both arms and legs affected), diplegia
(two limbs affected), or hemiplegia (one side of the body affected). Motor
symptoms can vary widely, ranging from minor difficulty with fine motor
movements such as grasping and manipulation of objects, to significant
muscular and motor control of all four limbs.
A few studies have looked at the effects of rhythm and instrument
playing on this population. When investigating the effects of RAS on adults
with CP (Kim et al., 2011), kinematic analysis revealed significant
increases in the anterior tilt of the pelvis and hip flexion at initial contact
with RAS training; however, there were no statistical differences in knee,
ankle, and foot kinematic parameters. Furthermore, Kim et al. (2012),
looked at the effects of RAS versus neurodevelopmental treatment (Bobath)
on gait patterned in adults with cerebral palsy over an intensive 3-week
training period. Findings indicated that RAS significantly increased
cadence, velocity, stride length, and step length, in addition to showing
significant increases in overall normalization of the gait on the gait
deviation index scores compared to the neurodevelopmental treatment
group. In contrast, the neurodevelopmental treatment group showed
significant decreases in cadence, velocity, stride length, and step length,
with a significant increase in step time; however, neurodevelopmental
treatment showed significant improvements in internal and external rotation
s of hip joints. In support of TIMP, Chong et al. (2013) explored finger
exercises on the keyboard as a tool to increase manual dexterity and
velocity in adults with cerebral palsy, finding improvements after twelve
30-minute TIMP sessions.
S C
Since the 1990s, a strong body of research evidence has set the foundation
for the use of rhythm and music as important tools in the development,
rehabilitation, and maintenance of sensorimotor function, particularly in the
treatment of neurologic disorders. Through external rhythmic cueing,
rhythmic entrainment optimizes the execution of a motor pattern by priming
the motor system and creating anticipatory rhythmic templates to allow for
optimal anticipation, motor planning, and execution of movement (Thaut,
McIntosh, & Hoemberg, 2014). While the temporal structures in music
remain the central elements when using music in the treatment of
sensorimotor dysfunction, other elements such as pitch, dynamics, and
harmony can also enhance and shape complex movements such as arm and
hand movements that are not intrinsically rhythmic. Neurologic Music
Therapy is a research-based system of clinical techniques guided by music
and neuroscience principles of music perception, cognition, and
performance. In the area of sensorimotor rehabilitation, three standardized
techniques, RAS, PSE, and TIMP have become well accepted in the
treatment of impairment and restoration of function based on current
knowledge of how music can aid in cortical reorganization, motor learning,
and neuromuscular re-education.
R
Altenmüller, E., Marco-Pallares, J., Münte, T. F., & Schneider, S. (2009). Neural reorganization
underlies improvement in stroke-induced motor dysfunction by music-supported therapy. Annals
of the New York Academy of Sciences 1169, 395–405.
Arnal, L. H. (2012). Predicting “when” using the motor system’s beta-band oscillations. Frontiers in
Human Neuroscience 6. Retrieved from https://doi.org/10.3389/fnhum.2012.00225
Arnal, L. H., Doelling, K. B., & Poeppel, D. (2015). Delta–beta coupled oscillations underlie
temporal prediction accuracy. Cerebral Cortex 25(9), 3077–3085.
Baram, Y., & Lenger, R. (2012). Gait improvement in patients with cerebral palsy by visual and
auditory feedback. Neuromodulation 15(1), 48–52.
Baram, Y., & Miller, A. (2007). Auditory feedback control for improvement of gait in patients with
multiple sclerosis. Journal of Neurological Sciences 254(1–2), 90–94.
Bengtsson, S. L., Ullén, F., Henrik Ehrsson, H., Hashimoto, T., Kito, T., Naito, E., … Sadato, N.
(2009). Listening to rhythms activates motor and premotor cortices. Cortex 45(1), 62–71.
Bernatzky, G., Bernatzky, P., Hesse, H. P., Staffen, W., & Ladurner, G. (2004). Stimulating music
increases motor coordination in patients afflicted by Morbus Parkinson. Neuroscience Letters 361,
4–8.
Bradt, J., Magee, W. L., Dileo, C., Wheeler, B. L., & McGilloway, E. (2010). Music therapy for
acquired brain injury. Cochrane Database of Systematic Reviews 7, CD006787.
doi:10.1002/14651858.CD006787.pub2
Brown, S. H., Thaut, M. H., Benjamin, J., & Cooke, J. D. (1993). Effects of rhythmic auditory cueing
on temporal sequencing of complex arm movements. Proceedings of the Society for Neuroscience
227(2).
Buetefish, C., Hummelsheim, H., Denzler, P., & Mauritz, K. H. (1995). Repetitive training of isolated
movements improves the outcome of motor rehabilitation of the centrally paretic hand. Journal of
Neurological Sciences 130(1), 59–68.
Bukowska, A. A. (2016). Influence of neurologic music therapy to improve the activity level in a
group of patients with PD. Nordic Journal of Music Therapy 25(1), 14.
Bukowska, A. A., Krężałek, P., Mirek, E., Bujas, P., & Marchewka, A. (2015). Neurologic music
therapy training for mobility and stability rehabilitation with Parkinson’s disease: A pilot study.
Frontiers in Human Neuroscience 9. Retrieved from https://doi.org/10.3389/fnhum.2015.00710
Cha, Y., Kim, Y., Hwang, S., & Chung, Y. (2014). Intensive gait training with rhythmic auditory
stimulation in individuals with chronic hemiparetic stroke: A pilot randomized controlled study.
Neurorehabilitation 35(4), 681–688.
Chadwick, D. M., & Clark, C. A. (1980). Adapting music instruments for the physically
handicapped. Music Educators Journal 67(3), 56–59.
Chong, H. J., Cho, S. R., Jeong, E., & Kim, S. J. (2013). Finger exercise with keyboard playing in
adults with cerebral palsy: A preliminary study. Journal of Exercise Rehabilitation 9(4), 420–425.
Chong, H. J., Cho, S. R., & Kim, S. J. (2014). Hand rehabilitation using MIDI keyboard playing in
adolescents with brain damage: A preliminary study. Neurorehabilitation 34(1), 147–155.
Conklyn, D., Stough, D., Novak, E., Paczak, S., Chemali, K., & Bethoux, F. (2010). A home-based
walking program using rhythmic auditory stimulation improves gait performance in patients with
multiple sclerosis: A pilot study. Neurorehabilitation and Neural Repair 24(9), 835–842.
Cope, T. E., Grube, M., Singh, B., Burn, D. J., & Griffiths, T. D. (2014). The basal ganglia in
perceptual timing: Timing performance in multiple system atrophy and Huntington’s disease.
Neuropsychologia 52(100), 73–81.
Crasta, J. E., Thaut, M. H., Anderson, C. W., Davies, P. L., & Gavin, W. J. (2018). Auditory priming
improves neural synchronization in auditory-motor entrainment. Neuropsychologia 117, 102–112.
de Dreu, M. J., van der Wilk, A. S., Poppe, E., Kwakkel, G., & van Wegen, E. E. (2012).
Rehabilitation, exercise therapy and music in patients with Parkinson’s disease: A meta-analysis of
the effects of music-based movement therapy on walking ability, balance and quality of life.
Parkinsonism & Related Disorders 18(Suppl. 1), 114–119.
de l’Etoile, S. K. (2008). The effect of rhythmic auditory stimulation on the gait parameters of
patients with incomplete spinal cord injury: An exploratory pilot study. International Journal of
Rehabilitation Research 31(2), 155–157.
Elliott, B. (1982). Guide to the selection of musical instruments with respect to physical ability and
disability. Saint Louis, MO: MMB Music, Inc.
Fernández-Miranda, J. C., Wang, Y., Pathak, S., Stefaneau, L., Verstynen, T., & Yeh, F. C. (2015).
Asymmetry, connectivity, and segmentation of the arcuate fascicle in the human brain. Brain
Structure & Function 220(3), 1665–1680.
Ford, M., Wagenaar, R., & Newell, K. (2007). The effects of auditory rhythms and instruction on
walking patterns in individuals post stroke. Gait and Posture 26(1), 150–155.
Fournier, K. A., Kimberg, C. I., Radonovich, K. J., Tillman, M. D., Chow, J. W., Lewis, M. H., …
Hass, C. J. (2010). Increased static and dynamic postural control in children with autism spectrum
disorders. Gait Posture 32(1): 6–9.
Frazzitta, G., Maestri, R., Uccellini, D., Bertotti, G., & Abelli, P. (2009). Rehabilitation treatment of
gait in patients with Parkinson’s disease with freezing: A comparison between two physical
therapy protocols using visual and auditory cues with or without treadmill training. Movement
Disorders 24(8), 1139–1143.
Freedland, R. L., Festa, C., Sealy, M., McBean, A., Elghazaly, P., Capan, A., … Rothman, J. (2002).
The effects of pulsed auditory stimulation on various gait measurements in persons with
Parkinson’s disease. Neurorehabilitation 17(1), 81–87.
Fujioka, T., Trainor, L. J., Large, E. W., & Ross, B. (2012). Internalized timing of isochronous sounds
is represented in neuromagnetic beta oscillations. Journal of Neuroscience 32(5), 1791–1802.
Ghai, S., Ghai, I., & Effenberg, A. O. (2017). Effect of rhythmic auditory cueing on gait in cerebral
palsy: A systematic review and meta-analysis. Neuropsychiatric Disease and Treatment 14, 43–59.
Ghai, S., Ghai, I., & Effenberg, A. O. (2018). Effect of rhythmic auditory cueing on aging gait: A
systematic review and meta-analysis. Aging and Disease 9(5), 901–923.
Ghai, S., Ghai, I., Schmitz, G., & Effenberg, A. O. (2018). Effect of rhythmic auditory cueing on
Parkinsonian gait: A systematic review and meta-analysis. Scientific Reports 8, 506.
Gillespie,L. D., Robertson, M. C., Gillespie, W. J., Sherrington, C., Gates, S., Clemson, L. M., &
Lamb, S. E. (2012). Interventions for preventing falls in older people living in the community.
Cochrane Database of Systematic Reviews 9, CD007146.
Grahn, J. A., & Brett, M. (2007). Rhythm and beat perception in motor areas of the brain. Journal of
Cognitive Neuroscience 19(5), 893–906.
Grau-Sánchez, J., Amengual, J. L., Rojo, N., Veciana de las Heras, M., Montero, J., Rubio, F., …
Rodríguez-Fornells, A. (2013). Plasticity in the sensorimotor cortex induced by music-supported
therapy in stroke patients: A TMS study. Frontiers in Human Neuroscience 7. Retrieved from
https://doi.org/10.3389/fnhum.2013.00494
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Howe, T. E., Lövgreen, B., Cody, F. W., Ashton, V. J., & Oldham, J. A. (2003). Auditory cues can
modify the gait of persons with early-stage Parkinson’s disease: A method for enhancing
Parkinsonian walking performance. Clinical Rehabilitation 17(4), 363–367.
Hurt, C. P., Rice, R. R., McIntosh, G. C., & Thaut, M. H. (1998). Rhythmic auditory stimulation in
gait training for patients with traumatic brain injury. Journal of Music Therapy 35(4), 228–241.
Hurt-Thaut, C. P. (2014). Rhythmic auditory stimulation to reduce falls in healthy elderly and
patients with Parkinson’s disease (Doctoral dissertation). UMI dissertation publishing, 3635683.
Johnell, O., Melton, L. J. III, Atkinson, E. J., O’Fallon, W. M., & Kurland, L. T. (1992). Fracture risk
in patients with Parkinsonism: A population-based study in Olmsted County, Minnesota. Age and
Ageing 21(1), 32–38.
Kadivar, Z., Corcos, D. M., Foto, J., & Hondzinski, J. M. (2011). Effect of step training and rhythmic
auditory stimulation on functional performance in Parkinson patients. Neurorehabilitation and
Neural Repair 25(7), 626–635.
Kim, J. S., & Oh, D. W. (2012). Home-based auditory stimulation training for gait rehabilitation of
chronic stroke patients. Journal of Physical Therapy Science 24(8), 775–777.
Kim, S., Kwak, E., Park, E., & Cho, S. (2012). Differential effects of rhythmic auditory stimulation
and neurodevelopmental treatment/Bobath on gait patterns in adults with cerebral palsy: A
randomized controlled trial. Clinical Rehabilitation 26(10), 904–914.
Kim, S. J., Kwak, E. E., Park, E. S., Lee, D. S., Kim, K. J., Song, J. E., & Cho, S. R. (2011). Changes
in gait patterns with rhythmic auditory stimulation in adults with cerebral palsy.
Neurorehabilitation 29(3), 233–241.
Kojovic, M., Pareés, I., Sadnicka, A., Kassavetis, P., Rubio-Agusti, I., Saifee, T. A., & Bhatia, K. P.
(2012). The brighter side of music in dystonia. Archives of Neurology 69(7), 917–919.
Luft, A. R., McCombe-Waller, S., Whitall, J., Forrester, L. W., Macko, R., Sorkin, J. D., … Hanley,
D. F. (2004). Repetitive bilateral arm training and motor cortex activation in chronic stroke: A
randomized controlled trial. Journal of the American Medical Association 292(15), 1853–1861.
Ma, H. I., Hwang, W. J., & Lin, K. C. (2009). The effects of two different auditory stimuli on
functional arm movement in persons with Parkinson’s disease: A dual-task paradigm. Clinical
Rehabilitation 23(3), 229–237.
McCombe Waller, S., Harris-Love, M., Liu, W., & Whitall, J. (2006). Temporal coordination of the
arms during bilateral simultaneous and sequential movements in patients with chronic hemiparesis.
Experimental Brain Research 168(3), 450–454.
McIntosh, G. C., Brown, S. H., Rice, R. R., & Thaut, M. H. (1997). Rhythmic auditory-motor
facilitation of gait patterns in patients with Parkinson’s disease. Journal of Neurology,
Neurosurgery, and Psychiatry 62(1), 22–26.
McIntosh, G. C., Rice, R. R., Hurt, C. P., & Thaut, M. H. (1998). Long-term training effects of
rhythmic auditory stimulation on gait in patients with Parkinson’s disease. Movement Disorders
13(2), 212 [Abstract].
Mak, M. (2006). Feed-forward audio-visual cues could enhance sit-to-stand in Parkinsonian patients.
Proceedings of the 4th World Congress for Neurorehabilitation, F1B-7.
Malcolm, M. P., Massie, C., & Thaut, M. H. (2009). Rhythmic auditory-motor entrainment improves
hemiparetic arm kinematics during reaching movements: A pilot study. Topics in Stroke
Rehabilitation 16(1), 69–79.
Merchant, H., Grahn, J., Trainor, L., Rohrmeier, M., & Fitch, W. T.(2015). Finding the beat: A neural
perspective across humans and non-human primates. Philosophical Transactions of the Royal
Society B: Biological Sciences 370(1664), 20140093.
Miller, R. A., Thaut, M. H., McIntosh, G. C., & Rice, R. R. (1996). Components of EMG symmetry
and variability in Parkinsonian and healthy elderly gait. Electroencephalography and Clinical
Neurophysiology/Electromyography and Motor Control 101(1), 1–7.
Molinari, M., Leggio, M., Filippini, V., Gioia, M., Cerasa, A., & Thaut, M. (2005). Sensorimotor
transduction of time information is preserved in subjects with cerebellar damage. Brain Research
Bulletin 67(6), 448–458.
Morris, G. S., Suteerawattananon, M., Etnyre, B. R., Jankovic, J., & Protas, E. J. (2004). Effects of
visual and auditory cues on gait in individuals with Parkinson’s disease. Journal of the
Neurological Sciences 219(1–2), 63–69.
Pacchetti, C., Mancini, F., Aglieri, R., Fundaro, C., Martignoni, E., & Nappi, G. (2000). Active music
therapy in Parkinson’s disease: An integrative model method for motor and emotional
rehabilitation, Psychosomatic Medicine 62(3), 386–393.
Paltsev, Y. I., & Elner, A. M. (1967). Change in the functional state of the segmental apparatus of the
spinal cord under the influence of sound stimuli and its role in voluntary movement. Biophysics
12, 1219–1226.
Pau, M., Corona, F., Pili, R., Casula, C., Sors, F., Agostini, T., … Murgia, M. (2016). Effects of
physical rehabilitation integrated with rhythmic auditory stimulation on spatio-temporal and
kinematic parameters of gait in Parkinson’s disease. Frontiers in Neurology 7. Retrieved from
https://doi.org/10.3389/fneur.2016.00126
Peng, Y. C., Lu, W. T., Wang, T. H., Chen, Y. L., Liao, H. F., Lin, K. H., & Tang, P. F. (2011).
Immediate effects of therapeutic music on loaded sit-to-stand movement in children with spastic
diplegia. Gait & Posture 33(2), 274–278.
Picelli, A., Camin, M., Tinazzi, M., Vangelista, A., Cosentino, A., Fiaschi, A., & Smania, N. (2010).
Three-dimensional motion analysis of the effects of auditory cueing on gait pattern in patients with
Parkinson’s disease: A preliminary investigation. Neurological Sciences 31(4), 423–430.
Richards, C. L., Malouin, F., Bedard, P. J., & Cioni, M. (1992). Changes induced by L-Dopa and
sensory cues on the gait of Parkinsonian patients. In M. Wollacot & F. Horak (Eds.), Posture and
gait: Control mechanisms (pp. 126–129). Eugene, OR: University of Oregon Books.
Ringenbach, S. D., Zimmerman, K., Chen, C. C., Mulvey, G. M., Holzapfel, S. D., Weeks, D. J., &
Thaut, M. H. (2014). Adults with Down syndrome performed repetitive movements fast with
continuous music cues. Journal of Motor Learning and Development 2(3), 47–54.
Robertson, S. D., Chua, R., Maraj, B. K., Kao, J. C., & Weeks, D. J. (2002). Bimanual coordination
dynamics in adults with Down syndrome. Motor Control 6(4), 388–407.
Robertson, S. D., Van Gemmert, A. W., & Maraj, B. K. (2002). Auditory information is beneficial for
adults with Down syndrome in a continuous bimanual task. Acta Psychologica 110(2), 213–229.
Roerdink, M., Bank, P. J., Peper, C. L., & Beek, P. J. (2011). Walking to the beat of different drums:
Practical implications for the use of acoustic rhythms in gait rehabilitation. Gait & Posture 33(1),
690–694.
Rojo, N., Amengual, J., Juncadella, M., Rubio, F., Camara, E., Marco-Pallares, J., … Altenmüller, E.
(2011). Music-supported therapy induces plasticity in the sensorimotor cortex in chronic stroke: A
single-case study using multimodal imaging (fMRI-TMS). Brain Injury 25(7–8), 787–793.
Rossignol, S., & Melvill Jones, G. (1976). Audiospinal influences in man studied by the H-reflex and
its possible role in rhythmic movement synchronized to sound. Electroencephalography &
Clinical Neurophysiology 41(1), 83–92.
Sameiro-Barbosa, C. M., & Geiser, E. (2016). Sensory entrainment mechanisms in auditory
perception: Neural synchronization cortico-striatal activation. Frontiers in Neuroscience 10.
Retrieved from https://doi.org/10.3389/fnins.2016.00361
Schaffert, N., & Mattes, K. J. (2015). Effects of acoustic feedback training in elite-standard para-
rowing. Journal of Sports Science 33(4), 411–418.
Schauer, M., & Mauritz, K. H. (2003). Musical motor feedback (MMF) in walking hemiparetic
stroke patients: Randomized trials of gait improvement. Clinical Rehabilitation 17(7), 713–722.
Schmahmann, J. D., & Pandya, D. N. (2008). Disconnection syndromes of basal ganglia, thalamus,
and cerebrocerebellar systems. Cortex 44(8), 1037–1066.
Schmitz, G., Kroeger, D., & Effenberg, A. O. (2014). A mobile sonification system for stroke
rehabilitation. Paper presented at the 20th International Conference on Auditory Display, New
York.
Schneider, S., Münte, T., Rodriguez-Fornells, A., Sailer, M., & Altenmüller, E. (2010). Music-
supported training is more efficient than functional motor training for recovery of fine motor skills
in stroke patients. Music Perception: An Interdisciplinary Journal 27(4), 271–280.
Seebacher, B., Kuisma, R., Glynn, A., & Berger, T. (2015). Rhythmic cued motor imagery and
walking in people with multiple sclerosis: A randomised controlled feasibility study. Pilot and
Feasibility Studies 1, 25. doi:10.1186/s40814-015-0021-3
Seebacher, B., Kuisma, R., Glynn, A., & Berger, T. (2016). The effect of rhythmic-cued motor
imagery on walking, fatigue and quality of life in people with multiple sclerosis: A randomised
controlled trial. Multiple Sclerosis Journal 23(2), 286–296.
Shahraki, M., Sohrabi, M., Torbati, H. T., Nikkhah, K., & NaeimiKia, M. (2017). Effect of rhythmic
auditory stimulation on gait kinematic parameters of patients with multiple sclerosis. Journal of
Medicine and Life 10(1), 33–37.
Son, H., & Kim, E. (2015). Kinematic analysis of arm and trunk movements in the gait of
Parkinson’s disease patients based on external signals. Journal of Physical Therapy Science
27(12), 3783–3786.
Song, G. B., & Ryu, H. J. (2016). Effects of gait training with rhythmic auditory stimulation on gait
ability in stroke patients. Journal of Physical Therapy Science 28(5), 1403–1406.
Song, J. H., Zhou, P. Y., Cao, Z. H., Ding, Z. G., Chen, H. X., & Zhang, G. B. (2015). Rhythmic
auditory stimulation with visual stimuli on motor and balance function of patients with Parkinson’s
disease. European Review for Medical and Pharmacological Sciences 19(11), 2001–2007.
Spaulding, S. J., Barber, B., Colby, M., Cormack, B., Mick, T., & Jenkins, M. E. (2013). Cueing and
gait improvement among people with Parkinson’s disease: A meta-analysis. Archives of Physical
Medicine and Rehabilitation 94(3), 562–570.
Stephan, K. M., Thaut, M. H., Wunderlich, G., Schicks, W., Tian, B., Tellmann, L., … Hoemberg, V.
(2002). Conscious and subconscious sensorimotor synchronization: Prefrontal cortex and the
influence of awareness. NeuroImage 15(2), 345–352.
Suh, J. H., Han, S. J., Jeon, S. Y., Kim, H. J., Lee, J. E., Yoon, T. S., & Chong, H. J. (2014). Effect of
rhythmic auditory stimulation on gait and balance in hemiplegic stroke patients.
Neurorehabilitation 34(1), 193–199.
Teki, S., Grube, M., & Griffiths, T. D. (2012). A unified model of time perception accounts for
duration-based and beat-based timing mechanisms. Frontiers in Integrative Neuroscience 5.
Retrieved from https://doi.org/10.3389/fnint.2011.00090
Teki, S., Grube, M., Kumar, S., & Griffiths, T. D. (2011). Distinct neural substrates of duration-based
and beat-based auditory timing. Journal of Neuroscience 31(10), 3805–3812.
Thaut, M. H. (2005). The future of music in therapy and medicine. Annals of the New York Academy
of Sciences 1060, 303–308.
Thaut, M. H., & Hoemberg, V. (Eds.) (2014). The Oxford handbook of neurologic music therapy.
Oxford: Oxford University Press.
Thaut, M. H., Hoemberg, V., Kenyon, G., & Hurt, C. P. (1998). Rhythmic entrainment of hemiparetic
arm movements in stroke patients. Proceedings of the Society for Neuroscience 653(7) [Abstract].
Thaut, M. H., & Kenyon, G. P. (2003). Rapid motor adaptations to subliminal frequency shifts during
syncopated rhythmic sensorimotor synchronization. Human Movement Science 22(3), 321–338.
Thaut, M. H., Kenyon, G. P., Hurt, C. P., McIntosh, G. C., & Hoemberg, V. (2002). Kinematic
optimization of spatiotemporal patterns in paretic arm training with stroke patients.
Neuropsychologia 40(7), 1073–1081.
Thaut, M. H., Lange, H., Miltner, R., Hurt, C. P., & Hoemberg, V. (1996). Rhythmic entrainment of
gait patterns in Huntington’s disease patients. Proceedings of the Society for Neuroscience 727(6)
[Abstract].
Thaut, M. H., Leins, A. K., Rice, R. R., Argstatter, H., Kenyon, G. P., McIntosh, G. C., & Fetter, M.
(2007). Rhythmic auditory stimulation improves gait more than NDT/Bobath training in near-
ambulatory patients early post-stroke: A single-blind, randomized trial. Neurorehabilitation and
Neural Repair 21(5), 455–459.
Thaut, M. H., McIntosh, G. C., & Hoemberg, V. (2014). Neurobiological foundations of neurologic
music therapy: Rhythmic entrainment and the motor system. Frontiers in Psychology 5. Retrieved
from https://doi.org/10.3389/fpsyg.2014.01185
Thaut, M. H., McIntosh, G. C., Prassas, S. G., & Rice, R. R. (1993). The effect of auditory rhythmic
cuing on stride and EMG patterns in hemiparetic gait of stroke patients. Journal of Neurologic
Rehabilitation 7(1), 9–16.
Thaut, M. H., McIntosh, G. C., & Rice, R. R. (1997). Rhythmic facilitation of gait training in
hemiparetic stroke rehabilitation. Journal of Neurological Sciences 151(2), 207–212.
Thaut, M. H., McIntosh, G. C., Rice, R. R., Miller, R. A., Rathbun, J., & Brault, J. M. (1996).
Rhythmic auditory stimulation in gait training for Parkinson’s disease patients. Movement
Disorders 11(2), 193–200.
Thaut, M. H., Miltner, R., Lange, H. W., Hurt, C. P., & Hoemberg, V. (1999). Velocity modulation
and rhythmic synchronization of gait in Huntington’s disease. Movement Disorders 14(5), 808–
819.
Thaut, M. H., Nickel, A., Kenyon, G. P., Meissner, N., & McIntosh, G. C. (2005). Rhythmic auditory
stimulation (RAS) for gait training in hemiparetic stroke rehabilitation: An international
multicenter study. Proceedings of the Society for Neuroscience 756(6).
Thaut, M. H., Rice, R. R., Braun Janzen, T., Hurt-Thaut, C. P., & McIntosh, G. C. (2018). Rhythmic
auditory stimulation for reduction of falls in Parkinson’s disease: A randomized controlled study.
Clinical Rehabilitation, July 23. doi:10.1177/0269215518788615
Thaut, M. H., Schicks, W., McIntosh, G. C., & Hoemberg, V. (2002). The role of motor imagery and
temporal cuing in hemiparetic arm rehabilitation. Neurorehabilitation and Neural Repair 16, 115.
Thaut, M. H., Schleiffers, S., & Davis, W. B. (1991). Analysis of EMG activity in biceps and triceps
muscle in a gross motor task under the influence of auditory rhythm. Journal of Music Therapy 28,
64–88.
Torres, E. B., Brincker, M., Isenhower, R. W., Yanovich, P., Stigler, K. A., Nurnberger, J. I., … José,
J. V. (2013). Autism: The micro-movement perspective. Frontiers in Integrative Neuroscience 7.
Retrieved from https://doi.org/10.3389/fnint.2013.00032
Torres, E. B., & Donnellan, A. M. (2015). Editorial for research topic “Autism: The movement
perspective.” Frontiers in Integrative Neuroscience 9. Retrieved from
https://doi.org/10.3389/fnint.2015.00012
Turova, V., Alves-Pinto, A., Ehrlich, S., Blumenstein, T., Cheng, G., & Lampe, R. (2017). Effects of
short-term piano training on measures of finger tapping, somatosensory perception and motor-
related brain activity in patients with cerebral palsy. Neuropsychiatric Disease and Treatment 13,
2705–2718.
Wang, T. H., Peng, Y. C., Chen, Y. L., Lu, T. W., Liao, H. F., Tang, P. F., & Shieh, J. Y. (2013). A
home-based program using patterned sensory enhancement improves resistance exercise effects for
children with cerebral palsy: A randomized controlled trial. Neurorehabilitation and Neural Repair
27(8), 684–694.
Whitall, J., McCombe Waller, S., Silver, K. H., & Macko, R. F. (2000). Repetitive bilateral arm
training with rhythmic auditory cueing improves motor function in chronic hemiparetic stroke.
Stroke 31(10), 2390–2395.
Willems, A. M., Nieuwboer, A., Chavret, F., Desloovere, K., Dom, R., Rochester, L., … Van Wegen,
E. (2006). The use of rhythmic auditory cues to influence gait in patients with Parkinson’s disease,
the differential effect for freezers and non-freezers, an explorative study. Disability and
Rehabilitation 28(11), 721–728.
Wood, B. H., Bilclough, J. A., Bowron, A., & Walker, R. W. (2002). Incidence and prediction of falls
in Parkinson’s disease: A prospective multidisciplinary study. Journal of Neurology, Neurosurgery
& Psychiatry 72(6), 721–725.
Yoo, J. (2009). The role of therapeutic instrumental music performance in hemiparetic arm
rehabilitation. Music Therapy Perspectives 27(1), 16–24.
Yoo, G. E., & Kim, S. J. (2016). Rhythmic auditory cueing in motor rehabilitation for stroke patients:
Systematic review and meta-analysis. Journal of Music Therapy 53(2), 149–177.
Zatorre, R. J., Halpern, A. R., & Herholz, S. C. (2012). Neuronal correlates of perception, imagery,
and memory for familiar tunes. Journal of Cognitive Neuroscience 24(6), 1382–1397.
CHAPT E R 29
NEUROLOGIC MUSIC
THERAPY FOR SPEECH
AND LANGUAGE
R E H A B I L I TAT I O N
Y U N E S . L E E, C O R E N E T H A U T, A N D C H A R L E N E
S A N TO N I
D
Motor speech disorders (MSDs) can occur due to neurologic impairments
affecting the planning, programming, control, or execution of speech.
MSDs include the dysarthrias and apraxia of speech (Duffy, 2005, p. 5).
Dysarthria is a collective term referring to a neuropathophysiologic
disruption in the activation and control (e.g., strength, speed, range of
motion, tone, coordination) of the muscles necessary for speech production.
Dysarthria therefore, can affect the respiratory, phonatory, resonatory,
articulatory, and prosodic aspects of speech. Several categories exist:
flaccid, spastic, hypokinetic, hyperkinetic, ataxic, upper motor neuron, and
mixed; all resulting from damage or disturbance to the upper or lower
motor neurons, basal ganglia, or cerebellum (Darley, Aronson, & Brown,
1969; Duffy, 2005).
Singing and speech share the same proprioceptive feedback system.
Guenther’s Directions into Velocities of Articulators (DIVA) model
describes a segmental theory of speech motor control which proposes that
speech segments are coded by the central nervous system (CNS) as
auditory-temporal and somatosensory-temporal goal regions, and that two
controls drive a speech sound map: feedforward and feedback. The
feedforward mechanism outlines how the CNS sends anticipatory pre-
programmed instructions about movements by relying on past experiences
in movement planning, execution, and error correction. The feedback
mechanism provides scaffolding for how speech movement is controlled
based on the sensory input the CNS receives, which may indicate deviation
from the planned movement (Guenther, 2006; Guenther & Vladusich, 2012;
Tourville & Guenther, 2011). In the domain of speech and language
rehabilitation, the task of singing could be theorized as able to induce
neuromotor retraining via the formation of new motor command
relationships within the feedback mechanism, that stimulate learning within
the feedforward mechanism, thereby causing the CNS to re-calibrate or
reset its motor program for communication. In addition, since singing
naturally lends itself to heightening various elements of speech production
as an augmentative form of vocal loading, respiratory shaping, resonant
voicing, exaggerated articulation, and prosodic phrasing, singing could also
be theorized as able to modulate motor neuron activity; carrying with it
implications for rehabilitation (Cohen, 1994; Natke, Donath, & Kalveram,
2003; Tonkinson, 1994). Hereafter lies a review of current singing-related
voice therapy strategies prescribed to specific motor speech disordered
populations and their outcomes. Elucidation of the practice of singing as a
therapeutic science with reproducible effects is the main construct of the
review.
There is a significant profusion of literature reporting positive outcomes
for utilizing singing tasks as a means of voice therapy in dysarthric
populations. In traumatic brain injury and stroke, singing-induced gains
have been documented in areas related to maximum phonation time,
intensity, speech rate, prosody, vocal range, and overall intelligibility
(Baker, Wigram, & Gold, 2005; Cohen, 1988, 1992; Kim & Jo, 2013;
Tamplin, 2008). Therapeutic outcomes using NMT speech techniques such
as VIT, TS, and OMREX in Parkinson’s disease have also revealed
significant improvements in the areas of hypomimia, vocal intensity,
fundamental frequency, maximum phonation time, prosody, articulation,
and better lung function test scores, overall (Caligiuri, 1989; Canavan,
Evans, Foy, Langford, & Proctor, 2012; DeStewart, Willemse, Maassen, &
Horstink, 2003; Di Benedetto et al., 2009; Elefant, Baker, Lotan, Lagesen,
& Skeie, 2012; Haneishi, 2001; Stegemöller, Radig, Hibbing, Wingate, &
Sapienza, 2017; Tanner, 2012; Tanner, Rammage, & Liu, 2016; Tautscher-
Basnett, Tomantschger, Keglevic, & Freimuller, 2006). Accordingly, in
earlier work by Bellaire, Yorkston, and Beukelman (1986), the modification
of the breathing pattern of mildly dysarthric speakers resulted in the
amelioration of prosodic repertoire. Similarly, in 1993, Cohen and Masse
applied a singing intervention to persons with neurogenic communication
disorders, symptomatic of multiple sclerosis, cerebral palsy, Parkinson’s
disease, and cerebrovascular accident. Findings revealed improvements in
intelligibility ratings, vocal intensity, and vocal range.
The significance of singing in human development has always had firm
roots in our evolutionary inheritance: Charles Darwin theorized that
Neanderthals originally communicated using a catalog of song-like
expressions lacking in words or meaning (Darwin, 1872/1988).
Furthermore, recent research provides description of a phenomenon known
as infant-directed speech, or musical speech as being catalytic for preverbal
communication and an important stage in language learning in the earliest
stages of life (Fernald, 1989; Trainor, Clark, Huntley, & Adams, 1997). As
such, singing’s current emergence (or re-emergence) as speech’s keen and
remunerative partner, implies that the pairing has been evident all along,
and that prescribing singing training to motor speech disordered populations
is, therefore, reflective of a more refined understanding of where we came
from.
A S
S D
V D
C
There are many parallels between both the structure and production of
speech, language, and music. All can be considered sensorimotor behaviors
that require a high level of control and dynamic interplay between several
brain processes in order to select, organize, articulate, and implement in a
time-sensitive manner. Because of the inherent timing, rhythm, pattern, and
melodic structures in both music and speech, music has the potential to
simulate normal speech patterns, and therefore act as a training and
retraining tool for people with speech and language disorders.
R
Abdi, S., Khalessi, M. H., Khorsandi, M., & Gholami, B. (2001). Introducing music as a means of
habilitation for children with cochlear implants. International Journal of Pediatric
Otorhinolaryngology 59(2), 105–113.
Aitken Dunham, D. J. (2010). Efficacy of using music therapy combined with traditional aphasia and
apraxia of speech treatments (Master’s dissertation). Western California University, North
Carolina.
Akanuma, K., Meguro, K., Satoh, M., Tashiro, M., & Itoh, M. (2016). Singing can improve speech
function in aphasics associated with intact right basal ganglia and preserve right temporal glucose
metabolism: Implications for singing therapy indication. International Journal of Neuroscience
126(1), 39–45.
Alm, P. A. (2004). Stuttering and the basal ganglia circuits: A critical review of possible relations.
Journal of Communication Disorders 37(4), 325–69.
American Psychiatric Association (2013). Diagnostic and statistical manual of mental disorders (5th
ed.). Washington, DC: APA.
Austin, S. F. (1997). Movement of the velum during speech and singing in classically trained singers.
Journal of Voice—Official Journal of the Voice Foundation 11(2), 212–221.
Baker, F., Wigram, T., & Gold, C. (2005). The effects of song-singing programme on the affective
speaking intonation of people with traumatic brain injury. Brain Injury 19(7), 519–28.
Ballard, K. J., Wambaugh, J. L., Duffy, J. R., Layfield, C., Maas, E., Mauszycki, S., & McNeil, M. R.
(2015). Treatment for acquired apraxia of speech: A systematic review of intervention research
between 2004 and 2012. American Journal of Speech Language Pathology 24(2), 316–337.
Bedoin, N., Brisseau, L., Molinier, P., Roch, D., & Tillmann, B. (2016). Temporally regular musical
primes facilitate subsequent syntax processing in children with specific language impairment.
Frontiers in Neuroscience 10. Retrieved from
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4913515/
Belin, P., Van Eeckhout, P., Zilbovicius, M., Remy, P., François, C., Guillaume, S., Chain, F., …
Samson, Y. (1996). Recovery from nonfluent aphasia after melodic intonation therapy: A PET
study. Neurology 47(6), 1504–1511.
Bellaire, K., Yorkston, K. M., & Beukelman, D. R. (1986). Modification of breath patterning to
increase naturalness of a mildly dysarthric speaker. Journal of Communication Disorders 19(4),
271–280.
Bhide, A., Power, A., & Goswami, U. (2013). A rhythmic musical intervention for poor readers: A
comparison of efficacy with a letter-based intervention. Mind, Brain, and Education 7(2), 113–
123.
Birch, P., Gümoes, B., Stavad, H., Prytz, S., Björkner, E., & Sundberg, J. (2002). Velum behavior in
professional classic operatic singing. Journal of Voice 16(1), 61–71.
Bloch, C. S., Hirano, M., & Gould, W. J. (1985). Symptom improvement of spastic dysphonia in
response to phonatory tasks. Annals of Otology, Rhinology, & Laryngology 94, 51–4.
Bonacina, S., Cancer, A., Lanzi, P. L., Lorusso, M. L., & Antonietti, A. (2015). Improving reading
skills in students with dyslexia: The efficacy of a sublexical training with rhythmic background.
Educational Psychology 6(1510), 1–8.
Bonakdarpour, B., Eftekharzadeh, A., & Ashayeri, H. (2003). Melodic intonation therapy in Persian
aphasic patients. Aphasiology 17(1), 75–95.
Boone, D. R. (1983). The voice and voice therapy (3rd ed.). Englewood Cliffs, NJ: Prentice Hall.
Boone, D. R., McFarlane, S. C., Von Berg, A. L., & Zraick, R. I. (2010). The voice and voice therapy
(8th ed.). Boston, MA: Pearson.
Boucher, V., Garcia, L. J., Fleurant, J., & Paradis, J. (2001). Variable efficacy of rhythm and tone in
melody based interventions: Implications for the assumption of a right-hemisphere facilitation in
non-fluent aphasia. Aphasiology 15(2), 131–149.
Breier, J. I., Randle, S., Maher, L. M., & Papanicolaou, A. C. (2010). Changes in maps of language
activity activation following melodic intonation therapy using magnetoencephalography: Two case
studies. Journal of Clinical and Experimental Neuropsychology 32(3), 309–314
Brendal, B., & Ziegler, W. (2008). Effectiveness of metrical pacing in the treatment of apraxia of
speech. Aphasiology 22(1), 1–26.
Busto-Crespo, O., Uzcanga-Lacabe, M., Abad-Marco, A., Berasategui, I., García, L., Maraví, E., …
Fernández-González, S. (2016). Longitudinal voice outcomes after voice therapy in unilateral
vocal fold paralysis. Journal of Voice 30(6), 767.e9–767.e15.
Caligiuri, M. P. (1989). The influence of speaking rate on articulatory hypokinesia in Parkinsonian
dysarthria. Brain & Language 36(3), 1493–1502.
Canavan, M., Evans, C., Foy, C., Langford, R., & Proctor, R. (2012). Can group singing provide
effective speech therapy for people with Parkinson’s disease? Arts and Health 4(1), 83–95.
Canga, B., Azoulay, R., Raskin, J., & Loewy, J. (2015). AIR: Advances in respiration—music
therapy in the treatment of chronic pulmonary disease. Respiratory Medicine 109(12), 1532–1539.
Casper, J. (2000). Confidential voice. In J. C. Stemple (Ed.), Voice therapy: Clinical studies (2nd ed.,
pp. 128–139). San Diego, CA: Singular Publishing Group.
Chen, J. K.-C., Chuang, A. Y. C., McMahon, C., Hsieh, J. C., Tung, T. H., & Li, L. P. (2010). Music
training improves pitch perception in prelingually deafened children with cochlear implants.
Pediatrics 125(4), e793–e800.
Chen, S. H., Hsiao, T. Y., Hsiao, L. C., Chung, Y. M., & Chiang, S. C. (2007). Outcome of resonant
voice therapy for female teachers with voice disorders: Perceptual, physiological, acoustic,
aerodynamic, and functional measurements. Journal of Voice 21(4), 415–425.
Cohen, N. S. (1988). The use of superimposed rhythm to decrease the rate of speech in a brain-
damaged adolescent. Journal of Music Therapy 25(2), 85–93.
Cohen, N. S. (1992). The effect of singing instruction on the speech production of neurologically
impaired persons. Journal of Music Therapy 29(2), 87–102.
Cohen, N. S. (1994). Speech and song: Implications for therapy. Music Therapy Perspectives 12(1),
8–14.
Cohen, N. S., & Masse, R. (1993). The application of singing and rhythmic instruction as a
therapeutic intervention for persons with neurogenic communication disorders. Journal of Music
Therapy 30(2), 81–99.
Colcord, R. D., & Adams, M. R. (1979). Voicing duration and vocal SPL changes associated with
stuttering reduction during singing. Journal of Speech and Hearing Research 22(3), 468–479.
Colton, R. H., & Casper, J. K. (1996). Understanding voice problems: A physiological perspective
for diagnosis and treatment (3rd ed.). Baltimore, MD: Lippincott.
Cooper, M. (1973). Modern techniques of vocal rehabilitation. Springfield, IL: Charles C. Thomas.
Cortese, M. D., Riganello, F., Arcuri, F., Pignataro, L. M., & Buglione, I. (2015). Rehabilitation of
aphasia: Application of melodic-rhythmic therapy to Italian language. Frontiers in Human
Neuroscience 9, 1–8. Retrieved from https://doi.org/10.3389/fnhum.2015.00520
Cunnington, R., Bradshaw, J. L., & Iansek, R. (1996). The role of the supplementary motor area in
the control of voluntary movement. Human Movement Science 15(5), 627–647.
Darley, F. L., Aronson, A. E., & Brown, J. R. (1969). Differential diagnostic patterns of dysarthria.
Journal of Speech and Hearing Research 12(2), 246–269.
Darwin, C. (1872/1988). The expression of the emotions in man and animals. Ed. P. Ekman. Oxford:
Oxford University Press.
Desai, V., & Mishra, P. (2012). Voice therapy outcome in puberphonia. Journal of Laryngology and
Voice 2(1), 26–29.
DeStewart, B. J., Willemse, S. C., Maassen, B. A. M., & Horstink, M. W. I. M. (2003). Improvement
of voicing in patients with Parkinson’s disease by speech therapy. Neurology 60(3), 498–500.
Di Benedetto, P., Cavazzon, M., Mondolo, F., Rugiu, G., Peratoner, A., & Biasutti, E. (2009). Voice
and choral singing treatment: A new approach for speech and voice disorders in Parkinson’s
disease. European Journal of Physical Rehabilitation Medicine 45(1), 13–19.
Duffy, J. R. (2005). Motor speech disorders: Substrates, differential diagnosis, and management. St.
Louis, MO: Elsevier Mosby.
Dworkin, J. P., Abkarian, G. G., & Johns, D. F. (1988). Apraxia of speech: The effectiveness of a
treatment regimen. Journal of Speech and Hearing Disorders 53(3), 280–294.
Elefant, C., Baker, F. A., Lotan, M., Lagesen, S. K., & Skeie, G. O. (2012). The effect of group music
therapy on mood, speech, and singing in individuals with Parkinson’s disease: A feasibility study.
Journal of Music Therapy 49(3), 278–302.
Eley, R., & Gorman, D. (2010). Didgeridoo playing and singing to support asthma management in
aboriginal Australians. Journal of Rural Health 26(1), 100–104.
Engen, R. L. (2003). The singer’s breath: Implications for treatment of persons with emphysema
(Dissertation). University of Iowa.
Farrugia, N., Benoit, C. E., Schwartze, M., Pell, M., Obrig, H., Dalla Bella, S., & Kotz, S. (2014).
Auditory cueing in Parkinson’s disease: Effects on temporal processing and spontaneous theta
oscillations. Procedia—Social and Behavioral Sciences 126, 104–105. Special Issue for
International Conference on Timing and Time Perception, March 31–April 3, Corfu, Greece.
Fernald, A. (1989). Intonation and communicative intent in mothers’ speech to infants. Is the melody
the message? Child Development 60(6), 1497–1510.
Flaugnacco, E., Lopez, L., Terribili, C., Montico, M., Zoia, S., & Schön, D. (2015). Music training
increases phonological awareness and reading skills in developmental dyslexia: A randomized
control trial. PloS ONE 10(9), e0138715.
Fowler, L. P., & Morris, R. J. (2007). Comparison of fundamental frequency nasalance between
trained singers and nonsingers for sung vowels. Annals of Otology, Rhinology, & Laryngology
116(10), 739–746.
Friederici, A. D., Kotz, S. A., Werheid, K., Hein, G., & Von Cramon, D. (2003). Syntactic
comprehension in Parkinson’s disease: Investigating early automatic and late integrational
processes using event-related brain potentials. Neuropsychology 17(1), 133–142.
Fu, Q.-J., Galvin, J. J., Wang, X., & Wu, J. L. (2015). Benefits of music training in Mandarin-
speaking pediatric cochlear implant users. Journal of Speech, Language, and Hearing Research
58(1), 163–169.
Fujii, S., & Wan, C. Y. (2014). The role of rhythm in speech and language rehabilitation: The SEP
hypothesis. Frontiers in Human Neuroscience 8, 1–15. doi:10.3389/fnhum.2014.00777
Gfeller, K. (2016). Music-based training for pediatric CI recipients: A systematic analysis of
published studies. European Annals of Otorhinolaryngology, Head and Neck Diseases, 12th
European Symposium on Pediatric Cochlear Implant (ESPCI 2015) 133(Suppl. 1), S50–S56.
Gfeller, K., Driscoll, V., Smith, R. S., & Scheperle, C. (2012). The music experiences and attitudes of
a first cohort of prelingually-deaf adolescents and young adults CI recipients. Seminars in Hearing
33(4), 346–360.
Glover, H., Kalinowski, J., Rastatter, M., & Stuart, A. (1996). Effect of instruction to sing on
stuttering frequency at normal and fast rates. Perceptual and Motor Skills 83(2), 511–522.
Gordon, R. L., Shivers, C. M., Wieland, E. A., Kotz, S. A., Yoder, P. J., & McAuley, J. D. (2015).
Musical rhythm discrimination explains individual differences in grammar skills in children.
Developmental Science 18(4), 635–644.
Grahn, J. A., & Brett, M. (2009). Impairment of beat-based rhythm discrimination in Parkinson’s
disease. Cortex 45(1), 54–61.
Guenther, F. H. (2006). Cortical interactions underlying the production of speech sounds. Journal of
Communication Disorders 39(5), 350–365.
Guenther, F. H., & Vladusich, T. (2012). A neural theory of speech acquisition and production.
Journal of Neurolinguistics 25(5), 408–422.
Habib, M., Lardy, C., Desiles, T., Commeiras, C., Chobert, J., & Besson, M. (2016). Music and
dyslexia: A new musical training method to improve reading and related disorders. Frontiers in
Psychology 7. Retrieved from
http://journal.frontiersin.org/article/10.3389/fpsyg.2016.00026/abstract
Haneishi, E. (2001). Effects of a music therapy voice protocol on speech intelligibility, vocal acoustic
measures, and mood of individuals with Parkinson’s disease. Journal of Music Therapy 38(4),
273–290.
Haro-Martínez, A. M., García-Concejero, V. E., López-Ramos, A., Maté-Arribas, E., López-Táppero,
J., Lubrini, G., … & Fuentes, B. (2017). Adaptation of melodic intonation therapy to Spanish: A
feasibility pilot study, Aphasiology 31(11), 1333–1343.
Healey, E. C., Mallard, A. R., & Adams, M. R. (1976). Factors contributing to the reduction of
stuttering during singing. Journal of Speech and Hearing Research 19, 475–480.
Helfrich-Miller, K. R. (1984). Melodic intonation therapy with developmentally apraxic children.
Seminars in Speech and Language 5, 119–126.
Hilton, M. P., Savage, J., Hunter, B., McDonald, S., Repanos, C., & Powell, R. (2013). Singing
exercises improve sleepiness and frequency of snoring among snorers: A randomised controlled
trial. International Journal of Otolaryngology and Head & Neck Surgery 2(3), 97–102.
Irons, J. Y., Kenny, D. T., McElrea, M., & Chang, A. B. (2012). Singing therapy for young people
with cystic fibrosis: A randomized controlled pilot study. Music and Medicine 4(3), 136–145.
Jamaly, S., Leidag, M., Schneider, H. W., Domanksi, U., Rasche, K., Schröder, M., & Nilius, G.
(2017). The effect of singing therapy compared to standard physiotherapeutic lung sport in COPD.
Pneumologie 71(S01), S1–S125.
Jennings, J. J., & Kuehn, D. P. (2008). The effects of frequency range, vowel, dynamic loudness
level, and gender on nasalance in amateur and classically trained singers. Journal of Voice 22(1),
75–89.
Jokel, R., De Nil, L. F., & Sharpe, A. K. (2007). Speech disfluencies in adults with neurogenic
stuttering associated with stroke and traumatic brain injury. Journal of Medical Speech-Language
Pathology 15(3), 243–261.
Keith, R., & Aronson, A. (1975). Singing as therapy for apraxia of speech and aphasia: Report of a
case. Brain & Language 2, 483–488.
Kim, M., & Tomaino, C. (2008). Protocol evaluation for effective music therapy for persons with
nonfluent aphasia. Topics in Stroke Rehabilitation 15(6), 555–569.
Kim, S. J., & Jo, U. (2013). Study of accent-based music speech protocol development for improving
voice problems in stroke patients with mixed dysarthria. Neurorehabilitation 32(1), 185–190.
Kostyk, B. E., & Putnam Rochet, A. (1998). Laryngeal airway resistance in teachers with vocal
fatigue: A preliminary study. Journal of Voice 12(3), 287–299.
Kotz, S. A., Frisch, S., Von Cramon, D. Y., & Friederici, A. D. (2003). Syntactic language
processing: ERP lesion data on the role of the basal ganglia. Journal of the International
Neuropsychological Society 9(7), 1053–1060.
Kotz, S. A., & Gunter, T. C. (2015). Can rhythmic auditory cuing remediate language-related deficits
in Parkinson’s disease? Annals of the New York Academy of Sciences 1337, 62–68.
Kotz, S. A., & Schmidt-Kassow, M. (2015). Basal ganglia contribution to rule expectancy and
temporal predictability in speech. Cortex 68, 48–60.
Kotz, S. A., & Schwartze, M. (2010). Cortical speech processing unplugged: A timely subcortico-
cortical framework. Trends in Cognitive Sciences 14(9), 392–399.
Kotz, S. A., & Schwartze, M. (2016). Motor timing and sequencing in speech production: A general-
purpose framework. In G. Hickok & S. L. Small (Eds.), Neurobiology of Language (pp. 717–724).
New York: Academic Press.
Kotz, S. A., Schwartze, M., & Schmidt-Kassow, M. (2009). Non-motor basal ganglia functions: A
review and proposal for a model of sensory predictability in auditory language perception. Cortex
45(8), 982–990.
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills.
Nature Reviews Neuroscience 11(8), 599–605.
Krauss, T., & Galloway, H. (1982). Melodic intonation therapy with language delayed apraxic
children. Journal of Music Therapy 19(2), 102–113.
Large, E. W., Herrera, J. A., & Velasco, M. J. (2015). Neural networks for beat perception in musical
rhythm. Frontiers in Systems Neuroscience 9. Retrieved from
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4658578/
Large, E. W., & Jones, M. R. (1999). The dynamics of attending: How people track time-varying
events. Psychological Review 106(1), 119–159.
Lebrun, Y. (1998). Clinical observations and experimental research on the study of stuttering.
Journal of Fluency Disorders 23(2), 119–122.
Lim, K. B., Kim, Y. K., Lee, H. J., Yoo, J., Hwang, J. Y., Kim, J. A., & Kim, S. K. (2013). The
therapeutic effect of neurologic music therapy and speech language therapy in post-stroke aphasic
patients. Annals of Rehabilitation Medicine 37(4), 556–562.
Limb, C. J., & Roy, A. T. (2014). Technological, biological, and acoustical constraints to music
perception in cochlear implant users. Hearing Research 308(Suppl. C), 13–26.
Limb, C. J., & Rubinstein, J. T. (2012). Current research on music perception in cochlear implant
users. Otolaryngologic Clinics of North America, Cochlear Implants: Adult and Pediatric 45(1),
129–140.
Lord, V. M., Hume, V. J., Kelly, J. L., Cave, P., Silver, J., Waldman, M., & Hopkinson, N. S. (2012).
Singing classes for chronic obstructive pulmonary disease: A randomized controlled trial. BMC
Pulmonary Medicine 12(1), 1–7.
Lortie, C. L., Rivard, J., Thibeault, M., & Tremblay, P. (2017). The moderating effect of frequent
singing on voice aging. Journal of Voice 31(1), 112.e1–112.e12.
Mainka, S., & Mallien, G. (2014). Rhythmic speech cueing (RSC). In M. H. Thaut & V. Hoemberg
(Eds.), Handbook of neurologic music therapy (pp. 150–160). Oxford: Oxford University Press.
Martikainen, A., & Korpilahti, P. (2011). Intervention for childhood apraxia of speech: A single-case
study. Child Language Teaching and Therapy 27(1), 9–20.
Mauszycki, S. C., & Wambaugh, J. L. (2008). The effects of rate control treatment on consonant
production accuracy in mild apraxia of speech. Aphasiology 22(7–8), 906–920.
Naeser, M. A., & Helm-Estabrooks, N. A. (1985). CT scan lesion localization and response to
melodic intonation therapy with nonfluent aphasia cases. Cortex 21(2), 203–223.
Natke, U., Donath, T. M., & Kalveram, K. T. (2003). Control of voice fundamental frequency in
speaking versus singing. Journal of the Acoustical Society of America 113(3), 1587–1593.
Ogawa, M., Hosokawa, K., Yoshida, M., Yoshii, T., Shiromoto, O., & Inohara, H. (2013). Immediate
effectiveness of humming on the supraglottic compression in subjects with muscle tension
dysphonia. Folia Phoniatrica et Logopaedica 65(3), 123–128.
Onofre, F., Ricz, H. M. A., Takeshita-Monaretti, T., Prado, M. Y. D. A., & Aguiar-Ricz, L. (2013).
Effect of singing training on total laryngectomees wearing a tracheoesophageal voice prosthesis.
Acta cirúrgica brasileira/Sociedade Brasileira para Desenvolvimento Pesquisa em Cirurgia 28,
119–125.
Petersen, B., Weed, E., Sandmann, P., Brattico, E., Hansen, M., Sørensen, S. D., & Vuust, P. (2015).
Brain responses to musical feature changes in adolescent cochlear implant users. Frontiers in
Human Neuroscience 9. Retrieved from
https://www.frontiersin.org/articles/10.3389/fnhum.2015.00007/full
Popovici, M. (1995). Melodic intonation therapy in the verbal decoding of aphasics. Revue Roumaine
de Neurologie et Psychiatrie 33, 57–97.
Przybylski, L., Bedoin, N., Krifi-Papoz, S., Herbillon, V., Roch, D., Léculier, L., … Tillmann, B.
(2013). Rhythmic auditory stimulation influences syntactic processing in children with
developmental language disorders. Neuropsychology 27(1), 121–131.
Rocca, C. (2012). A different musical perspective: Improving outcomes in music through
habilitation, education, and training for children with cochlear implants. Seminars in Hearing
33(4), 425–433.
Rochette, F., Moussard, A., & Bigand, E. (2014). Music lessons improve auditory perceptual and
cognitive performance in deaf children. Frontiers in Human Neuroscience 8, 488. Retrieved from
https://doi.org/10.3389/fnhum.2014.00488
Rosenberger, P. B. (1980). Dopaminergic systems and speech fluency. Journal of Fluency Disorders
5, 255–267.
Rowe, D. C., Van den Oord, E. J., Stever, C., Giedinhagen, L. N., Gard, J. M., Cleveland, H. H., …
Waldman, I. D. (1999). The DRD2 TaqI polymorphism and symptoms of attention deficit
hyperactivity disorder. Molecular Psychiatry 4(6), 580–586.
Roy, N., Weinrich, B., Grey, S. D., Tanner, K., Stemple, J. C., & Sapienza, C. M. (2003). Three
treatments for teachers with voice disorders: A randomised clinical trial. Journal of Speech,
Language, and Hearing Research 46, 670–688.
Saliba, J., Bortfeld, H., Levitin, D. J., & Oghalai, J. S. (2016). Functional near-infrared spectroscopy
for neuroimaging in cochlear implant recipients. Hearing Research 338(Suppl. C), 64–75.
Santoni, C., de Boer, G., Thaut, M., & Bressmann, T. (2018). Influence of altered auditory feedback
on oral-nasal balance in song. Journal of Voice. Manuscript in print.
Sauder, C., Roy, N., Tanner, K., Houtz, D. R., & Smith, M. E. (2010). Vocal function exercises for
presbylaryngis: A multidimensional assessment of treatment outcomes. Annals of Otology,
Rhinology, & Laryngology 119(7), 460–467.
Schlaug, G., Marchina, S., & Norton, A. (2008). From singing to speaking: Why singing may lead to
recovery of expressive language function in patients with Broca’s aphasia. Music Perception 25(4),
315–323.
Schlaug, G., Marchina, S., & Norton, A. (2009). Evidence for plasticity in white matter tracts of
chronic aphasic patients undergoing intense intonation-based speech therapy. New York Academy
of Sciences 1169, 385–94.
Seger, C. A., Spiering, B. J., Sares, A. G., Quraini, S. I., Alpeter, C., David, J., & Thaut, M. H.
(2013). Corticostriatal contributions to musical expectancy perception. Journal of Cognitive
Neuroscience 25(7), 1062–77.
Skingley, A., Page, S., Clift, S., Morrison, I., Coulton, S., Treadwell, P., … Shipton, M. (2014).
Singing for breathing: Participants’ perceptions of a group singing programme for people with
COPD. Arts & Health 6(1), 59–74.
Skodda, S., & Schlegel, U. (2008). Speech rate and rhythm in Parkinson’s disease. Movement
Disorders: Official Journal of the Movement Disorder Society 23(7), 985–992.
Smith, S., & Thyme, K. (1976). Statistic research on changes in speech due to pedagogic treatment
(the accent method). Folia Phoniatrica 28, 98–103.
Sparks, R. W., Helm, N., & Albert, M. (1974). Aphasia rehabilitation resulting from melodic
intonation therapy. Cortex 10(4), 303–316.
Sparks, R. W., & Holland, A. L. (1976). Method: Melodic intonation therapy for aphasia. Journal of
Speech and Hearing Disorders 41, 298–300
Stahl, B., Kotz, S. A., Henseler, I., Turner, R., & Geyer, F. (2011). Rhythm in disguise: Why singing
may not hold the key to recovery from aphasia. Brain 134(10), 3083–3093.
Stegemöller, E. L., Radig, H., Hibbing, P., Wingate, J., & Sapienza, C. (2017). Effects of singing on
voice, respiratory control and quality of life in persons with Parkinson’s disease. Disability and
Rehabilitation 39(6), 594–600.
Stemple, J. C., Glaze, L. E., & Klaben, B. G. (2009). Clinical voice pathology: Theory and
management (4th ed.). San Diego, CA: Singular Publishing Group.
Stemple, J. C., Lee, L., D’Amico, B., & Pickup, B. (1994). Efficacy of vocal function exercises as a
method of improving voice production. Journal of Voice 8(3), 271–278.
Stephan, K. M., Thaut, M. H., Wunderlich, G., Schicks, W., Tian, B., Tellmann, L., … Hömberg, V.
(2002). Conscious and subconscious sensorimotor synchronization: Cortex and the influence of
awareness. NeuroImage 15(2), 345–352.
Sundberg, J., Birch, P., Gümoes, B., Stavad, H., Prytz, S., & Karle, A. (2007). Experimental findings
on the nasal tract resonator in singing. Journal of Voice 21(2), 127–137.
Tamplin, J. (2008). A pilot study into the effect of vocal exercises and singing on dysarthric speech.
Neurorehabilitation 23(3), 207–216.
Tamplin, J., Baker, F. A., Grocke, D., Brazzale, D. J., Pretto, J. J., Ruehland, W. R., … Berlowitz, D.
J. (2013). Effect of singing on respiratory function, voice, and mood after quadriplegia: a
randomized controlled trial. Archives of Physical Medicine and Rehabilitation 94(3), 426–434.
Tanner, K., Roy, N., Merrill, R. M., & Power, D. (2005). Velopharyngeal port status during classical
singing. Journal of Speech, Language, and Hearing Research 48(6), 1311–1324.
Tanner, M. A. (2012). Voice improvement in Parkinson’s disease: Vocal pedagogy and voice therapy
combined (Doctoral dissertation). University of Alberta.
Tanner, M. A., Rammage, L., & Liu, L. (2016). Does singing and vocal strengthening improve vocal
ability in people with Parkinson’s disease? Arts & Health 8(3), 199–212.
Tautscher-Basnett, A., Tomantschger, V., Keglevic, S., & Freimuller, M. (2006). Group therapy for
individuals with Parkinson disease (PD) focusing on voice strengthening. Presentation to the 4th
World Congress for Neurorehabilitation.
Thaut, M. H., Thaut, C. P., & McIntosh, K. (2014). Melodic intonation therapy (MIT). In M. H.
Thaut & V. Hoemberg (Eds.), Handbook of neurologic music therapy (pp. 140–145). Oxford:
Oxford University Press.
Thaut, M. H., & Hoemberg, V. (Eds.). (2014). Handbook of neurologic music therapy. Oxford:
Oxford University Press.
The American Speech-Language-Hearing Association (2017). Spasmodic dysphonia (Website).
https://www.asha.org/public/speech/disorders/Spasmodic-Dysphonia
Thomson, J. M., Leong, V., & Goswami, U. (2012). Auditory processing interventions and
developmental dyslexia: A comparison of phonemic and rhythmic approaches. Reading and
Writing 26(2), 139–161.
Titze, I. R. (2006). Voice training and therapy with a semi-occluded vocal tract: Rationale and
scientific underpinnings. Journal of Speech, Language, and Hearing Research 49(2), 448–459.
Tonkinson, S. (1994). The Lombard effect in choral singing. Journal of Voice 8(1), 24–29.
Tourville, J. A., & Guenther, F. H. (2011). The DIVA model: A neural theory of speech acquisition
and production. Language and Cognitive Processes 26(7), 952–981.
Trainor, L. J., Clark, E. D., Huntley, A., & Adams, B. (1997). The acoustic basis of infant preferences
for infant-directed singing. Infant Behavior and Development 20(3), 383–396.
Van der Meulen, I., Van de Sandt-Koenderman, M. E., & Ribbers, G. M. (2012). Melodic intonation
therapy: present controversies and future opportunities. Archives of Physical Medicine and
Rehabilitation 93(1), S46–52.
Van Riper, C. (1982). The nature of stuttering (2nd ed.). Englewood Cliffs, NJ: Prentice Hall.
Verdolini, K., Druker D. G., Palmer, P. M., & Samawi, H. (1998). Laryngeal adduction in resonant
voice. Journal of Voice 12(3), 315–327.
Verdolini-Marston, K., Burke, M. K., Lessac, A., Glaze, L., & Caldwell, E. (1995). Preliminary study
of two methods of treatment for laryngeal nodules. Journal of Voice 9(1), 74–85.
Victor, M., & Ropper, A. H. (2001). Adams and Victor’s principles of neurology (7th ed.). New York:
McGraw-Hill.
Wambaugh, J. L., & Martinez, A. L. (2000). Effects of rate and rhythm control treatment on
consonant production accuracy in apraxia of speech. Aphasiology 14(8), 851–871.
Wambaugh, J. L., Nessler, C., Cameron, R., & Mauszycki, S. C. (2012). Acquired apraxia of speech:
The effects of repeated practice and rate/rhythm control treatments on sound production accuracy.
American Journal of Speech-Language Pathology 21(2), S5–S27.
Wan, C. Y., Zheng, X., Marchina, S., Norton, A., & Schlaug, G. (2014). Intensive therapy induces
contralateral white matter changes in chronic stroke patients with Broca’s aphasia. Brain &
Language 136(Suppl. C), 1–7.
Wiener, M., Lohoff, F. W., & Coslett, H. B. (2011). Double dissociation of dopamine genes and
timing in humans. Journal of Cognitive Neuroscience 23(10), 2811–2821.
Wiener, M., Lee, Y.-S., Lohoff, F. W., & Coslett, H. B. (2014). Individual differences in the
morphometry and activation of time perception networks are influenced by dopamine genotype.
NeuroImage 89, 10–22.
Wong, P. C. M., Ettlinger, M., & Zheng, J. (2013). Linguistic grammar learning and DRD2-TAQ-IA
polymorphism. PLoS ONE 8(5), e64983.
Wu, J. C., Maguire, G., Riley, G., Fallon, J., LaCasse, L., Chin, S., … Lottenberg, S. (1995). A
positron emission tomography [18F] deoxyglucose study of developmental stuttering. Neuroreport
6(3), 501–505.
Yanagisawa, E., Estill, J., Mambrino, L., & Talkin, D. (1991). Supraglottic contributions to pitch
raising: Videoendoscopic study with spectroanalysis. Annals of Otology, Rhinology, &
Laryngology 100(1), 19–30.
Yiu, E. M. L., & Ho, E. Y. Y. (2002). Short-term effect of humming on vocal quality. Asia Pacific
Journal of Speech, Language and Hearing 7(3), 123–137.
Yorkston, K. M., Beukelman, D. R., Strand, E. A., & Hakel, M. (2010). Management of motor speech
disorders in children and adults. Austin, TX: Pro-Ed Inc.
Ziegler, A., & Hapner, E. R. (2013). Phonation resistance training exercise (PhoRTE) therapy. In A.
Behrman & J. Haskell (Eds.), Exercises for voice therapy (pp. 147–148). San Diego, CA: Plural
Publishing.
Ziegler, A., Verdolini Abbott, K., Johns, M., Klein, A., & Hapner, E. R. (2014). Preliminary data on
two voice therapy interventions in the treatment of presbyphonia. The Laryngoscope 124(8),
1869–1876.
Zumbansen, A., Peretz, I., & Hébert, S. (2014). Melodic intonation therapy: Back to basics for future
research. Frontiers in Neurology 5, 1–11. Retrieved from https://doi.org/10.3389/fneur.2014.00007
CHAPT E R 30
NEUROLOGIC MUSIC
T H E R A P Y TA R G E T I N G
COGNITIVE AND
AFFECTIVE FUNCTIONS
S H A N TA L A H E G D E
C A F :
C F R
C R : T
M I C
F
M B P
M T : F S S
M N M
Addressing cognitive dysfunction via music is a recent frontier in music
therapy and has emerged as one of the most promising and innovative new
methods. Considering music as a form of therapy has a long history and has
been practiced across cultures over several centuries, dating back even to
prehistoric times (Thaut, 2015). Music has always been considered to
facilitate “overall well-being” and the “feel-good” factor by enhancing
overall emotional health (de l’Etoile, 2010). Music therapy has been
considered effective in the reduction of anxiety, depression, and agitation.
However, often the process underlying the mechanisms of change was
unclear. The changes were explained relying upon the popular and
prevailing psychological theories such as behavioral, psychoanalytic, and
humanistic schools of thought (de l’Etoile, 2010, 2016). Such explanations
have had limited contributions in understanding the underlying processes of
music in therapy which is crucial in standardizing techniques to target
specific functioning as well as in evaluating the specific outcome. With
advancement in neuroscientific investigations on music perception and
cognition and better understanding of the underlying neural correlates, there
has been a major change in music therapy principles and objectives. Music
therapy has shifted from its perspective as a social science model to a
neuroscience-based model (Thaut, 2005a; Thaut, McIntosh, & Hoemberg,
2014). An increasing number of controlled trials have examined the effects
of music-based interventions such as listening, singing, and actively
engaging in music by playing an instrument in neurorehabilitation
(Sihvonen et al., 2017).
A O N M
T T E
M
The recent advancement in the field of music therapy has been the
development of the neuroscience-based approach to music therapy practice
and research. “Neurologic Music Therapy” (NMT) is based on
neuroscientific modes of music perception, cognition, and production. NMT
consists of twenty standardized techniques targeting three main domains of
functioning, namely, sensorimotor, language, and cognitive-affective
dysfunctions that are the result of neurologic disease of the human nervous
system (Clair, Pasiali, & Lagasse, 2008; de l’Etoile, 2010; Thaut, 2005a).
Each of the techniques focuses on the influence of music on non-musical
domains of functioning with non-musical therapeutic goals.
NMT is based on two interrelated models to explain how music-based
interventions and training can access and modulate cognitive functions in a
neuropsychological rehabilitation context. The two models of the NMT are
the “Rational Scientific Mediating Model” (RSMM) and the
“Transformational Design Model” (TDM) (de l’Etoile, 2016; Hegde, 2014;
Thaut, McIntosh, & Hoemberg, 2014). The TDM is the clinical component
or practical extension of the scientific theory model of the RSMM.
There are three major areas which the various techniques of NMT
address:
The first two areas are covered in other chapters of this book. This
chapter will focus on various techniques targeting cognitive and affective
functions.
Enhanced neuroscientific understanding of music perception and
cognition along with advancement in non-invasive research tools to study
human brain functions over the last two to three decades have contributed
to linking music and cognitive functions as well examining the shared and
unique neural correlates underlying music and non-musical cognitive
processes. Until recently, lack of standardized methods was a major
drawback in the field of music therapy.
CR via NMT-based techniques is one of the recent developments in the
area of NMT in comparison to clinical research in the domain of
sensorimotor and speech and language functions (Gardiner & Horwitz,
2015; Thaut, 2010). Careful analysis of the literature indicates only a
handful of studies with a strong theoretical background examining the
effects of traditional music therapy on cognitive functions. NMT
techniques, developed on the basis of RSMM and TDM, are evidence-
based, theoretically grounded, and standardized in terminology and
methods. The six steps of the TDM, which is the clinical application model,
run parallel to the fundamental principles of CR. This comparison is
presented in tabular format in Table 1.
Table 1. A comparison of basic principles of cognitive remediation and clinical
application model of (TDM) of NMT
Key references for principles of CR: Ben-Yishay & Prigatano, 1990; Eack, 2012;
Prigatano, 1997; Raskin, 2010.
T NMT T
C E
S F D
R
Ben-Yishay, Y., Piasetsky, E., & Rattok, J. (1985). A systematic method for ameliorating disorders in
basic attention. In M. J. Meir, A. L. Benton, & L. Diller (Eds.), Neuropsychological rehabilitation
(pp. 165–82). New York: Guilford Press.
Ben-Yishay, Y., & Prigatano, G. P. (1990). Cognitive remediation. In M. Rosenthal, M. R. E. R.
Griffith, M. R. Bond, & J. D. Miller (Eds.), Rehabilitation of the adult and child with traumatic
brain injury (2nd ed., pp. 393–409). Philadelphia, PA: Davis.
Benedict, R. H., Harris, A. E., Markow, T., McCormick, J. A., Nuechterlein, K. H., & Asarnow, R. F.
(1994). Effects of attention training on information processing in schizophrenia. Schizophrenia
Bulletin 20, 537–546.
Bernardi, N. F., Cioffi, M. C., Ronchi, R., Maravita, A., Bricolo, E., Zigiotto, L., … Vallar, G. (2017).
Improving left spatial neglect through music scale playing. Journal of Neuropsychology 11(1),
135–158.
Biswas, A., Hegde, S., Jhunjhunwala, K., & Pal, P. K. (2016). Two sides of the same coin:
Impairment in perception of temporal components of rhythm and cognitive functions in
Parkinson’s disease. Basal Ganglia 6(1), 63–70.
Blood, A. J., & Zatorre, R. J. (2001). Intensely pleasurable responses to music correlate with activity
in brain regions implicated in reward and emotion. Proceedings of the National Academy of
Sciences 98(20), 11818–11823.
Bodak, R., Malhotra, P., Bernardi, N. F., Cocchini, G., & Stewart, L. (2014). Reducing chronic visuo-
spatial neglect following right hemisphere stroke through instrument playing. Frontiers in Human
Neuroscience 8, 413. Retrieved from https://doi.org/10.3389/fnhum.2014.00413
Bruel-Jungerman, E., Davis, S., & Laroche, S. (2007). Brain plasticity mechanisms and memory: A
party of four. Neuroscientist 13(5), 492–505.
Cappa, S. F., Benke, T., Clarke, S., Rossi, B., Stemmer, B., & Van Heugten, C. M. (2005). EFNS
guidelines on cognitive rehabilitation: Report of an EFNS task force. European Journal of
Neurology 12(9), 665–680.
Castro, M., Tillmann, B., Luaute, J., Corneyllie, A., Dailler, F., Andre-Obadia, N., & Perrin, F.
(2015). Boosting cognition with music in patients with disorders of consciousness.
Neurorehabilitation and Neural Repair 29(8), 734–742.
Cernich, A. N., Kurtz, S. M., Mordecai, K. L., & Ryan, P. B. (2010). Cognitive rehabilitation in
traumatic brain injury. Current Treatment Options in Neurology 12(5), 412–423.
Chanda, M. L., & Levitin, D. J. (2013). The neurochemistry of music. Trends in Cognitive Sciences
17(4), 179–193.
Chung, C. S., Pollock, A., Campbell, T., Durward, B. R., & Hagen, S. (2013). Cognitive
rehabilitation for executive dysfunction in adults with stroke or other adult non-progressive
acquired brain damage. Cochrane Database of Systematic Reviews 4, CD008391.
Cicerone, K. D. (2012). Facts, theories, values: Shaping the course of neurorehabilitation. The 60th
John Stanley Coulter memorial lecture. Archives of Physical Medicine and Rehabilitation 93(2),
188–191.
Cicerone, K. D., Dahlberg, C., Kalmar, K., Langenbahn, D. M., Malec, J. F., Bergquist, T. F., …
Morse, P. A. (2000). Evidence-based cognitive rehabilitation: Recommendations for clinical
practice. Archives of Physical Medicine and Rehabilitation 81(12), 1596–1615.
Cicerone, K. D., Langenbahn, D. M., Braden, C., Malec, J. F., Kalmar, K., Fraas, M., … Ashman, T.
(2011). Evidence-based cognitive rehabilitation: Updated review of the literature from 2003
through 2008. Archives of Physical Medicine and Rehabilitation 92(4), 519–530.
Clair, A. A., Pasiali, V., & Lagasse, B. (2008). Neurologic music therapy. In A. A. Darrow (Ed.),
Introduction to approaches in music therapy (2nd ed., pp. 153–171). Silver Spring, MD: American
Music Therapy Association.
Conway, C. M., Pisoni, D. B., & Kronenberger, W. G. (2009). The importance of sound for cognitive
sequencing abilities: The auditory scaffolding hypothesis. Current Directions in Psychological
Science 18(5), 275–279.
de Diego-Balaguer, R., Martinez-Alvarez, A., & Pons, F. (2016). Temporal attention as a scaffold for
language development. Frontiers in Psychology 7. Retrieved from
https://doi.org/10.3389/fpsyg.2016.00044
de l’Etoile, S. K. (2010). Neurologic music therapy: A scientific paradigm for clinical practice. Music
and Medicine 2(2), 78–84.
de l’Etoile, S. K. (2016). Processes of music therapy: Clinical and scientific rationales and models. In
S. Hallam, I. Cross, & M. Thaut (Eds.), The Oxford handbook of music psychology (2nd ed., pp.
805–818). Oxford: Oxford University Press.
Deutsch, D. (1982). Organizational processes in music. In M. Clynes (Ed.), Music, mind and brain
(pp. 119–131). New York: Plenum Press.
Diamond, P. T., Felsenthal, G., Macciocchi, S. N., Butler, D. H., & Lally-Cassady, D. (1996). Effect
of cognitive impairment on rehabilitation outcome. American Journal of Physical Medicine and
Rehabilitation 75(1), 40–43.
Diller, L., & Gordon, W. A. (1981). Interventions for cognitive deficits in brain-injured adults.
Journal of Consulting and Clinical Psychology 49(6), 822–834.
Drake, C., Jones, M. R., & Baruch, C. (2000). The development of rhythmic attending in auditory
sequences: Attunement, referent period, focal attending. Cognition 77(3), 251–288.
Eack, S. M. (2012). Cognitive remediation: A new generation of psychosocial interventions for
people with schizophrenia. Social Work 57(3), 235–246.
Elvevag, B., & Goldberg, T. E. (2000). Cognitive impairment in schizophrenia is the core of the
disorder. Critical Reviews in Neurobiology 14(1), 1–21.
Faralli, A., Bigoni, M., Mauro, A., Rossi, F., & Carulli, D. (2013). Noninvasive strategies to promote
functional recovery after stroke. Neural Plasticity 2013, 854597.
Fischer, C., Luaute, J., Adeleine, P., & Morlet, D. (2004). Predictive value of sensory and cognitive
evoked potentials for awakening from coma. Neurology 63(4), 669–673.
Fujii, S., & Wan, C. Y. (2014). The role of rhythm in speech and language rehabilitation: The SEP
hypothesis. Frontiers in Human Neuroscience 8. Retrieved from
http://dx.doi.org/10.3389/fnhum.2014.00777
Gardiner, J. C., & Horwitz, J. L. (2015). Neurologic music therapy and group psychotherapy for
treatment of traumatic brain injury: Evaluation of a cognitive rehabilitation group. Music Therapy
Perspectives 33(2), 193–201.
Gfeller, K. E. (1983). Musical mnemonics as an aid to retention with normal and learning disabled
students. Journal of Music Therapy 20(4), 179–189.
Gordon, W. A., Zafonte, R., Cicerone, K., Cantor, J., Brown, M., Lombard, L., … Chandna, T.
(2006). Traumatic brain injury rehabilitation: State of the science. American Journal of Physical
Medicine and Rehabilitation 85(4), 343–382.
Guilbert, A., Clement, S., & Moroni, C. (2017). A rehabilitation program based on music practice for
patients with unilateral spatial neglect: A single-case study. Neurocase 23(1), 12–21.
Harvey, P. D., Green, M. F., Keefe, R. S., & Velligan, D. I. (2004). Cognitive functioning in
schizophrenia: A consensus statement on its role in the definition and evaluation of effective
treatments for the illness. Journal of Clinical Psychiatry 65(3), 361–372.
Hegde, S. (2014). Music based cognitive remediation therapy for patients with traumatic brain injury.
Frontiers in Neurology 5. Retrieved from https://doi.org/10.3389/fneur.2014.00034
Hegde, S. (2017). A review of Indian research on cognitive remediation for schizophrenia. Asian
Journal of Psychiatry 25, 54–59.
Herholz, S. C., & Zatorre, R. J. (2012). Musical training as a framework for brain plasticity:
Behavior, function, and structure. Neuron 76(3), 486–502.
Hitch, G. J. (1996). Temporal grouping effects in immediate recall: A working memory analysis.
Quarterly Journal of Experimental Psychology Section A 49(1), 116–139.
Ho, Y. C., Cheung, M. C., & Chan, A. S. (2003). Music training improves verbal but not visual
memory: Cross-sectional and longitudinal explorations in children. Neuropsychology 17(3), 439–
450.
Hommel, M., Peres, B., Pollak, P., Memin, B., Besson, G., Gaio, J. M., & Perret, J. (1990). Effects of
passive tactile and auditory stimuli on left visual neglect. Archives of Neurology 47, 573–576.
Ishihara, M., Revol, P., Jacquin-Courtois, S., Mayet, R., Rode, G., Boisson, D., … Rossetti, Y.
(2013). Tonal cues modulate line bisection performance: Preliminary evidence for a new
rehabilitation prospect? Frontiers in Psychology 4, 704. Retrieved from
https://doi.org/10.3389/fpsyg.2013.00704
Jacobsen, J. H., Stelzer, J., Fritz, T. H., Chetelat, G., La Joie, R., & Turner, R. (2015). Why musical
memory can be preserved in advanced Alzheimer’s disease. Brain 138(8), 2438–2450.
Jakobson, L. S., Cuddy, L. L., & Kilgour, A. R. (2003). Time tagging: A key to musicians’ superior
memory. Music Perception: An Interdisciplinary Journal 20(3), 307–313.
Jäncke, L. (2009). Music drives brain plasticity. F1000 Biology Reports 1, 78. Retrieved from
http://www.F1000.com/Reports/Biology/content/1/78
Jones, M. R. (1976). Time, our lost dimension: Toward a new theory of perception, attention, and
memory. Psychological Review 83(5), 323–355.
Jones, M. R., & Boltz, M. (1989). Dynamic attending and responses to time. Psychological Review
96(3), 459–491.
Jones, M. R., Boltz, M., & Kidd, G. (1982). Controlled attending as a function of melodic and
temporal context. Perception & Psychophysics 32(3), 211–218.
Jones, M. R., Moynihan, H., Mackenzie, N., & Puente, J. (2002). Temporal aspects of stimulus-
driven attending in dynamic arrays. Psychological Science 13(4), 313–319.
Jung, H., Sontag, S., Park, Y. S., & Loui, P. (2015). Rhythmic effects of syntax processing in music
and language. Frontiers in Psychology 6, 1762. Retrieved from
https://doi.org/10.3389/fpsyg.2015.01762
Kern, P., Wolery, M., & Aldridge, D. (2007). Use of songs to promote independence in morning
greeting routines for young children with autism. Journal of Autism and Developmental Disorders
37(7), 1264–1271.
Keshavan, M. S., Vinogradov, S., Rumsey, J., Sherrill, J., & Wagner, A. (2014). Cognitive training in
mental disorders: Update and future directions. American Journal of Psychiatry 171(5), 510–522.
Kilgour, A. R., Jakobson, L. S., & Cuddy, L. L. (2000). Music training and rate of presentation as
mediators of text and song recall. Memory & Cognition 28(5), 700–710.
Kleim, J. A., Barbay, S., & Nudo, R. J. (1998). Functional reorganization of the rat motor cortex
following motor skill learning. Journal of Neurophysiology 80(6), 3321–3325.
Kleinstauber, M., & Gurr, B. (2006). Music in brain injury rehabilitation. Journal of Cognitive
Rehabilitation 24, 4–14.
Koelsch, S. (2009). A neuroscientific perspective on music therapy. Annals of the New York Academy
of Sciences 1169, 374–384.
Lam, R. W., Kennedy, S. H., McIntyre, R. S., & Khullar, A. (2014). Cognitive dysfunction in major
depressive disorder: Effects on psychosocial functioning and implications for treatment. Canadian
Journal of Psychiatry/Revue Canadienne de Psychiatrie 59(12), 649–654.
Large, E. W., & Jones, M. R. (1999). The dynamics of attending: How people track time-varying
events. Psychological Review 106(1), 119–159.
Levitin, D. J., & Tirovolas, A. K. (2009). Current advances in the cognitive neuroscience of music.
Annals of the New York Academy of Sciences 1156, 211–231.
Limb, C. J. (2006). Structural and functional neural correlates of music perception. The Anatomical
Record Part A: Discoveries in Molecular, Cellular, and Evolutionary Biology 288, 435–446.
McAuley, J. D., & Jones, M. R. (2003). Modeling effects of rhythmic context on perceived duration:
A comparison of interval and entrainment approaches to short-interval timing. Journal of
Experimental Psychology: Human Perception and Performance 29(6), 1102–1125.
Maratos, A. S., Gold, C., Wang, X., & Crawford, M. J. (2008). Music therapy for depression.
Cochrane Database of Systematic Reviews 1, CD004517.
Marder, S. R. (2006). Initiatives to promote the discovery of drugs to improve cognitive function in
severe mental illness. Journal of Clinical Psychiatry 67(7), e03.
Miller, J. E., Carlson, L. A., & McAuley, J. D. (2013). When what you hear influences when you see:
Listening to an auditory rhythm influences the temporal allocation of visual attention.
Psychological Science 24(1), 11–18.
Mossler, K., Chen, X., Heldal, T. O., & Gold, C. (2011). Music therapy for people with schizophrenia
and schizophrenia-like disorders. Cochrane Database of Systematic Reviews 4, CD004025.
Müller, T. (2002). Drug treatment of non-motor symptoms in Parkinson’s disease. Expert Opinion in
Pharmacotherapy 3(4), 381–388.
Müller, T. (2012). Drug therapy in patients with Parkinson’s disease. Translational
Neurodegeneration 1, 10. doi:10.1186/2047-9158-1-10
Münte, T. F., Altenmüller, E., & Jäncke, L. (2002). The musician’s brain as a model of
neuroplasticity. Nature Reviews Neuroscience 3, 473–478.
Nayak, S., Wheeler, B. L., Shiflett, S. C., & Agostinelli, S. (2000). Effect of music therapy on mood
and social interaction among individuals with acute traumatic brain injury and stroke.
Rehabilitation Psychology 45(3), 274–283.
Patel, A. D. (2010). Music, biological evolution, and the brain. In M. Bailar (Ed.), Emerging
disciplines (pp. 91–144). Houston, TX: Rice University Press.
Peretz, I. (2006). The nature of music from a biological perspective. Cognition 100(1), 1–32.
Peretz, I., & Zatorre, R. J. (2005). Brain organization for music processing. Annual Review of
Psychology 56, 89–114.
Peterson, D. A., & Thaut, M. H. (2007). Music increases frontal EEG coherence during verbal
learning. Neuroscience Letters 412(3), 217–221.
Podd, M. H. (2012). History of cognitive remediation. In M. H. Podd, Cognitive remediation for
brain injury and neurological illness (pp. 1–4). New York: Springer.
Prigatano, G. P. (1997). Learning from our successes and failures: Reflections and comments on
“Cognitive rehabilitation: How it is and how it might be.” Journal of the International
Neuropsychological Society 3(5), 497–499.
Raglio, A., Attardo, L., Gontero, G., Rollino, S., Groppo, E., & Granieri, E. (2015). Effects of music
and music therapy on mood in neurological patients. World Journal of Psychiatry 5(1), 68–78.
Raskin, S. A. (2010). Current approaches to cognitive rehabilitation. In C. Armstrong & L. Morrow
(Eds.), Handbook of medical neuropsychology (pp. 505–518). New York: Springer.
Rohling, M. L., Faust, M. E., Beverly, B., & Demakis, G. (2009). Effectiveness of cognitive
rehabilitation following acquired brain injury: A meta-analytic re-examination of Cicerone et al.’s
(2000, 2005) systematic reviews. Neuropsychology 23(1), 20–39.
Rund, B. R., & Borg, N. E. (1999). Cognitive deficits and cognitive training in schizophrenic
patients: A review. Acta Psychiatrica Scandinavica 100(2), 85–95.
Salimpoor, V. N., Zald, D. H., Zatorre, R. J., Dagher, A., & McIntosh, A. R. (2015). Predictions and
the brain: How musical sounds become rewarding. Trends in Cognitive Sciences 19(2), 86–91.
Särkämö, T., Altenmüller, E., Rodríguez-Fornells, A., & Peretz, I. (2016). Editorial: Music, brain,
and rehabilitation: Emerging therapeutic applications and potential neural mechanisms. Frontiers
in Human Neuroscience 10, 103. Retrieved from https://doi.org/10.3389/fnhum.2016.00103
Särkämö, T., Ripollés, P., Vepsäläinen, H., Autti, T., Silvennoinen, H. M., Salli, E., … Rodríguez-
Fornells, A. (2014). Structural changes induced by daily music listening in the recovering brain
after middle cerebral artery stroke: A voxel-based morphometry study. Frontiers in Human
Neuroscience 8, 245. Retrieved from https://doi.org/10.3389/fnhum.2014.00245
Särkämö, T., Tervaniemi, M., & Huotilainen, M. (2013). Music perception and cognition:
Development, neural basis, and rehabilitative use of music. Wiley Interdisciplinary Reviews:
Cognitive Science 4(4), 441–451.
Särkämö, T., Tervaniemi, M., Laitinen, S., Forsblom, A., Soinila, S., Mikkonen, M., … Hietanen, M.
(2008). Music listening enhances cognitive recovery and mood after middle cerebral artery stroke.
Brain 131(3), 866–876.
Särkämö, T., Tervaniemi, M., Laitinen, S., Numminen, A., Kurki, M., Johnson, J. K., & Rantanen, P.
(2014). Cognitive, emotional, and social benefits of regular musical activities in early dementia:
Randomized controlled study. Gerontologist 54(4), 634–650.
Sarnthein, J., Vonstein, A., Rappelsberger, P., Petsche, H., Rauscher, F. H., & Shaw, G. L. (1997).
Persistent patterns of brain activity: An EEG coherence study of the positive effect of music on
spatial-temporal reasoning. Neurological Research 19(2), 107–116.
Schlaug, G. (2009). Part VI introduction: Listening to and making music facilitates brain recovery
processes. Annals of the New York Academy of Sciences 1169, 372–373.
Schutz, L. E., & Trainor, K. (2007). Evaluation of cognitive rehabilitation as a treatment paradigm.
Brain Injury 21(6), 545–557.
Shaw, G. L., & Bodner, M. (1999). Music enhances spatial-temporal reasoning: Towards a
neurophysiological basis using EEG. Clinical Electroencephalography 30(4), 151–155.
Sihvonen, A. J., Särkämö, T., Leo, V., Tervaniemi, M., Altenmüller, E., & Soinila, S. (2017). Music-
based interventions in neurological rehabilitation. Lancet Neurology 16(8) 648–660.
Simmons-Stern, N. R., Budson, A. E., & Ally, B. A. (2010). Music as a memory enhancer in patients
with Alzheimer’s disease. Neuropsychologia 48(10), 3164–3167.
Sohlberg, M. M., & Mateer, C. A. (2001). Cognitive rehabilitation: An integrative
neuropsychological approach. New York: Guilford Press.
Spring, B. J., & Ravdin, L. (1992). Cognitive remediation in schizophrenia: Should we attempt it?
Schizophrenia Bulletin 18(1), 15–20.
Sun, J.-H., Tan, L., & Yu, J.-T. (2014). Post-stroke cognitive impairment: Epidemiology, mechanisms
and management. Annals of Translational Medicine 2(8), 80.
Sutoo, D., & Akiyama, K. (2004). Music improves dopaminergic neurotransmission: Demonstration
based on the effect of music on blood pressure regulation. Brain Research 1016(2), 255–262.
Thaut, M. H. (2005a). Rhythm, music and the brain: Scientific foundations and clinical applications.
New York: Routledge.
Thaut, M. H. (2005b). Neurologic music therapy in cognitive rehabilitation. In M. Thaut, Rhythm,
music and the brain: Scientific foundations and clinical applications (pp. 179–202). New York:
Routledge.
Thaut, M. H. (2005c). Neurologic music therapy in sensorimotor rehabilitation. In M. Thaut, Rhythm,
music and the brain: Scientific foundations and clinical applications (pp. 137–164). New York:
Routledge.
Thaut, M. H. (2005d). Neurologic music therapy in speech and language rehabilitation. In M. Thaut,
Rhythm, music and the brain: Scientific foundations and clinical applications (pp. 165–178). New
York: Routledge.
Thaut, M. H. (2010). Neurologic music therapy in cognitive rehabilitation. Music Perception 27(4),
281–285.
Thaut, M. H. (2014). Assessment and the transformational design model (TDM). In M. H. Thaut &
V. Hoemberg (Eds.), Handbook of neurologic music therapy (pp. 60–68). Oxford: Oxford
University Press.
Thaut, M. H. (2015). Music as therapy in early history. Progress in Brain Research 217, 143–158.
Thaut, M. H., Gardiner, J. C., Holmberg, D., Horwitz, J., Kent, L., Andrews, G., Donelan, B., &
McIntosh, G. R. (2009). Neurologic music therapy improves executive function and emotional
adjustment in traumatic brain injury rehabilitation. Annals of the New York Academy of Sciences
1169, 406–416.
Thaut, M. H., & Hoemberg, V. (Eds.). (2014). Handbook of neurologic music therapy. Oxford:
Oxford University Press.
Thaut, M. H., McIntosh, G. C., & Hoemberg, V. (2014). Neurologic music therapy: From social
science to neuroscience. In M. H. Thaut & V. Hoemberg (Eds.), Handbook of neurologic music
therapy (pp. 1–6). Oxford: Oxford University Press.
Thaut, M. H., Peterson, D. A., & McIntosh, G. C. (2005). Temporal entrainment of cognitive
functions: Musical mnemonics induce brain plasticity and oscillatory synchrony in neural
networks underlying memory. Annals of the New York Academy of Sciences 1060, 243–254.
Thaut, M. H., Peterson, D. A., McIntosh, G. C., & Hoemberg, V. (2014). Music mnemonics aid
verbal memory and induce learning: Related brain plasticity in multiple sclerosis. Frontiers in
Human Neuroscience 8, 395. Retrieved from https://doi.org/10.3389/fnhum.2014.00395
Thompson, R. G., Moulin, C. J., Hayre, S., & Jones, R. W. (2005). Music enhances category fluency
in healthy older adults and Alzheimer’s disease patients. Experimental Aging Research 31(1), 91–
99.
Volpe, B. T., & McDowell, F. H. (1990). The efficacy of cognitive rehabilitation in patients with
traumatic brain injury. Archives of Neurology 47, 220–222.
Wallace, W. T. (1994). Memory for music: Effect of melody on recall of text. Journal of
Experimental Psychology: Learning, Memory, & Cognition 20(6), 1471–1485.
Wan, C. Y., & Schlaug, G. (2010). Music making as a tool for promoting brain plasticity across the
life span. Neuroscientist 16(5), 566–577.
Wolfe, D. E., & Hom, C. (1993). Use of melodies as structural prompts for learning and retention of
sequential verbal information by preschool students. Journal of Music Therapy 30(2), 100–118.
Wykes, T., Huddy, V., Cellard, C., McGurk, S. R., & Czobor, P. (2011). A meta-analysis of cognitive
remediation for schizophrenia: Methodology and effect sizes. American Journal of Psychiatry
168(5), 472–485.
CHAPT E R 31
MUSICAL DISORDERS
I S A B E L L E R O YA L, S É B A S T I E N PA Q U E T T E, A N D
PA U L I N E T R A N C H A N T
P -B A
B F D
A A
The term “amusia” is often used to refer to its developmental form (i.e.,
congenital amusia), as previously described in this chapter. However,
amusia can also be acquired, often following a significant neurological
event (e.g., brain trauma or stroke). Because the severity and the location of
brain insults are typically unique to each individual, individuals with
acquired amusia present with very unique manifestations, often not limited
to a single aspect of music processing. As a result, the scientific literature
on acquired amusia has documented a wide array of music deficits (see
Clark, Golden, & Warren, 2015 for a review). Although studying both
congenital and acquired amusia allows for a better understating of how
music is processed by the brain, the study of acquired amusia is uniquely
positioned to inform us on deficient music processing networks resulting
from a neurological insult. A review of several acquired amusia case studies
suggests that lesions producing deficits in musical perception are
predominantly located in the right hemisphere (Stewart, von Kriegstein,
Warren, & Griffiths, 2006). However, some exceptions do exist. For
instance, there are reports describing acquired music perception
impairments that occurred following lesions to both hemispheres (Ayotte,
Peretz, Rousseau, Bard, & Bowjanowski, 2000; Liégeois-Chauvel, Peretz,
Babaï, Laguitton, & Chauvel, 1998; Peretz, 1990; Schuppert, Münte,
Wieringer, & Altenmüller, 2000).
M A
W H W L S
A ?
R
Albouy, P., Mattout, J., Bouet, R., Maby, E., Sanchez, G., Aguera, P. E., … Tillmann, B. (2013).
Impaired pitch perception and memory in congenital amusia: The deficit starts in the auditory
cortex. Brain 136(5), 1639–1661.
Albouy, P., Mattout, J., Sanchez, G., Tillmann, B., & Cacin, A. (2015). Altered retrieval of melodic
information in congenital amusia: Insights from dynamic causal modeling of MEG data. Frontiers
in Human Neuroscience 9. Retrieved from
http://journal.frontiersin.org/Article/10.3389/fnhum.2015.00020/abstract
Avidan, G., & Behrmann, M. (2009). Functional MRI reveals compromised neural integrity of the
face processing network in congenital prosopagnosia. Current Biology 19(13), 1146–1150.
Ayotte, J., Peretz, I., & Hyde, K. (2002). Congenital amusia: A group study of adults afflicted with a
music-specific disorder. Brain 125(2), 238–251.
Ayotte, J., Peretz, I., Rousseau, I., Bard, C., & Bowjanowski, M. (2000). Patterns of music agnosia
associated with middle cerebral artery infarcts. Brain 123(9), 1926–1938.
Bégel, V., Benoit, C.-E., Correa, A., Cutanda, D., Kotz, S. A., & Dalla Bella, S. (2017). “Lost in
time” but still moving to the beat. Neuropsychologia 94, 129–138.
Behrmann, M., & Avidan, G. (2005). Congenital prosopagnosia: Face-blind from birth. Trends in
Cognitive Sciences 9(4), 180–187.
Boets, B., Op de Beeck, H. P., Vandermosten, M., Scott, S. K., Gillebert, C. R., Mantini, D., …
Ghesquière, P. (2013). Intact but less accessible phonetic representations in adults with dyslexia.
Science 342(6163), 1251–1254.
Cirelli, L. K., Einarson, K. M., & Trainor, L. J. (2014). Interpersonal synchrony increases prosocial
behavior in infants. Developmental Science 17(6), 1003–1011.
Clark, C. N., Golden, H. L., & Warren, J. D. (2015). Acquired amusia. Handbook of Clinical
Neurology 129, 607–631.
Confavreux, C., Croisile, B., Garassus, P., Aimard, G., & Trillet, M. (1992). Progressive amusia and
aprosody. Archives of Neurology 49(9), 971–976.
Costanzo, F., Varuzza, C., Rossi, S., Sdoia, S., Varvara, P., Oliveri, M., … Menghini, D. (2016).
Reading changes in children and adolescents with dyslexia after transcranial direct current
stimulation. NeuroReport 27(5), 295–300.
Darwin, C. (1871). The descent of man, and selection in relation to sex. London: Murray.
Fujii, T., Fukatsu, R., Watabe, S. I., Ohnuma, A., Teramura, K., Kimura, I., … Kogure, K. (1990).
Auditory sound agnosia without aphasia following a right temporal lobe lesion. Cortex 26(2), 263–
268.
Gosselin, N., Paquette, S., & Peretz, I. (2015). Sensitivity to musical emotions in congenital amusia.
Cortex 71, 171–182.
Hirel, C., Lévêque, Y., Deiana, G., Richard, N., Cho, T.-H., Mechtouff, L., … Nighoghossian, N.
(2014). Amusie acquise et anhédonie musicale. Revue Neurologique 170(8–9), 536–540.
Hochman, M., & Abrams, K. (2014). Amusia for pitch caused by right middle cerebral artery infarct.
Journal of Stroke and Cerebrovascular Diseases 23(1), 164–165.
Honing, H. (2013). Structure and interpretation of rhythm in music. In D. Deutsch (Ed.), The
psychology of music (pp. 369–404). London: Academic Press.
Hyde, K. L., Lerch, J. P., Zatorre, R. J., Griffiths, T. D., Evans, A. C., & Peretz, I. (2007). Cortical
thickness in congenital amusia: When less is better than more. Journal of Neuroscience 27(47),
13028–13032.
Hyde, K. L., & Peretz, I. (2004). Brains that are out of tune but in time. Psychological Science 15(5),
356–360.
Hyde, K. L., Zatorre, R. J., Griffiths, T. D., Lerch, J. P., & Peretz, I. (2006). Morphometry of the
amusic brain: A two-site study. Brain 129(10), 2562–2570.
Hyde, K. L., Zatorre, R. J., & Peretz, I. (2011). Functional MRI evidence of an abnormal neural
network for pitch processing in congenital amusia. Cerebral Cortex 21(2), 292–299.
Koelsch, S. (2011). Toward a neural basis of music perception: A review and updated model.
Frontiers in Psychology 2. Retrieved from https://doi.org/10.3389/fpsyg.2011.00110
Koelsch, S., & Siebel, W. A. (2005). Towards a neural basis of music perception. Trends in Cognitive
Sciences 9(12), 578–584.
Leveque, Y., Fauvel, B., Groussard, M., Caclin, A., Albouy, P., Platel, H., & Tillmann, B. (2016).
Altered intrinsic connectivity of the auditory cortex in congenital amusia. Journal of
Neurophysiology 116(1), 88–97.
Liégeois-Chauvel, C., Peretz, I., Babaï, M., Laguitton, V., & Chauvel, P. (1998). Contribution of
different cortical areas in the temporal lobes to music processing. Brain 121(10), 1853–1867.
Lonsdale, A. J., & North, A. C. (2011). Why do we listen to music? A uses and gratifications
analysis. British Journal of Psychology 102(1), 108–134.
Loui, P., Alsop, D., & Schlaug, G. (2009). Tone deafness: A new disconnection syndrome? Journal
of Neuroscience 29(33), 10215–10220.
Lyon, G. R., Shaywitz, S. E., & Shaywitz, B. A. (2003). A definition of dyslexia. Annals of Dyslexia
53(1), 1–14.
Martínez-Molina, N., Mas-Herrero, E., Rodríguez-Fornells, A., Zatorre, R. J., & Marco-Pallarés, J.
(2016). Neural correlates of specific musical anhedonia. Proceedings of the National Academy of
Sciences 113(46), E7337–E7345.
Mas-Herrero, E., Zatorre, R. J., Rodríguez-Fornells, A., & Marco-Pallarés, J. (2014). Dissociation
between musical and monetary reward responses in specific musical anhedonia. Current Biology
24(6), 699–704.
Mathias, B., Lidji, P., Honing, H., Palmer, C., & Peretz, I. (2016). Electrical brain responses to beat
irregularities in two cases of beat deafness. Frontiers in Neuroscience 10. Retrieved from
https://doi.org/10.3389/fnins.2016.00040
Mendez, M. F., & Geehan, G. R. (1988). Cortical auditory disorders: Clinical and psychoacoustic
features. Journal of Neurology, Neurosurgery, and Psychiatry 51(1), 1–9.
Merriam, A. P. (1964). The anthropology of music. Evanston, IL: Northwestern University Press.
Moreau, P., Jolicœur, P., & Peretz, I. (2009). Automatic brain responses to pitch changes in
congenital amusia. Annals of the New York Academy of Sciences 1169, 191–194.
Neave, N., McCarty, K., Freynik, J., Caplan, N., Honekopp, J., & Fink, B. (2011). Male dance moves
that catch a woman’s eye. Biology Letters 7(2), 221–224.
Nettl, B. (2000). An ethnomusicologist contemplates universals in musical sound and musical
culture. In N. L. Wallin, B. Merker, & S. Brown (Eds.), The origins of music (pp. 463–472).
Cambridge, MA: MIT Press.
Nozaradan, S., Peretz, I., & Mouraux, A. (2012). Selective neuronal entrainment to the beat and
meter embedded in a musical rhythm. Journal of Neuroscience 32(49), 17572–17581.
Omigie, D., Müllensiefen, D., & Stewart, L. (2012). The experience of music in congenital amusia.
Music Perception: An Interdisciplinary Journal 30(1), 1–18.
Opitz, B., Rinne, T., Mecklinger, A., von Cramon, D. Y., & Schröger, E. (2002). Differential
contribution of frontal and temporal cortices to auditory change detection: fMRI and ERP results.
NeuroImage 15(1), 167–174.
Palmer, C., Lidji, P., & Peretz, I. (2014). Losing the beat: Deficits in temporal coordination.
Philosophical Transactions of the Royal Society B: Biological Sciences 369(1658), 20130405–
20130405.
Paquette, S., Fujii, S., Li, H. C., & Schlaug, G. (2017). The cerebellum’s contribution to beat interval
discrimination. NeuroImage, 163, 177–182.
Paquette, S., Li, H. C., Corrow, S. L., Buss, S. S., Barton, J., & Schlaug, G. (2018). Developmental
perceptual impairments: Cases when tone-deafness and prosopagnosia co-occur. Frontiers in
human neuroscience, 12, 438.
Peretz, I. (1990). Processing of local and global musical information by unilateral brain-damaged
patients. Brain 113(4), 1185–1205.
Peretz, I. (2001). Brain specialization for music. Annals of the New York Academy of Sciences 930,
153–165.
Peretz, I. (2002). Brain specialization for music. The Neuroscientist 8(4), 372–380.
Peretz, I. (2006). The nature of music from a biological perspective. Cognition 100(1), 1–32.
Peretz, I. (2016). Neurobiology of congenital amusia. Trends in Cognitive Sciences 20(11), 857–867.
Peretz, I., Ayotte, J., Zatorre, R. J., Mehler, J., Ahad, P., Penhune, V. B., & Jutras, B. (2002).
Congenital amusia: A disorder of fine-grained pitch discrimination. Neuron 33(2), 185–191.
Peretz, I., Brattico, E., Järvenpää, M., & Tervaniemi, M. (2009). The amusic brain: In tune, out of
key, and unaware. Brain 132(5), 1277–1286.
Peretz, I., Brattico, E., & Tervaniemi, M. (2005). Abnormal electrical brain responses to pitch in
congenital amusia. Annals of Neurology 58(3), 478–482.
Peretz, I., Champod, A. S., & Hyde, K. (2003). Varieties of musical disorders: The Montreal battery
of evaluation of amusia. Annals of the New York Academy of Sciences 999, 58–75.
Peretz, I., & Coltheart, M. (2003). Modularity of music processing. Nature Neuroscience 6(7), 688–
691.
Peretz, I., Cummings, S., & Dubé, M. P. (2007). The genetics of congenital amusia (tone deafness):
A family-aggregation study. American Journal of Human Genetics 81(3), 582–588.
Peretz, I., Gagnon, L., & Bouchard, B. (1998). Music and emotion: Perceptual determinants,
immediacy, and isolation after brain damage. Cognition 68(2), 111–141.
Peretz, I., Kolinsky, R., Tramo, M., Labrecque, R., Hublet, C., Demeurisse, G., & Belleville, S.
(1994). Functional dissociations following bilateral lesions of auditory cortex. Brain 117(6), 1283–
1301.
Peretz, I., & Vuvan, D. (2017). Prevalence of congenital amusia. European Journal of Human
Genetics 25, 625–630.
Phillips-Silver, J., Toivianen, P., Gosselin, N., Piché, O., Nozaradan, S., Palmer, C., & Petetz, I.
(2011). Born to dance but beat deaf: A new form of congenital amusia. Neuropsychologia 49(5),
961–969.
Ramus, F. (2014). Neuroimaging sheds new light on the phonological deficit in dyslexia. Trends in
Cognitive Sciences 18(6), 274–275.
Repp, B. H. (2005). Sensorimotor synchronization: A review of the tapping literature. Psychonomic
Bulletin & Review 12(6), 969–992.
Salimpoor, V. N., van den Bosch., I., Kovacevik, N., McIntosh, A. R., Dagher, A., & Zatorre, R. J.
(2013). Interactions between the nucleus accumbens and auditory cortices predict music reward
value. Science 340(6129), 216–219.
Särkämö, T., Tervaniemi, M., Soinila, S., Autti, T., Silvennoinen, H. M., Laine, M., & Hietanen, M.
(2009). Cognitive deficits associated with acquired amusia after stroke: A neuropsychological
follow-up study. Neuropsychologia 47(12), 2642–2651.
Schaal, N. K., Pfeifer, J., Krause, V., & Pollok, B. (2015). From amusic to musical? Improving pitch
memory in congenital amusia with transcranial alternating current stimulation. Behavioural Brain
Research 294, 141–148.
Schuppert, M., Münte, T. F., Wieringer, B. M., & Altenmüller, E. (2000). Receptive amusia:
Evidence for cross-hemispheric neural networks underlying music processing strategies. Brain
123(3), 546–559.
Sihvonen, A., Ripollés, P., Leo, V., Rodríguez-Fornells, A., Soinila, S., & Särkämö, T. (2016). Neural
basis of acquired amusia and its recovery after stroke. Journal of Neuroscience 36(34), 8872–
8881.
Sihvonen, A., Ripollés, P., Rodríguez-Fornells, A., Soinila, S., & Särkämö, T. (2017). Revisiting the
neural basis of acquired amusia: Lesion patterns and structural changes underlying amusia
recovery. Frontiers in Neuroscience 11. Retrieved from https://doi.org/10.3389/fnins.2017.00426
Sihvonen, A., Ripollés, P., Särkämö, T., Leo, V., Rodríguez-Fornells, A., Saunavaara, J., … Soinila,
S. (2017). Tracting the neural basis of music: Deficient structural connectivity underlying acquired
amusia. Cortex 97, 255–273.
Sowiński, J., & Dalla Bella, S. (2013). Poor synchronization to the beat may result from deficient
auditory-motor mapping. Neuropsychologia 51(10), 1952–1963.
Stewart, L., von Kriegstein, K., Warren, J. D., & Griffiths, T. D. (2006). Music and the brain:
Disorders of musical listening. Brain: A Journal of Neurology 129(Pt. 10), 2533–2553.
Thomas, C., Avidan, G., Humphreys, K., Jung, K. J., Gao, F., & Behrmann, M. (2008). Reduced
structural connectivity in ventral visual cortex in congenital prosopagnosia. Nature Neuroscience
12(1), 29–31.
Tranchant, P., & Vuvan, D. T. (2015). Current conceptual challenges in the study of rhythm
processing deficits. Frontiers in Neuroscience 9. Retrieved from
https://doi.org/10.3389/fnins.2015.00197
Tranchant, P., Vuvan, D. T., & Peretz, I. (2016). Keeping the beat: A large sample study of bouncing
and clapping to music. PloS ONE 11(7), e0160178.
Trehub, S. E. (2001). Musical predispositions in infancy. Annals of the New York Academy of
Sciences 930, 1–16.
Vuvan, D., Nunes-Silva, M., & Peretz, I. (2015). Meta-analytic evidence for the non-modularity of
pitch processing in congenital amusia. Cortex 69, 186–200.
Vuvan, D., Paquette, S., Mignault Goulet, G., Royal, I., Felezeu, M. & Peretz, I. (2017). The
Montreal protocol for identification of amusia. Behavior Research Methods 50.
Whiteford, K. L., & Oxenham, A. J. (2018). Learning for pitch and melody discrimination in
congenital amusia. Cortex 103, 164–178.
Wilbiks, J. M. P., Vuvan, D. T., Girard, P.-Y., Peretz, I., & Russo, F. A. (2016). Effects of vocal
training in a musicophile with congenital amusia. Neurocase 22(6), 526–537.
Wiltermuth, S. S., & Heath, C. (2009). Synchrony and cooperation. Psychological Science 20(1), 1–
5.
Zatorre, R. J., & Peretz, I. (2001). The biological foundations of music. Annals of the New York
Academy of Sciences 930, 281–299.
CHAPT E R 32
W H E N B L U E T U R N S TO
G R AY: T H E E N I G M A O F
M U S I C I A N ’ S D Y S TO N I A
D AV I D P E T E R S O N A N D E C K A RT A LT E N MÜ L L E R
Dystonia
M ’ dystonia is a particular form of dystonia. The concept of
dystonia has evolved over the past several decades. A recent update to its
definition reflects a multiyear effort to achieve some consensus on how it is
described:
Dystonia is a movement disorder characterized by sustained or intermittent muscle
contractions causing abnormal, often repetitive, movements, postures, or both. Dystonic
movements are typically patterned, twisting, and may be tremulous. Dystonia is often
initiated or worsened by voluntary action and associated with overflow muscle activation.
(Albanese et al., 2013)
Musician’s Dystonia
Musician’s dystonia (MD) is a specific type of focal dystonia. Onset can
range from approximately 18 years old to the seventh decade, but most
commonly occurs in the mid-30s (Altenmüller, 2003; Brandfonbrener,
1995; Brandfonbrener & Robson, 2004; Chang & Frucht, 2013; Conti,
Pullman, & Frucht, 2008; Jankovic & Shale, 1989; Lederman, 1991;
Schuele & Lederman, 2004a). MD is one of the most perplexing forms of
dystonia. Documenting a cohort of over 590 MD patients diagnosed
between 1994 and 2007 at the Institute of Music Physiology and Musicians’
Medicine of the Hanover University of Music, Drama, and Media
(Altenmüller, Baur, Hofmann, Lim, & Jabusch, 2012; Jabusch &
Altenmüller, 2006a) it has been associated with almost every instrument
and in several body regions. In every case MD involves impaired voluntary
motor control while a musician is playing the instrument. The symptoms
generally appear during movements that are extensively trained. It can
affect control of facial, lip, and tongue muscles (“embouchure dystonia,”
Frucht et al., 2001), lower limbs, or, in the majority of patients, the muscles
controlling the arm or hand. MD involving the hand is a form of FHD.
MD is sometimes referred to as “musician’s cramp” because it is often
described in conjunction with a form of FHD called “writer’s cramp.”
However, the term “cramp” can be misleading; MD rarely involves the
maximum intensity contractions associated with cramps (Pesenti, Barbieri,
& Priori, 2004; Tubiana, 2000). In the hand, MD is usually associated with
loss of fine control and coordination most commonly in heterogeneous
subsets of digits 2–5 (Charness, Ross, & Shefner, 1996; Frucht, 2009b;
Furuya, Tominaga, Miyazaki, & Altenmüller, 2015; Jankovic & Ashoori,
2008; Jankovic & Shale, 1989). The relative amount of excessive finger
flexion or extension depends on the type of instrument (Frucht, 2009b;
Conti et al., 2008). Flexion is more common than extension. If multiple
fingers are involved, they are usually adjacent fingers. Patients report a
feeling of loss of automaticity in previously automatic music performance
(Frucht, 2009b). Several examples of abnormal postural configurations are
depicted in Fig. 1. MD is painless in most, but not all, patients (Jabusch &
Altenmüller, 2006b). Indeed pain may suggest diagnosis of repetitive strain
injury or occupational fatigue syndrome rather than MD.
FIGURE 1. Representative posture patterns in musician’s dystonia.
Reproduced from Hans-Christian Jabusch and Eckart Altenmüller, Epidemiology,
phenomenology, and therapy of musician’s cramp. In Eckart Altenmüller, Mario
Wiesendanger, and Jürg Kesselring (Eds.), Music, motor control, and the brain, p. 267,
Figure 17.2, doi:10.1093/acprof:oso/9780199298723.001.0001, Copyright © Oxford
University Press 2006, reproduced by permission of Oxford University Press.
FIGURE 4. Treatment efficacy based on patient ratings. Black bars: deterioration. Gray bars: No
change. White bars: alleviation. Hatched bars: no reply. Trhx, trihexyphenidyl; BT, botulinum toxin.
Reproduced from Hans-Christian Jabusch, Dorothea Zschucke, Alexander Schmidt, Stephan
Schuele, and Eckart Altenmüller, Focal dystonia in musicians: Treatment strategies and long-
term outcome in 144 patients, Movement Disorders 20(12), 1623–1626,
https://doi.org/10.1002/mds.20631, Copyright © 2005 Movement Disorder Society.
Plasticity
Plasticity is an overused and sometimes misinterpreted term in the
neurosciences. In the simplest sense, it refers to the ability of the system to
change. It has manifestations at behavioral and several physiological levels,
and likely has bases at the circuit and molecular levels, historically couched
in terms of changes in the network of synapses among neurons, and the
molecular signaling pathways that modulate the strength of those synapses
at different timescales.
In the context of dystonia, plasticity can refer to the near-real time
adaptations of sensorimotor systems, as for example a greater response than
normal to paired associate stimulation protocols, whereby the relative
timing of sensory stimuli and exogenous stimulation of motor cortical areas
can strengthen that sensory system’s subsequent ability to evoke the motor
response within the same experimental session. Plasticity can also refer to
the systems-level homologues of synaptic plastic process of long-term
potentiation and depression (LTP and LTD), in the form of high- and low-
frequency repetitive stimulation with TMS to potentiate or depress
subsequent cortical excitability. At a drastically slower timescale, plasticity
can refer to the slow, often insidious changes in brain circuitry that likely
underlie the pathogenesis of dystonia. There are also the notions of
“metaplasticity” and “homeoplasticity” which, among other things, are
thought to regulate the primary plastic processes already discussed.
In the case of MD, one theory is that musical practice in healthy
musicians is associated with (indeed, likely relies upon) beneficial plastic
adaptation in the motor cortex, including for example a reduction in motor
thresholds and increase in motor excitability, and that in MD patients these
processes have progressed too far and begin to compromise, rather than
enhance, movement patterns (Altenmüller & Jabusch, 2010b). Thus
musicians may have (or take advantage of) greater primary plastic
capabilities and MD patients may have dysfunctional metaplastic processes
that regulate the primary plastic processes in an abnormal way. A great
mystery in MD research is how such dysfunctional plastic processes are
initiated and used in generating the disorder.
P T
Biological Predisposition
With relatively rare exceptions, most central nervous system disorders,
including most forms of dystonia, likely arise from some complex
combination of genetic and environmental factors. In the adult onset
dystonias, this is usually conceptualized as a genetic predisposition
followed by some environmental “trigger.” In this chapter, we formulate a
theoretical framework for the pathogenesis of MD that is organized in a
similar fashion: biological predisposition and “use patterns” (see Fig. 5).
Biological predispositions include of course genetics but also gender. “Use
patterns” is a broad umbrella term that refers to how the sensorimotor
systems are used over protracted periods of time. As such, it can be
considered a category of “environmental” factors.
FIGURE 6. Gender distribution of 591 patients and 2651 healthy musicians, in relative ratios.
Reproduced from Eckart Altenmüller, Volker Baur, Aurélie Hofmann, Vanessa K. Lim, and
Hans-Christian Jabusch, Musician’s cramp as manifestation of maladaptive brain plasticity:
Arguments from instrumental differences, Annals of the New York Academy of Sciences
1252, 259–265, doi:10.1111/j.1749-6632.2012.06456.x Copyright © 2012, John Wiley and
Sons.
Use Patterns
Several features of MD suggest that how the sensorimotor systems are used
in music performance may be a factor contributing to the development of
MD. Motor workload and movement complexity appear to be risk factors.
MD is more likely to appear in the hand with higher demands of
spatiotemporal precision, such as the right hand on keyboards and plucked
instruments, and the left hand in bowed string players. Among the string
instruments, MD appears to be more prevalent on string instruments with
shorter string lengths, such as the violin versus double-bass (Altenmüller et
al., 2012; Jabusch & Altenmüller, 2006b). In fact, the relative absence of
MD documented in double bassists may be due to a lack of simultaneous
finger action (Conti et al., 2008). Another apparent risk factor is the type of
musical performance. Classical musicians seem to be more at risk of
developing MD than pop or jazz musicians. A hypothesized reason why is
that classical music involves higher expectations of temporally precise
reproduction and less opportunity for improvisation than pop and jazz
(Jabusch & Altenmüller, 2006b). Professional classical musicians have also
typically undergone many tens of thousands of hours of practice, involving
movements that are extensively trained, representing an “oversampling” of
a disproportionately narrow part of the space of possible sensorimotor
mappings. Finally, in some cases, peripheral issues such as prolonged pain
syndromes, nerve entrapment, and subclinical range of motion limitations
(Charness et al., 1996; Leijnse, Hallett, & Sonneveld, 2015) that can
precede MD onset may induce compensatory changes in use patterns that
are pathogenic and would not otherwise occur in a healthy musician’s motor
repertoire. Furuya (Furuya & Hanakawa, 2016) suggests that such
dysfunctional adaptations of body representations in somatosensory and
motor systems may be an intermediate point on the path toward full MD
development. It should be noted that these are all retrospective
observational results and there have been no controlled, prospective studies
regarding these factors.
As with genetics and gender, the mechanisms by which use patterns
contribute to MD pathogenesis are unclear. But having a theoretical
framework can help provide a rational basis for future experimental
research.
F D
Better Assessment
Among the numerous unmet needs in dystonia (Albanese, 2017), one of the
designated research priorities in focal task-specific dystonias is more
precise methods for characterizing and assessing the phenotype (Richardson
et al., 2017). This is acutely evident for MD. For over a decade dystonia
researchers have suggested that a new rating scale for MD that is reliable,
valid, sensitive, and specific to MD is sorely needed (Jankovic and Ashoori,
2008; Spector & Brandfonbrener, 2007). Peterson and colleagues (Peterson
et al., 2013) comprehensively summarized the state of affairs in rating scale
use for MD. As alluded to in the section “Rating Scale Deficiencies,” none
of the existing scales have been completely and rigorously evaluated for
reliability and validity, sensitivity to change, practical use in a clinical
setting, and specifically tailored to MD. Exacerbating the concerns about
reliability and sensitivity to change, most of the existing rating scales are
based on human judgments, making them inherently subjective. Further
developments in objective rating instruments that would make them more
readily applicable in the clinical setting would help mitigate these issues.
As evident in the MD rating scale review (Peterson et al., 2013), there
appears to be no standard choice for rating scale(s) in MD research; most
studies use only one or two rating scales but the scales used vary widely
from one study to the next. When trying to compare and reconcile multiple
treatment studies, this makes it difficult to discern between treatment effects
and measurement effects. Likewise non-standard selection of rating scales
diminishes the collective research value of pathophysiology studies.
In summary, as depicted in Fig. 7, we suggest that the migration to a
more ideal mode for assessing MD severity would include two categories of
improvements: (1) standardizing the choice of rating scale(s) used across
studies, and (2) increasing the efficiency with which a small number of
rating scales can completely cover the conceptual space of measurements
one wants to make in the disorder. This mode would accelerate research
into both the pathophysiology of MD and improved treatments.
FIGURE 7. Toward ideal rating scale use. Upper left: current rating scale use is neither consistent
across studies nor efficient. Upper right: efficient rating scale use would minimize the number of
scales used in each study (in this example depiction, to two scales per study). Lower left: consistent
rating scale use would ensure that the same rating scales are used across all studies. Lower right: the
ideal scenario would be both consistent and efficient by pushing the envelope on both: a consistent
application of the minimal number of scales across all studies.
New Treatments
Another designated research priority in focal task-specific dystonias is
innovative clinical trial design that takes into account the tremendous
heterogeneity in the presentation of these dystonias from one patient to the
next (Richardson et al., 2017). Despite concerted efforts to evaluate an array
of new treatment approaches, most have involved small, unblinded,
retrospective studies. Clearly we need new trials that are controlled,
blinded, prospective, and randomized. Unfortunately, these designs are very
difficult to implement in the space of PMR treatments. This poses a creative
challenge for the field of MD, given the likely prominent role of PMR in
MD treatment. Relatedly, given the heterogeneity in MD manifestation,
treatments should logically be tailored to the individual patient. This is also
challenging in the context of the common research goal of reproducibility.
Yet there is a persistent need for new therapies for MD. Although many
patients manage to stay in the field with currently available therapies, most
have to make substantial compromises in their level of music performance.
And outcomes are particularly limiting for patients with embouchure
dystonia (Frucht et al., 2001; Jabusch & Altenmüller, 2006b).
Given the central theme of disordered temporal processing in MD, we
hypothesize that the highly refined timing characteristics of the
sensorimotor systems in professional musicians are critical to not only
understanding the disorder but also optimizing treatment. Simply put,
patients get stuck in patterns of inappropriate motor sequences and greater
attention to the critical role of time in related auditory perception and motor
performance could help them overcome these dysfunctional patterns. Future
treatment research should evaluate novel non-pharmacological
interventions for MD that are focused on timing and determine whether and
how their efficacy is related to psychophysical measures of temporal
processing, such as the TDT.
With respect to the motor performance aspect of timing, one form of
PMR-style intervention explicitly focused on timing is slow-down exercises
(SDE; Sakai, 2006). SDE is an MD rehabilitation strategy that has patients
slow down the tempo of their symptom-evoking performance pieces until
symptoms resolve, and then gradually increase the tempo back to the
original speed as long as symptoms do not return. To our knowledge,
groups independent of the original developers have not evaluated SDE. And
the original application of SDE required many weeks of retraining to
establish effects. Nevertheless, approaches like SDE merit further
systematic, controlled investigation by independent groups.
TMS offers a non-invasive method to not only study MD but also,
typically in the form of repetitive TMS (rTMS), modulate it. rTMS (and its
surface voltage counterpart transcranial Direct Current Stimulation, tDCS)
have been evaluated as potential treatments for FHD, mostly with writer’s
cramp (Cho & Hallett, 2016; Obeso, Cerasa, & Quattrone, 2016) but also
with MD (Furuya et al., 2014; Kieslinger, Holler, Bergmann, Golaszewski,
& Staffen, 2013). Because it inherently involves stimulation with temporal
precision at millisecond timescales, TMS provides a potentially
complementary physiologic counterpart to the PMR methods that operate at
behavioral timescales. Although research on their combined utility as FHD
therapy has thus far shown only mixed results (Kimberley, Schmidt, Chen,
Dykstra, & Buetefisch, 2015), we expect further trials in the near future.
Although stereotactic surgeries, especially deep brain stimulation (DBS)
in the globus pallidus internum (GPi), have demonstrated efficacy in
generalized dystonia, there are relatively fewer reported series in focal
dystonias, and to our knowledge none in MD. Future advances in DBS
technology, including closed loop designs, may be able to incorporate task
context (either behaviorally or physiologically) and facilitate symptom
reduction that is appropriately context-dependent for focal task-specific
dystonias. However, as with the botulinum toxin injections, the exquisite
spatiotemporal demands of music performance will make MD probably one
of the last indications for the treatment. Although lesion methods have
fallen out of favor with the advent of DBS, there is one successful reported
case of a thalamotomy for an MD patient that was refractory to oral
medications and botulinum toxin (Horisawa et al., 2017).
Endogenous cannabinoid receptors play an important role in, among
other things, synaptic plasticity processes in the basal ganglia. Thus, they
are a rational target for dystonia. Jabusch reported positive though transient
benefits from a single dose of THC in a pianist with MD (Jabusch,
Schneider, & Altenmüller, 2004b), but subsequent controlled studies in
more broadly defined movement disorders populations have since produced
mixed results (Kluger, Triolo, Jones, & Jankovic, 2015; Koppel et al.,
2014).
Since successful treatment is still a challenge, preventing musician’s
dystonia is important. Although prospective studies are lacking, avoidance
of triggering factors, such as chronic pain, overuse, anxiety, and mechanical
repetitions are important and may prevent manifestation of MD, especially
in those musicians with genetic susceptibility.
Research on Pathophysiology
The sensorimotor systems employed during music performance operate at
high rates and with great temporal precision. Thus future basic and clinical
research in MD should specifically measure and modulate the motor control
system with a focus on timing. The TDT is one obvious paradigm for
pursuing this. But the TDT is usually measured in the visual or
somatosensory domains. Yet the auditory modality is particularly important
for musicians. So future work should include the auditory TDT to measure
the temporal precision of auditory processing in MD patients.
Research into temporal processing in MD should also be carefully
integrated with previous themes in pathophysiology, such as altered
surround inhibition. We expect that the TDT deficits in dystonia that have
been interpreted as a time-domain version of reduced surround inhibition
(Tamura et al., 2009) are not independent from but actually interact with
alterations in the precision of “spatial” surround inhibition processes.
Recent evidence that surround inhibition can be modulated by attention
(Kuhn, Keller, Lauber, & Taube, 2018) provides a timely segue to our other
recommendation for future MD pathophysiology research: we should
allocate some resources to attention. Although difficult to assay in non-
human primates, attention may have been a factor in a monkey model of
FHD (Byl, Merzenich, & Jenkins, 1996). And differential use of attention
seems to be a factor in PMR approaches to MD (Brian Hays, personal
communication).
And attentional modulation has likely been a key element in what is
sometimes labeled as psychogenic dystonia because it is often based on
tests of distractibility. If a symptom is modulated when attention is directed
to or away from a motor function, is it still organic dystonia? Regardless of
what we call it, attention seems to play a role. Indeed allocation of attention
has itself been considered a form of action selection, and therefore
attentional focus can be considered part of the task-dependent aspect
inherent in most focal dystonias. Unfortunately, attention can be difficult to
measure and is often not considered in evaluating overt motor function. But
simple gaze monitoring may provide a reasonable first step.
The brain circuits mediating attention are widespread but likely rely
heavily upon thalamic systems that have thus far been relatively
understudied in neuroscience. Yet the thalamus is a central node mediating
communication among many motor systems including the cerebellum,
cortex, striatum, and of course brainstem. Indeed Hutchinson and
colleagues (Hutchinson et al., 2013) have postulated that the projection
from the superior colliculus to the striatum via the intralaminar nuclei of the
thalamus mediates the TDT. So future pathophysiology in research on
dystonias, including MD, should redirect attention to timing, attentional
focus, and the thalamus.
A
Peterson wishes to thank Dominique Sy for assistance generating Figs. 5
and 7. Peterson acknowledges partial support from the Dystonia Coalition
(NS065701 and TR001456), from the Office of Rare Diseases Research at
the National Center for Advancing Translational Sciences and the National
Institute of Neurological Disorders and Stroke, the Bachmann-Strauss
Dystonia & Parkinson Foundation, the Benign Essential Blepharospasm
Research Foundation, UCSD’s Kavli Institute for Brain and Mind, the
National Institute of Mental Health (NIMH 5T32-MH020002), the National
Science Foundation (the Temporal Dynamics of Learning Center, a Science
of Learning Center [SMA-1041755] and the program in Mind, Machines,
Motor Control [EFRI-1137279]), the Howard Hughes Medical Institute, and
the Congressionally Directed Medical Research Program (W81XWH-17-1-
0393). Altenmüller acknowledges support from BMBF Project Dystract:
100255620.
R
Albanese, A. (2017). Editorial: Unmet needs in dystonia. Frontiers in Neurology 8. Retrieved from
https://doi.org/10.3389/fneur.2017.00197
Albanese, A., Bhatia, K., Bressman, S. B., Delong, M. R., Fahn, S., Fung, V. S. C., … Teller, J. K.
(2013). Phenomenology and classification of dystonia: A consensus update. Movement Disorders
28(7), 863–873.
Altenmüller, E. (2003). Focal dystonia: Advances in brain imaging and understanding of fine motor
control in musicians. Hand Clinics 19(3), 523–538.
Altenmüller, E., Baur, V., Hofmann, A., Lim, V. K., & Jabusch, H. C. (2012). Musician’s cramp as
manifestation of maladaptive brain plasticity: Arguments from instrumental differences. Annals of
the New York Academy of Sciences 1252, 259–265.
Altenmüller, E., & Jabusch, H. C. (2009). Focal hand dystonia in musicians: Phenomenology,
etiology, and psychological trigger factors. Journal of Hand Therapy 22(2), 144–154; quiz 155.
Altenmüller, E., & Jabusch, H. C. (2010a). Focal dystonia in musicians: Phenomenology,
pathophysiology, triggering factors, and treatment. Medical Problems of Performing Artists 25(1),
3–9.
Altenmüller, E., & Jabusch, H. C. (2010b). Focal dystonia in musicians: Phenomenology,
pathophysiology and triggering factors. European Journal of Neurology 17(1), 31–36.
Altenmüller, E., Kesselring, J., & Wiesendanger, M. (2006). Music, motor control and the brain.
Oxford: Oxford University Press.
Altenmüller, E., & Müller, D. (2013). A model of task-specific focal dystonia. Neural Networks 48,
25–31.
Avanzino, L., Pelosin, E., Vicario, C. M., Lagravinese, G., Abbruzzese, G., & Martino, D. (2016).
Time processing and motor control in movement disorders. Frontiers in Human Neuroscience 10.
Retrieved from https://doi.org/10.3389/fnhum.2016.00631
Avanzino, L., Tinazzi, M., Ionta, S., & Fiorio, M. (2015). Sensory-motor integration in focal
dystonia. Neuropsychologia 79(Part B), 288–300.
Bara-Jimenez, W., Catalan, M. J., Hallett, M., & Gerloff, C. (1998). Abnormal somatosensory
homunculus in dystonia of the hand. Annals of Neurology 44(5), 828–31.
Bara-Jimenez, W., Shelton, P., Sanger, T. D., & Hallett, M. (2000). Sensory discrimination
capabilities in patients with focal hand dystonia. Annals of Neurology 47(3), 377–380.
Baur, V., Jabusch, H. C., & Altenmüller, E. (2011). Behavioral factors influence the phenotype of
musician’s dystonia. Movement Disorders 26(9), 1780–1781.
Bernstein, C. J., Ellard, D. R., Davies, G., Hertenstein, E., Tang, N. K. Y., Underwood, M., &
Sandhu, H. (2016). Behavioural interventions for people living with adult-onset primary dystonia:
A systematic review. BMC Neurology 16. doi:10.1186/s12883-016-0562-y
Bradley, D., Whelan, R., Walsh, R., Reilly, R. B., Hutchinson, S., Molloy, F., & Hutchinson, M.
(2009). Temporal discrimination threshold: VBM evidence for an endophenotype in adult onset
primary torsion dystonia. Brain 132(9), 2327–2335.
Brandfonbrener, A. G. (1995). Musicians with focal dystonia: A report of 58 cases seen during a 10-
year period at a performing-arts medical-center. Medical Problems of Performing Artists 10, 121–
127.
Brandfonbrener, A. G., & Robson, C. (2004). Review of 113 musicians with focal dystonia seen
between 1985 and 2002 at a clinic for performing artists. Advances in Neurology 94, 255–256.
Butler, J. S., Beiser, I. M., Williams, L., McGovern, E., Molloy, F., Lynch, T., … Hutchinson, M.
(2015). Age-related sexual dimorphism in temporal discrimination and in adult-onset dystonia
suggests GABAergic mechanisms. Frontiers in Neurology 6. Retrieved from
https://doi.org/10.3389/fneur.2015.00258
Byl, N. N., Merzenich, M. M., & Jenkins, W. M. (1996). A primate genesis model of focal dystonia
and repetitive strain injury. 1. Learning-induced dedifferentiation of the representation of the hand
in the primary somatosensory cortex in adult monkeys. Neurology 47(2), 508–520.
Candia, V., Wienbruch, C., Elbert, T., Rockstroh, B., & Ray, W. (2003). Effective behavioral
treatment of focal hand dystonia in musicians alters somatosensory cortical organization.
Proceedings of the National Academy of Sciences 100, 7942–7946.
Chang, F. C. F., & Frucht, S. J. (2013). Motor and sensory dysfunction in musician’s dystonia.
Current Neuropharmacology 11(1), 41–47.
Charness, M. E., Ross, M. H., & Shefner, J. M. (1996). Ulnar neuropathy and dystonic flexion of the
fourth and fifth digits: Clinical correlation in musicians. Muscle and Nerve 19(4), 431–437.
Chen, R., Classen, J., Gerloff, C., Celnik, P., Wassermann, E. M., Hallett, M., & Cohen, L. G. (1997).
Depression of motor cortex excitability by low-frequency transcranial magnetic stimulation.
Neurology 48(5), 1398–1403.
Cho, H. J., & Hallett, M. (2016). Non-invasive brain stimulation for treatment of focal hand dystonia:
Update and future direction. Journal of Movement Disorders 9(2), 55–62.
Cohen, L. G., & Hallett, M. (1988). Hand cramps: Clinical features and electromyographic patterns
in a focal dystonia. Neurology 38(7), 1005–1012.
Conti, A. M., Pullman, S., & Frucht, S. J. (2008). The hand that has forgotten its cunning: Lessons
from musicians’ hand dystonia. Movement Disorders 23(10), 1398–1406.
Delmaire, C., Krainik, A., Du Montcel, S. T., Gerardin, E., Meunier, S., Mangin, J. F., … Lehericy, S.
(2005). Disorganized somatotopy in the putamen of patients with focal hand dystonia. Neurology
64(8), 1391–1396.
Delmaire, C., Vidailhet, M., Wassermann, D., Descoteaux, M., Valabregue, R., Bourdain, F., …
Lehericy, S. (2009). Diffusion abnormalities in the primary sensorimotor pathways in writer’s
cramp. Archives of Neurology 66(4), 502–508.
Elbert, T., Candia, V., Altenmüller, E., Rau, H., Sterr, A., Rockstroh, B., … Taub, E. (1998).
Alteration of digital representations in somatosensory cortex in focal hand dystonia. Neuroreport
9(16), 3571–3575.
Fiorio, M., Gambarin, M., Valente, E. M., Liberini, P., Loi, M., Cossu, G., … Tinazzi, M. (2007).
Defective temporal processing of sensory stimuli in DYT1 mutation carriers: A new
endophenotype of dystonia? Brain 130(1), 134–142.
Fiorio, M., Tinazzi, M., Bertolasi, L., & Aglioti, S. M. (2003). Temporal processing of visuotactile
and tactile stimuli in writer’s cramp. Annals of Neurology 53(5), 630–635.
Frucht, S. J. (2009a). Embouchure dystonia: Portrait of a task-specific cranial dystonia. Movement
Disorders 24(12), 1752–1762.
Frucht, S. J. (2009b). Focal task-specific dystonia of the musicians’ hand: A practical approach for
the clinician. Journal of Hand Therapy 22(2), 136–142.
Frucht, S. J. (2015). Evaluating the musician with dystonia of the upper limb: A practical approach
with video demonstration. Journal of Clinical Movement Disorders 2. doi:10.1186/s40734-015-
0026-3
Frucht, S. J., Fahn, S., Greene, P. E., O’Brien, C., Gelb, M., Truong, D. D., … Ford, B. (2001). The
natural history of embouchure dystonia. Movement Disorders 16(5), 899–906.
Fuchs, T., Saunders-Pullman, R., Masuho, I., Luciano, M. S., Raymond, D., Factor, S., … Ozelius, L.
J. (2013). Mutations in GNAL cause primary torsion dystonia. Nature Genetics 45, 88-U128.
Furuya, S., & Hanakawa, T. (2016). The curse of motor expertise: Use-dependent focal dystonia as a
manifestation of maladaptive changes in body representation. Neuroscience Research 104, 112–
119.
Furuya, S., Nitsche, M. A., Paulus, W., & Altenmüller, E. (2014). Surmounting retraining limits in
musicians’ dystonia by transcranial stimulation. Annals of Neurology 75(5), 700–707.
Furuya, S., Tominaga, K., Miyazaki, F., & Altenmüller, E. (2015). Losing dexterity: Patterns of
impaired coordination of finger movements in musician’s dystonia. Scientific Reports 5. doi:
10.1038/srep13360
Gilio, F., Curra, A., Inghilleri, M., Lorenzano, C., Suppa, A., Manfredi, M., & Berardelli, A. (2003).
Abnormalities of motor cortex excitability preceding movement in patients with dystonia. Brain
126, 1745–1754.
Group, D. S. (2004). Rating scales for dystonia: Assessment of reliability of three scales. Advances in
Neurology 94, 329–336.
Hallett, M. (2006). Pathophysiology of writer’s cramp. Human Movement Science 25(4–5), 454–463.
Haslinger, B., Noe, J., Altenmüller, E., Riedl, V., Zimmer, C., Mantel, T., & Dresel, C. (2017).
Changes in resting-state connectivity in musicians with embouchure dystonia. Movement
Disorders 32(3), 450–458.
Hays, B. (1987). Painless hand problems of string-pluckers. Medical Problems of Performing Artists
2, 39–40.
Horisawa, S., Tamura, N., Hayashi, M., Matsuoka, A., Hanada, T., Kawamata, T., & Taira, T. (2017).
Gamma knife ventro-oral thalamotomy for musician’s dystonia. Movement Disorders 32(1), 89–
90.
Hutchinson, M., Kimmich, O., Molloy, A., Whelan, R., Molloy, F., Lynch, T., … O’Riordan, S.
(2013). The endophenotype and the phenotype: Temporal discrimination and adult-onset dystonia.
Movement Disorders 28(13), 1766–1774.
Ioannou, C. I., & Altenmüller, E. (2014). Psychological characteristics in musician’s dystonia: A new
diagnostic classification. Neuropsychologia 61, 80–88.
Ioannou, C. I., Furuya, S., & Altenmüller, E. (2016). The impact of stress on motor performance in
skilled musicians suffering from focal dystonia: Physiological and psychological characteristics.
Neuropsychologia 85, 226–236.
Jabusch, H. C., & Altenmüller, E. (2004). Anxiety as an aggravating factor during onset of focal
dystonia in musicians. Medical Problems of Performing Artists 19, 75–81.
Jabusch, H. C., & Altenmüller, E. (2006a). Epidemiology, phenomenology, and therapy of musician’s
cramp. In E. Altenmüller, J. Kesselring, & M. Wiesendanger (Eds.), Music, motor control and the
brain (pp. 265–282). Oxford: Oxford University Press.
Jabusch, H. C., & Altenmüller, E. (2006b). Focal dystonia in musicians: From phenomenology to
therapy. Advances in Cognitive Psychology 2(2–3), 207–220.
Jabusch, H. C., Müller, S. V., & Altenmüller, E. (2004a). Anxiety in musicians with focal dystonia
and those with chronic pain. Movement Disorders 19(10), 1169–1175.
Jabusch, H. C., Schneider, U., & Altenmüller, E. (2004b). Delta 9-tetrahydrocannabinol improves
motor control in a patient with musician’s dystonia. Movement Disorders 19, 990–991.
Jabusch, H. C., Vauth, H., & Altenmüller, E. (2004c). Quantification of focal dystonia in pianists
using scale analysis. Movement Disorders 19(2), 171–180.
Jankovic, J., & Ashoori, A. (2008). Movement disorders in musicians. Movement Disorders 23(14),
1957–1965.
Jankovic, J., & Shale, H. (1989). Dystonia in musicians. Seminars in Neurology 9, 131–135.
Kieslinger, K., Holler, Y., Bergmann, J., Golaszewski, S., & Staffen, W. (2013). Successful treatment
of musician’s dystonia using repetitive transcranial magnetic stimulation. Clinical Neurology and
Neurosurgery 115(9), 1871–1872.
Killian, O., McGovern, E. M., Beck, R., Beiser, I., Narasimham, S., Quinlivan, B., … Reilly, R. B.
(2017). Practice does not make perfect: Temporal discrimination in musicians with and without
dystonia. Movement Disorders 32(2), 1791–1792.
Kimberley, T. J., Schmidt, R. L. S., Chen, M., Dykstra, D. D., & Buetefisch, C. M. (2015). Mixed
effectiveness of rTMS and retraining in the treatment of focal hand dystonia. Frontiers in Human
Neuroscience 9. Retrieved from https://doi.org/10.3389/fnhum.2015.00385
Kluger, B., Triolo, P., Jones, W., & Jankovic, J. (2015). The therapeutic potential of cannabinoids for
movement disorders. Movement Disorders 30(3), 313–327.
Koppel, B. S., Brust, J. C. M., Fife, T., Bronstein, J., Youssof, S., Gronseth, G., & Gloss, D. (2014).
Systematic review: Efficacy and safety of medical marijuana in selected neurologic disorders.
Report of the Guideline Development Subcommittee of the American Academy of Neurology.
Neurology 82, 1556–1563.
Kuhn, Y. A., Keller, M., Lauber, B., & Taube, W. (2018). Surround inhibition can instantly be
modulated by changing the attentional focus. Scientific Reports 8(1). doi:10.1038/s41598-017-
19077-0
Lederman, R. J. (1991). Focal dystonia in instrumentalists: Clinical features. Medical Problems of
Performing Artists 6, 132–136.
Lederman, R. J. (2001). Embouchure problems in brass instrumentalists. Medical Problems of
Performing Artists 16, 53–57.
Lee, A., Eich, C., Ioannou, C. I., & Altenmüller, E. (2015a). Life satisfaction of musicians with focal
dystonia. Occupational Medicine 65(5) 380–385.
Lee, A., Tominaga, K., Furuya, S., Miyazaki, F., & Altenmüller, E. (2015b). Electrophysiological
characteristics of task-specific tremor in 22 instrumentalists. Journal of Neural Transmission
122(3), 393–401.
Lee, A., Voget, J., Furuya, S., Morise, M., & Altenmüller, E. (2016). Quantification of sound
instability in embouchure tremor based on the time-varying fundamental frequency. Journal of
Neural Transmission 123(5), 515–521.
Leijnse, J. N. A. L., Hallett, M., & Sonneveld, G. J. (2015). A multifactorial conceptual model of
peripheral neuromusculoskeletal predisposing factors in task-specific focal hand dystonia in
musicians: Etiologic and therapeutic implications. Biological Cybernetics 109(1), 109–123.
Lim, V. K., & Altenmüller, E. (2003). Musicians’ cramp: Instrumental and gender differences.
Medical Problems of Performing Artists 18, 21–26.
Marsden, C. D., & Sheehy, M. P. (1990). Writer’s cramp. Trends in Neurosciences 13(4), 148–153.
Meck, W. H., Penney, T. B., & Pouthas, V. (2008). Cortico-striatal representation of time in animals
and humans. Current Opinion in Neurobiology 18(2), 145–152.
Obeso, I., Cerasa, A., & Quattrone, A. (2016). The effectiveness of transcranial brain stimulation in
improving clinical signs of hyperkinetic movement disorders. Frontiers in Neuroscience 9.
Retrieved from https://doi.org/10.3389/fnins.2015.00486
Oga, T., Honda, M., Toma, K., Murase, N., Okada, T., Hanakawa, T., … Shibasaki, H. (2002).
Abnormal cortical mechanisms of voluntary muscle relaxation in patients with writer’s cramp: An
fMRI study. Brain 125(4), 895–903.
Pastor, M. A., Macaluso, E., Day, B. L., & Frackowiak, R. S. J. (2008). Putaminal activity is related
to perceptual certainty. NeuroImage 41(1), 123–129.
Paulig, J., Jabusch, H. C., Grossbach, M., Boullet, L., & Altenmüller, E. (2014). Sensory trick
phenomenon improves motor control in pianists with dystonia: Prognostic value of glove-effect.
Frontiers in Psychology 5. Retrieved from https://doi.org/10.3389/fpsyg.2014.01012
Peller, M., Zeuner, K. E., Munchau, A., Quartarone, A., Weiss, M., Knutzen, A., … Siebner, H. R.
(2006). The basal ganglia are hyperactive during the discrimination of tactile stimuli in writer’s
cramp. Brain 129(10), 2697–708.
Pesenti, A., Barbieri, S., & Priori, A. (2004). Limb immobilization for occupational dystonia: A
possible alternative treatment for selected patients. Advances in Neurology 94, 247–54.
Peterson, D. A., Berque, P., Jabusch, H. C., Altenmüller, E., & Frucht, S. J. (2013). Rating scales for
musician’s dystonia: The state of the art. Neurology 81(6), 589–598.
Peterson, D. A., & Sejnowski, T. J. (2017). A dynamic circuit hypothesis for the pathogenesis of
blepharospasm. Frontiers in Computational Neuroscience 11. Retrieved from
https://doi.org/10.3389/fncom.2017.00011
Peterson, D. A., Sejnowski, T. J., & Poizner, H. (2010). Convergent evidence for abnormal striatal
synaptic plasticity in dystonia. Neurobiology of Disease 37, 558–573.
Priori, A., Pesenti, A., Cappellari, A., Scarlato, G., & Barbieri, S. (2001). Limb immobilization for
the treatment of focal occupational dystonia. Neurology 57(3), 405–409.
Richardson, S. P., Altenmüller, E., Alter, K., Alterman, R. L., Chen, R., Frucht, S., … Hallett, M.
(2017). Research priorities in limb and task-specific dystonias. Frontiers in Neurology 8. Retrieved
from https://doi.org/10.3389/fneur.2017.00170
Ridding, M. C., Sheean, G., Rothwell, J. C., Inzelberg, R., & Kujirai, T. (1995). Changes in the
balance between motor cortical excitation and inhibition in focal, task specific dystonia. Journal of
Neurology, Neurosurgery & Psychiatry 59(5), 493–498.
Rosenkranz, K. (2010). Plasticity and intracortical inhibition in dystonia: Methodological
reconsiderations. Brain 133(6), e146.
Rosenkranz, K., Williamon, A., Butler, K., Cordivari, C., Lees, A. J., & Rothwell, J. C. (2005).
Pathophysiological differences between musician’s dystonia and writer’s cramp. Brain 128(4),
918–931.
Rosset-Llobet, J., Candia, V., Molas, S. F. I., Cubells, D., & Pascual-Leone, A. (2009). The challenge
of diagnosing focal hand dystonia in musicians. European Journal of Neurology 16(7), 864–869.
Rosset-Llobet, J., Fabregas-Molas, S., & Pascual-Leone, A. (2015). Effect of transcranial direct
current stimulation on neurorehabilitation of task-specific dystonia: A double-blind, randomized
clinical trial. Medical Problems of Performing Artists 30, 178–184.
Ruiz, M. H., Senghaas, P., Grossbach, M., Jabusch, H. C., Bangert, M., Hummel, F., … Altenmüller,
E. (2009). Defective inhibition and inter-regional phase synchronization in pianists with
musician’s dystonia: An EEG study. Human Brain Mapping 30(8), 2689–2700.
Sakai, N. (2006). Slow down exercise for the treatment of focal hand dystonia in pianists. Medical
Problems of Performing Artists 21, 25–28.
Satoh, M., Narita, M., & Tomimoto, H. (2011). Three cases of focal embouchure dystonia:
Classifications and successful therapy using a dental splint. European Neurology 66(2), 85–90.
Schmidt, A., Jabusch, H. C., Altenmüller, E., Hagenah, J., Bruggemann, N., Hedrich, K., … Klein, C.
(2006). Dominantly transmitted focal dystonia in families of patients with musician’s cramp.
Neurology 67(4), 691–693.
Schmidt, A., Jabusch, H. C., Altenmüller, E., Hagenah, J., Bruggemann, N., Lohmann, K., … Klein,
C. (2009). Etiology of musician’s dystonia: Familial or environmental? Neurology 72(14), 1248–
1254.
Schuele, S., & Lederman, R. J. (2004a). Long-term outcome of focal dystonia in string
instrumentalists. Movement Disorders 19(1), 43–48.
Schuele, S. U., & Lederman, R. J. (2004b). Occupational disorders in instrumental musicians.
Medical Problems of Performing Artists 19, 123–128.
Schultz, W., & Dickinson, A. (2000). Neuronal coding of prediction errors. Annual Review of
Neuroscience 23, 473–500.
Sohn, Y. H., & Hallett, M. (2004). Disturbed surround inhibition in focal hand dystonia. Annals of
Neurology 56(4), 595–599.
Spector, J. T., & Brandfonbrener, A. G. (2005). A new method for quantification of musician’s
dystonia: The frequency of abnormal movements scale. Medical Problems of Performing Artists
20, 157–162.
Spector, J. T., & Brandfonbrener, A. G. (2007). Methods of evaluation of musician’s dystonia:
Critique of measurement tools. Movement Disorders 22(3), 309–312.
Tamura, Y., Ueki, Y., Lin, P., Vorbach, S., Mima, T., Kakigi, R., & Hallett, M. (2009). Disordered
plasticity in the primary somatosensory cortex in focal hand dystonia. Brain 132(3), 749–755.
Tubiana, R. (2000). Musician’s focal dystonia. In R. Tubiana & P. C. Amadio (Eds.), Medical
problems of the instrumentalist musician. London and Malden, MA: Martin Dunitz.
Tubiana, R. (2003). Musician’s focal dystonia. Hand Clinics 19, 303–308.
Tubiana, R., & Chamagne, P. (1993). Medical professional problems of the upper-limb on musicians.
Bulletin de l’Académie Nationale de Médecine 177, 203–216.
Valdes, K., Naughton, N., & Algar, L. (2014). Sensorimotor interventions and assessments for the
hand and wrist: A scoping review. Journal of Hand Therapy 27(4), 272–286.
van der Steen, M. C., van Vugt, F. T., Keller, P. E., & Altenmüller, E. (2014). Basic timing abilities
stay intact in patients with musician’s dystonia. PloS ONE 9(3), e92906.
Vecchio, M., Malaguarnera, G., Giordano, M., Malaguarnera, M., Volti, G. L., Galvano, F., …
Malaguarnera, M. (2012). A musician’s dystonia. Lancet 379, 2116.
Williams, L. J., Butler, J. S., Molloy, A., McGovern, E., Beiser, I., Kimmich, O., … Reilly, R. B.
(2015). Young women do it better: Sexual dimorphism in temporal discrimination. Frontiers in
Neurology 6. Retrieved from https://doi.org/10.3389/fneur.2015.00160
Zeuner, K. E., Knutzen, A., Granert, O., Gotz, J., Wolff, S., Jansen, O., … Witt, K. (2015). Increased
volume and impaired function: The role of the basal ganglia in writer’s cramp. Brain and Behavior
5(2), e00301.
Zeuner, K. E., & Molloy, F. M. (2008). Abnormal reorganization in focal hand dystonia: Sensory and
motor training programs to retrain cortical function. Neurorehabilitation 23(1), 43–53.
SECTION VIII
T HE F U T U R E
CHAPT E R 33
MI C H A E L H . T H A U T A N D D O N A L D A . H O D G E S
I : T F
P
T N H F
R
This last and short chapter, therefore, will try to point to some significant
developments in the past regarding music and brain research that may hold
potential for future developments. We will also try to point to areas for
future research efforts, that in our understanding are underserved and in
need of more and new research strategies. And with due caution the chapter
will try to throw in some ideas of nonlinear leaps, motivated by so many
innovative accounts provided in the chapters of this book.
C
The ten horizons laid out in this chapter for potential future research should
not be read as predictors—see the introduction to this chapter—rather than
as areas of potential growth for the advancement of knowledge in music as
a language the human brain has created. The chapters in this book are trying
to present the most current state of knowledge about the functions and
operations in the brain associated with music. As we think and create music
subjectively it becomes an object of discovery and scientific research.
Therefore, the final mission of the book is to serve as a springboard for
future brain research in music, providing knowledge from very different
angles of scientific discovery to help shape new trajectories of pursuit to
understand music and our brains, that is, as the brain creates and engages in
music it is changed by engaging in music.
R
Berlyne, D. E. (1971). Aesthetics and psychobiology. New York: Appleton-Century-Crofts.
Conard, N. J., Malina, M., & Muenzel, S. C. (2009). New flutes document the earliest musical
tradition in southwestern Germany. Nature 460, 737–740.
Darwin, C. (1871). The descent of man and selection in relation to sex. London: John Murray.
Gruhn, W., & Rauscher, F. (Eds.). (2007). Neurosciences in Music Pedagogy. New York: Nova
Biomedical Books.
Guptill, C. A. (2011). The lived experience of professional musicians with playing-related injuries: A
phenomenological inquiry. Medical Problems of Performing Artists 26(2), 84–95.
Henechowicz, T., Chen, J., Cohen, L. G., & Thaut, M. H. (2018). Prevalence of BDNF
polymorphism in musicians: Evidence for compensatory motor learning strategies in music?
Proceedings of the Society for Neuroscience. In press.
Hodges, D. (2013). Music listeners, philosophers, and researchers. Physics of Life Reviews 10(3),
275–276.
Huron, D. (2016). Aesthetics. In S. Hallan, I. Cross, & M. Thaut (Eds.), The Oxford handbook of
music psychology (2nd ed., pp. 233–245). Oxford: Oxford University Press.
Kant, I. (1790). Kritik der Urteilskraft. Berlin: Lagarde.
Koshimori, Y., Strafella, A., Valli, M., Sharma, V., Cho, S. S., Houle, S., & Thaut, M. H. (2018).
Motor synchronization to rhythmic auditory stimulation (RAS) attenuates dopamine response in
the ventral striatum in young healthy adults. Proceedings of the Society for Neuroscience 670, 15.
Thaut, M. H. (2005). Rhythm, music, and the brain. New York: Taylor & Francis.
Thaut, M. H., Schweizer, T., Leggieri, M., Churchill, N., Fornazzari, L., & Fischer, C. (2018). Neural
basis for potential preservation of musical memory and effects on functional intra network
connectivity in early Alzheimer’s disease and mild cognitive dysfunction. Proceedings of the
Society for Neuroscience 741, 12.
Wittgenstein, L. (1967). Lectures and conversations on aesthetics, psychology, and religious belief.
Ed. C. Barret. Berkeley, CA: University of California Press.
I
A
Abrams, D. A. et al. 397
absolute pitch (AP) 240, 443, 461, 471
and autism 681–90, 683
abstract-feature mismatch negativity (afMMN) 196
academic achievement, effect of musical training 654–6
acetylcholine (Ach) receptors 350
acquired amusia 765–70
acquired brain injury (ABI) 700–1
Action Research Arm Test (ARAT) 701
Action Simulation for Auditory Perception (ASAP) model 225
active vs. passive exposure 32
activities of daily living (ADL) 745
adaptive convergent sequence evolution 443
Adhikari, B. M. et al. 509–10
adjacency matrices (Aij) 133
adrenocorticotropic hormone (ACTH) 342, 741
aesthetic experiences 366, 367–74, 371, 373
brain structures 375–6
emotions 73–4, 75–8, 80
future challenges 381–2
pleasure 105 see also pleasure in music
studies 377–81
aesthetic judgment 286, 290, 374, 376
affect, definitions 286–7
affective functions, neurologic music therapy 745
aging 623–4, 636–7
brain mechanisms 629–31
cognition in musicians 624–6
emotions and well-being 628–9, 635–6
and language 633–4
and listening to music 627–8
and memory 632–3
and motor functions 634–5, 705
and musical training 626–7, 656–7
pathologies 631–3, 635–6
singing therapy 725
Aitken Dunham, D. J. 719
Akan language 570
Albanese, A. et al. 776
allergic skin responses 348
Alluri, V. and Toiviainen, P. 377
Alm, P. A. 722–3
alpha-melanocyte-stimulating hormone 741
Altenmüller, E., Finger, S., and Boller, F. 3–4
Alzheimer’s disease (AD) 254, 625, 631–3, 634, 635–6
cognitive remediation 740
NMT 751
Ammirante, P. et al. 225
amplitude envelopes 151, 157, 158, 190–1
amplitude variation 153–5, 154
Amunts, K. et al. 202
amusia 5, 25, 400, 443, 461, 556
acquired 765–70
congenital 760–5, 769–70
and imagery 523–4
and language disorders 581
phenotypes 770–1
pitch-based 761–3
amygdala 102–4, 103, 299, 300, 301, 446
Anderson, B., Southern, B. D., and Powers, R. E. 217
angular gyrus 518
anhedonia 768–9
animals
brain plasticity 431
dystonia 477
rhythm perception 175–6
anterior cingulate cortex (ACC) 106, 300, 301, 490, 513–14
anterior pituitary 341–5
anticipatory musical imagery 526
anxiety 752
aphasia 393, 400, 523, 720–1
apraxia of speech (AOS) 718–20
aptitude 427–8
evolution 443–5
exceptional musicianship 671–4, 681–90
genetic influences 555–6
genetic markers 445–6
as genetic trait 440–3
innateness 808
and musical training 658–60, 659
without training 25
archaeological findings 29–30
arcuate fasciculus 428, 471, 634
arginine vasopressin (AV) 339, 340–1, 342
Arm Paresis Score 701
Arnal, L. et al. 407–8
arousal 286, 601
enhancing musical memory 246–7
art
archaeological findings 29–30
neuroaesthetics of 367–9
associative coding of emotions 288
Associative Mood and Memory Training (AMMT) 749
assortativity 129
asthma 727
asymmetric sampling in time hypothesis 406
atlas/region-of-interest (ROI) based networks 130, 131
attention 263–4
to pitch and harmonicity 266–7
selection and filtering 266
temporal 267–9
theories 264–6
training 747–8, 749–50
attention deficit (hyperactivity) disorder (ADD/ADHD) 607
audio-visual bounce effect (ABE) 157
audio-visual integration research 156–8
auditory areas, primary (A1) and secondary (A2) 464
auditory association areas (AAs) 464
auditory belt 221
auditory brainstem response (ABR) 550–1, 553
auditory core 221
auditory cortex 394, 428
auditory domain 470
auditory feedback 464
auditory-frontal networks 93–5
auditory gestalts 464
auditory-limbic networks 102–5
auditory–motor integration circuit 469
auditory-motor networks 95–102, 96
auditory parabelt 221
auditory pathways 90–3, 90, 217
auditory perception 153–9
autistic children 678–81, 678, 680
development 675–7, 677, 678–81, 678, 680
Auditory Perception Training (APT) 747
auditory processing 216–17
auditory sensory memory (ASM) 192–4
auditory sensory processing 517
auditory stream segregation 266
auditory system 188–94
Auditory Verbal Learning Test 751
Auerbach, S. 470
autism spectrum disorder (ASD) 581, 607, 705–6
and absolute pitch (AP) 681–90, 683
and development of auditory perception 678–81, 678, 680
and exceptional musical abilities 671–4, 681–90
autobiographical memory 245, 248, 302, 633
axons 27
Ayotte, J. et al. 766
B
Baba, A. et al. 527
Babbitt, M. 569
Bach, J. S. 197
background music 252–4, 627–8, 633
backward playing of notes 151
Baddeley, A. D. 531
Baddeley, A. D. and Logie, R. H. 531
Baer, L. et al. 426
Bailey, J. A. et al. 426
Baird, A. et al. 632–3
Balbag, M. A. et al. 626
Balkwill, L. L. and Thompson, W. F. 45
Bangert, M. and Altenmüller, E. 469
Bantu languages 570
Barabási, A.-L. 125
Baram, Y. and Miller, A. 704
Bard, P. 189
basal ganglia (BG) 223, 301, 426, 467, 473–4, 696, 716
rhythm perception 168–70
basal ganglia-thalamo-cortical network 100–2
bats 443
beat 165–6, 166, 187
continuation 101
extraction 592
perception 166–72, 177, 178–9
somatosensory perception 225
visuomotor perception 224
see also rhythm
Beat Alignment Task (BAT) 596
beat finding disorder 763–5
bebop 569
Beethoven, L. v, 145, 159
behavioral changes in musical training 272–3
behavioral studies, imagery and perception 522–3
Beisteiner, R. 5
Bellaire, K. et al. 718
Benedek, M. et al. 278
Bengtsson, S. L. et al. 27, 471, 488–9, 493–4, 498
Berger, H. 5, 268
Berkowitz, A. L. and Ansari, D. 489–90, 492, 518
Berlyne, D. E. 71
Berns, G. S. et al. 377
beta-endorphin 741 see also β-endorphins
Bhide, A. et al. 728
biological approaches vs. cultural approaches 19–22, 461–2
biological restrictions 21
bipolar disorder 451
birds, rhythm perception 175–6
birdsong 67–73, 81, 393–4, 441, 443, 444–5, 447, 449
Blondel, V. D. et al. 129
Blood, A. J. and Zatorre, R. J. 292, 377
blood-flow studies 292, 293, 296, 304
blue notes 213
Bogert, B. et al. 375
Bonacina, S. et al. 728
bonding 66
Boone, D. R. et al. 726–7
bootstrapping problem 573, 742
Bottiroli, S. et al. 628
Box and Block Test (BBT) 701
brain, co-evolution with hands 21
brain damage
effect on imagery and perception 523–4
pathological aging 631–2, 635–6
pitch perception 191
and rhythm processing 166, 169–70
brain-derived neurotrophic factor (BDNF) 753
brain development see development; plasticity
brain imaging studies of imagery and perception 525–6
brain injuries 700–1
cognitive remediation 740
singing therapy 718
brain scanning, technological advances 365
brainstem 301
brainstem reflex 289, 302
Brattico, E. 367, 377
Brattico, E. and Pearce, M. T. 377
Brattico, E. et al. 372, 373, 374, 375–6, 377
Brazilian music 603
BRECVEMAC framework 289–90, 294, 299, 301 see also ICINAS-BRECVEMAC
Bregman, A. S. 266
Brendal, B. and Ziegler, W. 719
Broadbent, D. E. 266
Broca’s aphasia 720–1
Broca’s area 202, 243, 277, 397, 400
Brotons, M. and Koger, S. M. 634
Brown, S. and Jordania, J. 30
Brown, S. et al. 377, 506–7, 513
Brust, J. 3
Bugos, J. A. et al. 627
C
Caetano, G. and Jousmäki, V. 220
Cage, J. 21
Cameron, D. J. et al. 49
Cappiani, L. 4
Carnātic music, tonality 48
Cason, N. et al. 404, 606
caudal subdivision 216
cave paintings 29–30
CBB (culture–behavior–brain) loop model 20, 20
CDK5 pathway 450
cello
harmonics 151, 152
multisensory perception 221
central activation 72
central nervous system (CNS) 717
central sulcus 465, 472
centrality analysis 128
cerebellum 100, 106, 243, 300, 301, 427, 467, 490, 696
rhythm perception 168
cerebral palsy (CP) 698, 699, 700, 706
Chan, M. F. et al. 629
Chapin, H. et al. 377
Chatterjee, A. and Vartanian, O. 381
Chen, J. K.-C. et al. 723–4
Cheung, V. et al. 201
Chiao, J. 20
Chikazoe, J. et al. 293
childhood apraxia of speech (CAS) 719–20
children see infants
chills, response to music 335, 380
Chinese IDyOM model 53–8
Chinese music, phrase boundary perception 49
Chobert, J. et al. 553
cholinergic systems 350
Chomsky, N. 392, 808
Chomsky hierarchy 196
Chong, H. J. et al. 706
Chopin 601
chord functions 200
chord transitions 194, 197–9, 198, 201–3
chords 147–8
CHRNA9 gene 446
chromosomes 445–6, 448
chronic obstructive pulmonary disease (COPD) 727
cingulate motor area (CMA) 465, 466–7, 515–16
Cirelli, L. K. et al. 604, 605
clapping, synchronous 605 see also synchronous movement
clarinet, harmonics 148–9, 149, 150
classical music 337, 549
clinical research model 744
Closure Positive Shift 49
clusters 127–8
cluttering 722
cochlea 188–9, 192
cochlear 446
cochlear implant (CI) users 147, 723–4
cognition
of creative musicians 278
and listening to music 627–8
and musical expertise 624–6
neural mechanisms 521
neurologic music therapy (NMT) 747–53
and neurological disorders 738
and short-term training 626–7
and training 647–50, 654–7, 659
cognitive control 271
cognitive decline in aging 623–4
cognitive dysfunction, music therapy 742–3
cognitive functions, neurologic music therapy 745
cognitive goal appraisal 290
cognitive integration 213–16
cognitive neuroscience of music 365, 369–74, 371, 373
brain structures 375–6
future challenges 381–2
studies 377–81
cognitive remediation/rehabilitation (CR) 738–40, 753
vs. Transformational Design Model of NMT 746
cognitive reserve, in aging 629–31
Cohen, J. 250
Cohen, N. S. and Masse, R. 718
Colcord, R. D. and Adams, M. R. 723
Coleman, O. 75–6
Common Patterns 30–1
community detection procedures 130
community structure analysis 129, 130, 131
comparative approach 393
compensatory approach to cognitive remediation 739
complexity 70, 248
composers of music 450–1
instructions 146
conductors 468
congenital amusia 760–5, 769–70
connectivity 122–3, 127
and improvisation 499–500
see also network metrics
connectivity analysis 133
consciousness disorders 750
Conserved Universals 30–1
contagion, emotional 290, 303
convergent analysis 449–50
Cook, N. 58
Cope, T. E. et al. 704
copy number variation (CNV) analysis 446
corpus callosum (CC) 421, 422, 425, 426, 471, 473
Corrigall, K. A. et al. 461
cortical sound processing 547–50
cortico-cerebellar network 97–100, 98, 102
corticospinal tract 471
corticotrophin-releasing hormone (CRH) 741
corticotropin releasing factor (CRF) 342
cortisol 341–5, 741
Costa-Giomi, E. 652
Cox, A. 530
cramp 475, 477, 777, 785–6 see also dystonia
Crasta, J. E. et al. 696
creativity 263, 274–5, 278–9
biological basis 450–1
definition 277–8
generative see improvisation
in musical improvisation 275–6
neural correlates 497–9
neuroimaging studies 276–7
personality and cognitive profiles 278
Critchley, M. and Henson, R. 3
cross-cultural research 32–3, 43–4, 58–60
memory 50–1
preferences 46–7
recognition of emotions 45–6
structural features 47–50
Cross, I. 571
Cross, I. and Morley, I. 19
crying, response to music 76–7
Cue Redundancy Model (CRM) 45, 48, 58
cultural approaches 22–32
vs. biological approaches 19–22, 461–2
cultural distance 44, 51–60
cultural distance hypothesis 32, 33, 52
culturally contextualized behaviors (CC-Behavior) 20
culturally voluntary behaviors (CV-Behavior) 20
Cupchik, G. C. et al. 367
cyclic form 568
cystic fibrosis 727
cytokines 348
D
Dai, L. et al. 340
Dalla Bella, S. et al. 601
dance 533–4
beat finding disorders 763–5
universality 30
see also synchronous movement
dance movement therapy (DMT) 336
Darwin, C. 718, 808
data-driven approaches to creativity 277–8
DC-EEG (direct current EEG) 5
de-expertise 475–6
de Manzano, O. and Ullén, F. 492–5, 513
Deacon, T. W. 571–2
deafness 189, 581, 723–4
deep brain stimulation (DBS), for musician’s dystonia (MD) 793
Default Mode Network 136–7
Default Network 277
Degé, F. and Schwarzer, G. 653
degenerative brain disorders, cognitive remediation 740
degenerative movement disorders 701–5
degree distribution 125–7
Dehaene-Lambertz, G. 579
deliberate practice 460–1
dementia 448, 631–3, 635–6
Demorest, S. M. and Morrison, S. J. 52
Demorest, S. M. et al. 51
dendrites 463
Dennett, D. C. 279
DePape, A.-M. R. et al. 681
depression 339, 342, 752
Desai, V. and Mishra, P. 726
Deutsch, D. 265
Deutsch, D. et al. 571, 582–3
development
of auditory perception 675–7, 677
autistic children 678–81, 678, 680
of the brain, effect of training 422–7, 426, 551–3, 557
co-development of language and music 576–8, 577
disorders
and amusia 770
in music and language 580–1
rehabilitation for 705–6
and rhythm 605–7
of language perception 574–6, 575
developmental coordination disorder (DCD) 607
Developmental Speech and Language Training Through Music (DSLM) 717
developmental stress hypothesis 68, 81
Diagnostic and Statistical Manual of Mental Disorders–5 (DSM–5) 672–3, 722
Diamond, A. 263
Different Trains (Reich) 570–1
diffusion tensor imaging (DTI) 5, 380, 421, 425, 470, 471
Ding, N. et al. 406
direct current stimulation (DCS), for dystonia 784
Directions into Velocities of Articulators (DIVA) 717
discrimination of sounds 394–6
disordered timing 786–7
disorders, language and music 580–1 see also development; neurological disorders
Distorted Tunes Test (DTT) 442
Divergent Thinking Test 278
Dobos, R. 29
Dobzhansky, T. 19
‘dock-in’ model of emotional recognition 45–6, 58
Doidge, N. 27
dolphins 443
Donnay, G. F. et al. 502–4, 517
dopamine (DA) 102, 104–5, 246, 334–6, 337, 431, 446, 448–9, 450–1, 451, 631, 741
and rhythm perception 170
dopamine imaging 807
dopamine signaling 795
dorsal cochlear nucleus (DCN) 338
dorsal pathways 217, 218
dorsal premotor cortex (PMd) 276, 488, 489, 490, 494–5, 515
dorsolateral prefrontal cortex (DLPFC) 276, 488, 514–15
Dowling, W. J. and Harwood, D. L. 148, 288
Down syndrome 699, 700
Drake, C. and Bertrand, D. 30
DRD2 polymorphism 716
DRD4 gene 450
drumming, synchronized 603–4 see also synchronous movement
DSM-5 672–3, 722
dual-pathway model 94
DUSP1 gene 447, 449–50
Dynamic Attending Theory 268
dysarthria 717–18
dyslexia 580, 606–7, 653
and amusia 770
neurologic music therapy (NMT) for 727–8
rhythm processing 173
dystonia
embouchure 477, 777, 779, 779
focal 475, 476–8, 776–7, 778, 784–6
focal hand (FHD) 777, 778, 779, 784, 785–6
musician’s (MD) 475–8, 479, 700, 777–82, 778, 779
future directions 791–6
pathogenic theory 787–90, 788, 794–6
pathophysiology 784–7, 794
and plasticity 787
treatments 782–4, 783, 791–3
types 776–7
Dystonia Study Group 780, 781–2
E
early right anterior negativity (ERAN) 199–200, 266, 277, 375, 548
earworms 528
East African music, rhythm perception 49
echoic memory 192–3
echolalia 679–80
echolocation 443
Edgren, J. 5
education in music 46, 442, 809
EEG (encephalography) 5, 376
improvisation 509–12
Einarson, K. M. and Trainor, L. J. 596
Eitan, Z. and Granot, R. Y. 530
El Haj, M. et al. 633
El Sistema music training 552, 554–5
electrophysical methods 468
electrophysiology, studies of imagery and perception 524
Eley, R. and Gorman, D. 727
Ellis, R. J. et al. 273
Elmer, S. et al. 402
embodied musical imagery 529–34
embouchure dystonia 477, 777, 779, 779
emotions 79, 285–6, 465
aesthetic 73–4, 75–8, 80
and aging 628–9, 635–6
auditory processing 190
cross-cultural studies on recognition 45–6
cultural specificity 32
definition 286
discreteness 293
empirical studies 291–304, 306–23
enhancing musical memory 246–7
induction 289–90, 297–300, 301–2
vs. perception 287, 292, 296
layers of expression 288–9, 288
neural responses 104
neurologic music therapy (NMT) 747–53
perception 287–9, 300
vs. induction 287, 292, 296
psychological mechanisms 287–90
regulation using cognitive remediation 741
and rhythm 600–1
sensitivity to 402
specific brain regions 293–6
universality 30
visual displays 213
empirical aesthetics 366, 367–74, 371, 373
brain structures 375–6
future challenges 381–2
studies 377–81
endocrine responses 339–45
endogenous cannabinoid receptors 793
endogenous opioid systems (EOSs) 336–8
enjoyment of music see pleasure in music
entrainment 407
entropy 277–8, 599–600
environment, influence on genetic expression 22–3, 23
environmental effects, vs. genetic effects 440
epilepsy 5, 766
episodic buffer 241
episodic memory 238, 245, 248, 290, 303
ERAN (early right anterior negativity) 199–200, 266, 277, 375, 548
Ericsson, K. A. et al. 460
error-driven learning 430
Escoffier, N. et al. 300
esthetic experiences see aesthetic experiences
esthetic judgment see aesthetic judgment
evaluative conditioning (EC) 290, 303
event-related desynchronizations (ERDs) 249–50
event-related potentials (ERPs) 5, 192, 375, 594
in infants 28
in monkeys 176
phrase boundary perception 49
tonality perception 48
visual rhythm perception 178
evolution
of brain and hands 21
learned song 74–5
of musical aptitude 443–5
Ewe language 570
Executive Control Network 277
executive functions (EFs) 255, 263, 269–70
in aging 657
and musical training 554–5, 649–50
training 748
transfer 269–74
expectations 267, 290, 303–4, 398
experience-dependent processes 25–6, 26
experience-expectant processes 25–6, 26, 26–7
expert performance 22–3, 23
expertise see musicians
explicit memory system 238, 244, 245–6, 255
exposure see familiarity
extinction 265
extracurricular activities 654–5
F
faculties of the mind 4, 4
Fahn-Marsden (FM) scale 782
falling, risk of 702, 705
familiarity
and cross-cultural research 32, 33
infant responses 28–9
and music memory 50
network connections 106
and preference 46–7, 58, 240
scale perception 47
far transfer 271–74, 646, 649
Farrugia, N. et al. 528
Fava, E. et al. 579
Fawcett, C. and Tunçgenç, B. 604–5
feature integration theory 264–5
feedback 430, 464, 465, 699, 717
feedforward and feedback connections 219
Fernández-Miranda, J. C. et al. 95
Ferreri, L. et al. 252, 628
ferrets 431
fetuses, response to musical stimulation 27–8
Fifth Symphony (Beethoven) 145, 159
filtering, attentional 266
flat tones 153–5, 154, 155
Flaugnacco, E. et al. 606–7, 728
fluency disorders 721–3
fMRI (functional magnetic resonance imaging) 5, 365, 469
of improvisation 488–506
network-based approaches 132–3, 134, 135
phrase boundary perception 49
rhythm perception 168–9
focal dystonias 475, 476–8, 776–7, 778, 784–6
focal hand dystonia (FHD) 777, 778, 779, 784, 785–6
Fodor, J. 392
foot tapping, neural basis 97
forgetfulness of self 77
formal institutional training in improvisation (FITI) 511–12
FOS gene 447, 449–50
Foster, N. A. and Valentine, E. R. 633
Foster, N. E. and Zatorre, R. J. 422
Fox, N. A. et al. 220
FOXP1 gene 445
FOXP2 gene 443–5, 450
fractional anisotropy (FA) 426, 470, 473
Fractionating Emotional Systems (FES) 45
François, C. and Schön, D. 403
free form jazz 75–6
free response generation 492–4
freestyle rap 276, 495–7
freezing episodes 703
French horn, harmonics 155
frequency following response (FFR) 190, 550–1
frequency range, biological restrictions 21
frequency tagging approach 171
Friederici, A. D. 201
Fritz, T. 45–6, 58
frontal cortex 93–5, 301
frontal gyrus 300
Früholz, S. et al. 377
Fu, Q.-J. et al. 724
Fujioka, T. et al. 172
functional near-infrared spectroscopy (fNIRS) 724
functions of music 31
fusiform face area (FFA) 770
fusiform gyrus 488–9
future of brain research in music 805–11
G
Gaab, N. et al. 239
Gagaku music 568
Gall, F. J. 4
Galvan, A. 25
gamelan music 568
garden path sentences 398, 582
Gaser, C. and Schlaug, G. 472
Gaston, E. 19
GATA2 gene 448, 448
Gaver, W. 671–3
gender ratios for musician’s dystonia (MD) 788–9, 789
gene–maturation–environment interactions 424
Generative Syntax Model (GSM) 197
Generative Theory of Tonal Music (GTTM) 197
genetic effects vs. environmental effects 440
genetic influences 20
on musical behavior 22–3, 23
genomic approaches 5, 439, 452
to aptitude 440–3, 445–6
convergent analysis 449–50
creativity 450–1
evolution 443–5
effect of music on transcriptome 447–9
genre 137, 568
German music 603
Gerry, D. W. et al. 595–6
Gervain, J. 580
Gestalt formation 192–4
gestures 223
Ghitza, O. 407
Gillespie, L. D. et al. 705
Gilmore, S. 225
Giraud, A. L. and Poeppel, D. 406
Glennie, Dame E. 189
Global Dystonia Scale (GDS) 782
globus pallidus internum (GPi) 793
Glover, H. et al. 723
Gooding, L. et al. 625
Goswami, U. 405
GPR98 gene 444
grahabēdham modulation 48
Grahn, J. A. and Brett, M. 169
Granert, O. et al. 473, 474
graph theory 123, 125, 131, 137
Grau-Sánchez, J. et al. 701
gray matter (GM) 420–2, 426, 429
density 472–3
pianists 474
GRIN2B gene 450
groove 169
group drumming 340, 344, 347, 348
Guenther, F. H. 717
Guerrieri, M. 145
guided imagery and music (GIM) therapy 344
H
Habib, M. et al. 606, 728
Habibi, A. et al. 273
hallucinations, musical 526–7
Halpern, A. R. 523
Halpern, A. R. and Müllensiefen, D. 245
Halpern, A. R. and O’Connor, M. G. 244
Halpern, A. R. et al. 531
Hambrick, D. Z. et al. 444
Han, S. and Ma, Y. 20
hand dystonia 477
handicap principle 67–9, 81
hands, co-evolution with brain 21
Hanna-Pladdy, B. and MacKay, A. 625
Hannon, E. E. and Trainor, L. J. 28
Hanslick, E. 79, 80, 370
harmonic dependencies 197–9, 198
harmonic expectancy violations 93
harmonicity, attention to 266–7
harmonics 146, 148–52, 149, 155
Harmony project 553
Hawaiian language 570
head movements, and rhythm perception 225–6
Healey, E. C. et al. 723
hearing mechanism 22
hedonic reversal 73–4, 80
Helfrich-Miller, K. R. 720
hemi-neglect 751–2
hemispheres 293
asymmetry 93, 106
hemispheric specialization 217
hemodynamic responses 248–9
Henson, R. 3
heritability 428, 442, 555–6
Heschl’s gyrus (HG) 420, 423, 464, 470
damage to 523
Hidalgo, C. et al. 405
hierarchical syntactic structures 196–203
Hilton, M. P. et al. 725
Hinton, G. et al. 71
hippocampus 104, 301, 465
Hmong language 570
Hofstadter, D. R. 197
homunculus 466
hormones 463, 741
Hubbard, T. L. 521–2, 530
hubs 127–9, 128
Hugo, V. 568
human characteristics 19
humanistic approach 364, 365
Huntington’s disease 703
Huron, D. 372, 810
Hurt-Thaut, C. P. 705
Husserl, E. 675
Hutchinson, M. et al. 794
Hyde, K. L. et al. 217, 471
hypothalamic-pituitary-adrenal (HPA) axis 341
I
ICINAS-BRECVEMAC 287–8, 288, 289–90, 294, 299, 301, 304, 305
iconic coding of emotions 288
identity decision 158
IDyOM (Information Dynamics of Music) 52–8
Ilari, B. 602–3
imagery 94–5, 534–6
embodied 529–34
involuntary 526–9
and perception
behavioral and psychophysical studies 522–3
effect of brain damage 523–4
physiological measures 524–6
Imagination, Tension, Prediction, Reaction, and Appraisal (ITPRA) 372
imitation 675–6
immediate early response genes (IEGs) 448–9
immune cells 347
immune system 347–9
immunoglobin A (IgA) 348–9
implicit memory system 238, 244, 245–6, 255
improvisation 487, 512–14
EEG studies 509–12
fMRI studies 488–506
jazz 276, 277, 278, 490–1, 502–6, 508–9
language areas 516–17
limbic processing 516
as model of creativity 275–6
motor regions 515–16
neuroimaging studies 276–7
parietal lobes 518–19
PET studies 506–7
sensory processing 517–18
similarity to speech 571
tDCS studies 508–9
individual variation
in enjoyment of music 380
in musical memory 247–8
infant-directed singing 596–8
infant-directed speech 573
infants
cortisol levels 345
development of auditory perception 675–7, 677
early right anterior negativity (ERAN) 200, 203
emotion perception 600–1
enculturation 32
language and music
co-development 576–8, 577
perception 571–6, 575
mimicry 675–6
response to musical stimulation 28–9
rhythm development 174–5
rhythm perception 48–9, 593–4, 595–6
scale perception 47
synchronous movement 603, 604–5
vestibular system 226
inferior colliculus (IC) 189–90, 338, 446
inferior frontal gyrus (IFG) 221, 375, 490, 516–17, 763, 764
inferior parietal lobule 490
information retrieval techniques 277
inhibition, abnormalities 476
innateness 808
inner ear and inner voice 531–2
instrument tones vs. pure tones 22
instruments
harmonics 148–52, 149, 155
universality 30
insula 301, 516
insular cortex 243
Intartaglia, B. et al. 401
integration 213–16, 214
integration windows 217
intelligence quotient (IQ), effect of musical training 272–3, 647–50, 655
internal rehearsal 243
interregional interactions 123–4
intervals
judging size by sight 218
memory 240–1
intrinsic coding of emotions 288
intrinsic features in musical memory 245–6
invariants 29
involuntary musical imagery 94, 526–9
Ishizu, T. and Zeki, S. 377
J
Jabusch, H. C. 793
Jackendoff, R. 567–8, 569
Jackson, J. 4
Jacobsen, T. and Beudt, S. 367
James, W. 263
Janata, P. 302
Janus, M. et al. 555
Japanese language 570
Jaschke, A. C. et al. 658
Javanese scales vs. Western scales 574
jazz 549, 571
blue notes 213
free form 75–6
improvisation 276, 277, 278, 490–1, 502–6, 508–9
jazz musicians, personalities 278
Jennings, J. J. and Kuehn, D. P. 726
Jentschke, S. and Koelsch, S. 403
Johnson, M. K. et al. 244
Jokel, R. et al. 722
Juslin, P. N. 286, 288, 290, 374, 375
Juslin, P. N. and Västfjäll, D. 372, 374
K
Kalveram, K. T. and Seyfarth, A. 255
Kämpfe, J. et al. 252
Kant, I. 810
Karma, K., (Karma music test (KMT)) 441, 442, 444, 445–6
KCTD8 gene 446
Keith, R. and Aronson, A. 719–20
Keller, P. E. 532
keyboard players 472
Kim, M. and Tomaino, C. 719
Kim, S. J. et al. 706
Kimata, H. 348
kinaesthetic feedback 465
kinaesthetic rhythm 177
Kindermusik classes 595–6
King, B. B. 213
Kleinmintz, O. M. et al. 278
Klimesch, W. 249, 252
Klimesch, W. et al. 250
Knoblauch, A. 4–5
knowledge-free structuring 194, 196
Kodaly and Orff 607, 728
Koelsch, S. 299, 377
Koelsch, S. et al. 105, 197, 202, 245, 301
Kojovic, M. et al. 700
Kolb, B. and Gibb, R. 27
Kornysheva, K. et al. 378
Korsakoff’s syndrome 244
Kotz, S. A. and Gunter, T. C. 716
Kotz, S. A. and Schwartze, M. 723
Koyama, Y. et al. 348
Kragness, H. E. and Trainor, L. J. 599
Kraus, N. 401
Kraus, N. et al. 553
Krauss, T. and Galloway, H. 720
Kreiner, H. and Eviatar, Z. 582
Kuhl, P. K. 44
Kühn, S. and Gallinat, J. 378
Kunert, R. et al. 399
Kussner, M. B. et al. 252
L
Lai, G. et al. 581
Lai, H. L. and Good, M. 629
Lang Lang 462
language
and aging 633–4
co-development with music 576–8, 577
development 676–7
discrimination of sounds 394–6
disorders 580–1, 715–28
entanglement with music 582
learning 195
meaning in 79
modular approach 392–4
vs. music 25, 187–8, 567–9
aesthetic experience 370
music of speech 569–71
and music training 270–1, 400–4, 652–4
perception
development 574–6, 575
innate abilities 571–4, 578–80
PET studies 506–7
phonemes 190–1, 550
processing 391, 396–400
rehabilitation 633–4
rhythm 173–4
similarities to music 396–400
stress-timed vs. syllable-timed 570
temporal focus 406–8
tonal languages 203, 240, 402, 570
Mandarin Chinese 551
and tonality perception 48
training vs. musical training 553, 555
use of music in language training 404–6
language acquisition 567, 583
Language Acquisition Device (LAD) 808
language areas, improvisation 516–17
language functions, neurologic music therapy 745
Larson, S. 530
laryngospasms 725
late positive component (LPC) 547–8
lateral prefrontal cortex 375
lateral regions 277
lateral sulcus 216
laterodorsal tegmental nucleus (LDT) 350
learning 194–5
enhanced by music 269–74
influence of background music 252–4
transfer between language and music 400–4
see also music training
learning-related changes in coherence (LRCC) 253–4
learning related synchronization (LRS) 751
LeBlanc, A. 46
Leder, H. et al. 381
left anterior negativity (LAN) 203
left auditory cortex 106
left posterior planum temporale 243
Lehne, M. and Koelsch, S. 378
Lerdahl, F. and Jackendoff, R. 197
Levitin, D. J. 528
Levitin, D. J. and Menon, V. 202
Liégeois-Chauvel, C. et al. 766
Limb, C. J. and Braun, A. R. 276, 490–1, 498, 500
limbic system 102–5, 516
Lindquist, K. A. et al. 132, 135, 294
Linked (Barabási) 125
Liu, C. et al. 378, 381, 495–7, 498, 500, 513
locus coeruleus (LC) 345–7
Lomber, S. G., Meredith, M. A., and Kral, A. 219
long-term depression (LTD) 787
long-term memory 238
long-term potentiation (LTP) 787
looped speech experiment 582–3
Lopata, J. A. et al. 511–12
Lortie, C. L. et al. 725
loudness judgments, visuomotor influences 223
lullabies 345, 596–7
vs. playsongs 597
universality 30
lyrical improvisation 495–7
lyrics 106, 569–70
M
Mafa people, Cameroon 46
Magic Flute, The (Mozart) 21
magnetic resonance imaging (MRI) 470
Mahmoudzadeh, M. et al. 579
major chords vs. minor chords 147
Makam music, Turkish 53–8
Mandarin Chinese language 551, 570
Mang, E. 676
Manning, F. C. and Schutz, M. 226
Manuck, S. and McCaffery, J. 22
MAOB protein imaging 807
Marie, C. et al. 550
marimba 156–7
Marr, D. 391, 406
Martinez-Molina, N. et al. 378
Mas-Herrero, E. et al. 378
Mathias, B. et al. 765
Mauszycki, S. C. and Wambaugh, J. L. 719
maximum-likelihood estimation (MLE) 215–16
McDermott, O. et al. 636
McGurk effect 221
McIntosh, G. C. et al. 702
McPherson, M. J. et al. 504–6
meaning 78, 79, 245
medial PFC 513
medial temporal area 221
mediating model 744
MEG (magnetoencephalography) 5, 365, 468
Mehr, S. A. et al. 28, 274, 651–2
melodic expectancy violations 48
melodic intonation therapy (MIT) 633–4, 715, 717, 719–20, 721
memory 237–8, 465
and aging 632–3
auditory sensory (ASM) 192–4
autobiographical 245, 248, 302, 633
cross-cultural research 50–1
during music listening 238–9, 239
effect of complexity 248
enhanced by music 250–5
episodic 238, 245, 248, 290, 303
and expertise 248
intervals 240–1
neural activity 248–50
neural networks 243–4
recognition of music 244–50, 255–6, 256
tonal working memory 241–4
tone 239–40
training 554, 649–50, 651, 652–3, 748, 750
Mendelian rules 444
Menninghaus, W. et al. 381
Menon, V. and Levitin, D. J. 378
mental practice and performance 532–3
mental training 469
Merchant, H. et al. 696
mere exposure effect 244
Merker, B. 80n, 91
mesial regions 276–7
mesolimbic reward pathway 246, 299n
metaphors, spatial and force 529–30
metaplasticity 428–9
meta-systems 502
meter 48–9, 165
perception in infants 593–4, 595–6
see also rhythm
methodological approach 393
metrical hierarchy 592–3, 595
Meyer, L. B. 601
middle cerebral artery (MCA) 767
middle temporal gyrus (MTG) 767
MIDI-based Scale Analysis (MSA) 780
mild cognitive impairment (MCI) 625
Milovanov, R. et al. 403
mimicry 530, 675–6
Minagawa-Kawai, Y. et al. 579–80
Mingus, C. 75–6
minor chords vs. major chords 147
mirror neurons 179, 303, 469, 530
mismatch negativity (MMN) 176, 193, 241, 375, 524, 548–9, 552–3, 724
abstract-feature (afMMN) 196
physical (phMMN) 196
statistical (sMMN) 195–6, 199
modularity (Q) 129–30, 131
modulation identification 48
Molnar-Szakacs, I. and Overy, K. 378
monkeys
auditory system 191
dystonia 477
rhythm perception 176
vibration receptors 220
monocyte chemoattractant protein (MCP) 348
Montag, C. et al. 378
Montreal Battery on Evaluation of Amusia (MBEA) 556, 761–2
Montreal Protocol for Identification of Amusia (MPIA) 762
mood 286
Moran, J. 571
Morcom, A. M. and Fletcher, P. C. 136–7
Moreno, S. et al. 273, 555, 653
Mote, J. 600–1
motherese 573
motor brain function 468–9
motor co-representations 469
motor cortex 95–102, 96, 221
motor cortico-basal-ganglia-thalamo-cortical (mCBGT) circuit 176
motor evoked potentials (MEPs) 169
motor functions, and aging 634–5
motor regions 301
musicians 421
primary and secondary 465, 466–7
role in improvisation 515–16
motor signals 408
motor speech disorders (MSDs) 717–23
motor system
rehabilitation for 695–707
in rhythm perception 166–70, 179–80
Moussard, A. et al. 625, 627, 633
movement, synchronous 602–5 see also dance; drumming
movement-based influences on rhythm perception 225–6
movement disorders 701–5
Mozart, W. A. 21
Mozart Effect 269, 274 see also learning
Müller, K. et al. 225
Müller, V. et al. 378
multidimensional scaling (MDS) 150
multifactorial gene–environment interaction model (MGIM) 23
multifactorial traits 440
multiple sclerosis (MS) patients 254, 698, 704–5, 751
multiple system atrophy (MSA) 704
multisensory nature of music training 430
multisensory perception 221, 227
pitch 218–21
rhythm 223–6
timbre 221–2
multisensory processing 212–16
rhythm 177–8
music, definition 187
music centers 5
music education 46, 442, 809
music processing, neural basis 90–106
Music Psychosocial Training and Counseling (MPC) 749
music-syntactic processing 200, 201–3
music therapy 344, 348, 634, 635–6
for cognitive dysfunction 742–3
neurologic (NMT) 743–6, 807
cognition and emotion 747–53
for language disorders 715–28
for motor system disorders 697–700, 707
music training 419
and academic achievement 654–6
and aptitude 658–60, 659
effect on brain development 424–7, 426, 551–3, 557
effect on brain function 467–70
and brain structure 420–2, 420, 470–4, 474
and cognition 647–50, 654–7, 659
in aging 626–7
demonstration and observation 469
effect on executive functions 554–5
and healthy aging 656–7
and language skills 400–4, 652–4
motor functions 468–9
nature vs. nurture debate 461–2
and plasticity 422–4, 429–32, 462–4, 467–8, 546–7
cortical sound processing 547–50
short-term 427–8
studies of 645–7
types 650
and visuospatial skills 651–2
see also practice
musical affect 534
musical anhedonia 768–9
Musical Attention Control Training (MACT) 748
musical behavior, genetic influences 22–3, 23
Musical Echoic Memory Training (MEMT) 748
Musical Executive Function Training (MEFT) 748
musical expectancy formation 194–6, 197, 202–3, 290, 303–4
Musical Language Acquisition Device (MLAD) 808
Musical Mnemonic Training (MMT) 748
Musical Neglect Training (MNT) 747, 751–2
musical response model 744
Musical Sensory Orientation Training (MSOT) 747
Musical Speech Stimulation (MUSTIM) 717, 720
musical structure building 197
musical systems 42–3
musicality 34
musicians
auditory brainstem response (ABR) 550–1, 553
brain structure 420–2, 420, 428–9
cognition and aging 624–6
cortical sound processing 547–50
with dyslexia 580
exceptional and autistic 671–4, 681–90
language skills 400–4
and memory 248, 250–2
motor co-representations 469
network connections 106
neural responses to piano tones 467–8
plasticity 478–80, 479
practice 429–30
somatosensory perception 468
musician’s dystonia (MD) 475–8, 479, 777–82, 778, 779
future directions 791–6
pathogenic theory 787–90, 788, 794–6
pathophysiology 784–7, 794
and plasticity 787
treatments 782–4, 783, 791–3
musicogenic epilepsy 5
myelination 27
N
N100 responses 5
Nair, D. G. et al. 300
Nakata, T. and Mitani, C. 601
Narme, P. et al. 636
nature vs. nurture 19–22
in music training 461–2
Neanderthals 369, 718
Neapolitan chord 266
near-infrared spectroscopy (NISR), in infants 28
Nettl, B. 31
network-based approaches 123–5
neuroimaging analysis 132–8, 134
network disorders, dystonia as 478
network generation 132–5
network metrics 125–31 see also connectivity
network science 5, 125
networks 89, 95–105
interactions 105–6
Neuhoff, J. 159
neural auditory pathways 90–3, 90
neural oscillations 598–9
neural plasticity see plasticity
neural pruning 26–7
neural resonance theory 170–2
neuroaesthetics 366, 367–74, 371, 373
brain structures 375–6
future challenges 381–2
studies 377–81
neurochemical responses to music 333–4, 350–2
cholinergic systems 350
dopamine see dopamine
endogenous opioid systems (EOSs) 336–8
neuroendocrine systems 339–45
norepinephrine (NE) systems 345–7
peripheral immune system 347–9
serotonin systems 338–9
neuroendocrine systems 339–45
neuroimaging analysis, network-based 132–8, 134
neurologic music therapy (NMT) 743–6, 807
cognition and emotion 747–53
for language disorders 715–28
for motor system disorders 697–700, 707
neurological disorders
cognitive functions 738
cognitive remediation (CR) for 739–40
rehabilitation for 695–707
speech 717–23
neurological markers, of congenital amusia 762–3
neuropsychiatric disorders 451
neurotransmitter imaging 807
Newman, M. E. J. 125
Nieminen, S. et al. 378
node parcellation 130, 131, 133
nodes 125–7, 125, 126, 131, 134
non-musical parallel model 744
norepinephrine (NE) systems 345–7
Norman-Haignere, S. et al. 569
nostalgia 245
notes, duration 156–7
novelty spectrum 72–4, 75–6
NR3C1 447
NRGN gene 447
O
oboe, harmonics 155
occipital gyri 489
olivocerebellar network 99
Onofre, F. et al. 725
onset of notes 151
Openness-to-Experience trait 649
OPERA (Overlap, Precision, Emotion, Repetition, Attention) hypothesis 264, 270, 403
opioid receptors 352, 741
Oral Motor and Respiratory Exercises (OMREX) 717, 718, 719
orbitofrontal cortex 106
Organ2/ASLSP (Cage) 21
oscillation-based models of speech perception 407
oscillatory functions 751
in rhythm perception 170–2
out-of-culture scale violations 48
overtones 148–9, 149
oxytocin (OT) 339–40, 341, 534, 741
P
Pacinian corpuscles 188–9
pain modulation 338
Pallesen, K. J. et al. 554
Pantev, C. et al. 22, 467
Papoušek, M. 676
parabelt 216, 221
parahippocampal gyrus 300
parahippocampus 301
parental music education 442
parietal areas 243
parietal lobe 465
role in improvisation 518–19
Parkinson’s disease (PD) 448, 634–5, 695, 698–700, 702–3
cognitive remediation 740
response to dopamine 335
rhythm perception 169–70, 179
speech deficits 716, 718
Parkinsonism 704
passive musical exposure 32, 627–8
Patel, A. 393, 400, 403
Patel, A. D. 264, 270, 582
pathogenic theory of musician’s dystonia (MD) 787–90, 788, 794–6
pathophysiology, for musician’s dystonia (MD) 784–7, 794
pathways
auditory 217
visual 218
Patterned Sensory Enhancement (PSE) 698–9, 701
PCC 518–19
PCDHA gene cluster 446
PCHD7 gene 446
PDGFRA gene 446
Pearce, M. T. 54, 378
Pearce, M. T. et al. 381
pedaling 472
pedunculopontine tegmental nucleus (PPT) 350
Pelowski, M. et al. 367
Perani, D. 567
Perani, D. et al. 579
perception 464–5
and imagery
behavioral and psychophysical studies 522–3
effect of brain damage 523–4
physiological measures 524–6
deficits see amusia
development 574–6, 575, 675–7, 677
autistic children 678–81, 678, 680
innate abilities 571–4, 578–80
training 747–8
without awareness 762–3
perceptual integration 213–16
perceptual magnetic effect 44
perceptual narrowing 595–6
percussion, expressive gestures 223
percussion instruments, note duration 156–7
percussive tones 153, 154
Pereira, C. S. et al. 378
Peretz, I. 296, 300, 304, 571
Peretz, I. and Coltheart, M. 393
Peretz, I. et al. 397–8, 766
perfect pitch see absolute pitch
performance 459–60
biological restrictions 21
brain regions 464, 465–7, 465
effect on human transcriptome 447
expert 22–3, 23
mental 532–3
and plasticity 460, 462–4, 478–80, 479
as therapy for neurological disorders 699–700
peripheral immune system 347–9
personality traits 649, 655
of creative musicians 278
PET (positron emission tomography) scans 5, 333–4, 352, 807
improvisation 506–7
Petersen, B. et al. 724
Peterson, D. A. et al. 780, 791
Petkov, C. I. et al. 191
Phillips, D. P., Hall, S. E., and Boehnke, S. E. 153
Phillips-Silver, J. and Trainor, L. J. 226, 594
Phillips-Silver, J. et al. 764–5
phonemes 569–71
similarity to timbre 190–1
phonological loop 241, 243
phonological store 243
PHOX2B gene 446
phrase boundary perception 49
phrenology 4, 4
physical medicine and rehabilitation (PMR) for dystonia 784, 792
physical mismatch negativity (phMMN) 196
physical responses to music, neural basis 97
physiological studies of imagery and perception 524–6
Piaf, Edith 462
pianists 489–90, 492–5, 499–502
gray matter 472–4, 474
mental performance 532–3
pedaling 472
piano
improvisation 276
temporal structure of notes 151
vibration detection 222
Picelli, A. et al. 703
Pinho, A. L. et al. 499–502, 518
Piper, A. 571
pitch
absolute vs. relative 681–3, 683
attention to 266–7
changes, detection 147
expectations 52–3
imagery and perception 522
interval representation 56
memory 239–40
metaphors 530
multisensory perception 218–21
perception 189–90, 191
acquired deficits 766
heritability 442–3
in infants 573–4
processing 194
in tonal languages 551
pitch-based amusia 761–3
pitch perception accuracy (PPA) test 446
planum temporale 194
plasticity 24–6, 26, 740
and aging 624, 627, 631
cognitive remediation 741–2
and musical performance 460, 462–4
and musician’s dystonia (MD) 475–8, 479, 787
and performance 478–80, 479
and training 422–4, 429–32, 462–4, 467–8, 546–7
cortical sound processing 547–50
playsongs vs. lullabies 597
pleasure in music 376, 760
anhedonia 768–9
individual variation 380
training 430–1
Poeppel, D. 217, 406
pontomesencephalic tegmentum (PMT) 350
posterior pituitary 339–41
power spectra 150–1, 152
PPP2R3A gene 448–9
practice
and brain structure changes 25, 421, 423
deliberate 460–1
effect on brain structure 470–4, 474
and expertise 429–30
and genetic influences 22–3, 23
mental 532–3
myelination 27
through observation 469
see also training
pre-supplementary motor area (pre-SMA) 276, 488–9, 494–5, 515
precentral gyrus 465
precise auditory timing hypothesis (PATH) 406
predictability 372
prediction 430
in rhythm 598–600
predictive coding model 221–2
Predominant Patterns 30–1
preference 286
cross-cultural research 46–7
network connections 106
prefrontal regions 277
prehistoric evidence of music 369
premotor area (PMA) 375, 465, 466
premotor cortex (PMC), rhythm perception 168
presbylaryngis 725
primary auditory cortex (A1/PAC) 191–2, 464, 470
primary motor cortex (M1) 169, 465, 466, 472
primary somatosensory area (S1) 465
priming of motor activity 696–7
private music tuition 650
proficiency, and memory tasks 250–2
progressive supranuclear palsy (PSP) 704
proopiomelanocortin (POMC) 741
prosody 570–1, 582
perception in infants 572–3
prosopagnosia 770
protocadherin15 (PCDH15) 449
Przybylski, L. et al. 606
Przysinda, E. et al. 199, 278
psychiatric disorders 451
cognitive remediation 740
psychological impact of music 78–81
psychological voice disorders 726
psychophysical studies, imagery and perception 522–3
psychosocial function, NMT 749
puberphonia 726
publications 6
pulse 568
pure tones 153
vs. instrument tones 22
putamen see basal ganglia
Putkinen, V. et al. 552–3
Pygmalion effect 251
R
Raaijmakers, J. G. and Shiffrin, R. M. 255
radioligands 333–4, 352
rāgamālikā modulation 48
Ramus, F. and Mehler, J. 570
random networks 126–8, 127
Range Universals 30–1
rap, freestyle 276, 495–7
rating scales for dystonia 780–2, 781, 791, 792
Rational Scientific Mediating Model (RSMM) 743–4, 749
real-world associations 156
recognition of music 244–50, 255–6, 256
recursion 197
Redies, C. 368
Redirected Phonation 726–7
Reelin pathway 445
refined auditory processing 467–8
refined somatosensory perception 468
region-of-interest (ROI)/atlas based networks 130, 131, 301, 494–5
regions 123
regular networks 126–7, 127
regularities 194–6, 202–3
Reich, S. 570–1
reinforcement learning (RL) 795
relative pitch (RP) processing 240–1, 681–3, 683
relaxation 629
repetition 568, 699
representations of music 568
Resonant Voice Therapy 726
respiratory disorders, therapy for 727
restorative approach to cognitive remediation 739
reward mechanisms 337
reward pathway, 299n
reward system (mesolimbic) 246, 374, 380, 431, 631
RGS2 gene 447
RGS9 gene 445
rhythm 592–3
for attention training 749–50
beat-based vs. non-beat based 165–6, 166
development 174–5
and developmental disorders 605–7
disorders 763–5
and emotion 600–1
in infant-directed singing 596–8
of language 173–4, 570–1
mirroring and joint action 179–80
multisensory perception 223–6
perception 48–9, 166–72
cross-modal investigations 177–8
evolution 175–6
in infants 572, 574, 575–6, 593–4, 595–6
processing abilities, individual differences 178–9
regularity 598–600
selective attention 268–9
and stuttering 722
synchronous movement 602–5
Rhythmic Auditory Stimulation (RAS) 696–8, 702–3, 706
rhythmic entrainment 289–90, 302–3, 374, 695–6, 698
rhythmic improvisation 276, 497–9
rhythmic intervention for dyslexia 727–8
rhythmic priming 404
rhythmic-reading training (RRT) 728
Rhythmic Speech Cueing (RSC) 717, 719
Richie, L. 570
right frontotemporal network 763, 764
Rochette, F. et al. 724
Rock, A. M. et al. 597
rock music 549
Rohrmeier, M. 197
Rohrmeier, M. and Cross, I. 194
Roland, P. E., Skinhøj, E., and Lassen, N. A. 5
roles of music 31
Rosen, D. S. et al. 508–9
Rosenblum, L. D. and Fowler, C. A. 223
Rosenkrantz, K. 785–6
Rubinov, M. and Sporns, O. 131
Russian language 570
Russo, F. A., Ammirante, P., and Fels, D. I. 222
S
Sachs, M. E. et al. 378
Saint-Georges, C. et al. 573
Sakamoto, M. et al. 635
Saldaña, H. M. and Rosenblum, L. D. 221
Salimpoor, V. N. and Zatorre, R. J. 379
Salimpoor, V. N. et al. 337, 378
same–different tests 5
Sammler, D. et al. 398–9, 400, 630
Samson, S. and Peretz, I. 244
Santoni, C. et al. 726
Särkämo, T. et al. 254
SAT scores 654
Sauder, C. et al. 725
scaffolding 462, 717, 742
scale-free networks 127
scale perception 47
Scaled Inclusivity 130–1
scales 187
Javanese vs. Western 574
scat singing 569
Schellenberg, E. G. 658
schemas 247–8
Schenker, H. 197
schizophrenia 451, 524, 527–8, 531
cognitive remediation 740
Schlaug, G. 471
Schlaug, G. et al. 425, 471, 634
Schneider, N. et al. 343, 470
Schön, D. et al. 398
Schopenhauer, A. 77, 80
Schulze, K. and Koelsch, S. 242
Schumann 21
Schutz, M. et al. 223
sea lions, rhythm perception 176
Search of Associative Memory (SAM) model 247, 253, 255
Seashore, C. (Seashore tests) 5, 441, 442, 444, 446
secondary auditory area (A2) 464
secondary motor areas 465
secretory IgA (S-IgA) 349
Seesjärvi, E. et al. 556
segmental dystonia 777
Seinfeld, S. et al. 627
selection, and filtering 266
selection theories of attention, early vs. late 264–6
Semal, C. et al. 242
semantic associative network model of memory formation 247
semantic memory 238, 244–5
sensorimotor domain 471
sensorimotor functions, neurologic music therapy 745
sensorimotor integration 477, 786
sensorimotor pathways 221
sensory deficits, language disorders 723–4
sensory memory 237
sensory perception 477
sensory processing, improvisation 517–18
serial-to-parallel conversion 238–9, 239
serotonin (5-HT) 246, 338–9, 741
sex differences, in endocrine levels 344
sexual selection theory 67–9
Shahin, A., Roberts, L., and Trainor, L. 28
Shannon, C. E. 277
shared-resources hypothesis 300
shared syntactic resource integration hypothesis (SSRIH) 270–1
short-latency intracortical inhibition (SICI) 785–6
short-term memory 238
sight, visual rhythm 177–8
sight and sound association 156–8
Silbo Gomero, whistled speech 582
sine-wave speech 582
singing
birds see songbirds
development in children 578
endocrine responses 343–4
infant-directed 596–8
and memory tasks 253–5
oxytocin levels 340
universality 30
singing therapy 717–18, 721, 722–3
for respiratory disorders 727
for voice disorders 725–7
Six Degrees (Watts) 125
skin, music detected in 189
Skinner, B. F. 392
sleep disorders 725
Slevc, L. R. and Okada, B. M. 271
Slevc, L. R. et al. 398
slow-down exercises (SDEs) 793
SMA 515, 554
small-world networks 126–7, 127, 129
Smith, H. 21
Smith, J. D. et al. 531
sMMN (statistical mismatch negativity) 195–6, 199
SNCA gene 447, 448, 448
social functions of music 602–5
socio-economic status (SES) 655, 659
Soley, G. and Hannon, E. E. 596
somatosensory cortex 221, 465
somatosensory influences
on pitch perception 220–1
on rhythm perception 224–5
on timbre perception 222
somatosensory perception 468
somatosensory processing 517–18
songbirds 67–73, 81, 393–4, 441, 443, 444–5, 447, 449
songs, combination of language and music 398–9
Sowiński, J. and Dalla Bella, S. 765
spasmodic dysphonia (SD) 725
specific language impairment (SLI) 580–1
Spector, J. T. and Brandfonbrener, A. G. 780
spectral properties 148–9, 149, 150–1, 152
spectrum envelope 190–1
speech
music of 569–71
transcribed into music 570–1
speech disorders 4
fluency 721–3
neurologic music therapy (NMT) for 715–28, 745
speech sounds 550
speech therapy 633–4
spinal cord injuries 698
sports science 809
Staal, F. 78
Stahl, B. et al. 405
Stanford-Binet IQ test 648
statistical learning 25
statistical mismatch negativity (sMMN) 195–6, 199
statistical structures 194–6, 202–3
Steele, C. J. et al. 426, 473
Steinbeis, N. and Koelsch, S. 379
Sternberg, R. J. et al. 263
Strait, D. L. et al. 190
stress hormones 341–5
stress-timed languages vs. syllable-timed languages 570
string musicians 468, 472–3
mental performance 532–3
stroke patients 254, 633–4, 635, 698, 699, 701
acquired amusia 767
cognitive remediation 740
singing therapy 718
Stroop tests 554–5
structural features, cross-cultural research 47–50
structure
effects of training on 470–4, 474
individual variation in understanding 247–8
stuttering 581, 722–3
substantia nigra pars compacta (SNpc) 334
Sun, L. et al. 199–200
superadditivity 215
superior colliculus 221
superior parietal lobule (SPL) 490, 517
superior temporal gyrus (STG) 216, 470, 488, 489, 763, 764
superior temporal sulcus 219, 221
supplementary motor area (SMA) 465, 466
rhythm perception 168
surgery
cortisol levels 343
immunoglobin A levels 349
surprise 372
Suzuki, M. et al. 379
Swift, J. 623
syllable-timed languages vs. stress-timed languages 570
syllables, rhythm 606–7
Symbolic Communication Training Through Music (SYCOM) 717
sympathetic nervous system (SNS) 345–7
synapses 463
development 26
synchronous movement 602–5 see also dance; drumming
synchrony with rhythm 174, 176 see also rhythm
synesthesia 528–9
syntax 393, 399–400
infant development 576–7
synthesized notes 148–9, 152, 157, 201
systematic approach 364–5
T
Tamplin, J. et al. 727
techno music 337
tempo
and emotion 600–1
memory for 245–6
temporal approach, to language and music 406–8
temporal attention 267–9
temporal brain areas, and memory 248
temporal discrimination threshold (TDT) 786–7, 790, 793, 794
temporal dynamics 146, 148–55, 149, 154, 157–9
temporal gyri 470–1
temporal lobes 464
temporal processing
disorders 786–7
universals 30
temporoparietal junction (TPJ) 488, 518
deactivation 492
tensor based morphometry (TBM) 5, 470
Tervaniemi, M. et al. 550
testosterone 344
thalamus 301
Thaut, M. H. et al. 100, 253, 696, 701, 702, 703
Therapeutic Instrumental Music Performance (TIMP) 699–700, 701, 706
Therapeutic Singing (TS) 717, 718, 719
Thompson, W. F. et al. 213, 218
Thomson, J. M. et al. 727
thresholding procedures 133
throat singing 571
Tierney, A. and Kraus, N. 405–6
timbre 149–51, 192
memory for 245–6
multisensory perception 221–2
perception in infants 572, 573–4
of phonemes 569–70
similarity to phonemes 190–1
time perception, acquired deficits 766–7
timing 427
mechanisms, absolute vs. relative 166
neural networks 97–102
Tinbergen, N. 391
Toccata in C Major (Schumann) 21
tone, memory 239–44
tonal languages 203, 240, 402, 551, 570
tonality perception 47–8
tone deafness see amusia
tone intervals, memory 240–1
tone onset 151
tone patterns 193
tonotopic maps 216
tonotopic organization 22, 90–1, 191
Torres, E. B. et al. 706
Toscanini 21
Touch-Cue Method (TCM) 720
trading fours 502–4
training 419
and academic achievement 654–6
and aptitude 658–60, 659
effect on brain development 424–7, 426, 551–3, 557
effect on brain function 467–70
and brain structure 420–2, 420, 470–4, 474
and cognition 647–50, 654–7, 659
in aging 626–7
demonstration and observation 469
effect on executive functions 554–5
and healthy aging 656–7
and language skills 400–4, 652–4
motor functions 468–9
nature vs. nurture debate 461–2
and plasticity 422–4, 429–32, 462–4, 467–8, 546–7
cortical sound processing 547–50
short-term 427–8
studies of 645–7
types 650
and visuospatial skills 651–2
see also practice
Trainor, L. J. 597
Trainor, L. J. and Adams, B. 594
Trainor, L. J. et al. 226, 240, 572, 601
Tramo, M. J. et al. 191
Tranchant, P. and Vuvan, D. T. 765
Tranchant, P. et al. 224
transcranial alternating current stimulation (tACS) 771
transcranial direct current stimulation (tDCS) 507–9, 771
transcranial magnetic stimulation (TMS) 5, 219, 250, 430
for amusia 771
for musician’s dystonia (MD) 784, 793
rhythm perception 169
transfer 269–74, 630, 646, 649
music and language 400–4
Transformational Design Model (TDM) 744–6, 749
transverse temporal gyrus 216
traumatic brain injury (TBI) 698, 700
cognitive remediation 740
singing therapy 718
trombone, harmonics 151, 152
Trost, W. et al. 379
trumpet, harmonics 148–9, 149, 150
Turkish IDyOM model 53–8
Turkish music 50–1
twin studies 23, 24–5, 556
aging 626
aptitude 442
music training 423
willingness to practice 444, 461
U
Ullén, F., Hambrick, D., and Mosing, M. 22–3
Ulrich 188
Unified Dystonia Rating Scale (UDRS) 782
unity assumption 158
universality of music 369–70
universals 29–31
unpleasant sounds 155–6, 339
use patterns 790
usefulness of music 275
V
valence 286, 601
Van den Heuvel, M. P. et al. 133
Van Wijk, B. C., Stam, C. J., and Daffertshofer, A. 133
Vaquero, L. et al. 426, 463
Vatakis, A. et al. 158
ventral pathways 217
ventral premotor cortex (vPMC) 201, 426, 490
ventral striatum 104–5, 301
ventral tegmental area (VTA) 334
ventriloquist effect 158
ventrolateral prefrontal cortex (PFC) 554
rhythm perception 168
Verghese, J. et al. 626
vertical structure 147
vestibular apparatus 188–9
vestibular cortex 221
vestibular system 225–6
vibration receptors 188–9, 220, 222, 224–5
Vienna Integrated Model of Art Perception (VIMAP) 369
Villarreal, M. F. et al. 497–9
viola, harmonics 155
visual cortex 221
visual imagery 290, 303
visual pathways 218
visual perception 213
visual processing 518
visual rhythm 177–8
visuomotor influences
on pitch perception 218–20
on rhythm perception 223–4
timbre perception 221–2
visuospatial skills, and musical training 651–2, 657
VLDLR gene 445
Vocal Intonation Therapy (VIT) 717, 718
vocal misuse disorders 726
voice disorders 724–7
voluntary musical imagery 94–5
voxel based morphometry (VBM) 5, 470, 767
voxel-based networks 130, 131
Vuust, P. and Kringelbach, M. L. 379
Vuust, P. and Witek, M. A. G. 408
Vuvan, D. T. et al. 199
W
Wager, T. D. et al. 292
Wallaschek, R. 5
Wambaugh, J. L. et al. 719
Wan, C. Y. et al. 715
Warren, J. D. et al. 192
Watts, D. J. 125
Wechsler Intelligence Scale for Children–III (WISC–III) 648
Welch, G. 676
well-being in aging 628–9, 635–6
Wernicke’s area 464
Weschler Preschool and Primary Scale of Intelligence–III (WPPSI–III) 653
Weschler Preschool and Primary Scale of Intelligence–Revised (WPPSI–R) 651
Western IDyOM model 53–8
Western music
modulation identification 48
phrase boundary perception 49
rhythm perception 49
Western scales vs. Javanese scales 574
Westernization of music 32
whistled speech 582
white matter (WM) 27, 420–1, 426, 428
density 463
imaging techniques 470
and music practice 471
Whitfield, I. 192
‘Who put the Bomp?’ song 569
Wiener, M. et al. 716
Wilkins, R. W. et al. 379
Williams Syndrome (WS) 340
Wilson, F. 21
Wilson, R. S. et al. 625
Witek, M. A. et al. 380
Wittgenstein, L. 810
Wong, P. C. M. et al. 190, 551, 716
Wood, B. H. et al. 702
working memory (WM) 237–8
effect of musical training 554
neural networks 243–4
tonal 241–4
training 649–50, 652–3
X
Xhosa language 570
Y
Yoruba language 570
Yoshida, K. A. et al. 594
Z
Zahavi, A. 67, 81
Zajonc, R. B. 244
Zatorre, R. J. 192
Zatorre, R. J. and Belin, P. 216
zebra finches 393–4, 441, 443, 444–5
Zeki, S. 367
Zentner, M. and Eerola, T. 602–3
Zentner, M. et al. 80
Zentner, M. R. 286
Zhang, J. et al. 199
Ziegler, A. et al. 725
ZNF223 gene 448
zygonic theory of musical-structural understanding 674–5, 679