ArticlePDF Available

Abstract and Figures

In this paper, I propose that embodied cognition in music has two distinct levels: 1) the apparent corporeal articulation of music by performers and listeners, which reflects either a desire to make visible their emotional responses to the music or rhythmic entrainment, and 2) the principal (though concealed) level of transient muscular reactions to the main coding aspects in music: the tonal relationships arranged in time. This paper argues that the apparent corporeal articulation with regard to the entrainment effect and dance (Leman & Maes, 2014) is more related to the multimodal integration that is characteristic of attending to such a multidisciplinary performing art as opera and ballet than to purely musical content. I also present empirical data on the perception of tonal distances (Korsakova-Kreyn & Dowling, 2014) and suggest an explanation of why listeners' intuitive navigation in tonal and temporal space lies at the heart of emotional responses to music, including corporeal articulation. In addition, the paper touches on the research into temporality in music, such as memory constraints in the perception of tonal structures (Tillmann & Bigand, 2004). The main emphasis of this paper is on the principal two dimensions of music: tonal relationships and time. Understanding the primacy of these dimensions is important for defining music cognition and music in general. The paper also identifies the need for collaboration among various subdisciplines in musicology and the cognitive sciences so as to further the development of the nascent field of embodied cognition in music. Expressive distinctions are easily encoded by the listeners through the verbal labels, but they are practically untranslatable by bodily mediation, when body expression is induced by the musical stimulus. Frances and Bruchon-Schweitzer (1983) The concept of embodied cognition is based on the understanding that our emotions, memory, speech, and imagination are inseparable from the experiences of our bodies. To say it differently, a mind is shaped by the motor and somatosensory experience of the body that houses that mind. Music, a very special form of communication between humans, illustrates two levels of embodied cognition. The first, " surface " level is the influence of visible bodily movement on music perception and cognition. The second, " deep " level deals with melodic morphology. Our minds read melodic information by comparing differences in the perceived tonal stability of melodic elements. The sense of stability is directly related to a physical sensation of perceived tension, which means that melodic morphology is based on a highly primitive principle of perception that involves changes in somatic tension (Radchenko et al., 2015)—changes that most likely include transient actions of the musculature in response to tonal and temporal patterns. This is why tonal music presents what is probably the most obvious and holistic example of embodied cognition. SPEECH AND EMBODIED COGNITION The theory of embodied cognition postulates that sensory information and motor activity are essential for understanding the surrounding world and for developing the abilities that are important for abstract reasoning (Foglia & Wilson, 2013). Because both memory and speech include sensorimotor representations, our imagination relies on previously experienced gestures and movements (Wellsby &
Content may be subject to copyright.
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
1
Embodied Cognition in Music
MARINA KORSAKOVA-KREYN
Touro College, New York, NY, USA
ABSTRACT: In this paper, I propose that embodied cognition in music has two
distinct levels: 1) the apparent corporeal articulation of music by performers and
listeners, which reflects either a desire to make visible their emotional responses to the
music or rhythmic entrainment, and 2) the principal (though concealed) level of
transient muscular reactions to the main coding aspects in music: the tonal
relationships arranged in time. This paper argues that the apparent corporeal
articulation with regard to the entrainment effect and dance (Leman & Maes, 2014) is
more related to the multimodal integration that is characteristic of attending to such a
multidisciplinary performing art as opera and ballet than to purely musical content. I
also present empirical data on the perception of tonal distances (Korsakova-Kreyn &
Dowling, 2014) and suggest an explanation of why listeners’ intuitive navigation in
tonal and temporal space lies at the heart of emotional responses to music, including
corporeal articulation. In addition, the paper touches on the research into temporality in
music, such as memory constraints in the perception of tonal structures (Tillmann &
Bigand, 2004). The main emphasis of this paper is on the principal two dimensions of
music: tonal relationships and time. Understanding the primacy of these dimensions is
important for defining music cognition and music in general. The paper also identifies
the need for collaboration among various subdisciplines in musicology and the
cognitive sciences so as to further the development of the nascent field of embodied
cognition in music.
KEYWORDS: music perception, embodied cognition, emotional processing in music
Expressive distinctions are easily encoded by the listeners through the verbal labels, but they are
practically untranslatable by bodily mediation, when body expression is induced by the musical stimulus.
Frances and Bruchon-Schweitzer (1983)
The concept of embodied cognition is based on the understanding that our emotions, memory, speech, and
imagination are inseparable from the experiences of our bodies. To say it differently, a mind is shaped by
the motor and somatosensory experience of the body that houses that mind. Music, a very special form of
communication between humans, illustrates two levels of embodied cognition. The first, “surface” level is
the influence of visible bodily movement on music perception and cognition. The second, “deep” level
deals with melodic morphology. Our minds read melodic information by comparing differences in the
perceived tonal stability of melodic elements. The sense of stability is directly related to a physical
sensation of perceived tension, which means that melodic morphology is based on a highly primitive
principle of perception that involves changes in somatic tension (Radchenko et al., 2015)changes that
most likely include transient actions of the musculature in response to tonal and temporal patterns. This is
why tonal music presents what is probably the most obvious and holistic example of embodied cognition.
SPEECH AND EMBODIED COGNITION
The theory of embodied cognition postulates that sensory information and motor activity are essential for
understanding the surrounding world and for developing the abilities that are important for abstract
reasoning (Foglia & Wilson, 2013). Because both memory and speech include sensorimotor
representations, our imagination relies on previously experienced gestures and movements (Wellsby &
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
2
Pexman, 2014). This is why peoples reactions to speech are faster when they see the physical movements
described by the speech (Glenberg & Kaschak, 2002). Imaging studies have shown that attending to a list
of adjectives and verbs generates activation in the motor cortex for verbs only, which demonstrates an
association between speech and specific muscular experience (James & Maouene, 2009). The perception of
abstract words such as courage and dignity produces changes in activation in the areas of the brain
associated with emotional processing (Kousta et al., 2011). Since emotion is an affective state of mind that
is accompanied by physiological changes, the perception of such abstract words may illustrate embodied
cognition as well.
PERCEIVED TENSION AND MELODIC MORPHOLOGY
When thinking about music, we must first of all consider melodic line and rhythm. In other words, the two
main dimensions of music are the space of tones and the arrow of time. Currently, the theoretical
background for studies in embodied cognition in music often resembles the theoretical underpinnings of
embodied cognition in linguistic language, which means emphasizing the effects of the observable bodily
movements of music performers and listeners on our understanding of music’s content (Leman & Maes,
2014). An image of a performing musician could indeed influence music perception, but a statement that
the musician’s corporeal articulation convey the content of music is not necessarily precise, considering
that such great musicians as Svyatoslav Richter and Glenn Gould believed that watching a pianist
performing music distracts from listening to the music and understanding its content.
There are, however, certain “covert” properties of melodies and harmonies that allow us to
describe music as a particularly significant case of embodied cognition. These properties manifest the
nature of music’s melodic morphology that is defined by perceived tension. For the theory of embodied
cognition in music, the most noteworthy aspect of melodic morphology is the phenomenon of intuitive
muscular actions of music listeners in response to the basic melodic elements that constitute musical
matter.
Among the varieties of human communication, music occupies a special place because it is able to
transmit complex information in the absence of symbolic representation. People use universally adopted
symbols to communicate mathematical thoughts, and people use semantic units (“constants of
consciousness,” Solntsev, 1974) to inform about things and actions and their qualia via linguistic language.
In music, however, there are neither words nor familiar visual images and, unlike the language of
mathematics and a foreign tongue, music cognition does not require special training (Fritz et al., 2009).
Nevertheless, musical compositions are able to elicit aesthetic emotions and convey general ideas, which
means that music incorporates a highly abstract way of transferring complex information.
The aim of musical information is to influence human psychological states. We know that
listening to music generates changes in vital signs (Bernardi et al., 2006) and neurochemistry (Chanda &
Levitin, 2013). These changes speak for the involvement of the autonomic nervous system during attending
to music, and they highlight the fundamentally intuitive nature of music perception. Merely averaging
physiological measures during music listening, however, does not explain the dynamic nature of musical
emotions (Krumhansl, 1997).
The method used to code and transfer information in music is very simple. This simplicity begins
with the parsimony of the basic elements used in the construction of musical matter (here we are speaking
of the European musical tradition that dominates the world today). Musical matter consists of seven
diatonic and five chromatic tones; these twelve tones are repeated in different registers, creating a musical
diapason. While the twelve tones differ in frequency (pitch), the most important quality of their
relationships is the difference in degree of attraction to a stable tonal center, the tonic. The difference in
tonal attraction is perceived as a difference in tension. Music teachers casually use the concept of tonal
tension when explaining the character of tonal relationships with regard to musical phrasing and
expressiveness. When people listen to music, the degrees of perceived tension define musical expectations
and generate a sense of melodic and harmonic motion in music.
The human mind reads musical information by using an appropriate system of reference that is
formed spontaneously by the mind as soon as we hear different musical sounds. This system of reference,
or perceptual schema, is known colloquially as a musical scale (see Figure 1). The nonlinear distances
between the tonic and the other tones of a scale represent melodic intervals that are characterized by the
intensity of tonal attraction, that is, by the degree of potential tonal energy. A diatonic scale has eight basic
intervals, ranging from unison to octave.
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
3
Fig. 1. The tones of a scale differ in pitch and in the degree of their attraction to a tonic. In the C major
scale, the leading tone B (diatonic step VII) is perceived as unstable (high potential energy), whereas the
tonic C represents a potential energy value of zero.
Music is constructed of melodic intervals and their combinations, chords. Considering that the distribution
of tones and their groupings in the tonal space can be explained in terms of differences in the intensity of
tonal attraction to a center of stability (the tonic), we can say that music, as a pattern of melodic
information, communicates a configuration of levels of tonal attraction that are arranged in time. The
character of the relationships among different tones allows us to discuss music in terms of phenomenal
gravity(Scruton, 1997) or the “tonal force field. Any conventional musical composition exists in a
certain key (tonality), which means a basic set of tones that are organized into a system of reference around
a given tonal center. The distance between two tonalitiesbetween their tonal centersis usually
illustrated using the circle of fifths.The greater the number of shared tones between the two scales (the
greater the amount of common information between them), the closer the two tonalities are on the circle of
fifths (see Figure 2).
Fig. 2. The greater the number of shared tones between two scales, the closer the two tonalities are on the
circle of fifths. Tonal distance increases gradually as the tonic of the second scale shifts away from the first
by intervals of a fifth, so that fewer and fewer tones are shared.
The expression tonal attraction belongs to a standard vocabulary for teaching music theory to students of
musical schools of all levels. Beginning in the 1980s, perceived tension in music became a focus of
empirical studies that used musical stimuli of various complexities, from the separate tones of a scale to
fragments from musical compositions (Krumhansl & Kessler, 1982; Bigand et al., 1996; Krumhansl, 1996;
Lehne et al., 2013). One of the important findings of the studies was a link between perceived tonal tension
and physical tension (Madsen & Fredrickson, 1993). This discovery elucidates a distinction between the
imaginary sense of gesture and movement in music and the concrete experience of perceived tonal tension
that underlies the musical imagery.
The combinations of tones that are perceived as tense are called dissonances, whereas those
combinations that are pleasing to the ear are called consonances. The brain’s activation for consonant and
dissonant sounds differs for one-day-old babies (Virtala et al., 2013), while two-month-old babies already
demonstrate a preference for consonances (Trainor, 2004). In other words, human beings are born equipped
to perceive the differences between the basic psychophysical characteristics of consonant and dissonant
sounds that are the foundation of tonal patterns and therefore of music perception. Moreover, the sensation
of musical dissonance and consonance is available to one-day-old chicks (Chiandetti & Vallortigara, 2011),
which suggests that the melodic matter of music uses mechanisms of perception that do not require
cognitive efforts of a higher order.
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
4
The consonant and dissonant qualia of combinations of musical sounds could be explained by
differences in the amount of effort required by the auditory system to process melodic information. The
hypothesis of a gradient of neuronal cost of auditory processing (Korsakova-Kreyn, 2005, 2014) proposes
that when musical sounds share important spectral information, namely the strongest harmonics in their
overtone series, the sharing makes it easier for the auditory system to process a combination of these
sounds, which generates an agreeable auditory experience. For instance, the “happy” triad in the major
mode and the sonically pleasant Pythagorean intervals of an octave, a fifth, and a fourth all present a
combination of tones that belong to the beginning of a given overtone series (counting octave equivalence).
At the level of neuronal nets, this redundancy of important spectral information may create the conditions
for more efficient processing. In other words, the psychophysics of melodic compoundsthe agreeable
quality of musical consonants and the perceived tension of dissonant soundsmay be related to the degree
of effort demanded from the auditory system to process the particular musical sound combinations.
The hypothesis of a gradient of neuronal cost of auditory processing corresponds to the principle
of less effort (Ferrero, 1894); here, the novelty lies in assigning the degree of pleasantness to the degree of
effort (or economy of dedicated resources) on the level of neuronal processing. Research into neuronal
activity in the brain stem in response to combinations of musical sounds has demonstrated that the
magnitude of brain stem frequency-following responses (FFR) was more robust and more coherent for a
consonant combination as compared with a dissonant combination (Bidelman & Krishnan, 2009). For
example, the magnitude of FFR for triads was distributed as follows: major > minor > diminished >
augmented. The study also found that the strongest neural pitch salience was produced for the unison, that
is, for two tones that have identical fundamental tones and overtone series.
Major-mode triads are known for their agreeable sonic quality and “happy” character. A major
triad is made up of a fundamental tone and its four strongest harmonics, which are generated at the very
beginning of the overtone series for that tone. This spectral characteristic emphasizes the importance of the
“hidden” dimension of overtones in music. When the harmonic series of different tones have their
beginning (strongest) overtones in common, sounding them together produces the consonant Pythagorean
intervals and a major triad, whereas when the tones are the “distant relatives” with regard to their overtone
series, this produces dissonances. It is precisely the interaction among different overtone series that
generates the tonal force field, in which melodic structures appear as “folds” of tonal space in the same
sense that material objects are “foldsand wrinkles” in 3-D space (Florensky, 1925/1993).
Conceptualizing melodies and other tonal structures as objects shaped by phenomenal tonal
gravity suggests the possibility of explaining music with the help of analytical geometry, that is, in terms of
melodic topology. The best candidate for this task seems to be a quaternionic and octonionic algebra that is
also believed to be important for explaining human intelligence (Goertzel et al., 2008). Taking into
consideration the hypothesis of supramodality in music perception (Korsakova-Kreyn, 2005), which
proposes that the perception of music on the level of gestalt involves abstraction from the modality of
incoming information during assessment of the configuration of homogeneous elements (tones) in a
dynamic field, the mathematical model of melodic objects could advance our understanding of abstract and
spatial reasoning.
UNIVERSAL LANGUAGE
A conventional musical composition includes stable and unstable tonal elements. The listeners intuition
for tonal instability and stability determines the tonal expectations that guide the perception of music and
creates a feeling of melodic and harmonic motion. Because tonal music affects listeners by means of
auditory patterns made up of different levels of perceived tonal stability or different levels of perceived
tension, it is safe to assume that music perception involves a corresponding pattern of somatic changes,
including a pattern of muscular tension and release.
A sensation of auditory comfort when listening to consonant sounds as compared with auditory
uneasiness in response to disagreeable dissonant sounds illustrates the main morphological principle of
music: perceived tension (Lerdahl & Krumhansl, 2007). The simplicity, if not primitivism of this principle,
explains music’s ability to reach people in various cognitive states, including children with autism and
elderly people with Alzheimer’s disease. Music has the ability to influence psychological states by utilizing
the simplest type of reaction of the living organism to its environment, which is variation in the degree of
experienced physical tension..
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
5
EMOTION IN MUSIC AND TONAL DISTANCES
Our exploratory study in affective responses to reorientation of the tonal system of reference, or tonal
modulation, revealed that perception of the fine aspects of tonal organization in music does not require
special musical training (Korsakova-Kreyn & Dowling, 2014). Testing the nonmusicians generated a map
of tonal relationships that corresponds essentially with the accepted theory of functional harmony (see
Figure 3). The study included two experiments. Experiment 1 used a set of 48 brief harmonic progressions
modulating to all available degrees and in all available modal conditions for the beginning and the ending
of the stimuli. Experiment 2 used a set of 24 harmonic progressions and a balanced set of real music
excerpts from the compositions of the First Viennese School and the Romantics. Experiment 2 explored
affective responses to modulation to three selected targets: the subdominant, the dominant, and a step 8,
and in the major-major condition only (the modulating stimuli began and ended in the major mode). The
perception of tonal distances was measured using bipolar adjective scales related to valence, synesthesia,
potency, and tension. The results demonstrated that increased tonal distance was felt as colder, darker, and
tenser than near tonal proximity (see Figure 4).
!"#$%
&'()%
*#+,-.%
/"#0%
1"223%
4")%
5(6"7"8.%
982(6"7"8.%
:+#$%
!";6#+8,%
4.#'8,%
!6"0%
<=>?%
<@>?%
?>?%
@>?%
=>?%
<A>?% <=>?% <@>?% ?>?% @>?% =>?% A>?%
Second'Component'
First'Component'
BC%1"#$'8+D%5#',#677+'87%E@=%4.627%F%B%G')"(%&'8)+H'87I%
J@%K%CL>=MN%J=%K%@?>OM%
G"P'#<G"P'#%
$+8'#<G"P'#%
G"P'#<$+8'#%
$+8'#<$+8'#%
!"#$<&'()%
*#+,-.</"#0%
1"223<4")%
5(6"7"8.<982(6"7"8.%
:+#$<!";6#+8,%
4.#'8,<!6"0%
@%
=%
A%
B%
L%
Q%
R%
C%
O%
@?%
@@%
?%
@%
=%
A%
B%
L%
Q%
R%
C%
O%
@?%
@@%
?%
@%
=%
A%
L%
Q%
R%
C% O%
@?%
@@%
?%
?%
@%
=%
A%
B%
L%
R%
Q%
C%
O%
@?% @@%
B%
?%<%@@%4.627%
Fig. 3. A PCA map of tonal distances for Experiment 1. The strongest variable commonality is related to
the major and minor modes. Target tonalities in the major mode (diamonds and squares) are all on the side
with the positive connotations (happy, pleasant, bright, and warm), whereas target tonalities in the minor
mode (triangles and dots) gravitate to the negative side (sad, unpleasant, dark, and cold). The participants
sensed different affective content in the different degrees of modulation. They recognized the importance of
the subdominant (5) and dominant (7), which are both close in terms of key proximity, and the
“pleasantness” of modulations in a major mode to the distant lowered second (up 1 semitone), the leading
tone (up 11 semitones), and the minor sixth (up 8 semitones)steps 1, 8, and 11 are in fact popular targets
of deceptive cadence. The participants indicated negative feelings about modulation to the tritone (often
termed the diabolo in musica, up 6 semitones) and the flattened leading tone (up 10 semitones).
The tonal patterns of musical compositions can be imagined as artfully arranged sequences of
auditory triggers that generate transient reactions of tension and relaxation. The accepted understanding of
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
6
embodied cognition presumes reiterating a previous modality-specific experience. Music is able to re-create
various psychological states in their dynamics by using artfully sequenced gradations of perceived tension.
This feeling of tension and relaxation is related to what is commonly called “life experience.” The episodes
of peoples life experiences are generally infused with emotional components of various valence and
intensity; these experiences are accompanied by somatic changes, including patterns of physical tension
and relaxation. The re-creation of psychological states in music listeners happens through integration of the
fleeting sensations of perceived tension encoded in the tonal relationships and the musical rhythms. An
artful sequencing of musical sounds and their combinations has the ability to transmit the process of
feeling. Moreover, in music we are dealing with aesthetic emotion, which includes the sense of beauty,
nonpragmatic interests, cultural constructs, personal experiences, and the idiosyncrasies of personality.
Fig. 4. Experiment 2: A PCA map of affective responses to tonal distances for Real Music excerpts (RMs).
Modulation to the distant step 8 was sensed as colder, darker, and tenser than modulation to the near
subdominant (5) and dominant (7) steps. Target-steps are numbered for the C major as the beginning
tonality.
APPARENT BODILY MOVEMENT AND MUSIC COGNITION: THE CRITIQUE OF
DOMINATING IDEOLOGY
The existing studies in embodied cognition in music focus on visible bodily movements during music
production and perception. This approach has a certain problem, however. The crux of the problem is that
the prevalent theory explains embodied cognition in music from the perspective of kinesthetic information
but without any explication of the causes of the sensations that give rise to bodily movements. For example,
when a discussion of the effects of musical phrasing on corporeal articulation does not mention the actual
musical source of melodic phrasing, which is the tonal relationships, this leads to equating musical phrasing
with phrasing in speech. This equation does not have a strong foundation, however. First, music does not
have cognitive constants similar to words. Second, melodic forms are defined by tonal gravity, which
makes music perception more akin to navigation in phenomenal tonal space and even quasispatial
reasoning than to speech. Complex musical compositions such as fugues and those written in sonata form
provide ample evidence for melodic objectoriented cognition. Studies in the perception of melodic
contour (Dowling, 1972; Korsakova-Kreyn & Dowling, under review) reaffirm the idea of a melody as a
melodic object that can still be available to perception when transformed in the ways that resemble
conventional visuospatial transformation (Shepard & Cooper, 1982). An astute critique of the idea of
corporeal articulation informing the meaning of music came from the research of the pioneers in the field of
embodied cognition, Frances and Bruchon-Schweitzer (1983), who wrote, [E]xpressive distinctions are
easily encoded by the listeners through the verbal labels, but they are practically untranslatable by bodily
mediation, when body expression is induced by the musical stimulus.”
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
7
There is indeed a shared primeval root for music and speech, which is vocal production. The
evolutionary bifurcation of vocal production into speech and music did not leave much space for relating
the formation of content in speech to formation of content in music, however. Speech provides high-level
precision of transfer of information: Speech operates with semantic units representing actions, things,
phenomena, and qualia. In contrast, music dwells on the emotional component of vocal production: Music
operates with musical sounds organized hierarchically according to perceived tension.
Apart from the shared auditory modality, another point of confusion for studies in music cognition
has been the concept of syntax in music. Some of the studies that suggest an affinity between music and
linguistic language mention linguistic syntax and musical syntax in the same breath. But these two kinds of
syntax have nothing in common. Whereas in linguistic language words are arranged according to the rules
of parts of speech (the rules of linguistic syntax), music has neither cognitive constants similar to words,
nor parts of speech. Instead, music uses a tonal gravitational pull.”
The conventional syntax of music (dictated by the gravitational pull) is expressed in a succinct
formula of tonal harmony made up of a sequence of four chords: a triad on a tonic (the expression of
stability), a triad on the subdominant tone (the expression of mild plagal instability), a triad on the
dominant tone (the expression of instability, thanks to the leading tone), and the tonic triad (a movement
back to stability). This formula of motion in tonal space came down to us from the 18th century, and it
explains the “Newtonian” tonal universe of the First Viennese School. The majority of the music we hear
today on headphones and in the concert hall relies on this formula; empirical studies have demonstrated
that people have an innate understanding of tonal harmony and that they intuitively apply the rules of tonal
harmony even when listening to a solo melody (Holleran et al., 1995). The tonal relationships, as explained
by the basic formula of tonal harmony, define melodic phrasing.
When we “feel” the completion of a musical phrase as a return to tonal stability, we experience an
emotional component of the primeval vocal production that was harnessed by the tonal hierarchy. These
two propertiesthe emotional component of voice production and the melodic line of speech (the latter
emphasizes the content of speech)exhaust the concrete shared properties between music and language.
Intoning the words in speech brings out the emotional component of human communication, but the
intoning can sharpen the meaning of words if and only if the words belong to a familiar vocabulary for the
listener. There is no real ground for linking linguistic syntax with musical syntax since the conventional
musical system, as Bigand et al. (2014) explain, is rooted in psychoacoustic properties of sound, and this
is not the case for linguistic syntax.”
The nonsymbolic, supremely abstract way of communication used in music stands in stark
contrast to the precision of linguistic language. The practice of translating the words into different
languages illustrates the reliability of the semantic certainty of words. Music offers precision of a very
different kind, related to the emotional aspect of human life. In the words of Susanne Langer (1942),
Because the forms of human feeling are much more congruent with music form than with the forms of
language, music can reveal the nature of feeling and a detail and truth that language cannot approach.”
Musical performance does, of course, demand physical effort and bodily movement. Skillful
playing on a musical instrument can give a tremendous pleasure to the performing musician, a pleasure that
is inseparable from the precision of conveyance of the melodic image and form. When the well-practiced
gestures inadvertently deliver wrong notes, it can be devastating for a musician and for the content of the
musical piece (even if the visible bodily movement looks the same for the wrong notes and the right notes).
Furthermore, it is the melodic thinking that dictates the visual appearance of the physical efforts of a
performing musician. In symmetry with the psychomotor program of the performer, the perception of
music consists of the integration of transitory moments of perceived tension; these moments are the result
of somatic responsiveness to details of the tonal relationships and rhythm.
Pedagogical practice does employ gestures to illustrate the expressive value of the sound volume
and melodic direction and to indicate the rounding of a musical phrase. Yet overall, neither music
pedagogy nor performing practice sees bodily movement as playing a significant role in expressing the
content of music. Unlike the great attention given to corporeal articulation in theater, where the meaning of
words can be modulated by gesture and by physical appearance in general, music pedagogy and performing
practices have been at best indifferent and at worst nonapproving towards expressive gesticulation during
instrumental performance. If music pedagogy had ever found an expressive value of bodily movements
(apart from solving specific technical problems related to playing a musical instrument), these movements
would have been studied and incorporated into pedagogical practicebut this is not the case.
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
8
From the nonexpert perspective, seeing a virtuoso musician playing a musical instrument is
exciting because of the ability of the musician to make music. For an expert musician, observing another
performing musician is usually of interest because the corporeal articulation can provide a hint as to how a
certain technical problem can be solved. For some listeners, the sight of an enthusiastically gesticulating
music performer can be entertaining, while for others it can be quite annoying and distracting from the
content of music. Differences in the interpretation of the same musical composition show that a deeply
expressive performance can be visually modest. Furthermore, playing the same melody on different
musical instruments produces different gestures that do not necessarily affect the meaningof the melody.
The problem of connecting observable gestures with musical content emerges as soon as we try to
compare the ordinary visuospatial perception that takes place in the 3-D world with the melodic and
harmonic reasoning used in the cyclical tonal space, which is devoid of symbolic representation. For
example, our research into tonal modulation revealed that the affective influence of melodic direction (an
aspect of music that is easy to illustrate with a gesture) depends on the overall tonal context. Specifically, a
melodic contourvisualized as a line in two-dimensional spaceexerts more affective influence in the
case of a musical phrase that modulates to a near key than it does in the case of a musical phrase that moves
to a distant key (currently, the 12-tone tonal space has its explanation in a toroidal model [Purwins, 2005]).
Unlike the readily available possibilities of connecting a certain gesture with a certain action in
speech, in music tonal tension and release are the minute constituents of a flow of sounds, where a specific
arrangement of the sounds in time generates a melodic and harmonic image that can be sensed as a gesture.
Other aspects of musicsuch as tempo and expressive timing (agogica and rubato), timbre, accentuation,
and variation in sound volumebelong to secondary qualia in relation to the tonal and temporal
relationships (Scruton, 1997). If a music performer’s gestures appear meaningful for the listeners, this is so
because these gestures skillfully follow the program of musical intentions encoded in the pattern of tonal
tension and release. This is why the empathic involvement and sense of gesture in music are related not so
much to visible corporeal articulation but, first of all, to tonal relationships arranged in time.
The unfolding of tonal and rhythmical patterns generates an emotion-forming process in music.
The corporeal articulation visible in music performance is simply the incidental by-product of playing a
musical instrument, one that generally tells us very littleand in a rather sporadic mannerabout the
content of a particular piece of music. Some of the gestures of a performing musician may occasionally
exert an affective influence on listeners, yet a rationale for the importance of seeing the performer in order
to understand the logic of emotion in music, particularly in a relatively complex musical composition, is
not clear. When a violinist makes a spectacular gesture after completing a virtuoso passage, this gesture is
locally specific and tells us little about the overall content of the musical composition. When, during the
rounding of a musical phrase, a pianist slows down the music’s pace and, on touching the final chord,
makes a visible transition from being physically tense to becoming physically relaxed, the causes of the
observable transformation reside in the tonal relationships that dictate the psychomotor program.
A discussion of musical content needs to be clear about the hierarchy of the musical aspects that,
when integrated, produce the complex phenomenon of aesthetic emotion. At the apex of this hierarchy are
the tonal and temporal relationships that are primary with regard to the secondary and tertiary properties in
music, such as timbre, sound volume, and the physical location of the sound source. For example, the first
few notes of a favorite melody can be initiated by a human voice, and then the melody can be picked up
and completed by a violin (or a trumpet, or a xylophone)yet the division of the melody among different
sound sources does not destroy the wholeness of the melodic object. This kind of distribution of a single
motif among different musical instruments can be found in Western classical chamber music and
symphonies, among other types of music. The source of musical sounds and their acoustical characteristics
such as spectral envelope (timbre), amplitude (sound volume), and register (diapason) constitute secondary
qualia with regard to the primary qualia of tonal relationships and the temporal organization of tones. Such
technology-based improvements as headphones and recontextualizations of tracks to obtain better sound
quality therefore do not belong to the musicological core.
The distinction between the acoustical characteristics of musical sounds and the tonal relationships
in the acousmatic space of melodic elements and musical forms is crucially important for describing music
and music cognition. The two main dimensions in musicthe tonal space and the arrow of time
(rhythm)are interwoven; together they constitute the tonal chronotope, the musical space populated by
melodic/harmonic structures. An intuition for tonal hierarchy and musical rhythm lets listeners recognize
music in an acoustic array; this intuition permits us to sense the difference between music and speech and
between music and the environmental noise emitted, for instance, by a working engine. When the acoustic
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
9
array “offers information about the identity of and location of vibratory events” (Kersten, 2014), this kind
of information is not imperative for music perception: The artifactual nature of music relies first and
foremost on the tonal and temporal organization of musical sounds.
Philosophic discourse on the temporality of musical experience and on the effect of musical
context (Phillips, 2014; Kon, 2014) could benefit from the studies addressing the “sliding temporal
window” in music perception (Tillmann & Bigand, 2004) and the time constraints for melodic perception
(for example, the concept of “melodic fission,” Dowling, 1973). The involvement of short-term memory
allows for the temporal integration of tonal events arranged in time in a specific order; the cognition of
melodic structures relies to a large extent on the sliding temporal window of music perception, which is
affected by the speed and complexity of tonal changes. For instance, the perception of reorientation of a
tonal scale on a different tonal center (tonal modulation) and the perception of melodic transformation,
when a transformed melody is compared to a template, both require a brief memory buffer.
The influence of tonal proximity on the estimate of duration of a modulating musical fragment
(Firmino et al., 2009) showed a fascinating connection between tonal distance and time perception in
music, when modulation to the furthest tonal distance elicited the shortest time estimation. An ongoing
EEG study by Radchenko et al. (in print) reveals the dependence of somatic changes on tonal proximity
and on the interaction of tonal proximity and the major and minor modes. These studies provide new,
important evidence for embodied cognition in music and indicate a new direction for understanding the
process of musical meaning formation.
CORPOREAL ARTICULATION AND MULTISENSORY INTEGRATION
Leman’s concept of the “human body as mediator in a musical meaning formation process” (2010) is
irrefutably fruitful for the field of music cognition. As a novel development in cognitive science, the field
of embodied cognition in music has been undergoing a process of clarification of criteria. Perhaps the
research into the effect of visible bodily movement on making sense of music actually addresses the
multisensory integration characteristic of attending to multidisciplinary performing art forms such as opera,
pop-music stage production, ballet, and the like.
The strongest argument for surface”-level embodied cognition in music comes from the studies in
the synchronization of peoples movements in response to tempo, meter, and rhythm (Naveda & Leman,
2009; Burger et al., 2013). Music has been used for millennia for accompanying important ceremonies and
for dancing and marching. Apart from the pleasantness of melody, the main value of music for collective
movement is musical meter and rhythm. The repetitiveness of musical meter makes it easier for a group of
people to move together (Repp & Su, 2013). The power of repetitiveness is so great that a melody can be
redundant with regard to eliciting synchronized movement among the listeners. Actually, synchronized
movement can be achieved without any melody. (Large & Gray [2015] showed that rhythmic entrainment
can be elicited from nonhuman primates.) Even for an art form as complex as ballet, until recently music
was recognized primarily as a handmaiden, a provider of musical meter (Martens, 1918). Among
outstanding attempts that have been made to illustrate music’s content with bodily movement are the
innovative works of the great contemporary choreographers Jiří Kylián and Nacho Duato Bárcia.
VISUAL GESTURES AND MEANING IN MUSIC
There is a special case in music performance when corporeal articulation is thoughtfully studied and
practiced with an eye to achieving technical virtuosity of execution. The art of conducting is defined by the
expressiveness and precision of gestures whose purpose is to assist a group of musicians in realizing the
shared goal of conveying the images and structure of a musical composition. A conductor’s gestures take
into consideration the specific nature of different musical instruments (such as brass versus stringed
instruments) and of the voice. These gestures are suggestive of the character of the music in relation to the
phrasing, variations in sound volume, melodic direction, and the details of tempo and rhythm, including
accentuation and grouping; all these aspects of conducting are dictated, first of all, by the tonal
relationships unfolding in time that generate musical images.
CONCLUSION
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
10
The proposed integrative theory of embodied cognition in music identifies two levels of bodily
involvement in music perception: the “surface” level of corporeal articulation of music and the “deep” level
of momentary somatic changes (including minute muscular actions) in response to a pattern of tonal and
temporal relationships. It is the joint task of the core areas of musicology (music theory, functional
analysis, and the analysis of musical form in historical context) and the science of music (particularly the
research into the psychophysics of melodic elements and the neurophysiology of psychomotor program in
music performance) to reinforce the theoretical foundation for the nascent understanding of music
perception as embodied cognition. Bearing in mind Panksepp’s proposal that “our subneocortical animalian
brain, with its many basic attentional, emotional, and motivational systems, may actually lie at the center of
our mental universe” (2004), research into music, the universally acknowledged language of emotion,
could prove to be crucially important for our understanding of the human mind.
ACKNOWLEDGMENTS
I would like to express my gratitude to Peter Borten for his valuable assistance with regard to the text of
this article.
NOTES
[1] Correspondence can be addressed to: Dr. Marina Korsakova-Kreyn, Touro College, E-mail:
marina.korsakova-kreyn@touro.edu or mnkors@gmail.com
REFERENCES
Bernardi, L., Porta, C., & Sleight, L. (2006). Cardiovascular, cerebrovascular, and respiratory changes
induced by different types of music in musicians and non-musicians: The importance of silence. Heart, 92,
445-452
Bidelman, G. M. & A. Krishnan. (2009). Neural correlates of consonance, dissonance, and the hierarchy of
musical pitch in the human brainstem. The Journal of Neuroscience, 29 (42), 1316513171
Bigand, E., Delbé, C., Poulin-Charronnat, B., Leman, M., & Tillmann, B. (2014). Empirical evidence for
musical syntax processing? Computer simulations reveal the contribution of auditory short-term memory.
Frontiers in Systems Neuroscience, 8: 94, doi: 10.3389/fnsys.2014.00094
Bigand, E., Parncutt, R., & Lerdahl, F. (1996). Perception of musical tension in short chord sequences: The
influence of harmonic function, sensory dissonance, horizontal motion, and musical training. Perception
and Psychophysics, 58, 125-141
Burger B., Thompson M. R., Luck G., Saarikallio S. and Toiviainen P. (2013) Influences of rhythm- and
timbre-related musical features on characteristics of music-induced movement. Frontiers of Psychology,
4(183), 1-10
Chanda, M. L. & D. J. Levitin (2013). The neurochemistry of music. Trends in Cognitive Sciences, 17 (4),
179-193
Chiandetti, C. & Vallortigara, G. (2011). Chicks like consonant music. Psychological science, 22(10),
1270-1273
Dowling W. J. (1972). Recognition of melodic transformations: Inversion, retrograde, and retrograde
inversion. Perception and Psychophysics, 12 (50), 417-421
Dowling, W. J. (1973). The perception of interleaved melodies. Cognitive Psychology, 5, 322-337
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
11
Ferrero, G. (1894). L’inertie mentale et la loi du moindre effort. Revue Philosophique de la France et de
l’Étranger, 37, 169182
Firmino, E. A., Bueno, J. L. O., & Bigand, E. (2009). Travelling through pitch space speeds up musical
time. Music Perception, 26(3), 205209
Florensky, P. A. (1925/1993). Анализ пространственности и времени в художественно-
изобразительных произведениях [Analysis of spatiality and time in the representational art]. Москва:
Прогресс [Moscow: Progress]
Foglia, L. & Wilson, R. A. (2013) Embodied cognition. WIREs Cognitive Science, doi: 10.1002/wcs.1226
Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., Friederici, A., & Koelsch, S.
(2009). Universal recognition of three basic emotions in music. Current Biology, 19 (7), 573-576
Frances, R. & Bruchon-Schweitzer, M. (1983). Musical expressions and body expression. Perceptual and
Motor Skills, 57, 587-595, pg. 594
Glenberg, A. M. & Kaschak, M. P. (2002). Grounding language in action. Psychonomic Bulletin & Review,
9, 558565
Goertzel, B., Aam, O., F. Smith, A. & Palmer, K. (2008). Mirror neurons, mirrorhouses, and the algebraic
structure of the self. Cybernetics and Human Knowing 15(1), 9-28
Holleran, S., Jones, M. & Butler, D. (1995). Perceived implied harmony. Journal of Experimental
Psychology, 21 (3), 737-753
James, K. H. & Maouene, J. (2009). Auditory verb perception recruits motor systems in the developing
brain: an fMRI investigation. Developmental Science, 12, F26F34
Kersten, L. (2014). Music and Cognitive Extension. Empirical Musicology Review, 9 (3-4), 193-202, pg.
194
Kon, M. (2014). The temporality of musical experience: philosophical models and embodiment. Empirical
Musicology Review, 9 (3-4), 213-223
Korsakova-Kreyn, M. (2005). Tonal processing and musical imagery: the tonal chronotope of music.
Dissertation Proposal, on file at The University of Texas at Dallas, web source: http://universe-of-
music.com/
Korsakova-Kreyn, M. (2014). Bach, Escher, and mental rotation: An empirical study in the perception of
visual and melodic congruency. In A. Kozbelt (Ed.), Proceedings of the Twenty-Third Biennial Congress of
the International Association of Empirical Aesthetics, 495-504
Korsakova-Kreyn, M. & Dowling, W. J. (under review). Mental Rotation in Visual and Musical Space:
Comparing Pattern Recognition in Different Modalities
Korsakova-Kreyn, M. & W. Jay Dowling. (2014). Emotional processing in music: Study in affective
responses to tonal modulation in controlled harmonic progressions and real music. Psychomusicology,
24(1), 4-20
Kousta, S.-T., Vigliocco, G., Vinson, D.P., Andrews, M. & Del Campo, E. (2011). The representation of
abstract words: why emotion matters. Journal of Experimental Psychology, 140(1), 14-34
Krumhansl, C. L. (1996). A perceptual analysis of Mozart’s piano sonata K. 282: Segmentation, tension,
and musical ideas. Music Perception, 13, 401432
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
12
Krumhansl, C. L. (1997). An exploratory study of musical emotions and psychophysiology. Canadian
Journal of Experimental Psychology, 51, 336-335
Krumhansl, C.L. & Kessler, E. (1982). Tracing the dynamic changes in perceived tonal organization in a
spatial representation of musical keys. Psychological Review, 89, 334368
Langer, S. (1942/2009). Philosophy in a New Key: A Study in the Symbolism of Reason, Rite and Art.
Harvard University Press. pg. 235
Large, E.W. & Gray, P.M. (2015). Spontaneous Tempo and Rhythmic Entrainment in a Bonobo. Journal of
Comparative Psychology, Advance online publication: http://dx.doi.org/10.1037/com0000011
Lehne, M., Rohrmeier,M., Gollmann, D.,& Koelsch, S. (2013). The influence of different structural
features on felt musical tension in two piano pieces by Mozart and Mendelssohn. Music Perception: An
Interdisciplinary Journal, 31(2), 171-185
Leman, M. (2010). An embodied approach to music semantics. Musicae Scientiae, 14(1), 43-67; doi:
10.1177/10298649100140S104
Leman, M. & Maes, P.-J. (2014). The Role of Embodiment in the Perception of Music. Empirical
Musicology Review, 9 (3-4), 236-246
Lerdahl, F. & Krumhansl, C. L. (2007). Modelling tonal tension. Music Perception, 24(4), 329-366
Madsen, C. K. & Fredrickson, W. E. (1993). The experience of musical tension: A replication of Nielsen's
research using the continuous response digital interface. Journal of Music Therapy, 30, 46-63
Martens, F. H. (1918). The attitude of the dancer toward music. The Musical Quarterly, 4(3), 440-449
Naveda, L. & Leman, M. (2009). A cross-modal heuristic for periodic pattern analysis of samba music and
dance. Journal of New Music Research, 38(3), 255283
Panksepp, J. (2005) Affective consciousness: Core emotional feelings in animals and humans.
Consciousness and Cognition, 14(1), 30-80
Phillips, M. (2014). A discussion of the effect of musical context on perception of succession of musical
events, and its relevance for philosophical models of temporal experience. Empirical Musicology Review, 9
(3-4), 230-235
Purwins H. (2005). Profiles of pitch classes circularity of relative pitch and key experiments, models,
computational music analysis, and perspectives, Ph.D. Dissertation. Elektrotechnik und Informatik der
Technischen Universität Berlin
Radchenko, G. S., Bakhchina, A. V., Parin, S. A., Polevaya, S. A., Korsakova-Kreyn, M. N., & Fedotchev,
A. I. (in press). Psychophysical correlates of the perception of motion in tonal space.
Repp, B.H. & Su, Y.H. (2013). Sensorimotor synchronization: a review of recent research (2006-2012).
Psychonomic Bulletin & Review, 20(3), 403-452
Scruton, R. (1997). The Aesthetics of Music. Oxford University Press
Shepard, R. & Cooper, L. (1982). Mental images and their transformations. Cambridge, MA: MIT Press
Korsakova-Kreyn, Embodied Cognition in Music, June 2015
13
Solntsev, V. M. (1974). К вопросу о семантике, или языковом значении (вместо предисловия) [On a
question of semantics or linguistic meaning]. Проблемы семантики [Problems of semantics], Москва:
Наука [Moscow: Science]
Tillmann, B. & E. Bigand (2004). The relative importance of local and global structures in music
perception. Journal of Aesthetics and Art Criticism 62, (2), 211222
Trainor, L.J. (2004). Are there critical periods for music development? Developmental Psychobiology, 46,
262-278
Virtala, P., Huotilainen, M., Partanen, E., Fellman, V. & Tervaniemi, M. (2013). Newborn infants' auditory
system is sensitive to Western music chord categories. Frontiers in Psychology, 4, 492. doi:
10.3389/fpsyg.2013.00492
Wellsby, M. & Pexman, P.M. (2014). Developing embodied cognition: insights from children’s concepts
and language processing. Frontiers in Psychology, 5, 1-10
... Therefore, it also accounts for music semantics and, indeed, some work has already been published following this cognitive perspective. Korsakova-Kreyn [8] emphasizes the role of modulations of physical tension in music semantics. These modulations are conveyed by tonal and temporal relationships in music artifacts, consisting of manipulations of tonal tension in a tonal framework (musical scale). ...
Preprint
Full-text available
Embodied cognition states that semantics is encoded in the brain as firing patterns of neural circuits, which are learned according to the statistical structure of human multimodal experience. However, each human brain is idiosyncratically biased, according to its subjective experience history, making this biological semantic machinery noisy with respect to the overall semantics inherent to media artifacts, such as music and language excerpts. We propose to represent shared semantics using low-dimensional vector embeddings by jointly modeling several brains from human subjects. We show these unsupervised efficient representations outperform the original high-dimensional fMRI voxel spaces in proxy music genre and language topic classification tasks. We further show that joint modeling of several subjects increases the semantic richness of the learned latent vector spaces.
Article
Full-text available
The development of the mind follows the path of biological evolution towards the accumulation and transmission of information with increasing efficiency. In addition to the cognitive constants of speech (Solntsev, 1974), which greatly improved the transmission of information, people have created computing devices, from the abacus to the quantum computer. The capabilities of computers classified as artificial intelligence are developing at a rapid pace. However, at the present stage, artificial intelligence (AI) lacks an emotion module, and this makes AI fundamentally different from human intelligence, since the life of the mind in humans cannot be separated from their feelings (Damasio, 2010; Panksepp, 1997). Consciousness itself is formed through the sensory and motor systems, that is, it is embodied (Foglia & Wilson, 2013), which means that our mental life is inseparable from our sensory motor experience (Wellsby & Pexman, 2014). Evolutionarily, our minds rely on ancient survival mechanisms that influence our decisions and choices. Hence, for example, the question whether the choice of Artificial Intelligence will always be favorable for humanity.
Chapter
Resonanzprozesse entstehen sowohl im Konzertsaal, zwischen MusikerInnen und Publikum, als auch in der psychotherapeutischen Beziehung. Doch wie lassen sich diese Schwingungen, gleichzeitig allgegenwärtig und schwer zu fassen, einfangen? Die BeiträgerInnen widmen sich dieser faszinierenden Schnittstelle zwischen Musik und Psychoanalyse aus interdisziplinärer Perspektive. So werden Resonanzprozesse u.a. aus kulturtheoretischer und klinisch-praktischer, aus psychoanalytischer und musikwissenschaftlicher Sicht zugänglich. Die AutorInnen gehen dabei Stimmungen und Erfahrungen, die in therapeutischen und musikalischen Resonanzprozessen mitschwingen, auf den Grund und forschen nach Zusammenhängen zwischen Biografie und künstlerischem Schaffen. Mit Beiträgen von Seong-u Bak, Aleksandar Dimitrijević, Ingrid Erhardt, Barbara Gindl, Sebastian Leikert, Christopher Mahlstedt, Pietro Massa, Wolfgang Mastnak, Wolfgang Tschacher und Herbert Will.
Article
Full-text available
Artificial intelligence and machine learning capabilities are developing at lightning speed. However, these machines lack the module of emotional processing, which makes artificial intelligence fundamentally different from human intelligence. The intellectual development of humans follows the path of the biological evolution towards the accumulation and transfer of information with increasing efficiency. Apart from the cognitive constant of speech, which dramatically enhanced the transfer of information, humans have advanced their computational power by using various devices, from the abacus to the quantum computer. Yet the mental lives of humans cannot be separated from feelings. The neurodynamics of emotion is still poorly understood; this is the greatest challenge to creating artificial intelligence on par with human intelligence. Keywords: artificial intelligence; embodied cognition; emotion; social matrix; triune brain
Research
Full-text available
This is my masters dissertation discussing the origin of music from a cognitive science perspective.
Thesis
Full-text available
We can conceptualize the art of music as the organization of some space governed by two unifying principles: musical " gravity " or tonal attraction and temporal (gross) structuring. The review presents the results of behavioral studies that give insight into primal auditory processing—for instance how our brains generate tonal schemata—and into the processing of such complex musical constituent as a melodic contour. It also presents imaging and behavioral studies showing what regions of the brain are vital to music perception and how the brains of musicians differ from the brains of non-musicians. We speculate that a particular region of the cortex may be generally involved in pattern recognition notwithstanding the modality of the patterns, namely that the superior parietal lobule (BA7) is perhaps involved in both mental rotation and music perception. 2
Article
Full-text available
The article reviewed in this commentary takes philosophical models of temporal experience as its starting point, in an exploration of how an "experience of succession" may be distinguished from a mere "succession of experience". It is proposed that context is the important factor in differentiating the two experiences, rather than duration. Context is accounted for in broad terms, with specific discussion of gesture, performance environment, and mental imagery. The discussion may usefully pave the way for future collaboration between philosophers and psychologists. However, there are multiple fundamental findings in music cognition research, vital to any consideration of the context in which musical experiences occur (e.g. meter, tonality, expectation, familiarity), that could be factored into this discussion. The ideas discussed could be developed with greater consideration of recent empirical studies in music perception. Perhaps then a theoretical model of the experience of succession of musical events could give rise to experimental hypotheses, which may then be tested in order to further refine such models.
Conference Paper
Full-text available
Among the most fascinating aspects of music are the quasispatial properties of tonal space and structures. For instance, any conventional melody can be visualized as a contour—as a combination of ups and downs. A melodic contour can be “bent” by tonal forces, and all its melodic intervals can be “mirrored.” In addition, a contour can be augmented and diminished in duration, which is conceptually similar to a proportional enlargement and diminution of a visual object. Certain musical compositions, specifically Baroque fugues, resemble divisions of the plane, such as in M. C. Escher’ tessellations and M. K. Čiurlionis’ paintings. When the tonal space of music is visualized as phenomenal gravity (Scruton, 1997), tonal melodies can be explained as melodic objects shaped by the tonal force field. We hypothesized that the cognitive processing of these transformations may draw on spatial abilities developed for visuospatial reasoning and that the perception of melodic transforms involves the same neural substrate in the parietal cortex (specifically BA 7), which is engaged in visuospatial processing. We conducted a behavioral study that compared the perception of congruency of 3D geometrical figures and congruency of tonal melodies. The visuospatial task used a set of stimuli from a study in mental rotation (Shepard and Metzler, 1971), whereas the melodic rotation task used a set of melodies selected from the clavier compositions of J. S. Bach. The same melodies were used for a control task on timbre judgment. Performance was positively correlated for visual and melodic spatial tasks (r = .40), but the pattern of correlations overall differed between sexes. Males performed better than females on all three tasks. The obtained results converge with previous investigations that found gender effect in the visuospatial mental rotation task, and offer new information about gender effect in music perception. The discovery of a positive correlation between melodic and visual spatial tasks provides supporting evidence for the hypothesis of supramodal processing in music perception and inspires new directions for discussing the aesthetic experience. Keywords: music perception, supramodal processing, mental rotation, melodic transformation, gender effects
Article
Full-text available
Since its breakthrough at the beginning of the twenty-first century, the embodied music cognition theory has inspired empirical research on the role of action-perception couplings in musical activities. The integration of novel technologies and analysis methods inspired empirical research advancing knowledge regarding the role of embodiment in music perception and musical signification processes. In this paper, we present recent and on-going research in the field of embodied music cognition, with a focus on studies conducted at IPEM, the research laboratory in systematic musicology at Ghent University, Belgium. Attention is devoted to encoding/decoding principles underlying musical expressiveness, synchronization and entrainment, and action-based effects on music perception. The discussed empirical findings demonstrate that embodiment is only one component in an interconnected network of sensory, motor, affective, and cognitive systems involved in music perception. Currently, these findings drive the embodiment theory towards a more dynamical approach in which the interaction between various internal processes and the external environment are of central importance. Additionally, this approach envisions practical outcomes in the field of music affect research, wellbeing, healing, sports, music engineering, and brain studies.
Article
Full-text available
p class="Default">Extended cognition holds that cognitive processes sometimes leak into the world (Dawson, 2013). A recent trend among proponents of extended cognition has been to put pressure on phenomena thought to be safe havens for internalists (Sneddon, 2011; Wilson, 2010; Wilson & Lenart, 2014). This paper attempts to continue this trend. It is argued that because music perception is as part of a locationally wide computational system, it is an extended process. In articulating the view, the work of J.J Gibson (1966, 1986) and Robert Wilson (1994b, 1995, 2004) is drawn on. The view is defended from objections and its implications outlined. The paper concludes with a comparison to Krueger’s (2014) view of the “musically extended emotional mind”. <br /
Article
Full-text available
The emergence of speech and music in the human species represent major evolutionary transitions that enabled the use of complex, temporally structured acoustic signals to coordinate social interaction. While the fundamental capacity for temporal coordination with complex acoustic signals has been shown in a few distantly related species, the extent to which nonhuman primates exhibit sensitivity to auditory rhythms remains controversial. In Experiment 1, we assessed spontaneous motor tempo and tempo matching in a bonobo (Pan paniscus), in the context of a social drumming interaction. In Experiment 2, the bonobo spontaneously entrained and synchronized her drum strikes within a range around her spontaneous motor tempo. Our results are consistent with the hypothesis that the evolution of acoustic communication builds upon fundamental neurodynamic mechanisms that can be found in a wide range of species, and are recruited for social interactions. (PsycINFO Database Record (c) 2015 APA, all rights reserved).
Article
Full-text available
The present series of experiments used either musical excerpts which subjects listened to and then associated to verbal categories (Exp. I) or videotape recordings of body expression sequences which subjects watched and associated with the same verbal categories. These videotapes were recorded by “actors” under two conditions, (a) after the actors had listened to the musical excerpts (Exp. II) and (b) following the induction in actors of the verbal categories which had been associated to the excerpts in Exp. I (Exp. III). The results demonstrate a partial equivalence between the verbal categories associated with the music in Exp. I and those associated with the videotape recording of Exps. II and III. This partial equivalence of the verbal categories is explained by the existence of a bodily core to musical expression.
Thesis
Die Beziehungen der 24 Dur- und moll-Tonarten lassen sich doppelt-zirkulär als Torus darstellen. Wir zeigen die Konvergenz der Herleitungen auf den Beschreibungsebenen a) des psychoakustischen Experiments, b) der geometrischen Musiktheorie und c) der Computersimulation des Musikhörens. Shepard (1964) verallgemeinernd wird zirkuläre Tonhöhenwahrnehmung experimentell für vollständige Obertonspektren nachgewiesen und mathematisiert als spectrum of pitch differences, der Übertragung der virtuellen Tonhöhe (Terhardt, 1982) auf Intervalle. Musikalische Beispiele von zirkulärer Tonhöhen-, Tempo- und Lautheitssteigerung werden analysiert. Die neue eingeführte constant quotient (CQ-)profile Methode berechnet aus einer Musikaufnahme für jedes Chroma die spezifische Intensität. Die Methode ist a) konsistent mit psychologischen Profilen (Krumhansl und Kessler, 1982), b) hocheffizient, c) echtzeitfähig, d) rauschresistent, e) anwendbar auf Transpositionen, f) bis auf die annähernd gleichschwebend temperierte Stimmung frei von musikalischen Voraussetzungen, g) erhaltend bezüglich wesentlicher musikalischer Merkmale (Stil, Komponist, Tendenz zur Chromatik und zur Verdurung / Vermollung) in hochkomprimierter Darstellung. Von Bach bis Schostakowitsch werden letztere Merkmale extrahiert durch Klassifikation (support vector machine, regularisierte Diskriminanzanalyse), Clustering und Visualisierung (Isomap). Leman (1995) modelliert das Erlernen von Tonartenbeziehungen als neuro-inspirierte toroidale selbstorganisierende Kohonen-Karte (1982), die mit künstlichen, von einem Ohrmodell vorverarbeiteten Kadenzen trainiert wird. Wir verallgemeinern den Ansatz a) durch Erweiterung der Stimuli auf Musikaufnahmen und b) durch Verzichten auf die toroidale Struktur als Voraussetzung. Die CQ-Profile von Bachs WTK evozieren sowohl in der Korrespondenzanalyse als auch in der Isomap den Quintenzirkel. Toroidale Modelle von Tonartenrelationen (TOMIR) bilden sich heraus a) in toroidalen Kohonen-Karten, trainiert mit Chopins Preludes, b) in der Korrespondenzanalyse der Chroma-Häufigkeitsprofile, extrahiert aus den Noten (Bachs WTK I) und c) in der neu eingeführten topographic ordering map, einer Mathematisierung der musiktheoretischen Herleitung des TOMIR aus Webers (1817) Tonartentabelle. Die Resultate sind konsistent untereinander und mit der Visualisierung musikpsychologischer Bewertungsprofile. Ferner werden fuzzy distance, räumliche und systematisierte synästhetische Visualisierungsmethoden auch auf mehreren synchronen Zeitskalen zum automatischen Verfolgen von Tonartenmodulationen vorgeschlagen. Außerdem werden Statistiken zur Tonartenpräferenz erhoben und darauf Beziehungen zwischen Komponisten visualisiert. Die vorgestellten Methoden sind relevant für die Musikwissenschaft, Information Retrieval (MPEG-7), automatische Analyse, Echtzeit-, und audio-basierte Musiktechnologie.
Article
This paper looks at a framework for music semantics, based on the concept of embodied music cognition (Leman, 2007). The first part presents the representational viewpoint on semantics and why one should aim at transcending this. The second part deals with different types of semantics, ranging from more mental to more corporeal ones. The third part discusses issues of semantics related to embodiment. It is argued that the embodiment approach, by focusing on the role of the human body as mediator in a musical meaning formation process, may be a useful extension, and may perhaps be an alternative to the representational approach. An embodied approach to music semantics is fully embedded within an empirical and evidence-based research approach. Within this approach, methodologies from both the natural sciences and the human sciences are combined. It focuses on the human body as a mediator between mind and physical environment, thus allowing a fully multimodal approach to music semantics.
Article
A basic issue about musical emotions concerns whether music elicits emotional responses in listeners (the 'emotivist' position) or simply expresses emotions that listeners recognize in the music (the 'cognitivist' position). To address this, psychophysiological measures were recorded while listeners heard two excerpts chosen to represent each of three emotions: sad, fear, and happy. The measures covered a fairly wide spectrum of cardiac, vascular, electrodermal, and respiratory functions. Other subjects indicated dynamic changes in emotions they experienced while listening to the music on one of four scales: sad, fear, happy, and tension. Both physiological and emotion judgments were made on a second-by-second basis. The physiological measures all showed a significant effect of music compared to the pre-music interval. A number of analyses, including correlations between physiology and emotion judgments, found significant differences among the excerpts. The sad excerpts produced the largest changes in heart rate, blood pressure, skin conductance and temperature. The fear excerpts produced the largest changes in blood transit time and amplitude. The happy excerpts produced the largest changes in the measures of respiration. These emotion-specific physiological changes only partially replicated those found for non-musical emotions. The physiological effects of music observed generally support the emotivist view of musical emotions.