ArticlePDF Available

Abstract and Figures

On the basis of previous results showing that music training positively influences different aspects of speech perception and cognition, the aim of this series of experiments was to test the hypothesis that adult professional musicians would learn the meaning of novel words through picture-word associations more efficiently than controls without music training (i.e., fewer errors and faster RTs). We also expected musicians to show faster changes in brain electrical activity than controls, in particular regarding the N400 component that develops with word learning. In line with these hypotheses, musicians outperformed controls in the most difficult semantic task. Moreover, although a frontally distributed N400 component developed in both groups of participants after only a few minutes of novel word learning, in musicians this frontal distribution rapidly shifted to parietal scalp sites, as typically found for the N400 elicited by known words. Finally, musicians showed evidence for better long-term memory for novel words 5 months after the main experimental session. Results are discussed in terms of cascading effects from enhanced perception to memory as well as in terms of multifaceted improvements of cognitive processing due to music training. To our knowledge, this is the first report showing that music training influences semantic aspects of language processing in adults. These results open new perspectives for education in showing that early music training can facilitate later foreign language learning. Moreover, the design used in the present experiment can help to specify the stages of word learning that are impaired in children and adults with word learning difficulties.
Content may be subject to copyright.
Professional Music Training and Novel Word Learning:
From Faster Semantic Encoding to Longer-lasting
Word Representations
Eva Dittinger
1
, Mylène Barbaroux
1
, Mariapaola DImperio
1
, Lutz Jäncke
2
,
Stefan Elmer
2
*, and Mireille Besson
1
*
Abstract
On the basis of previous results showing that music training
positively influences different aspects of speech perception and
cognition, the aim of this series of experiments was to test the
hypothesis that adult professional musicians would learn the
meaning of novel words through pictureword associations
more efficiently than controls without music training (i.e., fewer
errors and faster RTs). We also expected musicians to show fas-
ter changes in brain electrical activity than controls, in particular
regarding the N400 component that develops with word learn-
ing. In line with these hypotheses, musicians outperformed
controls in the most difficult semantic task. Moreover, although
a frontally distributed N400 component developed in both
groups of participants after only a few minutes of novel word
learning, in musicians this frontal distribution rapidly shifted
to parietal scalp sites, as typically found for the N400 elicited
by known words. Finally, musicians showed evidence for better
long-term memory for novel words 5 months after the main ex-
perimental session. Results are discussed in terms of cascading
effects from enhanced perceptiontomemoryaswellasin
terms of multifaceted improvements of cognitive processing
due to music training. To our knowledge, this is the first report
showing that music training influences semantic aspects of lan-
guage processing in adults. These results open new perspec-
tives for education in showing that early music training can
facilitate later foreign language learning. Moreover, the design
used in the present experiment can help to specify the stages of
word learning that are impaired in children and adults with
word learning difficulties.
INTRODUCTION
The aim of the present experiment was to examine the
influence of music training on word learning using both
behavioral and electrophysiological measures. On the ba-
sis of the evidence reviewed below, we tested the hy-
pothesis that musicians would be more efficient at
word learning than nonmusicians and that the differ-
ences would be reflected in their pattern of brain waves.
There is strong evidence from previous cross-sectional
studies comparing adult musicians and nonmusicians
that long-term music training promotes brain plasticity
(Münte, Altenmüller, & Jäncke, 2002) in modifying the
functional (Schneider et al., 2002; Pantev et al., 1998)
and structural (Elmer, Hänggi, Meyer, & Jäncke, 2013;
Gaser & Schlaug, 2003; Schneider et al., 2002) architec-
ture of the auditory pathway. Results of longitudinal stud-
ies, mostly in children, showed that music training can be
the cause of the observed effects (François, Chobert,
Besson, & Schön, 2013; Strait, Parbery-Clark, OConnell,
& Kraus, 2013; Chobert, François, Velay, & Besson, 2012;
Moreno et al., 2011; Hyde et al., 2009; Moreno et al.,
2009). Most importantly for the present purposes, there
is also evidence that music training improves different
aspects of speech processing (for review, see Asaridou &
McQueen, 2013; Besson, Chobert, & Marie, 2011; Kraus &
Chandrasekaran, 2010). These transfer effects possibly
arise because speech and music are auditory signals rely-
ing on similar acoustic cues (i.e., duration, frequency, in-
tensity, and timbre) and because they share, at least in
part, common neuronal substrates for auditory perception
(Peretz, Vuvan, Lagrois, & Armony, 2015; Jäncke, 2009)
and for higher-order cognitive processing (Rogalsky, Rong,
Saberi, & Hickok, 2011; Patel, 2008; Maess, Koelsch,
Gunter, & Friederici, 2001). For instance, music training fa-
cilitatesthe processing of a variety of segmental (Bidelman,
Weiss, Moreno, & Alain, 2014; Kühnis, Elmer, & Jäncke,
2014; Elmer, Meyer, & Jäncke, 2012; Chobert, Marie,
Fraois, Schön, & Besson, 2011; Musacchia, Sams, Skoe,
& Kraus, 2007) and suprasegmental speech attributes (Marie,
Delogu, Lampis, Olivetti Belardinelli, & Besson, 2011; Wong
& Perrachione, 2007) within native (Schön, Magne, &
Besson, 2004) and nonnative languages (Marques, Moreno,
Castro, & Besson, 2007). Moreover, both musically trained
children (Jentschke & Koelsch, 2009) and adults (Fitzroy &
Sanders, 2013) are more sensitive to violations of linguistic
and music syntax than participants without music training.
Perhaps most importantly, recent results also showed that
1
CNRS and Université Aix Marseille,
2
University of Zurich
*Shared last authorship.
© 2016 Massachusetts Institute of Technology Journal of Cognitive Neuroscience 28:10, pp. 15841602
doi:10.1162/jocn_a_00997
long-term music training positively improves cognitive func-
tions such as auditory attention (Strait, Slater, OConnell, &
Kraus, 2015), visual attention ( Wang, Ossher, & Reuter-
Lorenz, 2015), working and verbal memory (George &
Coch, 2011; Ho, Cheung, & Chan, 2003), executive func-
tions (Zuk, Benjamin, Kenyon, & Gaab, 2014; Moreno
et al., 2011; Pallesen et al., 2010), and general intelligence
(Schellenberg, 2004). These findings are not surprising in-
sofar as playing an instrument at a professional level is a
multidimensional task that, together with specific motor
abilities, requires acute auditory perception and focused at-
tention, code switching between the visual information on
the score andthe corresponding sounds, aswell as the abil-
ity to maintain auditory information in short- and long-term
memory. Taken together, these results are in line with dy-
namic models of human cognition (Friederici & Singer,
2015; Hagoort, 2014) positing that languageand possibly
musicare processed in interaction with other cognitive
functions.
Similar to playing music, word learning is also a multi-
dimensional task requiring both perceptive and higher-
order cognitive abilities. Let us take the example of Thai.
Thai is a tonal and a quantitative language that mainly
comprises monosyllabic words (as Mandarin Chinese
and other tonal languages) and in which both tonal (i.e.,
five tones) and vowel length contrasts are linguistically rel-
evant for understanding word meaning (e.g., /pa1/ low
tone with a short vowel means to findand /pa:1/ low
tone with a long vowel means forest; Gandour et al.,
2002). Thus, when it comes to learn novel words in Thai,
the learner has to focus attention on the acoustic stream
to discriminate spectral and temporal phonetic contrasts
and to build new phonological representations that can
then be associated with lexical meaning by recruiting
working, short-term, episodic, and semantic memory pro-
cesses. Thus, if music skills translate into improved audi-
tory perception and attention together with enhanced
working and verbal memory, it should be easier for musi-
cians to learn a language such as Thai.
The ERP method is one of the most eligible methods
to capture the fast temporal dynamics of word learning
and to examine brain plasticity, as reflected by changes
in the amplitude and/or latency of ERP components dur-
ing learning. Previous results in adults have shown that
the N400, a negative-going component that typically de-
velops between 300 and 600 msec poststimulus onset
(Kutas & Hillyard, 1980), increases in amplitude when
meaningless items acquired meaning. Specifically, results
showed N400 enhancements in native English speakers
after 14 hr of learning the meaning of novel French
words (McLaughlin, Osterhout, & Kim, 2004) and after
45 min of learning the meaning of rare words (e.g., clow-
der; Perfetti, Wlotko, & Hart, 2005). Moreover, if a novel
word (Borovsky, Elman, & Kutas, 2012; Borovsky, Kutas,
& Elman, 2010; Mestres-Missé, Rodriguez-Fornells, &
Münte, 2007) or pseudoword (Batterink & Neville,
2011) is presented in a strongly constrained and mean-
ingful context, even a single exposure can be sufficient
to build up initial word representations, an effect referred
to as fast mapping(Carey, 1978). An incubation-like pe-
riod and further exposures are then required for consol-
idation and integration into existing lexical networks
(Dumay & Gaskell, 2007). Thus, the N400 is taken as a
reliable index of word learning, reflecting the formation
of semantic representations.
Note though that the N400 component at the core of
the above-mentioned experiments clearly showed a more
frontal scalp distribution (Borovsky et al., 2010; Mestres-
Missé et al., 2007) than the centroparietal N400 typically
elicited by already known words (Kutas, Van Petten, &
Besson, 1988). This frontal N400 distribution is compati-
ble with results showing that prefrontal and temporal
brain regions are associated with the maintenance of novel
information in working or short-term memory and the for-
mation of new associations (Hagoort, 2014) and/or with
the initial building-up of word representations in episodic
memory (Rodriguez-Fornells, Cunillera, Mestres-Missé, &
De Diego-Balaguer, 2009; Wagner et al., 1998).
As mentioned above, most studies of music-to-language
transfer effects have focused on segmental, suprasegmen-
tal, and syntactic processing levels. On the basis of the re-
sults on word learning reviewed above, this study aimed at
going a step further so as to determine whether profes-
sional music training would also influence the semantic
level of processing, most often considered as language-
specific (but see Koelsch et al., 2004) by facilitating the
learning process of novel word meaning. The general hy-
pothesis was that the optimization of perceptual and cog-
nitive functions in professional musicians would positively
influence the speed and quality of word learning as re-
flected by a behavioral advantage for musicians (i.e., lower
error rates [ERRs] and faster RTs). Moreover, based on the
ERPs and word learning literature (Borovsky et al., 2010,
2012; Batterink & Neville, 2011; Mestres-Missé et al.,
2007; Perfetti et al., 2005; McLaughlin et al., 2004), we ex-
pected a frontally distributed N400 component to develop
in all participants during the early stages of novel word
learning. However, if the perceptual and cognitive compu-
tations involved in word learning were facilitated in musi-
cians, the development of the N400 component should be
faster in musicians than in controls in the learning phase.
By contrast, we expected the N400 to show a centroparietal
distribution when novel word learning was consolidated.
To test these general hypotheses, we used an ecolog-
ically valid experimental design inspired by Wong and
Perrachione (2007) and based on a series of four experi-
ments that comprised several tasks performed during
the main experimental session (see Figure 1AE). First, to
further test the hypothesis of improved auditory speech
discrimination in musicians compared with controls, par-
ticipants performed a phonological categorization task at
the beginning and at the end of the main experimental
session (see Figure 1A and E). On the basis of previous
results showing that musicians are advantaged when
Dittinger et al. 1585
the discrimination is most difficult (Diamond, 2013; Schön
et al., 2004), we expected musicians to outperform controls
in identifying phonemic contrasts that are not relevant for
lexical discrimination in French. Moreover, based on previ-
ous literature reporting that the N100 component reflects
encoding of auditory cues in the auditory-related cortex
(Kühnis et al., 2014) and is influenced by auditory attention
and perceptual learning (Seppänen, Hämäläinen, Pesonen,
& Tervaniemi, 2012; Woldorff & Hillyard, 1991), we ex-
pected this behavioral advantage to be accompanied by
an increased N100 amplitude in musicians.
Second, participants were asked to learn the meaning
of the novel words through pictureword associations
(see Figure 1B), a design that has often been used in word
learning experiments in children (Friedrich & Friederici,
2008; Torkildsen et al., 2008) and in adults (Dobel,
Lagemann, & Zwitserlood, 2009). No behavioral response
was required during this word learning phase, but ERPs
were recorded to test the main hypothesis that frontally
distributed N400 components would develop in both
groups of participants (François et al., 2013; Borovsky
et al., 2010; Rodriguez-Fornells et al., 2009) but with faster
temporal dynamics in musicians than in controls.
Third, to test for the efficacy of the learning phase, par-
ticipants performed a matching task and were asked to
decide whether a pictureword pair matched or mis-
matched the previously learned pairs (see Figure 1C).
Fourth, an important aspect was to determine whether
word learning was specific to the pictureword pairs
learned during the word learning phase or whether the
meaning of the newly learned words was already integrated
into semantic networks so that priming effects generalized
to new pictures. To this end, participants performed a se-
mantic task during which novel pictures that had not been
seen in the previous tasks were presented. They were
asked to decide whether the picture and the word were
semantically related or unrelated (see Figure 1D). In both
the matching and the semantic tasks and in both groups
of participants, we predicted that N400 amplitudes would
be larger for mismatch and semantically unrelated words
than for match and semantically related words (i.e., the typ-
ical N400 effect; Kutas & Hillyard, 1980), thereby showing
that participants had learned the meaning of the novel
words. Moreover, if the novel wordsmeanings were
already integrated into existing semantic networks at the
end of the word learning phase (Borovsky et al., 2012;
Figure 1. Experimental design. Participants performed a series of tasks in the main experimental session (AE): First, in the phonological categorization
task (A), nine natural Thai monosyllabic words had to be categorized based on voicing (Task 1), vowel length (Task 2), pitch (Task 3), or aspiration
contrasts (Task 4). Second, in the word learning phase (B), each word was paired with its respective picture. This phase included two separate blocks of
trials. Third, in the matching task (C), the words were presented with one of the pictures, either matching or mismatching the previously learned
associations. This task included two separate blocks of trials. Fourth, in the semantic task (D), the words were presented with novel pictures that were
either semantically related or unrelated to the novel words. Again, this task included two separate blocks of trials. Fifth, participants did again the four
tasks of the phonological categorization task (E). Finally, participants came back 5 months after the main session to perform again the matching and
semantic tasks (F).
1586 Journal of Cognitive Neuroscience Volume 28, Number 10
Batterink & Neville, 2011), we expected the N400 effect
(mismatchingmatching and unrelatedrelated words) in
the matching and semantic tasks to show the centroparie-
tal distribution typically found for already known words
(Kutas et al., 1988). Of main interest was to specify the spa-
tiotemporal dynamics of the N400 effect in musicians and
in nonmusicians. Finally, if we were tofind that music train-
ing influenced word learning, then we expected musical
ability to be positively correlated with word learning effi-
cacy, as reflected by behavioral measures and/or the
N400 effect in the matching and semantic tasks.
Finally, a subset of the participants was behaviorally re-
tested after 5 months (see Figure 1F) in the matching and
semantic tasks to evaluate the maintenance of novel
words in long-term memory. It was of interest to deter-
mine whether the behavioral advantages of musicians in
a variety of cognitive domains, as reviewed above, extend
to long-term memory. To the best of our knowledge, this
aspect has not yet been investigated.
In summary, this experimental design is relevant for
specifying whether music expertise influences the se-
mantic level of speech processing, an issue that, to our
knowledge, has not been addressed before. By analyzing
ERPs, we aimed at better understanding the dynamics of
word learning, how fast semantic processes develop, and
whether and how the N400 is influenced by music train-
ing. Showing that long-term music training with an early
start (as it is most often the case in professional musi-
cians) may facilitate foreign language learning later in life
should add evidence to the claim that music training has
important societal consequences for education (Besson
et al., 2011; Kraus & Chandrasekaran, 2010). Finally, this
experimental design is of potential interest for clinical re-
search: using several different tasks that call upon several
perceptual and cognitive functions (phonological catego-
rization, formation of pictureword associations, main-
taining these associations in short-term and long-term
memory and generalization of learning effects) within
the same patient may help specify the processing stages
that are deficient in adults or children with language
learning disorders.
METHODS
Participants
A total of 30 participants with 15 professional musicians
(MUS, eight women) and 15 controls without formal mu-
sic training (nonmusicians, NM, eight women) but in-
volved in a regular leisure activity (e.g., sports, dance,
theater) were paid to participate in the experimental ses-
sion lasting for 2.5 hr (including the application of the
Electrocap, psychometric measurements, and experimen-
tal tasks). The two groups did not differ in age (MUS:
mean age = 25.1 years, age range = 1930, SD =3.9;
NM: mean age = 25.7 years, age range = 1933, SD =
4.8; F(1, 28) = 0.02, p= .68). All participants were native
French speakers, had comparable education levels (uni-
versity degree) and socioeconomic background (criteria
of the National Institute of Statistics and Economic Stud-
ies; MUS: 4.4; NM: 4.9; F(1, 28) = 1.45, p=.24),andre-
ported no past or current audiological or neurological
deficits. MUS practiced their instruments for an average
of 17 years (range = 1124, SD = 4.1) and the musician
group included three pianists, two accordionists, four vio-
linists, one cellist, two guitarists, one hornist, one tubist,
and one flautist. None of the participants was bilingual,
but all spoke English as a second language and most par-
ticipants (except for 1 MUS and 3 NM) had a rudimentary
knowledge of a third language that was neither tonal nor
quantitative. The study was conducted in accordance with
the Helsinki declaration, and all participants gave their in-
formed consent before enrolling in the experiment.
Screening Measures
Cognitive Ability
Standardized psychometric tests were used to examine
short-term and working memory (forward and reverse
Digit Span, WISC-IV; Wechsler, 2003), visual attention
(NEPSY from Korkman, Kirk, & Kemp, 1998) and non-
verbal general intelligence (progressive matrices, PM47;
Raven, Corporation, & Lewis, 1962).
Musical Aptitude
Participants performed two musicality tests (adapted
from the MBEA battery; Peretz, Champod, & Hyde,
2003) consisting in judging whether pairs of piano melo-
dies were same or different, based either on melodic or
on rhythmic information.
Experimental Stimuli
Auditory Stimuli
Nine natural Thai monosyllabic words were selected for
the experiment: /ba1/, /pa1/, /p
h
a1/, /ba:1/, /pa:1/, /p
h
a:1/,
/ba:0/, /pa:0/, /p
h
a:0/.
1
These words varied in vowel dura-
tion, with short (/ba1/, /pa1/ and /p
h
a1/; 261 msec on
average) and long vowels (/ba:1/, /pa:1/, /p
h
a:1/, /ba:0/,
/pa:0/ and /p
h
a:0/; 531 msec on average), and in
fundamental frequency, with low-tone (/ba1/, /pa1/,
/p
h
a1/, /ba:1/, /pa:1/ and /p
h
a:1/; F0 = 175 Hz on average)
and midtone vowels (/ba:0/, /pa:0/ and /p
h
a:0/; F0 = 218 Hz
on average). Furthermore, words contained voicing
contrasts (/ba1/, /ba:1/ and /ba:0/, VOT = 144 msec vs.
/pa1/, /pa:1/ and /pa:0/, VOT = 3 msec) as well as aspiration
contrasts (/pa1/, /pa:1/ and /pa:0/, VOT = 3 msec vs. /p
h
a1/,
/p
h
a:1/ and /p
h
a:0/, VOT = 77 msec).
2
Stimuli were recorded
by a female ThaiFrench bilingual, ensuring that all words
were produced naturally. For each word, five versions were
digitally recorded to reproduce natural speech variability.
Sound pressure level was normalized across all words to
Dittinger et al. 1587
a mean level of 70 dB by using Praat software (Boersma
& Weenink, 2011).
Visual Stimuli
For the learning phase, nine pictures representing famil-
iar objects (i.e., bear, flower, key, chair, bell, eye, strawberry,
train, glass) were selected based on the standardized set of
260 pictures (that are matched for name and image agree-
ment, familiarity, and visual complexity) built by Snodgrass
and Vanderwart (1980).
3
The same pictures as in the learn-
ing phase were then presented in the matching task. For
the semantic task, 54 new pictures that the participants
had not seen before in the experiment and that were se-
manticallyrelated or unrelated to the meaning of the newly
learned words were chosen from the Internet by two of the
authors (ED and MB). Students from our university (n=
60; age range = 1925 years) were asked to rate the seman-
tic relatedness between new and old pictures (that is, those
previously presented during the word learning phase). Half
of the presented pairs were semantically related and the
other half were semantically unrelated, and this was con-
firmed by the studentsratings.
Experimental Tasks
Participants were tested individually in a quiet experimen-
tal room (i.e., Faraday cage), where they sat in a comfort-
able chair at about 1 m from a computer screen. Auditory
stimuli were presented through HiFi headphones (HD590,
Sennheiser Electronic GmBH, Wedemark, Germany) at
70-dB sound pressure level. Visual and auditory stimuli
presentation as well as the collection of behavioral data
were controlled by Presentation software ( Version 11.0,
Neurobehavioral Systems, Berkeley CA).
Main Experimental Session (See Figure 1AE)
Phonological categorization task. At the beginning
and at the end of the experiment, participants performed
four different phonological tasks that lasted for 2.3 min
each. All nine Thai monosyllabic words were presented
in each task, but participantswereaskedtocategorize
them based upon different features in each task: (1) the
voicing contrast (e.g., /ba1/ vs. /pa1/), (2) the vowel length
(e.g., short: /ba1/ vs. long /ba:1/), (3) pitch (e.g., low: /pa:1/
vs. high: /ba:0/), and (4) the aspiration contrast (e.g., /pa1/
vs. /pha1/; see Figure 1A and E). For each task, the contrast
was visually represented on the left (e.g., shortwith a
short line) and right (e.g., longwith a long line) half of
the screen and participants had to press one of two re-
sponse buttons according to the correct side (e.g., left
one for short and right one for long vowels), as quickly and
accurately as possible. Each word was presented 10 times
in a pseudorandomized order with the constraints of no
immediaterepetitionofthesamewordandnomore
than four successive same responses. Task order and re-
sponse side were counterbalanced across participants.
Word learning phase. Participants were asked to learn
the meaning of each word previously presented in the
phonological categorization task using pictureword as-
sociations. For instance, a drawing of a bear was followed
by the auditory presentation of the word /ba1/, and thus,
/ba1/ was the word for bear in our foreignlanguage
(see Figure 1B). Each of the nine pictureword pairs
was presented 20 times, resulting in 180 trials that were
pseudorandomly presented (i.e., no immediate repeti-
tion of the same association) in two blocks of 3 min each.
The picture was presented first and then followed after
750 msec by one of the nine words. Total trial duration
was 2000 msec. Two different lists were built, so that
across participants different pictures were associated
with different words. No behavioral response was re-
quired from the participants but they were told that sub-
sequent tests would evaluate whether they learned the
meaning of the novel words.
Matching task. One of the nine pictures was presented,
followed after 750 msec by an auditory word that
matched or mismatched the associations previously
learned in the word learning phase. For instance, where-
as the drawing of a bear followed by /ba1/ (i.e., bear) was
a match, the drawing of a strawberry followed by /ba1/
was a mismatch (see Figure 1C). Participants were asked
to press one of two response keys accordingly, as quickly
and accurately as possible. Response hand was counter-
balanced across participants. At the end of the trial, a row
of XXXX appeared on the screen, and participants were
asked to blink during this time period (1000 msec; total
trial duration: 3750 msec) to minimize eye movement
artifacts during word presentation. Each word was pre-
sented 20 times, half in match condition and half in mis-
match condition. The total of 180 trials was pseudorandomly
presented (i.e., no immediate repetition of the same asso-
ciation and no more than four successive same responses)
within two blocks of 5.6 min each.
Semantic task. One of the new pictures was presented,
followed after 1500 msec by an auditory word that was
semantically related or unrelated. For instance, although
the picture of a lock was semantically related to the pre-
viously learned word /pa:1/ (i.e., key), the picture of a
strawberry cake was semantically unrelated to /pa:1/ (see
Figure 1D). Participants were asked to press one of two
response keys accordingly, as quickly and accurately as
possible. A familiarization task including four trials was
administrated before starting the task. Response hand
was counterbalanced across participants. At the end of
the trial, a row of XXXX appeared on the screen, and par-
ticipants were asked to blink during this time period
(1000 msec; total trial duration = 4500 msec). Each word
was presented 12 times, but none of the new pictures
1588 Journal of Cognitive Neuroscience Volume 28, Number 10
were repeated, so that on eachtrial the word was associated
with a different related or unrelated picture. Half of the
pictureword pairs were semantically related, and half
were semantically unrelated. A total of 108 trials was pre-
sented pseudorandomly (i.e., no immediate repetition of
the same association and no more than four successive
same responses) within two blocks of 4 min each.
Long-term Memory Session (See Figure 1F)
To test for long-term memory effects on behavior (i.e.,
ERRs and RTs), participants performed again the matching
(always administered first) and semantic tasks 5 months af-
ter the main experimental session (no ERPs were recorded).
Because of a dropout rate of 33%, only 10 participants
were retested in each group. In the matching task, a total
of 270 trials were presented within three blocks. In the
semantic task, a total of 216 trials were presented in two
blocks (with a short pause within each block).
EEG Data Acquisition
The EEG was continuously recorded at a sampling rate of
512 Hz with a band-pass filter of 0102.4 Hz by using a
Biosemi amplifier system (BioSemi Active 2, Amsterdam,
The Netherlands) with 32 active Ag/Cl electrodes (Biosemi
Pintype) located at standard positions according to the
International 10/20 System (Jasper, 1958). The EOG was re-
corded from flat-type active electrodes placed 1 cm to the
left and right of the external canthi and from an electrode
beneath the right eye. Two additional electrodes were
placed on the left and right mastoids. Electrode imped-
ance was kept below 5 kΩ. EEG data were analyzed using
BrainVisionAnalyzersoftware(Version1.05.0005&
Version 2.1.0; Brain Products, München, Germany). All
data were re-referenced offline to the averaged left and
right mastoids, filtered with a bandpass filter from 1 to
30 Hz (slope of 24 dB/oct), and independent component
analysis and inverse independent component analysis
were used to identify and remove components associated
with vertical and horizontal ocular movements. Finally,
DC-detrend and removal of artifacts above a gradient cri-
terion of 10 μV/msec or a maxmin criterion of 100 μV
over the entire epoch were applied automatically. For
each participant, ERPs were time-locked to word onset,
segmented into 2700 msec epochs, including a 200-msec
baseline and averaged within each condition. Individual
averages were then averaged together to obtain the grand
average across all participants.
Statistical Analyses
ANOVAs were computed using Statistica software
(Version12.0,StatSoft,Inc.,Tulsa,OK).ForERRsand
RTs, ANOVAs always included Group (MUS vs. NM) as
between-subject factor as well as specific factors for each
task. As the phonological categorization task was the only
task that was performed both at the beginning and at the
end of the experiment, Order (preexperiment vs. postex-
periment) was included as a within-subject factor together
with Task (Voicing vs. Vowel length vs. Pitch vs. Aspira-
tion). The matching and semantic tasks were only per-
formed once but each experiment was divided into two
blocks of trials so that factors were Block (1 vs. 2) and Con-
dition (match vs. mismatch or related vs. unrelated). To
further increase clarity, factors were again specified at the
beginning of each task in the Results section. On the basis
of ERRs, four outliers (2 MUS and 2 NM, ±2 SD away from
the mean) were excluded from further analyses.
For ERPs, we analyzed the early stages of auditory pro-
cessing in the phonological categorization task using N100
peak amplitude measures. By contrast, during the word
learning phase, as well as in the matching and semantic
tasks, we focused on semantic processing and we analyzed
the mean amplitude of the N400 component. Effects on
the N200 were also analyzed using mean amplitudes. Be-
cause ERPs were only analyzed for correct responses and
because the ERP traces of the four outliers that were elim-
inated from behavioral analyses were similar to the grand
average in each group, all participants (i.e., 15 MUS and
15 NM) were included in the ERP analyses. ANOVAs always
included Group (MUS vs. NM) as a between-subject factor
and Laterality (left: F3, C3, P3; midline: Fz, Cz, Pz; right: F4,
C4, P4) and Anterior/Posterior (frontal: F3, Fz, F4; central:
C3, Cz, C4; parietal: P3, Pz, P4) as within-subject factors, to-
gether with specific factors for each task. As for behavior,
for the phonological categorization task, these factors were
Order (pre vs. post) and Task (Voicing vs. Vowel length vs.
Pitch vs. Aspiration). For the matching and semantic tasks,
factors were Block (1 vs. 2) and Condition (match vs. mis-
match or related vs. unrelated). Post hoc Tukey tests (re-
ducing the probability of Type I errors) were used to
determine the origin of significant main effects and interac-
tions. To simplify results presentation, we only report sig-
nificant results related to our hypotheses (full statistical
results can be seen in Table 1). Finally, correlation analyses
(Pearsons coefficient) were computed between error rates
in the musicality task with error rates or N400 effects in the
semantic task. General linear models (including Group as a
categorical factor, error rates in the musicality task as a con-
tinuous factor, and error rates or N400 effects in the seman-
tic task as a dependent factor) were used to test whether
the differences between the slopes and intercepts of the
two groups were significant.
RESULTS
Results are presented first for the screening measures of
cognitive ability and musical aptitude and second for the ex-
perimental tasks. For each experimental task, behavioral da-
ta are presented first followed by the ERPs data, except for
the word learning phase in which no behavioral data were
recorded and for the long-term memory tasks in which no
ERPs were recorded. Finally, for ERPs data (except in the
Dittinger et al. 1589
Table 1. Results of ANOVAs on the ERPs Data in Different Tasks of the Main Experimental Session
Word Learning Phase Matching Task Semantic Task
ALL MUS NM ALL MUS NM ALL MUS NM
FpFpFpFp F pFpF pFpFp
N400 (340550 msec)
G 2.72 .11 ––3.85 .06 ––0.93 .34 –– – –
G × B 0.12 .73 ––3.23 .08 ––0.02 .89 –– – –
G × L 0.41 .67 ––0.21 .82 ––0.21 .81 –– – –
G × R 0.13 .88 ––0.75 .48 ––3.14 .05 –– – –
G×B×L 4.65 .01 ––0.07 .93 ––0.37 .69 –– – –
G × B × R 1.87 .16 ––2.40 .10 ––0.01 .99 –– – –
G × L × R 0.67 .62 ––1.27 .29 ––2.23 .07 –– – –
G × B × L × R 0.82 .52 ––0.40 .81 ––0.60 .67 –– – –
G×C ––– – –1.61 .21 ––1.06 .31 –– – –
G×B×C ––– – –0.42 .52 ––0.84 .37 –– – –
G×C×L ––– – –1.25 .29 ––0.21 .81 –– – –
G×C×R ––– – –3.14 .05 ––0.04 .96 –– – –
G×B×C×L ––– – –0.83 .44 ––0.01 .99 –– – –
G×B×C×R ––– – –1.63 .21 ––0.15 .86 –– – –
G×C×L×R ––– – –1.82 .13 ––1.19 .32 –– – –
G×B×C×L×R ––– – –1.97 .43 ––
0.52 .72 –– – –
B4.25 .05 1.57 .23 2.72 .12 3.13 .09 <.001 .99 7.68 .02 1.69 .20 0.63 .44 1.11 .31
B×L 7.05 .002 16.38 <.001 0.10 .91 0.52 .60 0.28 .76 0.31 .74 0.81 .45 0.25 .78 0.91 .42
B × R 1.32 .27 3.29 .05 0.12 .89 1.46 .24 0.08 .93 5.08 .01 0.36 .70 0.10 .90 0.32 .73
B × L × R 0.42 .80 0.68 .61 0.54 .71 1.06 .38 0.84 .50 0.54 .71 0.65 .63 0.71 .59 0.51 .73
C––– – –0.01 .95 0.53 .48 1.38 .26 0.10 .75 0.65 .43 0.42 .53
C×B ––– – –1.17 .29 0.07 .79 2.06 .17 1.46 .24 0.04 .85 2.43 .14
C×L ––– – –9.72 <.001 5.36 .01 5.58 .009 1.41 .25 0.19 .83 2.25 .12
C×R ––– – –27.80 <.001 28.34 <.001 6.38 .005 19.18 <.001 8.98 <.001 10.22 <.001
C×B×L ––– – –0.47 .63 1.02 .38 0.39 .68 0.45 .64 0.43 .65 0.13 .88
C×B×R ––– – –2.43 .10 2.99 .07 0.09 .91 1.12 .33 1.36 .27 0.18 .84
C×B×L×R ––– – –1.29 .28 1.11 .36 1.15 .34 0.59 .67 0.57 .69 0.54 .71
N200 (230340 msec)
G 1.24 .28 ––5.56 .03 ––1.49 .23 –– – –
G × B 0.13 .73 ––0.04 .84 ––7.26 .01 –– – –
G × L 0.60 .55 ––0.27 .77 ––2.41 .10 –– – –
G × R 0.04 .96 ––0.87 .43 ––2.56 .09 –– – –
G × B × L 1.05 .36 ––0.03 .97 ––0.29 .75 –– – –
G × B × R 0.70 .50 ––0.69 .51 ––0.85 .43 –– – –
G × L × R 0.99 .42 ––1.55 .19 ––1.76 .14 –– – –
G × B × L × R 0.68 .61 ––0.20 .94 ––0.69 .60 –– – –
1590 Journal of Cognitive Neuroscience Volume 28, Number 10
phonological categorization task where the N100 compo-
nent is of main interest), analysis of the N400 component
is presented first, followed by analyses of the N200.
Screening Measures
Cognitive Ability
Psychometric data were evaluated by means of univariate
ANOVAs. Results showed no significant Group differences
regarding general reasoning abilities (i.e., progressive
matrices, PM47; F(1, 28) = 1.37, p= .25), verbal working
memory (reverse digit span; F(1, 28) = 2.88, p=.10),
nor visual attention (F(1, 28) = 3.17, p=.09).Bycon-
trast, MUS (mean = 7.6, SD = 0.30) showed better
short-term memory abilities than NM (mean = 6.7, SD =
0.30; direct digit span; F(1, 28) = 5.53, p= .03).
Musical Aptitude
A 2 × 2 ANOVA (i.e., 2 Groups × 2 Tasks) showed that
MUS made fewer errors (6.7%, SD =2.0)thanNM
(17.6%, SD =2.0;maineffectofGroup:F(1, 28) =
14.71, p< .001), and all participants performed better
on the rhythmic (9.8%, SD = 1.6) than on the melodic
task (14.4%, SD = 2.0; main effect of Task: F(1, 28) =
4.19, p= .05) with no Group × Task interaction.
Experimental Tasks
Phonological Categorization Task
Behavioral data. Results of 2 × 2 × 4 ANOVAs [i.e., 2
Groups (MUS vs. NM) × 2 Orders (pre vs. post) × 4
Tasks (Voicing vs. Vowel length vs. Pitch vs. Aspiration)]
showed that MUS (6.6%, SD = 2.2) made overall fewer
errors compared with NM (19.1%, SD = 2.2; main effect
of Group: F(1, 24) = 16.29, p< .001; Figure 2A). The
influence of music training was largest in the pitch (MUS:
4.1%, SD = 4.8; NM: 25.8%, SD =4.9;Tukey,p<.001)and
aspiration tasks (MUS: 7.6%, SD =5.5;NM:28.7%,SD =
5.5; Tukey, p< .001; Group × Task interaction: F(3,
72) = 11.82, p< .001). Finally, only NM improved their
level of performance from pre to post in the pitch task
(Group × Task × Order interaction: F(3, 72) = 3.31,
Table 1. (continued)
Word Learning Phase Matching Task Semantic Task
ALL MUS NM ALL MUS NM ALL MUS NM
FpFpFpFp F pFpF pFpFp
G×C ––– – –2.19 .15 ––1.49 .23 –––
G×B×C ––– – –0.28 .60 ––0.004 .95 –––
G×C×L ––– – –1.85 .17 ––0.64 .53 –––
G×C×R ––– – –1.41 .25 ––0.88 .42 –––
G×B×C×L ––– – –0.14 .87 ––0.37 .70 –––
G×B×C×R ––– – –0.31 .74 ––0.50 .61 –––
G×C×L×R ––– – –1.51 .21 ––2.19 .08 –––
G×B×C×L×R ––– – –1.99 .10 ––0.69 .60 –––
B11.61 .002 6.50 .02 5.11 .04 2.28 .14 0.81 .38 1.55 .23 0.19 .67 5.36 .04 2.35 .15
B×L 4.44 .02 4.14 .03 0.82 .45 1.62 .21 0.77 .47 0.90 .42 1.24 .30 0.78 .47 0.75 .48
B × R 0.93 .40 1.25 .30 0.47 .63 0.30 .74 0.20 .82 1.33 .28 0.26 .77 0.58 .57 0.52 .60
B × L × R 0.51 .73 0.26 .91 1.33 .27 0.71 .58 0.37 .83 0.62 .65 0.41 .80 1.40 .25 0.09 .99
C––– – –16.53 <.001 15.22 .002 3.38 .09 4.16 .05 4.87 .05 0.37 .55
C×B ––– – –0.72 .40 0.93 .35 0.05 .82 4.32 .05 2.00 .18 2.33 .15
C×L ––– – –3.02 .06 4.04 .03 1.18 .32 1.59 .21 1.60 .22 0.22 .80
C×R ––– – –20.06 <.001 15.12 <.001 6.12 .006 6.83 .002 5.05 .01 1.89 .17
C×B×L ––– – –0.21 .81 0.20 .82 0.16 .85 0.28 .76 0.57 .57 0.10 .90
C×B×R ––– – –1.34 .27 1.13 .34 0.26 .77 5.25 .008 4.97 .01 1.17 .33
C×B×L×R ––– – –4.25 .003 3.98 .007 1.88 .13 1.53 .20 1.24 .30 0.94 .45
Although the Condition × Anterior/posterior interactions are significant in musicians (MUS) and in nonmusicians (NM) for the N400 component, the effects are reversed
in both groups (typical N400 effect over parietal sites in MUS, inversed N400 effect over frontal sites in NM). Significant effects are printed in italics, and exact levels of
significance are indicated except when the pvalues are inferior to .001 (<.001). G = Group; B = Block; C = Condition; L = Laterality; R = Anterior/Posterior.
Dittinger et al. 1591
p= .03; NM: pre: 30.3%, SD = 4.1, post: 21.3%, SD =3.3,
Tukey, p= .02; MUS: pre: 4.5%, SD = 4.1, post: 3.7%, SD =
3.3, Tukey, p= .99). Analyses of RTs did not reveal sig-
nificant Group differences (Figure 2B).
Electrophysiological data. N100 amplitudes were eval-
uated by means of a 2 × 2 × 4 × 3 × 3 ANOVA (i.e., 2
Groups×2Orders×4Tasks×3Lateralitypositions
[left vs. midline vs. right] × 3 Anterior/Posterior posi-
tions [frontal vs. central vs. parietal]). Results revealed
a significant Group × Task × Laterality interaction effect
(F(6, 168) = 3.19, p= .005). Separate ANOVAs for each
group showed that only MUS were characterized by a
larger N100 in the aspiration task (MUS: 6.50 μV, SD =
2.59 and NM: 5.57 μV, SD = 1.81) compared with
the other three tasks over the left hemisphere and at
midline electrodes (Task × Laterality interaction: MUS:
5.47 μV, SD = 1.78; F(6, 84) = 3.13, p= .008, Tukey,
both ps < .001; NM: 5.01 μV, SD = 1.74; F<1;Fig-
ure 3A and B).
Word Learning Phase
Electrophysiological data. TheN400aswellasthe
N200 were evaluated by means of 2 × 2 × 3 × 3 ANOVAs
(i.e., 2 Groups × 2 Blocks [1 vs. 2] × 3 Laterality × 3
Anterior/Posterior positions).
For all participants and in line with previous results, the
N400 component was larger over frontal (2.63 μV, SD =
0.87) and central (2.52 μV, SD = 0.76) sites compared
with parietal sites (1.29 μV, SD =0.66;Tukey,both
ps < .001; main effect of Anterior/Posterior: F(2, 56) =
25.25, p<.001).Inaddition,theGroup×Block×Later-
ality interaction effect was significant (F(2, 56) = 4.65,
Figure 2. Percentages of errors (ERRs) and RTs in the different tasks are shown for musicians (MUS) in red and for nonmusician controls (NM) in
black. For the phonological categorization task (A and B), results are illustrated for the premeasurement (beginning) and postmeasurement (end
of main experimental session), separately for each task (voicing, vowel length, pitch, and aspiration). For the matching task (C and D), results
for Match (solid lines) and Mismatch (MM: dotted lines) words are illustrated in the two blocks of the main experimental session and in the
three blocks of the long-term memory session. For the semantic task (E and F), results for semantically Related (solid lines) and Unrelated
(dotted lines) words are illustrated in the two blocks of the main experimental session and of the long-term memory session.
1592 Journal of Cognitive Neuroscience Volume 28, Number 10
p= .01). Separate group analyses showed that only
MUS showed significantly increased amplitudes from
Block 1 to Block 2. This effect was localized over the
left hemisphere and midline electrodes (MUS: Block 1:
1.54 μV, SD = 0.76 and Block 2: 2.16 μV, SD =
0.79; Block × Laterality interaction: F(2, 28) = 16.38,
p< .001; Tukey, both ps < .001 and NM: Block 1:
2.34 μV, SD = 1.36 and Block 2: 2.91 μV, SD =1.38;
main effect of Block: F(1, 14) = 2.72, p=.12;Fig-
ure 4A and B).
Analyses of the N200 component did not reveal signif-
icant Group differences (main effect of Group: F(1, 28) =
1.24, p= .28), but all participants showed significantly
increased amplitudes from Block 1 (1.13 μV, SD =
1.22) to Block 2 (1.79 μV, SD = 0.99; main effect of
Block: F(1, 28) = 11.61, p= .002; Figure 4A and B).
Matching Task
Behavioral data. Results of three-way ANOVAs (i.e., 2
Groups × 2 Blocks × 2 Conditions [match vs. mismatch])
showed that ERRs did not significantly differ between the
two groups (main effect of Group: F(1, 24) = 2.19, p=
.15). However, all participants committed overall fewer er-
rors for match (14.2%, SD = 2.6) compared with mismatch
words (19.4%, SD = 2.2) and fewer errors in Block 2
(15.2%, SD = 2.1) than in Block 1 (18.5%, SD = 2.2; main
effect of Condition: F(1, 24) = 7.68, p= .01; main effect of
Block: F(1, 24) = 9.27 p= .006; Figure 2C). In line with
ERRs, analyses of RTs did not reveal between-group differ-
ences (main effect of Group: F< 1), but overall faster RTs
for match (1041 msec, SD = 64) than for mismatch words
(1080 msec, SD = 70; main effect of Condition: F(1, 24) =
5.90, p= .02) and faster RTs in Block 2 (994 msec, SD =
61) than in Block 1 (1128 msec, SD = 73; main effect of
Block: F(1, 24) = 60.45, p< .001; Figure 2D).
Electrophysiological data. The N400 as well as the N200
component were evaluated by means of2 × 2 × 2 × 3 × 3
ANOVAs (2 Groups × 2 Blocks × 2 Conditions × 3 Later-
ality × 3 Anterior/Posterior positions).
Analysis of the N400 revealed a significant Group × Con-
dition × Anterior/Posterior interaction effect (F(2, 56) =
3.14, p= .05). Results of separate group analyses showed
larger N400 amplitudes in MUS for mismatch (0.10 μV,
SD = 1.82) compared with match words over centroparietal
regions (0.67 μV, SD = 1.58; Condition × Anterior/Posterior
interaction: F(2, 28) = 28.34, p< .001; Tukey, central: p=.02;
parietal: p< .001). The opposite pattern was found in NM
with larger N400 for match (1.34 μV, SD = 1.42) than for
Figure 3. Phonological
categorization. (A) N100
components at the Central (Cz)
electrode are compared between
tasks for musicians (MUS) and
for nonmusician controls (NM).
In this and subsequent figures,
time in milliseconds is in
abscissa and the amplitude of the
effects in microvolt is in ordinate.
Time zero corresponds to word
onset and negativity is plotted
upwards. Latency windows
for statistical analyses are
indicated with gray dotted lines
and the level of significance is
represented by stars with *p<
.05, **p< .01, and ***p<.001
(red stars for MUS and black
stars for NM). (B) Topographic
voltage distribution maps
illustrate the N100s to the
words separately for each task
and for MUS and NM. Voltage
values are scaled from 8to
+8 μV.
Dittinger et al. 1593
mismatch words over frontocentral sites (0.89 μV, SD =
1.53; Condition × Anterior/Posterior interaction: F(2, 28)
= 6.38, p= .005; Tukey, frontal: p= .001; central: p=
.03; Figure 5A and B).
The N200 amplitude was overall smaller in MUS (0.15 μV,
SD = 0.60) compared with NM (2.14 μV, SD = 0.60; main
effect of Group: F(1, 28) = 5.56, p= .03) and the N200
effect (i.e., mismatch minus match words) was more widely
distributed in MUS compared with NM (Group × Block ×
Condition × Laterality × Anterior/Posterior interaction: F(4,
112) = 1.99, p= .10; Figure 5A and B). MUS showed larger
N200 amplitudes for mismatch (P4: 1.98 μV, SD =0.76)
than for match words (P4: 0.01 μV, SD = 0.68) over centro-
parietal scalp sites with largest differences over midline and
right hemisphere (Condition × Anterior/Posterior interac-
tion: F(2, 28) = 15.12, p< .001; Condition × Laterality in-
teraction: F(2, 28) = 4.04, p= .03). In addition, the N200
effect was larger in Block 2 (P4: 2.22 μV, SD = 0.50) than
in Block 1 (P4: 1.66 μV, SD = 0.63) over midline and right
centroparietal sites (Condition × Block × Laterality ×
Anterior/Posterior interaction: F(4, 56) = 3.98, p=.007).
NM also showed an N200 effect that was localized over pa-
rietal sites (1.06 μV, SD = 1.41; Condition × Anterior/
Posterior interaction: F(2, 28) = 6.12, p=.006).
Figure 4. Word learning phase. (A) ERPs recorded at left parietal (P3), central (Cz), and right frontal sites (F4) in Block 1 (solid line) and Block 2
(dotted line) are overlapped for all participants (ALL: black lines, top), and separately below for musicians (MUS: red lines) and nonmusician
controls (NM: black lines). (B) Topographic voltage distribution maps of the differences between the two blocks (Block 2 minus Block 1) are illustrated
for the ERP components of interest (N200, N400), separately for MUS and for NM. Voltage values are scaled from 1.0 to +1.0 μV.
1594 Journal of Cognitive Neuroscience Volume 28, Number 10
Semantic Task
Behavioral data. Results of three-way ANOVAs (i.e., 2
Groups × 2 Blocks × 2 Conditions [related vs. unrelated])
showed that MUS (23.6%, SD = 2.0) made overall fewer
errors than NM (30.5%, SD = 2.0; main effect of Group:
F(1, 24) = 5.82, p= .02), and all participants made fewer
errors for unrelated (22.6%, SD =2.6)thanforrelated
words (31.5%, SD = 2.9; main effect of Condition: F(1,
24) = 11.24, p= .003; Figure 2E). Moreover, all par-
ticipants made fewer errors in Block 2 (30.2%, SD = 2.3)
than in Block 1 (23.9%, SD = 2.4; main effect of Block:
F(1, 24) = 12.37, p= .002). RTs were faster for related
(1210 msec, SD = 72) than for unrelated words (1342 msec,
SD = 71; main effect of Condition: F(1, 24) = 41.32, p<
.001) and faster in Block 2 (1159 msec, SD = 70) than in
Block 1 (1393 msec, SD = 75; main effect of Block: F(1,
24) = 88.92, p< .001), with no between-group differences
(main effect of Group: F<1;Figure2F).
Electrophysiological data. TheN400aswellasthe
N200 component were evaluated by means of 2 × 2 ×
2 × 3 × 3 ANOVAs (i.e., 2 Groups × 2 Blocks × 2 Con-
ditions × 3 Laterality × 3 Anterior/Posterior positions).
N400 analyses revealed a significant Group × Anterior/
Posterior interaction effect (F(2, 56) = 3.14, p= .05). As
typically reported in the literature (Kutas et al., 1988), the
N400 was larger for semantically unrelated (2.17 μV, SD =
1.93) compared with related words (3.29 μV, SD = 1.66)
over parietal sites in MUS (Condition × Anterior/Posterior
interaction: F(2, 28) = 8.98, p< .001; Tukey, parietal: p<
.001). By contrast, a reversed N400 effect was found in NM
with larger N400 for related (2.09 μV, SD = 1.60) than
for unrelated words (1.19 μV, SD = 1.06) over frontal
sites (Condition × Anterior/Posterior; F(2, 28) = 10.22,
p< .001; Tukey, frontal: NM: p= .002, Figure 6A and B).
The N200 amplitude was larger in Block 2 than in
Block 1 in MUS only (Group × Block: (F(1, 28) = 7.26,
p= .01; MUS: Block 1: 1.58 μV, SD = 2.62 and Block 2:
0.89 μV, SD =2.68;F(1, 14) = 5.35, p=.04andNM:
Block 1: 0.11 μV, SD = 3.15 and Block 2: 0.39 μV,
SD = 2.58; F(1, 14) = 2.35, p= .15; Figure 6A and B).
In addition, the N200 was also larger for unrelated than
for related words in MUS but not in NM (main effect of
Figure 5. Matching task. (A) Left: ERPs recorded at frontal (Fz) and parietal (Pz) sites are overlapped for Match (solid lines) and Mismatch (dotted lines)
words for all participants across the two blocks of trials (ALL: black lines). Central and right: ERPs are presented separately for Block 1 and Block 2 and
for musicians (MUS: red lines) and nonmusician controls (NM: black lines). (B) Difference waves (Mismatch minus Match) are overlapped for MUS (red)
and NM (black) separately for Block 1 and for Block 2 at Fz and Pz. Topographic voltage distribution maps of the Mismatch minus Match differences are
illustrated for N200 and N400 components separately and for MUS and NM in Block 1 and Block 2. Voltage values are scaled from 2.5 to +1.5 μV.
Dittinger et al. 1595
Condition: MUS: Related: 1.88 μV, SD = 2.67 and Unre-
lated: 0.59 μV, SD = 3.03; F(1, 14) = 4.87, p=.05andNM:
Related: 0.30 μV, SD = 3.63 and Unrelated: 0.02 μV,
SD = 2.23; F< 1; Group × Condition × Laterality ×
Anterior/Posterior interaction: F(4, 112) = 2.19, p=.08),
and this effect was larger over central and parietal sites
(Condition × Anterior/Posterior interaction: F(2, 28) =
5.05, p=.01;Tukey,bothps<.001).
Figure 6. Semantic task. (A) Left: ERPs recorded at frontal (Fz) and parietal sites (Pz) are overlapped for semantically Related (solid lines) and
Unrelated (dotted lines) words for all participants across the two blocks of trials (ALL: black lines). Central and right panels: ERPs are presented
separately for Block 1 and Block 2 and for musicians (MUS: red lines) and nonmusician controls (NM: black lines). (B) Difference waves (Unrelated
minus Related) are overlapped for MUS (red) and NM (black) separately for Block 1 and Block 2 at Fz and Pz. Topographic voltage distribution maps
of the Unrelated minus Related differences are illustrated for N200 and N400 components separately and for MUS and NM in Block 1 and Block 2.
Voltage values are scaled from 2.5 to +1.0 μV.
Figure 7. Correlation analyses.
Correlations between the
percentages of error in the
musicality test (ERRs Musicality)
and the sizes of the N400 effect
(Unrelated minus Related) in
the semantic task are illustrated
for musicians (MUS: red) and
for nonmusician controls
(NM: black). Dotted lines
represent the 95% confidence
interval of the correlation line.
1596 Journal of Cognitive Neuroscience Volume 28, Number 10
Long-term Memory: Matching Task (Behavior)
Results of three-way ANOVAs (i.e., 2 Groups × 3 Blocks ×
2 Conditions [match vs. mismatch]) showed that MUS
(20.3%, SD = 2.8) made fewer errors compared with NM
(28.6%, SD = 2.9; main effect of Group: F(1, 19) = 4.19,
p= .05; Figure 2C). Moreover, all participants improved
their level of performance from Block 1 (31.4%, SD =
2.6) to Blocks 2 and 3 (22.2%, SD = 3.4 and 19.8%, SD =
3.6, respectively; main effect of Block: F(2, 38) = 26.40, p<
.001). No significant between-group differences were
found on RTs but in both groups, RTs were faster in Block
3(965msec,SD = 67) than in Block 2 (1049 msec, SD =
72) and slowest in Block 1 (1168 msec, SD = 82; main ef-
fect of Block: F(2, 38) = 36.48, p< .001; Figure 2D).
Long-term Memory: Semantic Task (Behavior)
Results of three-way ANOVAs (i.e., 2 Groups × 2 Blocks ×
2 Conditions [related vs. unrelated]) showed that between-
group differences were not significant on ERRs nor on
RTs. However, all participants made fewer errors for un-
related (21.5%, SD = 6.0) than for related words (29.7%,
SD = 6.3; main effect of Condition; F(1, 19) = 6.60, p=
.02) and were faster for related (1055 msec, SD = 106)
than for unrelated words (1151 msec, SD = 119; main
effect of Condition; F(1, 19) = 14.54, p= .001; Figure 2E
and F).
Relationships between Musical Aptitude, Behavioral
Data, and Brain Activity
Error rates in the musicality test were correlated with the
size of the N400 effects in the semantic task for MUS (r=
.74, p= .01) but not for NM (r= .01, p= .97;
Figure 7A and B). Moreover, the slopes of the correla-
tions were significantly different for the two groups (main
effect of Group: F(1, 26) = 7.36, p= .01) as well as the cor-
relation intercepts (Group × Musicality interaction: F(2,
26) = 6.52, p= .005).
DISCUSSION
Summary of Results
By using an ecologic valid experimental design we tested
the general hypothesis that professional musicians would
learn the meaning of novel words more efficiently than
control participants. Overall, both behavioral and electro-
physiological data support this hypothesis. Behaviorally,
musicians performed better than controls in the musicality
and phonological categorization tasks. In addition, al-
though all participants performed similarly in the matching
task, musicians made significantly fewer errors in the se-
mantic task. Finally, after 5 months, musicians remem-
bered more words than controls as reflected by lower
error rates in the matching task.
The electrophysiological markers of word learning also
clearly differed between the two groups. Although con-
trol participants showed similar N100 amplitudes in all
phonological categorization tasks, musicians showed an
increase in N100 amplitude when categorizing the most
difficult aspiration contrast. Most importantly and in line
with the word learning literature, both groups showed
enhanced N400 amplitudes over frontal scalp sites when
learning the meaning of novel words. However, only
musicians were additionally characterized by larger left-
lateralized N400s in the second block compared with
the first block of the word learning phase. Finally, only
musicians showed the typical centroparietal distribution
of the N400 effect in the matching and semantic tasks.
By contrast, nonmusicians showed reversed N400 effects
in both tasks over frontal sites. These findings are dis-
cussed in detail below. It is, however, important to note
that cross-sectional studies, as the one reported here,
are necessary to first demonstrate differences between
musically trained and untrained participants before de-
signing longitudinal experiments to test for the causality
of the reported effects.
Spatiotemporal Dynamics in the Learning Phase
Results showed clear evidence for fast brain plasticity, as
reflected by the rapid development of the N400 in both
groups of participants after only 3 min of novel word
learning (Block 1), that is after 10 repetitions of each
pictureword association (see Figure 4). This finding
extends previous results on word learning showing
N400 enhancements when learning a second language
(McLaughlin et al., 2004), the meaning of rare words
(Perfetti et al., 2005), and when learning the meaning
of novel words or pseudowords from highly constrained
sentence contexts (Borovsky et al., 2010, 2012; Batterink
& Neville, 2011; Mestres-Missé et al., 2007). Importantly,
and in line with previous work in adults (Borovsky et al.,
2012; Mestres-Missé et al., 2007) and in children (François
et al., 2013; Friedrich & Friederici, 2008), in both musi-
cians and controls the N400 component to novel words
was larger frontocentrally than parietally. These results
are compatible with previous findings, suggesting that pre-
frontal and temporal brain regions are associated with the
maintenance of novel information in working memory
(Hagoort, 2014) and the acquisition of word meaning
(Rodriguez-Fornells et al., 2009).
Importantly, however, the N400 increase from the first
to the second block of novel word learning was only sig-
nificant in musicians (see Figure 4A). In line with our hy-
pothesis, this is taken to suggest faster encoding of novel
word meaning in musicians than in controls. In addition,
word learning in musicians was accompanied by a rapid
shift of the N400 component distribution from frontal to
centroparietal sites. This shift in distribution is in line with
the hypothesis that musicians have already integrated novel
words representations in semantic memory (Batterink &
Dittinger et al. 1597
Neville, 2011) and that the N400 centroparietal distribution
in musicians reflects access to semantic memory (Kutas &
Federmeier, 2011). By contrast, in control participants, the
N400 component remained larger frontally throughout the
entire learning phase, suggesting that nonmusicians had
not yet integrated the novel wordsmeaning into estab-
lished semantic networks. This interpretation can be di-
rectly tested in future experiments by increasing the
number of repetitions of pictureword associations. Under
such conditions, a frontal to centroparietal shift in N400
distribution should also be found in nonmusicians.
Evidence for Rapidly Established Representations
of Novel Words in the Matching and
Semantic Tasks
As expected, all participants, whether musicians or con-
trols, were able to learn the nine pictureword associa-
tions, as reflected by error rates below chance level (i.e.,
50% errors) in both the matching (mean = 17%) and se-
mantic tasks (mean = 27%; see Figure 2C and E). To follow
the dynamics of word learning within each task, results
were compared between the first and the second block
of trials. Interestingly, the level of performance increased
with repetition in both tasks, indicating a still ongoing
learning effect. Moreover, both musicians and controls
showed clear matching effects with lower error rates and fas-
ter RTs for match compared with mismatch words (Boddy
& Weinberg, 1981). However, and in contrast to typical se-
mantic priming effects (Meyer & Schvaneveldt, 1971), both
groups of participants made more errors for semantically
related words than for semantically unrelated words. Al-
though unexpected, this result may reflect a response bias
towards rejection (i.e., considering the word as unrelated
to the picture) as a consequence of task difficulty generat-
ing higher uncertainty (Gigerenzer, 1991). In other words,
when participants were not certain whether the pictures
and the words were semantically related (e.g., honey
and bear), they tended to respond that they were unre-
lated. By contrast, results for RTs conform to the literature
(i.e., faster RTs for semantically related than for unrelated
words; Meyer & Schvaneveldt, 1971). Although the pres-
ence of a speedaccuracy trade-off limits the interpretation
of this priming effect, faster RTs for semantically related
words are indicative that new pictures that had not been
seen before in the experiment did activate the representa-
tions of semantically related newly learned words.
Turning to the influence of music training, musicians
and nonmusician controls performed similarly in the
matching task but musicians outperformed controls in
the semantic task. This contrastive pattern of results sug-
gests that the two tasks tap into different memory sys-
tems. To decide whether the newly learned word
matched the picture in the matching task, participants
had to retrieve the specific pictureword associations
that were stored in episodic memory during the word
learning phase. By contrast, in the semantic task partici-
pants had to retrieve general information from semantic
memory because the novel pictures that were presented
before the newly learned words had not been seen be-
fore in the experiment. In line with the centroparietal
shift in N400 distribution observed at the end of the word
learning phase, the finding that musicians outperformed
nonmusicians in the semantic task is taken as evidence
that musicians had already integrated the novel words
meanings into semantic memory so that priming effects
generalized to new pictures.
ERPs in the matching and semantic tasks also add sup-
port to this interpretation. In musicians and for both
tasks, the N400 over centroparietal sites was larger for
unexpected (mismatch/unrelated) than for expected
words (match/related; Figures 5 and 6). This sensitivity
to word characteristics and this scalp distribution corre-
spond to the N400 component, typically considered as the
electrophysiological marker of the integration of novel
wordsmeanings into semantic memory (Borovsky et al.,
2012; Batterink & Neville, 2011; Mestres-Missé et al., 2007)
and as reflecting the activity in a multimodal long-term
memory system that is induced by a given input stimulus
during a delimited time window as meaning is dynamically
constructed(Kutas & Federmeier, 2011, p. 22; see also
Steinbeis & Koelsch, 2008, for N400s elicited by single
chords incongruous with the preceding context). By con-
trast, in nonmusician controls, the N400 effect was re-
versed over frontocentral sites in both tasks, with larger
N400 for expected (match/related) than for unexpected
words (mismatch/unrelated). This finding was surprising
based on previous results showing that the N400 is larger
for unrelated than for related words in both lexical deci-
sion tasks (Borovsky et al., 2012) and semantic priming ex-
periments (Mestres-Misséetal.,2007).Becausethe
amount of music training was not controlled in these ex-
periments, a possible explanation is that some of the par-
ticipants had musical skills, hence influencing the results.
However, stimulus and task differences are more likely to
account for this discrepancy. Most experiments on ERPs
and word learning in adults used similar designs with novel
words embedded in sentence contexts. Here and as in ex-
periments conducted with children (Friedrich & Friederici,
2008; Torkildsen et al., 2008) and with adults (Dobel et al.,
2009), participants learned novel words through picture
word associations. It is thus possible that this learning
mode was more demanding than learning the meaning
of novel words from sentence contexts. Along these lines,
we interpret the reversed N400 effect in nonmusicians in
the matching and semantic tasks as showing that nonmu-
sicians had not yet fully learned the pictureword associa-
tions (as reflected by the frontal distribution of the N400 in
these tasks that is similar to the frontal N400 found during
word learning) and that they were still building up new ep-
isodic memory traces based on the correct information
provided by the matching words (as reflected by larger
N400 increase from the first to the second block for match
than for mismatch words; see Figure 5A). As mentioned
1598 Journal of Cognitive Neuroscience Volume 28, Number 10
above, these interpretations can be tested by increasing the
number of trials in the learning phase as well as in the
matching and semantic tasks and by comparing different
word learning designs (e.g., novel words in sentence con-
texts vs. pictureword associations).
Finally, although the main focus of this experiment was
on the modulations of N400 amplitude with novel word
learning, results also revealed interesting effects on the
N200 component. In both groups, the N200 components
in the matching task were larger for mismatching than for
matching words over parietal sites (see Figure 5). Insofar
as the N200 has been associated with categorization pro-
cesses (Friedrich & Friederici, 2008), mismatching words
possibly required more effortful categorization processes
than matching words. However, there is also evidence
that the N200 component reflects early contextual influ-
ences (van den Brink, Brown, & Hagoort, 2001) and pho-
nological processing (Connolly & Phillips, 1994). Thus,
mismatching words possibly elicited larger N200 than
matching words because they were unexpected based
on the picture context and/or at the phonological level.
In the semantic task, the increase in N200 amplitude
for semantically unrelated compared with related words
was only found in musicians, again suggesting that word
categorization was possibly more efficient in musicians
and/or that musicians were more sensitive to the context
(i.e., the picture) or to the phonological characteristics of
novel words than nonmusicians.
Evidence for Long-lasting Representations of
Novel Words
To our knowledge, this is the first experiment comparing
long-term memory (after 5 months) for novel words in
musicians and controls. Clearly, results need to be con-
sidered with caution because of the high dropout rate.
Nevertheless, they point to improved long-term memory
in musicians compared with controls in the matching
task (lower error rates) that was always performed first.
The memory traces of the words that have been learned
5 months before therefore seem stronger in musicians
than in nonmusicians. By contrast, no between-group dif-
ferences were found in the semantic task possibly be-
cause both groups of participants similarly benefited
from the reactivation of memory traces during the match-
ing task. Taken together, these results point to long-
lasting effects of rapidly established word representations
during word learning, and they open new perspectives
to further test for the influence of music training on
long-term memory.
Evidence for Transfer Effects from Music Training
to Word Learning
In summary, behavioral and electrophysiological data
showed that music training improved several aspects of
word learning. How can we account for the influence
of music training on novel word learning?
Cascading Effects from Perception to Word Learning
The first interpretation, in terms of cascading effects, is
that enhanced auditory perception and auditory atten-
tion (Strait et al., 2015) in musicians drive the facilitation
observed in word learning through different subsequent
steps (i.e., building up new phonological representations
and attaching meaning to them, storing this new informa-
tion in short- and long-term memory). In support of this
interpretation, the error rate in the musicality test was cor-
related with the size of the N400 effect in the semantic task
in musicians but not in controls (Figure 7), thereby clearly
pointing to a relationship between auditory perception/
attention and word learning.
Moreover, in line with the hypothesis of improved
speech discrimination in musicians when the task is most
difficult (Diamond, 2013; Schön et al., 2004), musicians
outperformed control participants in the phonological
categorization of tonal and aspiration contrasts, but both
groups performed equally well for simple voicing con-
trasts (/Ba/ vs. /Pa/). Thus, music training was most useful
to discriminate phonemes that are contrastive in Thai but
that do not belong to the French phonemic repertoire
(Figure 2A). It should be noted in this respect that, al-
though vowel length is not contrastive in French, controls
nevertheless performed as well as musicians in this task,
possibly because the natural difference in vowel duration
was large enough (270 msec) to be easily perceived by non-
musician controls. Taken together, these findings add sup-
port to the hypothesis of transfer effects between music
and speech (Asaridou & McQueen, 2013; Besson et al.,
2011), with near transfer for acoustic cues common to mu-
sic and speech such as frequency and with far transfer to
unfamiliar linguistic cues such as aspiration.
Note also that the N100 amplitude in musicians was en-
hanced for the nonnative aspiration contrast (see Figure 3).
Because this task was more difficult for French native
speakers than the other tasks, as revealed by behavioral
data, the increased N100 may reflect increased focused
attention (Strait et al., 2015) and mobilization of neuro-
nal resources in musicians. An alternative but com-
plementary interpretation is that this increase in N100
amplitude reflected increased neural synchronicity and
structural connectivity in musicians, who are typically more
sensitive to the acoustic-phonetic properties of speech
sounds than nonmusicians who showed similar N100 com-
ponents in all tasks (Bidelman et al., 2014; Elmer et al.,
2013; Chobert et al., 2012; Musacchia et al., 2007; Wong
& Perrachione, 2007).
Finally, and in line with the cascade hypothesis, it may
be that the very nature of the stimuli (monosyllabic Thai
words) is at least partly responsible of the musicians ad-
vantage in the semantic task. Because musicians are
more sensitive than nonmusicians to variations in pitch,
Dittinger et al. 1599
duration, and VOT, this possibly helped them build pre-
cise phonological representations that were then more
easily associated to a novel word meaning. One way to
test for this interpretation would be to use a similar de-
sign but with written instead of spoken words thereby
potentially reducing the musiciansacoustic-phonetic
advantage.
Multidimensional Facilitation Effects
The second interpretation, that is by no means contradic-
tory but rather complementary to the cascading interpre-
tation, is that the multidimensional nature of music
training independently improved the several functions
that are necessary for word learning. There is already
some evidence in the literature for enhanced working
and verbal memory (George & Coch, 2011; Ho et al.,
2003) and executive functions (Zuk et al., 2014; Moreno
et al., 2011; Rogalsky et al., 2011; Pallesen et al., 2010)
with musicianship. In line with these results, musicians
showed enhanced auditory short-term memory (digit
span; George & Coch, 2011). Moreover, as reported in
previous experiments, we also found significant be-
tween-group differences in the ERP component related
to semantic integration and memory (N400; Mestres-
Missé et al., 2007; Perfetti et al., 2005; McLaughlin
et al., 2004). However, in contrast to previous results,
we found no evidence for increased working memory
(reversed digit span; George & Coch, 2011) or for im-
proved nonverbal intelligence (PM47) in adult musicians.
Note that the assessment of general cognitive abilities
was quite limited in the present experiment because of
time constraints. It is thus possible that general cognitive
abilities other than the ones tested here could have influ-
enced the observed between-group differences in word
learning. Future experiments will aim at including more
tests targeting at elucidating different cognitive functions
(selective and sustained attention, working, short-term,
and long-term memory, executive functions). Finally, it
could also be argued that musicians performed better
than nonmusicians because they were generally more
motivated (Corrigall, Schellenberg, & Misura, 2013). Al-
though this is difficult to control for, it is important to
note that participants were not informed about the aim
of the experiment until the end of the session. Only that
we were interested in language, music, and the brain.
Thus, as they discovered task by task what the experi-
mental session was about, it is unlikely that general
motivation accounted for the present behavioral and
electrophysiological effects.
Conclusions
In summary, by recording ERPs during both learning and
test phases (matching and semantic tasks) of novel word
learning, results revealed fast changes in brain activity after
only a few minutes of exposure to pictureword associa-
tions. Moreover, these changes were stronger in musi-
cians than in controls. Specifically, the frontoparietal
shift of the N400 in the word learning phase (i.e., without
motor responses) only developed in musicians, which we
interpret as an electrophysiological signature of fast map-
ping(Carey, 1978). To our knowledge, this is the first re-
port showing that music training influences semantic
processing. As a future step, we plan to use a longitudinal
approach with nonmusician controls trained with music so
as to test for the causal influence of music training on
word learning. Finally, these results also open new per-
spectives to further investigate the influence of music
training on long-term memory for applications of music
training in the domain of native and second language
learning (Moreno, Lee, Janus, & Bialystok, 2015; Chobert
& Besson, 2013) and for using this type of experimental
design in clinical research to specify the stages of word
learning that are most deficient.
Acknowledgments
We would like to thank all the participants, Chotiga Pattamadilok
for registering the auditory stimuli, and Benjamin Furnari for
his help with data acquisition and analyses. The work of E. D.
M. Ba., M. D. I., and M. B., carried out within the Labex BLRI
(ANR-11-LABX-0036), has benefited from support from the
French government, managed by the French National Agency for
Research (ANR), under the program Investissements dAvenir
(ANR-11-IDEX-0001-02). E. D. was supported by a doctoral
fellowship from the BLRI and M. Ba. by a doctoral fellowship
from the French Ministry of Research and Education.
Reprint requests should be sent to Eva Dittinger, Laboratoire de
Neuroscience Cognitives, Université Aix Marseille, Centre Saint
Charles, 3 Place Victor Hugo, Marseille, France, 13331, or via
e-mail: eva.dittinger@blri.fr.
Notes
1. Following phonetic transcription in Thai, 1 refers to low-
tone, 0 to midtone, p
h
to aspirated voicing, and the colon to
long vowel duration.
2. Voice Onset Time (VOT) is defined as the interval between
the noise burst produced at consonant release and the wave-
form periodicity associated with vocal cord vibrations (Lisker
& Abramson, 1967) .
3. Pictures were chosen from the Snodgrass and Vanderwart
(1980) picturesset but were found on the Internet to ensure
sufficient resolution and quality.
REFERENCES
Asaridou, S. S., & McQueen, J. M. (2013). Speech and music
shape the listening brain: Evidence for shared domain-general
mechanisms. Frontiers in Psychology, 4, 321.
Batterink, L., & Neville, H. (2011). Implicit and explicit
mechanisms of word learning in a narrative context: An
event-related potential study. Journal of Cognitive
Neuroscience, 23, 31813196.
Besson, M., Chobert, J., & Marie, C. (2011). Transfer of training
between music and speech: Common processing, attention,
and memory. Frontiers in Psychology, 2, 94.
1600 Journal of Cognitive Neuroscience Volume 28, Number 10
Bidelman, G. M., Weiss, M. W., Moreno, S., & Alain, C. (2014).
Coordinated plasticity in brainstem and auditory cortex
contributes to enhanced categorical speech perception
in musicians. European Journal of Neuroscience,
40, 26622673.
Boddy, J., & Weinberg, H. (1981). Brain potentials, perceptual
mechanism and semantic categorization. Biological
Psychology, 12, 4361.
Boersma, P., & Weenink, D. (2011). Praat: Doing phonetics
by computer [computer program]. Retrieved from
www.praat.org.
Borovsky, A., Elman, J., & Kutas, M. (2012). Once is enough:
N400 indexes semantic integration of novel word meanings
from a single exposure in context. Language Learning
and Development, 8, 278302.
Borovsky, A., Kutas, M., & Elman, J. (2010). Learning to use
words: Event-related potentials index single-shot contextual
word learning. Cognition, 116, 289296.
Carey, S. (1978). The child as word learner. In M. Halle,
J. Bresnan, & G. A. Miller (Eds.), Linguistic theory
and psychological reality (pp. 264293). Cambridge,
MA: MIT Press.
Chobert, J., & Besson, M. (2013). Musical expertise and second
language learning. Brain Sciences, 3, 923940.
Chobert, J., François, C., Velay, J. L., & Besson, M. (2012).
Twelve months of active musical training in 8- to 10-year-old
children enhances the preattentive processing of syllabic
duration and voice onset time. Cerebral Cortex, 24, 956967.
Chobert, J., Marie, C., François, C., Schön, D., & Besson, M.
(2011). Enhanced passive and active processing of syllables
in musician children. Journal of Cognitive Neuroscience,
23, 38743887.
Connolly, J. F., & Phillips, N. A. (1994). Event-related potential
components reflect phonological and semantic processing
of the terminal words of spoken sentences. Journal of
Cognitive Neuroscience, 6, 256266.
Corrigall, K. A., Schellenberg, E. G., & Misura, N. M. (2013).
Music training, cognition, and personality. Frontiers in
Psychology, 4, 222.
Diamond, A. (2013). Executive functions. Annual Review
Psychology, 64, 135168.
Dobel, C., Lagemann, L., & Zwitserlood, P. (2009). Non-native
phonemes in adult word learning: Evidence from the N400m.
Philosophical Transactions of the Royal Society, Series B,
Biological Sciences, 364, 36973709.
Dumay, N., & Gaskell, M. G. (2007). Sleep-associated changes in
the mental representation of spoken words. Psychological
Science, 18, 3539.
Elmer, S., Hänggi, J., Meyer, M., & Jäncke, L. (2013). Increased
cortical surface area of the left planum temporale in
musicians facilitates the categorization of phonetic and
temporal speech sounds. Cortex, 49, 28122821.
Elmer, S., Meyer, M., & Jäncke, L. (2012). Neurofunctional and
behavioral correlates of phonetic and temporal
categorization in musically trained and untrained subjects.
Cerebral Cortex, 22, 650658.
Fitzroy, A. B., & Sanders, L. D. (2013). Musical expertise
modulates early processing of syntactic violations in
language. Frontiers in Psychology, 3, 603.
François, C., Chobert, J., Besson, M., & Schön, D. (2013). Music
training for the development of speech segmentation.
Cerebral Cortex, 23, 20382043.
Friederici, A. D., & Singer, W. (2015). Grounding language
processing on basic neurophysiological principles. Trends
in Cognitive Sciences, 19, 329338.
Friedrich, M., & Friederici, A. D. (2008). Neurophysiological
correlates of online word learning in 14-month-old infants.
NeuroReport, 19, 17571761.
Gandour, J., Wong, D., Lowe, M., Dzemidzic, M.,
Satthamnuwong, N., Tong, Y., et al. (2002). A crosslinguistic
fMRI study of spectral and temporal cues underlying
phonological processing. Journal of Cognitive Neuroscience,
14, 10761087.
Gaser, C., & Schlaug, G. (2003). Brain structures differ
between musicians and nonmusicians. Journal of
Neuroscience, 23, 92409245.
George, E. M., & Coch, D. (2011). Music training and
working memory: An ERP study. Neuropsychologia,
49, 10831094.
Gigerenzer, G. (1991). How to make cognitive illusions
disappear: Beyond heuristics and biases.European
Review of Social Psychology, 2, 83115.
Hagoort, P. (2014). Nodes and networks in the neural
architecture for language: Brocas region and beyond.
Current Opinion in Neurobiology, 28, 136141.
Ho, Y. C., Cheung, M. C., & Chan, A. S. (2003). Music training
improves verbal but not visual memory: Cross-sectional
and longitudinal explorations in children. Neuropsychology,
17, 439450.
Hyde, K. L., Lerch, J., Norton, A., Forgeard, M., Winner, E.,
Evans, A. C., et al. (2009). Musical training shapes structural
brain development. Journal of Neuroscience, 29,
30193025.
Jäncke, L. (2009). Music drives brain plasticity. F1000 Biology
Reports, 1, 16.
Jasper, H. H. (1958). The tentwenty system of the International
Federation. Clinical Neurophysiology, 10, 371375.
Jentschke, S., & Koelsch, S. (2009). Musical training modulates
the development of syntax processing in children.
Neuroimage, 47, 735744.
Koelsch, S., Kasper, E., Sammler, D., Schulze, K., Gunter, T., &
Friederici, A. D. (2004). Music, language and meaning: Brain
signatures of semantic processing. Nature Neuroscience,
7, 302307.
Korkman, M., Kirk, U., & Kemp, S. (1998). NEPSY: A developmental
neuropsychological assessment. San Antonio, TX: The
Psychological Corporation.
Kraus, N., & Chandrasekaran, B. (2010). Music training for
the development of auditory skills. Nature Reviews
Neuroscience, 11, 599605.
Kühnis, J., Elmer, S., & Jäncke, L. (2014). Auditory evoked
responses in musicians during passive vowel listening are
modulated by functional connectivity between bilateral
auditory-related brain regions. Journal of Cognitive
Neuroscience, 26, 27502761.
Kutas, M., & Federmeier, K. D. (2011). Thirty years and
counting: Finding meaning in the N400 component of the
event related brain potential (ERP). Annual Review of
Psychology, 62, 621647.
Kutas, M., & Hillyard, S. A. (1980). Reading senseless sentences:
Brain potentials reflect semantic incongruity. Science,
207, 203205.
Kutas, M., Van Petten, C., & Besson, M. (1988). Event-related
potential asymmetries during the reading of sentences.
Electroencephalogry and Clinical Neurophysiology,
69, 218233.
Lisker, L., & Abramson, A. S. (1967). Some effects of context
on voice onset time in English stops. Language ans Speech,
10, 128.
Maess, B., Koelsch, S., Gunter, T. C., & Friederici, A. D. (2001).
Musical syntaxis processed in the area of Broca: An
MEG-study. Nature Neuroscience, 4, 540545.
Marie, C., Delogu, F., Lampis, G., Olivetti Belardinelli, M., &
Besson, M. (2011). Influence of musical expertise on
segmental and tonal processing in Mandarin Chinese.
Journal of Cognitive Neuroscience, 23, 27012715.
Dittinger et al. 1601
Marques, C., Moreno, S., Castro, S. L., & Besson, M. (2007).
Musicians detect pitch violation in a foreign language better
than nonmusicians: Behavioural and electrophysiological
evidence. Journal of Cognitive Neuroscience, 19, 14531463.
McLaughlin, J., Osterhout, L., & Kim, A. (2004). Neural correlates
of second-language word learning: Minimal instruction
produces rapid change. Nature Neuroscience, 7, 703704.
Mestres-Missé, A., Rodriguez-Fornells, A., & Münte, T. F. (2007).
Watching the brain during meaning acquisition. Cerebral
Cortex, 17, 18581866.
Meyer, D. E., & Schvaneveldt, R. W. (1971). Facilitation in
recognizing pairs of words: Evidence of a dependence
between retrieval operations. Journal of Experimental
Psychology, 20, 227234.
Moreno, S., Bialystok, E., Barac, R., Schellenberg, E. G., Cepeda,
N. J., & Chau, T. (2011). Short-term music training enhances
verbal intelligence and executive function. Psychological
Science, 22, 14251433.
Moreno, S., Lee, Y., Janus, M., & Bialystok, E. (2015).
Short-term second language and music training induces
lasting functional brain changes in early childhood.
Child Development, 86, 394406.
Moreno, S., Marques, C., Santos, A., Santos, M., Castro, S. L.,
& Besson, M. (2009). Musical training influences linguistic
abilities in 8-year-old children: More evidence for brain
plasticity. Cerebral Cortex, 19, 712723.
Münte, T. F., Altenmüller, E., & Jäncke, L. (2002). The
musicians brain as a model of neuroplasticity. Nature
Reviews Neuroscience, 3, 473478.
Musacchia, G., Sams, M., Skoe, E., & Kraus, N. (2007). Musicians
have enhanced subcortical auditory and audiovisual
processing of speech and music. Proceedings of the National
Academy of Sciences, U.S.A., 104, 1589415898.
Pallesen, K. J., Brattico, E., Bailey, C. J., Korvenoja, A., Koivisto, J.,
Gjedde, A., et al. (2010). Cognitive control in auditory working
memory is enhanced in musicians. PLoS One, 5, 11120.
Pantev, C., Oostenveld, R., Engelien, A., Ross, B., Roberts, L. E.,
& Hoke, M. (1998). Increased auditory cortical representation
in musicians. Nature, 23, 811814.
Patel, A. D. (2008). Music, language, and the brain. New York:
Oxford University Press.
Peretz, I., Champod, A. S., & Hyde, K. (2003). Varieties of musical
disorders. The Montreal Battery of Evaluation of Amusia.
Annals of the New York Academy of Sciences, 999, 5875.
Peretz, I., Vuvan, D., Lagrois, M.-É., & Armony, J. L. (2015).
Neural overlap in processing music and speech. Philosophical
Transactions of the Royal Society, Series B, Biological
Sciences, 370, 20140090.
Perfetti, C. A., Wlotko, E. W., & Hart, L. A. (2005). Word learning
and individual differences in word learning reflected in
event-related potentials. Journal of Experimental Psychology:
Learning, Memory, and Cognition, 31, 12811292.
Raven, J. C., Corporation, P., & Lewis, H. K. (1962). Coloured
progressive matrices: Sets A, AB, B. London, UK: Oxford
Psychologist Press.
Rodriguez-Fornells, A., Cunillera, T., Mestres-Missé, A., &
De Diego-Balaguer, R. (2009). Neuropsychological mechanisms
involved in language learning in adults. Philosophical
Transactions of the Royal Society of London, Series B,
Biological Sciences, 364, 37113735.
Rogalsky, C., Rong, F., Saberi, K., & Hickok, G. (2011).
Functional anatomy of language and music perception:
Temporal and structural factors investigated using functional
magnetic resonance imaging. Journal of Neuroscience, 31,
38433852.
Schellenberg, E. G. (2004). Music lessons enhance IQ.
Psychological Science, 15, 511514.
Schneider, P., Scherg, M., Dosch, H. G., Specht, H. J., Gutschalk,
A., & Rupp, A. (2002). Morphology of Heschls gyrus reflects
enhanced activation in the auditory cortex of musicians.
Nature Neuroscience, 5, 688694.
Schön, D., Magne, C., & Besson, M. (2004). The music of speech:
Music training facilitates pitch processing in both music and
language. Psychophysiology, 41, 341349.
Seppänen, M., Hämäläinen, J., Pesonen, A. K., & Tervaniemi, M.
(2012). Music training enhances rapid neural plasticity of N1
and P2 source activation for unattended sounds. Frontiers in
Human Neuroscience, 6, 43.
Snodgrass, J. G., & Vanderwart, M. (1980). A standardized
set of 260 pictures: Norms for name agreement, image
agreement, familiarity, and visual complexity. Journal of
Experimental Psychology. Human Learning and Memory,
6, 174215.
Steinbeis, N., & Koelsch, S. (2008). Comparing the processing
of music and language meaning using EEG and fMRI provides
evidence for similar and distinct neural representations.
PLoS One, 3, 2226.
Strait, D. L., Parbery-Clark, A., OConnell, S., & Kraus, N. (2013).
Biological impact of preschool music classes on processing
speech in noise. Developmental Cognitive Neuroscience, 6,
5160.
Strait, D. L., Slater, J., OConnell, S., & Kraus, N. (2015). Music
training relates to the development of neural mechanisms of
selective auditory attention. Developmental Cognitive
Neuroscience, 12, 94104.
Torkildsen, J. V. K., Svangstu, J. M., Friis Hansen, H., Smith, L.,
Simonsen, H. G., Moen, I., et al. (2008). Productive
vocabulary size predicts event-related potential correlates of
fast mapping in 20-month-olds. Journal of Cognitive
Neuroscience, 20, 12661282.
van den Brink, D., Brown, C. M., & Hagoort, P. (2001).
Electrophysiological evidence for early contextual influences
during spoken-word recognition: N200 versus N400 effects.
Journal of Cognitive Neuroscience, 13, 967985.
Wagner, A. W., Schacter, D. L., Rotte, M., Koutstaal, W., Maril, A.,
Dale, A. M., et al. (1998). Building memories: Remembering
and forgetting of verbal experiences as predicted by brain
activity. Science, 281, 11881190.
Wang, X., Ossher, L., & Reuter-Lorenz, P. A. (2015). Examining
the relationship between skilled music training and attention.
Consciousness and Cognition, 36, 169179.
Wechsler, D. (2003). Wechsler Intelligence Scale for
children (4th ed.). San Antonio, TX: The Psychological
Corporation.
Woldorff, M., & Hillyard, S. A. (1991). Modulation of early
auditory processing during selective listening to rapidly
presented tones. Electroencephalography and Clinical
Neurophysiology, 79, 170191.
Wong, P. C. M., & Perrachione, T. K. (2007). Learning pitch
patterns in lexical identification by native English-speaking
adults. Applied Psycholinguistics, 28, 565585.
Zuk, J., Benjamin, C., Kenyon, A., & Gaab, N. (2014). Behavioral
and neural correlates of executive functioning in musicians
and non-musicians. PLoS One, 9, 99868.
1602 Journal of Cognitive Neuroscience Volume 28, Number 10
... Given that during associative learning the activation of one item can trigger the mnemonic representation of the related element, predictive processes are inextricability linked to memory (Bar, 2009). Drawing on this background, learning the meaning of novel words through picture-word associations (Dittinger et al., 2016;Dittinger et al., 2017;Dobel et al., 2009) is a prime example of associative learning that opens up the possibility of examining the neural basis of predictive processes. In particular, based on previous EEG studies that provided empirical evidence for the suitability of the auditory N200, N400 and LPC components as biomarkers of word learning (Bakker et al., 2015;Borovsky et al., 2010;de Diego-Balaguer et al., 2016;Dittinger et al., 2016;Dittinger et al., 2017;Francois et al., 2017;Mestres-Misse et al., 2007;Takashima et al., 2017;Zwitserlood et al., 2018), one would expect that the same EEG indices are likewise sensitive to the preactivation of learned words in response to pictures preceding the associated words. ...
... Drawing on this background, learning the meaning of novel words through picture-word associations (Dittinger et al., 2016;Dittinger et al., 2017;Dobel et al., 2009) is a prime example of associative learning that opens up the possibility of examining the neural basis of predictive processes. In particular, based on previous EEG studies that provided empirical evidence for the suitability of the auditory N200, N400 and LPC components as biomarkers of word learning (Bakker et al., 2015;Borovsky et al., 2010;de Diego-Balaguer et al., 2016;Dittinger et al., 2016;Dittinger et al., 2017;Francois et al., 2017;Mestres-Misse et al., 2007;Takashima et al., 2017;Zwitserlood et al., 2018), one would expect that the same EEG indices are likewise sensitive to the preactivation of learned words in response to pictures preceding the associated words. This perspective is based on the assumption that if a specific component is involved in tracking particular aspects of auditory word learning (e.g., phonological categorization, lexical-semantic processing, etc.), then memory-related pre-activation of determined features should principally be manifested in a modulation of the same ERPs. ...
... Such a top-down modulation of auditory-related ERPs has been shown, for example, for both the imagery of syllables and musical items (Meyer et al., 2007). Furthermore, during word learning, these ERP components typically shift along an anterior-posterior topographical axis, with frontal manifestations at early stages of word learning and posterior distributions when learning is stabilized (Bakker et al., 2015;Borovsky et al., 2010;Dittinger et al., 2016;Dittinger et al., 2017;Elmer et al., 2021). Importantly, since faithful predictions rely on the retrieval of already learned associations, one would expect that genuine predictive processes are most likely detectable at central and posterior scalp sites. ...
Article
Full-text available
Human beings continuously make use of learned associations to generate predictions about future occurrences in the environment. Such memory-related predictive processes provide a scaffold for learning in that mental representations of foreseeable events can be adjusted or strengthened based on a specific outcome. Learning the meaning of novel words through picture-word associations constitutes a prime example of associative learning because pictures preceding words can trigger word prediction through the pre-activation of the related mnemonic representations. In the present electroencephalography (EEG) study, we used event-related potentials (ERPs) to compare neural indices of word pre-activation between a word learning condition with maximal prediction likelihood and a non-learning control condition with low prediction. Results revealed that prediction-related N400 amplitudes in response to pictures decreased over time at central electrodes as a function of word learning, whereas late positive component (LPC) amplitudes increased. Notably, N400 but not LPC changes were also predictive of word learning performance, suggesting that the N400 component constitutes a sensitive marker of word pre-activation during associative word learning.
... Individuals with musical training are reported to have enhanced cognitive abilities including working memory, attention, and executive functioning (Schellenberg & Weiss, 2013;Talamini et al., 2017) than those without musical training. Moreover, they appear to show an aptitude for learning languages, often outperforming their non-musically trained counterparts in various linguistic tasks including word learning (Chobert & Besson, 2013;Dittinger et al., 2016;Kilgour et al., 2000). Thus, given the enhanced cognitive resources/abilities and linguistic abilities associated with musical training, it is possible that musical training may modulate the effect of music on word learning. ...
... Our results indicated that across conditions young adults performed better than older adults, unsurprisingly, given that older adults tend to have poorer declarative memory (Korchin & Basowitz, 1957;Meijer et al., 2008;Service & Craik, 1993). There was also a marginal positive relationship between word learning performance and years of musical training across conditions, which is also expected given that musical training appears to enhance general cognitive abilities and is related to higher linguistic aptitude (Chobert & Besson, 2013;Dittinger et al., 2016;Kilgour et al., 2000;Schellenberg & Weiss, 2013;Talamini et al., 2017). The older adults generally had less musical training experience than the younger adults in our sample, and so while we took that into consideration in our model (i.e., we mean-centred their musical experience by age group), it is possible that the effect of musical training may be stronger if there were more musically trained older adults in our sample. ...
Article
Full-text available
While anecdotal evidence suggests that music may facilitate verbal memory, empirical evidence for this is less clear. Here, we examined whether learners’ characteristics such as age, working memory (WM), and musical training may influence the effect of music on word learning. Young and older adults learned novel word-referent mappings presented in three music conditions (spoken in the presence of background music, sung in-key, and sung out-of-key) and a control condition (spoken in quiet) and their performance was assessed immediately after learning. We found that whereas age and, to an extent, musical training had a general effect on word learning, WM modulated the effect of music: performance was worse in the music conditions relative to the control condition for learners with lower WM whereas the opposite pattern was observed for those with higher WM. Our results thus highlight the importance of considering individual characteristics in determining the effect of music on verbal memory.
... Yet, the progress toward a comprehensive account of the musical mind and its cognitive make-up has been impeded by overreliance on correlational evidence (e.g., Dittinger et al., 2016Dittinger et al., , 2017Dittinger et al., , 2019Kühnis et al., 2013;Pinheiro et al., 2015;Silas et al., 2022;Swaminathan & Gopinath, 2013). To date, the exact mechanisms underlying possible causal relationships between musical training, language, and general cognitive benefits have remained poorly captured (Schellenberg, 2020). ...
Article
Full-text available
A growing body of research has been studying cognitive benefits that arise from music training in childhood or adulthood. Many studies focus specifically on the cognitive transfer of music training to language skill, with the aim of preventing language deficits and disorders and improving speech. However, predicted transfer effects are not always documented and not all findings replicate. While we acknowledge the important work that has been done in this field, we highlight the limitations of the persistent dichotomy between musicians and nonmusicians and argue that future research would benefit from a movement towards skill-based continua of musicianship instead of the currently widely practiced dichotomization of participants into groups of musicians and nonmusicians. Culturally situated definitions of musicianship as well as higher awareness of language diversity around the world are key to the understanding of potential cognitive transfers from music to language (and back). We outline a gradient approach to the study of the musical mind and suggest the next steps that could be taken to advance the field.
... Music experience and expertise are often associated with language ability, as adult or child musicians outperform their non-musician counterparts on a variety of tasks including verbal memory (Chan et al., 1998), unfamiliar spoken lyric recall (Kilgour et al., 2000), verbal sequencing (Piro & Ortiz, 2009), syllable discrimination (Zuk, 2014), speech perception and segmentation in noise (Bidelman & Yoo, 2020;Du & Zatorre, 2017), semantic encoding (Dittinger et al., 2016;, and learning of foreign languages (Yang et al., 2014). ...
Article
Full-text available
Listening to sung words rather than spoken words can facilitate word learning and memory in adults and school-aged children. To explore the development of this effect in young children, this study examined word learning (assessed as forming word-object associations) in 1-2-year-olds and 3-4-year-olds, and word long term memory (LTM) in 4-5-year-olds several days after the initial learning. In an intermodal preferential looking paradigm, children were taught a pair of words utilizing adult-directed speech (ADS) and a pair of sung words. Word learning performance was better with sung words than with ADS words in 1-2-year-olds (Experiments 1a, 1b), 3-4-year-olds (Experiment 1a), and 4-5-year-olds (Experiment 2b), revealing a benefit of song in word learning in all age ranges recruited. We also examined whether children successfully learned the words by comparing their performance against chance. The 1-2-year-olds only learned sung words, but the 3-4-year-olds learned both sung and ADS words, suggesting that the reliance on music features in word learning observed at ages 1-2 decreased with age. Furthermore, song facilitated the word mapping–recognition processes. Results on children’s LTM performance showed that the 4-5-year-olds’ LTM performance did not differ between sung and ADS words. However, the 4-5-year-olds reliably recalled sung words but not spoken words. The reliable LTM of sung words arose from hearing sung words during the initial learning rather than at test. Finally, the benefit of song on word learning and the reliable LTM of sung words observed at ages 3-5 cannot be explained as an attentional effect.
... Even for the infants, connectedness with the experimenters manipulated by interpersonal motor synchrony would predict the subsequent helping behavior [23]. Additionally, those who perceived a stronger connectedness with nature were more likely to engage in pro-environment behaviors [24][25][26]. All of these findings imply that connectedness can promote individuals' prosocial behavior and it may mediate the link between awe and prosocial behavior. ...
Article
Full-text available
Awe is an emotion frequently experienced by individuals in different cultures. When individuals experience awe, they would feel a sense of connectedness to other people or nature arises, and shift their attention to the outside world, which would increase empathy for others in need and, in turn, improve their prosocial tendencies. To test this proposal, we applied a cross-sectional study using a questionnaire survey to collect a sample of 1545 (Nfemale = 988) in Asia, aged between 16 and 71 years old (M = 22.81, SD = 7.80). The Structural Equation Model and bootstrapping method were used to test the mediation effects of connectedness and empathy between awe and prosocial tendency. Results showed that dispositional awe positively predicted a prosocial tendency, which could be partially explained by the multiple mediation effects of connectedness and empathy, after controlling for the effect of the small self. The findings deepen researchers’ understanding of the underlying mechanisms of the relationship between awe and prosociality and have practical implications for cultivating awe and prosocial behavior.
... Further, the type and modality of the tasks used to evaluate cognitive and executive functions can also affect the results. Specifically, it has been proposed that the benefits of music practice on cognition are stronger on more complex or cognitively demanding tasks (Blain, Talamini, Fornoni, Bidet-Caulet, & Caclin, 2022;Dittinger et al., 2016). In sum, although it remains to be determined to what extent and under what conditions choral singing has a positive effect on cognition, the literature indicates that singing has a positive effect on specific aspects of executive functions. ...
Article
The notion that lifestyle factors, such as music-making activities, can affect cognitive functioning and reduce cognitive decline in aging is often referred to as the mental exercise hypothesis. One ubiquitous musical activity is choir singing. Like other musical activities, singing is hypothesized to impact cognitive and especially executive functions. Despite the commonness of choir singing, little is known about the extent to which singing can affect cognition in adulthood. In this cross-sectional group study, we examined the relationship between age and four auditory executive functions to test hypotheses about the relationship between the level of mental activity and cognitive functioning. We also examined pitch discrimination capabilities. A non-probabilistic sample of 147 cognitively healthy adults was recruited, which included 75 non-singers (mean age 52.5 ± 20.3; 20–98 years) and 72 singers (mean age 55.5 ± 19.2; 21–87 years). Tests of selective attention, processing speed, inhibitory control, and working memory were administered to all participants. Our main hypothesis was that executive functions and age would be negatively correlated, and that this relationship would be stronger in non-singers than singers, consistent with the differential preservation hypothesis. The alternative hypothesis - preserved differentiation – predicts that the difference between singers and non-singers in executive functions is unaffected by age. Our results reveal a detrimental effect of age on processing speed, selective attention, inhibitory control and working memory. The effect of singing was comparatively more limited, being positively associated only with frequency discrimination, processing speed, and, to some extent, inhibitory control. Evidence of differential preservation was limited to processing speed. We also found a circumscribed positive impact of age of onset and a negative impact of singing experience on cognitive functioning in singers. Together, these findings were interpreted as reflecting an age-related decline in executive function in cognitively healthy adults, with specific and limited positive impacts of singing, consistent with the preserved differentiation hypothesis, but not with the differential preservation hypothesis.
Chapter
This study analyzes the incidence of the type of grouping strategy (homogeneous or heterogeneous groups) on a virtual learning environment for development of reading comprehension and writing expression skills and for inclusion self-perception of college students with phonological dyslexia diagnosed. A quasi-experimental design is established that considers writing an individual essay as pre- and post-test measures; and writing group essays in Blackboard Collaborative as the treatment or test. Concretely, it is selected a sample of eight groups (four homogeneous and four heterogeneous) made up of three members (one diagnosed with phonological dyslexia). The results showed that dyslexic students integrated into heterogeneous groups when working cooperatively, and felt more included in their groups. This had positive repercussions on the individual essays carried out by these students, as they move from a low competence level to a medium competence level.
Article
Full-text available
El llenguatge parlat i la música constitueixen dos sistemes de comunicació complexos, que presenten una gran diversitat cultural però que, malgrat això, es construeixen de manera relativament similar. Ambdós sorgeixen de la combinació d’elements més simples: fonemes, paraules i frases que es van encadenant en el llenguatge, i sons, melodies, harmonies, ritmes i timbres en la música; es regeixen per una sèrie de normes, també culturalment diverses però que en tots els casos serveixen per facilitar la comprensió del missatge que transmeten; permeten comunicar estats interns, inclosos els emocionals; reflecteixen aspectes de la realitat i al mateix temps permeten alterar mentalment els models de realitat; en la seva gènesi dinàmica hi ha implicats processos creatius, i tant la seva producció com comprensió se sustenten en les capacitats funcionals del cervell. Tenen alguna cosa en comú la música i el llenguatge parlat pel que fa a la seva gènesi, comprensió i interpretació, dins el cervell? O, dit d’una altra manera, comparteixen bases neuronals? En cas afirmatiu, els podem utilitzar en contextos educatius per potenciar-los ambdós de manera sinèrgica? En aquest article es fa una revisió de l’estat actual sobre aquest tema i es discuteix quina és la base biològica, tant evolutiva com neuronal, de la música i el llenguatge parlat; es veurà com interactuen sinèrgicament dins el cervell, i es proposarà des de la neuroeducació de quina manera es poden utilitzar en contextos educatius per potenciar-los mútuament, amb els avantatges formatius, cognitius i de construcció de la identitat que això suposa.
Article
Listening to sung words rather than spoken words can facilitate word learning and memory in adults and school-aged children. To explore the development of this effect in young children, this study examined word learning (assessed as forming word-object associations) in 1-2-year-olds and 3-4-year-olds, and word long term memory (LTM) in 4-5-year-olds several days after the initial learning. In an intermodal preferential looking paradigm, children were taught a pair of words utilizing adult-directed speech (ADS) and a pair of sung words. Word learning performance was better with sung words than with ADS words in 1-2-year-olds (Experiments 1a, 1b), 3-4-year-olds (Experiment 1a), and 4-5-year-olds (Experiment 2b), revealing a benefit of song in word learning in all age ranges recruited. We also examined whether children successfully learned the words by comparing their performance against chance. The 1-2-year-olds only learned sung words, but the 3-4-year-olds learned both sung and ADS words, suggesting that the reliance on music features in word learning observed at ages 1-2 decreased with age. Furthermore, song facilitated the word mapping-recognition processes. Results on children's LTM performance showed that the 4-5-year-olds' LTM performance did not differ between sung and ADS words. However, the 4-5-year-olds reliably recalled sung words but not spoken words. The reliable LTM of sung words arose from hearing sung words during the initial learning rather than at test. Finally, the benefit of song on word learning and the reliable LTM of sung words observed at ages 3-5 cannot be explained as an attentional effect.
Article
Full-text available
Learning new words in an unfamiliar language is a complex endeavor that requires the orchestration of multiple perceptual and cognitive functions. Although the neural mechanisms governing word learning are becoming better understood, little is known about the predictive value of resting-state (RS) metrics for foreign word discrimination and word learning attainment. In addition, it is still unknown which of the multistep processes involved in word learning have the potential to rapidly reconfigure RS networks. To address these research questions, we used electroencephalography (EEG), measured forty participants, and examined scalp-based power spectra, source-based spectral density maps and functional connectivity metrics before (RS1), in between (RS2) and after (RS3) a series of tasks which are known to facilitate the acquisition of new words in a foreign language, namely word discrimination, word-referent mapping and semantic generalization. Power spectra at the scalp level consistently revealed a reconfiguration of RS networks as a function of foreign word discrimination (RS1 vs. RS2) and word learning (RS1 vs. RS3) tasks in the delta, lower and upper alpha, and upper beta frequency ranges. Otherwise, functional reconfigurations at the source level were restricted to the theta (spectral density maps) and to the lower and upper alpha frequency bands (spectral density maps and functional connectivity). Notably, scalp RS changes related to the word discrimination tasks (difference between RS2 and RS1) correlated with word discrimination abilities (upper alpha band) and semantic generalization performance (theta and upper alpha bands), whereas functional changes related to the word learning tasks (difference between RS3 and RS1) correlated with word discrimination scores (lower alpha band). Taken together, these results highlight that foreign speech sound discrimination and word learning have the potential to rapidly reconfigure RS networks at multiple functional scales.
Article
Full-text available
From an early age, musicians learn complex motor and auditory skills (e.g., the translation of visually perceived musical symbols into motor commands with simultaneous auditory monitoring of output), which they practice extensively from childhood throughout their entire careers. Using a voxel-by-voxel morphometric technique, we found gray matter volume differences in motor, auditory, and visual-spatial brain regions when comparing professional musicians (keyboard players) with a matched group of amateur musicians and non-musicians. Although some of these multiregional differences could be attributable to innate predisposition, we believe they may represent structural adaptations in response to long-term skill acquisition and the repetitive rehearsal of those skills. This hypothesis is supported by the strong association we found between structural differences, musician status, and practice intensity, as well as the wealth of supporting animal data showing structural changes in response to long-term motor training. However, only future experiments can determine the relative contribution of predisposition and practice.
Article
Full-text available
Neural overlap in processing music and speech, as measured by the co-activation of brain regions in neuroimaging studies, may suggest that parts of the neural circuitries established for language may have been recycled during evolution for musicality, or vice versa that musicality served as a springboard for language emergence. Such a perspective has important implications for several topics of general interest besides evolutionary origins. For instance, neural overlap is an important premise for the possibility of music training to influence language acquisition and literacy. However, neural overlap in processing music and speech does not entail sharing neural circuitries. Neural separability between music and speech may occur in overlapping brain regions. In this paper, we review the evidence and outline the issues faced in interpreting such neural data, and argue that converging evidence from several methodologies is needed before neural overlap is taken as evidence of sharing. © 2015 The Author(s) Published by the Royal Society. All rights reserved.
Article
In the first comprehensive study of the relationship between music and language from the standpoint of cognitive neuroscience, the author challenges the widespread belief that music and language are processed independently. Since Plato's time, the relationship between music and language has attracted interest and debate from a wide range of thinkers. Recently, scientific research on this topic has been growing rapidly, as scholars from diverse disciplines including linguistics, cognitive science, music cognition, and neuroscience are drawn to the music-language interface as one way to explore the extent to which different mental abilities are processed by separate brain mechanisms. Accordingly, the relevant data and theories have been spread across a range of disciplines. This book provides the first synthesis, arguing that music and language share deep and critical connections, and that comparative research provides a powerful way to study the cognitive and neural mechanisms underlying these uniquely human abilities.
Article
While many aspects of cognition have been investigated in relation to skilled music training, surprisingly little work has examined the connection between music training and attentional abilities. The present study investigated the performance of skilled musicians on cognitively demanding sustained attention tasks, measuring both temporal and visual discrimination over a prolonged duration. Participants with extensive formal music training were found to have superior performance on a temporal discrimination task, but not a visual discrimination task, compared to participants with no music training. In addition, no differences were found between groups in vigilance decrement in either type of task. Although no differences were evident in vigilance per se, the results indicate that performance in an attention-demanding temporal discrimination task was superior in individuals with extensive music training. We speculate that this basic cognitive ability may contribute to advantages that musicians show in other cognitive measures. Copyright © 2015 Elsevier Inc. All rights reserved.
Presents a standardized set of 260 pictures for use in experiments investigating differences and similarities in the processing of pictures and words. The pictures are black-and-white line drawings executed according to a set of rules that provide consistency of pictorial representation. They have been standardized on 4 variables of central relevance to memory and cognitive processing: name agreement, image agreement, familiarity, and visual complexity. The intercorrelations among the 4 measures were low, suggesting that they are indices of different attributes of the pictures. The concepts were selected to provide exemplars from several widely studied semantic categories. Sources of naming variance, and mean familiarity and complexity of the exemplars, differed significantly across the set of categories investigated. The potential significance of each of the normative variables to a number of semantic and episodic memory tasks is discussed. (34 ref) (PsycINFO Database Record (c) 2006 APA, all rights reserved).
Article
In animal models the neural basis of cognitive and executive processes has been studied extensively at various hierarchical levels from microcircuits to distributed functional networks. This work already provides compelling evidence that diverse cognitive functions are based on similar basic neuronal mechanisms. More recent data suggest that even cognitive functions realized only in human brains rely on these canonical neuronal mechanisms. Here we argue that language, like other cognitive functions, depends on distributed computations in specialized cortical areas forming large-scale dynamic networks and examine to what extent empirical results support this view. Copyright © 2015 Elsevier Ltd. All rights reserved.