ArticlePDF Available

Evidence of degraded representation of speech in noise, in the aging midbrain and cortex

Authors:

Abstract and Figures

Humans have a remarkable ability to track and understand speech in unfavorable conditions, such as in background noise, but speech understanding in noise does deteriorate with age. Results from several studies have shown that in younger adults, low frequency auditory cortical activity reliably synchronizes to the speech envelope, even when the background noise is considerably louder than the speech signal. However, cortical speech processing may be limited by age-related decreases in the precision of neural synchronization in the midbrain. To better understand the neural mechanisms contributing to impaired speech perception in older adults, we investigated how aging affects midbrain and cortical encoding of speech when presented in quiet and in the presence of a single competing talker. Our results suggest that central auditory temporal processing deficits in older adults manifest in both the midbrain and in the cortex. Specifically, midbrain frequency following responses to a speech syllable are more degraded in noise in older adults than in younger adults. This suggests a failure of the midbrain auditory mechanisms needed to compensate for the presence of a competing talker. Similarly, in cortical responses, older adults show larger reductions than younger adults in their ability to encode the speech envelope when a competing talker is added. Interestingly, older adults showed an exaggerated cortical representation of speech in both quiet and noise conditions, suggesting a possible imbalance between inhibitory and excitatory processes, or diminished network connectivity, that may impair their ability to efficiently encode speech.
Content may be subject to copyright.
Evidence of degraded representation of speech in noise, in the aging midbrain
and cortex
Alessandro Presacco,
1,2
Jonathan Z. Simon,
2,3,4,5
and Samira Anderson
1,2
1
Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland;
2
Neuroscience and Cognitive
Science Program, University of Maryland, College Park, Maryland;
3
Department of Electrical and Computer Engineering,
University of Maryland, College Park, Maryland;
4
Department of Biology, University of Maryland, College Park, Maryland;
and
5
Institute for Systems Research, University of Maryland, College Park, Maryland
Submitted 16 May 2016; accepted in final form 12 August 2016
Presacco A, Simon JZ, Anderson S. Evidence of degraded rep-
resentation of speech in noise, in the aging midbrain and cortex. J
Neurophysiol 116: 2346 –2355, 2016. First published August 17,
2016; doi:10.1152/jn.00372.2016.—Humans have a remarkable abil-
ity to track and understand speech in unfavorable conditions, such as
in background noise, but speech understanding in noise does deteri-
orate with age. Results from several studies have shown that in
younger adults, low-frequency auditory cortical activity reliably syn-
chronizes to the speech envelope, even when the background noise is
considerably louder than the speech signal. However, cortical speech
processing may be limited by age-related decreases in the precision of
neural synchronization in the midbrain. To understand better the
neural mechanisms contributing to impaired speech perception in
older adults, we investigated how aging affects midbrain and cortical
encoding of speech when presented in quiet and in the presence of a
single-competing talker. Our results suggest that central auditory
temporal processing deficits in older adults manifest in both the
midbrain and in the cortex. Specifically, midbrain frequency follow-
ing responses to a speech syllable are more degraded in noise in older
adults than in younger adults. This suggests a failure of the midbrain
auditory mechanisms needed to compensate for the presence of a
competing talker. Similarly, in cortical responses, older adults
show larger reductions than younger adults in their ability to
encode the speech envelope when a competing talker is added.
Interestingly, older adults showed an exaggerated cortical repre-
sentation of speech in both quiet and noise conditions, suggesting
a possible imbalance between inhibitory and excitatory processes,
or diminished network connectivity that may impair their ability to
encode speech efficiently.
aging; electrophysiology; midbrain; cortex; hearing
NEW & NOTEWORTHY
We investigate the underlying neurophysiology of age-
related auditory temporal processing deficits in normal-
hearing listeners using natural speech (in noise). Two
neurophysiological techniques are used—magnetoen-
cephalography and EEG—to investigate two different
brain areas— cortex and midbrain—within each subject.
Older adults have more exaggerated cortical speech rep-
resentations than younger adults in both quiet and noise.
Midbrain speech representations depend more critically on
noise level and synchronize more weakly in older adults
than younger.
THE ABILITY TO TRACK AND UNDERSTAND speech in the presence of
interfering speakers is one of the most complex communication
challenges experienced by humans. In a complex auditory
scene, both humans and animals show an innate ability to
detect and recognize individual auditory objects, an important
component in the process of stream segregation. The ability to
transform the noise-corrupted acoustic signal into a neural
representation suitable for speech recognition may occur in the
auditory cortex via adaptive neural encoding (Ding et al. 2014;
Ding and Simon 2012, 2013). Specifically, low-frequency
auditory cortical activity recorded with magnetoencephalogra-
phy (MEG) reliably synchronizes to the slow temporal modu-
lations of speech, even when the energy of the background
noise is considerably higher than the speech signal and even
when the background noise is also speech (Ding and Simon
2012). However, the accuracy of cortical speech processing
may also be affected by the precision of neural synchronization
in the auditory midbrain, as seen in studies that compare
cortical responses with those using the frequency following
response (FFR), believed to arise primarily from the midbrain
(Chandrasekaran and Kraus 2010). For example, noise has a
greater impact on the robustness of cortical speech processing
in children (with learning impairments) who have delayed peak
latencies in FFRs to a speech syllable (King et al. 2002). In
normal-hearing young adults, earlier peak latencies in the FFR
are associated with larger N1 amplitudes in cortical responses
to speech in noise, and larger N1 amplitudes are associated
with a better ability to recognize sentences in noise (Parbery-
Clark et al. 2011). Furthermore, Bidelman et al. (2014) dem-
onstrated that age-related temporal speech-processing deficits
arising from the midbrain may be compensated by a stronger
cortical response. Recent work from Chambers et al. (2016)
showed that a combination of profound cochlear denervation
and desynchronization can result in absence of wave I in the
brain stem but not in the cortex, suggesting compensatory
central gain increases that help restore the representation of the
auditory object in auditory cortex. Whereas these studies ex-
amined age- and hearing loss-related changes in midbrain and
cortical encoding of vowels and tones presented in quiet, the
comparison between midbrain and cortical encoding of speech
syllables and sentences presented in competing single-talker
speech has not yet been investigated in either younger or older
adults.
Such auditory temporal processing deficits are of great
relevance, since communication difficulties for older adults
have a significant social impact, with strong correlations seen
Address for reprint requests and other correspondence: A. Presacco, Dept. of
Otolaryngology, Univ. of California, Irvine. Medical Sciences D, Room 404,
Irvine, CA, 92697 (e-mail: presacca@uci.edu).
J Neurophysiol 116: 2346 –2355, 2016.
First published August 17, 2016; doi:10.1152/jn.00372.2016.
2346 0022-3077/16 Copyright © 2016 the American Physiological Society www.jn.org
between hearing loss and depression (Carabellese et al. 1993;
Herbst and Humphrey 1980; Kay et al. 1964; Laforge et al.
1992) and cognitive impairment (Gates et al. 1996; Lin et al.
2013; Uhlmann et al. 1989). Although audibility is an impor-
tant factor in the older adult’s ability to understand speech
(Humes and Christopherson 1991; Humes and Roberts 1990),
the use of hearing aids often does not improve speech under-
standing in noise, perhaps because increased audibility cannot
restore temporal precision degraded by aging. Several electro-
physiological studies in humans and animals support the hy-
pothesis that degraded auditory temporal processing may play
a role in explaining speech-in-noise problems experienced by
older adults (Alain et al. 2014; Anderson et al. 2012; Clinard
and Tremblay 2013; Lister et al. 2011; Parthasarathy and
Bartlett 2011; Presacco et al. 2015; Ross et al. 2010; Soros
et al. 2009).
To investigate further the neural mechanisms underlying
age-related deficits in speech-in-noise understanding, this cur-
rent study evaluated the effects of aging on temporal synchro-
nization of speech in the presence of a competing talker in both
cortex and midbrain. To de-emphasize the effects of audibility,
only clinically normal-hearing listeners were included in both
the younger and older age groups. We posit several hypotheses.
First, in responses arising from midbrain, we hypothesize that
younger adults encode speech with greater neural fidelity,
reflected by higher amplitude responses and higher stimulus-
to-response and quiet-to-noise correlations than older adults
when the signal is presented in quiet and in noise. This
hypothesis was driven by the results of the above-mentioned
studies, showing more robust and less jittered responses in
quiet in younger adults (Anderson et al. 2012; Clinard and
Tremblay 2013; Mamo et al. 2016; Presacco et al. 2015) and an
age-related effect of noise (Parthasarathy et al. 2010). In
contrast, for cortical responses, we hypothesize that older
adults will show an over-representation of the response both in
quiet and noise. This hypothesis is driven by evidence showing
age-related increases in amplitude (Alain et al. 2014; Soros et
al. 2009) and in latency (Tremblay et al. 2003) of the main
peaks of auditory cortical responses. Finally, we hypothesize
that within an individual, better speech-in-noise understanding
(at the behavioral level) correlates with greater fidelity of
neural encoding of speech, regardless of age.
MATERIALS AND METHODS
The experimental protocol and all procedures were reviewed and
approved by the Institutional Review Board of the University of
Maryland. Participants gave written, informed consent, according to
principles set forth by the University of Maryland’s Institutional
Review Board, and were paid for their time.
Participants
Participants comprised 17 younger adults (18 –27 yr, means SD
22.23 2.27, 3 men) and 15 older adults (61–73 yr, means SD
65.06 3.30, 5 men), recruited from the Maryland; Washington,
D.C.; and Virginia areas. All participants had clinically normal hear-
ing (Fig. 1), defined as follows: 1) air conduction thresholds 25 dB
hearing level from 125 to 4,000 Hz bilaterally and 2) no interaural
asymmetry (15 dB hearing-level difference at no more than 2
adjacent frequencies). Participants had a normal intelligence quotient
[scores 85 on the Wechsler Abbreviated Scale of Intelligence (Zhu
and Garcia 1999)] and were not significantly different on intelligence
quotient [F
(1,30)
0.660, P0.423] and sex (Fisher’s exact, P
0.05). Because of the established effects of musicianship on subcor-
tical auditory processing (Bidelman and Krishnan 2010; Parbery-
Clark et al. 2012), professional musicians were excluded. In addition,
the older adults were screened for dementia on the Montreal Cognitive
Assessment (Nasreddine et al. 2005). All participants spoke English
as their first language, and none of them were tonal language speakers.
Speech Intelligibility
The Quick Speech-in-Noise test (QuickSIN) (Killion et al. 2004)
was used to quantify the ability to understand speech presented in
noise composed of four-talker babble.
EEG: Stimuli and Recording
A 170-ms/da/(Anderson et al. 2012) was synthesized at a 20-kHz
sampling rate with a Klatt-based synthesizer (Klatt 1980). The stim-
Fig. 1. Audiogram (mean 1 SE) of the grand averages of both
ears of younger (gray) and older (black) adults. All participants
have clinically normal hearing. HL, hearing level.
2347AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
ulus was presented at an 80 peak-dB sound-pressure level diotically
with alternating polarities at a rate of 4 Hz through electromagneti-
cally shielded insert earphones (ER·1; Etymotic Research, Elk Grove
Village, IL) via Xonar Essence One (ASUS, Taipei, Taiwan) using
Presentation software (Neurobehavioral Systems, Berkeley, CA). A
single-competing female talker narrating A Christmas Carol by
Charles Dickens was used as the background noise. FFRs were
recorded in quiet and in noise [signal-to-noise ratios (SNRs): 3, 0,
3, and 6 dB] at a sampling frequency of 16,384 Hz using the
ActiABR-200 acquisition system (BioSemi B.V., Amsterdam, Neth-
erlands) with a standard vertical montage of five electrodes (Cz active,
2 forehead ground common mode sense/driven right leg electrodes,
earlobe references) and with an online 100- to 3,000-Hz bandpass
filter. During the recording session (1 h), participants sat in a
recliner and watched a silent, captioned movie of their choice to
facilitate a relaxed yet wakeful state. Artifact-free sweeps (2,300)
were recorded for each condition from each participant.
Data analysis. Data recorded with BioSemi B.V. were analyzed in
MATLAB (version R2011b; MathWorks, Natick, MA) after being
converted into MATLAB format with the function pop_biosig from
EEGLab (Scott Makeig, Swartz Center for Computational Neurosci-
ence, University of California, San Diego, CA) (Delorme and Makeig
2004). Sweeps with amplitude in the 30-
V range were retained and
averaged in real time and then processed offline. The time window for
each sweep was 47 to 189 ms, referenced to the stimulus onset.
Responses were digitally bandpass filtered offline from 70 to 2,000 Hz
using a fourth-order Butterworth filter to minimize the effects of
cortical low-frequency oscillations (Galbraith et al. 2000; Smith et al.
1975). A final average response was created by averaging the sweeps
of both polarities to minimize the influence of cochlear microphonic
and stimulus artifact on the response and simultaneously maximize
the envelope response (Aiken and Picton 2008; Campbell et al. 2012;
Gorga et al. 1985). Root-mean-square (RMS) values were calculated
for the transition (18 68 ms) and steady-state (68 –170 ms) regions.
Correlation (Pearson’s linear correlation) between the envelope re-
sponse in quiet and noise was calculated for each subject to estimate
the extent to which noise affects the FFR. Pearson’s linear correlation
was also used to quantify the stimulus-to-response correlation in the
steady-state region, during which the response more reliably follows
the stimulus. For this analysis, the envelope of the analytic signal of
the stimulus was extracted and then band-pass filtered using the same
filter as for the response. Average spectral amplitudes over 20 Hz bins
were also calculated from each response using a fast Fourier transform
(FFT) with zero padding and 1 Hz interpolated frequency resolution
over the transition and steady-state regions for the fundamental
frequency (F
0
) and the first two harmonics. An additional analysis
signal was created by subtracting and then averaging the sweeps of the
two polarities to enhance the temporal fine structure (TFS) (Aiken and
Picton 2008). One younger adult was removed from the TFS analysis
because of corruption by stimulus artifact. Average spectral ampli-
tudes over 20 Hz bins were calculated for the TFS from each response
using a FFT with zero padding and 1 Hz interpolated frequency
resolution over the transition and steady-state regions for the frequen-
cies of 400 and 700 Hz, which represent the two main peaks of interest
from the two time regions (Anderson et al. 2012).
MEG: Stimuli and Recording
Participants were asked to attend to one of two stories (foreground)
presented diotically while ignoring the other one. The stimuli for the
foreground consist of segments from the book, The Legend of Sleepy
Hollow by Washington Irving, whereas the stimuli for the background
were the same as were used in the EEG experiment. The foreground
was spoken by a male talker, whereas the background story was
spoken by a female talker. Additional stimuli using a background
narration in an unfamiliar language were also presented, but the
responses to those stimuli are not analyzed here. Each speech mixture
was constructed, as described by Ding and Simon (2012), by digitally
mixing two speech segments into a single channel with a duration of
1 min. Five different conditions were recorded: quiet and 3, 0, 3,
and 6 dB SNR. Four different segments from the same foreground
story were used to minimize the possibility that the clarity of the
stories could affect the performance of the subjects. The same seg-
ment was played for quiet and 6 dB. To maximize the level of
attention of the subject on the foreground segment, participants were
asked beforehand to count the number of times a specific word or
name was mentioned in the story. The sounds, 70 dB sound-
pressure level when presented with a solo speaker, were delivered to
the participants’ ears with 50 sound tubing (E-A-RTONE 3A;
Etymotic Research), attached to E-A-RLINK foam plugs inserted into
the ear canal. The entire acoustic delivery system was equalized to
give an approximately flat transfer function from 40 to 4,000 Hz,
thereby encompassing the range of the delivered stimuli. Neuromag-
netic signals were recorded using a 157-sensor whole-head MEG
system (Kanazawa Institute of Technology, Nonoichi Ishikawa, Ja-
pan) in a magnetically shielded room, as described in Ding and Simon
(2012).
Data analysis. Three reference channels were used to measure and
cancel the environmental magnetic field by using time shift-principal
component analysis (de Cheveigné and Simon 2007). MEG data were
analyzed offline using MATLAB. The 157 raw MEG data channel
responses were first filtered between 2 and 8 Hz, with an order 700
windowed (Hamming) linear-phase finite impulse response filter, then
decomposed using nspatial filters into nsignal components (where
n157) using the denoising source separation (DSS) algorithm (de
Cheveigné and Simon 2008; Särelä and Valpola 2005). The first six
DSS component filters were then used for the analysis. The filtering
range of 2– 8 Hz was chosen based on previous results showing the
absence of intertrial coherence above 8 Hz (Ding and Simon 2013)
and the importance of the integrity of the modulation spectrum above
1 Hz to understand spoken language (Greenberg and Takayuki 2004).
The signal components used for analysis were then re-extracted from
the raw data for each trial, spatially filtered using the six DSS filters
just constructed, band-pass filtered between 1 and 8 Hz (Ding and
Simon 2012) with a second-order Butterworth filter, and averaged
over trials. Reconstruction of the envelope was performed using a
linear reconstruction matrix estimated via the Boosting algorithm
(David et al. 2007; Ding et al. 2014; Ding and Simon 2013). Success
of the reconstruction is measured by the linear correlation between the
reconstructed and actual speech envelope. The reconstructed envelope
was obtained from the unmixed speech of the single speaker to which
the participant was instructed to attend, not from the acoustic stimulus
mixture. The envelope was computed as the 1- to 8-Hz band pass-
filtered magnitude of the analytic signal. Data were analyzed using
three different time windows for this reconstruction model: 500, 350,
and 150 ms. The choice to narrow the integration window down to
150 ms is based on previous results, showing that the ability to track
the speech envelope substantially worsens as the window decreases
down to 100 ms (Ding and Simon 2013). These values refer to the
time shift imposed on our data with respect to the onset of the speech
and to the corresponding integration window of our reconstruction
matrix. Specifically, if processing time for younger and older adults
is the same, then their performance should follow the same pattern
as the integration window changes. Conversely, if older adults
require more time to process the information because of the
possible presence of temporal processing deficits, then the narrow-
ing of the integration window should negatively affect their per-
formance more than for younger adults. The noise floor was
calculated by using the neural response recorded from each con-
dition to reconstruct the speech envelope of a different stimulus
than was used during this response.
2348 AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
Statistical Analyses
All statistical analyses were conducted in SPSS version 21.0 (IBM,
Armonk, NY). Fisher’s z transformation was applied to all of the
correlation values calculated for the midbrain and cortical analysis
before any statistical analysis. Split-plot ANOVAs were used to test
for age-group condition interactions for the RMS values of the FFR
response in the time domain, for the stimulus-to-response correlations
of the FFR, and for correlation values calculated for the cortical data.
The Greenhouse-Geisser test was used when the Mauchly’s sphericity
test was violated. Paired t-tests were used for within-subject group
analysis for the correlation values and amplitudes for the cortical data,
whereas one-way ANOVAs were used to analyze the RMS amplitude
values of the FFR, stimulus-to-noise correlation of the FFR, FFT of
the FFR, quiet-noise correlations, and the correlation values for the
cortical data. The nonparametric Mann-Whitney U-test was used in
place of the one-way ANOVA when Levene’s test for Equality of
Variances was violated. Two-tailed Spearman’s rank correlation (
)
was used to evaluate the relationships among speech-in-noise scores,
midbrain, cortical parameters, and pure-tone average. The false dis-
covery rate procedure (Benjamini and Hochberg 1995) was applied to
control for multiple comparisons where appropriate.
RESULTS
Speech Intelligibility (QuickSIN)
Younger adults (means SD ⫽⫺0.57 1.13 dB SNR
loss) scored significantly better [F
(1,30)
10.613, P0.003]
than older adults (means SD 0.8 1.25 dB SNR loss) on
the QuickSIN test, suggesting that older adults’ performance in
noise may decline compared with younger adults, even when
audiometric thresholds are clinically normal.
Midbrain (EEG)
Amplitude analysis. Figure 2 shows the grand average of
FFRs of the stimulus envelope of younger and older adults in
quiet and in one of the four noise conditions tested (6 dB).
Figure 3Adisplays the RMS values for each condition tested in
younger and older adults in the transition and steady-state
regions. In both regions, the RMS values of the responses in
noise of younger and older adults are significantly higher than
the RMS calculated for the noise floor (all, P0.007).
TRANSITION REGION. A one-way ANOVA showed that
younger adults have significantly higher RMS values in quiet
[F
(1,30)
4.255, P0.048]. When all of the noise conditions
were collapsed together, one-way ANOVA showed significant
differences between younger and older adults [F
(1,126)
5.150,
P0.025; Fig. 3B]. The follow-up results of paired t-tests
suggest that noise significantly decreases response amplitude in
both younger and older adults in all of the noise conditions
tested (all, P0.01). Repeated-measures ANOVA showed a
condition age interaction between quiet and noise at 3dB
[F
(1,30)
6.264, P0.018] and 6dB[F
(1,30)
6.696, P
0.015] but not at the other conditions tested [F
(1,30)
1.125,
P0.297 and F
(1,30)
0.333, P0.568 for 3 and 0 dB,
respectively]. Repeated-measures ANOVA showed significant
differences across noise conditions in younger [F
(3,48)
13.384, P0.001] but not in older [F
(3,48)
0.885, P
0.457] adults (Fig. 3A).
STEADY-STATE REGION. A one-way ANOVA showed that
younger adults have significantly higher RMS values than
older adults in quiet [F
(1,30)
6.877, P0.014]. The fol-
Fig. 2. Grand average (n17 for younger and n15 for
older adults) of the response to the stimulus envelope for
younger (left) and older [right; quiet dark lines; noise (6
dB) light lines] adults. Statistical analyses carried out on
individual subjects show that in both the transition and
steady-state regions, noise resulted in a significant decrease
(P0.01 and 0.05 for the transition and steady-state
region, respectively) in the amplitude response for both
younger and older adults at all of the conditions tested. Higher
RMS values were also found in younger adults in both
regions (P0.05).
Fig. 3. RMS values 1 SE of the envelope for the
conditions (Q Quiet, 3⫽⫹3 dB, 0 0 dB, 3
3 dB, and 6⫽⫺6 dB) tested in younger (gray bars)
and older (black bars) adults. A: average RMS for each
single condition. B: average RMS collapsed across all
noise conditions tested. Younger adults had significantly
higher RMS values in quiet in both the transition and the
steady-state regions. An RMS group-interaction effect
was noted in the transition at 3 and 6 dB but not in the
steady-state region. Repeated-measures ANOVA, applied
to the 4 noise conditions, shows significant differences in
younger adults in both the transition and steady-state
regions but not in older adults. Noise minimally affects
older adults, likely because their response in quiet is
already degraded. *P0.05, ***P0.001.
2349AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
low-up results of paired t-tests suggest that noise significantly
decreases response amplitude in both younger and older adults in
all of the noise conditions tested (all, P0.05). Repeated-
measures ANOVA showed no condition age interaction be-
tween quiet and noise at any of the conditions tested [F
(1,30)
0.072, P0.791; F
(1,30)
0.000, P0.986; F
(1,30)
2.574,
P0.119; and F
(1,30)
3.197, P0.084 for 3; 0; 3; and 6
dB, respectively]. Repeated-measures ANOVA showed signifi-
cant differences across noise conditions in younger [F
(3,48)
19.847, P0.001] but not in older [F
(3,48)
0.874, P0.462]
adults (Fig. 3A). When all of the noise conditions were collapsed
together, a follow-up one-way ANOVA showed significant dif-
ferences between younger and older adults [F
(1,126)
27.364,
P0.001; Fig. 3B].
CORRELATION ANALYSIS. To analyze the robustness of the
response in noise, we linearly correlated (Pearson correlation)
the average response obtained in quiet with that recorded in
noise for both the transition and steady-state regions for
each subject. Repeated-measures ANOVA showed no sig-
nificant noise condition age interaction in either the
transition [F
(3,90)
1.129, P0.342] or the steady-state
[F
(3,90)
1.015, P0.390] region. When all of the noise
conditions were grouped together, a follow-up Mann-Whit-
ney U-test showed significantly higher Fisher-transformed r
values in younger adults in the steady-state [U(128)
1,272, Z⫽⫺3.667, P0.001] but not in the transition
[U(128) 1,675, Z⫽⫺1.743, P0.081] region.
STIMULUS-TO-RESPONSE CORRELATION. Repeated-measures
ANOVA showed a significant noise condition age inter-
action between quiet and noise at all of the noise conditions
tested [F
(1,30)
5.915, P0.021; F
(1,30)
4.302, P
0.047; F
(1,30)
5.786, P0.023; and F
(1,30)
8.318, P
0.007 for 3; 0; 3; and 6 dB, respectively]. A one-way
ANOVA showed that the younger adults’ correlation values
were significantly higher than those of older adults in all of
the noise conditions tested [F
(1,30)
7.768, P0.009;
F
(1,30)
5.535, P0.025; F
(1,30)
5.166, P0.030; and
F
(1,30)
8.838, P0.006 for 3; 0; 3; and 6 dB,
respectively] but not in quiet [U(62) 114, Z⫽⫺0.510,
P0.628].
Frequency analysis (envelope). TRANSITION REGION. A one-
way ANOVA showed no significant amplitude differences in
the transition region at F
0
and at the first two harmonics in all
of the conditions tested (all, Padjusted threshold), except
the second harmonic at 3dB[F
(1,30)
9.046, P0.005].
Repeated-measures ANOVA showed significant differences
across noise conditions only in younger adults and only at the
second harmonic [F
(3,48)
6.141, P0.007].
STEADY-STATE REGION. A one-way ANOVA showed that
younger adults’ F
0
amplitude is significantly higher than that of
older adults in all of the noise conditions tested [F
(1,30)
9.287, P0.005; F
(1,30)
9.598, P0.004; F
(1,30)
6.518,
P0.016; and F
(1,30)
8.901, P0.006 for 3; 0; 3; and
6 dB, respectively] but not in quiet [F
(1,30)
3.390, P
0.076]. No significant differences were found at the second
harmonic (all, P0.05), whereas the amplitude of the third
harmonic is significantly higher in younger adults than in older
adults in all of the noise conditions tested [F
(1,30)
12.744,
P0.001; F
(1,30)
9.259, P0.005; F
(1,30)
4.318, P
0.046; and F
(1,30)
7.517, P0.010 for 3; 0; 3; and 6
dB, respectively] and in quiet [F
(1,30)
26.771, P0.001].
Repeated-measures ANOVA showed significant amplitude dif-
ferences across noise conditions only in younger adults at the
F
0
[F
(3,48)
3.987, P0.013] and at the first [F
(3,48)
3.065,
P0.037] and second [F
(3,48)
8.421, P0.001] harmonics.
Frequency analysis (TFS). TRANSITION REGION. A one-way
ANOVA showed no significant group differences in amplitude
at 400 or 700 Hz at any of the conditions tested (all, P0.05).
Repeated-measures ANOVA showed significant differences
across noise conditions only in younger adults and only at 400
Hz [F
(3,45)
4.406, P0.008] but not for 700 Hz in the
younger adults or for either frequency in the older adults (all,
P0.05).
STEADY-STATE REGION. A one-way ANOVA showed that the
400-Hz amplitude in younger was significantly higher than that
of older adults but only in quiet [F
(1,29)
12.908, P0.001]
and at 0 dB SNR [F
(1,29)
9.654, P0.004]. No significant
differences were found at 700 Hz at any SNR (all, P0.05)
or in quiet (P0.05, Mann-Whitney test). Repeated-measures
ANOVA showed significant amplitude differences across noise
conditions in the younger adults at 400 Hz [F
(3,45)
4.802,
P0.006] but not for 700 Hz in the younger adults or for
either frequency in the older adults (all, P0.05).
Cortex (MEG)
Reconstruction of the attended speech envelope. The ability
to reconstruct the low-frequency speech envelope from cortical
activity is a measure of the fidelity of the neural representation
of that speech envelope (Ding and Simon 2012). Figure 4A
shows an example of reconstruction of the speech envelope of
the foreground in noise (6 dB) from a representative younger
and older adult. Figure 4Bplots the rvalues for each partici-
pant, at each tested condition, in ascending order. Figure 5A
displays the grand average SE of the reconstruction accuracy
for younger and older adults for all of the conditions tested.
One-way ANOVA showed significantly higher correlation val-
ues in older adults compared with younger adults in quiet
[F
(1,30)
14.923, P0.001] and in all of the noise conditions
tested [Fig. 5A;F
(1,30)
13.315, P0.001; F
(1,30)
13.374,
P0.001; F
(1,30)
15.331, P0.001; and F
(1,30)
6.195,
P0.019 for 3; 0; 3; and 6 dB, respectively]. All of the
reconstruction values were significantly higher than the noise
floor (all, P0.01). Since the difference between older and
younger adults is minimized at 6 dB, this condition was used
to analyze the effect of the integration window on the fidelity
of the reconstruction of the speech envelope. Results from a
split-plot ANOVA, applied to the three integration windows
used (Fig. 5B) for the analysis, revealed a reconstruction
window age-group interaction in quiet [F
(2,60)
9.332, P
0.004] but not in noise [F
(2,60)
0.105, P0.802]. Repeated-
measures ANOVA, applied to 500, 350, and 150 ms integra-
tion windows, shows significant differences in older adults in
both quiet [F
(2,32)
14.954, P0.001] and noise [F
(2,32)
5.048, P0.037] but not in younger adults [F
(2,32)
4.213,
P0.048 and F
(2,32)
1.195, P0.302 in quiet and noise,
respectively]. A follow-up paired t-test of the foreground
reconstructed in quiet and noise at 500 vs. 350 ms and 500 vs.
150 ms showed that the reconstruction accuracy of younger
adults is not significantly affected by the integration windows
in noise [t(16) 0.366, P0.719 and t(16) 1.162, P
0.262 for 500 vs. 350 ms and 500 vs. 150 ms, respectively],
whereas in quiet, the 500-ms integration window had signifi-
2350 AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
cantly lower values than 350 ms but not than 150 ms [t(16)
3.722, P0.002 and t(16) 0.973, P0.345 for 500 vs.
350 ms and 500 vs. 150 ms, respectively]. Conversely, older
adults’ ability to track the speech envelope of the foreground is
significantly reduced at 350 and 150 ms in both quiet [t(14)
0.248, P0.807 and t(14) 3.779, P0.002 for 500 vs.
350 ms and 500 vs. 150 ms, respectively] and noise [t(14)
2.064, P0.058 and t(14) 2.512, P0.0248 for 500 vs.
350 ms and 500 vs. 150 ms, respectively].
Reconstruction of the unattended speech envelope. Repeated-
measures ANOVA showed a significant correlation age
interaction across the four noise conditions tested [F
(3,90)
2.909,
P0.039]. A one-way ANOVA showed significantly higher
reconstruction accuracy in older adults at all of the noise condi-
tions tested except 3dB[F
(1,30)
3.487, P0.072; F
(1,30)
4.99, P0.033; F
(1,30)
7.523, P0.01; and F
(1,30)
19.251,
P0.001 for 3; 0; 3; and 6 dB, respectively]. All of the
reconstruction values were significantly higher than the noise floor
(all, P0.01).
Relationships among Behavioral, Midbrain, and Cortical
Data
Two-tailed Spearman’s rank correlation coefficient was used
to study the correlations among the following measurements:
speech-in-noise score, cortical decoding accuracy in quiet and
in noise with an integration window of 500 ms, and the
quiet-to-noise correlation value in the steady-state region of
midbrain responses. No significant correlations were found in
either younger or older adults in any of the relationships tested.
DISCUSSION
The results of this study provide support for most, but not
all, of the initial hypotheses. Behavioral data showed that older
adults do have poorer speech understanding in noise than
younger adults, despite their normal, audiometric hearing
thresholds. In midbrain, noise suppresses the response in
younger adults to a greater extent than in older adults, whereas
the fidelity of the reconstruction of speech in cortex is higher
Fig. 4. A: example of the reconstruction of the speech
envelope of the foreground for a representative
younger (top) and older (bottom) adult in noise (6
dB SNR). The waveforms have been standardized for
visualization purposes. B: scatter plots of rvalues for
each participant at each condition tested, plotted in
ascending order.
Fig. 5. Reconstruction accuracy 1 SE of the speech envelope
of the foreground for younger and older adults. A: results in
quiet and in all of the noise conditions tested. The black
horizontal line shows the noise floor. Older adults’ reconstruc-
tion accuracy is significantly higher in quiet (P0.001).
However, as a completing talker is added to the task, the
differences between the 2 age groups are reduced. B: recon-
struction accuracy in quiet and at 6 dB for the 3 integration
windows tested: 500, 350, and 150 ms. Significant differences
across the 3 integration windows were found only in older
adults in both quiet (P0.001) and noise (P0.05). *P
0.05, **P0.01, ***P0.001.
2351AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
in older than in younger adults. Differently from what was
initially hypothesized, no significant associations were found
between behavioral and electrophysiological data and between
midbrain and cortex.
Midbrain (EEG)
Amplitude response. The greater amplitude decrease in noise
in younger adults compared with older adults was unexpected.
However, an RMS age-group interaction was only signifi-
cant in the transition region and may be explained by reduced
audibility in the high frequencies in older adults, given that the
transition region is characterized by the presence of a high-
frequency burst. These results are consistent with an earlier
study that suggested that older adults’ high-frequency hearing
loss might disrupt their ability to encode the high-frequency
components of a syllable (Presacco et al. 2015).
Not surprisingly, in younger adults, the loss of amplitude
between quiet and noise conditions was also larger in the
steady-state region, although no significant RMS age inter-
action was observed. The lack of significant differences ob-
served in the steady-state region is consistent with results
reported by Parthasarathy et al. (2010), where amplitude mod-
ulation following responses differed in younger and older rats
only under specific SNR conditions. Specifically, they ob-
served that at the highest SNR, there were no significant
differences at any of the modulation frequencies tested, but
with a 10-dB loss of SNR, the amplitude modulation following
responses of younger rats tended to decrease substantially,
whereas older rats’ responses showed negligible changes. This
is consistent with results showing significant differences across
noise conditions only in younger adults. Additionally, previous
studies have shown that hearing loss may lead to an exagger-
ated representation of the envelope in midbrain (Anderson et
al. 2013; Henry et al. 2014). Despite having clinically normal,
audiometric thresholds up to 4 kHz, most of our older adults
have a mild, sensorineural hearing loss at higher frequencies (6
and 8 kHz). This mild hearing loss might have potentially
contributed to generating an amplitude response big enough to
reduce the RMS age interaction in the steady-state region.
Interestingly, the frequency domain analysis shows significant
differences across noise conditions only in younger adults in
both the transition (second harmonic in the envelope and 400
Hz in the TFS) and in the steady-state regions (fundamental
and both harmonics in the envelope and 400 Hz in the TFS),
consistent with observations of Parthasarathy et al. (2010).
Robustness of the envelope to noise. The correlation analysis
supported the initial hypothesis that younger adults’ responses
should be more robust to noise than those of older adults.
Younger adults showed significantly higher correlations when
all of the noise conditions were collapsed. The higher robust-
ness of the envelope to noise in younger adults is also con-
firmed by the results of the stimulus-to-response correlation,
which shows that the ability of older adults’ responses to
follow the stimulus is significantly worse than that of younger
adults in noise. These differences between the two age groups
may be due to disruption of periodicity in the encoded speech
envelope, which has been suggested to cause a decrease in
word identification (Pichora-Fuller et al. 2007).
Cortex (MEG)
Reconstruction of the speech envelope. The results of the
reconstruction of the speech envelope show that older adults
had higher correlation values both in quiet and in noise. An
enhanced reconstruction in older adults, both in quiet and in
noise, is consistent with studies showing an exaggerated rep-
resentation of cortical responses in older adults, both with and
without hearing loss. Specifically, Alain et al. (2014), Lister et
al. (2011), and Soros et al. (2009) report abnormally higher
amplitude for the P1 and N1 peaks in normal-hearing older
adults compared with normal-hearing younger adults, in agree-
ment with results from previous studies that showed that aging
might alter inhibitory neural mechanisms in the cortex (de
Villers-Sidani et al. 2010; Hughes et al. 2010; Juarez-Salinas et
al. 2010; Overton and Recanzone 2016). The P1 and N1 peaks
reflect different auditory mechanisms. Specifically, P1, occur-
ring 50 ms after the stimulus onset, originates in Heschl’s
gyrus and can be modulated by stimulus rate, intensity, and
modulation depth (Ross et al. 2000). Conversely, N1, occurring
100 ms after stimulus onset, originates in the Planum tempo-
rale and has been shown to be modulated by attention (Oka-
moto et al. 2011). Therefore, this exaggerated response might
be a reflection of changes in the way that the acoustical stimuli
are processed and as a consequence, in the level of attention
required to process them. Interestingly, Chambers et al. (2016)
recently showed that recovery from profound cochlear dener-
vation in rats leads to cortical spike responses higher than the
baseline recorded before inducing auditory neuropathy; this
finding reinforces the possibility that auditory neuropathy
could play a critical role in the over-representation of an
auditory stimulus. It is also possible that peripheral hearing
loss contributes to problems in the speech-encoding process, as
several studies have shown that this cortical neural enhance-
ment is exacerbated by hearing loss (Alain et al. 2014; Trem-
blay et al. 2003). However, no significant correlation was
found (two-tailed Spearman’s rank correlation) between the
pure-tone average for the frequencies between 2 and 8 kHz and
cortical reconstruction (all, P0.05). The above-mentioned,
exaggerated cortical response, which can take the form of both
better cortical reconstruction and higher peak amplitude (P1
and N1), is perhaps counterintuitive and in disagreement with
the concept of “stronger is better,” as observed in the midbrain.
However, if we assume that a decrease of inhibition leads to
larger neural currents, then we can hypothesize that this neural
enhancement is mainly the result of imbalance between excit-
atory and inhibitory mechanisms.
As higher cognitive processes affect the cortical representa-
tion of the speech signal, the higher reconstruction in older
adults may be related to an inefficient use of cognitive re-
sources and an associated decrease in cortical network connec-
tivity reported in older adults (Peelle et al. 2010). Decreased
cortical network connectivity would result in neighboring cor-
tical areas processing the same stimulus redundantly instead of
cooperatively. Such an overuse of neural resources would lead
to the over-representation observed here. Decreased connec-
tivity may translate to using significantly more energy to
accomplish a task that younger adults can complete with much
less effort. This explanation would be in agreement with
several studies showing that overuse of cognitive resources
2352 AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
leads to poorer performance on a secondary task (Anderson
Gosselin and Gagné 2011; Tun et al. 2009; Ward et al. 2016).
Importantly, the addition of a competing talker caused a
substantial drop of decoding accuracy in older adults, who
required a much longer integration time than younger adults.
This finding is consistent with several psychoacoustic (Fitzgib-
bons and Gordon-Salant 2001; Gordon-Salant et al. 2006) and
electrophysiological (Alain et al. 2012; Lister et al. 2011)
studies, demonstrating that older adults’ responses are affected
to a greater degree than younger adults when temporal param-
eters are varied. Specifically, older adults required longer time
to process specific temporal acoustic cues, such as voice-onset
time, vowel duration, silence duration, and transition duration
(Gordon-Salant et al. 2008). The degradation of the cortical
response from quiet to noise observed in both age groups is
also consistent with previous results showing that the evoked
response seen in quiet is affected by the presence of noise
(Billings et al. 2015). Specifically, cortical response amplitude
decreases as SNR decreases in both younger and older adults,
consistent with the current findings showing a reduction in
reconstruction accuracy within each group. Interestingly, even
with an integration window as narrow as 150 ms, older adults
still show evidence of enhanced reconstruction of the speech
envelope. These results contribute to understanding the signif-
icant group differences observed in the reconstruction of the
background noise. Older adults’ difficulty in understanding
speech in noise may also partially arise from the reduced
ability to suppress unattended stimuli, as suggested by an
over-representation also present in the reconstruction of the
background talker. Note also that even when low, the rvalues
in this study are unlikely to be tied to noise-floor effects:
reconstruction values in both age groups are well above the
noise floor and are consistent with previously published data
(Ding and Simon 2012).
Effect of Hearing Threshold Differences and Cognitive
Decline on Cortical Results
The possibility that the over-representation of the response
of older adults in quiet might be due to significant differences
in the hearing thresholds cannot be ruled out. In fact, even
though the older adults that we tested had clinically normal
hearing, all of their thresholds were significantly higher than
younger adults (P0.05), a typical occurrence for the major-
ity of aging studies. Cochlear synaptopathy has also been
suggested to result from aging and to be a possible contributor
to difficulties in understanding speech-in-noise (Sergeyenko et
al. 2013). Furthermore, several studies have also shown the
existence of age-related cognitive declines (Anderson Gosselin
and Gagné 2011; Pichora-Fuller et al. 1995; Surprenant 2007;
Tun et al. 2009) that may play an important role in compro-
mising attentional resources believed to be critical for proper
representation of the auditory object (Shamma et al. 2011).
Relationships Among Behavioral, Midbrain,
and Cortical Data
The absence of correlations among behavioral and electro-
physiological measurements suggests the possibility that our
behavioral measurements might not completely account for the
presence of temporal processing deficits in the central auditory
system. Caution should be used when interpreting the results
due to important factors. 1) Behavioral data were collected
with four-talker babble as the background noise, whereas
cortical and subcortical data were recorded using a single-
competing talker. A single-competing talker may draw the
subjects’ attention away from the target to a greater extent than
would four-talker babble, given the fact that multiple talkers
generate speech without meaning (little informational mask-
ing) (Larsby et al. 2008; Tun et al. 2002). 2) Several studies
have also shown that the performance in a task varies depend-
ing on different features of the masker (i.e., spectral differ-
ences, SNR level, etc.) (Calandruccio et al. 2010; Larsby et al.
2008). The speech materials used for the electrophysiological
recording were not equated for spectral differences with the
speech material used for the speech-in-noise test.
No significant association was found between midbrain and
cortical results. Even though previous results showed relation-
ships between weak speech encoding in the midbrain (FFR)
and an over-representation of the cortical response (Bidelman
et al. 2014), a more recent animal study (Chambers et al. 2016)
suggests that the absence of auditory brain stem response wave
I does not necessarily lead to an absence of cortical spike
response, suggesting compensatory central gain increases that
could help restore the representation of the auditory object at
the cortical level. This finding may also explain the lack of
association between midbrain and cortex findings. It could also
be argued that the absence of correlation between midbrain and
cortex could be linked to the different stimuli used for the EEG
(speech syllable/da/) and MEG (1 min of speech) task. Addi-
tionally, midbrain and cortical responses were filtered in dif-
ferent frequency ranges to reflect the frequency differences in
the responses emerging from different parts of the auditory
systems. The use of different stimuli was necessitated by the
larger number of trials required to obtain clear responses from
midbrain. Whereas for the cortical analysis, three runs were
sufficient to obtain a clear response above the noise floor, in the
midbrain, a minimum of 2,000 runs was needed, making the
use of stimuli longer than 170 ms not feasible for this long
experiment. Finally, subjects were passively listening to the
auditory stimuli in the EEG experiment, whereas in the MEG,
subjects were actively engaged in listening to the target
speaker.
Concluding Remarks
The results of our studies add compelling evidence to the
notion that age-related temporal processing deficits are a key
factor in explaining speech comprehension problems experi-
enced by older adults, particularly in noisy environments.
Auditory midbrain responses revealed an age-related failure to
encode speech syllables in quiet, which reduces the ability to
cope with the presence of a background talker. Whereas
younger adults adapt to the presence of noise and changes in its
loudness, older adults’ midbrain responses seem to be less
affected by different SNRs, suggesting a failure to encode
properly both the target and the irrelevant speech. This result is
likely not due to the noise-floor effect, as all of the RMS values
calculated in noise were significantly higher than the RMS of
the prestimulus. Our study also reveals an over-representation
of the cortical response, consistent with previous studies (Alain
et al. 2014; Lister et al. 2011; Soros et al. 2009); this neural
enhancement is reduced with the addition of a competing
2353AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
talker, suggesting that larger cortical responses are not bene-
ficial and might, in fact, represent a failure of the brain to
process speech properly. Critically, we were unable to find any
significant correlations between midbrain and cortex measures.
We believe this result brings additional support to recent
findings that suggest that cortical plasticity may partially re-
store temporal processing deficits at lower levels of the audi-
tory system (Chambers et al. 2016), although we cannot ex-
clude the possibility that a lack of correlation may reflect
differences in the stimuli used to elicit the midbrain and
cortical responses.
This apparent lack of relationship between midbrain and
cortex further highlights the relevance of this study, which is
the importance of investigating simultaneously different areas
of the auditory system to understand better the mechanisms
underlying age-related degradation of speech representation.
ACKNOWLEDGMENTS
The authors are grateful to Natalia Lapinskaya for excellent technical
support.
GRANTS
Funding for this study was provided by the University of Maryland College
Park (UMCP) Department of Hearing and Speech Sciences, UMCP AD-
VANCE Program for Inclusive Excellence (NSF HRD1008117), and National
Institute on Deafness and Other Communication Disorders (Grants
R01DC008342, R01DC014085, and T32DC-00046).
DISCLOSURES
The authors declare no competing financial interests.
AUTHOR CONTRIBUTIONS
A.P., J.Z.S., and S.A. conception and design of research; A.P. performed
experiments; A.P. analyzed data; A.P., J.Z.S., and S.A. interpreted results of
experiments; A.P. and S.A. prepared figures; A.P., J.Z.S., and S.A. drafted
manuscript; A.P., J.Z.S., and S.A. edited and revised manuscript; A.P., J.Z.S.,
and S.A. approved final version of manuscript.
REFERENCES
Aiken SJ, Picton TW. Envelope and spectral frequency-following responses
to vowel sounds. Hear Res 245: 35– 47, 2008.
Alain C, McDonald K, Van Roon P. Effects of age and background noise on
processing a mistuned harmonic in an otherwise periodic complex sound.
Hear Res 283: 126 –135, 2012.
Alain C, Roye A, Salloum C. Effects of age-related hearing loss and
background noise on neuromagnetic activity from auditory cortex. Front
Syst Neurosci 8: 8, 2014.
Anderson S, Parbery-Clark A, White-Schwoch T, Drehobl S, Kraus N.
Effects of hearing loss on the subcortical representation of speech cues. J
Acoust Soc Am 133: 3030 –3038, 2013.
Anderson S, Parbery-Clark A, White-Schwoch T, Kraus N. Aging affects
neural precision of speech encoding. J Neurosci 32: 14156 –14164, 2012.
Anderson Gosselin P, Gagné JP. Older adults expend more listening effort
than young adults recognizing speech in noise. J Speech Lang Hear Res 54:
944 –958, 2011.
Benjamini Y, Hochberg Y. Controlling the false discovery rate: a practical
and powerful approach to multiple testing. J R Statist Soc B 57: 289 –300,
1995.
Bidelman GM, Krishnan A. Effects of reverberation on brainstem represen-
tation of speech in musicians and non-musicians. Brain Res 1355: 112–125,
2010.
Bidelman GM, Villafuerte JW, Moreno S, Alain C. Age-related changes in
the subcortical-cortical encoding and categorical perception of speech.
Neurobiol Aging 35: 2526 –2540, 2014.
Billings CJ, Penman TM, McMillan GP, Ellis EM. Electrophysiology and
perception of speech in noise in older listeners: effects of hearing impair-
ment and age. Ear Hear 36: 710 –722, 2015.
Calandruccio L, Dhar S, Bradlow AR. Speech-on-speech masking with
variable access to the linguistic content of the masker speech. J Acoust Soc
Am 128: 860 – 869, 2010.
Campbell T, Kerlin JR, Bishop CW, Miller LM. Methods to eliminate
stimulus transduction artifact from insert earphones during electroencepha-
lography. Ear Hear 33: 144 –150, 2012.
Carabellese C, Appollonio I, Rozzini R, Bianchetti A, Frisoni GB, Frattola
L, Trabucchi M. Sensory impairment and quality of life in a community
elderly population. J Am Geriatr Soc 41: 401– 407, 1993.
Chambers AR, Resnik J, Yuan Y, Whitton JP, Edge AS, Liberman MC,
Polley DB. Central gain restores auditory processing following near-com-
plete cochlear denervation. Neuron 89: 867– 879, 2016.
Chandrasekaran B, Kraus N. The scalp-recorded brainstem response to
speech: neural origins and plasticity. Psychophysiology 47: 236 –246, 2010.
Clinard CG, Tremblay KL. Aging degrades the neural encoding of simple
and complex sounds in the human brainstem. J Am Acad Audiol 24:
590 –599; quiz 643–594, 2013.
David SV, Mesgarani N, Shamma SA. Estimating sparse spectro-temporal
receptive fields with natural stimuli. Network 18: 191–212, 2007.
de Cheveigné A, Simon JZ. Denoising based on spatial filtering. J Neurosci
Methods 171: 331–339, 2008.
de Cheveigné A, Simon JZ. Denoising based on time-shift PCA. J Neurosci
Methods 165: 297–305, 2007.
de Villers-Sidani E, Alzghoul L, Zhou X, Simpson KL, Lin RC, Merzenich
MM. Recovery of functional and structural age-related changes in the rat
primary auditory cortex with operant training. Proc Natl Acad Sci USA 107:
13900 –13905, 2010.
Delorme A, Makeig S. EEGLAB: an open source toolbox for analysis of
single-trial EEG dynamics including independent component analysis. J
Neurosci Methods 134: 9 –21, 2004.
Ding N, Chatterjee M, Simon JZ. Robust cortical entrainment to the speech
envelope relies on the spectro-temporal fine structure. Neuroimage 88:
41– 46, 2014.
Ding N, Simon JZ. Adaptive temporal encoding leads to a background-
insensitive cortical representation of speech. J Neurosci 33: 5728 –5735,
2013.
Ding N, Simon JZ. Emergence of neural encoding of auditory objects while
listening to competing speakers. Proc Natl Acad Sci USA 109: 11854 –
11859, 2012.
Fitzgibbons PJ, Gordon-Salant S. Aging and temporal discrimination in
auditory sequences. J Acoust Soc Am 109: 2955–2963, 2001.
Galbraith GC, Threadgill MR, Hemsley J, Salour K, Songdej N, Ton J,
Cheung L. Putative measure of peripheral and brainstem frequency-follow-
ing in humans. Neurosci Lett 292: 123–127, 2000.
Gates GA, Cobb JL, Linn RT, Rees T, Wolf PA, D’Agostino RB. Central
auditory dysfunction, cognitive dysfunction, and dementia in older people.
Arch Otolaryngol Head Neck Surg 122: 161–167, 1996.
Gordon-Salant S, Yeni-Komshian G, Fitzgibbons P. The role of temporal
cues in word identification by younger and older adults: effects of sentence
context. J Acoust Soc Am 124: 3249 –3260, 2008.
Gordon-Salant S, Yeni-Komshian GH, Fitzgibbons PJ, Barrett J. Age-
related differences in identification and discrimination of temporal cues in
speech segments. J Acoust Soc Am 119: 2455–2466, 2006.
Gorga M, Abbas P, Worthington D. Stimulus calibration in ABR measure-
ments. In: The Auditory Brainstem Response, edited by Jacobsen J. San
Diego: College Hill, 1985, p. 49 62.
Greenberg S, Takayuki A. What are the essential cues for understanding
spoken language? IEICE Trans Inf Syst 87: 1059 –1070, 2004.
Henry KS, Kale S, Heinz MG. Noise-induced hearing loss increases the
temporal precision of complex envelope coding by auditory-nerve fibers.
Front Syst Neurosci 8: 20, 2014.
Herbst KG, Humphrey C. Hearing impairment and mental state in the elderly
living at home. Br Med J 281: 903–905, 1980.
Hughes LF, Turner JG, Parrish JL, Caspary DM. Processing of broadband
stimuli across A1 layers in young and aged rats. Hear Res 264: 79 – 85,
2010.
Humes LE, Christopherson L. Speech identification difficulties of hearing-
impaired elderly persons: the contributions of auditory processing deficits. J
Speech Hear Res 34: 686 – 693, 1991.
2354 AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
Humes LE, Roberts L. Speech-recognition difficulties of the hearing-im-
paired elderly: the contributions of audibility. J Speech Hear Res 33:
726 –735, 1990.
Juarez-Salinas DL, Engle JR, Navarro XO, Recanzone GH. Hierarchical
and serial processing in the spatial auditory cortical pathway is degraded by
natural aging. J Neurosci 30: 14795–14804, 2010.
Kay DW, Beamish P, Roth M. Old age mental disorders in Newcastle upon
Tyne. I. A study of prevalence. Br J Psychiatry 110: 146 –158, 1964.
Killion MC, Niquette PA, Gudmundsen GI, Revit LJ, Banerjee S. Devel-
opment of a quick speech-in-noise test for measuring signal-to-noise ratio
loss in normal-hearing and hearing-impaired listeners. J Acoust Soc Am 116:
2395–2405, 2004.
King C, Warrier CM, Hayes E, Kraus N. Deficits in auditory brainstem
pathway encoding of speech sounds in children with learning problems.
Neurosci Lett 319: 111–115, 2002.
Klatt DH. Software for a cascade/parallel formant synthesizer. J Acoust Soc
Am 67: 971–995, 1980.
Laforge RG, Spector WD, Sternberg J. The relationship of vision and
hearing impairment to one-year mortality and functional decline. J Aging
Health 4: 126 –148, 1992.
Larsby B, Hallgren M, Lyxell B. The interference of different background
noises on speech processing in elderly hearing impaired subjects. Int J
Audiol 47, Suppl 2: S83–S90, 2008.
Lin FR, Yaffe K, Xia J, Xue QL, Harris TB, Purchase-Helzner E,
Satterfield S, Ayonayon HN, Ferrucci L, Simonsick EM. Hearing loss
and cognitive decline in older adults. JAMA Intern Med 173: 293–299, 2013.
Lister JJ, Maxfield ND, Pitt GJ, Gonzalez VB. Auditory evoked response to
gaps in noise: older adults. Int J Audiol 50: 211–225, 2011.
Mamo SK, Grose JH, Buss E. Speech-evoked ABR: effects of age and
simulated neural temporal jitter. Hear Res 333: 201–209, 2016.
Nasreddine ZS, Phillips NA, Bedirian V, Charbonneau S, Whitehead V,
Collin I, Cummings JL, Chertkow H. The Montreal Cognitive Assess-
ment, MoCA: a brief screening tool for mild cognitive impairment. JAm
Geriatr Soc 53: 695– 699, 2005.
Okamoto H, Stracke H, Bermudez P, Pantev C. Sound processing hierarchy
within human auditory cortex. J Cogn Neurosci 23: 1855–1863, 2011.
Overton JA, Recanzone GH. Effects of aging on the response of single
neurons to amplitude modulated noise in primary auditory cortex of Rhesus
macaque. J Neurophysiol 115: 2911–2923, 2016.
Parbery-Clark A, Anderson S, Hittner E, Kraus N. Musical experience
offsets age-related delays in neural timing. Neurobiol Aging 33:
1483.e1481–1483.e1484, 2012.
Parbery-Clark A, Marmel F, Bair J, Kraus N. What subcortical-cortical
relationships tell us about processing speech in noise. Eur J Neurosci 33:
549 –557, 2011.
Parthasarathy A, Bartlett EL. Age-related auditory deficits in temporal
processing in F-344 rats. Neuroscience 192: 619 – 630, 2011.
Parthasarathy A, Cunningham PA, Bartlett EL. Age-related differences in
auditory processing as assessed by amplitude-modulation following re-
sponses in quiet and in noise. Front Aging Neurosci 2: 152, 2010.
Peelle JE, Troiani V, Wingfield A, Grossman M. Neural processing during
older adults’ comprehension of spoken sentences: age differences in re-
source allocation and connectivity. Cereb Cortex 20: 773–782, 2010.
Pichora-Fuller MK, Schneider BA, Daneman M. How young and old adults
listen to and remember speech in noise. J Acoust Soc Am 97: 593– 608, 1995.
Pichora-Fuller MK, Schneider BA, Macdonald E, Pass HE, Brown S.
Temporal jitter disrupts speech intelligibility: a simulation of auditory aging.
Hear Res 223: 114 –121, 2007.
Presacco A, Jenkins K, Lieberman R, Anderson S. Effects of aging on the
encoding of dynamic and static components of speech. Ear Hear 36:
e352– e363, 2015.
Ross B, Borgmann C, Draganova R, Roberts LE, Pantev C. A high-
precision magnetoencephalographic study of human auditory steady-state
responses to amplitude-modulated tones. J Acoust Soc Am 108: 679 – 691,
2000.
Ross B, Schneider B, Snyder JS, Alain C. Biological markers of auditory gap
detection in young, middle-aged, and older adults. PLoS One 5: e10101,
2010.
Särelä J, Valpola H. Denoising source separation. J Mach Learn Res 6:
233–272, 2005.
Sergeyenko Y, Lall K, Liberman MC, Kujawa SG. Age-related cochlear
synaptopathy: an early-onset contributor to auditory functional decline. J
Neurosci 33: 13686 –13694, 2013.
Shamma SA, Elhilali M, Micheyl C. Temporal coherence and attention in
auditory scene analysis. Trends Neurosci 34: 114 –123, 2011.
Smith JC, Marsh JT, Brown WS. Far-field recorded frequency-following
responses: evidence for the locus of brainstem sources. Electroencephalogr
Clin Neurophysiol 39: 465– 472, 1975.
Soros P, Teismann IK, Manemann E, Lutkenhoner B. Auditory temporal
processing in healthy aging: a magnetoencephalographic study. BMC Neu-
rosci 10: 34, 2009.
Surprenant AM. Effects of noise on identification and serial recall of
nonsense syllables in older and younger adults. Neuropsychol Dev Cogn B
Aging Neuropsychol Cogn 14: 126 –143, 2007.
Tremblay KL, Piskosz M, Souza P. Effects of age and age-related hearing
loss on the neural representation of speech cues. Clin Neurophysiol 114:
1332–1343, 2003.
Tun PA, McCoy S, Wingfield A. Aging, hearing acuity, and the attentional
costs of effortful listening. Psychol Aging 24: 761–766, 2009.
Tun PA, O’Kane G, Wingfield A. Distraction by competing speech in young
and older adult listeners. Psychol Aging 17: 453– 467, 2002.
Uhlmann RF, Larson EB, Rees TS, Koepsell TD, Duckert LG. Relationship
of hearing impairment to dementia and cognitive dysfunction in older adults.
JAMA 261: 1916 –1919, 1989.
Ward CM, Rogers CS, Van Engen KJ, Peelle JE. Effects of age, acoustic
challenge, and verbal working memory on recall of narrative speech. Exp
Aging Res 42: 126 –144, 2016.
Zhu J, Garcia E. The Wechsler Abbreviated Scale of Intelligence (WASI).
New York: Psychological Corp., 1999.
2355AGING EFFECTS OF NEURAL PROCESSING OF SPEECH IN NOISE
J Neurophysiol doi:10.1152/jn.00372.2016 www.jn.org
... www.nature.com/scientificreports/ older adults-typically alters speech processing 8,9,27,28,30 . This is surprising, especially considering the common pool model of cognitive processing resources 31 , which suggests that cognitive decline and hearing loss may compete for limited cognitive and perceptual resources, thereby reducing neural encoding of speech features in affected individuals. ...
... This is noteworthy since recognizing words in continuous speech is a complex task requiring substantial cognitive resources, which can become even more challenging in the presence of hearing loss 38,39 . It is intriguing that we did not see these interactions in the higher-order linguistic models, nor did we observe a significant main effect of PTA on neural encoding accuracy, which is commonly reported in other studies 27,28,30 . We also found that while investigating the word-level segmentation mTRF model at the response signal level, hearing ability did not affect the response signal power, contrasting with the encoding accuracy results. ...
... The neural responses appeared relatively homogeneous across participants, regardless of their cognitive status, suggesting that neural encoding of speech in natural continuous settings may not effectively indicate early cognitive decline in older adults. Regarding hearing ability, our findings diverged from existing studies, where hearing loss often correlated with enhanced cortical speech tracking measures in older adults 27,28,30 . The relatively good hearing ability in our sample, with only seven individuals exhibiting a PTA exceeding 34 dB HL , which can be considered as moderate hearing loss 43 , might explain the lack of a significant association between hearing ability and neural encoding measures. ...
Article
Full-text available
The multivariate temporal response function (mTRF) is an effective tool for investigating the neural encoding of acoustic and complex linguistic features in natural continuous speech. In this study, we investigated how neural representations of speech features derived from natural stimuli are related to early signs of cognitive decline in older adults, taking into account the effects of hearing. Participants without ( n=25n = 25 n = 25 ) and with ( n=19n = 19 n = 19 ) early signs of cognitive decline listened to an audiobook while their electroencephalography responses were recorded. Using the mTRF framework, we modeled the relationship between speech input and neural response via different acoustic, segmented and linguistic encoding models and examined the response functions in terms of encoding accuracy, signal power, peak amplitudes and latencies. Our results showed no significant effect of cognitive decline or hearing ability on the neural encoding of acoustic and linguistic speech features. However, we found a significant interaction between hearing ability and the word-level segmentation model, suggesting that hearing impairment specifically affects encoding accuracy for this model, while other features were not affected by hearing ability. These results suggest that while speech processing markers remain unaffected by cognitive decline and hearing loss per se, neural encoding of word-level segmented speech features in older adults is affected by hearing loss but not by cognitive decline. This study emphasises the effectiveness of mTRF analysis in studying the neural encoding of speech and argues for an extension of research to investigate its clinical impact on hearing loss and cognition.
... Consistent with these findings, studies have demonstrated a reduction in brainstem FFRs elicited by low-to-mid frequency (i.e., 100-1000 Hz) periodic stimuli among older participants (Clinard and Cotter, 2015;T. A. Johnson and Brown, 2005;Vander Werff and Burns, 2011), including those with clinically-normal or near-normal audiometric thresholds (Anderson et al., 2012;Mamo et al., 2016;Märcher-Rørsted et al., 2022;Presacco et al., 2016). However, age-related deficits in central auditory processing, such as neural desynchronization at the auditory brainstem level (Pichora-Fuller et al., 2007;Robert Frisina and Frisina, 1997;Schneider et al., 1998), may also contribute to the decline in FFRs associated with aging. ...
... Steady-state evoked potentials have also been proposed as a potential marker of cochlear neural loss (Bharadwaj et al., 2015;Encina-Llamas et al., 2019;Keshishzadeh et al., 2020;Märcher-Rørsted et al., 2022;Parthasarathy and Kujawa, 2018;Prendergast et al., 2019;Shaheen et al., 2015). In a recent study, we showed a reduced magnitude of the brainstem FFR in response to tone stimulation in older humans assumed to present a larger degree of AN degeneration despite near-normal clinical thresholds, (Märcher-Rørsted et al., 2022), consistent with other previous work (Clinard and Cotter, 2015;Mamo et al., 2016;Presacco et al., 2016). Based on simulated AN responses using a computational AN model, Märcher-Rørsted et al. (2022) suggested that the reduced FFR in older listeners could be driven by age-related AN degeneration, despite the fact that the FFR is generated in the auditory brainstem (Henry, 1995;Snyder and Schreiner, 1984). ...
... These mechanisms have been described as having a bidirectional effect at different rates: showing enhanced synchronization in responses Table 2 Summary statistics (PLV, SNR, and spectral amplitudes) of the responses to the 10-ms pure tones at 516, 1032, and 3096 Hz with the mastoid-to-vertex and TM-tomastoid electrode configurations. to stimuli at low presentation rates (i.e., clicks at 12 Hz, Goossens et al., 2016;Presacco et al., 2016;Purcell et al., 2004), but decreased synchronization to stimulation with higher rates (i.e., 500 and 1000 Hz, Anderson et al., 2012;Purcell et al., 2004 see also review Herrmann and Butler, 2021). Our results are consistent with rate-dependent hyper-synchronous activity. ...
... To clarify the mechanisms that underlie age-related SPiN decline, the relationship between SPiN performance in young and older adults and the structure/function of these regions has been investigated extensively (Tremblay, Brisson, & Deschamps, 2021;Rogers et al., 2020;Presacco, Simon, & Anderson, 2016, 2019Tremblay, Perron, et al., 2019;Manan, Yusoff, Franz, & Mukari, 2017;Du, Buchsbaum, Grady, & Alain, 2016;Eckert et al., 2016;Peelle & Wingfield, 2016;Bilodeau-Mercure, Lortie, Sato, Guitton, & Tremblay, 2015;Vaden, Kuchinsky, Ahlstrom, Dubno, & Eckert, 2015;Erb & Obleser, 2013;Tremblay, Dick, & Small, 2013;Sheppard, Wang, & Wong, 2011;Wong, Ettlinger, Sheppard, Gunasekera, & Dhar, 2010;Wong et al., 2009;Hwang, Li, Wu, Chen, & Liu, 2007). Neuroimaging studies have shown positive correlations between activity within the dorsal stream and SPiN performance in older and younger adults (Fitzhugh, Schaefer, Baxter, & Rogalsky, 2021;Bilodeau-Mercure et al., 2015;Eckert et al., 2008). ...
... Furthermore, studies have reported a decreased gray matter and reduced BOLD signal in the STC for older adults performing SPiN tasks Peelle & Wingfield, 2016;Sheppard et al., 2011;Harris, Dubno, Keren, Ahlstrom, & Eckert, 2009;Hwang et al., 2007). These studies, combined with studies showing evidence of degraded speech representations in the central auditory system (e.g., Presacco et al., 2016;Presacco, Jenkins, Lieberman, & Anderson, 2015;Tremblay, Piskosz, & Souza, 2003), suggest that agerelated speech perception decline is associated with a reduced efficiency in early auditory processing mechanisms (Peelle, 2018;Peelle & Wingfield, 2016). However, most of the studies have not tested the relationship between SPiN performance and the structural and functional integrity of the STC in older adults; thus, causal conclusions cannot be drawn. ...
Article
Full-text available
Healthy aging is associated with reduced speech perception in noise (SPiN) abilities. The etiology of these difficulties remains elusive, which prevents the development of new strategies to optimize the speech processing network and reduce these difficulties. The objective of this study was to determine if sublexical SPiN performance can be enhanced by applying TMS to three regions involved in processing speech: the left posterior temporal sulcus, the left superior temporal gyrus, and the left ventral premotor cortex. The second objective was to assess the impact of several factors (age, baseline performance, target, brain structure, and activity) on post-TMS SPiN improvement. The results revealed that participants with lower baseline performance were more likely to improve. Moreover, in older adults, cortical thickness within the target areas was negatively associated with performance improvement, whereas this association was null in younger individuals. No differences between the targets were found. This study suggests that TMS can modulate sublexical SPiN performance, but that the strength and direction of the effects depend on a complex combination of contextual and individual factors.
... Herrmann et al. (2019) also observed enhanced 4 Hz EEG phase-locking in older adults compared to younger adults during 4 Hz AM auditory stimulation. Evoked cortical responses to brief sound stimuli (Alain et al., 2022;Henry et al., 2017;Herrmann et al., 2019Herrmann et al., , 2016 and M/EEG measures of synchronization to low-frequency speech envelope features (Decruy et al., 2018;Presacco et al., 2019Presacco et al., , 2016aPresacco et al., , 2016b have similarly been reported to increase with the age of participants. However, MEG and EEG cannot readily differentiate between AM synchronization at slow AM rates and long-latency auditory evoked potentials (Edwards and Chang, 2013). ...
... Several recent human M/EEG studies have reported increased cortical synchronization to slow AM with pronounced 4 Hz AM in older listeners compared with younger listeners (Decruy et al., 2018;Goossens et al., 2019Goossens et al., , 2016Herrmann et al., 2019;Presacco et al., 2016aPresacco et al., , 2016b. Using fully-modulated sinusoidal AM stimuli, Goossens et al. (2016) found that auditory steady-state responses (ASSRs) to ~4 Hz AM noise stimuli were increased in older adults compared with younger adults, all with audiometric thresholds ≤ 25 dB hearing level (HL) at octave frequencies from 125 Hz up to 4 kHz. ...
Article
Full-text available
Age-related alterations in the auditory system have been suggested to affect the processing of temporal envelope amplitude modulations (AM) at different levels of the auditory hierarchy, yet few studies have used functional magnetic resonance imaging (fMRI) to study this noninvasively in humans with high spatial resolution. In this study, we utilized sparse-sampling fMRI at 3 Tesla (3T) to investigate regional blood oxygenation level-dependent (BOLD) responses to AM noise stimuli in 65 individuals ranging in age from 19 to 77 years. We contrasted BOLD responses to AM noise stimuli modulated at 4 Hz or 80 Hz with responses to unmodulated stimuli. This allowed us to derive functional measures of regional neural sensitivity to the imposed AM. Compared to unmodulated noise, slowly varying 4 Hz AM noise stimuli elicited significantly greater BOLD responses in the left and right auditory cortex along the Heschl’s gyrus (HG). BOLD responses to the 80 Hz AM stimuli were significantly greater than responses to unmodulated stimuli in putatively primary auditory cortical regions in the lateral HG. BOLD responses to 4 Hz AM stimuli were significantly greater in magnitude than responses to 80 Hz AM stimuli in auditory cortical regions. We find no discernible effects of age on the functional recruitment of the auditory cortex by AM stimuli. While the results affirm the involvement of the auditory cortex in processing temporal envelope rate information, they provide no support for age-related effects on these measures. We discuss potential caveats in assessing age-related changes in responses to AM stimuli in the auditory pathway.
... Consequently, exploring the neural underpinnings of multi-talker speech comprehension in hearing-impaired listeners could yield valuable insights into the challenges faced by both normal-hearing and hearing-impaired individuals in this scenario. Studies have reported abnormally enhanced responses to fluctuations in acoustic envelope in the central auditory system of older listeners (e.g., Goossens et al., 2016;Parthasarathy et al., 2019;Presacco et al., 2016) and listeners with peripheral hearing loss (Goossens et al., 2018;Millman et al., 2017). As older listeners also suffer from suppression of task-irrelevant sensory information due to reduced cortical inhibitory control functions (Gazzaley et al., 2005(Gazzaley et al., , 2008, it is possible that impaired speech comprehension in a cocktail party situation arises from an attentional deficit linked to aging (Du et al., 2016;Presacco et al., 2016). ...
... Studies have reported abnormally enhanced responses to fluctuations in acoustic envelope in the central auditory system of older listeners (e.g., Goossens et al., 2016;Parthasarathy et al., 2019;Presacco et al., 2016) and listeners with peripheral hearing loss (Goossens et al., 2018;Millman et al., 2017). As older listeners also suffer from suppression of task-irrelevant sensory information due to reduced cortical inhibitory control functions (Gazzaley et al., 2005(Gazzaley et al., , 2008, it is possible that impaired speech comprehension in a cocktail party situation arises from an attentional deficit linked to aging (Du et al., 2016;Presacco et al., 2016). However, younger listeners with EHF hearing loss have also reported difficulty understanding speech in a multi-talker environment (Motlagh Zadeh et al., 2019). ...
Preprint
Full-text available
Comprehending speech requires deciphering a range of linguistic representations, from phonemes to narratives. Prior research suggests that in single-talker scenarios, the neural encoding of linguistic units follows a hierarchy of increasing temporal receptive windows. Shorter temporal units like phonemes and syllables are encoded by lower-level sensory brain regions, whereas longer units such as sentences and paragraphs are processed by higher-level perceptual and cognitive areas. However, the brain's representation of these linguistic units under challenging listening conditions, such as a cocktail party situation, remains unclear. In this study, we recorded electroencephalogram (EEG) responses from both normal-hearing and hearing-impaired participants as they listened to individual and dual speakers narrating different parts of a story. The inclusion of hearing-impaired listeners allowed us to examine how hierarchically organized linguistic units in competing speech streams affect comprehension abilities. We leveraged a hierarchical language model to extract linguistic information at multiple levels--phoneme, syllable, word, phrase, and sentence--and aligned these model activations with the EEG data. Our findings showed distinct neural responses to dual-speaker speech between the two groups. Specifically, compared to normal-hearing listeners, hearing-impaired listeners exhibited poorer model fits at the acoustic, phoneme, and syllable levels as well as the sentence levels, but not at the word and phrase levels. These results suggest that hearing-impaired listeners experience disruptions at both shorter and longer temporal scales, while their processing at medium temporal scales remains unaffected.
... Most studies focused on neural entrainment to amplitude modulated sounds show that older adults entrain more strongly and in a more stereotyped (less flexible) way to metronomic stimuli like those we used here (Goossens et al., 2016;Herrmann et al., 2019;Purcell et al., 2004). A similar pattern was observed for entrainment to the amplitude envelope of speech (Decruy et al., 2020;Presacco et al., 2016). A mixed pattern of results were reported for frequency modulated sounds; however, the existing data suggest that these differences might depend on parameters such as modulation rate and depth (Henry et al., 2017;Boettcher et al., 2002), which we will not further address here. ...
Article
Full-text available
Dynamic attending theory proposes that the ability to track temporal cues in the auditory environment is governed by entrainment, the synchronization between internal oscillations and regularities in external auditory signals. Here, we focused on two key properties of internal oscillators: their preferred rate, the default rate in the absence of any input; and their flexibility, how they adapt to changes in rhythmic context. We developed methods to estimate oscillator properties (Experiment 1) and compared the estimates across tasks and individuals (Experiment 2). Preferred rates, estimated as the stimulus rates with peak performance, showed a harmonic relationship across measurements and were correlated with individuals’ spontaneous motor tempo . Estimates from motor tasks were slower than those from the perceptual task, and the degree of slowing was consistent for each individual. Task performance decreased with trial-to-trial changes in stimulus rate, and responses on individual trials were biased toward the preceding trial’s stimulus properties. Flexibility, quantified as an individual’s ability to adapt to faster-than-previous rates, decreased with age. These findings show domain-specific rate preferences for the assumed oscillatory system underlying rhythm perception and production, and that this system loses its ability to flexibly adapt to changes in the external rhythmic context during aging.
... Past studies have observed enhanced cortical speech tracking in ageing compared to young adults. This suggests a compensatory role of increased speech-brain coupling to counteract the deleterious effect of age or of hearing loss on speech comprehension (Schmitt et al., 2022;Gillis et al., 2022a;Presacco et al., 2019;Presacco et al., 2016;Decruy et al., 2019;Decruy et al., 2020). As a corollary of this relationship, typically observed cross-sectionally, one might expect an individual's neural filtering strength to be connected not only to present but also to future trajectories of listening behaviour. ...
Article
Full-text available
Preserved communication abilities promote healthy ageing. To this end, the age-typical loss of sensory acuity might in part be compensated for by an individual’s preserved attentional neural filtering. Is such a compensatory brain–behaviour link longitudinally stable? Can it predict individual change in listening behaviour? We here show that individual listening behaviour and neural filtering ability follow largely independent developmental trajectories modelling electroencephalographic and behavioural data of N = 105 ageing individuals (39–82 y). First, despite the expected decline in hearing-threshold-derived sensory acuity, listening-task performance proved stable over 2 y. Second, neural filtering and behaviour were correlated only within each separate measurement timepoint (T1, T2). Longitudinally, however, our results raise caution on attention-guided neural filtering metrics as predictors of individual trajectories in listening behaviour: neither neural filtering at T1 nor its 2-year change could predict individual 2-year behavioural change, under a combination of modelling strategies.
... The dataset description contains an overview of the data collection protocol and file naming convention. More details and the analysis of the data are found in the publication [98]. ...
Article
Full-text available
The rapidly evolving landscape of artificial intelligence (AI) and machine learning has placed data at the forefront of healthcare innovation. Electroencephalography (EEG) has gained significant attention for its potential to revolutionize healthcare applications. However, the effective utilization of EEG data in advancing medical diagnoses and treatment hinges on the availability and quality of relevant datasets. In this context, we conducted a scoping review to explore the wealth of EEG datasets designed for healthcare applications. This review serves as a critical exploration of the current landscape, aiming to identify datasets related to healthcare conditions while assessing their reusability. Our findings highlight both the opportunities provided by the wealth of open access EEG datasets available as well as any limitations associated with their use. As AI increasingly relies on high-quality, well labelled data, barriers impeding the sharing and utilization of EEG data for healthcare (such as lack of comprehensive documentation or adherence to FAIR principles) must be addressed so as to leverage the potential of advanced deep learning models to unlock new possibilities for diagnosis and analysis of a wide array of medical conditions.
Article
The assessment of mental effort is increasingly relevant in neurocognitive and life span domains. Pupillometry, the measure of the pupil size, is often used to assess effort but has disadvantages. Analysis of eye movements may provide an alternative, but research has been limited to easy and difficult task demands in younger adults. An effort measure must be sensitive to the whole effort profile, including “giving up” effort investment, and capture effort in different age groups. The current study comprised three experiments in which younger (n = 66) and older (n = 44) adults listened to speech masked by background babble at different signal-to-noise ratios associated with easy, difficult, and impossible speech comprehension. We expected individuals to invest little effort for easy and impossible speech (giving up) but to exert effort for difficult speech. Indeed, pupil size was largest for difficult but lower for easy and impossible speech. In contrast, gaze dispersion decreased with increasing speech masking in both age groups. Critically, gaze dispersion during difficult speech returned to levels similar to easy speech after sentence offset, when acoustic stimulation was similar across conditions, whereas gaze dispersion during impossible speech continued to be reduced. These findings show that a reduction in eye movements is not a byproduct of acoustic factors, but instead suggest that neurocognitive processes, different from arousal-related systems regulating the pupil size, drive reduced eye movements during high task demands. The current data thus show that effort in one sensory domain (audition) differentially impacts distinct functional properties in another sensory domain (vision).
Article
Full-text available
Background/study context: A common goal during speech comprehension is to remember what we have heard. Encoding speech into long-term memory frequently requires processes such as verbal working memory that may also be involved in processing degraded speech. Here the authors tested whether young and older adult listeners' memory for short stories was worse when the stories were acoustically degraded, or whether the additional contextual support provided by a narrative would protect against these effects. Methods: The authors tested 30 young adults (aged 18-28 years) and 30 older adults (aged 65-79 years) with good self-reported hearing. Participants heard short stories that were presented as normal (unprocessed) speech or acoustically degraded using a noise vocoding algorithm with 24 or 16 channels. The degraded stories were still fully intelligible. Following each story, participants were asked to repeat the story in as much detail as possible. Recall was scored using a modified idea unit scoring approach, which included separately scoring hierarchical levels of narrative detail. Results: Memory for acoustically degraded stories was significantly worse than for normal stories at some levels of narrative detail. Older adults' memory for the stories was significantly worse overall, but there was no interaction between age and acoustic clarity or level of narrative detail. Verbal working memory (assessed by reading span) significantly correlated with recall accuracy for both young and older adults, whereas hearing ability (better ear pure tone average) did not. Conclusion: The present findings are consistent with a framework in which the additional cognitive demands caused by a degraded acoustic signal use resources that would otherwise be available for memory encoding for both young and older adults. Verbal working memory is a likely candidate for supporting both of these processes.
Article
Full-text available
Classical models of speech recognition assume that a detailed, short-term analysis of the acoustic signal is essential for accurately decoding the speech signal and that this decoding process is rooted in the phonetic segment. This paper presents an alternative view, one in which the time scales required to accurately describe and model spoken language are both shorter and longer than the phonetic segment, and are inherently wedded to the syllable. The syllable reflects a singular property of the acoustic signal - the modulation spectrum - which provides a principled, quantitative framework to describe the process by which the listener proceeds from sound to meaning. The ability to understand spoken language (i.e., intelligibility) vitally depends on the integrity of the modulation spectrum within the core range of the syllable (3-10 Hz) and reflects the variation in syllable emphasis associated with the concept of prosodic prominence ("accent"). A model of spoken language is described in which the prosodic properties of the speech signal are embedded in the temporal dynamics associated with the syllable, a unit serving as the organizational interface among the various tiers of linguistic representation.
Article
Sensory organ damage induces a host of cellular and physiological changes in the periphery and the brain. Here, we show that some aspects of auditory processing recover after profound cochlear denervation due to a progressive, compensatory plasticity at higher stages of the central auditory pathway. Lesioning >95% of cochlear nerve afferent synapses, while sparing hair cells, in adult mice virtually eliminated the auditory brainstem response and acoustic startle reflex, yet tone detection behavior was nearly normal. As sound-evoked responses from the auditory nerve grew progressively weaker following denervation, sound-evoked activity in the cortex—and, to a lesser extent, the midbrain—rebounded or surpassed control levels. Increased central gain supported the recovery of rudimentary sound features encoded by firing rate, but not features encoded by precise spike timing such as modulated noise or speech. These findings underscore the importance of central plasticity in the perceptual sequelae of cochlear hearing impairment. Video Abstract https://www.cell.com/cms/asset/982b93be-cdd5-47f1-8763-09fe12abaf94/mmc4.mp4 Loading ... (mp4, 53.06 MB) Download video
Article
Objectives: Speech perception in background noise is difficult for many individuals, and there is considerable performance variability across listeners. The combination of physiological and behavioral measures may help to understand sources of this variability for individuals and groups and prove useful clinically with hard-to-test populations. The purpose of this study was threefold: (1) determine the effect of signal-to-noise ratio (SNR) and signal level on cortical auditory evoked potentials (CAEPs) and sentence-level perception in older normal-hearing (ONH) and older hearing-impaired (OHI) individuals, (2) determine the effects of hearing impairment and age on CAEPs and perception, and (3) explore how well CAEPs correlate with and predict speech perception in noise. Design: Two groups of older participants (15 ONH and 15 OHI) were tested using speech-in-noise stimuli to measure CAEPs and sentence-level perception of speech. The syllable /ba/, used to evoke CAEPs, and sentences were presented in speech-spectrum background noise at four signal levels (50, 60, 70, and 80 dB SPL) and up to seven SNRs (-10, -5, 0, 5, 15, 25, and 35 dB). These data were compared between groups to reveal the hearing impairment effect and then combined with previously published data for 15 young normal-hearing individuals to determine the aging effect. Results: Robust effects of SNR were found for perception and CAEPs. Small but significant effects of signal level were found for perception, primarily at poor SNRs and high signal levels, and in some limited instances for CAEPs. Significant effects of age were seen for both CAEPs and perception, while hearing impairment effects were only found with perception measures. CAEPs correlate well with perception and can predict SNR50s to within 2 dB for ONH. However, prediction error is much larger for OHI and varies widely (from 6 to 12 dB) depending on the model that was used for prediction. Conclusions: When background noise is present, SNR dominates both perception-in-noise testing and cortical electrophysiological testing, with smaller and sometimes significant contributions from signal level. A mismatch between behavioral and electrophysiological results was found (hearing impairment effects were primarily only seen for behavioral data), illustrating the possible contributions of higher order cognitive processes on behavior. It is interesting that the hearing impairment effect size was more than five times larger than the aging effect size for CAEPs and perception. Sentence-level perception can be predicted well in normal-hearing individuals; however, additional research is needed to explore improved prediction methods for older individuals with hearing impairment.
Article
The speech-evoked auditory brainstem response (sABR) provides a measure of encoding complex stimuli in the brainstem, and this study employed the sABR to better understand the role of neural temporal jitter in the response patterns from older adults. In experiment 1, sABR recordings were used to investigate age-related differences in periodicity encoding of the temporal envelope and fine structure components of the response to a/da/speech token. A group of younger and a group of older adults (n=22 per group) participated. The results demonstrated reduced amplitude of the fundamental frequency and harmonic components in the spectral domain of the recorded response of the older listeners. In experiment 2, a model of neural temporal jitter was employed to simulate in a group of young adults (n=22) the response patterns measured from older adults. A small group of older adults (n=7) were also tested under the jitter simulation conditions. In the young adults, the results showed a systematic reduction in the response amplitude of the most robust response components as the degree of applied jitter increased. In contrast, the older adults did not demonstrate significant response reduction when tested under jitter conditions. The overall pattern of results suggests that older adults have reduced neural synchrony for encoding periodic, complex signals at the level of the brainstem, and that this reduced synchrony can be modeled by simulating neural jitter via disruption of the temporal waveform of the stimulus.