Article

Incidental auditory learning and memory-guided attention: Examining the role of attention at the behavioural and neural level using EEG

Authors:
To read the full-text of this research, you can request a copy directly from the authors.

Abstract

The current study addressed the relation between awareness, attention, and memory, by examining whether merely presenting a tone and audio-clip, without deliberately associating one with other, was sufficient to bias attention to a given side. Participants were exposed to 80 different audio-clips (half included a lateralized pure tone) and told to classify audio-clips as natural (e.g., waterfall) or manmade (e.g., airplane engine). A surprise memory test followed, in which participants pressed a button to a lateralized faint tone (target) embedded in each audio-clip. They also indicated if the clip was (i) old/new; (ii) recollected/familiar; and (iii) if the tone was on left/right/not present when they heard the clip at exposure. The results demonstrate good explicit memory for the clip, but not for tone location. Response times were faster for old than for new clips but did not vary according to the target-context associations. Neuro-electric activity revealed an old-new effect at midline-frontal sites and a difference between old clips that were previously associated with the target tone and those that were not. These results support the attention-dependent learning hypothesis and suggest that associations were formed incidentally at a neural level (silent memory trace or engram), but these associations did not guide attention at a level that influenced behavior either explicitly or implicitly.

No full-text available

Request Full-text Paper PDF

To read the full-text of this research,
you can request a copy directly from the authors.

... Although research suggests that memory for past events influences how we respond to events in the present, the degree of behavioral benefit varies and can even be absent (Fischer et al., 2020;Giesbrecht et al., 2013;Preston and Gabrieli, 2008;Schankin and Schubö, 2009). Variability in behavior offers an interesting scenario in which to understand when, and how, memory prepares action. ...
... To elicit a behavioral response, associations must be both learned and then useable for behavior (e.g., Elward et al., 2021). Fischer et al. (2020), for instance, demonstrated that incidental associative LTM between sound clips and a lateralized tone does not speed up response times to the same tone presented in a subsequent surprise memory test, despite neural evidence of learning (neural trace). Further, Batterink and Zhang (2022) demonstrated that statistical learning of embedded words in a continuous stream of speech occurred during sleep, but that no behavioral benefits were detected during subsequent wake. ...
Article
Does memory prepare us to act? Long-term memory can facilitate signal detection, though the degree of benefit varies and can even be absent. To dissociate between learning and behavioral expression of learning, we used high-density electroencephalography to assess memory retrieval and response processing. At learning, participants heard everyday sounds. Half of these sounds were paired with an above-threshold lateralized tone, such that it was possible to form incidental associations between the sound and the location of the tone. Importantly, attention was directed to either the sound (Experiment 1) or the tone (Experiment 2). Participants then completed a novel detection task that separated cued retrieval from response processing. At retrieval, we observed a striking brain-behavior dissociation. Learning was observed neurally in both experiments. Behaviorally, however, signal detection was only facilitated in Experiment 2, for which there was an accompanying explicit memory for tone presence. Further, implicit neural memory for tone location correlated with the degree of response preparation, but not response execution. Together, the findings suggest 1) that attention at learning affects memory-biased action and 2) that memory prepared action via both explicit and implicit associative memory, with the latter triggering response preparation.
... As demonstrated in behavioral studies, both the explicit and implicit long-term memory for the location of a target modulates subsequently deployed auditory spatial attention [90]. Evidence from AEP investigations indicates that implicit encoding of auditory spatial cues is at play [91]. ...
Article
Full-text available
Auditory spatial cues contribute to two distinct functions, of which one leads to explicit localization of sound sources and the other provides a location-linked representation of sound objects. Behavioral and imaging studies demonstrated right-hemispheric dominance for explicit sound localization. An early clinical case study documented the dissociation between the explicit sound localizations, which was heavily impaired, and fully preserved use of spatial cues for sound object segregation. The latter involves location-linked encoding of sound objects. We review here evidence pertaining to brain regions involved in location-linked representation of sound objects. Auditory evoked potential (AEP) and functional magnetic resonance imaging (fMRI) studies investigated this aspect by comparing encoding of individual sound objects, which changed their locations or remained stationary. Systematic search identified 1 AEP and 12 fMRI studies. Together with studies of anatomical correlates of impaired of spatial-cue-based sound object segregation after focal brain lesions, the present evidence indicates that the location-linked representation of sound objects involves strongly the left hemisphere and to a lesser degree the right hemisphere. Location-linked encoding of sound objects is present in several early-stage auditory areas and in the specialized temporal voice area. In these regions, emotional valence benefits from location-linked encoding as well.
... Additionally, during the procedural learning phase, a fronto-occipital network, synchronized with low frequency gamma oscillations, has been identified (Chaumon et al., 2009). Furthermore, Fischer et al. (2020) observed significant modulations over the fronto-central scalp region during the cue period in an auditory memory-guided attention task. These modulations may be the result of reflections from a source located in auditory cortices. ...
Article
Prior knowledge and long‐term memory can guide our attention to facilitate search for and detection of subtle targets embedded in a complex scene. A number of neuropsychological and experimental studies have investigated this effect, yet results in the field remain mixed, as there is a lack of consensus regarding the neural correlates thought to support memory‐guided attention. The purpose of this systematic review and meta‐analysis was to identify a common set of brain structures involved in memory‐guided attention. Statistical analyses were computed on functional magnetic resonance imaging (fMRI) studies that presented participants with a task that required them to detect a target or a change embedded in repeated and novel complex visual displays. After a systematic search, 10 fMRI studies met the selection criteria and were included in the analysis. The results yielded four significant clusters. Activity in right inferior parietal (Brodmann area [BA] 9) and right superior parietal (BA 7) lobes suggests involvement of a fronto‐parietal attention network, while activity in left mid‐cingulate cortex (BA 23) and right middle frontal gyrus (BA 10) suggests involvement of a fronto‐parietal control network. These findings are consistent with the notion that fronto‐parietal circuits are important for interfacing retrieved memories with attentional systems to guide search. This article is categorized under: Psychology > Memory Psychology > Learning Psychology > Attention
Article
The meme-centric memetic automaton (MA) was recently proposed as an adaptive entity or a software agent wherein memes are defined as the building blocks of knowledge. The conceptualization of MA has led to the development of a large number of potentially rich meme-inspired designs that form a cornerstone of memetic computation as tools for problem-solving. In this study, we investigate the use of memetic multi-agent systems to develop more intelligent and human-like autonomous agents by taking MA as the essential backbone of the agent. Taking inspiration from a psychological Broadbent-Treisman Attenuation Model, we propose an attention intensity control method in meme expression for enhancing agents’ perception of the value of all kinds of information captured from the environment, hence leading to a greater capability of meme knowledge generalization. Our particular focus is placed on the design of meme selection for more effective knowledge transmission across the population. To this end, we introduce a bidirectional imitation strategy based on agents’ estimation of the importance and/or uncertainty of decision making in a dynamic environment. Experiments on a minefield navigation simulation as well as a commercial video game demonstrate the superior performance of our proposed method compared to state-of-the-art methods.
Article
Full-text available
A large number of reports have been published on stochastic independence between implicit and explicit measures of memory. This is often taken to imply that different memory systems mediate implicit and explicit memory performance. In these cases, stochastic independence is inferred from contingency analysis of overall success rates in two memory tasks when performance in one or both of the tasks is, to a large extent, mediated by factors other than memory. Typically, the difference between performance with studied and nonstudied items is not large in implicit memory tasks. It is argued that this must be taken into account when evaluating the contingency analysis. A method is presented for estimating the relevant joint and conditional probabilities, assuming that the aspects of performance in the two tasks that are related to memory are dependent to the maximum possible extent. The method is applied to a number of published studies, and it is shown that the difference between these estimated probabilities and those given by stochastic independence is too small to allow any conclusion to be drawn about memory systems from contingency analysis of data reported in these studies.
Article
Full-text available
The traditional explanation of experiments related to conditioning of verbal behavior in terms of operant conditioning, i.e., learning without awareness, was questioned. It was hypothesized that such conditioning to cues given by E might be based on S's awareness of the cue and assumed meaning of such a stimulus. 2 experiments were performed, 1 where plural nouns were reinforced, and a 2nd to test the generalization of this acquired pattern on a word association test. The usual findings re: conditioning of plural nouns was replicated and there seemed to be a carry-over of the set to the word association test. However, some Ss seemed to be aware of the meaning of such a reinforcer as "um-hm." Moreover, response set was also seen to affect conditionability.
Article
Full-text available
Auditory long-term memory has been shown to facilitate signal detection. However, the nature and timing of the cognitive processes supporting such benefits remain equivocal. We measured neuroelectric brain activity while young adults were presented with a contextual memory cue designed to assist with the detection of a faint pure tone target embedded in an audio clip of an everyday environmental scene (e.g., the soundtrack of a restaurant). During an initial familiarization task, participants heard such audio clips, half of which included a target sound (memory cue trials) at a specific time and location (left or right ear), as well as audio clips without a target (neutral trials). Following a one-hour or twenty-four-hour retention interval, the same audio clips were presented, but now all included a target. Participants were asked to press a button as soon as they heard the pure tone target. Overall, participants were faster and more accurate during memory than neutral cue trials. The auditory contextual memory effects on performance coincided with three temporally and spatially distinct neural modulations, which encompassed changes in the amplitude of event-related potential as well as changes in theta, alpha, beta and gamma power. Brain electrical source analyses revealed greater source activity in memory than neutral cue trials in the right superior temporal gyrus and left parietal cortex. Conversely, neutral trials were associated with greater source activity than memory cue trials in the left posterior medial temporal lobe. Target detection was associated with increased negativity (N2), and a late positive (P3b) wave at frontal and parietal sites, respectively. The effect of auditory contextual memory on brain activity preceding target onset showed little lateralization. Together, these results are consistent with contextual memory facilitating retrieval of target-context associations and deployment and management of auditory attentional resources to when the target occurred. The results also suggest that the auditory cortices, parietal cortex, and medial temporal lobe may be parts of a neural network enabling memory-guided attention during auditory scene analysis.
Article
Full-text available
The brain networks supporting speech identification and comprehension under difficult listening conditions are not well specified. The networks hypothesized to underlie effortful listening include regions responsible for executive control. We conducted meta-analyses of auditory neuroimaging studies to determine whether a common activation pattern of the frontal lobe supports effortful listening under different speech manipulations. Fifty-three functional neuroimaging studies investigating speech perception were divided into three independent Activation Likelihood Estimate analyses based on the type of speech manipulation paradigm used: Speech-in-noise (SIN, 16 studies, involving 224 participants); spectrally degraded speech using filtering techniques (15 studies involving 270 participants); and linguistic complexity (i.e., levels of syntactic, lexical and semantic intricacy/density, 22 studies, involving 348 participants). Meta-analysis of the SIN studies revealed higher effort was associated with activation in left inferior frontal gyrus (IFG), left inferior parietal lobule, and right insula. Studies using spectrally degraded speech demonstrated increased activation of the insula bilaterally and the left superior temporal gyrus (STG). Studies manipulating linguistic complexity showed activation in the left IFG, right middle frontal gyrus, left middle temporal gyrus and bilateral STG. Planned contrasts revealed left IFG activation in linguistic complexity studies, which differed from activation patterns observed in SIN or spectral degradation studies. Although there were no significant overlap in prefrontal activation across these three speech manipulation paradigms, SIN and spectral degradation showed overlapping regions in left and right insula. These findings provide evidence that there is regional specialization within the left IFG and differential executive networks underlie effortful listening.
Article
Full-text available
In human scalp EEG recordings, both sustained potentials and alpha-band oscillations are present during the delay period of working memory tasks and may therefore reflect the representation of information in working memory. However, these signals may instead reflect support mechanisms rather than the actual contents of memory. In particular, alpha-band oscillations have been tightly tied to spatial attention and may not reflect location-independent memory representations per se. To determine how sustained and oscillating EEG signals are related to attention and working memory, we attempted to decode which of 16 orientations was being held in working memory by human observers (both women and men). We found that sustained EEG activity could be used to decode the remembered orientation of a stimulus, even when the orientation of the stimulus varied independently of its location. Alpha-band oscillations also carried clear information about the location of the stimulus, but they provided little or no information about orientation independently of location. Thus, sustained potentials contain information about the object properties being maintained in working memory, consistent with previous evidence of a tight link between these potentials and working memory capacity. In contrast, alpha-band oscillations primarily carry location information, consistent with their link to spatial attention.
Article
Full-text available
Long-term memory (LTM) has been shown to bias attention to a previously learned visual target location. Here, we examined whether memory-predicted spatial location can facilitate the detection of a faint pure tone target embedded in real world audio clips (e.g., soundtrack of a restaurant). During an initial familiarization task, participants heard audio clips, some of which included a lateralized target (p = 50%). On each trial participants indicated whether the target was presented from the left, right, or was absent. Following a 1 hr retention interval, participants were presented with the same audio clips, which now all included a target. In Experiment 1, participants showed memory-based gains in response time and d'. Experiment 2 showed that temporal expectations modulate attention, with greater memory-guided attention effects on performance when temporal context was reinstated from learning (i.e., when timing of the target within audio clips was not changed from initially learned timing). Experiment 3 showed that while conscious recall of target locations was modulated by exposure to target-context associations during learning (i.e., better recall with higher number of learning blocks), the influence of LTM associations on spatial attention was not reduced (i.e., number of learning blocks did not affect memory-guided attention). Both Experiments 2 and 3 showed gains in performance related to target-context associations, even for associations that were not explicitly remembered. Together, these findings indicate that memory for audio clips is acquired quickly and is surprisingly robust; both implicit and explicit LTM for the location of a faint target tone modulated auditory spatial attention. (PsycINFO Database Record
Article
Full-text available
Selective attention determines the effectiveness of implicit contextual learning (e.g., Jiang and Leung, 2005). Visual foreground-background segmentation, on the other hand, is a key process in the guidance of attention (Wolfe, 2003). In the present study, we examined the impact of foreground-background segmentation on contextual cueing of visual search in three experiments. A visual search display, consisting of distractor ‘L’s and a target ‘T’, was overlaid on a task-neutral cuboid on the same depth plane (Experiment 1), on stereoscopically separated depth planes (Experiment 2), or spread over the entire display on the same depth plane (Experiment 3). Half of the search displays contained repeated target-distractor arrangements, whereas the other half was always newly generated. The task-neutral cuboid was constant during an initial training session, but was either rotated by 90° or entirely removed in the subsequent test sessions. We found that the gains resulting from repeated presentation of display arrangements during training (i.e., contextual-cueing effects) were diminished when the cuboid was changed or removed in Experiment 1, but remained intact in Experiments 2 and 3 when the cuboid was placed in a different depth plane, or when the items were randomly spread over the whole display but not on the edges of the cuboid. These findings suggest that foreground-background segmentation occurs prior to contextual learning, and only objects/arrangements that are grouped as foreground are learned over the course of repeated visual search.
Article
Full-text available
The last decade has seen dramatic technological and conceptual changes in research on episodic memory and the brain. New technologies, and increased use of more naturalistic observations, have enabled investigators to delve deeply into the structures that mediate episodic memory, particularly the hippocampus, and to track functional and structural interactions among brain regions that support it. Conceptually, episodic memory is increasingly being viewed as subject to lifelong transformations that are reflected in the neural substrates that mediate it. In keeping with this dynamic perspective, research on episodic memory (and the hippocampus) has infiltrated domains, from perception to language and from empathy to problem solving, that were once considered outside its boundaries. Using the component process model as a framework, and focusing on the hippocampus, its subfields, and specialization along its longitudinal axis, along with its interaction with other brain regions, we consider these new developments and their implications for the organization of episodic memory and its contribution to functions in other domains.
Article
Full-text available
ALBERT BREGMAN'S (1990) BOOK AUDITORY Scene Analysis: The Perceptual Organization of Sound has had a tremendous impact on research in auditory neuroscience. Here, we outline some of the accomplishments. This review is not meant to be exhaustive, but rather aims to highlight milestones in the brief history of auditory neuroscience. The steady increase in neuroscience research following the book's pivotal publication has advanced knowledge about how the brain forms representations of auditory objects. This research has far-reaching societal implications on health and quality of life. For instance, it helped us understand why some people experience difficulties understanding speech in noise, which in turn has led to development of therapeutic interventions. Importantly, the book acts as a catalyst, providing scientists with a common conceptual framework for research in such diverse fields as speech perception, music perception, neurophysiology and computational neuroscience. This interdisciplinary approach to research in audition is one of this book's legacies.
Article
Full-text available
In visual search, participants detect and subsequently discriminate targets more rapidly when these are embedded in repeatedly encountered distractor arrangements, an effect termed contextual cueing (Chun & Jiang Cognitive Psychology, 36, 28-71, 1998). However, whereas previous studies had explored contextual cueing exclusively in visual search, in the present study we examined the effect in tactile search using a novel tactile search paradigm. Participants were equipped with vibrotactile stimulators attached to four fingers on each hand. A given search array consisted of four stimuli (i.e., two items presented to each hand), with the target being an odd-one-out feature singleton that differed in frequency (Exps. 1 and 2) or waveform (Exp. 3) from the distractor elements. Participants performed a localization (Exps. 1 and 2) or discrimination (Exp. 3) task, delivering their responses via foot pedals. In all three experiments, reaction times were faster when the arrangement of distractor fingers predicted the target finger. Furthermore, participants were unable to explicitly discriminate repeated from nonrepeated tactile configurations (Exps. 2 and 3). This indicates that the tactile modality can mediate the formation of configural representations and use these representations to guide tactile search.
Article
Full-text available
Sounds are ephemeral. Thus, coherent auditory perception depends on "hearing" back in time: retrospectively attending that which was lost externally but preserved in short-term memory (STM). Current theories of auditory attention assume that sound features are integrated into a perceptual object, that multiple objects can coexist in STM, and that attention can be deployed to an object in STM. Recording electroencephalography from humans, we tested these assumptions, elucidating feature-general and feature-specific neural correlates of auditory attention to STM. Alpha/beta oscillations and frontal and posterior event-related potentials indexed feature-general top-down attentional control to one of several coexisting auditory representations in STM. Particularly, task performance during attentional orienting was correlated with alpha/low-beta desynchronization (i.e., power suppression). However, attention to one feature could occur without simultaneous processing of the second feature of the representation. Therefore, auditory attention to memory relies on both feature-specific and feature-general neural dynamics. Copyright © 2015 the authors 0270-6474/15/351307-12$15.00/0.
Article
Full-text available
Despite a growing acceptance that attention and memory interact, and that attention can be focused on an active internal mental representation (i.e., reflective attention), there has been a paucity of work focusing on reflective attention to 'sound objects' (i.e., mental representations of actual sound sources in the environment). Further research on the dynamic interactions between auditory attention and memory, as well as its degree of neuroplasticity, is important for understanding how sound objects are represented, maintained, and accessed in the brain. This knowledge can then guide the development of training programs to help individuals with attention and memory problems. This review article focuses on attention to memory with an emphasis on behavioral and neuroimaging studies that have begun to explore the mechanisms that mediate reflective attentional orienting in vision and more recently, in audition. Reflective attention refers to situations in which attention is oriented toward internal representations rather than focused on external stimuli. We propose four general principles underlying attention to short-term memory. Furthermore, we suggest that mechanisms involved in orienting attention to visual object representations may also apply for orienting attention to sound object representations.
Article
Full-text available
Abstract A neuropsychological model of memory is proposed that incorporates Fodor's (1983) idea of modules and central systems. The model has four essential components: (1) a non-frontal neocortical component that consists of perceptual (and perhaps interpretative semantic) modules that mediate performance on item-specific, implicit tests of memory, (2) a modular medial temporal/hippocampal component that mediates encoding, storage, and retrieval on explicit, episodic tests of memory that are associative/cue dependent, (3) a central system, frontal-lobe component that mediates performance on explicit tests that are strategic and on procedural tests that are rule-bound, and (4) a basal ganglia component that mediates performance on sensorimotor, procedural tests of memory. The usefulness of the modular/central system construct is explored and evidence from studies of normal, amnesic, agnosic, and demented people is provided to support the model.
Article
Full-text available
A 2-process theory of human information processing is proposed and applied to detection, search, and attention phenomena. Automatic processing is activation of a learned sequence of elements in long-term memory that is initiated by appropriate inputs and then proceeds automatically--without S control, without stressing the capacity limitations of the system, and without necessarily demanding attention. Controlled processing is a temporary activation of a sequence of elements that can be set up quickly and easily but requires attention, is capacity-limited (usually serial in nature), and is controlled by the S. A series of studies, with approximately 8 Ss, using both reaction time and accuracy measures is presented, which traces these concepts in the form of automatic detection and controlled search through the areas of detection, search, and attention. Results in these areas are shown to arise from common mechanisms. Automatic detection is shown to develop following consistent mapping of stimuli to responses over trials. Controlled search was utilized in varied-mapping paradigms, and in the present studies, it took the form of serial, terminating search. (60 ref) (PsycINFO Database Record (c) 2012 APA, all rights reserved)
Article
Full-text available
In everyday situations, we often rely on our memories to find what we are looking for in our cluttered environment. Recently, we developed a new experimental paradigm to investigate how long-term memory (LTM) can guide attention and showed how the pre-exposure to a complex scene in which a target location had been learned facilitated the detection of the transient appearance of the target at the remembered location [Summerfield, J. J., Rao, A., Garside, N., & Nobre, A. C. Biasing perception by spatial long-term memory. The Journal of Neuroscience, 31, 14952–14960, 2011; Summerfield, J. J., Lepsien, J., Gitelman, D. R., Mesulam, M. M., & Nobre, A. C. Orienting attention based on long-term memory experience. Neuron, 49, 905–916, 2006]. This study extends these findings by investigating whether and how LTM can enhance perceptual sensitivity to identify targets occurring within their complex scene context. Behavioral measures showed superior perceptual sensitivity (d′) for targets located in remembered spatial contexts. We used the N2pc ERP to test whether LTM modulated the process of selecting the target from its scene context. Surprisingly, in contrast to effects of visual spatial cues or implicit contextual cueing, LTM for target locations significantly attenuated the N2pc potential. We propose that the mechanism by which these explicitly available LTMs facilitate perceptual identification of targets may differ from mechanisms triggered by other types of top–down sources of information.
Article
Full-text available
Effects of full and divided attention during study on explicit and implicit memory performance were investigated in two experiments. Study time was manipulated in a third experiment. Experiment 1 showed that both similar and dissociative effects can be found in the two kinds of memory test, depending on the difficulty of the concurrent tasks used in the divided-attention condition. In this experiment, however, standard implicit memory tests were used and contamination by explicit memory influences cannot be ruled out. Therefore, in Experiments 2 and 3 the process dissociation procedure was applied. Manipulations of attention during study and of study time clearly affected the controlled (explicit) memory component, but had no effect on the automatic (implicit) memory component. Theoretical implications of these findings are discussed.
Article
Full-text available
An area of research that has experienced recent growth is the study of memory during perception of simple and complex auditory scenes. These studies have provided important information about how well auditory objects are encoded in memory and how well listeners can notice changes in auditory scenes. These are significant developments because they present an opportunity to better understand how we hear in realistic situations, how higher-level aspects of hearing such as semantics and prior exposure affect perception, and the similarities and differences between auditory perception and perception in other modalities, such as vision and touch. The research also poses exciting challenges for behavioral and neural models of how auditory perception and memory work.
Article
Full-text available
Numerous studies have shown that musicians outperform nonmusicians on a variety of tasks. Here we provide the first evidence that musicians have superior auditory recognition memory for both musical and nonmusical stimuli, compared to nonmusicians. However, this advantage did not generalize to the visual domain. Previously, we showed that auditory recognition memory is inferior to visual recognition memory. Would this be true even for trained musicians? We compared auditory and visual memory in musicians and nonmusicians using familiar music, spoken English, and visual objects. For both groups, memory for the auditory stimuli was inferior to memory for the visual objects. Thus, although considerable musical training is associated with better musical and nonmusical auditory memory, it does not increase the ability to remember sounds to the levels found with visual stimuli. This suggests a fundamental capacity difference between auditory and visual recognition memory, with a persistent advantage for the visual domain.
Article
Full-text available
Voluntary orienting of visual attention is conventionally measured in tasks with predictive central cues followed by frequent valid targets at the cued location and by infrequent invalid targets at the uncued location. This implies that invalid targets entail both spatial reorienting of attention and breaching of the expected spatial congruency between cues and targets. Here, we used event-related functional magnetic resonance imaging (fMRI) to separate the neural correlates of the spatial and expectancy components of both endogenous orienting and stimulus-driven reorienting of attention. We found that during endogenous orienting with predictive cues, there was a significant deactivation of the right Temporal-Parietal Junction (TPJ). We also discovered that the lack of an equivalent deactivation with nonpredictive cues was matched to drop in attentional costs and preservation of attentional benefits. The right TPJ showed equivalent responses to invalid targets following predictive and nonpredictive cues. On the contrary, infrequent-unexpected invalid targets following predictive cues specifically activated the right Middle and Inferior Frontal Gyrus (MFG-IFG). Additional comparisons with spatially neutral trials demonstrated that, independently of cue predictiveness, valid targets activate the left TPJ, whereas invalid targets activate both the left and right TPJs. These findings show that the selective right TPJ activation that is found in the comparison between invalid and valid trials results from the reciprocal cancelling of the different activations that in the left TPJ are related to the processing of valid and invalid targets. We propose that left and right TPJs provide "matching and mismatching to attentional template" signals. These signals enable reorienting of attention and play a crucial role in the updating of the statistical contingency between cues and targets.
Article
Full-text available
Five experiments explored the effect of acoustic changes between study and test on implicit and explicit memory for spoken words. Study-test changes in the speaker's voice, intonation, and fundamental frequency produced significant impairments of auditory priming on implicit tests of auditory identification and stem completion but had little or no effect on explicit recall and recognition tests (Experiments 1-4). However, study-test changes in overall decibel level had no effect on priming on an auditory stem-completion test or on cued-recall performance (Experiment 5). The results are consistent with the idea that fundamental frequency information is represented in a perceptual representation system that plays an important role in auditory priming.
Article
Full-text available
One presentation of a word to a subject is enough to change the way in which the word is processed subsequently, even when there is no conscious (explicit) memory of the original presentation. This phenomenon is known as implicit memory. The neural correlates of implicit memory have been studied previously, but they have never been compared with the correlates of explicit memory while holding task conditions constant or while using a procedure that ensured that the neural correlates were not 'contaminated' by explicit memory. Here we use scalp-recorded event-related brain potentials to identify neural activity associated with implicit and explicit memory during the performance of a recognition memory task. Relative to new words, recently studied words produced activity in three neuroanatomically and functionally dissociable neural populations. One of these populations was activated whether or not the word was consciously recognized, and its activity therefore represents a neural correlate of implicit memory. Thus, when task and memory contamination effects are eliminated, the neural correlates of explicit and implicit memory differ qualitatively.
Article
Full-text available
Global context plays an important, but poorly understood, role in visual tasks. This study demonstrates that a robust memory for visual context exists to guide spatial attention. Global context was operationalized as the spatial layout of objects in visual search displays. Half of the configurations were repeated across blocks throughout the entire session, and targets appeared within consistent locations in these arrays. Targets appearing in learned configurations were detected more quickly. This newly discovered form of search facilitation is termed contextual cueing. Contextual cueing is driven by incidentally learned associations between spatial configurations (context) and target locations. This benefit was obtained despite chance performance for recognizing the configurations, suggesting that the memory for context was implicit. The results show how implicit learning and memory of visual context can guide spatial attention towards task-relevant aspects of a scene.
Article
Full-text available
Skill learning and repetition priming are considered by some to be supported by separate memory systems. The authors examined the relationship between skill learning and priming in 3 experiments using a digit entering task, in which participants were presented with unique and repeated 5-digit strings with controlled sequential structure. Both skill learning and priming were observed across a wide range of skill levels. Performance reflected the effects of learning at 3 different levels of stimulus structure, calling into question a binary dichotomy between item-specific priming and general skill learning. Two computational models were developed which demonstrated that previous dissociations between skill learning and priming can occur within a single memory system. The experimental and computational results are interpreted as suggesting that skill learning and priming should be viewed as 2 aspects of a single incremental learning mechanism.
Article
Full-text available
As event-related brain potential (ERP) researchers have increased the number of recording sites, they have gained further insights into the electrical activity in the neural networks underlying explicit memory. A review of the results of such ERP mapping studies suggests that there is good correspondence between ERP results and those from brain imaging studies that map hemodynamic changes. This concordance is important because the combination of the high temporal resolution of ERPs with the high spatial resolution of hemodynamic imaging methods will provide a greatly increased understanding of the spatio-temporal dynamics of the brain networks that encode and retrieve explicit memories.
Article
Full-text available
It is widely hypothesized that separate recollection and familiarity processes contribute to recognition memory. The present research measured event-related brain potentials (ERPs) from 128 head locations to identify patterns of brain activity related to recollection and familiarity. In two experiments, subjects performed a recognition memory task requiring discrimination between previously studied words, similar words that changed plurality between study and test, and new words (following Hintzman & Curran, 1994). The FN400 ERP component (300-500 msec) varied with the familiarity of words (new > studied = similar). The parietal component (400-800 msec) was associated with the recollection of plurality (studied > similar = new). Differences in the timing and spatial topography of the FN400 and parietal effects support the view that familiarity and recollection arise from distinct neurocognitive processes.
Article
Full-text available
One of the most debated questions in visual attention research is what factors affect the deployment of attention in the visual scene? Segmentation processes are influential factors, providing candidate objects for further attentional selection, and the relevant literature has concentrated on how figure-ground segmentation mechanisms influence visual attention. However, another crucial process, namely foreground-background segmentation, seems to have been neglected. By using a change blindness paradigm, we explored whether attention is preferentially allocated to the foreground elements or to the background ones. The results indicated that unless attention was voluntarily deployed to the background, large changes in the color of its elements remained unnoticed. In contrast, minor changes in the foreground elements were promptly reported. Differences in change blindness between the two regions of the display indicate that attention is, by default, biased toward the foreground elements. This also supports the phenomenal observations made by Gestaltists, who demonstrated the greater salience of the foreground than the background.
Article
The neural substrate of memory The ability to form memory is an essential trait that allows learning and the accumulation of knowledge. But what is a memory? There has been a long history of searching for the neuronal substrate that forms memory in the brain, and the emerging view is that ensembles of engram cells explain how memories are formed and retrieved. In a Review, Josselyn and Tonegawa discuss the evidence for engram cells as a substrate of memory, particularly in rodents; what we have learned so far about the features of memory, including memory formation, retrieval over time, and loss; and future directions to understand how memory becomes knowledge. Science , this issue p. eaaw4325
Article
Increased travel worldwide has led to an escalation of road traffic accidents, particularly among tourists driving in unfamiliar, opposite traffic flow driving scenarios. Ability to allocate attention to driving-relevant information and regions is predicted to be the main cause of tourist accidents, with a lack of attention directed to areas of space that are inhibited in familiar traffic conventions but relevant in overseas driving. This study investigated the influence of habit and expectancy on driver behaviour and allocation of attention in familiar (left-hand traffic; LHT) and unfamiliar (right-hand traffic; RHT) contexts. Twenty-eight drivers from the UK were presented with video clips of driving taken in the UK and in Poland and asked to judge whether it was safe to enter a roundabout in each clip. Half were given information about differences in LHT and RHT situations prior to the task. Judgement performance was not influenced by this information, however accuracy was higher for LHT and the RHT task was rated more difficult, supporting the notion that driving in unfamiliar surroundings is more effortful. In LHT both groups made more fixations to the right side of each roundabout, however in RHT, whilst the control group allocated attention in the same way, the intervention group made significantly more fixations to the left. Pre-drive preparatory information can therefore increase attention to the most relevant areas of space in unfamiliar driving contexts. This has implications for drive tourism and it is suggested that such information is made more explicit to drivers.
Chapter
Successful communication and navigation in cocktail party situations depends on complex interactions among an individual’s sensory, cognitive, and social abilities. Older adults may function well in relatively ideal communication situations, but they are notorious for their difficulties understanding speech in noisy situations such as cocktail parties. However, as healthy adults age, declines in auditory and cognitive processing may be offset by compensatory gains in ability to use context and knowledge. From a practical perspective, it is important to consider the aging auditory system in multitalker situations because these are among the most challenging situations for older adults. From a theoretical perspective, studying age-related changes in auditory processing provides a special window into the relative contributions of, and interactions among sensory, cognitive, and social abilities. In the acoustical wild, younger listeners typically function better than older listeners. Experimental evidence indicates that age-related differences in simple measures such as word recognition in quiet or noise are largely due to the bottom-up effects of age-related auditory declines. These differences can often be eliminated when auditory input is adjusted to equate the performance levels of listeners on baseline measures in quiet or noise. Notably, older adults exhibit enhanced cognitive compensation, with performance on auditory tasks being facilitated by top-down use of context and knowledge. Nevertheless, age-related differences can persist when tasks are more cognitively demanding and involve discourse comprehension, memory, and attention. At an extreme, older adults with hearing loss are at greater risk for developing cognitive impairments than peers with better hearing.
Article
Amnesia refers specifically to the focal disruption of anterograde and retrograde memory, with no other major impairments or clinical features. The study of focal lesion patients has greatly advanced our conceptualization of memory systems and their neuroanatomical basis. Although these distinctions are still very relevant, recent research is challenging such distinctions, and the independence of memory systems that they imply. The authors review the major memory systems with relevant patient findings, and subsequently discuss new insights into how different types of memory are affected in amnesia. The authors finish with a discussion on the implications of recent research for theories of amnesia.
Article
Implicit learning is generally characterized as learning that proceeds both unintentionally and unconsciously. Here are some examples: 1 Reber (1967), who coined the term ‘implicit learning’, asked participants to study a series of letter strings such as VXVS for a few seconds each. Then he told them that these strings were all constructed according to a particular set of rules (that is, a grammar; see Figure 8.1) and that in the test phase they would see some new strings and would have to decide which ones conformed to the same rules and which ones did not. Participants could make these decisions with better-than-chance accuracy but had little ability to describe the rules. For example, participants could not recall correctly which letters began and ended the strings. Reber described his results as a ‘peculiar combination of highly efficient behavior with complex stimuli and almost complete lack of verbalizable knowledge about them ’ (p. 859). 2 In the 1950s, a number of studies asked people to generate words ad libitum and established that the probability with which they would produce, say, plural nouns was increased if each such word was reinforced by the experimenter saying ‘umhmm ’ (e.g. Greenspoon, 1955). This result occurred in subjects apparently unable to report the reinforcement contingency. 3 Svartdal (1991) presented participants with brief trains of between 4 and 17 auditory clicks. Participants immediately had to press a response button exactly the same number of times and were instructed that feedback would be presented when the number of presses matched the number of clicks. In fact, though, feedback was contingent on speed of responding: for some
Article
Many daily activities involve looking for something. The ease with which these searches are performed often allows one to forget that searching represents complex interactions between visual attention and memory. Although a clear understanding exists of how search efficiency will be influenced by visual features of targets and their surrounding distractors or by the number of items in the display, the role of memory in search is less well understood. Contextual cueing studies have shown that implicit memory for repeated item configurations can facilitate search in artificial displays. When searching more naturalistic environments, other forms of memory come into play. For instance, semantic memory provides useful information about which objects are typically found where within a scene, and episodic scene memory provides information about where a particular object was seen the last time a particular scene was viewed. In this paper, we will review work on these topics, with special emphasis on the role of memory in guiding search in organized, real-world scenes. © 2015 New York Academy of Sciences.
Article
G*Power (Erdfelder, Faul, & Buchner, 1996) was designed as a general stand-alone power analysis program for statistical tests commonly used in social and behavioral research. G*Power 3 is a major extension of, and improvement over, the previous versions. It runs on widely used computer platforms (i.e., Windows XP, Windows Vista, and Mac OS X 10.4) and covers many different statistical tests of the t, F, and chi2 test families. In addition, it includes power analyses for z tests and some exact tests. G*Power 3 provides improved effect size calculators and graphic options, supports both distribution-based and design-based input modes, and offers all types of power analyses in which users might be interested. Like its predecessors, G*Power 3 is free.
Article
A number of ways of taxonomizing human learning have been proposed. We examine the evidence for one such proposal, namely, that there exist independent explicit and implicit learning systems. This combines two further distinctions, (1) between learning that takes place with versus without concurrent awareness, and (2) between learning that involves the encoding of instances (or fragments) versus the induction of abstract rules or hypotheses. Implicit learning is assumed to involve unconscious rule learning. We examine the evidence for implicit learning derived from subliminal learning, conditioning, artificial grammar learning, instrumental learning, and reaction times in sequence learning. We conclude that unconscious learning has not been satisfactorily established in any of these areas. The assumption that learning in some of these tasks (e.g., artificial grammar learning) is predominantly based on rule abstraction is questionable. When subjects cannot report the “implicitly learned” rules that govern stimulus selection, this is often because their knowledge consists of instances or fragments of the training stimuli rather than rules. In contrast to the distinction between conscious and unconscious learning, the distinction between instance and rule learning is a sound and meaningful way of taxonomizing human learning. We discuss various computational models of these two forms of learning.
Article
Driver inattention is thought to cause many automobile crashes. However, the research on attention is fragmented, and the applied research on driving and attention is further split between three largely independent traditions: the experimental research, the differential crash rate research, and the automation research. The goal of this review is to provide a conceptual framework to unify the research—a framework based on the combination of two fundamental dimen-sions of attentional selection: selection with and without conscious awareness (controlled and automatic), and selection by innate and acquired cognitive mechanisms (exogenous and endogenous). When applied to studies chosen to represent a broad range within the experimental literature, it reveals links between a variety of factors, including inexperience, inebriation, distracting stimuli, heads-up displays, fatigue, rumination, and secondary tasks such as phone conversations. This framework also has clear implications for the differential crash literature and the study of automated systems that support or replace functions of the driver. We conclude that driving research and policy could benefit from consideration of the different modes of attentional selection insofar as they integrate literatures, reveal directions for future research, and predict the effectiveness of interventions for crash-prevention.
Article
Priming and recollection are expressions of human memory mediated by different brain events. These brain events were monitored while people discriminated words from nonwords. Mean response latencies were shorter for words that appeared in an earlier study phase than for new words. This priming effect was reduced when the letters of words in study-phase presentations were presented individually in succession as opposed to together as complete words. Based on this outcome, visual word-form priming was linked to a brain potential recorded from the scalp over the occipital lobe about 450 ms after word onset. This potential differed from another potential previously associated with recollection, suggesting that distinct operations associated with these two types of memory can be monitored at the precise time that they occur in the human brain.
Article
The anterior prefrontal cortex is usually associated with high-level executive functions. In contrast, we show anterior prefrontal involvement in implicit change detection processes. A variant of the contextual cueing paradigm was used, in which repeated distractor configurations are implicitly learned and facilitate target search. After only six repetitions, the target location was changed in displays with repeated distractor configurations. We observed selective post-change signal increases in the anterior prefrontal cortex in repeated, but not novel displays. The data support the view that the anterior prefrontal cortex is involved in implicit change detection. This change detection is not dependent on extensive prior learning. Thus, anterior prefrontal involvement in complex cognitive tasks may be due to more basic processes than previously thought.
Article
Image regions corresponding to partially hidden objects are enclosed by two types of bounding contour: those inherent to the object itself (intrinsic) and those defined by occlusion (extrinsic). Intrinsic contours provide useful information regarding object shape, whereas extrinsic contours vary arbitrarily depending on accidental spatial relationships in scenes. Because extrinsic contours can only degrade the process of surface description and object recognition, it is argued that they must be removed prior to a stage of template matching. This implies that the two types of contour must be distinguished relatively early in visual processing and we hypothesize that the encoding of depth is critical for this task. The common border is attached to and regarded as intrinsic to the closer region, and detached from and regarded as extrinsic to the farther region. We also suggest that intrinsic borders aid in the segmentation of image regions and thus prevent grouping, whereas extrinsic borders provide a linkage to other extrinsic borders and facilitate grouping. Support for these views is found in a series of demonstrations, and also in an experiment where the expected superiority of recognition was found when partially sampled faces were seen in a back rather than a front stereoscopic depth plane.
Article
A new hypothesis about the role of focused attention is proposed. The feature-integration theory of attention suggests that attention must be directed serially to each stimulus in a display whenever conjunctions of more than one separable feature are needed to characterize or distinguish the possible objects presented. A number of predictions were tested in a variety of paradigms including visual search, texture segregation, identification and localization, and using both separable dimensions (shape and color) and local elements or parts of figures (lines, curves, etc. in letters) as the features to be integrated into complex wholes. The results were in general consistent with the hypothesis. They offer a new set of criteria for distinguishing separable from integral features and a new rationale for predicting which tasks will show attention limits and which will not.
Article
Many theorists assume that the cognitive system is composed of a collection of encapsulated processing components or modules, each dedicated to performing a particular cognitive function. On this view, selective impairments of cognitive tasks following brain damage, as evidenced by double dissociations, are naturally interpreted in terms of the loss of particular processing components. By contrast, the current investigation examines in detail a double dissociation between concrete and abstract work reading after damage to a connectionist network that pronounces words via meaning and yet has no separable components (Plaut & Shallice, 1993). The functional specialization in the network that gives rise to the double dissociation is not transparently related to the network's structure, as modular theories assume. Furthermore, a consideration of the distribution of effects across quantitatively equivalent individual lesions in the network raises specific concerns about the interpretation of single-case studies. The findings underscore the necessity of relating neuropsychological data to cognitive theories in the context of specific computational assumptions about how the cognitive system operates normally and after damage.
Article
Results are reviewed from several neuromagnetic studies which characterize the temporal dynamics of neural sources contributing to the visual evoked response and effects of attention on these sources. Different types of pattern-onset stimuli (< or = 2 degrees) were presented sequentially to a number of field locations in the right visual field. Multiple dipole models were applied to a sequence of instantaneous field distributions constructed at 10 ms intervals. Best-fitting source parameters were superimposed on Magnetic Resonance images (MRI) of each subject to identify the anatomical structure(s) giving rise to the surface patterns. At least three sources, presumably corresponding to different visual areas, were routinely identified from 80-150 ms following the onset of visual stimulation. This observation was consistent across subjects and studies. The temporal sequence and strength of activation of these sources, however, were dependent upon the specific stimulus parameters used to evoke the response (e.g., eccentricity) and on the relevance of the stimulus to the subject. In addition, our results provide evidence for the recurrence of activity in striate and extrastriate regions, following the initial cycle of responses.
Article
The electrophysiological correlates of recognition memory for new associations were investigated in two experiments. In both experiments subjects first studied unrelated word pairs. At test, they were presented with old words in the same pairing as at study (same pairs), old words in a different pairing from study (rearranged pairs), and pairs of new words. In Experiment 1 the test requirement was to discriminate between old and new pairs and, for any pair judged old, to then judge whether the pair was the same or rearranged. In Experiment 2 the requirement was merely to discriminate between old and new pairs. Event-related potentials (ERPs) were recorded for correctly classified same, rearranged and new pairs. The ERPs elicited by same pairs exhibited a similar pattern of effects in both experiments. Relative to the ERPs to new pairs, these effects took the form of sustained positive shifts with two distinct scalp maxima, over the left temporo-parietal and right frontal scalp respectively. ERPs to rearranged pairs showed effects which were similar in scalp topography, but markedly smaller in magnitude. This pattern of ERP effects closely resembles that found previously for test items defined as recollected on the basis of their attracting a successful source judgement. The findings therefore suggest that associative recognition memory shares some of the recollective processes that are engaged by the requirement to retrieve contextual information about a study episode. The findings from Experiment 2 indicate that the processes associated with the recollection of associated pairs are engaged regardless of whether the retrieval of associative information is an explicit task requirement.
Article
Event-related potentials (ERPs) were recorded while subjects performed a memory retrieval task requiring old/new judgements to visually presented old (previously studied) and new words. For words judged old, subjects made two binary forced-choice context (hereafter source) judgements, denoting the voice (male/female) and task (action/liking) with which the test word had been associated at study. By separating the ERPs according to the accuracy of the voice and task judgements, it was possible to test the prediction that the differences between ERPs to correctly identified old and new words at parietal scalp sites (parietal old/new effects) are sensitive to the amount or quality of information that is retrieved from episodic memory (Rugg, M.D., Cox, C.J.C., Doyle, M.C., Wells, T., 1995. Event-related potentials and the recollection of low and high frequency words. Neuropsychologia 33, 471-484). In keeping with this proposal, the magnitude of the parietal old/new effects co-varied with the number of accurate source judgements. This finding is consistent with proposals that the parietal old/new effect indexes recollection in a graded fashion.
Article
Spatial selective attention and spatial working memory have largely been studied in isolation. Studies of spatial attention have provided clear evidence that observers can bias visual processing towards specific locations, enabling faster and better processing of information at those locations than at unattended locations. We present evidence supporting the view that this process of visual selection is a key component of rehearsal in spatial working memory. Thus, although working memory has sometimes been depicted as a storage system that emerges 'downstream' of early sensory processing, current evidence suggests that spatial rehearsal recruits top-down processes that modulate the earliest stages of visual analysis.
Article
How do we find a target item in a visual world filled with distractors? A quarter of a century ago, in her influential 'Feature Integration Theory (FIT)', Treisman proposed a two-stage solution to the problem of visual search: a preattentive stage that could process a limited number of basic features in parallel and an attentive stage that could perform more complex acts of recognition, one object at a time. The theory posed a series of problems. What is the nature of that preattentive stage? How do serial and parallel processes interact? How does a search unfold over time? Recent work has shed new light on these issues.
Article
Under incidental learning conditions, spatial layouts can be acquired implicitly and facilitate visual search (contextual-cueing effect). We examined whether the contextual-cueing effect is specific to the visual modality or transfers to the haptic modality. The participants performed 320 (experiment 1) or 192 (experiment 2) visual search trials based on a typical contextual-cueing paradigm, followed by haptic search trials in which half of the trials had layouts used in the previous visual search trials. The visual contextual-cueing effect was obtained in the learning phase. More importantly, the effect was transferred from visual to haptic searches; there was greater facilitation of haptic search trials when the spatial layout was the same as in the previous visual search trials, compared with trials in which the spatial layout differed from those in the visual search. This suggests the commonality of spatial memory to allocate focused attention in both visual and haptic modalities.
Article
Humans process a visual display more efficiently when they encounter it for a second time, showing learning of the display. This study tests whether implicit learning of complex visual contexts depends on attention. Subjects searched for a white target among black and white distractors. When the locations of the target and the attended set (white distractors) were repeated, search speed was enhanced, but when the locations of the target and the ignored set (black distractors) were repeated, search speed was unaffected. This suggests that the expression of learning depends on attention. However, during the transfer test, when the previously ignored set now was attended, it immediately facilitated performance. In contrast, when the previously attended set now was ignored, it no longer enhanced search speed. We conclude that the expression of visual implicit learning depends on attention but that latent learning of repeated information does not.
Article
The electrophysiological correlates of recollection were investigated with a modified Remember/Know task in which subjects signaled whether they fully or partially recollected visual object information in each study episode. A positive-going ERP deflection--the left parietal old/new effect--was sensitive to the amount of information recollected, demonstrating greater amplitude when elicited by test items associated with full relative to partial recollection. These findings support prior proposals that the left parietal ERP old/new effect is sensitive to the amount of information recollected from episodic memory. An early-onsetting (ca. 150 ms), left frontal old/new effect differentiated items accorded correct old versus correct new responses regardless of whether the items were endorsed as familiar or recollected. This finding extends the range of circumstances under which early, frontally distributed old/new effects occur, and adds weight to previous suggestions that these effects are a neural correlate of familiarity-driven recognition memory.