Peter Hagoort’s research while affiliated with Radboud University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (497)


Figure 2: Summary of the behavioural performance. Points on the left of the violin plot represent individual participants' performance. Points on the violin plot represent the mean, error bars standard error of the mean. Voice performance shows how many acceptable sentences were produced in each condition. Word position reports accuracy values in the memory of words for each position in a sentence.
Figure 3: Whole-brain results for the generalization performance of the bag-of-nouns model, broad roles and narrow roles model in production and comprehension. All results are thresholded at p < 0.005, p < 0.05 FDR cluster corrected.
Figure 4: Generalization performance for the bag-of-nouns model (that separates the semantic categories of musician and athlete), the broad roles model and the narrow roles model, in six brain regions. LATL: left anterior temporal lobe. LPTL = left posterior temporal lobe. LTPJ: left temporoparietal junction. LIFGtri: pars triangularis of the left inferior frontal gyrus. LIFGop: pars opercularis of the left inferior frontal gyrus. LPreC: left precentral gyrus.
Figure 5: Generalization performance for the bag-of-verbs model that includes of all verbs and the narrow roles model, in six brain regions. LATL: left anterior temporal lobe. LPTL = left posterior temporal lobe. LTPJ: left temporo-parietal junction. LIFGtri: pars triangularis of the left inferior frontal gyrus. LIFGop: pars opercularis of the left inferior frontal gyrus. LPreC: left precentral gyrus.
Figure 6: Generalization performance for the categorical bag-of-nouns model (that separates the semantic categories of musician and athlete), the specific bag-of-nouns model and the nounspecifical broad roles model, in six brain regions. LATL: left anterior temporal lobe. LPTL = left
Neural Encoding of Semantic Structures During Sentence Production
  • Preprint
  • File available

November 2024

·

3 Reads

Laura Giglio

·

Peter Hagoort

·

The neural representations for compositional processing have so far been mostly studied during sentence comprehension. In an fMRI study of sentence production, we investigated the brain representations for compositional processing during speaking. We used a rapid serial visual presentation sentence recall paradigm to elicit sentence production from the conceptual memory of an event. With voxel-wise encoding models, we probed the specificity of the compositional structure built during the production of each sentence, comparing an unstructured model of word meaning without relational information with a model that encodes abstract thematic relations and a model encoding event-specific relational structure. Whole-brain analyses revealed that sentence meaning at different levels of specificity was encoded in a large left fronto-parieto-temporal network. A comparison with semantic structures composed during the comprehension of the same sentences showed similarly distributed brain activity patterns. An ROI analysis over left fronto-temporal language parcels showed that event-specific relational structure above word-specific information was encoded in the left inferior frontal gyrus (LIFG). Overall, we found evidence for the encoding of semantic meaning during sentence production in a distributed brain network and for the encoding event-specific semantic structures in the LIFG.

Download


Social Anxiety Alters Theory of Mind Activation and Intersubject Neural Variability During Movie Viewing

November 2024

·

14 Reads

Margot Mangnus

·

Saskia B. J. Koch

·

Robin Devillers

·

[...]

·

Social anxiety is characterized by an intense fear of judgment in social situations, yet the underlying mechanisms driving this condition remain poorly understood. One hypothesis holds that specific alterations in Theory of Mind (ToM) affect the ability to interpret others’ thoughts and emotions. Another hypothesis proposes that broader interpretive biases lead individuals to perceive social cues as overly significant, even in neutral settings. We investigated these possibilities by measuring brain activity, pupil responses, and heart rates in socially anxious individuals and matched controls as they viewed ‘Partly Cloudy’, an animated film known to engage the ToM network during specific scenes. While overall brain activity during ToM-related scenes was similar across groups, socially anxious participants exhibited reduced activation in the left posterior superior temporal sulcus (pSTS), a key area for ToM processing. Additionally, intersubject correlation analysis revealed a distinct neural response pattern in the socially anxious group, marked by uniform responses in sensory regions and heightened variability in higher-order cortical areas. This pattern persisted throughout the film and occurred without changes in heart rate or pupil responses, indicating a neural processing bias that manifests even in non-evaluative settings. These findings provide a neural basis for ToM alterations and broader interpretive biases in social anxiety, supporting cognitive-behavioral models and suggesting novel targets for intervention.


Perception of Emotions in Human and Robot Faces: Is the Eye Region Enough?

October 2024

·

33 Reads

The increased interest in developing next-gen social robots has raised questions about the factors affecting the perception of robot emotions. This study investigates the impact of robot appearances (humanlike, mechanical) and face regions (full-face, eye-region) on human perception of robot emotions. A between-subjects user study (N = 305) was conducted where participants were asked to identify the emotions being displayed in videos of robot faces, as well as a human baseline. Our findings reveal three important insights for effective social robot face design in Human-Robot Interaction (HRI): Firstly, robots equipped with a back-projected, fully animated face - regardless of whether they are more human-like or more mechanical-looking - demonstrate a capacity for emotional expression comparable to that of humans. Secondly, the recognition accuracy of emotional expressions in both humans and robots declines when only the eye region is visible. Lastly, within the constraint of only the eye region being visible, robots with more human-like features significantly enhance emotion recognition.


Relevance of Prosodic Focus and Lexical Stress for Discourse Comprehension in Turkish: Evidence from Psychometric and Electrophysiological Data

October 2024

·

32 Reads

Journal of Cognitive Neuroscience

Prosody underpins various linguistic domains ranging from semantics and syntax to discourse. For instance, prosodic information in the form of lexical stress modifies meanings and, as such, syntactic contexts of words as in Turkish kaz-má “pickaxe” (noun) versus káz-ma “do not dig” (imperative). Likewise, prosody indicates the focused constituent of an utterance as the noun phrase filling the wh-spot in a dialogue like What did you eat? I ate−−−−. In the present study, we investigated the relevance of such prosodic variations for discourse comprehension in Turkish. We aimed at answering how lexical stress and prosodic focus mismatches on critical noun phrases—resulting in grammatical anomalies involving both semantics and syntax and discourse-level anomalies, respectively—affect the perceived correctness of an answer to a question in a given context. To that end, 80 native speakers of Turkish, 40 participating in a psychometric experiment and 40 participating in an EEG experiment, were asked to judge the acceptability of prosodic mismatches that occur either separately or concurrently. Psychometric results indicated that lexical stress mismatch led to a lower correctness score than prosodic focus mismatch, and combined mismatch received the lowest score. Consistent with the psychometric data, EEG results revealed an N400 effect to combined mismatch, and this effect was followed by a P600 response to lexical stress mismatch. Conjointly, these results suggest that every source of prosodic information is immediately available and codetermines the interpretation of an utterance; however, semantically and syntactically relevant lexical stress information is assigned more significance by the language comprehension system compared with prosodic focus information.



Nonlinear dendritic integration supports Up-Down states in single neurons

September 2024

·

8 Reads

Changes in the activity profile of cortical neurons are due to phenomena at the scale of local and long-range networks. Accordingly, the states of cortical neurons and their, often abrupt, transitions – a phenomenon known as Up/Down states – are attributed to variations in the afferent neurons’ activity. However, cellular physiology and morphology may also play a role. This study examines the impact of dendritic nonlinearities, in the form of voltage-gated NMDA receptors, on the response of cortical neurons to balanced excitatory/inhibitory synaptic inputs. Using a neuron model with two segregated dendritic compartments, we compare cells with and without dendritic nonlinearities. Our analysis shows that NMDA receptors boost somatic firing in the balanced condition and increase the correlation of membrane potentials across the three compartments of the neuron model. Then we introduce controlled fluctuations in excitatory inputs and quantify the ensuing bimodality of the somatic membrane potential. We show that dendritic nonlinearities are crucial for detecting these fluctuations and initiating Up-Down states whose shape and statistics closely resemble electrophysiological data. Our results provide new insights into the mechanisms underlying cortical bistability and highlight the complex interplay between dendritic integration and network dynamics in shaping neuronal behavior. Significance statement In several physiological states, such as sleep or quiet wakefulness, the membrane of cortical cells shows a stereotypical bistability. The cell is either fully depolarized and ready to spike or in a silent, hyperpolarized state. This dynamics, known as Up-Down states, has often been attributed to changes in the network activity. However, whether cell-specific properties, such as dendritic nonlinearity, have a role in driving the neuron’s bistability remains unclear. This study examines the issue using a model of a pyramidal cell and reveals that the presence of dendritic NMDA receptors, drives the up-down states in response to small fluctuations in the network activity.



Fig. 1. Main effect of Congruency-N400 time window. ERP waveforms of the congruent (C+; black) and incongruent (C−; red) conditions at the critical word (onset at 0 ms) for the main effect of Congruency (Top Left), and when the critical word was in focus (F+; Top Right) or not in focus (F−; Bottom Left). All waveforms diverge from around 330 to 690 ms after the onset of the critical word, and the topographical distribution of the main effect (Bottom Right, first topography; centro-parietal maximum) is highly similar to the difference observed when the critical word was in focus (Bottom Right, second topography) and when it was not (Bottom Right, third topography). Negative is plotted up; shaded regions in the waveforms indicate SEM; orange shaded rectangles indicate time windows corresponding to statistically significant effects; scalp topographies depict the mean amplitude over the time interval corresponding to the effects; The waveforms represent the average of the electrodes that contribute to the N400 effect identified in the cluster-based permutation statistics; black filled circles in the scalp plots indicate electrodes for a statistically significant effect.
Fig. 2. Main effect of Congruency-Post-N400 time window. ERP waveforms of the congruent (C+; black) and incongruent (C−; red) conditions at the critical word (onset at 0 ms) for the main effect of Congruency (Top Left), and when the critical word was in focus (F+; Top Right) or not in focus (F−; Bottom Left). All waveforms diverge from around 835 ms to the end of the time window of interest. Negative is plotted up; shaded regions in the waveforms indicate SEM; orange shaded rectangles indicate time windows corresponding to statistically significant effects; teal shaded rectangles indicate time windows corresponding to marginal effects; scalp topographies depict the mean amplitude over the time interval corresponding to the effects; The waveforms represent the average of the electrodes that contribute to the effect in the cluster-based permutation statistics; black stars in the scalp plots indicate electrodes for a marginal effect-filled circles indicate electrodes for a statistically significant effect.
Fig. 3. Main effect of Focus. ERP waveforms of the in focus (F+; green) and not in focus (F−; purple) conditions at the critical word (onset at 0 ms) for the main effect of Focus (Top Left), and when the critical word was semantically congruent with the preceding sentence context (C+; Top Right) or not (C−; Bottom Left). All waveforms diverge from around 285 to 690 ms after the onset of the critical word. Negative is plotted up; shaded regions in the waveforms indicate SEM; orange shaded rectangles indicate time windows corresponding to statistically significant effects; scalp topographies depict the mean amplitude over the time interval corresponding to the effects; The waveforms represent the average of the electrodes that contribute to the cluster identified in the cluster-based permutation statistics; black filled circles in the scalp plots indicate electrodes for a statistically significant effect.
Information structure in Makhuwa: Electrophysiological evidence for a universal processing account

July 2024

·

88 Reads

Proceedings of the National Academy of Sciences

There is evidence from both behavior and brain activity that the way information is structured, through the use of focus, can up-regulate processing of focused constituents, likely to give prominence to the relevant aspects of the input. This is hypothesized to be universal, regardless of the different ways in which languages encode focus. In order to test this universalist hypothesis, we need to go beyond the more familiar linguistic strategies for marking focus, such as by means of intonation or specific syntactic structures (e.g., it-clefts). Therefore, in this study, we examine Makhuwa-Enahara, a Bantu language spoken in northern Mozambique, which uniquely marks focus through verbal conjugation. The participants were presented with sentences that consisted of either a semantically anomalous constituent or a semantically nonanomalous constituent. Moreover, focus on this particular constituent could be either present or absent. We observed a consistent pattern: Focused information generated a more negative N400 response than the same information in nonfocus position. This demonstrates that regardless of how focus is marked, its consequence seems to result in an upregulation of processing of information that is in focus.


Connectivity of Fronto-Temporal Regions in Syntactic Structure Building During Speaking and Listening

July 2024

·

5 Reads

·

1 Citation

Neurobiology of Language

The neural infrastructure for sentence production and comprehension has been found to be mostly shared. The same regions are engaged during speaking and listening, with some differences in how strongly they activate depending on modality. In this study, we investigated how modality affects the connectivity between regions previously found to be involved in syntactic processing across modalities. We determined how constituent size and modality affected the connectivity of the pars triangularis of the left inferior frontal gyrus (LIFG) and of the left posterior temporal lobe (LPTL) with the pars opercularis of the LIFG, the anterior temporal lobe (LATL) and the rest of the brain. We found that constituent size reliably increased the connectivity across these frontal and temporal ROIs. Connectivity between the two LIFG regions and the LPTL was enhanced as a function of constituent size in both modalities, and it was upregulated in production possibly because of linearization and motor planning in the frontal cortex. The connectivity of both ROIs with the LATL was lower and only enhanced for larger constituent sizes, suggesting a contributing role of the LATL in sentence processing in both modalities. These results thus show that the connectivity among fronto-temporal regions is upregulated for syntactic structure building in both sentence production and comprehension, providing further evidence for accounts of shared neural resources for sentence-level processing across modalities.


Citations (65)


... With respect to the cognitive neuroscience of syntax, the distinction between vertical and horizontal syntactic information has been supported via direct cortical recordings, being partitioned into closely neighboring portions of frontotemporal cortex and licensing distinct parsing strategies to isolate (Murphy et al. 2024b; see also Lopopolo et al. 2021). Posterior temporal cortex is increasingly being seen here as the critical code in the core syntax network (Biondo et al. 2024;Fahey et al. 2024;Matchin et al. 2024;Matchin & Hickok 2020;Yu et al. 2024), and is the earliest region to jointly resolve vertical and horiztonal dimensions (Murphy et al. 2024b). ...

Reference:

Shadow of the (Hierarchical) Tree: Reconciling Symbolic and Predictive Components of the Neural Code for Syntax
Multiple dimensions of syntactic structure are resolved earliest in posterior temporal cortex
  • Citing Article
  • September 2024

Progress in Neurobiology

... For instance, the same stem can be pronounced with different tones (marked with superscript 1 and 2) depending on the suffix that follows, as seen in the words 1 mugg-en 'the mug' and 2 mugg-ar 'mugs' [1,2]. Recent behavioral results suggest that these tone accents might also be involved in word comprehension [3]. Their role in lexicosemantic processing in the brain, however, remains unexplored. ...

The role of pitch accent in discourse comprehension and the markedness of Accent 2 in Central Swedish

... Instead, we suggest a rigorous dissection 9 of, for example, auditory 14 , lexical 13 , and task-modulated aspects 62 of this frequency-tagged readout. A dissociation between some of these factors has been found with neuro-computational language models in the temporalspatial domain 19,[63][64][65] . We believe that it will be helpful to extend this line of research into the frequency domain. ...

Neural dynamics express syntax in the time domain during natural story listening

... From a critical standpoint, these and any theory of language emergence and evolution should not ignore the nature of multimodal communication and should discuss within their theoretical framework the strong synchrony between the vocal and manual articulators in spoken and signed interactions. There is an enormous amount of empirical work showing that speech and gesture are tightly unified in naturalistic conversations (Kelly et al., 2014) and that the brain supports the processing and integration of spoken and visual information in language-specific regions (Hagoort & Özyürek, 2024). In the realm of sign languages, there is a wealth of work demonstrating that mouthings and mouth actions are tightly aligned with the manual component of signs and denote critical phonological and lexical information (Bank, 2015;Lewin & Schembri, 2011). ...

Extending the Architecture of Language From a Multimodal Perspective Extending the Architecture of Language From a Multimodal Perspective  CITE DETAILS RELATIONS

Topics in Cognitive Science

... In contrast to the wealth of knowledge the field has accumulated about single word production, very little is known about the type of speech that is unique to our species: sentences [17][18][19]. While some work has overcome the obstacles to studying production with non-invasive neural measures [e.g., 15,[17][18][19][20][21][22][23][24][25][26][27], these studies have largely focused on differences between comprehension and production rather than drawing on progress in word production research. ...

Diverging neural dynamics for syntactic structure building in naturalistic speaking and listening
  • Citing Article
  • March 2024

Proceedings of the National Academy of Sciences

... Giani et al. used two frequencies to tag visual stimulation (size-or luminance-modulated) and auditory stimulation (frequency-or amplitude-modulated), and they found no IM response with simultaneous stimulation on the two senses (Giani et al., 2012). However, Drijvers et al. (2021) and Seijdel et al. (2024) employed two frequencies to separately tag visual (gesture) and auditory (speech) stimulation of a video. When the visual and auditory stimulation were matched (e.g., simultaneous presentation of the action of driving and the audio of the word "drive"), there was a clear IM response. ...

Attention Drives Visual Processing and Audiovisual Integration During Multimodal Communication

The Journal of Neuroscience : The Official Journal of the Society for Neuroscience

... The emotional dialogue generation task aims to imbue conversational agents with human-like emotions, enabling them to recognize, understand, and express emotions, providing users with more personalized and diverse responses. Emotional dialogue generation primarily involves two stages: emotion recognition in conversation [14][15][16] and emotional expression in dialogue [17][18][19]. The first step is to identify the emotions implied in the user's current dialogue, and then generate content that accurately expresses those emotions. ...

Real-time emotion generation in human-robot dialogue using large language models

Frontiers in Robotics and AI

... Traditional tests examine language with repetitive, standardized and isolated tasks, missing social interactions and multimodal data [65]. For example, combining VR, EEG and eye-tracking in naturalistic environments such as supermarket and restaurant, can help predict language comprehension and identify neurophysiological markers [67]. ...

Combining EEG and 3D-eye-tracking to study the prediction of upcoming speech in naturalistic virtual environments: A proof of principle
  • Citing Article
  • November 2023

Neuropsychologia

... Word skipping has been studied extensively in the context of sentence constraint (see Staub, 2015), word length (see Brysbaert & Vitu, 1998), and lexical frequency (see Kliegl et al., 2004); however, up to this point, there has been no way of measuring how word skipping relates to lexical processing and semantic access. Because one of the best predictors of word skipping is word length, even when controlling for lexical frequency (Kliegl et al., 2004;Heilbron et al., 2023), some researchers have argued that skipping is primarily driven by oculomotor constraints (e.g., saccade targeting based on the spatial layout of words in a sentence; e.g., McConkie & Yang, 2003a, 2003bVitu, 2003) rather than deeper lexical processing (Brysbaert & Vitu, 1998). However, the fact that we found a larger N400 response to anomalies when a word is skipped (in high-constraint sentences) suggests that word skipping is associated with more extensive lexical processing during parafoveal preview. ...

Lexical Processing Strongly Affects Reading Times But Not Skipping During Natural Reading

Open Mind

... Another very recent methodological advance, named RIFT, has been used to study the integration of auditory and visual inputs in the brain in more naturalistic contexts (Drijvers et al., 2021;Seijdel et al., 2022Seijdel et al., , 2023. Using this technique, researchers can "tag" a certain signal at a specific high frequency by modulating its luminance (for visual signals) or amplitude (for auditory signals). ...

Attention drives visual processing and audiovisual integration during multimodal communication