ArticlePDF Available

Abstract

Gaze direction, a cue of both social and spatial attention, is known to modulate early neural responses to faces e.g. N170. However, findings in the literature have been inconsistent, likely reflecting differences in stimulus characteristics and task requirements. Here, we investigated the effect of task on neural responses to dynamic gaze changes: away and toward transition (resulting or not in eye contact). Subjects performed, in random order, social (away/toward them) and non-social (left/right) judgment tasks on these stimuli. Overall, in the non-social task, results showed a larger N170 to gaze aversion than gaze motion toward the observer. In the social task, however, this difference was no longer present in the right hemisphere, likely reflecting an enhanced N170 to gaze motion toward the observer. Our behavioral and ERP data indicate that performing social judgments enhances saliency of gaze motion toward the observer, even those that did not result in gaze contact. These data and that of previous studies suggest two modes of processing visual information: a 'Default mode' that may focus on spatial information; a 'Socially Aware mode' that might be activated when subjects are required to make social judgments. The exact mechanism that allows switching from one mode to the other remains to be clarified. © The Author (2015). Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.
1
Social decisions affect neural activity to perceived dynamic gaze
Marianne Latinus
1,2
, Scott A. Love
1,2
, Alejandra Rossi
1
, Francisco Parada
1
, Lisa Huang
1
,
Laurence Conty
3
, Nathalie George
,4,5,6,7
, Karin James
1
, Aina Puce
1,*
1
Psychological and Brain Sciences, Indiana University, Bloomington IN, USA,
2
Institut de Neuroscience de la Timone, UMR7289, CNRS-Aix-Marseille Université,
Marseille, France.
3
Laboratory of Psychopathology and Neuropsychology (LPN, EA2027), Paris 8 University,
Saint-Denis, France.
4
Inserm, U 1127 et Centre MEG-EEG, Paris, France,
5
Université Pierre et Marie Curie-Paris 6, UM 75, Paris, France.
6
CNRS, U7225 et Centre MEG-EEG, Paris, France
7
Institut du Cerveau et de la Moelle Epinière (ICM), Social and Affective Neuroscience Lab
and Centre MEG-EEG - CENIR, Paris, France
*Corresponding author:
Aina Puce, PhD,
Psychological and Brain Sciences,
Indiana University
1101 East 10
th
St,
Bloomington, IN 47401, USA
Tel: 1 812 650 2213
FAX: 1 812 855 4691
Email: ainapuce@indiana.edu
Running title: Social decisions affect neural activity
© The Author (2015). Published by Oxford University Press. For Permissions, please email:
journals.permissions@oup.com
Social Cognitive and Affective Neuroscience Advance Access published April 28, 2015
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
2
Abstract
Gaze direction, a cue of both social and spatial attention, is known to modulate early neural
responses to faces e.g. N170. However, findings in the literature have been inconsistent,
likely reflecting differences in stimulus characteristics and task requirements. Here, we
investigated the effect of task on neural responses to dynamic gaze changes: away and
toward transition (resulting or not in eye contact). Subjects performed, in random order,
social (away/toward them) and non-social (left/right) judgment tasks on these stimuli.
Overall, in the non-social task, results showed a larger N170 to gaze aversion than gaze
motion toward the observer. In the social task, however, this difference was no longer
present in the right hemisphere, likely reflecting an enhanced N170 to gaze motion toward
the observer. Our behavioral and ERP data indicate that performing social judgments
enhances saliency of gaze motion toward the observer, even those that did not result in
gaze contact. These data and that of previous studies suggest two modes of processing
visual information: a ‘Default mode’ that may focus on spatial information; a ‘Socially Aware
mode’ that might be activated when subjects are required to make social judgments. The
exact mechanism that allows switching from one mode to the other remains to be clarified.
Keywords: direct gaze, averted gaze, N170, task modulation, social and non-social context
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
3
Introduction
As social primates, we continually monitor the behaviors of others so that we can
appropriately respond in a social interaction. Our ability to do that depends critically on
decoding our visual environment, including important information carried by the face, the
eyes and includes gaze changes. An individual’s gaze direction transmits a wealth of
information not only as to their focus of spatial attention, but also about their intention to
approach or withdraw, therefore conveying both visuospatial and social information to the
observer. In this respect, it is important to note crucial differences between direct and
averted gaze. Direct gaze mainly signals that the observer is the likely recipient of a directed
behavior, and is indicative of the intention to start a communicative interaction; thus direct
gaze mainly conveys social information to the observer (Senju and Johnson, 2009). On the
contrary, averted gaze transmits both social and spatial information to the observer. On the
one hand, gaze cueing experiments indicate that averted gaze serves as a powerful stimulus
for altering the observer’s focus of visuospatial attention; yet there appears to be a
difference in how the brain treats visual cues consisting of eyes versus arrows. Lesions to
the right superior temporal sulcus or the amygdala disrupted gaze, but not arrow cueing
(Akiyama et al., 2006; Akiyama et al., 2007). In contrast to arrows, averted gaze also
conveys a range of social meanings including for instance, shyness, dishonesty, the
intentionality of the gazer, and their emotional state (Adams and Kleck, 2005; Fox, 2005;
Calder et al., 2007). Consequently, modulations of brain activity by gaze direction have
been accounted for by either a change in social (Puce and Schroeder, 2010; Caruana et al.,
2014) and/or visuospatial attention (Grossmann et al., 2007; Hadjikhani et al., 2008;
Straube et al., 2010). This suggests that gaze processing may be sensitive to task-based
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
4
manipulations of participants’ attention towards either a social or a spatial dimension.
Consistently, the task being performed by participants is known to be increasingly
important in the processing of social stimuli such as gaze or facial expressions (Graham and
Labar, 2012).
Not surprisingly, it is believed that the human brain possesses specialized mechanisms for
the processing of gaze and other important information conveyed by the eyes (Langton et
al., 2000; Itier and Batty, 2009). Indications of specialized processes dedicated to the
perception of gaze come from functional magnetic resonance imaging (fMRI) and event-
related potential (ERP) studies (Puce et al., 1998; Wicker et al., 1998; Puce et al., 2000;
George et al., 2001). Notably, the face-sensitive N170 (Bentin et al., 1996) also shows
sensitivity to static eyes, typically being larger and later for eyes shown in isolation (e.g.,
Bentin et al., 1996; Itier et al., 2006; Nemrodov and Itier, 2011). Furthermore, the N170
response to eyes matures more rapidly than that of faces (Taylor et al., 2001a), leading
researchers to describe the N170 as a potential early marker of eye gaze processing (Taylor
et al., 2001a; Taylor et al., 2001b; Itier et al., 2006; Nemrodov and Itier, 2011). Interestingly,
while in 4-month-old infants early brain activity is greater to gaze contact than to averted
gaze (Farroni et al., 2002), in adults, the modulation of N170 by gaze direction seems to
vary as a function of task demand and stimulus (Puce et al., 1998; Conty et al., 2007;
Ponkanen et al., 2011). Indeed, studies that have measured N170 modulations by gaze
direction in adults reported no consistent results (Itier and Batty, 2009). Some studies
reported a larger N/M170 to averted gaze (Puce et al., 2000; Watanabe et al., 2002; Itier et
al., 2007a; Caruana et al., 2013), some to direct gaze (Conty et al., 2007; Ponkanen et al.,
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
5
2011) while others reported no modulations of the N170 by gaze direction (Taylor et al.,
2001b; Schweinberger et al., 2007; Brefczynski-Lewis et al., 2011; Myllyneva and Hietanen,
2015). Inconsistent results in the study of gaze perception are also reported in fMRI
(Calder et al., 2007; Nummenmaa and Calder, 2009). Discrepancies between studies have
been attributed mainly to task and stimulus factors (Itier and Batty, 2009 11; Nummenmaa
and Calder, 2009; Puce and Schroeder, 2010). Gaze perception studies have used either
passive viewing tasks (Puce et al., 2000; Watanabe et al., 2002; Caruana et al., 2013), or
“social” judgment tasks, where participants report whether the gaze was oriented away or
toward them (Conty et al., 2007; Itier et al., 2007a).
Moreover, gaze perception studies have used a diversity of stimuli with varying head
orientation, either front-view (e.g., Puce et al., 2000), or/and ¾-viewed (e.g., Kawashima et
al., 1999; Conty et al., 2007; Itier et al., 2007a), and varying angles of gaze deviations (from
5 to 30 degrees; e.g., Schweinberger et al., 2007). Importantly, the majority of studies
manipulated gaze in static displays, even though gaze is rarely static in natural situations,
and social information important for non-verbal communication is often conveyed via
dynamic gaze changes. The use of dynamic stimuli can pose a challenge in
neurophysiological studies because they may not have clear onsets, and can potentially
elicit a continuous and dynamic neural response (see Ulloa et al., 2012). To overcome this
problem, apparent face motion stimuli, which allow eliciting clear ERPs to dynamic
stimulation, were developed (Puce et al., 2000; Conty et al., 2007). Apparent face motion
stimuli have a precise stimulus onset for performing traditional ERP analyses, while
conserving the dynamic and more ecological aspects of perception.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
6
The purpose of the current study was to investigate whether the modulations of the N170
by gaze direction depend on task demands, using a varied set of eye gaze transitions. To
that aim, we used an apparent motion paradigm in a trial structure identical to Conty et al.
(2007) and a subset of their stimuli. We generated a series of 6 potential viewing
conditions: three motions away from the participants and three motions toward the
participants. The 6 conditions included full gaze transition between an extreme and a direct
gaze, mimicking conditions used in Puce et al. (2000), two conditions starting at an
intermediate gaze position, mimicking those of Conty et al. (2007) and an additional 2
conditions, ending on the intermediate gaze position, to ensure a balanced stimulus design.
Importantly, we ran two task versions on the same subject group using the same stimuli in
the same experimental session. In Task 1 subjects made a ‘social’ judgment identifying if the
gaze moved toward or away from them, as in Conty et al. (2007). In Task 2, subjects made a
‘non-social’ judgment where they indicated if the gaze change was to their left or right. We
thus explicitly examined how ERPs to viewing eye gaze changes were influenced by the task
performed by the participants. We hypothesized that gaze transition away from the
participants will lead to larger N170 than gaze motion toward the participants, at least in
the non-social task, regardless of the size of the gaze transition.
1. Materials and methods
1.1 Subjects
Overall, 26 subjects from the general Indiana University (Bloomington, USA) community
took part in the experiment. All provided written informed consent in a study that was
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
7
approved by the Indiana University Institutional Review Board (IRB 1202007935). All
subjects were paid $US25 for their participation. Four individuals generated
electroencephalographic (EEG) data that contained excessive head or eye movement
artifacts, and hence were excluded from subsequent data analysis. Therefore, a total of 22
subjects (11 female; mean age±SD: 26.23±3.44 years) contributed data to this study. All, but
one, subjects were right handed (mean handedness±SD: +54.77±31.94), as assessed by the
Edinburgh Handedness Inventory (Oldfield, 1971). All subjects were free from a history of
neurological or psychiatric disorders and had normal, or corrected-to-normal, vision.
1.2 Stimuli
Stimuli consisted of the frontal face views from Conty et al. (2007) that were presented
using an identical trial structure. A total of forty 8-bit RGB color frontal view faces (20
males) were presented with direct gaze (Direct), 15° (Intermediate) or 30° (Extreme)
averted gaze positions. There were a total of six images per face: one with a direct gaze, one
with an averted horizontal gaze of 15° to the right, one with an averted horizontal gaze of
30° to the right, and mirror images of each. Apparent gaze motion was created from the
static images by presenting two images sequentially (Fig. 1A).
A total of 6 gaze apparent motion conditions were generated (see Fig. 1A, B):
(1) Direct to extreme gaze [Dir-Ext];
(2) Extreme to direct gaze [Ext-Dir];
(3) Intermediate to extreme [Int-Ext];
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
8
(4) Intermediate to direct [Int-Dir];
(5) Direct to intermediate [Dir-Int];
(6) Extreme to intermediate [Ext-Int].
The large number of conditions in this study allows us to reconcile potential differences
between previous published studies. In the current study, conditions (1) and (2) were
identical to those previously studied in Puce et al. (2000), whereas (3) and (4) were a
subset of those used in Conty et al. (2007). Conditions (5) and (6) were not used in either of
the previous studies, but were added to the current study so that a balanced experimental
design could be created. For the sake of brevity, throughout the manuscript we refer to
these groupings of pairs of conditions subsequently as “Full Transition”, “Intermediate-to-
Endpoint”, and “Endpoint-to-Intermediate”. Note that conditions (1), (3) and (5)
correspond to gaze transition made away from the subjects, while conditions (2), (4) and
(6) correspond to gaze transition made toward the subjects, resulting (2,4) or not (6) in eye
contact.
1.3 Design
Each subject completed two tasks in a recording session: in the Social Task subjects pressed
one of two response buttons to indicate whether the viewed gaze transition was moving
Away or Toward them. In the Non-social Task, a gaze transition was judged relatively as
either moving towards either their Left or their Right. The order of the two tasks was
counterbalanced across subjects. For each task, 480 trials were broken up into four runs of
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
9
120 trials each to allow rest for subjects between runs, so that they could keep their face
and eye movements to a minimum.
A single trial consisting of the presentation of two stimuli had the following structure: the
first image of each trial was presented for 800, 900 or 1000ms (randomized) on a black
background. It was immediately replaced by a second image, which differed from the first
one only by its gaze direction, creating an apparent motion stimulus. The second image
remained on the screen for 1100ms. Trials were separated by an 800ms white fixation
cross appearing on a black background (Fig. 1A). Each of the six stimulus conditions was
presented a total of 80 times in randomized order, for a total of 480 trials per task.
1.4 Data acquisition
Each subject was fitted with a 256-electrode HydroCel Geodesic Sensor Net (EGI Inc.,
Eugene, OR, USA). Electrodes were adjusted as needed to keep impedances below 60KΩ,
consistent with manufacturer’s recommendations. Half way through the experimental
session an additional impedance check was performed, and impedances were adjusted as
needed. Continuous EEG recordings were made during both tasks using a gain of 5000 with
a set of EGI Net Amps 300 neurophysiological amplifiers using NetStation 4.4 data
acquisition software and were stored for off-line analysis. EEG data were recorded with
respect to the vertex using a sampling rate of 500 Hz and a band pass filter of 0.1-200 Hz.
Once the EEG set-up was complete, subjects sat in a comfortable chair in a darkened room
2.75m away from a 160cm monitor (Samsung SyncMaster P63FP, Refresh Rate of 60Hz)
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
10
mounted on a wall at eye level. Stimuli were presented with a visual angle of 7.0x8.6 deg
(horizontal X vertical) using Presentation V14 software (Neurobehavioral Systems, San
Francisco, CA).
1.5 Data analysis
1.5.1 Behavioral data
Response time and accuracy data, collected with Presentation, were exported to Matlab
2012 (The Mathworks Inc., Natick, MA, USA). Mean response times and accuracy were
calculated for each condition, task and subject.
1.5.2 EEG data preprocessing
EEG data preprocessing was performed in NetStation EEG software (EGI Inc., Eugene, OR,
USA) and EEGlab (Delorme and Makeig, 2004), following recommended guidelines; detailed
information is presented in supplementary material. Continuous EEG data were epoched
into 1.6s epochs, including a 518ms pre-stimulus (the 2
nd
face of the apparent motion
stimulus) onset
1
.
1.5.3 Event-related Potentials (ERPs)
For each task, an average ERP was generated for each subject and condition; average
number of trials included in the average ERP were greater than 60 in all conditions and
1
An advisory notice from the EGI EEG system manufacturer has informed us about an 18ms delay between
real-time acquisition (to which events are synchronized) and the EEG signal. Consequently, a post-hoc latency
factor of 18ms was applied to all ERP latencies.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
11
tasks (repeated-measures ANOVAs, all p>0.05). ERP peak analyses were conducted on
individual subject averages for each of the 12 conditions (6 apparent motion x 2 tasks).
N170 latencies and amplitudes were measured from the ERPs averaged over a 9-electrode
cluster (Fig 1C), centred on the electrode where the grand average (collapsed for
conditions) was maximal between 142 and 272ms post-stimulus. In order to investigate
ERP effects other than the N170, we further tested for experimental effects at all time-
points and electrodes using the LIMO EEG toolbox (spatial-temporal analyses, Pernet et al.,
2011 - presented in Supplementary Material)
1.6 Statistical Analysis
Statistical analyses of behavioral data and ERP peak amplitudes and latencies were
performed using IBM SPSS Statistics V20 (IBM Corp., Armonk, NY, USA).
1.6.1 Behavioral data
To directly compare our results with those of previous studies, we ran three separate 2-way
repeated-measure ANOVAs on the data in this current study explicitly comparing the
conditions identical to Puce et al. 2000 (Dir-Ext/Ext-Dir), Conty et al. 2007 (Int-Ext/Int-Dir)
and new conditions not previously tested (Ext-Int/Dir-Int – Fig. 1B); these ANOVAs are
respectively referred to as the Full transition ANOVA, Intermediate-to-Endpoint ANOVA or
Endpoint-to-Intermediate ANOVA. All three ANOVAs had 2 within-subjects factors: Task
and Condition (Dir-Ext/Ext-Dir or Int-Ext/Int-Dir or Dir-Int/Ext-Int). A significant effect
was identified at the p<0.05; significant interactions were further explored using paired t-
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
12
test. Results of an omnibus ANOVA with all conditions are presented in Supplementary
Material.
1.6.2 Event-related potentials (ERPs)
A mixed-design ANOVA using within-subject factors of Task, Condition and Hemisphere and
a between-subjects factor of gender was performed to identify significant differences in
N170 amplitude and latency. A significant effect was identified at the p<0.05 level using a
Greenhouse-Geisser sphericity correction, when relevant; pairwise comparisons were
Bonferroni corrected. Statistical analyses of N170 latency are reported in supplementary
material.
Using logic similar to the analysis of the behavioral data, we additionally ran three separate
3-way ANOVAs comparing the condition subgroups i.e. the Full transition ANOVA,
Intermediate-to-Endpoint ANOVA and Endpoint-to-Intermediate ANOVA. All three ANOVAs
had 3 within-subjects factors: Task, Condition and Hemisphere (Left, Right). A significant
effect was identified at the p<0.05 level; significant interactions were further explored using
paired t-test.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
13
2. Results
2.1 Behavioral results
2.1.1 Accuracy
Accuracy results are displayed in Figure 2A. The Full Transition ANOVA, i.e., comparison
between direct-to-extreme and extreme-to-direct gaze changes, revealed an effect of task
(F(1,21)=5.34; p=0.031; η
2
=0.20), and a two-way interaction (F(1,21)=12.13; p=0.002;
η=0.37). Subjects performed better in the non-social task than in the social task, in
particular for gaze aversions. In the non-social task, accuracy was higher for gaze aversion
than gaze changes toward the participants (t(21)=3.82; p=0.001); in the social task,
accuracy was not modulated by gaze transition direction (t(21)=-0.751; p=0.461). The
Intermediate-to-Endpoint ANOVA, i.e., comparisons between Int-Ext and Int-Dir, again
showed a main effect of task (F(1,21)=4.58; p=0.044; η
2
=0.17): accuracy was better in the
non-social task than in the social task, regardless of the gaze direction. No other effect or
interaction was found. Finally, the Endpoint-to-Intermediate ANOVA, namely comparisons
between Ext-Int and Dir-Int changes, showed a main effect of task (F(1,21)=7.48; p=0.012;
η
2
=0.26), condition (F(1,21)=5.49; p=0.029; η
2
=0.21) and a two-way interaction
(F(1,21)=7.20; p=0.014; η
2
=0.26). Again, accuracy differed as a function of gaze transition
direction in the non-social task (higher for gaze aversion; t(21)=2.82; p=0.01) but not in the
social task (t(21)=0.44; p=0.666).
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
14
2.1.2 Response Times
Response time (RT) data are displayed in Figure 2B. The Full Transition ANOVA, revealed
main effects of task (F(1,21)=33.42; p<0.001; η
2
=0.61), condition (F(1,21)=19.09; p<0.001;
η
2
=0.48) and an interaction (F(1,21) =7.65; p=0.012; η
2
=0.267). RTs were overall shorter in
the non-social than in the social task, and, in the social task, for gaze changes made toward
the subjects. The two-way interaction between task and condition revealed that RTs were
significantly shorter for toward gaze transition in the social task (t(21)=4.64; p<0.001), but
less so in the non-social task (t(21)=2.76; p=0.012). The Intermediate-to-Endpoint ANOVA
revealed main effects of task (F(1,21)=30.81; p<0.001; η
2
=0.59) and condition
(F(1,21)=20.62; p<0.001; η
2
=0.49). Again RTs were shorter in the non-social task and for
eye gaze changes toward the subjects. Finally, the Endpoint-to-Intermediate ANOVA
showed effects similar to that of the Full Transition ANOVA: main effect of task
(F(1,21)=31.94; p<0.001; η
2
=0.60), condition (F(1,21)=12.83; p=0.002; η
2
=0.38) and an
interaction (F(1,21)=5.58; p=0.028; η
2
=0.21). RTs were faster in the non-social task, and for
gaze changes made toward the subjects. RTs were only significantly faster to toward gaze
transition in the social task (t(21)=4.07; p=0.001; non-social task: t(21)=1.56; p=0.13).
2.2 Event Related Potentials (ERPs)
Overall, a very clear triphasic positive-negative-positive ERP complex was observed in all
conditions at the posterior scalp bilaterally (Figs. 3-5). The negative ERP corresponded to
the N170, consistent with previous studies (Puce et al., 2000; Conty et al., 2007) and peaked
around 200ms. So as to better compare with previous studies, averaged ERPs recorded for
9-electrode occipitotemporal clusters over each hemisphere for all conditions were
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
15
generated (Figs. 3–5, S1). For all conditions, N170 was earlier (see Supplementary Material)
and larger in the right electrode cluster. Below we describe the results of statistical tests on
ERP amplitudes as a function of task and condition.
2.2.1 N170 Amplitude: Omnibus ANOVA
Histograms of N170 amplitudes and latencies as a function of condition are presented in
Supplementary Materials (Fig. S1). Our ANOVA had factors of Task, Condition, Hemisphere,
and Gender (or participant). Participants’ gender affected N170 amplitude differently for
the two tasks (Task x Gender interaction: F(1,20)=5.577; p=0.028; η
2
=0.218). There was no
difference in N170 amplitude between male and female participants in the social task;
however, N170 was larger in male participants in the non-social task. The omnibus mixed-
factor ANOVA also revealed main effects of condition (F(4.09,81.83)=13.017; p<0.001;
η
2
=0.394), hemisphere (F(1,20)=15.515; p=0.001; η
2
=0.437) and a three-way interaction
between task, condition, and hemisphere (F(4.05,81.11)= 5.125; p=0.001; η
2
=0.204).
N170 was also larger to eye motion made toward an extreme position, i.e., away from the
observer (Dir-Ext and Int-Ext changes, which did not differ), compared to eye motion made
from an extreme position, i.e., toward the observer (Ext-Dir/Ext-Int, which did not differ).
Other eye gaze motion directions (Dir-Int/Int-Dir) led to N170 with intermediate
amplitudes. The three-way interaction between task, condition and hemisphere was further
explored by running two 2-way repeated-measure ANOVAs: one per hemisphere. In the left
hemisphere (LH) there was a main effect of condition (F(4.09,86.08)=11.31; p<0.001;
η
2
=0.35): N170 was larger for motion away from the participants which did not differed
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
16
significantly (Dir-Ext, Int-Ext, Dir-Int; all p>0.29); it was the smallest for motion toward the
participants (Ext-Int and Ext-Dir, which did not differed significantly). In the right
hemisphere (RH), there was a main effect of condition (F(4.21,88.35)=7.65; p<0.001;
η
2
=0.27) and, a significant interaction between task and condition (F(78.67,3.74)=3.83;
p=0.008; η
2
=0.15). Overall, N170 amplitudes were larger for gaze aversion to an extreme
averted position than for gaze motion ending with gaze contact; motion toward an
intermediate position led to the smallest N170, with a larger N170 to the motion away from
the observer (Dir-Int).
We further explored the interaction in RH, by running a one-way ANOVA for each task: both
revealed a significant effect of condition (non-social task: F(3.58,75.07)=6.46; p<0.001;
η
2
=0.24; social task: F(3.53,73.99)=6.09; p<0.001; η
2
=0.23). In the non-social task, N170
were larger for motion away from the participants, and smaller for motion toward the
participant. In the social task, N170 amplitudes were the smallest for gaze changes toward
an intermediate position: gaze aversion toward an intermediate averted gaze (Dir-Int) yield
the smallest N170. Consequently, while in the non-social task, in RH, N170 amplitude
appears larger for gaze transition away from the subjects, in the social task, the pictures is
less clear cut, with the N170 being the smallest for an away condition (Dir-Int).
2.2.2 N170 amplitude: ANOVAs on condition groupings based on previous studies
In the Full Transition ANOVA (Fig. 3), overall N170 was larger in the right hemisphere
(F(1,20)=14.98; p=0.001; η
2
=0.43) and for gaze aversions (Dir-Ext; F(1,20)=13.70; p=0.001;
η
2
=0.41). A three-way interaction between task, condition, and hemisphere (F(1,20)=4.69;
p=0.042; η
2
=0.19) indicated that, while the away condition evoked a larger N170 than the
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
17
toward condition in both hemispheres when subjects were involved in a non-social
judgment (LH: t(21)=-2.79; p=0.011; RH: t(21)=-2.88; p=0.009), this difference disappeared
in RH in the social task (LH: t(21)=-5.30; p<0.001; RH: t(21)=-0.96; p=0.35). This absence of
difference in RH during the social judgments was attributed to an enhanced N170 to the
stimuli showing a gaze change toward the subjects. A Task by Gender interaction
(F(1,20)=4.75; p=0.041; η
2
=0.19) revealed a larger N170 for male subjects than female
subjects in the social task, while no differences were observed in the non-social task.
The Intermediate-to-Endpoint ANOVA (Fig. 4) again showed a larger N170 over the right
hemisphere (F(1,20)=14.74; p=0.001; η
2
=0.42) and for gaze changes away from subjects
(F(20, 1)=7.54; p=0.012; η
2
=0.27). A three-way interaction between task, condition, and
hemisphere (F(20, 1)=10.35; p=0.004; η
2
=0.34) showed that the modulation of N170
amplitude by gaze direction was significant in RH, while subjects were involved in the non-
social task (t(21)=-2.78, p=0.011), but not in the social task (t(21)=-1.10, p=0.284), likely
reflecting an enhanced N170 amplitude for gaze changes toward subjects. In LH, the
opposite was true: the difference between conditions was not significant during the non-
social task (t(21)=-0.11, p=0.916), but N170 was significantly smaller for gaze changes
toward the subjects in the social task (t(21)=-3.30, p=0.003). There was no interaction
between task and gender (F(1,20)=3.10; p=0.094).
Finally, the Endpoint-to-Intermediate ANOVA (Fig. 5) again showed that N170 was larger in
RH (F(1,20)=14.80; p=0.001; η
2
=0.43), and for the motion away from subjects
(F(1,20)=11.72; p=0.003; η
2
=0.37). As observed in the Full Transition ANOVA, a 3-way
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
18
interaction between task, condition, and hemisphere (F(1,20)=4.70; p=0.042; η
2
=0.19)
revealed was no difference in N170 amplitude evoked by away and toward gaze changes in
RH, when subjects performed a social judgment (t(21)=1.116; p=0.28). A Task by Gender
interaction (F(1,20)=6.70; p=0.018; η
2
=0.25) revealed a larger N170 for male subjects than
female subjects in the social task, and no differences were observed in the non-social task.
3. Discussion
Our aim in this study was to test if brain activity to gaze changes was sensitive to task
demands, as mainly signaled by N170 ERP characteristics. This question is important given
inconsistencies found in the gaze perception literature irrespective of whether EEG/MEG or
fMRI was used as the imaging modality (Itier and Batty, 2009 11; Nummenmaa and Calder,
2009). Differences between studies could have arisen from differences in task requirements
or stimuli used (e.g., size of the gaze transition). Hence, here we studied the same group of
subjects as we varied task (social vs. non-social judgment) and used the same trial structure
with a subset of stimuli previously used in Conty et al. (2007), and analogous stimulus
conditions used in two previous studies. Our data showed clear main effects of task, gaze
direction (away/toward) and interaction effects between these two variables. These effects
were present irrespective of gaze position onset and the size of the gaze transition, and
irrespective of whether the motion toward the viewer involved eye contact.
Discrepancies between studies may potentially be explained by differences in the degree of
gaze aversion, which ranged from 5 to 30 degrees (Puce et al., 2000; Conty et al., 2007;
Schweinberger et al., 2007). In our current study, gaze aversion was produced with
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
19
different degrees, nonetheless when all conditions were compared in a single statistical
analysis, N170 amplitude was not modulated by motion transition size. N170 amplitudes
were not significantly different for conditions with the same direction of motion, but with
different degrees of motion excursion. Rather, the N170 response pattern was more
“categorical”, indicating that the observed modulations in N170 amplitude occurred when
gaze changed to look (further) away from the observer. This observation is consistent with
an fMRI study showing sensitivity of the anterior superior temporal sulcus to overall gaze
direction that was independent from gaze angle (Calder et al., 2007).
We performed separate statistical analyses on the current data, based on groups of
conditions that were used in each of our previous studies (Puce et al., 2000; Conty et al.,
2007), so as to enable interpretation of observed differences. We replicated the findings of
Puce et al. (2000) where extreme gaze aversions elicited larger N170s relative to direct
gaze in a passive viewing task (using the non-social task here as a comparison). For non-
social judgments, we observed similar N170 amplitude modulation effects for all away and
all toward conditions, irrespective of the starting/ending position of the gaze transition. Gaze
direction facilitates target detection by directing attention toward the surrounding space
(Itier and Batty, 2009). Larger N170s to averted gaze could reflect a shift of attention
toward the surrounding space, cueing the observer to a potentially more behaviorally
relevant part of visual space. Spatial cueing has previously been shown to modulate early
ERPs (P1/N1; Holmes et al., 2003; Jongen et al., 2007); notably, the N170 is enhanced for
cued/attended targets (McDonald et al., 2003; Pourtois et al., 2004; Carlson and Reinke,
2010). Thus, our results suggest an increased salience of spatial cueing over social
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
20
processing, at least with the frontal face views used here, in situations where no explicit
social judgment needs to be made. It should be noted that these cueing effects occur earlier
in time relative to other ERP effects related to spatial cueing seen in the literature. Two
types of known ERP negativity elicited to spatial cueing in Posner-like (Posner et al., 1980)
visuo-spatial cueing paradigms: the posterior Early Directing Attention Negativity (EDAN),
and the Anterior Directing Attention Negativity (ADAN) (Harter et al., 1989; Yamaguchi et
al., 1994; Hopf and Mangun, 2000; Nobre et al., 2000), typically occur in the 200-400ms and
300-500ms range, respectively.
Surprisingly, task had no effect on the N170 measured in the left hemisphere: left N170s
were always larger for gaze aversion, irrespective of whether a social or non-social decision
was being made, consistent with previous reports (Hoffman and Haxby, 2000; Itier et al.,
2007b; Caruana et al., 2013). Notably, however, in the social task, the modulations of N170s
by gaze aversion disappeared in the right hemisphere: N170s were not significantly
different between conditions due to the occurrence of an enhanced N170 for gaze
transitions toward the participants. This effect was seen for all gaze transitions toward
participants, regardless of whether the gaze change ended with direct eye contact (and
independent of reference electrode see Supplementary Materials). Our results are
consistent with those of Itier and collaborators (2007b) who used a social task with static
face onset displays; they report larger N170 to averted than direct gaze, for front-view faces
only, and mainly in the left hemisphere. However, our results are inconsistent with Conty et
al., 2007 who reported significantly larger N170s to direct gaze than to gaze aversion, in
particular for deviated head view. It should be noted that these two studies included stimuli
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
21
with different head positions in addition to gaze changes, and it is possible that these
additional conditions may have further modulated N170 activity (e.g., see Itier et al.,
2007b).
Head orientation has been shown to modulate gaze perception in behavioral paradigms
(Langton et al., 2000). In particular, incongruence between head and gaze direction can
decrease participant’s performance in spatial judgments (Langton et al., 2000). Thus,
including different head views may have put greater emphasis on the processing of gaze
direction in Conty et al. (2007), because the gaze transitions were displayed under different
configurations and therefore judging their direction (away/toward them) required
deepened processing of the eye region. Following Conty et al.’s hypothesis, our results
indicate that, in the case of explicit social judgments, N170 modulations reflect processes of
toward motion transition, even though the observer does not necessarily become the focus
of attention. It should be noted that defining social attention is particularly difficult and the
definition varies between research groups: some scientists will make the argument that our
tasks are actually both social as there is a human face in all conditions, for others both our
tasks may be considered as spatial. It is possible that task differences could reflect a self-
referential processing effect, as it the spatial computation may seem to be made relatively
to the self (in the “social” task), or absolutely (non-social task). However, we believe that
the left/right judgment could also be regarded as self-centered as it is dependent on the
participant’s right and left.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
22
Overall our observations are consistent with previous studies showing that N170 is
modulated by top-down influences (Bentin and Golland, 2002; Jemel et al., 2003; Latinus
and Taylor, 2006). Our spatial-temporal analysis showed that ERP effects that were
modulated by task began as early as 148ms and lasted for 300ms. This modulation was
mostly seen on frontal electrodes, in line with the idea that task engages top-down
influences arising from higher order regions. The modulation of an ERP as a function of
social decision might be indicative of whether the brain is in a ‘Socially Aware’ mode or not.
In the non-social task, involving visuospatial judgments, we observed larger N170s to
transitions away from the observer, irrespective of whether the transition was made from
gaze contact. This augmented N170 to gaze aversion replicates previous results observed
under passive viewing (Puce et al., 2000; Caruana et al., 2013) suggesting that in the case of
viewing eye gaze, the implicit working mode of the brain may not be social. Changes in gaze
direction, in particular aversion from a direct gaze position, can signal a shift in spatial
attention toward a specific location, inducing the observer to shift his/her attention toward
the same location (Puce and Perrett, 2003; Hadjikhani et al., 2008; Straube et al., 2010). Our
data seems to indicate that the ‘Default’ mode of the brain might be to process gaze as an
indication of a shift in attention toward a specific visuospatial location. In contrast, in an
explicit ‘Socially Aware’ mode, the N170 is augmented in conditions with toward gaze
transitions. Our data suggest that the social meaning of direct gaze arises from being in an
explicit social mode where social context is the most salient stimulus dimension. This study
raises questions regarding what exactly is an explicit social context. A study by Pönkanen et
al. (2011) reported larger N170 to direct than averted gaze with real persons, but not with
photography; yet, the same team further demonstrated that these effects were dependent
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
23
on the mental attribution from the observer (Myllyneva and Hietanen, 2015). Taken
together with our results, this may suggest than in ecological situations, such as in a face-to-
face conversation, the presence of real faces rather than photographs may be sufficient in
generating a social context, allowing the switching of processing mode from a default to an
explicit socially aware mode.
Converging with the proposal that gaze processing is non-social by default, participants had
more accurate performance and faster response times in the non-social task. As stated
previously, gaze direction is thought to facilitate target detection – target detection is more
likely to occur in the non-social task, than in the social task in which the observer is the
target of a directed behavior. Alternatively, social judgment may result in a slower
attentional disengagement from the face leading to slower RTs and increased error rates in
the social task (Itier et al., 2007b). Indeed, slower RTs have been reported for faces with
direct gaze, and were thought to reflect an enhanced processing of faces with direct gaze
(Vuilleumier et al., 2005), or a slower attentional disengagement from faces with direct gaze
(Senju and Hasegawa, 2005). Consistent with this explanation, RTs for ‘towards’ gaze
transitions in the social task were faster than gaze aversion, which may reflect a smaller
attentional disengagement in gaze change toward the participants in the social task. These
smaller RTs were consistent with our neurophysiological data, in that social judgments
produced relatively enhanced N170 amplitudes to gaze transitions directed toward the
subject.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
24
We have made the claim here that our data are consistent with a ‘socially aware’ and ‘non-
social’ bias for information processing in the brain, and that this bias can be conferred via
top-down mechanisms (e.g. task demand) or via bottom-up mechanisms (e.g. redeployment
of visuospatial attention). An alternative possibility is that non-social mechanisms related
to the re-allocation of visual attention to another part of visual space – a possibility that
requires future testing. Studies that systematically investigate the hypothesis that social
context influences the perception of gaze transition direction are needed. One way of
providing an explicitly social context might be to include multisensory stimulation, where
auditory cues such as non-speech and speech vocalizations and other (non-human)
environmental sounds might be presented as a face generates gaze transitions away from
the viewer. This would also make the stimuli more ecologically valid. Multisensory cues
with explicitly social and non-social dimensions could potentially differentiate between
different ‘top-down’ processing modes (including social and non-social ones). So that this
issue can be teased out participants would have to complete a series of tasks, where they
might focus on one particular type of auditory cue at a given time. If our claim that
communicative intent in incoming stimuli switches the brain into a “socially aware” mode is
true, then we should observe the largest N170 to gaze transitions accompanied by speech
stimuli, with those to non-speech vocalizations being smaller, and those to environmental
sounds being smaller still when subjects are explicitly given a social task e.g. detect a word
target such as hello for example.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
25
4. Conclusions
This dataset reconciles data from two studies on apparent gaze motion with seemingly
opposite results. We report modulations of the N170 by gaze transition direction
dependent on the task performed by the subjects. N170 was larger to gaze aversions in both
hemispheres when subjects were involved in a non-social task, mimicking previous results
with passive viewing, suggesting that the brain’s Default mode may not be ‘social’. Focusing
subjects’ attention to social aspects of the stimuli, by requiring explicit social judgments, led
to an enhanced N170 to toward gaze transitions in the right hemisphere, irrespective of the
ending position of the gaze motion. This could reflect an increased salience of toward gaze
motion in a ‘social’ context, and the brain operating in a ‘Socially Aware’ mode.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
26
5. References
Adams, R. B., Jr. & Kleck, R. E. (2005) 'Effects of direct and averted gaze on the perception of
facially communicated emotion', Emotion, 5(1), pp. 3-11.
Akiyama, T., Kato, M., Muramatsu, T., Saito, F., Umeda, S. & Kashima, H. (2006) 'Gaze but not
arrows: a dissociative impairment after right superior temporal gyrus damage',
Neuropsychologia, 44(10), pp. 1804-1810.
Akiyama, T., Kato, M., Muramatsu, T., Umeda, S., Saito, F. & Kashima, H. (2007) 'Unilateral
amygdala lesions hamper attentional orienting triggered by gaze direction', Cereb
Cortex, 17(11), pp. 2593-2600.
Bentin, S., Allison, T., Puce, A., Perez, E. & McCarthy, G. (1996) 'Electrophysiological Studies
of Face Perception in Humans', Journal of Cognitive Neuroscience, 8, pp. 551-565.
Bentin, S. & Golland, Y. (2002) 'Meaningful processing of meaningless stimuli: the influence
of perceptual experience on early visual processing of faces', Cognition, 86(1), pp.
B1-14.
Brefczynski-Lewis, J. A., Berrebi, M. E., McNeely, M. E., Prostko, A. L. & Puce, A. (2011) 'In the
blink of an eye: neural responses elicited to viewing the eye blinks of another
individual', Front Hum Neurosci, 5, p. 68.
Calder, A. J., Beaver, J. D., Winston, J. S., Dolan, R. J., Jenkins, R., Eger, E. & Henson, R. N.
(2007) 'Separate coding of different gaze directions in the superior temporal sulcus
and inferior parietal lobule', Curr Biol, 17(1), pp. 20-25.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
27
Carlson, J. M. & Reinke, K. S. (2010) 'Spatial attention-related modulation of the N170 by
backward masked fearful faces', Brain Cogn, 73(1), pp. 20-27.
Caruana, F., Cantalupo, G., Lo Russo, G., Mai, R., Sartori, I. & Avanzini, P. (2014) 'Human
cortical activity evoked by gaze shift observation: an intracranial EEG study', Hum
Brain Mapp, 35(4), pp. 1515-1528.
Caruana, F., Cantalupo, G., Russo, G. L., Mai, R., Sartori, I. & Avanzini, P. (2013) 'Human
cortical activity evoked by gaze shift observation: An intracranial EEG study', Hum
Brain Mapp.
Conty, L., N'Diaye, K., Tijus, C. & George, N. (2007) 'When eye creates the contact! ERP
evidence for early dissociation between direct and averted gaze motion processing',
Neuropsychologia, 45(13), pp. 3024-3037.
Delorme, A. & Makeig, S. (2004) 'EEGLAB: an open source toolbox for analysis of single-trial
EEG dynamics including independent component analysis', Journal of neuroscience
methods, 134(1), pp. 9-21.
Farroni, T., Csibra, G., Simion, F. & Johnson, M. H. (2002) 'Eye contact detection in humans
from birth', Proc Natl Acad Sci U S A, 99(14), pp. 9602-9605.
Fox, E. (2005) 'The role of visual processes in modulating social interactions', Visual
Cognition, 12(1), pp. 1-11.
George, N., Driver, J. & Dolan, R. J. (2001) 'Seen gaze-direction modulates fusiform activity
and its coupling with other brain areas during face processing', Neuroimage, 13(6 Pt
1), pp. 1102-1112.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
28
Graham, R. & Labar, K. S. (2012) 'Neurocognitive mechanisms of gaze-expression
interactions in face processing and social attention', Neuropsychologia, 50(5), pp.
553-566.
Grossmann, T., Johnson, M. H., Farroni, T. & Csibra, G. (2007) 'Social perception in the infant
brain: gamma oscillatory activity in response to eye gaze', Soc Cogn Affect Neurosci,
2(4), pp. 284-291.
Hadjikhani, N., Hoge, R., Snyder, J. & de Gelder, B. (2008) 'Pointing with the eyes: the role of
gaze in communicating danger', Brain Cogn, 68(1), pp. 1-8.
Harter, M. R., Miller, S. L., Price, N. J., Lalonde, M. E. & Keyes, A. L. (1989) 'Neural processes
involved in directing attention', J Cogn Neurosci, 1(3), pp. 223-237.
Hoffman, E. A. & Haxby, J. V. (2000) 'Distinct representations of eye gaze and identity in the
distributed human neural system for face perception', Nat Neurosci, 3(1), pp. 80-84.
Holmes, A., Vuilleumier, P. & Eimer, M. (2003) 'The processing of emotional facial
expression is gated by spatial attention: evidence from event-related brain
potentials', Brain Research Cogn Brain Research, 16(2), pp. 174-184.
Hopf, J. M. & Mangun, G. R. (2000) 'Shifting visual attention in space: an electrophysiological
analysis using high spatial resolution mapping', Clin Neurophysiol, 111(7), pp. 1241-
1257.
Itier, R. J., Alain, C., Kovacevic, N. & McIntosh, A. R. (2007a) 'Explicit versus implicit gaze
processing assessed by ERPs', Brain Res, 1177, pp. 79-89.
Itier, R. J. & Batty, M. (2009) 'Neural bases of eye and gaze processing: the core of social
cognition', Neurosci Biobehav Rev, 33(6), pp. 843-863.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
29
Itier, R. J., Latinus, M. & Taylor, M. J. (2006) 'Face, eye and object early processing: what is
the face specificity?', Neuroimage, 29(2), pp. 667-676.
Itier, R. J., Villate, C. & Ryan, J. D. (2007b) 'Eyes always attract attention but gaze orienting is
task-dependent: evidence from eye movement monitoring', Neuropsychologia, 45(5),
pp. 1019-1028.
Jemel, B., Pisani, M., Calabria, M., Crommelinck, M. & Bruyer, R. (2003) 'Is the N170 for faces
cognitively penetrable? Evidence from repetition priming of Mooney faces of familiar
and unfamiliar persons', Brain Research Cogn Brain Research, 17(2), pp. 431-446.
Jongen, E. M., Smulders, F. T. & Van der Heiden, J. S. (2007) 'Lateralized ERP components
related to spatial orienting: discriminating the direction of attention from processing
sensory aspects of the cue', Psychophysiology, 44(6), pp. 968-986.
Kawashima, R., Sugiura, M., Kato, T., Nakamura, A., Hatano, K., Ito, K., Fukuda, H., Kojima, S.
& Nakamura, K. (1999) 'The human amygdala plays an important role in gaze
monitoring. A PET study', Brain, 122 ( Pt 4), pp. 779-783.
Langton, S. R., Watt, R. J. & Bruce, I. I. (2000) 'Do the eyes have it? Cues to the direction of
social attention', Trends Cogn Sci, 4(2), pp. 50-59.
Latinus, M. & Taylor, M. J. (2006) 'Face processing stages: impact of difficulty and the
separation of effects', Brain Res, 1123(1), pp. 179-187.
McDonald, J. J., Teder-Salejarvi, W. A., Di Russo, F. & Hillyard, S. A. (2003) 'Neural substrates
of perceptual enhancement by cross-modal spatial attention', Journal of Cognitive
Neuroscience, 15(1), pp. 10-19.
Myllyneva, A. & Hietanen, J. K. (2015) 'There is more to eye contact than meets the eye. ',
Cognition, 134, pp. 100–109.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
30
Nemrodov, D. & Itier, R. J. (2011) 'The role of eyes in early face processing: a rapid
adaptation study of the inversion effect', Br J Psychol, 102(4), pp. 783-798.
Nobre, A. C., Gitelman, D. R., Dias, E. C. & Mesulam, M. M. (2000) 'Covert visual spatial
orienting and saccades: overlapping neural systems', Neuroimage, 11(3), pp. 210-
216.
Nummenmaa, L. & Calder, A. J. (2009) 'Neural mechanisms of social attention', Trends Cogn
Sci, 13(3), pp. 135-143.
Oldfield, R. C. (1971) 'The assessment and analysis of handedness: the Edinburgh
inventory', Neuropsychologia, 9(1), pp. 97-113.
Pernet, C. R., Chauveau, N., Gaspar, C. & Rousselet, G. A. (2011) 'LIMO EEG: a toolbox for
hierarchical LInear MOdeling of ElectroEncephaloGraphic data', Comput Intell
Neurosci, 2011, p. 831409.
Ponkanen, L. M., Alhoniemi, A., Leppanen, J. M. & Hietanen, J. K. (2011) 'Does it make a
difference if I have an eye contact with you or with your picture? An ERP study', Soc
Cogn Affect Neurosci, 6(4), pp. 486-494.
Posner, M. I., Snyder, C. R. & Davidson, B. J. (1980) 'Attention and the detection of signals', J
Exp Psychol, 109(2), pp. 160-174.
Pourtois, G., Grandjean, D., Sander, D. & Vuilleumier, P. (2004) 'Electrophysiological
correlates of rapid spatial orienting towards fearful faces', Cereb Cortex, 14(6), pp.
619-633.
Puce, A., Allison, T., Bentin, S., Gore, J. C. & McCarthy, G. (1998) 'Temporal cortex activation
in humans viewing eye and mouth movements', J Neurosci, 18(6), pp. 2188-2199.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
31
Puce, A. & Perrett, D. (2003) 'Electrophysiology and brain imaging of biological motion',
Philos Trans R Soc Lond B Biol Sci, 358(1431), pp. 435-445.
Puce, A. & Schroeder, C. E. (2010) 'Multimodal Studies Using Dynamic Faces', in C. Curio, H.
H. Bulthoff & M. A. Giese (eds), Dynamic Faces: Insights from Experiments and
Computation, Cambridge, MA, USA, MIT Press.
Puce, A., Smith, A. & Allison, T. (2000) 'ERPs evoked by viewing facial movements.', Cog
Neuropsychol, 17, pp. 221-239.
Schweinberger, S. R., Kloth, N. & Jenkins, R. (2007) 'Are you looking at me? Neural
correlates of gaze adaptation', Neuroreport, 18(7), pp. 693-696.
Senju, A. & Hasegawa, T. (2005) 'Direct gaze captures visuospatial attention', Visual
Cognition, 12(1), pp. 127-144.
Senju, A. & Johnson, M. H. (2009) 'The eye contact effect: mechanisms and development',
Trends Cogn Sci, 13(3), pp. 127-134.
Straube, T., Langohr, B., Schmidt, S., Mentzel, H. J. & Miltner, W. H. (2010) 'Increased
amygdala activation to averted versus direct gaze in humans is independent of
valence of facial expression', Neuroimage, 49(3), pp. 2680-2686.
Taylor, M. J., Edmonds, G. E., McCarthy, G. & Allison, T. (2001a) 'Eyes first! Eye processing
develops before face processing in children', Neuroreport, 12(8), pp. 1671-1676.
Taylor, M. J., Itier, R. J., Allison, T. & Edmonds, G. E. (2001b) 'Direction of gaze effects on
early face processing: eyes-only versus full faces', Brain Research Cogn Brain
Research, 10(3), pp. 333-340.
Ulloa, J. L., Puce, A., Hugueville, L. & George, N. (2012) 'Sustained neural activity to gaze and
emotion perception in dynamic social scenes', Soc Cogn Affect Neurosci.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
32
Vuilleumier, P., George, N., Lister, V., Armony, J. & Driver, J. (2005) 'Effects of perceived
mutual gaze and gender on face processing and recognition memory', Visual
Cognition, 12(1), pp. 85-101.
Watanabe, S., Miki, K. & Kakigi, R. (2002) 'Gaze direction affects face perception in humans',
Neurosci Lett, 325(3), pp. 163-166.
Wicker, B., Michel, F., Henaff, M. A. & Decety, J. (1998) 'Brain regions involved in the
perception of gaze: a PET study', Neuroimage, 8(2), pp. 221-227.
Yamaguchi, S., Tsuchiya, H. & Kobayashi, S. (1994) 'Electroencephalographic activity
associated with shifts of visuospatial attention', Brain, 117 ( Pt 3), pp. 553-562.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
33
TABLE
Table 1: Behavioral Results (group mean +/- standard errors of the mean (s.e.m.)) as a
function of Task and Condition.
Dir
-
Ext
Ext
-
Dir
Int
-
Ext
Int
-
Dir
Dir
-
Int
Ext
Int
Accuracy (%) ±
±±
± s.e.m.
Non-
Social
98.35
± 0.79
95.17
± 1.33
96.53
± 1.04
96.02
± 1.18
97.78
± 0.76
93.24
± 2.17
Social
94.15
± 0.97
94.83
± 1.16
93.52
± 0.95
94.20
± 1.27
92.50
± 1.22
92.10
± 1.31
RTs (ms) ±
±±
± sem
Non
-
Social 434 ± 16 412 ± 19 466 ± 14 443 ± 18 440 ± 16 429 ± 20
Social 535 ± 24 489 ± 23 563 ± 19 523 ± 23 545 ± 24 512 ± 24
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
34
FIGURE CAPTIONS
Figure 1: Methods. A. Time line for individual trial structure. A first static face is
displayed on the screen. Gaze direction in that first face can be direct (illustrated),
intermediate or extreme. The first face is then replaced by a second static face, in which the
gaze direction, different from gaze direction in the first face, can be direct, intermediate or
extreme (illustrated) in order to create apparent gaze motion. Subjects were instructed to
respond while the second face was still on screen. B. Example of the different apparent
motion conditions. The light gray box highlights gaze transition made away from the
subjects. The dark gray box highlights gaze transition made toward the subjects. C.
Location of the electrodes of interest (red dots). Data illustrated is the average ERPs
across conditions at the latency of the N170.
Figure 2: Behavioral Results. Red bars: non-social task; blue bars: social task. Light
colored bars illustrate gaze transition made away from the subjects. Dark colored bars
highlight transition made toward the subjects. Error bars represent standard error of mean
(s.e.m). A. Accuracy. *: p< 0.05. Note that accuracy is plotted between 80 and 100%. B.
Response Times. *: p< 0.01.
Figure 3: Grand average ERPs for the Full Transitions (Direct-to-Extreme and
Extreme-to-Direct – conditions similar to Puce et al. (2000)). Top panels, red lines: non-
social task; bottom panels, blue lines: social task. Light colored lines illustrate gaze
transition made away from the subjects. Dark colored lines highlighted transition made
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
35
toward the subjects. Shaded areas represent the 95% confidence interval built using
bootstrap (n=1000) with replacement of the data under H1.
Figure 4: Grand average ERPs for the Intermediate-to-Endpoint transitions (IE and ID
– conditions similar to Conty et al. (2007)). Top panels, red lines: non-social task; bottom
panels, blue lines: social task. Light colored lines illustrate gaze transition made away from
the subjects. Dark colored lines highlighted transition made toward the subjects. Shaded
areas represent the 95% confidence interval built using bootstrap (n=1000) with
replacement of the data under H1.
Figure 5: Grand average ERPs for the Endpoint-to-Intermediate transitions (Direct-
to-Intermediate and Extreme-to-Intermediate). Top panels, red lines: non-social task;
bottom panels, blue lines: social task. Light colored lines illustrate gaze transition made
away from the subjects. Dark colored lines highlighted transition made toward the subjects.
Shaded areas represent the 95% confidence interval built using bootstrap (n=1000) with
replacement of the data under H1.
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
110x72mm (300 x 300 DPI)
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
75x112mm (300 x 300 DPI)
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
150x120mm (300 x 300 DPI)
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
150x120mm (300 x 300 DPI)
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
150x120mm (300 x 300 DPI)
at University of Glasgow on May 22, 2015http://scan.oxfordjournals.org/Downloaded from
... Larger amplitudes for N170 have been found in a number of studies in response to faces with averted gaze as compared to direct gaze, that is, when eyes appear to look directly at the participant/observer (Caruana et al., 2014;Itier et al., 2007;Latinus et al., 2015;Puce et al., 2000;Rossi et al., 2015). There are also reports of larger N170 amplitudes for direct as compared to averted gaze (Conty et al., 2007(Conty et al., , 2012Watanabe et al., 2006). ...
... In line with previous findings, such as by Itier et al. (2007) for both initial gaze position and change gazes, by Latinus et al. (2015) for dynamic gaze changes, and by Stephani et al. (2020) for gaze-contingent stimulus presentations, N170 amplitude was larger when the eyes were looking away from the observers than when aiming at them. The N170 is interpreted as reflecting the structural encoding of faces (Eimer, 2000). ...
... Although the interaction of gaze and hemisphere did not reach significance, scalp topographies showed that a larger N170 for gaze aversion was observed in the right than in the left hemisphere. This resembles the findings of Latinus et al. (2015) in a social task and several other studies (for review see Eimer, 2011), which found a larger gaze effect in the right hemisphere. ...
Thesis
Full-text available
The eye gaze and its direction are important and relevant non-verbal cues for the establishment of social interactions and the perception of others’ emotional facial expressions. Gaze direction itself, whether eyes are looking straight at the viewer (direct gaze) or whether they look away (averted gaze), affects our social attention and emotional response. This implies that both emotion and gaze have informational values, which might interact at early or later stages of neurocognitive processing. Despite the suggestion of a theoretical basis for this interaction, the shared signal hypothesis (Adams & Kleck, 2003), there is a lack of structured electrophysiological investigations into the interactions between emotion and gaze and their neural correlates, and how they vary across populations. Addressing this need, the present doctoral dissertation used event-related brain potentials (ERPs) to study responses to emotional expressions and gaze direction in a novel paradigm combining static and dynamic gaze with facial expressions. The N170 and EPN were selected as ERP components believed to reflect gaze perception and reflexive attention, respectively. Three different populations were investigated. Study 1, in a normal sample, investigated the amplitudes of the ERP components elicited by the initial presentation of faces and subsequent changes of gaze direction in half of the trials. In Study 2, based on the atypical face processing and diminished responses to eye gaze in autism, the ERPs and eye movements were examined in two samples of children varying in the severity of their autism traits. In Study 3, In a large sample, I addressed the putatively increased sensitivity in emotion processing and response to eye gaze in mothers during their postpartum period with a particular focus on infant's faces. Taken together, the results from three studies demonstrate that in social interactions, the emotional effects of faces are modulated by dynamic gaze direction.
... Larger amplitudes for N170 have been found in a number of studies in response to faces with averted gaze as compared to direct gaze, that is, when eyes appear to look directly at the participant (Caruana et al., 2014;Itier et al., 2007;Latinus et al., 2015;Puce et al., 2000;Rossi et al., 2015). Yet, there are also reports of larger N170 amplitudes for direct as compared to averted gaze (Conty et al., 2007(Conty et al., , 2012Watanabe et al., 2006), and several studies found no modulation of the N170 by gaze direction (Brefczynski-Lewis et al., 2011;Myllyneva & Hietanen, 2015;Ponkanen et al., 2011;Schweinberger et al., 2007;Taylor et al., 2001). ...
... Some of these inconsistencies may be explained by the properties of the task . As suggested by Latinus et al. (2015), social tasks, in which the participant indicates whether or not the face makes eye contact, may attenuate gaze effects on the N170 in contrast to emotional tasks, where expressions are classified for emotion, or spatial tasks, when gaze direction has to be judged. In addition to task | 3 of 18 BAGHERZADEH-AZBARI et al. requirements, also head orientation, gaze deviations, and static versus dynamic gaze have been discussed as causing inconsistencies in the gaze perception literature (Conty et al., 2007;Itier et al., 2007;Puce et al., 2000, for a recent review see Tautvydaitė et al., 2022). ...
... For the N170, we expected a larger amplitude for averted relative to direct gaze in both initial gaze and gaze change phases; based on Itier et al. (2007) who observed larger N170 amplitudes for averted gaze in static images and the studies by Latinus et al. (2015) and Puce et al. (2000) who found the same effect in changing gaze. In a previous study using our change detection task, we observed larger N170 amplitudes in response to averting gaze than to gaze that turns toward observer (Stephani et al., 2020). ...
Article
Full-text available
This is an open access article under the terms of the Creative Commons Attribution-NonCommercial-NoDerivs License, which permits use and distribution in any medium, provided the original work is properly cited, the use is non-commercial and no modifications or adaptations are made. Abstract The emotional expression and gaze direction of a face are important cues for human social interactions. However, the interplay of these factors and their neural correlates are only partially understood. In the current study, we investigated ERP correlates of gaze and emotion processing following the initial presentation of faces with different emotional expressions (happy, neutral, angry) and an averted or direct gaze direction as well as following a subsequent change in gaze direction that occurred in half of the trials. We focused on the time course and scalp topography of the N170 and EPN components. The N170 amplitude was larger to averted than direct gaze for the initial face presentation and larger to gaze changes from direct to averted than from averted to direct in response to the gaze change. For the EPN component in response to the initial face presentation, we replicate classic effects of emotion, which did not interact with gaze direction. As a major new finding, changes from direct to averted gaze elicited an EPN-like effect when the face showed a happy expression. No such effect was seen for angry expressions. We conclude that happy faces reflexively attract attention when they look at the observer rather than away from the observer. These results for happy expressions are in line with the shared signal hypothesis that posits a better processing of expressions if their approach or avoidance tendency is consistent with gaze direction. However, the shared signal hypothesis is not supported by the present results for angry faces.
... This was true both for receiver non-mutual gaze onsets (where receivers were not looking at their partners and thus may have failed to detect their partners' gaze shift), but also for the receiver-mutual condition (where receivers were directly gazing towards their partner at the time of the gaze shift). Evidence from adult ERP studies in which dynamic changes in gaze are simulated, through the presentation of a series of static images on a screen suggest that human adults are sensitive to 'dynamic' changes in gaze 106,107 . However, these simulated changes in gaze are still far from the continuous way that gaze is processed during real social interactions, and it is likely that the effects observed in these studies are largely driven by more low-level properties of the simulation (e.g., retinal stimulation evoked by the presentation of a series of static images) rather than reflecting the actual processing of the gaze shift. ...
Article
Full-text available
Temporal coordination during infant-caregiver social interaction is thought to be crucial for supporting early language acquisition and cognitive development. Despite a growing prevalence of theories suggesting that increased inter-brain synchrony associates with many key aspects of social interactions such as mutual gaze, little is known about how this arises during development. Here, we investigated the role of mutual gaze onsets as a potential driver of inter-brain synchrony. We extracted dual EEG activity around naturally occurring gaze onsets during infant-caregiver social interactions in N = 55 dyads (mean age 12 months). We differentiated between two types of gaze onset, depending on each partners’ role. ‘Sender’ gaze onsets were defined at a time when either the adult or the infant made a gaze shift towards their partner at a time when their partner was either already looking at them (mutual) or not looking at them (non-mutual). ‘Receiver’ gaze onsets were defined at a time when their partner made a gaze shift towards them at a time when either the adult or the infant was already looking at their partner (mutual) or not (non-mutual). Contrary to our hypothesis we found that, during a naturalistic interaction, both mutual and non-mutual gaze onsets were associated with changes in the sender, but not the receiver’s brain activity and were not associated with increases in inter-brain synchrony above baseline. Further, we found that mutual, compared to non-mutual gaze onsets were not associated with increased inter brain synchrony. Overall, our results suggest that the effects of mutual gaze are strongest at the intra-brain level, in the ‘sender’ but not the ‘receiver’ of the mutual gaze.
Article
Several studies showed a positive effect of stories on Theory of Mind (ToM) performance. The aim of the present exploratory study was to investigate whether and how a specific aspect of narrative, i.e., character perspective, modulates the brain activation in response to a ToM task and improve the accuracy. Fifty participants were divided in three groups based on the text assigned: first-person perspective group (1 G; n = 16), third-person perspective group (3 G; n = 18) and a scientific essay group (EG; n = 16). The electroencephalographic and behavioral responses to eyes expressions, taken from the “Reading the Mind in the Eyes” test, were recorded pre-(T0) and post-(T1) reading task. The main results showed a greater N100 on left fronto-central electrodes and a greater P220–400 on right temporo-parietal electrodes in response to eye expressions at T1 compared to T0 in 3 G. A lower N220–400 was found on right fronto-central in response to eye expressions at T1 compared to T0 in 1 G and 3 G. The results suggest that, although reading first- and third-person stories modulates self-processes in a similar way, third-person stories involve an early stage of processing and a more extended neural network including anterior-posterior brain sites.
Article
Full-text available
The two visual pathway description of Ungerleider and Mishkin changed the course of late 20th century systems and cognitive neuroscience. Here, I try to reexamine our laboratory's work through the lens of the Pitcher and Ungerleider new third visual pathway. I also briefly review the literature related to brain responses to static and dynamic visual displays, visual stimulation involving multiple individuals, and compare existing models of social information processing for the face and body. In this context, I examine how the posterior STS might generate unique social information relative to other brain regions that also respond to social stimuli. I discuss some of the existing challenges we face with assessing how information flow progresses between structures in the proposed functional pathways and how some stimulus types and experimental designs may have complicated our data interpretation and model generation. I also note a series of outstanding questions for the field. Finally, I examine the idea of a potential expansion of the third visual pathway, to include aspects of previously proposed “lateral” visual pathways. Doing this would yield a more general entity for processing motion/action (i.e., “[inter]action”) that deals with interactions between people, as well as people and objects. In this framework, a brief discussion of potential hemispheric biases for function, and different forms of neuropsychological impairments created by focal lesions in the posterior brain is highlighted to help situate various brain regions into an expanded [inter]action pathway.
Article
Full-text available
Artificially created human faces play an increasingly important role in our digital world. However, the so-called uncanny valley effect may cause people to perceive highly, yet not perfectly human-like faces as eerie, bringing challenges to the interaction with virtual agents. At the same time, the neurocognitive underpinnings of the uncanny valley effect remain elusive. Here, we utilized an electroencephalography (EEG) dataset of steady-state visual evoked potentials (SSVEP) in which participants were presented with human face images of different stylization levels ranging from simplistic cartoons to actual photographs. Assessing neuronal responses both in frequency and time domain, we found a non-linear relationship between SSVEP amplitudes and stylization level, that is, the most stylized cartoon images and the real photographs evoked stronger responses than images with medium stylization. Moreover, realness of even highly similar stylization levels could be decoded from the EEG data with task-related component analysis (TRCA). Importantly, we also account for confounding factors, such as the size of the stimulus face’s eyes, which previously have not been adequately addressed. Together, this study provides a basis for future research and neuronal benchmarking of real-time detection of face realness regarding three aspects: SSVEP-based neural markers, efficient classification methods, and low-level stimulus confounders.
Article
Full-text available
The integration of neuroscience and psychotherapy research has long been a topic of interest in the field of mental health. A significant challenge in psychotherapy research is understanding the dyadic interaction between patient and therapist. This interaction is complex, emerging from a myriad of multi-level factors such as gestures, verbal communication, mentalization, and environmental influences. This article aims to present a roadmap for the future integration of neuroscience and psychotherapy research, addressing the complexities of human interaction. We introduce the 4E/MoBI approach, a framework that combines theoretical and methodological tools to study the dynamics of the brain, body, and environment in real-world settings. This approach emphasizes the use of physiological systems (e.g., heart and brain), behavioral interactions (e.g., conversations and eye-tracking), and environmental video recordings. Additionally, the scalable experimental design (SED) heuristic is discussed as a method to blend controlled experiments with real-world scenarios, allowing for the parametric testing of neurobehavioral markers. As a practical demonstration of the SED heuristic within the 4E/MoBI framework, a concrete experimental example using the N170 event-related potential (ERP) component is presented. While the N170 ERP component is not posited as the foundational marker in the field, it serves to illustrate the application of hypothesis-driven designs and analyses. The 4E/MoBI approach offers a promising avenue for the integration of neuroscience and psychotherapy research. By addressing existing gaps, such as the physiology and phenomenology of expertise in psychotherapy, this framework can foster a virtuous relationship between the two disciplines, paving the way for more comprehensive and nuanced understandings of therapeutic interactions.
Article
Full-text available
joint attention gaze cueing paradigm eyetracking mobile brain/body imaging real-life situations Joint attention has been long investigated using behavioral measures. However, there is a need to know more about humans' processing of gaze cueing, linking both mind, brain, and body dynamics. This review aims at presenting past and current research concerning mechanisms of joint attention, encompassing electroencephalographic (EEG), oculomotor, heart rate (HR), facial elec-tromyography, and skin conductance responses (SCR) results and extracting the most important methodological factors diversifying effects, that lead to partial inconsistent knowledge. Particular focus is given to the gaze-cueing effect. Literature analysis reveals four main experimental procedure factors diversifying results. Effects reported in the review show sensitivity to the gaze cueing duration time, ecological accuracy, research paradigm, and contextual variables. Three brain, body, and eye response levels were proposed adequately to the conceptual framework for systems engaged in the joint attention phenomenon. As a synthesis of the review, we have presented inte-grative methodological scheme leading to the revealing holistic reaction pattern in gaze cueing research. Analyzing one type of psychophysiological data gives limited and incomplete insight into the mechanism of imitation of eye movement in response to gaze-cueing. In these cases, the oculomotor, cognitive and affective processes are usually separated. The review indicates the need for simultaneous data collection and the analysis of joint brain/body/eye activity to understand the phenomenon of joint attention in complex and multi-level ways. As a new research paradigm, the mobile brain/body imaging (MoBI) framework allows capturing how a person responds to gaze cueing in realistic, real-life situations.
Preprint
Full-text available
Artificially created human faces play an increasingly important role in our digital world. However, the so-called uncanny valley effect may cause people to perceive highly, yet not perfectly human-like faces as eerie, bringing challenges to the interaction with virtual agents. At the same time, the neurocognitive underpinnings of the uncanny valley effect remain elusive. Here, we utilized an electroencephalography (EEG) dataset of steady-state visual evoked potentials (SSVEP) in which participants were presented with human face images of different stylization levels ranging from simplistic cartoons to actual photographs. Assessing neuronal responses both in frequency and time domain, we found a non-linear relationship between SSVEP amplitudes and stylization level, that is, the most stylized cartoon images and the real photographs evoked stronger responses than images with medium stylization. Moreover, realness of even highly similar stylization levels could be decoded from the EEG data with task-related component analysis (TRCA). Importantly, we also account for confounding factors, such as the size of the stimulus face’s eyes, which previously have not been adequately addressed. Together, this study provides a basis for future research and neuronal benchmarking of real-time detection of face realness regarding three aspects: SSVEP-based neural markers, efficient classification methods, and low-level stimulus confounders.
Article
Full-text available
Neural processes associated with two aspects of visual-spatial attention were investigated with event-related potential (ERPs): those that direct spatial attention to a given point in space and those that modulate the processing of sensory input after attention has been directed. The subjects were 6- to 9-year-old children (51 boys and 35 girls). An arrow cue directed attention from the central to peripheral visual field; targets were then flashed in the attended or ignored visual field 600 msec after the cue. The directing of attention to the left vs. right visual field was associated with hemispheric differences in slow potentials prior to the presentation of the targets. The earliest potential, which started about 200 msec after the cue and was negative over the hemisphere contralateral to the direction of attention, was greatest over the parietal area and appeared to reflect processes directing attention per se. The last potential, which peaked 60 msec after the target and was positive over the hemisphere contralateral to the direction of attention, was greatest over the occipital-parietal region. It appeared to reflect the modulation of cortical excitability in the regions receiving input from the relevant and irrelevant visual fields. The effects of spatial attention on P1, N1, and P3 ERP components following the targets replicated previous results. Boys appeared more aroused (as indicated by CNVs) and reflected faster and greater selective processing (as indicated by reaction time, and N1-P1 latency and amplitude) than girls.
Article
Full-text available
It is now recognized that visual processes modulate many social interactions. For example, the eye gaze of another person is a powerful cue to guide attention to a particular part of the visual field. Conversely, a direct gaze may indicate potential threat or the opportunity for a sexual encounter. In addition, the social or affective significance of a stimulus, as well as the mood state of the observer, can have profound effects on basic attentional and perceptual processes. This special issue is aimed at elucidating the role of visual processes in social interactions by linking work on the basic cognitive mechanisms mediating vision with work on the social and emotional context in which the processing takes place. This introduction provides a brief overview and then summarizes the papers making up the special issue and attempts to place them in a wider theoretical framework.
Article
Full-text available
To understand social interactions we must decode dynamic social cues from seen faces. Here we used magnetoencephalography (MEG) to study the neural responses underlying the perception of emotional expressions and gaze direction changes as depicted in an interaction between two agents. Subjects viewed displays of paired faces that first established a social scenario of gazing at each other (mutual attention) or gazing laterally together (deviated group attention) and then dynamically displayed either an angry or happy facial expression. The initial gaze change elicited a significantly larger M170 under the deviated than the mutual attention scenario. At around 400 ms after the dynamic emotion onset, responses at posterior MEG sensors differentiated between emotions, and between 1000 and 2200 ms left posterior sensors were additionally modulated by social scenario. Moreover, activity on right anterior sensors showed both an early and prolonged interaction between emotion and social scenario. These results suggest that activity in right anterior sensors reflects an early integration of emotion and social attention, while posterior activity first differentiated between emotions only, supporting the view of a dual route for emotion processing. Altogether, our data demonstrate that both transient and sustained neurophysiological responses underlie social processing when observing interactions between others.
Article
Recent studies have shown enhanced brain and autonomic responses to seeing a face with a direct gaze. Interestingly, greater responses to eye contact vs. averted gaze have been observed when showing “live” faces as stimuli but not when showing pictures of faces on a computer screen. In this study, we provide unequivocal evidence that the differential responses observed in the “live” condition are dependent on the observer’s mental attributions. Results from two experiments showed that eye contact resulted in greater autonomic and brain responses compared to averted gaze if a participant believed that the stimulus person sitting on the other side of an electronic shutter was able to see him or her through the shutter. Gaze direction had no effects if participants believed that the transparency from their side was blocked. The results suggest that mental attributions exert a powerful modulation on the processing of socially relevant sensory information.
Article
This study investigated whether the direct gaze of others influences attentional disengagement from faces in an experimental situation. Participants were required to fixate on a centrally presented face with varying gaze directions and to detect the appearance of a peripheral target as quickly as possible. Results revealed that target detection was delayed when the preceding face was directly gazing at the subject (direct gaze), as compared with an averted gaze (averted gaze) or with closed eyes (closed eyes). This effect disappeared when a temporal gap was inserted between the offset of the centrally presented face and the onset of a peripheral target, suggesting that attentional disengagement contributed to the delayed response in the direct gaze condition. The response delay to direct gaze was not found when the contrast polarity of eyes in the facial stimuli was reversed, reinforcing the importance of gaze perception in delayed disengagement from direct gaze.
Article
While is widely accepted that the posterior temporal region is activated during the observation of faces showing gaze shifts, it is still unclear whether its activity is stronger while observing direct or averted gaze. Furthermore, despite its assessed role in social cognition, studies describing an enhanced activity of the posterior temporal region during the observation of gaze aversion interpreted this activity in terms of spatial attention toward the target direction. This spatial attention interpretation is not easily reconcilable with the role of the posterior temporal region in social cognition, and an overarching view of its global cognitive function would be much more preferable. Here we used intracranial EEG to assess the precise spatial localization of the gaze shifts coding in the posterior temporal region, to assess its selectivity for direct versus averted gaze and to distinguish between a spatial-attentional and a social interpretations of gaze aversion. We found stronger activation during gaze aversion than direct gaze and lateral side switch observation, the latter indicating that the crucial aspect of gaze aversion is the prior presence of the eye contact and its interruption, and not the gaze direction. These results suggest a more social-oriented interpretation based on the view that among humans, gaze aversion signals a negative relational evaluation in social interaction. Hum Brain Mapp, 2013. © 2013 Wiley Periodicals, Inc.