ArticlePDF Available

Eye-gaze information input based on pupillary response to visual stimulus with luminance modulation

Authors:

Abstract and Figures

This study develops an information-input interface in which a visual stimulus targeted by a user’s eye gaze is identified based on the pupillary light reflex to periodic luminance modulations of the object. Experiment 1 examines how pupil size changes in response to periodic luminance modulation of visual stimuli, and the results are used to develop an algorithm for information input. Experiment 2a examines the effectiveness of interfaces with two objects. The results demonstrate that 98% accurate identification of the gaze targeted object is possible if the luminance modulation frequencies of two objects differ by at least 0.12 Hz. Experiment 2b examines the accuracy of a gaze directed information input method based on a keyboard configuration with twelve responses. The results reveal that keyboard input is possible with an average accuracy of 85% for luminance modulation frequencies from 0.75 to 2.75 Hz. The proposed pupillometry based information-input interface offers several advantages, such as low burden on users, minimal invasiveness, no need for training or experience, high theoretical validity, and no need for calibration. Thus, the pupillometry method presented herein has advantages for practical use without requiring the eye’s position to be calibrated. Additionally, this method has a potential for the design of interfaces that allow patients with severely limited motor function to communicate with others.
Content may be subject to copyright.
RESEARCH ARTICLE
Eye-gaze information input based on pupillary
response to visual stimulus with luminance
modulation
Yumiko MutoID
1,2¤
*, Hideka Miyoshi
1
, Hirohiko Kaneko
1
*
1Dept. of Information and Communications Engineering, Tokyo institute of Technology, Yokohama,
Kanagawa, Japan, 2Brain Science Institute, Tamagawa university, Tokyo, Japan
¤Current address: Brain Science Institute, Tamagawa university, Tokyo, Japan
*muto@lab.tamagawa.ac.jp (YM); kaneko.h.ab@m.titech.ac.jp (HK)
Abstract
This study develops an information-input interface in which a visual stimulus targeted by a
user’s eye gaze is identified based on the pupillary light reflex to periodic luminance modula-
tions of the object. Experiment 1 examines how pupil size changes in response to periodic
luminance modulation of visual stimuli, and the results are used to develop an algorithm for
information input. Experiment 2a examines the effectiveness of interfaces with two objects.
The results demonstrate that 98% accurate identification of the gaze targeted object is pos-
sible if the luminance modulation frequencies of two objects differ by at least 0.12 Hz. Exper-
iment 2b examines the accuracy of a gaze directed information input method based on a
keyboard configuration with twelve responses. The results reveal that keyboard input is pos-
sible with an average accuracy of 85% for luminance modulation frequencies from 0.75 to
2.75 Hz. The proposed pupillometry based information-input interface offers several advan-
tages, such as low burden on users, minimal invasiveness, no need for training or experi-
ence, high theoretical validity, and no need for calibration. Thus, the pupillometry method
presented herein has advantages for practical use without requiring the eye’s position to
be calibrated. Additionally, this method has a potential for the design of interfaces that allow
patients with severely limited motor function to communicate with others.
Introduction
To regulate the amount of light entering the eye, the pupil diameter changes as a function of
the brightness of the object being viewed [1] [2] [3] [4], and the characteristics of this pupillary
light reflex (PLR) have been explored in many studies [e.g., [5] [6] [7] [8]]. The present study
develops an information-input interface in which a visual object targeted by the user’s eye gaze
is detected based on the pupillary light reflex to periodic luminance modulations of the object.
In addition to conventional information-input devices such as the keyboard and mouse, sev-
eral methods have been proposed that use biological signals such as eye-gaze position and
brain waves [see, e.g., [9] [10]]. Unfortunately, the signals produced by such methods have
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 1 / 18
a1111111111
a1111111111
a1111111111
a1111111111
a1111111111
OPEN ACCESS
Citation: Muto Y, Miyoshi H, Kaneko H (2020) Eye-
gaze information input based on pupillary response
to visual stimulus with luminance modulation.
PLoS ONE 15(1): e0226991. https://doi.org/
10.1371/journal.pone.0226991
Editor: Yoko Hoshi, Preeminent Medical Phonics
Education & Research Center, Hamamatsu
University School of Medicine, JAPAN
Received: April 30, 2019
Accepted: December 10, 2019
Published: January 9, 2020
Copyright: ©2020 Muto et al. This is an open
access article distributed under the terms of the
Creative Commons Attribution License, which
permits unrestricted use, distribution, and
reproduction in any medium, provided the original
author and source are credited.
Data Availability Statement: We have uploaded
the data underlying the findings described in our
manuscript to GitHub (https://github.com/Yumiko-
data/data.git) as possible as we can.
Funding: This research was supported by
Adaptable and Seamless Technology transfer
Program through Target-driven R&D (A-STEP)
from Japan Science and Technology Agency (JST).
Competing interests: The authors have declared
that no competing interests exist.
some disadvantages for the intended purpose: The brain-wave techniques generally require
a contact sensor, which is extremely expensive and complicated signal analysis to extract the
desired information. Conventional eye-gaze input methods require not only user training but
also calibration to relate the signal from the device to eye position. In addition, even if individ-
ual adjustments are made beforehand, measurements of eye position are not possible in some
cases. In contrast, pupillometry offers various advantages for information input. To begin, it
requires only a relatively inexpensive noncontact sensor. In addition, pupillary response to
light is sufficiently stable and not only requires no calibration to relate the signal to the change
in light intensity, but also requires no user training to improve signal stability. Accordingly,
pupillometry is a promising technique for information input, especially for patients with
extremely limited motor functions, such as those suffering from the locked-in syndrome.
To communicate intentions and as an information-input interface, particularly among
the patients with extremely limited motor function, a method has been proposed that uses the
pupillary response to a cognitive load, such as two-digit mental arithmetic [11]. This method
is used mainly for patients with locked-in syndrome. The study posed yes-no questions and
obtained approximately 3.6 selections per minute with 90% accuracy for healthy participants
and approximately 2.7 selections per minute with 70% accuracy for patients with severe motor
disabilities. However, this accuracy and response frequency are insufficient for practical use,
and the technique subjects the patients to an excessive cognitive load. Recently, Matho
ˆt et al.
[12] proposed an information-input method to estimate the position of covert visual attention
foci that employed pupillometry and that uses visual stimuli consisting of two circles of black
and white that flash alternately. Participants selected a letter by looking at it covertly. The let-
ters were shown within circles with oscillating brightness. Small changes in pupil size reflected
changes in the luminance of the attended stimulus [13], which enabled the stimulus selected
by the participant to be identified in real time independently of eye movement. This algorithm
was based on the finding that PLR could be modulated via covert attention [14] [15] and used
as a communication method by patients with complete loss of motor control. However, multi-
ple numbers and characters must be displayed simultaneously if it is to be used for actual
human–computer interaction (HCI) applications simultaneously. Fundamentally, this
method is based on two response options; for example, selecting one character out of eight
requires repeating three times the process of selection between two patterns, which requires
an excessively long time for information input, making for an inefficient information-input
technique. For use in real-life situations, multiple numbers and characters must be displayed
simultaneously.
Herein, we propose a method to determine gaze direction based on the frequency of the
PLR response to periodic luminance modulations of various visual stimuli. The visual stimulus
consists of circles whose luminance is modulated at various frequencies and that demark
pieces of input information (numerals or characters). Specifically, the user may be presented
with multiple pieces of input information; for example, the numerals 0 to 9. Each numeral is
demarked by a circle whose luminance is modulated at a unique frequency (i.e., a frequency
that differs from that of the circles around the other nine numerals). As the observer gazes at a
circle, the frequency of pupillary oscillation is measured, which allows the item with the same
luminance-modulation frequency to be identified as the target of the gaze. This approach
determines a unique target of the gaze based only on the pupil data and without directly mea-
suring eye gaze. It can be used with multiple simultaneous stimuli that correspond to various
numbers and characters. Previous research used square- or sine-wave-modulated luminance
to explore the pupillary response to visual stimuli and clarify its basic characteristics in order
to understand the mechanisms and construct a PLR model [see, e.g., [16] [17] [18] [19] [20]].
However, it remains unclear how the pupil responds to visual stimuli consisting of multiple
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 2 / 18
circles whose luminance is modulated at different frequencies. In addition, no reports exist of
the PLR applied to an information-input interface based on items whose luminance is modu-
lated at different frequencies.
This study consists of two parts: First, we investigate the basic characteristics of pupillary
response to a periodic luminance modulation (Experiment 1). Second, we propose an algo-
rithm for a communication system that uses pupillometry and show its effectiveness when
applied to an information-input interface in which multiple visual items are presented simulta-
neously (Experiment 2). As shown in Fig 1, experiment 2a proposes an interface with two
response options (yes or no) and experiment 2b proposes an interface in the form of a numeric
keypad with twelve response options. Experiment 1 presents a single circular visual stimulus
in the center of a screen and modulates its luminance for each trial at one of the following fre-
quencies: 0.25, 0.5, 0.75, 1.0, 1.25, 1.5, 1.75, 2.0, 2.25, 2.5, 2.75 or 3.0 Hz (Fig 1a). The time-
domain pupil-diameter data are analyzed by using a discrete Fourier transform (DFT) to
quantify the characteristics of the pupil response to the luminance-modulation frequency.
This allows us to identify a suitable range of luminance-modulation frequency and to propose
an algorithm to estimate the luminance-modulation frequency of the gaze-targeted item based
only on the pupil data gathered while the patient gazes at the input interface. Experiment 2
applies the algorithm developed in experiment 1. Experiment 2a allows us to evaluate the effec-
tiveness of an input interface that offers two response options (yes or no) and experiment 2b
does the same but for an interface with 12 response options (numerals or characters). Experi-
ment 2a presents two items on a screen (Fig 1b) and the circle gazed at by the observer is
identified by comparing the measured frequency of pupil-size oscillation with the luminance-
modulation frequency of the various items in the interface. Experiment 2b presents twelve
items (characters or numbers) on a screen (Fig 1c) and uses the same approach to identify
the gaze-targeted item. Herein we evaluate the accuracy of this input identification method.
Fig 1. Visual stimuli for (a) experiment 1, (b) experiment 2a, and (c) experiment 2b.
https://doi.org/10.1371/journal.pone.0226991.g001
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 3 / 18
Herein, participants were instructed to pay attention to a visual stimulus while looking at it.
This is referred to as overt attention. Previous studies [11] [12] focused on covert visual atten-
tion and intentionally excluded the need for motor (eye) movement. In contrast, our proposed
system was designed to use overt attention as a source of new input data.
Experiment 1: Basic characteristics of pupillary response to
periodic luminance modulation
Method
1. Participants. Ten observers participated in experiment 1, all of which were students
and staff (nine males and one female, aged 23 to 29) affiliated with the Tokyo Institute of Tech-
nology. All had either normal or corrected to normal vision. The experiment was conducted
with the approval of the Ethics Committee of the Tokyo Institute of Technology. Written
informed consent was obtained from each participant.
2. Apparatus. Participants were seated in a chair in a dark room in front of a CRT moni-
tor positioned at a viewing distance of 50 cm (Sony Corporation, GDM-F400, 1280 ×1024 pix-
els, 19 inch). A head-chin rest was used to fix the head position of each participant. Pupil size
and eye position of the left eye were recorded with an eye-movement measurement system
consisting of infrared illumination and a CCD camera (iRecHS2 system, [21]) with a data-
sampling frequency of 333 Hz.
3. Visual stimuli. The visual stimulus used in the experiment consisted of a 3˚diameter
circle on the gray background (16.4 cd/m
2
) of the screen. The luminance of the circle (y
RGB
)
was modulated as follows:
yRGB ¼Asinð2pftÞ þ 127;ð1Þ
where fis the luminance-modulation frequency [Hz] of the circle and A is the amplitude of
the luminance modulation and is fixed at 128, which is half of the maximum RGB value. The
visual stimuli were generated and controlled by using MATLAB (MathWorks, Inc.) and a
function library (Psychtoolbox-3) installed on a Mac Book Air running OS X Yosemite (Apple
Inc.). The gamma value of the CRT monitor was determined to be 2.8 based on measurements
with a luminance meter (CS-100A Luminance and Color Mete, KONICA MINOLTA, INC.)
and the gamma correction was applied to linear RGB values. The darkest stimulus was 0.99
cd/m
2
and the brightest was 99.4 cd/m
2
.
4. Procedure. The circle was presented in the center of the screen (Fig 1a). The RGB val-
ues of this circle were modulated by a sine function to create a grayscale luminance modula-
tion [see Eq (1)]. Twelve luminance-modulation frequencies were tested: 0.25, 0.5, 0.75, 1.0,
1.25, 1.5, 1.75, 2.0, 2.25, 2.5, 2.75, and 3.0 Hz. The frequencies were selected based on the study
[22], which shows that a temporal series of luminance stimuli could be synchronized with
modulations in pupil size at frequencies up to 3.15 Hz in participants aged from 20 to 30 years.
Each trial began by a signal from the participant, and each stimulus was presented for 10 s.
Each session consisted of 12 trials, each with a different frequency, and with the 12 frequencies
presented in random order. Two sessions were conducted for each participant and no practice
trials were held prior to the experiment. Participants started each trial by pressing a start but-
ton after closing their eyes briefly and resting for what they deemed sufficient time between
each trial. Participants were also instructed to pay attention to the visual stimulus while look-
ing at it.
5. Analysis. The initial second of pupil data was excluded from the analysis because pupil
size is known to be influenced by the onset of visual stimulus [6]. The data were then smoothed
by applying a moving-average filter with the duration of 120 ms, that is each mean is calculated
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 4 / 18
over a sliding window of length 40 data. After this preprocessing, the frequency characteristics
of the data were explored by applying a DFT [23]. Specifically, the frequency with the highest
power was determined based on the periodogram power spectral density (PSD), which was
estimated from the frequency components obtained from the DFT of the measured and nor-
malized time-domain data of pupil size. The frequency steps in the DFT were approximately
0.01 Hz [24]. A periodogram is a nonparametric method for estimating PSD and is defined
below. Comparative reduction of noise components is an advantage of this technique and
involves squaring the amplitude of the DFT of a sample and scaling it for comparison with
other methods in which a simple spectrum is calculated. Specifically, this is done as follows:
PxxðfÞ ¼ 1
LFsjX
L1
n¼0
xLðnÞej2pfn=Fsj2;ð2Þ
where F
s
is the sampling frequency and P
xx
(f) is the PSD of a signal x
L
(n) with length L. All
calculations were done offline using the Matlab signal processing toolbox (MathWorks, Inc.).
There are two more reasons why we applied the DFT analysis: First, the use of visual stimuli in
which the luminance was sinusoidally modulated led us to expect a similar sinusoidal oscilla-
tion of the pupil diameter. If the DFT reveals a strong frequency component in the sinusoidally
modulated data, this frequency could serve as an appropriate index of synchronization. Sec-
ond, given two or more visual items with different luminance-modulation frequencies (Experi-
ment 2), the influence of items not targeted by the gazed of the patient may be quantified by
comparing the PSD of the luminance-modulation frequencies of these items.
Results and discussion
1. Frequency analysis of pupillary response. In experiment 1, we investigated the time
domain data of pupil response to an isolated visual stimulus presented at the center of a screen
and with one of 12 different luminance-modulation frequencies. Fig 2 shows examples for one
participant of the measured time-domain data for pupil diameter (right column) in response
to the 12 luminance-modulation frequencies (left column). Fig 3 shows the PSD obtained as a
function of frequency by applying a DFT to the data of Fig 2. The red lines indicate the lumi-
nance-modulation frequency for the given visual stimulus.
These results show that, for all luminance-modulation frequencies f, the PSD peak, which
indicates the main frequency at which the pupil size is modulated, is approximately the same
as the luminance-modulation frequency of the gaze-targeted visual stimulus. In other words,
the modulation of pupil size is synchronized with the luminance modulation of the visual stim-
ulus, as expected. However, the magnitude of the PSD peak depends on the luminance-modu-
lation frequency of the visual stimulus. Therefore, to quantify the change in PSD frequency, we
averaged the PSD peak over all the participants for each luminance-modulation frequency (Fig
4). The results show that the PSD peak decreases with increasing luminance-modulation fre-
quency. A nonlinear regression analysis shows that this relationship may be approximated
with high explanatory power by the exponential function
fðxÞ ¼ aexpðbxÞ ð3Þ
Fitting Eq (3) to the data gives a = 3.545, b = 2.844 (R
2
= 0.97, root mean square
error = 0.09). This decrease in PSD peak with increasing frequency is attributed to the funda-
mental characteristics of the pupil response. Numerous previous studies have shown that
the amplitude of pupil-size modulation depends on the luminance modulation frequency
of the visual stimulus. The amplitude of pupil-size modulation is small in response to rapid
luminance modulations[see, e.g., [6]]. Because the PSD analysis applied herein is based on the
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 5 / 18
amplitude of pupil-size modulations [Eq(2)], the exponential decrease in Fig 4 is attributed to
the differences in PSD amplitude at each frequency. Therefore, a meaningful comparison of
PSD magnitudes in response to luminance-modulation frequencies would require a correction
for the fundamental difference in the pupil response amplitude.
2. Estimation of oscillation frequency f of visual stimulus. This study estimates the
luminance-modulation frequency fof the visual stimulus from the measured oscillations in
pupil size. First, for each trial, we calculated the PSD for each of the 12 frequencies (f= 0.25,
0.5, 0.75, 1.0, 1.25, 1.5, 1.75, 2.0, 2.25, 2.5, 2.75, or 3.0 Hz). Next, the PSD for each frequency
Fig 2. Experiment 1: (a) luminance-modulated visual stimuli and (b) change in pupil size [pix] as a function of
time. The initial second of data is excluded from the analysis. See text for details.
https://doi.org/10.1371/journal.pone.0226991.g002
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 6 / 18
Fig 3. Experiment 1: Examples of DFT of time-domain pupil-size data; the results are for a single participant. The red line
shows the luminance-modulation frequency of the gaze-targeted visual stimulus.
https://doi.org/10.1371/journal.pone.0226991.g003
Fig 4. Experiment 1: Average PSD of subjects as a function of luminance-modulation frequency of gaze-targeted visual
stimulus.
https://doi.org/10.1371/journal.pone.0226991.g004
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 7 / 18
was corrected by multiplying by the weight
o¼1
fðxÞ¼1
aexpð bxÞ;ð4Þ
which is the reciprocal of Eq (3), because the PSD varies exponentially with modulation fre-
quency, as shown in Fig 4. We determine the parameters aand bin the exponential function
f(x) [Eq (3)] from a nonlinear regression analysis of the data (see Fig 4). After this correction,
the PSD results for the 12 frequencies are compared to estimate the frequency f0that produces
the largest PSD (PSD
max
). If the estimated frequency f0corresponds to the luminance-modula-
tion frequency fof the gaze-targeted visual stimulus, then the estimate of f0is considered
successful. Table 1 lists the estimation accuracy averaged over the 10 participants for each fre-
quency. An accuracy of 90% or greater is obtained at all frequencies except for 0.5 and 3.0 Hz,
which have accuracies of 80% and 30%, respectively. These results reveal that the proposed
method allows us to estimate the gaze point with high accuracy (90%) if the luminance-modu-
lated frequency of the visual stimulus is between 0.75 and 2.75 Hz. Table 2 is the estimation
confusion matrix. The first column shows the frequency of the estimated stimulus (Hz),
whereas the actual frequency is shown in a horizontal row. Estimation error was highest at 3
Hz, which indicated that 3 Hz was not suitable for application in the system. An upper limit of
2.75 Hz was more appropriate.
Experiment 2: Information-input interface and its evaluation
Method
1. Participants. Experiment 2a used ten male participants aged 22 to 29 years and experi-
ment 2b used eight participants (five males and three females) aged 21 to 28 years, all of whom
Table 1. Experiment 1: Accuracy for each frequency.
Frequency[Hz] 0.25 0.5 0.75 1.0 1.25 1.5 1.75 2.0 2.25 2.5 2.75 3.0
Accuracy[%] 90 80 90 90 90 90 100 90 100 100 100 30
https://doi.org/10.1371/journal.pone.0226991.t001
Table 2. Experiment 1: Confusion matrix.
Estimated frequency [Hz]
0.25 0.5 0.75 1.0 1.25 1.5 1.75 2.0 2.25 2.5 2.75 3.0
Actual frequency [Hz]
0.25 91 0 0 0 0 0 0 0 0 0 0
0.5 0 80 0 0 1 0 0 0 0 0 1
0.75 0 0 9000000001
1 0 0 0 90 0 0 0 0 0 0 1
1.25 0 0 0 0 90000001
1.5 0 0 0 0 0 90 0 0 0 0 1
1.75 0 0 0 0 0 0 10 00000
2 0 0 0 0 0 0 0 90 0 0 1
2.25 0 0 0 0 0 0 0 0 10 0 0 0
2.5 0 0 0 0 0 0 0 0 0 10 0 0
2.75 0 0 0 0 0 0 0 0 0 0 10 0
3 3 0 0 1 0 0 0 0 1 2 0 3
https://doi.org/10.1371/journal.pone.0226991.t002
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 8 / 18
were students or staff affiliated with the Tokyo Institute of Technology. All participants had
either normal or corrected to normal vision. The two experiments were conducted on different
dates and times. One participant participated in both experiments. The experiments were
approved by the Ethics Committee of the Tokyo Institute of Technology. A written informed
consent was obtained from each study participant.
2. Apparatus and analysis. The apparatus for experiment 2 and its analysis are the same
as for experiment 1.
3. Procedure.
1. Experiment 2a: Communication interface with two response options
This interface is intended to be used for a communication device that accepts only two
answers, such as yes or no. Two circles separated by a 1˚viewing angle appear side by side
on the screen (Fig 1b) and serve as visual stimuli. The left circle is modulated one of the fol-
lowing eight luminance-modulation frequencies: f= 0.75, 1, 1.25, 1.5, 1.75, 2, 2.25, and 2.5
Hz. The luminance-modulation frequency of the right visual stimulus was 0.06, 0.12, or
0.25 Hz higher than that of the left stimulus. For example, when the left circle was modu-
lated at 0.75 Hz, the right circle was modulated at 0.81, 0.87, or 1.0 Hz. In a given experi-
mental session, each combination of stimulus frequencies was presented once, so 3×8 = 24
trials were conducted in total. The luminance-modulation frequencies of the left and right
circles differed by 0.06, 0.12, or 0.25 Hz to determine the smallest possible frequency differ-
ence Δffor which the gaze position of the pupil could be differentiated. We assumed that
this combination of frequencies could be used to examine how the luminance modulation
of the adjacent visual object affects the results. The participants were instructed to direct
their gaze at the left or right circle. We refer to the condition under which eye gaze was
directed at the right (left) circle as the R (L) condition. Each participant took part in two
sessions so that the experiment comprised a total of 48 trials. All of the combinations of sti-
muli and instructions were presented in random order. Each session was divided into two
sets of 12 trials, with each set of trials conducted separately. In each trial, first, a fixation
point was presented at the center of a gray background. After participants pressed the start
button to begin the trial, an “R” or “L” was displayed, instructing the participants to gaze at
either the right or left circle. The luminance of the circles was then modulated for 8 s while
the participants directed their gaze at the designated circle. After closing their eyes to rest as
needed, the participants pressed the start button for the next trial. No training was done in
advance of the experiments.
2. Experiment 2b: Information-input interface with 12 options
This interface is intended to be used as an input device (e.g., a numeral or character key-
board) with a small number of keys. Twelve circles were presented simultaneously on the
screen (see Fig 1c), with a 1˚viewing-angle separation between the edges of adjacent cir-
cles. The luminance-modulation frequencies of the circles were, from left to right, 0.58,
0.70, 0.82 for the top row, 0.94, 1.06, 1.18 for the second row, 1.30, 1.42, 1.54 for the third
row, and 1.66, 1.78, 1.90 Hz for the bottom row. These 12 frequencies were separated
by 0.12 Hz. At the center of each circle appeared a numeral from 0 through 9, the word
“SPACE,” and the symbols “<” or “” with a viewing-angle size of 0.5˚and a fixed lumi-
nance of 16.4 cd/m
2
. The participants started each trial by pressing the start button. First,
at the center of the gray screen appeared for 2 s a numeral or character, which instructed
the participant to gaze at the corresponding circle during the following trial. The 12 sti-
muli (Fig 1c) were then presented for 8 s each, during which time the participants gazed
at the predesignated circle. Prior to this experiment, each participant took part in a prac-
tice session.
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 9 / 18
4. Results and discussion.
1. Experiment 2a: Communication interface with two response options
The goal of experiment 2a was to identify at which circle the participants gazed based
on pupil-size oscillation measured with a visual display. Fig 5 shows typical PSD results
obtained with 0.01 Hz intervals. These results were constructed from the DFT of 7 s of
time-series data of pupil response for one participant gazing at the right circle. The results
show the PSD for differences in luminance-modulation frequencies between the two circles
of Δf= 0.06, 0.12, and 0.25 Hz. The blue and red lines indicate the PSD at the luminance-
modulation frequency of the left and right circles, respectively. These results show that, in
all of the trials, the PSD of the luminance-modulation frequency (red line) of the gaze-tar-
geted circle (i.e., the right circle) exceeded that of the left circle (blue line). This indicates
Fig 5. Experiment 2a: Example of DFT of time-domain pupil modulation results for oneparticipant and for each
frequency condition in which eye gaze was directed at the right circle.
https://doi.org/10.1371/journal.pone.0226991.g005
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 10 / 18
that, even when the difference between the luminance-modulation frequencies of the gaze-
targeted stimulus and adjacent (i.e., non-gaze-targeted) stimuli is relatively small (0.06,
0.12, 0.25 Hz), it is possible to identify the gaze-targeted stimulus based on the pupil
response.
Fig 6 shows the accuracy with which the gaze-targeted visual stimulus is identified
(called the “identification accuracy”) averaged over all participants for each frequency com-
bination. The gaze-targeted stimulus is identified by comparing the PSD of the luminance-
modulation frequencies of the two circles. Based on these results, we conclude that the
gaze-targeted object can be identified with at least 85% accuracy except just one condition
(Δf= 0.06Hz, luminance-modulation frequency = 2Hz) tested in this experiment. Fig 7
Fig 6. Average identification accuracy as a function of luminance-modulation frequency.
https://doi.org/10.1371/journal.pone.0226991.g006
Fig 7. Experiment 2a: Average identification accuracy as a function of left-right frequency difference Δf.
https://doi.org/10.1371/journal.pone.0226991.g007
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 11 / 18
shows the identification accuracy as a function of the frequency difference Δfbetween the
left and right circles for all base frequencies. The gaze-targeted stimulus was identified by
comparing the PSD of the luminance-modulation frequencies of the two circles. In other
words, we assume that, when a participant gazed at a circle with a certain luminance-modu-
lation frequency, the PSD for the given frequency increased.
For each frequency condition and for participants following the instructions regarding
gaze position, we calculated the accuracy of the identification of the gaze-targeted stimu-
lus based on the pupil response. The identification accuracy averaged over participants
was 89.4%, 98.8%, and 98.1% for the frequency difference Δfof 0.06, 0.12, and 0.25 Hz,
respectively. We used a single-factor analysis of variance to investigate the quantitative
effect of the frequency difference between gaze-targeted and non-gaze-targeted visual sti-
muli and found a significant effect [F (2,47) = 9.28, p <0.01]. The results of multiple com-
parisons based on Fisher’s protected least significant difference method reveal that the
differences in the results for Δf= 0.06 Hz and Δf= 0.12 Hz, and for Δf= 0.06 Hz and
Δf= 0.25 Hz are significant at the level of 1%. Based on these analyses, a reliable estimate
(98% accuracy) from the present method requires a minimum difference in luminance-
modulation frequency between the two circles of 0.12 Hz. In experiment 2a, we identified
the gaze-targeted circles (L circle or R circle) by comparing the frequency of the PSD peak
of the pupil-size oscillation with the luminance-modulation frequencies of two circles
without applying the correction (4). As shown in Fig 6, the identification accuracy is very
high. In fact, recalculating the prediction with the correction (4) included does not change
the results, which we attribute to the small difference Δfin the luminance-modulation fre-
quencies of the two circles. For greater differences Δfin the luminance-modulation fre-
quencies, the correction method should be included for a more reliable comparison
between the two PSDs.
2. Experiment 2b: Information-input interface with 12 options
Experiment 2b simultaneously compared 12 circles with different luminance-modulation
frequencies and examined the identification accuracy of the gaze-targeted circle. Fig 8
shows typical PSDs calculated with 0.01 Hz, DFT sampling intervals for time-domain data
spanning 7 s of pupil response. The results are given for each frequency condition for one
participant. The red lines indicate the luminance-modulation frequency of the gaze-tar-
geted circle. These results show that, in most of the cases, one of the PSD peaks corresponds
to the luminance-modulation frequency of the gaze-targeted stimulus. In other words, even
when simultaneously presenting 12 visual stimuli with different luminance-modulation fre-
quencies, the pupil size synchronizes to the luminance-modulation frequency of the gaze-
targeted stimulus.
However, more often than not, the PSD peak at the luminance-modulation frequency of
the gaze-targeted stimulus is a minor PSD peak. Therefore, we apply the data-correction
method developed in experiment 1 to increase the identification accuracy. Specifically, the
PSD functions shown in Fig 8 are multiplied by the weight ωgiven in Eq (4), and the fre-
quency with the highest corrected PSD is taken as the frequency for identifying the gaze-
targeted stimulus. Fig 9 shows the identification accuracy for each of the 12 luminance-
modulation frequencies averaged over all participants, and Table 3 lists the accuracy aver-
aged over all frequencies for each participant. These results indicate that information input
is possible for all of participants tested with at least at 60% accuracy. Note that, upon exclud-
ing the results for 0.58 Hz because the identification accuracy is quite low (62.5%) at this
frequency, the average accuracy reaches 87.5%.
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 12 / 18
Discussion
This study proposes an information-input interface that uses pupillometry and DFT analysis
and focuses on the synchronization of pupil-size oscillation with luminance modulation of visual
stimuli. We experiment with an interface with two response choices and an information-input
Fig 8. Experiment 2b: Typical PSD as a function of frequency for each frequency condition for a single
participant.
https://doi.org/10.1371/journal.pone.0226991.g008
Fig 9. Experiment 2b: Identification accuracy averaged over all participants for the 12 luminance-modulation frequencies.
https://doi.org/10.1371/journal.pone.0226991.g009
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 13 / 18
interface with 12 response choices and evaluate the accuracy with which information can be
input using the proposed method. Although pupil size is well known to oscillate in response to
periodic luminance modulation of visual stimuli [see, e.g., [16] [17] [18] [19] [20]], no reports
exist that detail the correspondence between the pupil-oscillation frequency and luminance-
modulation frequency of stimuli in an information-input interface. In addition to the proposed
information-input interface, we report several important findings regarding the pupil response
to visual stimuli with various luminance-modulation frequencies.
In experiment 1, we examine the basic frequency characteristics of the pupil response by
exposing the participants to a single circular object with a luminance-modulation frequency
from 0.25 to 3.0 Hz. The analysis results are consistent with those from previous studies, which
reported, for participants aged 20 to 30 years, synchronization between the frequency of oscil-
lations in participant pupil size and the luminance-modulation frequency up to 3.15 Hz [see,
e.g., [22]]. The PSD obtained from the DFT of the time-domain data depends on the frequency
condition and, based on the results of a nonlinear regression analysis, can be approximated by
an exponential function with a high degree of explanatory power. This result is attributed to
the pupil-size modulation lessening with increasing luminance frequency. The PSD results
were thus corrected for this effect in order to compare the PSD for stimuli with different lumi-
nance-modulation frequencies and select the frequency that produces the greatest PSD. Upon
applying this correction method [see Eq (4)], the accuracy with which the gaze-targeted stimu-
lus is identified based on its luminance-modulation frequency is 90% or more for frequencies
from 0.75 to 2.75 Hz. Identification accuracy is 90% at 0.25 Hz and 80% at 0.5 Hz. At 3 Hz, the
identification accuracy drops to 30%. From these considerations, we conclude that luminance-
modulation frequencies from 0.75 to 2.25 Hz are optimal for a gaze-input system. The reason
why the identification accuracy was low in the less than 0.5 Hz, would be that the change of
pupil size is affected by the change in heart beat and respiration due to changes in the activa-
tion of the sympathetic nervous system that occur in the range of 0.15 to 0.40 Hz [25] [26]
[27]. From 0.75 to 2.75 Hz, the visual stimulus may be identified with the proposed methods
based on the gaze with an accuracy of 90%. Furthermore, we propose an algorithm to generate
input information based on the data obtained.
In experiment 2a, we examined the effectiveness of an interface with two response choices.
Participants gazed at one of the two visual stimuli in the display and the gaze-targeted stimulus
was identified by comparing the measured frequency of pupil-size modulation with the lumi-
nance-modulation frequency of the gaze-targeted stimulus. The results demonstrate an identi-
fication accuracy of 89.4% for the difference Δf= 0.06 Hz between the luminance-modulation
frequencies of the stimuli, which rises to 98% or more for Δf= 0.12 and 0.25 Hz. These results
demonstrate the effectiveness of the proposed method and indicate that the pupil diameter
becomes synchronized to the luminance modulation of the gaze-targeted stimulus rather
than to that of adjacent, non-gaze-targeted stimuli. It was recently shown that the PLR is
also affected by the luminance modulation of nearby, non-gaze-targeted stimuli [12] [28] [29].
Therefore, we consider that the proposed method is more efficient and reliable when partici-
pants direct both gaze and attention to a given visual stimulus.
Experiment 2b investigated the effectiveness of the proposed algorithm applied to an infor-
mation-input interface containing 12 items presented in a keyboard configuration. Based on
Table 3. Experiment2b: Identification accuracy for each participant averaged over all luminance-modulation frequencies.
Participant No. A B C D E F G H Average
Accuracy rate [%] 100 66.7 100 83.3 100 75 66.7 91.7 85.4
https://doi.org/10.1371/journal.pone.0226991.t003
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 14 / 18
the results of experiment 1, the luminance of the visual stimuli was modulated at frequencies
ranging from 0.58 to 1.9 Hz. The results show that the identification accuracy averaged over all
participants was 85.4%, which indicates that the pupil-size oscillation becomes synchronized
to the luminance-modulation frequency of the gaze-targeted stimulus (3˚viewing angle), even
if neighboring stimuli have differing luminance-modulation frequencies with a separation of
1˚. However, the identification accuracy for the lowest luminance-modulation frequency of
0.58 Hz is only 62.5%, so we conclude that an appropriate range of luminance-modulation fre-
quency for visual stimuli is from 0.7 to 1.9 Hz for a reliable and accurate information-input
interface. Some further issues remain to be discussed to improve the proposed method. The
first issue pertains to the individual differences observed in the accuracy with which the stimuli
are identified. Although three participants had identification accuracies of 100% in experiment
2b, two participants had identification accuracies around 67%. One possible cause for this
individual difference is a difference in experience with the interface. This issue is especially
important when developing a human interface for handicapped people. We must develop an
interface that all patients can use with ease without special experience or training. The second
issue to consider is the interface design, which includes the stimulus configuration and the
algorithm for choosing a gaze-targeted object. In experiment 2b, we arranged the 12 stimuli so
that the luminance-modulation frequency increased from left to right and from top to bottom.
However, this arrangement may not be optimal because the configuration of the stimuli and
the procedure design should make the algorithm more accurate and allow the development of
an easy-to-use interface. In addition, in this study, the luminance of the visual stimulus was
modulated by a sine wave to reduce participant stress. However, a square-wave modulation
may improve the accuracy because pupil size reportedly changes more under square-wave
modulation of the luminance of the stimulus [18]. Finally, the third issue is how attention
and gaze position affect the ability to direct the gaze at an object whose luminance is being
modulated. As mentioned above, recent reports claim that the PLR is affected by the lumi-
nance of a visual stimulus in a neighboring position, even if the eye is not directed at that posi-
tion [28] [29].
In the present study, the participants were instructed to direct both eye and attention to the
visual stimulus, which helped to accurately identify the gaze-targeted object without being dis-
tracted by the luminance modulation of adjacent stimuli. However, if distracting objects are in
the field of view, the accuracy with which gaze-targeted objects are identified might decrease.
Conversely, the use of attention only without directing gaze toward an object may suffice as
input. Such an interface would be helpful for locked-in patients. Therefore, further detailed
investigations are warranted to look into how eye gaze and attention can be exploited to
improve the information-input interface.
Changes in pupil size are known to depend on several factors. Psychological factors, such as
alertness level [6], emotions [30] [31] [32], and mental workload [33] [34], affect the mediation
of pupil size by the autonomic nervous system. Additionally, the mean steady-state pupil size
reportedly decreases with age [35]. The impacts of these factors must be investigated to employ
our proposed system in the future. This will require experiments involving many participants
with different demographic characteristics, such as gender, age, and living environment.
Compared with other techniques for information-input interfaces that use biological sig-
nals, such as EEG, MEG, or optical topography, pupillometry has some important benefits.
The proposed method for implementing an information-input interface using pupillometry
puts a low burden on users, has minimal invasiveness, requires no training or experience,
has high theoretical validity, and requires no calibration. However, online analysis will
be required for real-life application of this method. In fact, we have begun performing
online analysis. The results appear to be reliable [36], although we have not yet collected a
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 15 / 18
sufficient amount of data. In the near future, we hope to develop a complete system that
functions online.
Conclusions
This study demonstrates an information-input interface that uses the synchronization of pupil
oscillation with the luminance modulation of visual stimuli to identify the stimuli. The method
identifies the gaze-targeted object from among many objects with differing luminance-modu-
lation frequencies. The use of pupillometry for uncalibrated eye tracking in HCI has several
possible practical applications. This method may be helpful for the design of interfaces that
allow patients with severely limited motor function to communicate clearly with others.
Acknowledgments
We thank Shoma Ichino, Junki Someya and Tsubasa Tanaka for their contribution in various
stages of this project.
Author Contributions
Conceptualization: Yumiko Muto, Hideka Miyoshi, Hirohiko Kaneko.
Data curation: Yumiko Muto, Hideka Miyoshi.
Formal analysis: Yumiko Muto.
Funding acquisition: Hirohiko Kaneko.
Investigation: Yumiko Muto, Hideka Miyoshi, Hirohiko Kaneko.
Methodology: Yumiko Muto, Hideka Miyoshi, Hirohiko Kaneko.
Writing – original draft: Yumiko Muto.
Writing – review & editing: Yumiko Muto, Hirohiko Kaneko.
References
1. Bouma H. Size of the Static Pupil as a Function of Wave-length and Luminosity of the Light Incident on
the Human Eye. Nature. 1962; 193:690. https://doi.org/10.1038/193690a0 PMID: 13871842
2. Crawford BH. The dependence of pupil size upon external light stimulus under static and variable condi-
tions. Royal Society of London Proceedings Series B Biological Sciences. 1936; 121(823):376–395.
https://doi.org/10.1098/rspb.1936.0072
3. Woodhouse JM, Campbell FW. The role of the pupil light reflex in aiding adaptation to the dark. Vision
Research. 1975; 15(6):649–653. https://doi.org/10.1016/0042-6989(75)90279-5 PMID: 1138479
4. Woodhouse JM. The effect of pupil size on grating detection at various contrast levels. Vision Research.
1975; 15(6):645–648. https://doi.org/10.1016/0042-6989(75)90278-3 PMID: 1138478
5. Loewenfeld IE. Mechanisms of reflex dilatation of the pupil. Documenta Ophthalmologica. 1958; 12
(1):185–448. https://doi.org/10.1007/bf00913471 PMID: 13609524
6. Loewenfeld IE. The Pupil: Anatomy, Physiology and Clinical Applications. Iowa State University Press.
1993;.
7. Barbur JL. A Study of Pupil Response Components in Human Vision. In: Basic and Clinical Perspec-
tives in Vision Research. Boston, MA: Springer US; 1995. p. 3–18. Available from: http://link.springer.
com/10.1007/978-1-4757-9362-8_1.
8. Reeves P. The Response of the Average Pupil to Various Intensities of Light. Journal of the Optical
Society of America. 1920; 4(2):35–43. https://doi.org/10.1364/JOSA.4.000035
9. Farwell LA, Donchin E. Talking off the top of your head: toward a mental prosthesis utilizing event-
related brain potentials. Electroencephalography and Clinical Neurophysiology. 1988; 70(6):510–523.
https://doi.org/10.1016/0013-4694(88)90149-6 PMID: 2461285
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 16 / 18
10. Fazel-Rezai R, Allison BZ, Guger C, Sellers EW, Kleih SC, Ku¨bler A. P300 brain computer interface:
current challenges and emerging trends. Frontiers in neuroengineering. 2012; 5:14. https://doi.org/10.
3389/fneng.2012.00014 PMID: 22822397
11. Stoll J, Chatelle C, Carter O, Koch C, Laureys S, Einha
¨user W. Pupil responses allow communication in
locked-in syndrome patients. Current Biology. 2013; 23(15):R647–R648. https://doi.org/10.1016/j.cub.
2013.06.011 PMID: 23928079
12. Matho
ˆt S, Melmi JB, van der Linden L, Van der Stigchel S. The Mind-Writing Pupil: A Human-Computer
Interface Based on Decoding of Covert Attention through Pupillometry. PLOS ONE. 2016; 11(2):
e0148805. https://doi.org/10.1371/journal.pone.0148805 PMID: 26848745
13. Matho
ˆt S, Van der Stigchel S. New Light on the Mind’s Eye: The Pupillary Light Response as Active
Vision. Current Directions in Psychological Science. 2015; 24(5):374–378. https://doi.org/10.1177/
0963721415593725 PMID: 26494950
14. Binda P, Pereverzeva M, Murray SO. Attention to bright surfaces enhances the pupillary light reflex.
The Journal of neuroscience: the official journal of the Society for Neuroscience. 2013; 33(5):2199–204.
https://doi.org/10.1523/JNEUROSCI.3440-12.2013
15. Matho
ˆt S, van der Linden L, Grainger J, Vitu F. The pupillary light response reveals the focus of covert
visual attention. PloS one. 2013; 8(10). https://doi.org/10.1371/journal.pone.0078168 PMID: 24205144
16. Myers GA, Gannon JA, Stark LW. Level dependent signal flow in the light pupil reflex. Biological Cyber-
netics. 1993; 68(3):235–240. https://doi.org/10.1007/bf00224857 PMID: 8452892
17. Barbur JL, Moro S, Harlow JA, Lam BL, Liu M. Comparison of pupil responses to luminance and colour
in severe optic neuritis. Clinical Neurophysiology. 2004; 115(11):2650–2658. https://doi.org/10.1016/j.
clinph.2004.05.022 PMID: 15465455
18. Moro S. A study of pupil response components in human vision. Unpublished Doctoral thesis, City Uni-
versity London. 2004;.
19. Silva B, Sfer A, D’Urso Villar MA, Issolio LA, Colombo EM. Pupil dynamics with periodic flashes: effect
of age on mesopic adaptation. Journal of the Optical Society of America A. 2016; 33(8):1546–1552.
https://doi.org/10.1364/JOSAA.33.001546
20. Zangemeister WH, Gronow T, Grzyska U. Pupillary responses to single and sinusoidal light stimuli in
diabetic patients. Neurology international. 2009; 1(1):e19–e19. https://doi.org/10.4081/ni.2009.e19
PMID: 21577356
21. Matsuda K, Nagami T, Sugase Y, Takemura A, Kawano K. A Widely Applicable Real-Time Mono/
Binocular Eye Tracking System Using a High Frame-Rate Digital Camera. In: Kurosu M, editor. Human-
Computer Interaction. User Interface Design, Development and Multimodality. Cham: Springer Inter-
national Publishing; 2017. p. 593–608.
22. Alexandridis E, Manner M. [Frequency of the pupillary response following flicker stimuli (author’s
transl)]. Albrecht von Graefes Archiv fur klinische und experimentelle Ophthalmologie Albrecht von
Graefe’s archive for clinical and experimental ophthalmology. 1977; 202(3):175–80. https://doi.org/10.
1007/bf00407866 PMID: 301718
23. Press W H Teukolsky S A, Vetterling W T F BP. Numerical recipes in C: the art of scientific computing.
Cambridge University Press. 1992;.
24. Kay SM, Marple SL. Spectrum analysis-A modern perspective. Proceedings of the IEEE. 1981; 69
(11):1380–1419. https://doi.org/10.1109/PROC.1981.12184
25. Calcagnini G, Lino S, Censi F, Cerutti S. Cardiovascular autonomic rhythms in spontaneous pupil fluctu-
ations. In: Computers in Cardiology 1997; 1997. p. 133–136.
26. Yoshida H, Yana K, Okuyama F, Tokoro T. Time-varying properties of respiratory fluctuations in pupil
diameter of human eyes. Methods of information in medicine. 1994; 33(1):46–8. https://doi.org/10.
1055/s-0038-1634990 PMID: 8177078
27. Yoshida H, Mizuta H, Gouhara T, Suzuki Y, Yana K, Okuyama F. Statistical properties of simulta-
neously recorded fluctuations in pupil diameter and heart rate. In: Proceedings of 17th International
Conference of the Engineering in Medicine and Biology Society. vol. 1; 1995. p. 165–166 vol.1.
28. Matho
ˆt S, van der Linden L, Grainger J, Vitu F. The Pupillary Light Response Reveals the Focus of
Covert Visual Attention. PLOS ONE. 2013; 8(10):e78168. https://doi.org/10.1371/journal.pone.
0078168 PMID: 24205144
29. Naber M, Alvarez G, Nakayama K. Tracking the allocation of attention using human pupillary oscilla-
tions. Frontiers in Psychology. 2013; 4(919). https://doi.org/10.3389/fpsyg.2013.00919 PMID:
24368904
30. Partala T, Surakka V. Pupil size variation as an indication of affective processing. International
Journal of Human Computer Studies. 2003; 59(1-2):185–198. https://doi.org/10.1016/S1071-5819(03)
00017-X
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 17 / 18
31. Bradley MM, Miccoli L, Escrig MA, Lang PJ. The pupil as a measure of emotional arousal and auto-
nomic activation. Psychophysiology. 2008; 45(4):602–607. https://doi.org/10.1111/j.1469-8986.2008.
00654.x PMID: 18282202
32. Oliva M, Anikin A. Pupil dilation reflects the time course of emotion recognition in human vocalizations.
Scientific Reports. 2018; 8(1). https://doi.org/10.1038/s41598-018-23265-x
33. Van Orden KF, Limbert W, Makeig S, Jung TP. Eye activity correlates of workload during a visuospatial
memory task. Human factors. 2001; 43(1):111–21. https://doi.org/10.1518/001872001775992570
PMID: 11474756
34. Ahlstrom U, Friedman-Berg FJ. Using eye movement activity as a correlate of cognitive workload. Inter-
national Journal of Industrial Ergonomics. 2006; 36(7):623–636. https://doi.org/10.1016/j.ergon.2006.
04.002
35. Hasegawa S, Ishikawa S. [Age changes in pupillary light reflex. A demonstration by means of a pupill-
ometer]. Nippon Ganka Gakkai zasshi. 1989; 93(10):955–61. PMID: 2603847
36. Muto Y, Kaneko H. Proposal of “YES/NO” communication method based on frequency analysis of pupil
response. Human Interface Symposium (HI2018), 25D. 2018;.
Eye-gaze information input based on pupillary response
PLOS ONE | https://doi.org/10.1371/journal.pone.0226991 January 9, 2020 18 / 18
... Hybrid BCI can improve the classification accuracy, increase the number of commands, and shorten the detection time of the BCI system by combining two or more patterns (at least one of which is a brain signal) (Hong and Jawad, 2017). Recently, pupillary responses (PR), such as the pupillary light reflex, have been used as the second pattern in addition to EEG due to the low user burden, non-invasiveness, and no need for training (Muto et al., 2020). Pupil diameter changes steadily with the illuminance of the observed object to regulate the amount of light entering the eye (Crawford, 1936;Woodhouse, 1975;Woodhouse and Campbell, 1975), and the modulation frequency of PR is synchronized with the luminance-modulation frequency of the visual stimulus. ...
... Pupil diameter changes steadily with the illuminance of the observed object to regulate the amount of light entering the eye (Crawford, 1936;Woodhouse, 1975;Woodhouse and Campbell, 1975), and the modulation frequency of PR is synchronized with the luminance-modulation frequency of the visual stimulus. The amplitude of PR decreases as the stimulation frequency increases (Muto et al., 2020), and the consistent, measurable PR can be induced at the flickering frequency up to 2.3 Hz (Naber et al., 2013). Compared with the detection of gaze position Yao et al., 2018), the measurement of PR does not require system calibration. ...
... They also established a binary communication based on PR and achieved an accuracy of 100% at 10 bpm and 96% at 15 bpm. Muto et al. (2020) realized an information input interface with 12 options (from 0.58 to 1.90 Hz, with an interval of 0.12 Hz) based on PR. The averaged power spectral density (PSD) peak decreased with increasing luminance-modulation frequency, and the averaged classification accuracy reached 85.4% with a data length of 7 s. ...
Article
Full-text available
Brain-computer interface (BCI) based on steady-state visual evoked potential (SSVEP) has been widely studied due to the high information transfer rate (ITR), little user training, and wide subject applicability. However, there are also disadvantages such as visual discomfort and “BCI illiteracy.” To address these problems, this study proposes to use low-frequency stimulations (12 classes, 0.8–2.12 Hz with an interval of 0.12 Hz), which can simultaneously elicit visual evoked potential (VEP) and pupillary response (PR) to construct a hybrid BCI (h-BCI) system. Classification accuracy was calculated using supervised and unsupervised methods, respectively, and the hybrid accuracy was obtained using a decision fusion method to combine the information of VEP and PR. Online experimental results from 10 subjects showed that the averaged accuracy was 94.90 ± 2.34% (data length 1.5 s) for the supervised method and 91.88 ± 3.68% (data length 4 s) for the unsupervised method, which correspond to the ITR of 64.35 ± 3.07 bits/min (bpm) and 33.19 ± 2.38 bpm, respectively. Notably, the hybrid method achieved higher accuracy and ITR than that of VEP and PR for most subjects, especially for the short data length. Together with the subjects’ feedback on user experience, these results indicate that the proposed h-BCI with the low-frequency stimulation paradigm is more comfortable and favorable than the traditional SSVEP-BCI paradigm using the alpha frequency range.
... For a pupillary light reflex-based BCI with singlefrequency coding, Muto et al achieved over 98% discrimination performance when the difference in the blinking frequency between the stimuli was larger than ∆f = 0.12 Hz [53]. When the range of the stimulation frequency was 0.7-1.9 ...
Article
Full-text available
Objective: Recently, pupil oscillations synchronized with steady visual stimuli were used as input for an interface. The proposed system, inspired by a brain-computer interface based on steady-state visual evoked potentials, does not require contact with the participant. However, the pupil oscillation mechanism limits the stimulus frequency to 2.5 Hz or less, making it hard to enhance the information transfer rate. Approach: Here, we compared multiple conditions for stimulation to increase the information transfer rate (ITR) of the pupil vibration-based interface, which were called monocular-single, monocular-superposed, and binocular-independent conditions. The binocular-independent condition stimulates each eye at different frequencies respectively and mixes them by using the visual stereoscopic perception of users. The monocular-superposed condition stimulates both eyes by a mixed signal of two different frequencies. We selected the shape of the stimulation signal, evaluated the amount of spectral leakage in the monocular-superposed and binocular-independent conditions, and compared the power spectrum density at the stimulation frequency. Moreover, 5, 10, and 15 patterns of stimuli were classified in each condition. Main results: A square wave, which causes an efficient pupil response, was used as the stimulus. Spectral leakage at the beat frequency was higher in the monocular-superposed condition than in the binocular-independent one. The power spectral density of stimulus frequencies was greatest in the monocular-single condition. Finally, we could classify the 15-stimulus pattern, with ITRs of 14.4 (binocular-independent, using 5 frequencies), 14.5 (monocular-superimposed, using 5 frequencies), and 23.7 bits/min (monocular-single, using 15 frequencies). There were no significant differences for the binocular-independent and monocular-superposed conditions. Significance: This paper shows a way to increase the number of stimuli that can be simultaneously displayed without decreasing ITR, even when only a small number of frequencies are available. This could lead to the provision of an interface based on pupil oscillation to a wider range of users.
Article
Full-text available
The processing of emotional signals usually causes an increase in pupil size, and this effect has been largely attributed to autonomic arousal prompted by the stimuli. Additionally, changes in pupil size were associated with decision making during non-emotional perceptual tasks. Therefore, in this study we investigated the relationship between pupil size fluctuations and the process of emotion recognition. Participants heard human nonverbal vocalizations (e.g., laughing, crying) and indicated the emotional state of the speakers as soon as they had identified it. The results showed that during emotion recognition, the time course of pupil response was driven by the decision-making process. In particular, peak pupil dilation betrayed the time of emotional selection. In addition, pupil response revealed properties of the decisions, such as the perceived emotional valence and the confidence in the assessment. Because pupil dilation (under isoluminance conditions) is almost exclusively promoted by norepinephrine (NE) release from the locus coeruleus (LC), the results suggest an important role of the LC-NE system during emotion processing.
Article
Full-text available
The purpose of this study is to determine the pupillary dynamics with periodical flashes from a peripheral glare source, in similar conditions to night driving, while focusing on dependence with age. We measured two groups of people: youth and adults. Maximum pupil size decreases due to periodic flashes. Latency does not present significant differences. The reduction of pupil size is greater for older adults. The presence of a peripheral and periodic glare source modifies the pupil size. This leads to a reduction of retinal illuminance, which is greater for older adults.
Article
Full-text available
We present a new human-computer interface that is based on decoding of attention through pupillometry. Our method builds on the recent finding that covert visual attention affects the pupillary light response: Your pupil constricts when you covertly (without looking at it) attend to a bright, compared to a dark, stimulus. In our method, participants covertly attend to one of several letters with oscillating brightness. Pupil size reflects the brightness of the selected letter, which allows us-with high accuracy and in real time-to determine which letter the participant intends to select. The performance of our method is comparable to the best covert-attention brain-computer interfaces to date, and has several advantages: no movement other than pupil-size change is required; no physical contact is required (i.e. no electrodes); it is easy to use; and it is reliable. Potential applications include: communication with totally locked-in patients, training of sustained attention, and ultra-secure password input.
Article
Full-text available
The eye’s pupils constrict (shrink) in brightness, and dilate (expand) in darkness. The pupillary light response was historically considered a low-level reflex without any cognitive component. Here we review recent studies that have dramatically changed this view: The light response depends not only on a stimulus' brightness, but also on whether you are aware of the stimulus, whether you are paying attention to it, and even whether you are thinking about it. We highlight the link between the pupillary light response and eye-movement preparation: When you intend to look at a bright stimulus, a pupillary constriction is prepared along with the eye movement, before the eyes set in motion. This preparation allows the pupil to rapidly change its size as your eyes move from bright to dark objects and back again. We discuss the implications of these recent advances for our understanding of the subtle yet important role that pupillary responses play in vision.
Article
Full-text available
The muscles that control the pupil are richly innervated by the autonomic nervous system. While there are central pathways that drive pupil dilations in relation to arousal, there is no anatomical evidence that cortical centers involved with visual selective attention innervate the pupil. In this study, we show that such connections must exist. Specifically, we demonstrate a novel Pupil Frequency Tagging (PFT) method, where oscillatory changes in stimulus brightness over time are mirrored by pupil constrictions and dilations. We find that the luminance-induced pupil oscillations are enhanced when covert attention is directed to the flicker stimulus and when targets are correctly detected in an attentional tracking task. These results suggest that the amplitudes of pupil responses closely follow the allocation of focal visual attention and the encoding of stimuli. PFT provides a new opportunity to study top-down visual attention itself as well as identifying the pathways and mechanisms that support this unexpected phenomenon.
Article
Full-text available
The pupillary light response is often assumed to be a reflex that is not susceptible to cognitive influences. In line with recent converging evidence, we show that this reflexive view is incomplete, and that the pupillary light response is modulated by covert visual attention: Covertly attending to a bright area causes a pupillary constriction, relative to attending to a dark area under identical visual input. This attention-related modulation of the pupillary light response predicts cuing effects in behavior, and can be used as an index of how strongly participants attend to a particular location. Therefore, we suggest that pupil size may offer a new way to continuously track the focus of covert visual attention, without requiring a manual response from the participant. The theoretical implication of this finding is that the pupillary light response is neither fully reflexive, nor under complete voluntary control, but is instead best characterized as a stereotyped response to a voluntarily selected target. In this sense, the pupillary light response is similar to saccadic and smooth pursuit eye movements. Together, eye movements and the pupillary light response maximize visual acuity, stabilize visual input, and selectively filter visual information as it enters the eye.
Article
Full-text available
For patients with severe motor disabilities, a robust means of communication is a crucial factor for their well-being [1]. We report here that pupil size measured by a bedside camera can be used to communicate with patients with locked-in syndrome. With the same protocol we demonstrate command-following for a patient in a minimally conscious state, suggesting its potential as a diagnostic tool for patients whose state of consciousness is in question. Importantly, neither training nor individual adjustment of our system's decoding parameters were required for successful decoding of patients' responses.
Conference Paper
We have developed a new mono/binocular eye tracking system by using an IEEE1394b or USB-3.0 digital camera that provides high sensitivity, high resolution, high frame-rate and no rolling shutter distortion. Our goal is to provide a system that is friendly to researchers who conduct experiments. The system is non-invasive and inexpensive and can be used for mice, marmosets, monkeys, and humans. It has adopted infrared light to illuminate an eye (eyes). The reflected image of the infrared light on the cornea and the black image of the pupil are captured by the camera. The center of the pupil and the center of the corneal reflection are calculated and tracked over time. The movement of the head is compensated by the reflection. Since the high resolution camera has a 2048 horizontal pixels resolution, we can capture the images of both eyes simultaneously and calculate the parameters of the two eyes at each frame. The gaze position data can be read out on-line via computer network and/or DAC (digital analog converter). The adoption of the Windows 10 as the operation system makes this eye tracking system user-friendly. Because of the high frame-rate of the digital camera, the sampling rate of the system can be as high as 700 Hz and the latency less than 4 ms.
Chapter
The accepted classical view of the pupil response to light is that the ambient light level determines largely the steady-state size of the pupil (Lowenstein et al, 1964) and that rapid increments in light flux on the retina cause a brisk constriction of the pupil, that is often described as the dynamic pupil light reflex response (Stark & Shermann, 1957; Alexandridis, 1985; Lowenfeld, 1993). The afferent pathways involved in the control of the pupil in man have been associated with subcortical projections and this is consistent with clinical observations which suggest that the pupils continue to respond normally to sudden changes in room illumination even when the patients are corticaliy blind (Brindley et al, 1969). The response of the pupil often represents an important measurement in neurological and ophthalmologic examinations, but the poor understanding of the visual pathways involved and the kind of stimulus characteristics that cause pupillary responses limit the usefulness and potential use of such tests. Animal studies have contributed significantly to our understanding of the afferent neural pathways that subserve the Pupil Light Reflex response (PLR) in man and helped to reinforce the accepted classical view. Lesion studies combined with measurements of pupil responses before and after surgery in monkey have demonstrated the direct invol vement of the pretectal olivary nuclei and the Edinger-Westphal nuclei in both the steady state control of pupil size and the dynamic PLR response (Pierson & Carpenter, 1974). In rat, the control of the pupil response is somewhat less complicated and functionally more consistent with the regulation of retinal illuminance (Clarke & Ikeda, 1985a, b). The pupil diameter follows an almost linear relationship when plotted as a function of the logarithm of retinal illuminance over a wide range.