Conference PaperPDF Available

First Eye Movement Verification and Identification Competition at BTAS 2012

Authors:

Abstract and Figures

This paper presents the results of the first eye movement verification and identification competition. The work provides background, discusses previous research, and describes the datasets and methods used in the competition. The results highlight the importance of very careful eye positional data capture to ensure meaningfulness of identification outcomes. The discussion about the metrics and scores that can assist in evaluation of the captured data quality is provided. Best identification results varied in the range from 58.6% to 97.7% depending on the dataset and methods employed for the identification. Additionally, this work discusses possible future directions of research in the eye movement-based biometrics domain.
Content may be subject to copyright.
First Eye Movement Verification and Identification
Competition at BTAS 2012
Paweł Kasprowski
Institute of Informatics
Silesian University of Technology
Gliwice, Poland
kasprowski@polsl.pl
Oleg V. Komogortsev, Alex Karpov
Department of Computer Science
Texas State University-San Marcos
San Marcos, USA
{ok11,ak26}@txstate.edu
Abstract—This paper presents the results of the first eye
movement verification and identification competition. The work
provides background, discusses previous research, and describes
the datasets and methods used in the competition. The results
highlight the importance of very careful eye positional data
capture to ensure meaningfulness of identification outcomes. The
discussion about the metrics and scores that can assist in
evaluation of the captured data quality is provided. Best
identification results varied in the range from 58.6% to 97.7%
depending on the dataset and methods employed for the
identification. Additionally, this work discusses possible future
directions of research in the eye movement-based biometrics
domain.
Index Terms—eye movements, biometrics, competition
I. INTRODUCTION
There are several ways to use eye related information for
biometric purposes, e.g., iris [4][26], face recognition [41],
retina [29], periocular information [39]. One of the additional
biometric modalities related to the eye is biometrics based on
eye movements. This biometric modality was suggested
approximately 10 years ago [11][38], however, relatively few
publications were written on this topic so far. To facilitate
research in this area we have decided to organize an eye
movement biometric competition.
The competition provided a common ground in a form of
several datasets to benchmark the eye movement biometric
methods derived by the participants. Our subsequent work
with the results and the datasets allowed providing
recommendations related to the eye movement data collection,
measuring eye movement quality, and deciding when to
record samples from the subjects to ensure meaningfulness of
current and future benchmarking results.
This paper is organized as follows: Section II provides
information related to the eye movements and their origin,
also the information about eye movement data recording and
assessing its quality; Section III briefly describes previous
work related to the eye movement biometrics; Section IV
provides the details of the datasets employed in the
competition; Section V describes the competition protocol;
Section VI outlines the results; Section VII provides a
discussion about results and related data quality; Section VIII
follows with a conclusion and future work.
II. EYE MOVEMENTS
A. Overview
Human eyes differ substantially from a common digital
camera [22]. One of the differences is a non-uniform picture
quality across the visual field. Specifically, the fovea – the
high visual acuity zone of human retina is just approximately
two degrees of the visual angle [28]. The acuity of vision
sharply drops outside of the fovea. The center of the fovea -
foveola is the central part of the retina that provides highest
visual acuity. Visual axis can be extended from the foveola to
the object perceived by the eye, thus creating a point of regard
also called a gaze point. Non-uniformity of human vision
necessitates eye movements with a goal of capturing visual
information from the surrounding world.
Among different eye movements exhibited by the Human
Visual System (HVS) following two types are most relevant to
this work: fixation – eye is stable toward the object of interest,
saccade – rapid eye rotation between fixation points.
Miniature eye movements such as tremor, drift, and micro
saccades [42] are a part of a fixation and keep an eye in
constant motion. Constant movement is necessary due to the
motion sensing nature of the light perceiving cells, which
require constant excitation to translate the light to the neuronal
signal [22]. Artificial stabilization of the eye globe leads to the
loss of vision. Saccades are the fastest movements in the
human body with velocities reaching several hundred degrees
per second [37]. The eye is blind during saccades.
Eye movements may be divided into voluntary and
involuntary. Voluntary eye movements are result of our will,
making it possible to control the focus of our attention.
Involuntary eye movement is a reflexive action, automatic
response to some stimulus, for instance sudden stimulus
movement near the edge of vision.
Physiologically, the eye movements are made possible by a)
oculomotor plant – eye globe, three pairs of extraocular
muscles, and surrounding tissues [37], b) different brain areas
and mechanisms are responsible for programming oculomotor
plant [37]. Extensive description of the related anatomical
structures and their functionality are beyond the scope of this
This is a pre-print. The paper will be published in the proceedings of the IEEE
Fifth International Conference on Biometrics: Theory, Applications and Systems
(BTAS 2012)
work.
B. Gaze Data Recording and Quality Assessment
Eye movements are recorded by a device called an eye
tracker, which reports raw eye positional data at specified
sampling frequency [22]. Main characteristics of the eye
tracking equipment are: a) positional accuracy – the difference
between the reported and the actual gaze point, b) precision –
minimum amount of gaze shift detectable by the equipment, c)
sampling frequency. Detailed survey of the different eye
tracking approaches can be found here [23].
Currently, two main metrics are widely accepted as quality
indicators of captured raw eye positional data: calibration
error and data loss.
Calibration error is determined during a calibration
procedure. Calibration procedure is very important in the eye
tracking research. Its goal is to train eye tracking software to
estimate eye gaze position for every eye image captured by the
image sensor. This goal is achieved by a presentation of pre-
set target points that are usually uniformly distributed on the
visual screen and requesting a subject to look at these
predefined gaze locations [22]. Subsequently, when actual
recording takes place, the gaze locations that fall outside of
the initial calibration points are interpolated by various
algorithms [23]. Calibration error indicates the average
positional difference between the coordinates of the pre-set
calibration points and the coordinates of the estimated gaze
locations for those points. Calibration error varies depending
on a subject and experiment setup. Calibration error is
expected to be close to equipment’s positional accuracy,
however, quite often the calibration error can be several times
larger than positional accuracy reported by an eye-tracking
vendor. Please note that calibration error can be also termed as
accuracy or positional accuracy in the eye tracking literature
(e.g. [27][35]).
Data loss is the amount of gaze samples reported as invalid
by an eye tracker. Data loss is usually caused by blinking,
head movements, changes in stimulus or surrounding lighting,
and squinting. In cases when gaze points fall outside of the
recording boundaries (e.g., computer screen), which usually
happens due to poor calibration, these gaze points are marked
as invalid. Please note that not all eye trackers are capable of
marking invalid gaze points. In such cases invalid gaze points
should be found and marked by the experimenter to compute
resulting data loss.
Usually smallest calibration error and data loss for the eye
tracking systems are achieved when subject’s head is fixated
by a chin rest at an optimal distance from the image sensor
(usually 40-70cm.). However, modern advances in the table
and head mounted eye tracking systems allow to collect
acceptable data quality when the head is not fixated.
More detailed discussion about gaze data quality can be
found here [27].
C. Classification of Captured Gaze Data Into Fixations and
Saccades
Raw eye positional signal should be classified into fixations
and saccades (and other eye movement types when stimulus
properties are likely to invoke them) in cases when it is
necessary to assess performance of the Human Visual System
(HVS) or to employ eye movement characteristics for
biometric purposes. Several algorithms exist for classification
purposes [33]. Classification of the raw gaze data into fixation
and saccades also allows assessing the meaningfulness of the
captured data via behavior scores that are discussed next.
D. Behavior Scores
Behavior scores provide the capability to assess the
performance of the HVS, which is represented by the results
provided by the eye movement classification algorithms
[35][33]. Behavior scores support the idea that the HVS
performance of a normal person matches signal characteristics
encoded in the stimulus. Cases where HVS performance
drastically differs from the ideal response of a normal person
might be indicative of pathologies (e.g., head trauma), poor
quality of the recorded signal, or/and failed eye movement
classification. Currently, behavior scores can be computed
only for pulse-step or pulse-step-ramp stimulus (moving dot of
light) with known characteristics [33][31].
Following behavior scores are employed for data quality
control in this work: Fixation Quantitative Score (FQnS) –
measures the amount of detected fixational behavior in
response to a stimulus, Fixation Qualitative Score (FQlS)
compares the spatial proximity of the classified fixation signal
to the presented stimulus signal. The FQlS usually highly
correlates with the calibration error. Saccade Quantitative
Score (FQnS) – measures the amount of detected saccadic
behavior in response to a stimulus. Detailed discussion on the
behavior scores and their ideal values can be found here [33].
III. EYE MOVEMENT BIOMETRICS: PREVIOUS WORK
Related work in the eye movement biometrics field can be
approximately divided into four general categories:
a) Use of the raw eye positional signal and its derivatives
[11][1][12][15]. Standard techniques for feature extraction are
used including first/second derivatives of the signal, Wavelet
transform, Fourier transform, and other frequency related
transformations of the signal. Frequently methods such as
Principal Component Analysis are employed to reduce the
number of features. Template matching is done using such
algorithms as K Nearest Neighbors, Naive Bayes, C45
Decision Trees and Support Vector Machines.
b) Use of oculomotor plant characteristics (OPC) [17],
where OPC are extracted with the help of a mathematical
model of the eye and OPC templates are matched by statistical
methods such as Hotelling’s T-square test;
c) Inference of brain control strategies, and mechanisms
responsible for guidance of visual attention via analysis of
complex eye movement patterns (CEM) [24][25]. In CEM
approach features are represented by several individual and
aggregated characteristics that are fundamental to HVS with
template matching occurring via probabilistic and distance
related approaches. It is possible to include in this category
approach investigated by Rigas and colleagues [20], where
Minimum Spanning Tree structures representing sequences of
fixation/saccades and distances between these structures are
employed for distinguishing individuals.
d) There is also effort that investigates combined
performance of OPC with CEM [36], and also multimodal
ocular biometrics approach where OPC, CEM, and iris
information are combined together [34].
IV. DATASETS FOR THE COMPETITION
Two types of the datasets were employed for the
competition: uncalibrated and calibrated. Uncalibrated
datasets were collected with a goal of minimizing data capture
time, i.e., without calibration procedure or equipment
adjustments. Therefore, captured data quality could not be
controlled. Calibrated datasets were captured with every
precaution to obtain highest possible data quality, i.e.,
equipment was adjusted and re-adjusted when necessary to
ensure minimum calibration error and data loss. Depending on
a subject, necessary iterative equipment adjustments,
subsequent calibration, and its verification could sometimes
increase data capture times considerably. Table I provides
technical details for each database. Next two sections provide
additional details.
A. Uncalibrated Datasets
Dataset A. Step stimulus was presented in a form of a
jumping dot interpolated on the 3x3 grid. The stimulus
consisted of eleven dot position changes giving twelve
consecutive dot positions. Subjects were given instructions to
follow the dot. First dot appeared in the middle of the screen.
After 1600 ms the dot in the middle disappeared and for 20 ms
a screen was blank. Subsequently, a dot appeared in the upper
right corner. The sequence continued until all locations of the
3x3 grid were visited. Dot movements on the grid were
interspersed with dots presented at the central screen location.
Figure 1 provides additional stimulus details. Figure 2
presents a sample from the recorded signal.
Maximum of 10 recordings per day were conducted per
subject. Figure 3 presents a histogram of time intervals
between the recordings.
Dataset B. Step stimulus was presented in a form of a
jumping dot interpolated on the 2x2 grid. Dot’s position
changed after every 550 ms. Additionally there was 550 ms
"break" with no dot (black screen) in the middle of
stimulation.
Maximum of 10 recordings per day were conducted per
subject. Figure 3 presents a histogram of time intervals
between the recordings.
B. Calibrated Datasets
Dataset C. For each recording session the step stimulus
was presented as a vertical jumping dot, consisting of a grey
disc sized approximately 1 degree of visual angle with a small
black point in the center. The dot performed 120 vertical
jumps with amplitude of 20 degrees of the visual angle. At
each spot the dot was displayed for 1s. The first two
recordings for each subject were conducted during the same
day with an interval of approximately 15 minutes; two more
recording were conducted one week later during a single day
with an interval of approximately 15 minutes. Chin rest was
employed to stabilize subjects' heads during the recording.
Figure 2 presents a sample from the recorded signal. Figure 3
presents a histogram of time intervals between the recordings.
Dataset D. For each recording session the dot appeared 100
times each time at random location with only requirement that
at the end of the 100 appearances the spatial placement of the
dots on the screen would be close to uniform. Other recording
parameters are the same as for the Dataset C. Videos depicting
stimulus for Datasets C and D can be found at
www.emvic.org. Figure 3 presents a histogram of time
intervals between the recordings. Datasets C and D are a part
of a larger biometric database that can be downloaded here
[32].
Fig. 1. Dataset A stimulus and timeline.
Fig. 2. Excerpt of the recorded signal from the Dataset A and Dataset C.
Fig. 3. Time interval between the recordings of each individual. The data
from all individuals is aggregated into the graph.
0%
10%
20%
30%
40%
50%
60%
70%
80%
10
sec
30
sec
60
sec
1
hour
1
day
1
week
Time
Interval
Between
Recordings
Dataset
A
Dataset
B
Dataset
C
Dataset
D
TABLE I. TECHNICAL DETAILS FOR EACH DATASET. BEHAVIOR SCORES
FOR DATASETS A AND B WERE COMPUTED AFTER POST CALIBRATION [13].
PRIOR TO SCORE COMPUTATION SOME RECORDINGS IN DATASET S A AND B
WITH VERY POOR DATA QUALITY WERE REMOVED. DATA LOSS FO R DATASET S
A AND B WERE COMPUTED WHEN SIGNAL WAS DETECTED OUT OF SCREEN
BOUNDARIES.
Characteristic Dataset A Dataset B Dataset C Dataset D
# of subjects 37 75 29 27
Total # of
recordings 978 4168 116 108
Recordings per
subject 4-158 5-172 4 4
Recording
duration 8.2s 8.2s 120s 100s
Maximum
time between
the first and
the last
recording
3 months 4 months 12 days 12 days
Experimental setup
Eye Tracker
(ET) Ober2 Ober2
Eye Link
1000
Eye Link
1000
ET accuracy N/A N/A 0.25°-0.5° 0.25°-0.
ET precision N/A N/A 0.01° 0.01°
ET sampling
rate 250Hz 250Hz 1000Hz 1000Hz
ET type head
mounted
head
mounted remote remote
Chin rest no no yes yes
Display size 17’’ 17’’ 30’’ 30’’
Distance head
to screen 700mm 700mm 685mm 685mm
Data Quality
Average
calibration
error (SD)
N/A N/A 0.73° (0.39) 0.75° (0.55)
Average Data
Loss (SD)
9.06%
(14.53)
35.19%
(25.45)
2.88%
(0.04)
2.88%
(0.04)
Behavior scores
Ideal_FQnS 66% 63% 74% 75%
FQnS (SD) 19% (13.4) 10% (9) 59% (9.2) 65% (8.2)
Ideal_FQlS 0° 0° 0° 0°
FQlS (SD) 2.14° (0.55) 2.11° (0.59) 1.11° (0.42) 1.38° (0.33)
Ideal_SQnS 100% 100% 100% 100%
SQnS (SD) 116% (34) 149% (50) 108% (62) 116% (31)
C. Separation into Training and Testing Sets
Datasets A and B were separated into the training set that
contained 65% of the recordings and the testing set that
contained 35% of the recordings. The assignment of
recordings to each set was made by random stratified
sampling.
Datasets C and D were separated into the training set and
testing sets via 50% / 50% split. The testing set contained the
recordings that were performed during the first week of the
experiments and thus each subject was represented by two
recordings. The training set contained the recordings that were
performed during the second week of experiments and
contained remaining two recordings for each subject.
V. COMPETITION PROTOCOL
The aim of the competitors was to build their classification
models using labeled recordings in the training sets and then
try to use those models to classify unlabeled recordings in the
testing sets.
In many similar competitions (e.g. Fingerprint Verification
Competition [43]) competitors should send the organizers an
application that is able to take given input data and produce
output in the specified format. Usually, when ranking the
performance of such applications both the execution time and
the accuracy are taken into the consideration. However, we
decided to simplify the submission process. Competitors had
to send only a file in pre-defined format that contained the
identity of each individual in the testing set. Submission
format allowed for strong classification (only one identity for
each recording) or weak (multiple identities each marked with
a probability value).
The competition was broken into two parts: the main
competition (www.emvic.org) that consisted of all four
datasets and an additional competition on Kaggle web service
(http://www.kaggle.com/c/emvic). The Kaggle’s
part was simplified and consisted of the dataset A only with a
slightly different submission file format as required by the
host.
The main competition required for every recording a list of
probabilities that a given recording belongs to a specific
subject (sid) in format sid1:prob,sid2:prob. The
number of sid:prob pairs was not specified. It could be
just one sid:1 for the strong classifier case or a list of all
sids and probabilities in case of the weak classifier. We
encouraged competitors to send sets of sid:prob pairs to
enable the extraction of various performance parameters,
however, very few participants sent this information.
The competition employed rank-1 identification accuracy
benchmark marked as ACC1. It was computed as the number
of records classified correctly to the whole number of records.
Correct classification is when correct sid is marked by the
highest probability.
For Kaggle competition the file format was slightly
different, but contained the same information. Log loss metric
was suggested by Kaggle project administration and was
adopted as a performance benchmark.
Log loss is defined as:
()
∑∑
==
= N
i
M
jjiji yy
N
loss 11 ,, ^log
1
log (1)
where N is the number of samples, M is the number of
subjects, log is the natural logarithm, yi,j^ is the posterior
probability that the jth subject generated the ith sample, and yi,j
is the ground truth (yi,j=1 means that the jth subject generated
the ith sample, yi,j=0 indicates otherwise).
Log loss metric guarantees that submissions in which
correct sids are marked by the high probability (not necessary
the highest one) get better score.
Kaggle competition participants were limited to two daily
submissions. Table II indicates that substantial number of
attempts (frequently more than 20) was required to obtain best
results. To minimize over-fitting only 25% of the test dataset
A (so called “public” part) was given for these trials.
Competitors sending their solutions were aware of their public
score only. The final score was calculated using the whole
testing dataset. Whole testing Dataset A frequently yielded
worse log loss scores.
VI. COMPETITION RESULTS
There were overall 49 competitors in Kaggle hosted
competition and 45 registered users in the main competition.
There were 524 submissions sent in Kaggle competition and
106 in the main. Tables II and III present the summary of the
results.
As a part of the competition participants filled out a survey
where they had an opportunity to provide an account of
methods and tools they employed to achieve their results.
Most of the participants (especially “kagglers”) treated the
data in the individual recordings as a sequence of numbers and
treated those sequences via general signal processing and data
mining algorithms. Except raw eye positional signal only
features related to the first and second derivate of the signal (i.
e. velocity and acceleration) were considered. Extraction of
fixations and saccades from raw eye positional signal and
employment of any features related to those events was not
reported. The only reported exception was signal parsing into
segments via fixations.
Because the number of attributes extracted from the signal
was substantial, most of the competitors used some techniques
to reduce dimensionality. The most popular techniques
involved dividing a recording into some subsets (typically 8 to
16) and summarizing these subsets by their characteristics,
e.g., average velocity, average spatial location. Among others,
methods such as Singular Value Decomposition or/and
Principal Component Analyses led to a reasonable reduction
of dimensionality. Among various techniques employed for
the template matching SVM, Random Forest and k Nearest
Neighbors methods were most popular and produced highest
identification accuracy results.
TABLE II. TOP FIVE RESULTS OF THE KAGGLES COMPETITION. EACH
SCORE REPRESENTS LOG LOSS.
Team Method Entries Public score Final score
IRIG kNN 28 0,23 0,23
Killian O. Random Forest 18 0,12 0,37
Dorothy Random Forest +
LDA 24 0,18 0,48
zeon n/a 24 0,33 0,52
GeLo n/a 10 0,33 0,59
TABLE III. TOP RESULTS FOR MAIN COMPETITION. BEST RESULTS FOR
EACH DATASET ARE BOLDED.
Participant Methodology ACC1
Dataset A
Michal Hradiš, Brno
University of Technology
2D histogram speed and
direction, SVM 97,55%
Ioannis Rigas, University of
Patras, Greece
Multivariate Wald-Wolfowitz
test, kNN 96,63%
Nguyen Viet Cuong,
Nation.Univ. of Singapore
Bayesian Network with Mel-
frequency cepstral coefficients
features
93,56%
Dataset B
Michal Hradiš, Brno
University of Technology
2D histogram speed and
direction, SVM 95,11%
Ioannis Rigas, University of
Patras, Greece
Multivariate Wald-Wolfowitz
test, kNN 90,43%
Nguyen Viet Cuong, National
Univ of Singapore
SVM with Mel-frequency
cepstral coefficients features 90,43%
Dataset C
Michal Hradiš, Brno
University of Technology
Nearest neighbor with X2
distance. 2D histograms on short
windows - speed and direction.
58,62%
Nguyen Viet Cuong, National
Univ. of Singapore
SVM with Mel-frequency
cepstral coefficients features +
PCA
37,93%
Ioannis Rigas, University of
Patras, Greece
Multivariate Wald-Wolfowitz
test, kNN 25,86%
Dataset D
Michal Hradiš, Brno
University of Technology
Nearest neighbor with X2
distance. 2D histograms on short
windows - speed and direction.
66,67%
Nguyen Viet Cuong, National
Univ. of Singapore
SVM with Mel-frequency
cepstral coefficients features +
split data to 10 segments
48,15%
VII. DISCUSSION
A. Result Differences for Calibrated and Uncalibrated
Datasets
Competition results (Table III) indicate that there are
substantial biometric accuracy differences between
uncalibrated and calibrated datasets presented in this work. It
is very important to discuss possible causes for these
differences.
1) Calibration Impact
The difference between the ideal and computed behavior
scores discussed above for the uncalibrated datasets indicate
that the recorded data contains substantial amount of noise and
positional error. Similar conclusion can be made after the
manual inspection of the data in the post-calibrated form, e.g.,
it can be seen from Fig. 2 that for Dataset A the eye positional
signal is quite far from the stimulus and that signals from each
eye are far apart. These outcomes highlight current danger in
employing uncalibrated data for the biometric purposes. The
eye movement data collected in the uncalibrated form might
contain unique noise introduced by a plethora of individual
subject related parameters such as equipment position, head
position, lightning, etc. that might bias identification
outcomes.
2) Impact of Recording Patterns
Histograms presented on Fig. 3 for the uncalibrated datasets
indicate that the majority of the recordings were conducted
with very close temporal proximity of each other, i.e., 95%
recordings of Dataset A and 83.83% recordings of the Dataset
B were conducted within 60 seconds of each other. We
hypothesize that such recording arrangement made it possible
for the unique subjects’ related noise characteristics discussed
above to be translated to the high identification accuracy.
3) Impact of Other Factors
Other factors that might have contributed to the
identification accuracy differences: 1) uncalibrated datasets
contained the information from both eyes, while calibrated
datasets contained the information from one eye only, 2)
uncalibrated datasets contained more recordings per subject,
which coupled together with the unique noise and recording
patterns produced very high identification accuracy, 3) Dataset
A was available to the participants for the longest period of
time. Substantial amount of re-submissions was done for the
Dataset A and B, providing more opportunities to improve the
results.
For calibrated datasets particular low identification
accuracy result for dataset C can be in part explained by the
fixed vertical stimulus, which reduced the amount of HVS
performance variability that exist between subjects in case of
non-fixed stimulus. This hypothesis is supported by the fact
that for random stimulus presented for the dataset D the
identification accuracy was improved dramatically. We
hypothesize that random stimulus provided an opportunity to
capture subject related HVS variability in the periphery,
where, for example, the impact of nonlinearities present in the
oculomotor plant structure are more pronounced.
B. Recommendations for Creating Future Eye Movement
Databases
We encourage following recommendations when creating
eye movement databases for the biometric testing.
1) Monitoring & Reporting Data Quality
Perform a calibration prior to a recording of each individual
record. Make sure that the average calibration error does not
exceed 1.5 degrees of the visual angle for each eye that is
being recorded. In cases when calibration error exceeds 1.5° it
is important to re-adjust equipment’s settings and re-calibrate
a person to ensure that calibration error below suggested value
is obtained prior to the actual recording. Some additional
information about the impact of calibration error on the
resulting eye movement biometric performance is discussed
here [25]. We suggest that data loss should be kept at a
minimum by careful adjustment of the recording setup. We
also hypothesize that data loss up to 10% is reasonable to
ensure high validity of the captured data. However, currently,
there are no detailed studies that measure performance
tradeoffs between the data loss and the corresponding eye
movement biometric accuracy.
In cases when stimulus performance is fixed (i.e.,
predefined pulse-step stimulus) we suggest reporting behavior
scores. Such information would allow assessing: 1) quality of
the recorded data, 2) performance of the eye movement
classification methods in cases when separation of the raw eye
positional signal into fixations and saccades is necessary, 3)
HVS performance. In cases when normal population is
recorded (no HVS pathologies) the behavior scores tend to be
close to their ideal values which are discussed in detail in
[33][35].
2) Temporal Separation of Individual Records
We already discussed the impact of the recordings’
temporal proximity on the accuracy results where temporal
clustering coupled together with the unique noises (e.g.,
Datasets A and B) might lead to the artificially high
identification accuracy. There are no hard guidelines in the
biometric literature that suggest ideal temporal proximity
between the recordings of an individual to validate a new
biometric modality. We can hypothesize that validation
strategies where the times of the recordings for each person
have a uniform temporal spread, e.g., recordings done once
every hour in a day, once per day, once per month, would
allow to minimize biases related to the equipment setup,
stress, fatigue, illness, and drug related effects.
VIII. SUMMARY & FUTURE WORK
Eye movement based biometrics is an emerging field. The
organized competition allowed exploring current state of the
art in terms of the applicable biometric algorithms, outlining
dangers associated with the different approaches related to the
data capture, and providing suggestions for the future creation
of the eye movement biometric databases.
The results of the competition indicate that the information
about the eye movements can be exploited for the biometric
purposes. Current accuracy results can be compared to the
accuracy of the early face recognition systems [30][40],
however we hypothesize that future work will improve current
performance levels. Additional work is required to statistically
measure between and within subject variability of the eye
movement related characteristics and the stability of the eye
movement traits in longitudinal recording where impact of
fatigue, aging, stress, and possible effect of various substances
is present.
As a result of the data processing associated with the
competition it was possible to find current dangers associated
with recording of the uncalibrated eye positional data.
Uncalibrated data makes it very difficult to control the quality
of the data capture, therefore making it possible to record
unique noises coming from the specific adjustments of the
equipment necessary to record a subject. Equipment
calibration and quality measurement might require additional
time and effort to capture the data, however, it is, currently,
the only way to carefully control what is being recorded and to
ensure the physiological validity of the data. When recording a
subject multiple times we suggest performing and validating a
calibration prior to each recording. Recordings should not be
clustered into a single recording session where a large number
of recordings of the same subject are clustered together. In the
best case, recordings should be spread over multiple days with
equal amount of recordings per day/month/year to avoid the
impact of biases introduced by the equipment setup and other
factors.
Competitions have the advantage of providing a common
ground for comparing the performance of the different
biometric methods and techniques. Therefore, we are planning
to organize future competitions that would provide such
opportunity and will push forward the state of the art of the
eye movement-based biometrics.
IX. ACKNOWLEDGMENT
Data collection for Datasets C and D and the competition
related work was supported in part by the National Institute of
Standards and Technology (NIST) under Grant
#60NANB10D213, and Texas State University-San Marcos
via internal grants.
REFERENCES
[1] R. Bednarik, T. Kinnunen, A. Mihaila and P. Fränti. Eye-
movements as a biometric, 14 Scandinavian Conference on
Image Analysis, Lecture Notes in Computer Science, Springer-
Verlag, vol. 3540, pp. 780-789, 2005.
[2] C.S. Campbell, P.P. Maglio. A robust algorithm for reading
detection. Proceedings of the ACM Workshop on Perceptual
User Interfaces, 2002.
[3] S. Dareddy. Eye Movemements Verification and Identification
Competition Survey, 2012, unpublished.
[4] J. Daugman. High Confidence Visual Recognition of Persons by
a Test of Statistical Independence. IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 15, no. 11 1993.
[5] A. Duchowski. A Breadth-First Survey of Eye Tracking
Applications. Behavior. Research Methods, Instruments &
Computers (BRMIC), 34(4), 2002.
[6] R. Engbert, A. Longtin, R. Kliegl. Complexity of Eye
Movements in Reading. International Journal of Bifurcation and
Chaos in Applied Sciences and Engineering, Vol. 14, No. 2,
2004
[7] J. M. Henderson, A. Hollingworth. Eye Movements and Visual
Memory: Detecting Changes to Saccade Targets in Scenes.
Michigan State University, Visual Cognition Lab, Rye Lab
Technical Report Tech Report, 2001.
[8] A. J. Hornof, T. Halverson. Cleaning up systematic error in eye
tracking data by using required fixation locations. Behavior
Research Methods, Instruments, and Computers, 34, 2002.
[9] G. K. Hung, Models of Oculomotor Control. World Scientific
Publishing Co., 2001.
[10] E. Javal. Physiologie de la lecture et de l’écriture. Paris: Félix
Alcan, 1905.
[11] P. Kasprowski, J. Ober. Eye movement tracking for human
identification, 6th World Conference BIOMETRICS’2003,
London, 2003.
[12] P. Kasprowski, J. Ober. Eye Movement in Biometrics,
Proceedings of Biometric Authentication Workshop, European
Conference on Computer Vision in Prague 2004, LNCS 3087,
Springer-Verlag., 2004.
[13] P. Kasprowski. Human identification using eye movements.
Doctoral thesis. http://www.kasprowski.pl/phd_ kasprowski.pdf.
Silesian Unversity of Technology, Poland, 2004.
[14] P. Kasprowski, J. Ober. Enhancing eye movement based
biometric identification method by using voting classifiers. SPIE
Defence & Security Symposium, SPIE Proceedings, Orlando,
Florida, 2005.
[15] T. Kinnunen, F. Sedlak, R. Bednarik. Towards task-independent
person authentication using eye movement signals. Proceedings
of the 2010 Symposium on Eye-Tracking Research &
Applications, ACM New York, NY, USA, 2010.
[16] O. Komogortsev, S. Jayarathna, C. R. Aragon, M. Mahmoud.
Biometric Identification via an Oculomotor Plant Mathematical
Model. Proceedings of the 2010 Symposium on Eye-Tracking
Research & Applications, ACM New York, NY, USA, 2010.
[17] O. V. Komogortsev, A. Karpov, L. Price, C. Aragon. Biometric
Authentication via Oculomotor Plant Characteristic.
Proceedings of the IEEE/IARP International Conference on
Biometrics (ICB), pp. 1-8, 2012.
[18] D. Norton, L. W. Stark. Scanpaths in eye movements during
pattern perception, Science, 171 308-311, 1971.
[19] M. Nishigaki, D. Arai. A user authentication based on human
reflexes using blind spot and saccades response, International
Journal of Biometrics, Vol. 1, No. 2, pp. 173-189, 2008.
[20] I. Rigas, G. Economou, S. Fotopoulos. Biometric identification
based on the eye movements and graph matching techniques,
Pattern Recognition Letters, Volume 33, Issue 6, 15 April 2012,
Pages 786-792, ISSN 0167-8655, 2012.
[21] B. S. Schnitzer, E. Kowler. Eye movements during multiple
readings of the same text. Vision Research, 46(10): 1611-1632,
2006.
[22] A. Duchowski. Eye Tracking Methodology: Theory and
Practice, Springer, 2007.
[23] D. W. Hansen, J. Qiang. In the Eye of the Beholder: A Survey
of Models for Eyes and Gaze. IEEE Transactions on Pattern
Analysis and Machine Intelligence 32(3): 478-500, 2010.
[24] C. Holland, O. V. Komogortsev. Biometric Identification via
Eye Movement Scanpaths in Reading, IEEE International Joint
Conference on Biometrics (IJCB), 2011.
[25] C. Holland, O. V. Komogortsev. Biometric Verification via
Complex Eye Movements: The Effects of Environment and
Stimulus. IEEE Fifth International Conference on Biometrics:
Theory, Applications and Systems (BTAS 2012).
[26] K. Hollingsworth, K. W. Bowyer, et al. All Iris Code Bits are
Not Created Equal. First IEEE International Conference on
Biometrics: Theory, Applications, and Systems, 2007. BTAS
2007.
[27] K. Holmqvist, M. Nystrom, et al. Eye tracker data quality: what
it is and how to measure it. Proceedings of the Symposium on
Eye Tracking Research and Applications. Santa Barbara,
California, ACM: 45-52, 2012.
[28] D. Irwin. Visual Memory Within and Across Fixations. Eye
Movements and Visual Cognition (Springer Series in
Neuropsychology). K. Raymer. New-York: pp. 146-165, 1992
[29] A. Jain, L. Hong et al.. Biometric identification. Commun. ACM
43(2): 90-98, 2000
[30] T. Kanade. Picture Processing System by Computer Complex
and Recognition of Human Faces. Ph. D. , Kyato University,
1973
[31] O. Komogortsev, A. Karpov. Automated classification and
scoring of smooth pursuit eye movements in the presence of
fixations and saccades. Behavior Research Methods: 1-13.,
2012.
[32] O. V. Komogortsev. Eye Movement Biometric Database v1.
2011, from http://www.cs.txstate.edu/~ok11/embd_v1.html.
[33] O. V. Komogortsev, D. V. Gobert, et al. Standardization of
Automated Analyses of Oculomotor Fixation and Saccadic
Behaviors. IEEE Transactions on Biomedical Engineering
57(11): 2635-2645, 2010.
[34] O. V. Komogortsev, C. Holland, et al. Multimodal Ocular
Biometrics Approach: A Feasibility Study. IEEE Fifth
International Conference on Biometrics: Theory, Applications
and Systems (BTAS 2012).
[35] O. V. Komogortsev, A. Karpov, et al. CUE: Counterfeit-
resistant Usable Eye-based Authentication via Oculomotor Plant
Characteristics and Complex Eye Movement Patterns. SPIE
Defence Security+Sensing Conference on Biometric
Technology for Human Identification IX, 2012.
[36] O. V. Komogortsev, A. Karpov et al. Biometric Authentication
via Oculomotor Plant Characteristic. IEEE/IARP International
Conference on Biometrics (ICB), 2012.
[37] R. J. Leigh, D. S. Zee. The Neurology of Eye Movements,
Oxford University Press, 2006.
[38] A. Maeder, C. Fookes. A visual attention approach to personal
identification. Eighth Australian and New Zealand Intelligent
Information, 2003.
[39] U. Park, A. Ross et al. Periocular biometrics in the visible
spectrum: a feasibility study. Proceedings of the 3rd IEEE
international conference on Biometrics: Theory, applications
and systems. Washington, DC, USA, IEEE Press: 153-158,
2009.
[40] M.A.Turk, A.P. Pentland. Face recognition using eigenfaces.
IEEE Computer Society Conference on Computer Vision and
Pattern Recognition (CVPR), 1991.
[41] L. Wiskott. Face recognition by elastic bunch graph matching,
1997.
[42] L. Yarbus. Eye Movements and Vision. Moscow, Institute for
Problems of Information Transmission Academy of Sciences of
the USSR, 1967.
[43] R. Cappelli, M. Ferrara, D. Maltoni F. Turroni. Fingerprint Veri-
fication Competition at IJCB2011 Proceedings of International
Joint Conference of Biometrics, 2011.
... The results of the eye movement verification and identification competitions held in 2012 and 2014 also confirm the above observations [27], [28]. The first competition employed data collected using low-level stimuli (so-called "jumping points"), while the second one was based on data related to high-level tasks (viewing pictures of faces). ...
... In this way, it is possible to collect both parameters related to fixations (focusing the gaze on a point) and saccades (when the user moves the eye from point to point) [29]. The "jumping points" approach was adopted by many authors, usually proving to be effective [26], [27], [31]. ...
... Solutions close to the subject of this article exploit information concerning the size of the pupil, the number of fixations and their durations, the parameters of saccadic movements, etc., to determine the user's identity [11]- [14], [19], [27], [32]. For example, in the paper [33], the feature vector contained 60 features (out of the 73 collected). ...
Article
Full-text available
Devices capable of tracking the user’s gaze have become significantly more affordable over the past few years, thus broadening their application, including in-home and office computers and various customer service equipment. Although such devices have comparatively low operating frequencies and limited resolution, they are sufficient to supplement or replace classic input interfaces, such as the keyboard and mouse. The biometric application we researched verifies a user’s identity based on parameters acquired with a low-cost eye tracker. The use of the eye-tracking device in bank booths has many advantages, including the fact that eye trackers are contactless devices, which, especially in the light of the Covid pandemic, has increasing importance, in addition to providing a solution for confirming the liveness of the user. This paper describes an experiment in which 20 features extracted from eye movement data related mainly to saccades and fixations are used as a complementary biometric modality to authenticate clients at banking kiosks. Data were collected from 39 subjects while operating a banking system using engineered biometric kiosk prototypes. Authentication performance employing eye-movement tracking and parameterizing was compared for two classifiers: Support Vector Machine (SVM) and extreme Gradient Boosting (XGBoost). The results showed that the XGBoost-based classifier outperformed the SVM-based one regarding equal error rates (6.76% to 8% vs. 16.21 to 18.78%). Similar differences were obtained for true acceptance rates at different false acceptance rates (0.1 and 0.01), where the SVM-based classifier achieved a maximum of 81.08% and the XGBoost-based achieved 98.65%. Finally, prospects for the broader application of eye movement tracking as a biometric modality are discussed.
... During the years 2012-2015, three eye movement verification and identification competitions were conducted. The first one [6] was held in 2012. Four data set types were employed: two uncalibrated and two calibrated. ...
... Additionally, in some research, the data sets were collected during a one-day experiment [10][11][12], which could facilitate subjects' recognition. In others, the data were recorded with one-week or one-year intervals, yet unsatisfactory outcomes were obtained [6][7][8], or complex classification setup and stimuli were used [9,11]. It was the motivating factor for the exploration of new possibilities in this field. ...
... The feature vectors described earlier were used in the classification process as an input to several classification algorithms. For the research purpose, the methods applied in the previously conducted studies [4,6,11,21] were utilized to check their performance on a different data set. ...
Article
Full-text available
The paper presents studies on biometric identification methods based on the eye movement signal. New signal features were investigated for this purpose. They included its representation in the frequency domain and the largest Lyapunov exponent, which characterizes the dynamics of the eye movement signal seen as a nonlinear time series. These features, along with the velocities and accelerations used in the previously conducted works, were determined for 100-ms eye movement segments. 24 participants took part in the experiment, composed of two sessions. The users’ task was to observe a point appearing on the screen in 29 locations. The eye movement recordings for each point were used to create a feature vector in two variants: one vector for one point and one vector including signal for three consecutive locations. Two approaches for defining the training and test sets were applied. In the first one, 75% of randomly selected vectors were used as the training set, under a condition of equal proportions for each participant in both sets and the disjointness of the training and test sets. Among four classifiers: kNN (k = 5), decision tree, naïve Bayes, and random forest, good classification performance was obtained for decision tree and random forest. The efficiency of the last method reached 100%. The outcomes were much worse in the second scenario when the training and testing sets when defined based on recordings from different sessions; the possible reasons are discussed in the paper.
... Multimodal authentication using SVM is proposed in [47]. The EEG and eye tracking data were used from [48] and [49], respectively. The datasets were combined by considering the similarity of data to produce a fused dataset of hypothetical subjects. ...
Article
Full-text available
An electroencephalogram (EEG) is a measurement that reflects the overall electrical activity in the brain. EEG signals are effective for biometric authentication and robust against malware attacks and any kind of fraud activities due to the uniqueness of the signals. Significant progress in research on EEG-based authentication has been achieved in the last few years, with machine learning being extensively used for classifying EEG signals. However, to the best of our knowledge, there has been no investigation into the overall progress made in such research. In this paper, the literature on the various factors involved in state-of-the-art biometric authentication systems is reviewed. We provide a thorough comparison of different machine learning biometric authentication techniques. The comparison criteria include the research objectives, machine learning algorithms, computational complexity, source of brainwaves, feature extraction methods, number of channels, and so on. Alongside the discussion of existing works, directions for future research are suggested to improve authentication accuracy. This paper provides an in-depth discussion of different advanced biometric authentication techniques, and a vivid picture of state-of-the-art machine learning-based biometric authentication techniques using EEG.
... The model was parameterized by using the obtained data, and the identification of a person is sought in the parameter space. In 2012 the Eye Movement Verification and Identification Competition (EMVIC) at BTAS 2012 was held [37]. The competition used publicly available eye movement databases. ...
Article
Full-text available
Human eyes are in constant motion. Even when we fix our gaze on a certain point, our eyes continue to move. When looking at a point, scientists have distinguished three different fixational eye movements (FEM)—microsaccades, drift and tremor. The main goal of this paper is to investigate one of these FEMs—microsaccades—as a source of information for biometric analysis. The paper argues why microsaccades are preferred for biometric analysis over the other two fixational eye movements. The process of microsaccades’ extraction is described. Thirteen parameters are defined for microsaccade analysis, and their derivation is given. A gradient algorithm was used to solve the biometric problem. An assessment of the weights of the different pairs of parameters in solving the biometric task was made.
... In previous studies, authentication was performed by applying a PIN code using a person's way of looking, and obtaining eye information with eyetracker [44, 45] as well as authentication by writing a graphic password via the way one looks and extracting information of a person's gaze, then finding the pattern of a person's gaze with an eyetracker [46] and eventually obtaining a pattern of a person's gaze with an eyetracker while following a jump point that is a point of light or a colored dot [47,48]. ...
Article
Full-text available
In recent years, security concerns, including document fraud and identity theft, terrorism, and cybercrime have led to the growth and development of biometric technologies, and in general, all activities related to biometric research have gained momentum. This research presents a more intelligent methodology for authenticating and verifying one's identity while reading and typing a text. Threats to e-commerce security are wreaking havoc on online trade via a variety of nefarious methods. Thus, this method can be implemented to furtively identify these attacks and impede them, using a variety of texts to study. We refer to the authentication as a behavioral biometric technology based on how people look when typing in which the visual information was implemented to verify the identity of 35 users while typing via using an eye tracker device. Hence, we have employed the two-layer perceptron neural network to model the learning phase in the object recognition process. The fusion of eye tracker device information and two-layer perceptron neural network coefficients was done at the decision level, and we have achieved an average of 98.97%, 98.76% and 98.86% for sensitivity, specificity and accuracy, respectively. By the combination of biometric technology products as well as modern computer technology, it is easy to perform monitoring, management, systems integration, automated management, and security applications.
... In this subsection, we will present the authentica- A multimodal authentication using SVM is proposed in [47]. The EEG data and the eye tracking data are used from [48], and [49], respectively. ...
Preprint
Full-text available
An EEG is a measurement that reflects the overall electrical activity in the brain. EEG signals are effective for biometric authentication and robust against malware attacks and any kind of fraud activities because of the uniqueness of the signals. A lot of progress in research of EEG based authentication is achieved in the last few years. Especially, machine learning has been extensively used in the last couple of years in classifying EEG signals for authentication. However, to the best of our knowledge, there is no investigation of the overall progress in such kind of research. In this paper, the state-of-the-art literature on various factors of biometric authentication systems has been reviewed. We provide a thorough comparison among different machine learning based biometric authentication techniques. The comparison criteria include objectives of research, machine learning algorithms, computational complexity, source of brainwaves, feature extraction methods, number of channels, and so on. Alongside the discussion of the existing works, we suggest some future research directions for improving authentication accuracy. This paper provides a thorough discussion of different advanced biometric authentication techniques, and can be a good source of grasping a vivid picture of the state-of-the-art machine learning based biometric authentication techniques using EEG.
Article
Eye-tracking technology has gained significant attention in recent years due to its wide range of applications in human-computer interaction, virtual and augmented reality, and wearable health. Traditional RGB camera-based eye-tracking systems often struggle with poor temporal resolution and computational constraints, limiting their effectiveness in capturing rapid eye movements. To address these limitations, we propose EyeTrAES, a novel approach using neuromorphic event cameras for high-fidelity tracking of natural pupillary movement that shows significant kinematic variance. One of EyeTrAES's highlights is the use of a novel adaptive windowing/slicing algorithm that ensures just the right amount of descriptive asynchronous event data accumulation within an event frame, across a wide range of eye movement patterns. EyeTrAES then applies lightweight image processing functions over accumulated event frames from just a single eye to perform pupil segmentation and tracking (as opposed to gaze-based techniques that require simultaneous tracking of both eyes). We show that these two techniques boost pupil tracking fidelity by 6+%, achieving IoU~=92%, while incurring at least 3x lower latency than competing pure event-based eye tracking alternatives [38]. We additionally demonstrate that the microscopic pupillary motion captured by EyeTrAES exhibits distinctive variations across individuals and can thus serve as a biometric fingerprint. For robust user authentication, we train a lightweight per-user Random Forest classifier using a novel feature vector of short-term pupillary kinematics, comprising a sliding window of pupil (location, velocity, acceleration) triples. Experimental studies with two different datasets (capturing eye movement across a range of environmental contexts) demonstrate that the EyeTrAES-based authentication technique can simultaneously achieve high authentication accuracy (~=0.82) and low processing latency (~=12ms), and significantly outperform multiple state-of-the-art competitive baselines.
Article
Eye-tracking technology is being increasingly integrated into mixed reality devices. Although critical applications are being enabled, there are significant possibilities for violating user privacy expectations. We show that there is an appreciable risk of unique user identification even under natural viewing conditions in virtual reality. This identification would allow an app to connect a user's personal ID with their work ID without needing their consent, for example. To mitigate such risks we propose a framework that incorporates gatekeeping via the design of the application programming interface and via software-implemented privacy mechanisms. Our results indicate that these mechanisms can reduce the rate of identification from as much as 85% to as low as 30%. The impact of introducing these mechanisms is less than 1.5° error in gaze position for gaze prediction. Gaze data streams can thus be made private while still allowing for gaze prediction, for example, during foveated rendering. Our approach is the first to support privacy-by-design in the flow of eye-tracking data within mixed reality use cases.
Conference Paper
Full-text available
Growing efforts have been concentrated on the development of alternative biometric recognition strategies, the intended goal to increase the accuracy and counterfeit-resistance of existing systems without increased cost. In this paper, we propose and evaluate a novel biometric approach using three fundamentally different traits captured by the same camera sensor. Considered traits include: 1) the internal, non-visible, anatomical properties of the human eye, represented by Oculomotor Plant Characteristics (OPC); 2) the visual attention strategies employed by the brain, represented by Complex Eye Movement patterns (CEM); and, 3) the unique physical structure of the iris. Our experiments, performed using a low-cost web camera, indicate that the combined ocular traits improve the accuracy of the resulting system. As a result, the combined ocular traits have the potential to enhance the accuracy and counterfeit-resistance of existing and future biometric systems.
Conference Paper
Full-text available
This paper presents an objective evaluation of the effects of stimulus type and eye tracking specifications on the accuracy of biometric verification based on complex eye movement patterns (CEM). Five stimulus types (simple, complex, cognitive, random, textual), six spatial accuracy tiers (0.5°, 1.0°, 1.5°, 2.0°, 2.5°, 3.0°), and six temporal resolution tiers (1000 Hz, 500 Hz, 250 Hz, 120 Hz, 75 Hz, 30 Hz) are evaluated to identify their effects. The results suggest the use of eye tracking equipment capable of 0.5° spatial accuracy and 250 Hz temporal resolution for biometric purposes, though biometric accuracy remains achievable for systems capable of at least 1.0° spatial accuracy and 30 Hz temporal resolution. While not conclusive, the complex and textual pattern stimuli provided the greatest accuracy, with little difference between the remaining stimuli.
Conference Paper
Full-text available
This paper presents an objective evaluation of various eye movement-based biometric features and their ability to accurately and precisely distinguish unique individuals. Eye movements are uniquely counterfeit resistant due to the complex neurological interactions and the extraocular muscle properties involved in their generation. Considered biometric candidates cover a number of basic eye movements and their aggregated scanpath characteristics, including: fixation count, average fixation duration, average saccade amplitudes, average saccade velocities, average saccade peak velocities, the velocity waveform, scanpath length, scanpath area, regions of interest, scanpath inflections, the amplitude-duration relationship, the main sequence relationship, and the pairwise distance between fixations. As well, an information fusion method for combining these metrics into a single identification algorithm is presented. With limited testing this method was able to identify subjects with an equal error rate of 27%. These results indicate that scanpath-based biometric identification holds promise as a behavioral biometric technique.
Article
Full-text available
This paper presents an objective evaluation of the effects of eye tracking specification and stimulus presentation on the biometric viability of complex eye movement patterns (CEM). Six spatial accuracy tiers (0.5°, 1.0°, 1.5°, 2.0°, 2.5°, 3.0°), six temporal resolution tiers (1000 Hz, 500 Hz, 250 Hz, 120 Hz, 75 Hz, 30 Hz), and five stimulus types (simple, complex, cognitive, textual, random) are evaluated to identify acceptable conditions under which to collect eye movement data. The results suggest the use of eye tracking equipment capable of at least 0.5° spatial accuracy and 250 Hz temporal resolution for biometric purposes, while stimulus had little effect on the biometric viability of eye movements.