ArticlePDF Available

A Robust Algorithm for Eye Detection on Gray Intensity Face without Spectacles

Authors:

Abstract and Figures

This paper presents a robust eye detection algorithm for gray intensity images. The idea of our method is to combine the respective advantages of two existing techniques, feature based method and template based method, and to overcome their shortcomings. Firstly, after the location of face region is detected, a feature based method will be used to detect two rough regions of both eyes on the face. Then an accurate detection of iris centers will be continued by applying a template based method in these two rough regions. Results of experiments to the faces without spectacles show that the proposed approach is not only robust but also quite efficient.
Content may be subject to copyright.
A Robust Algorithm for Eye Detection
on Gray Intensity Face without Spectacles
Kun Peng, Liming Chen,
LIRIS, Département MI, Ecole Centrale de Lyon
36 avenue Guy de Collongue, BP 163, 69131 Ecully, France
Su Ruan,
Equipe Image, CReSTIC, Département GE&II, IUT de Troyes
9 rue de Quebec, 10026 Troyes,
France
and
Georgy Kukharev
Faculty of Computer Science and Information Technology, Technical University of Szczecin
Zolnierska 49, 71-210 Szczecin,
Poland
ABSTRACT
This paper presents a robust eye detection algorithm
for gray intensity images. The idea of our method is
to combine the respective advantages of two existing
techniques, feature based method and template based
method, and to overcome their shortcomings. Firstly,
after the location of face region is detected, a feature
based method will be used to detect two rough
regions of both eyes on the face. Then an accurate
detection of iris centers will be continued by
applying a template based method in these two
rough regions. Results of experiments to the faces
without spectacles show that the proposed approach
is not only robust but also quite efficient.
Keywords: Eye detection, Face detection, Face
recognition, Image processing, Pattern recognition
1. INTRODUCTION
As one of the salient features of the human face,
human eyes play an important role in face
recognition and facial expression analysis. In fact,
the eyes can be considered salient and relatively
stable feature on the face in comparison with other
facial features. Therefore, when we detect facial
features, it is advantageous to detect eyes before the
detection of other facial features. The position of
other facial features can be estimated using the eye
position [1]. In addition, the size, the location and
the image-plane rotation of face in the image can be
normalized by only the position of both eyes.
Eye detection is divided into eye position detection
[1, 2] and eye contour detection [3, 15, 16]. (The
second plays an important role in applications such
as video conferencing and vision assisted user
interface [2]). However, most algorithms for eye
contour detection, which use the deformable
template proposed by Yuille et al. [3], require the
detection of eye positions to initialize eye templates.
Thus, eye position detection is important not only
for face recognition and facial expression analysis
but also for eye contour detection. In this paper eye
detection means eye position detection.
Related work
The existing work in eye position detection can be
classified into two categories: active infrared (IR)
based approaches and image-based passive
approaches. Eye detection based on active remote IR
illumination is a simple yet effective approach. But
they all rely on an active IR light source to produce
the dark or bright pupil effects. In other words, these
methods can only be applied to the IR illuminated
eye images. It’s certain that these methods would not
be widely used, because in many real applications
the face images are not IR illuminated.
Thus this paper only focuses on the image-based
passive methods, which can be broadly classified
into three categories: template based methods [3-6],
appearance based methods [7-9] and feature based
methods [10-14]. In the template based methods, a
generic eye model, based on the eye shape, is
designed firstly. Template matching is then used to
search the image for the eyes. While these methods
can detect eyes accurately, they are normally
time-consuming.
The appearance based methods [7-9] detect eyes
JCS&T Vol. 5 No. 3 October 2005
127
perkpeng@hotmail.com
based on their photometric appearance. These
methods usually need to collect a large amount of
training data, representing the eyes of different
subjects, under different face orientations, and under
different illumination conditions. These data are
used to train a classifier such as a neural network or
the support vector machine and detection is achieved
via classification.
Feature based methods explore the characteristics
(such as edge and intensity of iris, the color
distributions of the sclera and the flesh) of the eyes
to identify some distinctive features around the eyes.
Although these methods are usually efficient, they
lack accuracy for the images which have not high
contrast. For example, these techniques may mistake
eyebrows for eyes.
In summary, the image-based eye detection
approaches locate the eyes by exploiting eyes
differences in appearance and shape from the rest of
the face. The special characteristics of the eye such
as dark pupil, white sclera, circular iris, eye corners,
eye shape, etc. are utilized to distinguish the human
eyes from other objects. However, these approaches
lack either efficiency or accuracy, and they are not
ideal for some real applications.
Principle of our method
In this paper, we propose a robust algorithm for eye
detection on gray intensity face, based on combining
the feature base methods and template based
approaches. Combining the respective strengths of
different complementary techniques and overcoming
their shortcomings, the proposed method uses firstly
the feature based method to find out broadly the two
regions of eyes in a face, and the template based
method is then used to locate the center of iris
accurately.
The template based approaches are usually
time-consuming. Its inefficiency comes from two
main factors. Firstly, in order to improve the
accuracy, these methods have to match the whole
face with an eye template pixel by pixel. Secondly,
as we don’t know the size of eyes for an input face
image, we need to repeat the matching process with
eye templates of different sizes. That is to say, we
have to perform the template matching several
times.
So the solution to improve the efficiency of this
algorithm focuses on two points: reducing the area
in the face image for template matching and cutting
down the times of this type of matching. In fact, our
method firstly detects the two rough regions of eyes
in the face using a feature based method. Thus the
following template matching will be performed only
in these two regions which are much smaller than
the whole face. In addition, we can evaluate the size
of eye template according to the size of these two
regions. In other words, profiting from possibility of
evaluating the size of eyes, our algorithm performs
the template matching just once. Altogether, the
proposed method combines the accuracy of template
based methods and the efficiency of feature based
methods.
Outline of the paper
The remainder of the paper is organized as follows:
The details of eyes detection algorithm are described
in Section 2. Section 3 is devoted to the experiments
on the ORL face database. The discussion and
conclusion are given in the last section.
2. PROPOSED METHOD
Architecture
Currently, there are a lot of promising face detection
methods [17-19]. This paper therefore assumes that
(1) a rough face region has been located or the
image consists of only one face, and (2) eyes in face
image can be seen.
The architecture of the proposed approach is shown
in Fig. 1. When a face image is presented to the
system, face detection will be firstly performed to
locate the rough face region. The second step, which
uses an efficient feature based method, is to locate
two rough regions of eyes in the face. In the same
time, on the basis of these two regions, the sizes of
two eyes will be evaluated, and the templates of eyes
will be created according to the estimated sizes.
Finally, the precise locations of the two centers of
iris will be found out after template matching is
applied in these two rough regions.
JCS&T Vol. 5 No. 3 October 2005
128
Fig. 1. Flowchart of proposed method
Detection of eyes’ regions
When the rough face region is detected, as we have
said, an efficient feature based method will be
sequentially applied to locate the rough regions of
both eyes which will be used to the following
affining detection. Fig. 2 shows the processes of the
proposed method:
The first step is to calculate the gradient image (b) of
the rough face region image (a). Then we apply a
horizontal projection to this gradient image. As we
know that the eyes locate in the upper part of the
face and that the pixels near the eyes are more
changeful in value comparing with the other parts of
face, it is obvious that the peak of this horizontal
projection in the upper part can give us the
horizontal position of eyes. According to this
horizontal position and the total height of the face,
we can easily line out a horizontal region (c) in
which the eyes locate.
And then we perform a vertical projection to all
pixels in this horizontal region of image (c), and a
peak of this projection can be found near the vertical
center of face image. In fact, the position of this
vertical peak can be treated as the position of
vertical center of face (d), because the area between
both eyes is most bright in the horizontal region.
In the same time, a vertical projection will be done
to the gradient image (b). There are two peaks of
projection near the right and left boundary of face
image which correspond to right and left limit of the
face (e). In addition, from these two vertical limit
lines, the width of face can be easily estimated.
Combining all results from (c), (d) and (e), we can
get an image segmented like (f). Finally, based on
the result of (f) and the estimated width of face, the
regions of both eyes can be lined out (g).
Creation of eye templates
After the two rough regions of eyes are detected,
template matching will be used to locate the precise
positions of iris centers in these regions. Because the
matching region reduces from the whole face to the
two rough regions, the efficiency of algorithm is
well improved.
Obviously, the first obligatory step for a template
matching is to create a template. It’s easy to find out
eye templates which can be obtained from a real face
image. But the template can’t be directly used for
matching, because the size of the eye in the template
is not same as that in the input image. A simple
solution for this problem is to perform the process of
matching several times, and each time we will use
the template with different size. But this method is
very ineffective.
Concerning our algorithm, in order to improve the
efficiency, the size of the eyes will be estimated
automatically. Thus the process of matching can be
only performed just once. As we have said, the
width of face is already estimated (see Fig. 2), and
the size of eye template can be easily decided
according to the width of face and the geometric
structure of human face. The last image (g) in Fig. 2
shows two eye templates (at two top corners) created
basing on the estimated eye sizes.
JCS&T Vol. 5 No. 3 October 2005
129
Fig. 2. Detection of eyes’ regions
Localization of iris centers
Suppose that we have a template g[i, j] and we wish
to detect its instances in an image f[i, j]. An obvious
thing to do is to place the template at a location in an
image and to detect its presence at that point by
comparing intensity values in the template with the
corresponding values in the image. Since it is rare
that intensity values will match exactly, we require a
measure of dissimilarity between the intensity values
of the template and the corresponding values of the
image. Several measures may be defined:
[, ]
max
ij R
f
g
,
[, ]ij R
f
g
,
2
[, ]
()
ij R
g
,
where R is the region of the template.
The sum of the squared errors is the most popular
measure. In the case of template matching, this
measure can be computed indirectly and
computational cost can be reduced. We can simplify:
222
[, ] [, ] [, ] [, ]
() 2
ij R ij R ij R ij R
f
gfg
∈∈
−= +
∑∑
fg
.
Now if we assume that f and g are fixed, then
gives a measure of mismatch. A reasonable
strategy for obtaining all locations and instances of
the template is to shift the template and use the
match measure at every point in the image. Thus, for
an
m
×
n
template, we compute:
fg
11
[, ] [ ,] [ , ]
mn
kl
M
ij gklfi kj l
==
=
++
∑∑
,
where
k
and
l
are the displacements with respect to
the template in the image. This operation is called
the
cross-correlation
between
f
and
g
.
Our aim will be to find the locations that are local
maxima or are above a certain threshold value.
However, a minor problem in the above computation
was introduced when we assumed that
f
and
g
are
constant. When applying this computation to images,
the template
g
is constant, but the value of
f
will be
varying. The value of
M
will then depend on
f
and
hence will not give a correct indication of the match
at different locations. This problem can be solved by
using normalized crosscorrelation. The match
measure
M
then can be computed using:
11
1/2
2
11
[, ] [ ,] [ , ]
[, ]
[, ]
[,]
mn
fg
kl
fg
mn
kl
Cij gklfikjl
Cij
Mij
fikjl
==
==
=
++
=
++
⎩⎭
∑∑
∑∑
.
Fig. 3 shows an example of using template matching
to locate the iris centers. In left side, the first line
displays two templates of eyes which are created
according to the sizes estimated. The images in the
JCS&T Vol. 5 No. 3 October 2005
130
second line are the rough eyes’ regions in which the
template matching will be applied. The result of
template matching using cross-correlation is shown
by the images of the third line. And the image in left
side shows the final result of eye position detection.
Fig. 3. Template matching
3. EXPERIMENTAL RESULTS
In this section, we present the experimental results
of our algorithms. We use the images in the ORL
database, a well-known free database of faces, to do
our experiments. In this database, there are
completely photographs of 40 persons, of which
each one has 10 various views. The 10 views of the
same person include faces looking to the right, to the
left, downward and upward (see the first line of Fig.
4). All faces in this database are presented by images
in gray-level with the size of 92×112.
We made experiments using all faces without
spectacles which concerns 227 face images and 29
persons. The success rate of proposed algorithm for
all 227 faces is 95.2%. Fig. 4 shows examples of the
images for which the proposed algorithm could
correctly detect the irises of both eyes. In the first
line, there are five face views of the same person.
And the images in the second line are faces of five
different persons.
Fig. 4. Examples for faces without spectacles
The execution time of the proposed algorithm is
about 0.982 second on average by a PC whose CPU
is Pentium IV, 1.8 GHz. It’s remarkable that this
execution time is reckoned for a program written in
Matlab. Obviously, the execution time would be
reduced a lot if the program is transplanted from
Matlab to C or C++.
We also made experiments to the faces with
spectacles, but the results are very unsatisfied. Fig. 5
shows some examples of these experiments. The
images in the first line are five examples for faces
with spectacles for which the proposed algorithm
could correctly detect the iris centers of both eyes.
And the second line gives five faces for which the
proposed algorithm failed the correct detection of
the iris centers. After comparing and analyzing the
detection results, we found out that the false
detection is mainly due to the reflection of the
spectacles. That is to say, the reflection of the
spectacles changes the intensity values of pixels
around eyes, which leads to a false template
matching.
Fig. 5. Examples for faces with spectacles
4. CONCLUSION
A robust eye detection method for gray intensity
faces is reported in this paper. The proposed
algorithm combines two existing techniques: feature
base method and template based method. The
proposed algorithm firstly makes use of feature
based methods to detect two rough regions of eye.
The precise locations of iris centers are then detected
by performing template matching in these two
regions.
The proposed method has been tested by images
from ORL face database. Experimental results show
that this method works well with the faces without
spectacles. For 227 faces without spectacles, the
detection accuracy is 95.2%. In addition, the average
execution time of proposed algorithm shows that this
approach is also quite efficient.
However, the proposed method doesn’t work so well
for the faces with spectacles. Experimental results
show that the false detection is mainly due to the
reflection of spectacle. In view of above limitation,
the future work will be concentrated on improving
JCS&T Vol. 5 No. 3 October 2005
131
the detection accuracy for the faces with spectacles
by reducing the effect of reflection.
REFERENCES
[1]
F. R. Brunelli, T. Poggio, Face recognition:
features versus templates, IEEE Trans. Pattern
Anal. Mach. Intell. 15 (10) (1993) 1042–1052.
[2]
Y. Tian, T. Kanade, J.F. Cohn, Recognizing
upper face action units for facial expression
analysis, Proceedings of the IEEE Conference
on Computer Vision and Pattern Recognition,
Hilton Head Island, South Carolina, 2000, Vol.
1, pp. 294–301.
[3]
A.L. Yuille, P.W. Hallinan, D.S. Cohen, Feature
extraction from faces using deformable
templates, Int. J. Comput. Vision 8 (2) (1992)
99–111.
[4]
X. Xie, R. Sudhakar, H. Zhuang, On improving
eye feature extraction using deformable
templates, Pattern Recognit. 27 (1994)
791–799.
[5]
K.M. Lam, H. Yan, Locating and extracting the
eye in human face images, Pattern Recognit. 29
(1996) 771–779.
[6]
M. Nixon, Eye spacing measurement for facial
recognition, in: Proc. of the Society of
Photo-Optical Instrument Engineers, 1985.
[7]
A. Pentland, B. Moghaddam, T. Starner,
View-based and modular eigenspaces for face
recognition, in: Proc. IEEE Conf. on Computer
Vision and Pattern Recognition (CVPR’94),
Seattle, WA, 1994.
[8]
W. min Huang and R. Mariani, Face detection
and precise eyes location, in: Proc. Int. Conf.
on Pattern Recognition (ICPR’00), 2000.
[9]
J. Huang, H. Wechsler, Eye detection using
optimal wavelet packets and radial basis
functions (rbfs), Int. J. Pattern Recognit. Artif.
Intell. 13 (7) (1999) 1009–1025.
[10]
G.C. Feng, P.C. Yuen, Variance projection
function and its application to eye detection for
human face recognition, Int. J. Comput. Vis. 19
(1998) 899–906.
[11]
G.C. Feng, P.C. Yuen, Multi-cues eye detection
on gray intensity image, Pattern Recognit. 34
(2001) 1033–1046.
[12]
S. Kawato, J. Ohya, Real-time detection of
nodding and head-shaking by directly detecting
and tracking the between-eyes, in: Proc. 4th
IEEE Int. Conf. on Automatic Face and Gesture
Recognition, 2000, pp. 40–45.
[13]
Y. Tian, T. Kanade, J.F. Cohn, Dual-state
parametric eye tracking, in: Proc. 4th IEEE Int.
Conf. On Automatic Face and Gesture
Recognition, 2000.
[14]
S.A. Sirohey, A. Rosenfeld, Eye detection in a
face image using linear and nonlinear filters,
Pattern Recognit. 34 (2001) 1367–1391.
[15]
G. Chow, X. Li, Toward a system for automatic
facial feature detection, Pattern Recognition 26
(1993) 1739–1755.
[16]
K.M. Lam, H. Yan, Locating and extracting the
eye in human face images, Pattern Recognition
29 (5) (1996) 771–779.
[17]
H.A. Rowley, S. Baluja, T. Kanade, Neural
network-based face detection, IEEE Trans.
Pattern Anal. Mach. Intell. 20 (1) (1998)
23–38.
[18]
K.K. Sung, T. Poggio, Example-based learning
for viewbase human face detection, IEEE Trans,
Pattern Anal. Mach. Intell. 20 (1) (1998)
39–51.
[19]
P.C. Yuen, G.C. Feng, J.P. Zhou, A contour
detection method: initialization and contour
model, Pattern Recognition Lett. 20 (2) (1999)
141–148.
JCS&T Vol. 5 No. 3 October 2005
132
Received: Mar. 2005. Accepted: May 2005.
... The authors conducted a series of experiments to validate the performance of the eye-tracking system. In [79], a robust algorithm for eye recognition in gray images is presented. The primary objective of the method is to integrate the respective advantages of two methods, the feature-based method and the template-based method, while overcoming their respective disadvantages. ...
Article
Full-text available
In the contemporary era, dizziness is a prevalent ailment among patients. It can be caused by either vestibular neuritis or a stroke. Given the lack of diagnostic utility of instrumental methods in acute isolated vertigo, the differentiation of vestibular neuritis and stroke is primarily clinical. As a part of the initial differential diagnosis, the physician focuses on the characteristics of nystagmus and the results of the video head impulse test (vHIT). Instruments for accurate vHIT are costly and are often utilized exclusively in healthcare settings. The objective of this paper is to review contemporary methodologies for accurately detecting the position of pupil centers in both eyes of a patient and for precisely extracting their coordinates. Additionally, the paper describes methods for accurately determining the head rotation angle under diverse imaging and lighting conditions. Furthermore, the suitability of these methods for vHIT is being evaluated. We assume the maximum allowable error is 0.005 radians per frame to detect pupils' coordinates or 0.3 degrees per frame while detecting the head position. We found that for such conditions, the most suitable approaches for head posture detection are deep learning (including LSTM networks), search by template matching, linear regression of EMG sensor data, and optical fiber sensor usage. The most relevant approaches for pupil localization for our medical tasks are deep learning, geometric transformations, decision trees, and RASNAC. This study might assist in the identification of a number of approaches that can be employed in the future to construct a high-accuracy system for vHIT based on a smartphone or a home computer, with subsequent signal processing and initial diagnosis.
... These approaches are generally grouped into three categories: image matching, machine learning and image processing based methods. Image matching based methods [1,2] first construct eye templates, and then compare them with subimages in the areas to be detected. The sub-image with the highest matching score is considered as the eye. ...
Conference Paper
In this paper we present a new robust approach for eye localization adopted into lip reading in mobile environments, where the input image is assumed to contain a single face. Firstly, we segment eye candidates regions using intensity information in YCbCr color space. Then coupled regions of eye candidates including assumed eye brows are extracted with the restrictions of morphology and geometry characteristics of eyes in frontal face. GMM is applied next to validate exact eye couple.
... Eye blinking detector is a system that could detects blinking eyes that were resulted from fatigue and drowsiness. Peng, Chen, Ruan, and Kukharev (2005) have proposed an eye detection algorithm which was based on gray intensity images. In the algorithm, point template was stored and later will be matched with existing template. ...
Article
Full-text available
Road accidents are incidents that should be avoided. One of the contributing factors of road accidents occurrence is drowsiness while driving due to fatigue. In this project, fatigue and drowsiness of a person can be detected by looking at the eye area. Drowsy situations are dangerous especially when driving a vehicle over long distances. When a person starts to feel drowsy, the eyes will start to blink more frequently. This characteristic can be used to monitor a driver’s fitness level. In this project the Viola-Jones algorithm using MATLAB cascade object detector was used to detect the presence of blinking eyes. The algorithm is consisted of three phases which are, selection of Haar-like characteristics, integrating a picture into the whole, and Classifiers in a cascading fashion. A number of 29 samples images consisting of the condition fatigue face and non-fatigue face was used. The obtained results obtained were calculated based on the accuracy value of the detected blinking eye. The result shows that both the total Area Under the Curve (AUC) values for faces with fatigue situations and non-fatigue situations are above acceptable values which is 0.5. This indicates both classifications are acceptable and can be used to detect the presence of blinking eyes which represent drowsiness.
... Facial image processing using 3D-CNN to accurately decode the emotional expressions from videos, a robust video feature representation scheme is required which can incorporate both the global facial gestures and local facial gestures which involves the active contour information of different facial parts including lip-contour and eye-contour (with eye-brow). The reason behind choosing the lip and eye-contour as local facial gestures is straight-forward as there exist many literature which validates the fact that the detection of the changes in lip (Halder et al., 2011) and eye-contours (Zheng et al., 2014) play a significant role in emotion recognition (Liu et al., 2010), (Peng et al., 2005). As shown in Fig. 2, a three-stream 3D-CNN is designed based on the model proposed by Huang et al. in 2018(Huang et al., 2018, which can extract both the spatial and temporal image features from the video clip of facial expression changes during playing android games. ...
Article
With the fastest growing popularity of gaming applications on android phone, analyzing emotion changes of steadfast android-gamers have become a study of utmost interest among most of the psychologists. Recently, some android games are producing negative impacts to the gamers; even in the worst cases the effect is becoming life-threatening too. Most of the existing research works are based on psychological view-point of exploring the impact (positive/negative) of playing android games for the child and adult age-group. However, the online recognition of emotional state changes of the android-gamers while playing video games may be relatively unexplored. To fill this void, the present study proposes a novel method of identifying the emotional state changes of android-gamers by decoding their brain signals and facial images simultaneously during playing video games. Besides above, the second novelty of the paper lies in designing a multimodal fusion method between brain signals and facial images for the said application. To address this challenge, the paper proposes a fused type-2 fuzzy deep neural network (FT2FDNN) which integrates the brain signal processing approach by a general type-2 fuzzy reasoning algorithm with the flavor of the image/video processing approach using a deep convolutional neural network. FT2FDNN uses multiple modalities to extract the similar information (here, emotional changes) simultaneously from the type-2 fuzzy and deep neural representations. The proposed fused type-2 fuzzy deep learning paradigm demonstrates promising results in classifying the emotional changes of gamers with high classification accuracy. Thus the proposed work explores a new era for future researchers.
... For 227 faces without spectacles, the detection accuracy is 95.2%. In addition, the average execution time of proposed algorithm shows that this approach is also quite efficient [33]. ...
Article
In recent years, face recognition has emerged as an important problem in the domain of computer vision. FR has become a popular area for research, mainly due to increasing security demands and its potential, commercial and law enforcement applications. Hence, this paper focuses on developing a technique that provides a solution for an efficient face recognition system in different applications. Such as including biometric verification, passport verification, cybercrime, find missing person, etc. As compared with other biometrics systems using fingerprint/ palm-print and iris, face has distinct advantages because of its non-contact process. Here mainly three techniques were proposed and implemented. Feature extraction was achieving using Walsh Hadamard Transform & Local Binary Pattern [WLBP]. LBP is a simple and efficient in the illumination, rotation and grayscale variance. Another approach WHT algorithm is with odd dimension matrix. Automated frontal face recognition was realizing using Principal Component Analysis [PCA]. This paper involves the design of an efficient age invariant facial recognition system [AIFR]. The AIFR system is also expanding into a pose and Illumination invariant, which was implemented and tested on facial images for subjects with different poses and lighting conditions. While using face database of more than 93 individuals consisting of over 718 facial images to test face feature extraction, face recognition. Execution time for face recognition is average of four seconds. Recognition Rate is more than 99%.
Article
Gesture recognition is a technology through which inputs are given to the devices through human gestures. It is a human-machine interaction. By implementing the gesture recognition technology a device can identify and respond to the various gestures of a human. Gestures are movements or positions of the person's hands, face, fingers, to convey any information. These are non-verbal communication. Hand gestures can be mainly used for expressing ones ideas. This hand recognition approach can be mainly divided into Data-Glove Based and Vision Based approaches. Facial Gesture Recognition uses the different face expressions to interact with the machine. Nose tip being the highest point on the face can be considered as an important feature to indicate the head pose.
Conference Paper
This paper explores the identification of human face and eye in still frontal color pictures. First; the appropriate pre-processing stage is done. It requires image refresh, RGB to grayscale conversion, image binarization, noise reduction and removal of small artifacts. The proposed algorithm is then extended to the position the face by detecting the edges the face using the detection of pixel color variations in the binary image. Finally, a normalized cross-relation is applied to determine the precise location of the eyes within the localized region of the face. proposed algorithm is robust enough to achieve approximately 94% accuracy and can locate a face in single face images.
Conference Paper
Full-text available
We present a neural network-based upright frontal face detection system. A retinally connected neural network examines small windows of an image and decides whether each window contains a face. The system arbitrates between multiple networks to improve performance over a single network. We present a straightforward procedure for aligning positive face examples for training. To collect negative examples, we use a bootstrap algorithm, which adds false detections into the training set as training progresses. This eliminates the difficult task of manually selecting nonface training examples, which must be chosen to span the entire space of nonface images. Simple heuristics, such as using the fact that faces rarely overlap in images, can further improve the accuracy. Comparisons with several other state-of-the-art face detection systems are presented, showing that our system has comparable performance in terms of detection and false-positive rates.
Conference Paper
Full-text available
We develop an automatic system to analyze subtle changes in upper face expressions based on both permanent facial features (brows, eyes, mouth) and transient facial features (deepening of facial furrows) in a nearly frontal image sequence. Our system recognizes fine-grained changes in facial expression based on Facial Action Coding System (FACS) action units (AUs). Multi-state facial component models are proposed for tracting and modeling different facial features, including eyes, brews, cheeks, and furrows. Then we convert the results of tracking to detailed parametric descriptions of the facial features. These feature parameters are fed to a neural network which recognizes 7 upper face action units. A recognition rate of 95% is obtained for the test data that include both single action units and AU combinations
Article
Few approaches to automated facial recognition have employed geometric measurement of characteristic features of a human face. Eye spacing measurement has been identified as an important step in achieving this goal. Measurement of spacing has been made by application of the Hough transform technique to detect the instance of a circular shape and of an ellipsoidal shape which approximate the perimeter of the iris and both the perimeter of the sclera and the shape of the region below the eyebrows respectively. Both gradient magnitude and gradient direction were used to handle the noise contaminating the feature space. Results of this application indicate that measurement of the spacing by detection of the iris is the most accurate of these three methods with measurement by detection of the position of the eyebrows the least accurate. However, measurement by detection of the eyebrows' position is the least constrained method. Application of these techniques has led to measurement of a characteristic feature of the human face with sufficient accuracy to merit later inclusion in a full package for automated facial recognition.
Article
This paper presents a novel eye detection method for gray intensity image. The precise eye position can be located if the eye windows are accurately detected. The proposed method uses multi-cues for detecting eye windows from a face image. Three cues from the face image are used. Each cue indicates the positions of the potential eye windows. The "rst cue is the face intensity because the intensity of eye regions is relatively low. The second cue is based on the estimated direction of the line joining the centers of the eyes. The third cue is from the response of convolving the proposed eye variance "lter with the face image. Based on the three cues, a cross-validation process is performed. This process generates a list of possible eye window pairs. For each possible case, variance projection function is used for eye detection and veri"cation. A face database from MIT AI laboratory, which contains 930 face images with di!erent orientations and hairstyles captured from di!erent people, is used to evaluate the proposed system. The detection accuracy is 92.5%.
Article
We present a new approach for eye detection using the variance projection function. The variance projection function is developed and employed to locate landmarks of the human eye which are then used to guide the detection of the eye position and shape. A number of eye images are selected to evaluate the capability of the proposed method and the results are encouraging. q 1998 Elsevier Science B.V. All rights reserved.
Article
An improved method of extracting eye features from facial images using eye templates is described. It retains all advantages of the deformable template method originally proposed by A. L. Yuille, P. W. Hallinan and D. S. Cohen (Int. J. Comput. Vision 99–111 (1989)) and rectifies some of its weaknesses. This is achieved by the following modifications. First, the original eye template and the overall energy function to represent the most salient features of the eye are modified. Secondly, in order to simplify the issue of selecting weights for the energy terms, the value of each energy term is normalized in the range 0–1 and only two different weights are assigned. This weighting schedule does not require expert knowledge therefore it is more user friendly. Thirdly, all parameters of the template are changed simultaneously during the minimization process rather than using a sequential procedure. This scheme prevents some parameters of the eye template from being overly changed, helps the algorithm to converge to the global minimum, and reduces the processing time. The selection of initial parameters of the eye template is based on an eye window obtained in preprocessing. Experimental results are presented to demonstrate the efficacy of the algorithm. A comparison study of various processing schemes is also given.
Article
A model-based methodology is proposed to detect facial features from a front-view ID-type picture. The system is composed of three modules: context (i.e. face location), eye, and mouth. The context module is a low resolution module which defines a face template in terms of intensity valley regions. The valley regions are detected using morphological filtering and 8-connected blob coloring. The objective is to generate a list of hypothesized face locations ranked by face likelihood. The detailed analysis is left for the high resolution eye and mouth modules. The aim for both is to confirm as well as refine the locations and shapes of their respective features of interest. The detection is done via a two-step modelling approach based on the Hough transform and the deformable template technique. The results show that facial features can be located very quickly with Adequate or better fit in over 80% of the images with the proposed system.
Article
In this paper, a new contour detection method based on the snake model is developed and reported. The proposed method consists of two steps. The first step is to locate the initial snake contour and a novel initialization algorithm has been developed. In the second step, an improved snake algorithm is developed to locate the final contour(s). Images with single and multiple objects are selected to evaluate the capability of the proposed method and the results are encouraging.
Article
This paper describes two methods of eye detection in a face image. The face is first detected as a large flesh-colored region, and anthropometric data are then used to estimate the size and separation of the eyes. When a linear filtering method, using filters based on Gabor wavelets, was then applied to detect the eyes in the gray-level image of the face, the detection rate was good (80% on one dataset, 95% on another), but there were many false alarms. A nonlinear filtering method was therefore developed to detect the corners of the eyes in the color image of the face. This method gave a 90% detection rate with no false alarms.
Article
Facial feature extraction is an important step in automated visual interpretation and human face recognition. Among the facial features, the eye plays the most important part in the recognition process. The deformable template can be used in extracting the eye boundaries. However, the weaknesses of the deformable template are that the processing time is lengthy and that its success relies on the initial position of the template. In this paper, the head boundary is first located in a head-and-shoulders image. The approximate positions of the eyes are estimated by means of average anthropometric measures. Corners, the salient features of the eyes, are detected and used to set the initial parameters of the eye templates. The corner detection scheme introduced in this paper can provide accurate information about the corners. Based on the corner positions, we can accurately locate the templates in relation to the eye images and greatly reduce the processing time for the templates. The performance of the deformable template is assessed with and without using the information on corner positions. Experiments show that a saving in execution time of about 40% on average and a better eye boundary representation can be achieved by using the corner information.