b - Chander Kant

advertisement
Biometric Based User Authentication through
Facial Expression
A
Dissertation
Submitted in the partial fulfillment of the requirement
For the Award of Degree
Of
Master of Technology
In Computer Science and Engineering
Supervised By:
Submitted By:
Dr. Chander Kant
Assistant Professor
Annu
M. Tech (FinalYear)
Roll No. 49035
Department of Computer Science & Applications
Kurukshetra University, Kurukshetra
2013
1
DEPARTMENT OF COMPUTER SCIENCE &APPLICATIONS
KURUKSHETRA UNIVERSITY KURUKSHETRA
HARYANA (INDIA)
No: ____________
Dated: __________
DECLARATION
I, Annu, a student of Master of Technology (Computer Science & Engineering), in the
Department
of
Computer
Science
and
Applications,
Kurukshetra
University,
Kurukshetra, Roll no. 49035, for the session 2011-2013, hereby, declare that the
dissertation entitled “Biometric Based User Authentication through Facial
Expression” has been completed after the theory examination of 3rd semester.
The matter embodied in this Dissertation has not been submitted to any other institute or
university for the award of any degree to the best of my knowledge and belief.
Date:
Annu
2
DEPARTMENT OF COMPUTER SCIENCE & APPLICATIONS
KURUKSHETRA UNIVERSITY KURUKSHETRA
HARYANA (INDIA)
Dr. Chander Kant
No: ____________
(Assistant Professor)
Dated: __________
CERTIFICATE
This is to certify that the dissertation entitled “Biometric Based User Authentication
through Facial Expression” submitted by Ms. Annu for the award of degree of Master
of Technology in Computer Science and Engineering Roll No. 49035, Kurukshetra
University, Kurukshetra is a record of bonafide research work carried by her, under my
supervision and guidance. The dissertation, in my opinion worthy for consideration for
the award of Degree of Master of Technology in accordance with the regulations of
Kurukshetra University, Kurukshetra. The matter entitled in this dissertation has not been
submitted to any other institute or university for the award of any degree to the best of
my knowledge and belief.
(Dr. Chander Kant)
Assistant Professor,
Department of Computer Science & Applications
Kurukshetra University, Kurukshetra
3
DEPARTMENT OF COMPUTER SCIENCE & APPLICATIONS
KURUKSHETRA UNIVERSITY KURUKSHETRA
HARYANA (INDIA)
Prof. Suchita Upadhyaya
No: ____________
(Chairperson)
Dated: __________
CERTIFICATE
It is certified that Ms. Annu is a bonafide student of Master of Technology in Computer
Science and Engineering, Roll no. 49035, Kurukshetra University, Kurukshetra. He has
undertaken the dissertation entitled “Biometric Based User Authentication through
Facial Expression” under the supervision of Dr. Chander Kant.
(Prof. Suchita Upadhyaya)
Chairperson Department of Computer Science & Applications
Kurukshetra University, Kurukshetra
4
Acknowledgement
No one can do everything all without someone’s help and guidance. Here, I would like to
take a moment to thank those who have helped and inspired me throughout this journey
and without their guidance and support I might not have been able to complete this
research work.
First of all I would like to express my sincere gratitude to my supervisor
Dr. Chander Kant, Assistant Professor, Department of Computer Science &
Applications, Kurukshetra University, Kurukshetra, for his constant support, inspiration
and guidance in both academic and personal life. I am extremely grateful to him for
being an excellent advisor and a wonderful teacher. His attention to detail and
completeness to every aspect of the work, including research, presentation and technical
writing has helped me tremendously to improve my skills. In the course of my master
study, he has helped me with helpful ideas, guidance, comments and criticisms. His
constant motivation, support and infectious enthusiasm have guided me throughout my
dissertation work. I have been privileged to work under his supervision, and I truly
appreciate this help. His encouraging words have often pushed me to put in my best
possible efforts.
I express my sincere thanks to Dr. Suchita Upadhyaya, Chairperson, Department of
Computer Science & Application, Kurukshetra University, Kurukshetra and other staff
members for their support and encouragement.
Last, but not least, I would like to express my gratitude and appreciation to all of my
family for their support, encouragement, faith, understanding and help. Without them,
this work would not have been possible.
Annu
5
List of Publications
Publications in Conference:
-
“Challenges and Scope of Face Recognition” published in the proceedings of All
India Seminar on Information Security on February 25-26, 2013 by DCRUST,
Murthal, Sonepat, under the aegis of Computer Engineering Division Board, IEI.
Publications in the International Journal:
-
“Liveness Detection in Face recognition using Euclidean Distances” published
in “International Journal for advance Research in Engineering and Technology
ISSN: 2320-6802” Vol. 1, Issue IV, May 2013, pp 1-5.
-
“A Novel Approach for Facial Expression Recognition Using Euclidean
Distances” is accepted in “International Journal of Engineering and Advanced
Technology (IJEAT) ISSN: 2249-8958”, Vol. 2, Issue V, June 2013.
6
Table of Contents
List of Figures……………………………………………………………….........10
List of Tables …………………………………………………………………….11
Abstract………………………………………………………...............................12
Dissertation Outline ……………………………………………….………..……13
Chapter1 Introduction to Biometrics………………………………………………….14
1.1 Biometric Technology……………………………………………………………….14
1.2 Biometric Techniques………………………………………………………………..18
1.2.1 Physical Characteristics Based Techniques………………………………………..18
1.2.2 Behavioral Characteristics Based Techniques……………………………………..23
1.3 Comparison of Various Biometric Technologies……………………………………25
1.4 Advantages of using Biometrics…………………………………………………….27
1.5 Challenges of Biometric System…………………………………………………….27
1.6 Multimodal Biometrics………………………………………………………………28
1.7 Soft Biometrics………………………………………………………………………29
Chapter 2 Face Recognition System…………………………………………………..31
2.1. Motivation…………………………………………………………………………...32
2.2 Why face Recognition is difficult……………………………………………………33
2.3 Automatic Face Detection……………………………………………………………35
2.4 Face Detection Approaches………………………………………………………….35
2.4.1 Face Detection by Template Matching……………………………………….……36
2.4.2 Skin Distribution Model………………………………………………….………..36
2.4.3 Face Detection by Neural Network………………………………………………..36
2.4.4 Face Detection by Eigen Face Method………………………………………….…37
2.5 Image Processing………………………………………………………………….....38
7
2.6 Fusion of Face with Other Biometric Traits……………………………………........40
2. 7 Various Biometric Fusion Techniques………………………………………………41
2.7.1 Fusion at the Image Level………………………………………………………….42
2.7.2 Fusion at the Feature Extraction Level…………………………………………….43
2.7.3 Fusion at the Rank Level………………………………………………………….44
2.7.4 Fusion at the Matching Score Level……………………………………………….45
2.7.5 Fusion at the Decision Level……………………………………………………….47
2.8 Summary……………………………………………………………………………..48
Chapter 3 Literature Survey…………………………………………………………. 49
3.1 Security and Privacy Issues in Biometric Authentication…………………………..68
3.1.2 Biometric System Concerns……………………………………………………….68
3.1.2 Vulnerability of Biometric Authentication System……………………………….69
Chapter 4 Conceptual Models for Face Recognition…………………………………74
4.1 Spoofing in Biometric system……………………………………………………….76
4.1.1 Spoofing in face Recognition System………………………………………….….76
4.2 Face Image Acquisition……………………………………………………………...77
4.3 Proposed face Recognition Algorithm for Image Acquisition………………………78
4.3.1 Architecture of the proposed approach…………………………………………….78
4.3.2 Euclidean Distance Test……………………………………………………………80
4.3.2.1 Euclidean Distance………………………………………………….……………81
4.3.2.2 Algorithm of the proposed approach………………………….…………………81
4.3.2.3 Comparison………………………………………………………………………82
4.4 Summary………………………………………………………………….………….83
Chapter 5 Facial Expression Recognition…………………………………………….84
5.1 Face Detection / Localization…………………………………………………….….84
5.1.1 Color models for skin color classification…………………………………………85
8
5.1.2 Algorithm for detecting a face……………………………………………………..88
5.2 Facial Expression Recognition………………………………………………………90
5.2.1 Facial Action Coding System……………………………………………………...90
5.2.2 Proposed Work…………………………………………………………………….93
5.2.2.1 Algorithm of the proposed approach…………………………………………….93
5.2.2.2 Eigen Face Method………………………………………………………………95
5.2.2.3 Calculating Eigen Faces………………………………………………………….97
5.3 Summary……………………………………………………………………………..98
Chapter 6 Face Recognition Applications…………………………………………….99
6.1 Government use applications……………………………………………….………..99
6.1.1 Law enforcement…………………………………………………………………..99
6.1.2 Security / counterterrorism……………………………………………………….100
6.1.3 Immigration……………………………………………………………………….101
6.1.4 Correctional institutions/prisons …………………………………………………101
6.1.5 Legislature ………………………………………………………………………..101
6.2 Commercial use…………………………………………………………………….102
6.2.1 Banking……………………………………………………………………...……102
6.2.2 Pervasive computing…………………………………………………………...…101
6.2.3 Voter verification…………………………………………………………………103
6.2.4 Healthcare………………………………………………………………...………104
6.2.5 Day care…………………………………………………………………..………104
6.2.6 Missing children/runaways……………………………………………………….104
6.2.7 Residential security……………………………………………………………….104
6.2.8 Internet, E-commerce…………………………………………………..…………104
6.2.9 Benefit payment……………………………………………………..……………105
6.3 Other areas where face recognition used………………………………...…………105
Chapter 7 Conclusion and Future Scope…………………………………………….106
Bibliography…………………………………………...………………………………108
9
List of Figures
Figure 1.1 Enrollment Process in Biometric System
Figure 1.2 Verification and Identification Process
Figure 1.3 Minutiae point of a fingerprint
Figure 1.4 Stricture of hand geometry
Figure 1.5 Structure of iris
Figure 1.6 Image of a face
Figure 1.7 Image of retina
Figure 1.8 Image of voice rhythm
Figure 1.9 Keystroke pattern
Figure 1.10 Example of a signature
Figure 2.1 Steps for face recognition system applications
Figure 2.2 Methods for face detection
Figure 2.3 Biometric fusion systems
Figure 2.4 Authentication Process Flow
Figure 2.5 Fusion at the Image level
Figure 2.6 Fusion at the Feature Extraction level
Figure 2.7 Fusion at the Matching Score level
Figure 2.8 Fusions at the Decision Level
Figure 3.1 Block diagram of a typical biometric authentication system
Figure 4.1 Techniques used in face recognition
Figure 4.2 Block Diagram of image acquisition using CCD camera
Figure 4.3 Architecture of the proposed approach
Figure 5.1 RGB color space
Figure 5.2 The RGB color space within the YCbCr color space
Figure 5.3 HSI color space
Figure 5.4 Steps of a face detection algorithm
Figure 5.5 Facial action coding system
Figure 5.6 Flow Chart of the proposed approach
10
List of Tables
Table 1.1 Performance Comparison of various Biometric Traits
Table 5.1 FACS Action Units
Table 5.2 Miscellaneous Actions Units
11
Abstract
Biometric Based
Expression
User
Authentication
through
Facial
The dissertation addresses certain special concepts of face recognition system. Compared
with other biometric technologies, such as iris, speech, fingerprint recognition, face
recognition can be easily considered as the most reliable biometric trait in all. Face
recognition is a particular type of biometric system that can be used to reliably identify a
person by analyzing the facial features of a person’s image. A biometric system verifies
user identity through comparison of a certain behavioral or psychological characteristic
possessed by the user. Face recognition is used worldwide for the security purposes.
User authentication systems that are based on knowledge such as password or physical
tokens such as ID card are not able to meet strict security performance requirements of a
number of modern computer applications. These applications generally use computer
networks (e.g., Internet), affect a large portion of population, and control financially
valuable tasks (e.g., e-commerce). Biometrics-based authentication systems are good
alternatives to the traditional methods. These systems are more reliable as biometric data
cannot be lost, forgotten, or guessed and more user-friendly because we don’t need to
remember or carry anything. The main functional component of the existing face
recognition system consists of image capturing, face detection, and then features
extraction and finally matching. If it matches correctly then accept the user otherwise
rejected. This dissertation gives a general presentation of the face recognition technology
and also the facial expression recognition with its advantages as well as challenges.
12
Dissertation Outline
Chapter 1
Chapter 1 gives the introduction part of the biometric authentication system, parameter
characteristics, advantages of biometric system, and challenges of biometric system,
multimodal biometric system, soft biometrics and image processing.
Chapter 2
Chapter 2 gives the introduction of the face recognition and detection, face detection
approaches, difficulties in face recognition and fusion of face with other biometric traits.
Chapter 3
Chapter 3 gives the literature survey, security and privacy issues related to the use of
biometrics. Solutions and previous work in this field are presents as well.
Chapter 4
Chapter 4 will discussed about existing face recognition techniques and the proposed
work which includes a method to check the liveness of a person.
Chapter 5
Chapter 5 will discuss about the facial expression recognition, Facial action coding
system, and the proposed work to recognize a person’s expression and then
authenticating the user.
Chapter 6
Chapter 6 presents current applications of face recognition, face recognition survey and
some areas where face recognition is used.
Chapter 7
Chapter 7 gives the conclusion and future scope of face recognition technology.
13
Chapter 1
Introduction to Biometrics
Now-a-days, biometric recognition is a common and reliable way to authenticate any
human being based on his physiological or behavioral biometrics [1]. A physiological
biometric traits is stable in their biometric like fingerprint, iris pattern, facial feature,
hand geometry, gait pattern etc. whereas behavioral biometric traits is related to the
behavior of person such as signature, speech pattern, keystroke pattern. Facial recognition
system is a computer application for automatically identifying or verifying a person from
a digital image or a video frame from a video source.
Face recognition is not a new idea but it has received substantial attention over the last
three decades due to its value both in understanding how FR process works in humans as
well as in addressing many challenging real-world applications, including de-duplication
of identity documents (e.g. passport, driver license), access control and video
surveillance. While face recognition in controlled conditions (frontal face of cooperative
users and controlled indoor illumination) has already achieved impressive performance
over large-scale galleries [2], there still exist many challenges for face recognition in
uncontrolled environments, such as partial occlusions, large pose variations, and extreme
ambient illumination. Local facial features have played an important role in forensic
applications for matching face images. These features include any salient skin region that
appears on the face. Scars, moles, and freckles are representative examples of the local
facial features [3].
1.1 Biometric Technology
Biometric refers to the automatic identification of a person based on his or her physical or
behavioral characteristics. This identification method is preferred over traditional
methods involving passwords and PINs (personal identification numbers) for several
14
reasons, including the person to be identified is required to be physically present at the
point of identification and identification based on biometric techniques avoids the need to
remember a password or carry a token. Now-a-days, with the increased use of computers
as the means of transportation of information technology, restrict access to sensitive or
personal data is essential. Biometric authentication has seen considerable progress in
reliability and accuracy, with some of the traits offering good performance.
The terms “Biometrics” and “Biometry” have been used since the early 20th century to
refer to the field of improvement of arithmetical and mathematical methods applicable to
data analysis problems in the biological sciences [4]. The need for biometrics can be
found in central, state and local governments, in the military and in commercial
applications. Now-a-days, worldwide network security infrastructures, government IDs,
secure electronic banking, investing and other financial communication, retail sales, law
enforcement and health and social services are already benefiting from these
technologies. Biometric based authentication applications include workstation, network
and domain access, single sign in, application login, data protection from illegal access,
remote access to resources, transaction security and web security. Trust in this type of
electronic communication is the necessary to the successful growth of the global financial
system. Utilizing biometrics for personal authentication is appropriate, well-situated and
considerably more accurate than current methods such as utilization of passwords or
PINs.
A biometric system may operate either in an ‘identification’ system or a ‘verification’
(authentication) system. Before the system can be put into verification or identification
mode, a system database consisting of biometric templates must be created through the
process of enrollment.
Enrollment- is the process where a user’s initial biometric sample(s) are collected,
assessed, processed, and stored for ongoing use in a biometric system as shown in Figure
1.1. Basically, user enrollment is a process that is responsible for registering individuals
in the biometric system storage. During the enrollment process, the biometric
characteristics of a person are first captured by a biometric scanner to produce a sample.
Some systems collect multiple samples of a user and then either select the best image or
15
fuse multiple images or create a composite template. If users are experiencing problems
with a biometric system then they have to re-enroll to gather higher quality data.
Figure1.1 Enrollment Process in Biometrics System
Biometric system provides two main functionalities viz. verification and identification.
Figure 1.2 shows the flow of information in verification and identification systems.
Identification- One-to-Many Correspondence: Biometrics can be used to determine a
person’s identity even without his knowledge or permission. The user’s input is
compared with the templates of all the persons enrolled in the database and the identity of
the person whose template has the highest degree of similarity with the user’s input is
output by the biometric system. Typically, if the highest similarity between the input and
all the templates is less than a fixed minimum threshold, the system rejects the input,
which implies that the user presenting the input is not one among the enrolled users. For
example, scanning a crowd with a camera and using biometric recognition technology,
one can determine matches against a known database. Identification is the initial stage to
identify the user through his biometric trait. The data of user stored for ongoing use in a
biometric system permanently.
Verification- One-to-One correspondence: Biometrics can also be used to verify a
person’s identity and the system verifies whether the claim is genuine. If the user’s input
and the template of the claimed identity have a high degree of similarity, then the claim is
accepted as “genuine”. Otherwise, the claim is rejected and the user is considered as
“fraud”. For example, one can grant physical access to a secure area in a building by
16
using finger scans or can grant access to a bank account at an ATM by using retinal scan.
Figure 1.2 shows the flow of information in verification and identification systems.
Figure1.2 Verification and Identification Process
Biometric is necessary due to these characteristics:

Links the event to a particular user ( a password or symbol may be used by
somebody other than the approved user)

Is convenient (nothing to carry or memorize)

Accurate ( it provides for positive authentication)

Fast and scalable

Easy to use and easily understandable

Is becoming socially acceptable and

Cost effective
17
1.2 Biometric Techniques
There are many different techniques available to identify/verify a person based on their
biometric characteristics as suggested by U.K. Biometric Working Group [UKBWG,
2003]. These techniques can be divided into physical characteristics and behavioral
characteristics based techniques.
1.2.1 Physical characteristics based Techniques
Biometrics techniques based on physical characteristics of human being such as finger
print, hand geometry; palm print etc are called physical characteristics based techniques.
Following are examples of biometric techniques based on physical characteristics.
Fingerprint Recognition: Among all the biometric techniques, fingerprints based
identification is the oldest method which has been successfully used in several
applications. The fingerprint itself consists of patterns originate on the tip of the finger,
thus making it a physical biometric. Fingerprints are known to be unique and absolute for
each person and the basic characteristics of fingerprints do not change with time. The
distinctiveness of a fingerprint can be determined by the patterns of ridges and minutiae
points on the surface of the finger. These unique patterns of lines can either be in a loop,
whorl or arch pattern. The most common method involves recording and comparing the
fingerprint’s “minutiae points”. Minutiae points can be considered the uniqueness of an
individual’s fingerprint. The major Minutia points in fingerprint are: ridge ending,
bifurcation, and short ridge or dot as shown in Figure 1.3.
(a) Ridges Ending
(b) Ridges Bifurcation
Figure 1.3 Minutiae points in fingerprint
18
(c) Dot
The ridge ending is the point at which a ridge terminates (see Figure 1.3 a). Bifurcations
are points at which a single ridge splits into two ridges (see Figure 1.3 b). Short ridges or
dots are ridges which are significantly shorter than the average ridge length on the
fingerprint (see Figure 1.3 c). Some examples of the use of fingerprint devices in general
areas are:

Fight the abuse of social services like social security.

Permitting logins based on fingerprints.

Fight against criminal immigration.

Attendance management system in industry, colleges or companies.
Hand Geometry: This biometric approach uses the geometric form of the hand for
confirming the user’s identity. Specific features of a hand must be combined to assure
dynamic verification, since human hands are not unique. Individual hand features are not
descriptive enough for identification. Characteristics such as finger curves, thickness and
length, the height and width of the back of the hand, the distances between joints and the
overall bone structure are usually extracted. Those characteristics are pretty much
determined and mostly do not change in a range of years. The basic structure of hand
geometry is shown in figure 1.4.
Figure 1.4 Structure of hand geometry
As the hand geometry readers are big, rugged, quick and handy, they are well suited to
use in warehouses, manufacturing facilities and other industrial locations that have the
space to comfortably house them. Hand geometry readers are good for time-andattendance applications (replacing punch clocks) where their simplicity and rapid cycle
times are big assets and their lackluster accuracy rates are not major liabilities. The list
19
showing below following examples in real world areas where hand scan identification is
or was used:

Users at the Olympic Games 1996 were identified with hand scans.

In a lot of cases access to military plants is granted upon successful hand scan
identification.

Airport personnel at the San Francisco Airport are identified by hand scans.
Iris Recognition: Biometrics is the science of measuring human characteristics that are
stable and unique among users. Iris recognition is the process of verify a human being by
the pattern of iris. The iris is the area of the eye where the pigmented or colored circle,
usually black, brown or blue rings the dark pupil of the eye. As compared to other
biometric traits the iris is more secured and protected. Iris recognition is an approach to
biometric based verification and identification of people [5]. The future of the iris
recognition system is better in fields that demand rapid identification of the users in a
dynamic environment [6]. Iris patterns are extremely complex. The basic structure of iris
is shown below in figure 1.5.
Figure 1.5 Structure of iris
In this technique, the user places him so that he can see his own eye's reflection in the
device. The user may be able to do this from up to 2 feet away or may need to be as close
as a couple of inches depending on the device. Verification time is generally less than 5
seconds, though the user will only need to look into the device for a couple of moments.
To prevent a fake eye from being used to fool the system, these devices may vary the
light shone into the eye and watch for pupil dilation.
20
Face Recognition: Face recognition (FR) is the problem of verifying or identifying a
face from its image. User face detection plays an important role in applications such as
video observation, human computer interfaces, face recognition and face image databases
[7]. Local facial features have played an important role in forensic applications for
matching face images. The use of local features has become more popular due to the
development of higher resolution sensors, an increase in face image database size, and
improvements in image processing and computer vision algorithms. Local features
provide a unique capability to investigate, annotate, and exploit face images in forensic
applications by improving both the accuracy and the speed of face-recognition systems.
This information is also necessary for forensic experts to give testimony in courts of law
where they are expected to conclusively identify suspects [8]. The image of face is shown
in figure 1.6.
Figure 1.6 Face Recognition
To enable this biometric technology it requires having at least one video camera, PC
camera or a single-image camera. On the other hand, this biometric technique still has to
deal with a lot of problems. Finding a face in a picture where the location, the direction,
the background and the size of a face is variable is a very hard task and many algorithms
have been worked on to solve this problem.
Retina Recognition: Along with iris recognition technology, retina scan is possibly the
most accurate and reliable biometric technology. It is also among the most difficult to use
and requires well trained and is supposed as being quite to highly invasive. The users
have to be cooperative and patient to achieve an accurate performance.
21
Figure 1.7 Retina Recognition
Basically the retina is a thin curve on the back of the eye which senses light and transmits
impulses through the optic nerve to the brain as shown in figure 1.7. Retinal scanning
analyses the layer of blood vessels at the back of the eye. Blood vessels used for
biometric identification are located along the neural retina which is the outermost of the
retina’s four cell layers. Research has proven that the patterns of blood vessels on the
back of the human eye were unique from person to person. It even been proven that these
patterns, even between twins, were indeed unique. This pattern also does not change over
the course of a lifetime. The retinal scanner requires the user to place their eye into some
sort of device and then asks the user to look at a particular mark so that the retina can be
clearly imaged. Scanning involves using a low-intensity light source and an optical
coupler and can read the patterns at a great level of accuracy. This process takes about 10
to 15 seconds in total. There is no known way to replicate a retina, and a retina from a
dead person would deteriorate too fast to be useful, so no extra precautions have been
taken with retinal scans to be sure the user is a living human being.
Vein pattern recognition: Vascular patterns are best described as a picture of the veins
in a person's hand or face. The thickness and location of these veins are believed to be
unique enough to an individual to be used to verify a person's identity. The most common
form of vascular pattern readers is hand-based, requiring the user to place their hand on a
curved reader that takes an infrared scan. This scan creates a picture that can then be
compared to a database to verify the user's stated identity.
22
1.2.2 Behavioral Characteristics based Techniques:
Those Biometrics techniques which are based on the behavior of human being such as
voice, signature, gait, keystroke etc. are called behavioral characteristics based
techniques. Following are examples of biometric techniques based on behavioral
characteristics.
Voice Recognition: Mostly, the voice biometric solutions can be used through a typical
telephone or microphone equipment to the computer. In order to identify or authenticate
users, most voice biometric solution creates a voice print of the user, a template of the
person’s unique voice characteristics created when the user enrolls with the system.
During enrollment the user has to select a passphrase or repeat a sequence of numbers or
words. The passphrase should be in the length of 1 to 5 seconds. The problem with short
passphrases is that they have not enough data for identification. Longer passphrases have
too much information and takes too much time. The user has to repeat the passphrase or
the sequence of numbers several times. This makes the enrollment process long lasting
than with other biometric technologies. All successive attempts to access the system
require the user to speak, so that their live voice sample may be compared against the
pre-recorded template. Voice rhythm is shown below in figure 1.8.
Figure 1.8 Voice rhythm pattern
A voice biometric sample is a numerical model of the sound, pattern and rhythm of a
user’s voice. The main problem occurring in the voice is that the user’s voice changes
over time along with the growth of a user or when someone has got a cold or another
disease. Background noise can also be a disturbing factor which does not gives the
accurate result.
23
Keystroke Dynamics: Keystroke dynamics uses the manner and rhythm in which a user
types characters/password or phrase on the keyboard or keypad. The system then records
the timing of the typing and compares the password itself and the timing to its database.
Here, verification takes less than 5 seconds. Keystroke dynamics is the process of
analyzing the way a user types at a terminal by monitoring the keyboard inputs thousands
of times per second in an attempt to identify users based on normal typing rhythm
patterns. The keystroke pattern shown as below in figure 1.9.
Figure 1.9 Keystroke pattern
Signature Recognition: Signature verification is the process that is used to recognize a
user’s handwritten signature. Dynamic signature verification uses behavioral biometrics
of a handwritten signature to validate the identity of a person. This can be achieved by
analyzing the shape, speed, stroke, and pen pressure and timing information during the
act of signing. The example of signature is shown below in figure 1.10.
Figure 1.10 Signature Recognition
On the other hand, there is the simple signature comparison which only takes into
account what the signature looks like. So with dynamic signature verification, it is not the
shape or look of the signature that is meaningful, there are the changes in the speed,
pressure and timing that occur during the act of signing, thus making it virtually
impossible to duplicate those features. The main difficulty with this technology is to
distinguish between the reliable part of a signature, these are the characteristics of the
24
static image, and the behavioral parts of a signature, which vary with each signing
process. Comparing many signatures made by one user reveals the fact that a user’s
signature is never completely the same and can vary considerably over a user lifetime.
Allowing these variations in the system, while providing the best protection against fake
is a big problem faced by this biometric technology. The financial industry sometimes
uses signature verification for money transactions.
1.3 Comparison of Various Biometric Technologies:
Performance of a biometric measure is usually referred to in terms of the false accept rate
(FAR), the false non match or reject rate (FRR), and the failure to enroll rate (FTE or
FER). The FAR measures the percent of invalid users who are incorrectly accepted as
genuine users, while the FRR measures the percent of valid users who are rejected as
impostors. A number of biometric characteristics may be captured in the first phase of
processing. However, automated capturing and automated comparison with previously
stored data requires that the biometric characteristics satisfy the following characteristics:
1. Universal: Every person must possess the characteristic or attribute. The attribute must
be one that is universal and rarely lost to accident or disease.
2. Uniqueness/singularity: Each expression of the attribute must be unique to the
individual. The characteristics should have sufficient unique properties to distinguish one
person from any other. Height, weight, hair and eye color are all attributes that are unique
assuming a particularly precise measure, but do not offer enough points of differentiation
to be useful for more than categorizing.
3. Permanence/Invariance of Properties: They should be constant over a long period of
time. The attribute should not be subject to significant differences based on age either
episodic or chronic disease.
4. Collectability/Measurability: The properties should be suitable for capturing without
waiting time and must be easy to gather the attribute data inactively.
25
5. Performance: it is the measurement of accuracy, speed, and robustness of technology
used.
6. Acceptability: The capturing should be possible in a way acceptable to a large
percentage of the population. Excluded are particularly invasive technologies, i.e.
technologies which require a part of the human body to be taken or which apparently
damage the human body.
7. Circumvention: Ease of use of a substitute.
There are also some other parameters which are very important during the analysis of a
biometric trait. These are:
Reducibility: The captured data should be capable of being reduced to a file which is
easy to handle.
Reliability and Tamper-resistance: The attribute should be impractical to mask or
manipulate. The process should ensure high reliability and reproducibility.
Privacy: The process should not violate the privacy of the person.
Comparable: Should be able to reduce the attribute to a state that makes it digitally
comparable to others. The less probabilistic the matching involved, the more authoritative
the identification.
Inimitable: The attribute must be irreproducible by other means. The less reproducible
the attribute, the more likely it will be reliable.
Table 1.1 below shows a comparison of various biometric systems in terms of above
mentioned parameters. A. K. Jain ranks each biometric based on the categories as being
low, medium or high. A low ranking indicates poor performance in the evaluation
criterion whereas a high ranking indicates a very good performance.
26
Table 1.1 Performance Comparison of various Biometric Traits
1.4 Advantages of Using Biometrics:

Easier fraud detection

Better than password/PIN or smart cards

No need to memorize passwords

Requires physical presence of the person to be identified

Unique Physical or behavioral characteristic

Cannot be borrowed, stolen or forgotten

Cannot leave it at home
1.5
Challenges with Biometric System
The human face is not a unique rigid object. There are billions of different faces and each
of them can assume a variety of deformations. These variations can be classified as
follows:
a) Inter-personal variations due to
1. Race: It signifies the complexion of the person according to which a person can be
distinguished from other.
27
2. Identity: It tells about the name, address, phone no. of the person which distinguishes
him from the other person.
3. Genetics: Genetic code of each person is different. Because of the genes every
person has different face and gender.
b) Intra-personal variations due to
1. Deformations: These are the result of injury or accident on the face.
2. Expressions: This shows the mood of the person. From expressions it is easy to
determine whether the person is happy or sad.
3. Aging: With age wrinkles appears on the face. The wrinkles change the formation of
the face to a great extent.
4. Facial hairs: Man have moustaches and beard which change the look of the face when
shaven.
5. Cosmetics: Cosmetic surgery has become one of the widely used techniques to
enhance your facial features.
Each of these variations can give rise to a new field of research. Among each of these
facial expressions is a most widely studied technique from past many years.
1.6 Multimodal Biometrics
To enhance the security of the system Biometric follow a new techniques called as
Multimodal biometric. Multimodal Biometric means Biometric authentication by
scanning more number of characteristics.
A multimodal biometric system uses multiple modalities to capture different types of
biometrics. This allows the integration of two or more types of biometric recognition and
verification systems in order to meet stringent performance requirements or combine
several weak biometric measurements to engineer a strong optionally continuous
biometric system. The multimodal system could be, for instance, a combination of
fingerprint verification, face recognition, voice verification and keystroke dynamics or
any other combination of biometrics. This enhanced structure takes advantage of the
28
proficiency of each individual biometric and can be used to overcome some of the
limitations of a single biometric. A multimodal system can combine any number of
independent biometrics and overcome some of the limitations presented by using just one
biometric as the verification tool. This is important if the quality scores from individual
systems are not very high. A multimodal system, which combines the conclusions made
by a number of unrelated biometrics indicators, can overcome many of these limitations.
Also it is more difficult to forge multiple biometric characteristics than to forge a single
biometric characteristic.
1.7 Soft Biometrics
Soft biometric traits are the characteristics of human being that provide some information
about the user, but lack of the distinctiveness and permanence to sufficiently differentiate
any two users. Soft biometric traits embedded in a face (e.g., gender and facial marks) are
ancillary information and are not fully distinctive by themselves in the facial recognition
tasks. However, this information can be explicitly combined with facial matching score to
improve the overall facial recognition accuracy and efficiency and helps the forensic
investigators. The characteristics of human being like gender, height, weight and age can
also be used for the identification purpose. Although these characteristics are not unique
and reliable, yet they provide some useful information about the user. These
characteristics are known as soft biometric traits and these can be integrated with the
primary biometric identifiers like fingerprint, face, iris, signature for the identification
purpose. Unimodal biometric systems make just use of a single biometric trait for
individual recognition. It is difficult to achieve very high recognition rates using
unimodal systems due to problems like noisy sensor data and non-universality or lack of
distinctiveness of the selected biometric trait. Multimodal biometric systems address
some of these problems by combining evidence obtained from multiple sources. A
multimodal biometric system that utilizes a number of different biometric identifiers like
face, fingerprint, hand-geometry, and iris can be more robust to noise and minimize the
problem of non-universality and lack of distinctiveness. However, the problem with
multimodal system is that it will require a longer verification time thereby causing some
inconvenience to the users. A possible solution to the problem of designing a trustworthy
29
and user-friendly biometric system is to use additional information about the user like
height, weight, age, gender, ethnicity, and eye color to improve the performance of the
primary biometric system.
The rest of the dissertation is divided into the six Chapters.
Chapter 2 discusses the face recognition and detection system, face detection approaches,
difficulties of face recognition system, fusion of face with other biometric traits.
Chapter 3 discusses the literature survey and analyses the existing work in the field of
face recognition and detection in general.
Chapter 4 will discuss about the proposed works, existing face recognition and detection
techniques and to check the liveness of a user.
Chapter 5 will discuss about facial expression recognition system, facial action coding
system and the proposed method for recognizing the facial expression and the Eigen face
method.
Chapter 6 presents the current applications of face recognition, face recognition survey
and some areas where face recognition is used.
At last, Chapter 7 presents the conclusion of dissertation and gives the scope of future
work.
30
Chapter 2
Face Recognition System
Face recognition systems are part of facial image processing applications and their
significance as a research area is increasing recently. These systems use biometric
information of the humans and are applicable easily instead of fingerprint, iris, signature
etc., because these types of biometrics are not much suitable for non-collaborative
people. Face detection is essential front end for a face recognition system. Face detection
locates and segments face regions from cluttered images, either obtained from video or
still image. Detection application is used to find position of the faces in a given image
[9].Numerous techniques have been developed to detect faces in a single image[10][11].It
has numerous applications in areas like surveillance and security control systems, content
based image retrieval, video conferencing and intelligent human computer interfaces.
Most of the current face recognition systems presume that faces are readily available for
processing. However, we do not typically get images with just faces. We need a system
that will segment faces in cluttered images. With a portable system, we can sometimes
ask the user to pose for the face identification task.
Figure 2.1 Steps for face recognition system applications
The first step for face recognition system is to acquire an image from a camera. Second
step is face detection from the acquired image. As a third step, face recognition that takes
the face images from output of detection part. Final step is person identity as a result of
recognition part. An illustration of the steps for the face recognition system is shown in
figure 2.1.
31
In addition to creating a more cooperative target, we can interact with the system in order
to improve and monitor its detection. With a portable system, detection seems easier. The
task of face detection is seemingly trivial for the human brain, yet it still remains a
challenging and difficult problem to enable a computer /mobile phone/PDA to do face
detection. This is because the human face changes with respect to internal factors like
facial expression, beard, mustache glasses etc and it is also affected by external factors
like scale, lightning conditions, and contrast between face, background and orientation of
face.
Face detection remains an open problem. Many researchers have proposed different
methods addressing the problem of face detection. In a recent survey face detection
technique is classified in to feature based and image based. The feature based techniques
use edge information, skin color, motion and symmetry measures, feature analysis,
snakes, deformable templates and point distribution. Image based techniques include
neural networks, linear subspace method like Eigen faces, fisher faces etc. The problem
of face detection in still images is more challenging and difficult when compared to the
problem of face detection in video since emotion information can lead to probable
regions where face could be located.
2.1 Motivation
Face detection plays an important role in today’s world. They have many real world
applications like human/computer interface, surveillance, authentication and video
indexing. However research in this field is still young. Face recognition depends heavily
on the particular choice of features used by the classifier One usually starts with a given
set of features and then attempts to derive an optimal subset (under some criteria) of
features leading to high classification performance with the expectation that similar
performance can also be displayed on future trials using novel (unseen) test data.
Interactive Face Recognition is divided in to several phases. It includes:

Creating drivers for the handheld device that link with the application with the
captured image.
32

A face detection program is run inside the handheld device which detects the face
from the image.

The obtained face is transmitted through wireless network.

The server performs the face recognition and is transmitted back.
2.2 Why Face Recognition is Difficult?
The greatest difficulty of face recognition, compared to other biometrics, stems from the
immense variability of the human face. The facial appearance depends heavily on
environmental factors .for example, the lighting conditions, background scene and head
pose. It also depends on facial hair, the use of cosmetics, jewellery and piercing.

Pose: Variation due to the relative camera-face pose (frontal, 45 degree, profile,
upside down), and some facial features such as Facial Expression or the nose may
become partially or wholly occluded.

Presence or absence of structural components: Facial features such as beards,
mustaches, and glasses may or may not be present, and there is a great deal of
variability amongst these components including shape, color, and size. Local
facial mark features such scars, moles, and freckles play an important role for
matching face images in forensic applications [12]. Local facial mark features
provide a unique capability to investigate, annotate, and exploit face images in
forensic applications by improving both the accuracy and the matching speed of
face-recognition systems. This information is also necessary for forensic experts
to give testimony in courts of law where they are expected to conclusively
identify suspects.

Facial expression: The appearance of faces is directly affected by a person's
facial expression.

Occlusion: Faces may be partially occluded by other objects. In an image with a
group of people, some faces may partially occlude other faces.
33

Image orientation: Face images directly vary for different rotations about the
camera's optical axis.

Imaging conditions: When the image is formed, factors such as lighting (spectra,
source distribution and intensity) and camera characteristics (sensor response,
lenses) affect the appearance of a face.

Facial Aging: Many face recognition scenarios exhibit a significant age
difference between the probe and gallery images of a subject. Facial aging is a
complex process that affects both the shape and texture (e.g., skin tone or
wrinkles) of a face. This aging process also appears in different manifestations in
different age groups. In addition to facial aging, there are other factors that
influence facial appearance as well (e.g. pose, lighting, expression, occlusion)
which makes it difficult to study the aging pattern using these two public domain
longitudinal face databases.

Forensic Sketch Recognition: When no photograph of a suspect is available, a
forensic sketch is often generated. Forensic sketches are an artist rendition of a
person’s facial appearance that is derived from an eye witness description.
Forensic sketches have a long history of use, where traditionally they have been
disseminated to media outlets and law enforcement agencies in the hopes that
someone will recognize the person in the sketch. Forensic sketches can be
misleading due to errors in witness memory recall that cause inaccuracies in the
sketch drawn by a forensic artist. Because a significant amount of time is needed
to generate a single forensic sketch, they generally represent culprits who
committed the most heinous crimes (e.g. murder and sexual assault). Thus, the
ability to match forensic sketches to mug shot databases is of great importance.

Face Recognition in Video: Face recognition in video has gained importance due
to the widespread deployment of surveillance cameras. The ability to
automatically recognize faces in video streams will facilitate a method of human
identification using the existing networks of surveillance cameras. However, face
images in video often contain non-frontal poses of the face and may undergo
severe lighting changes.
34
2.3 Automatic Face Detection
a) Facial measures have historically been of interest to a small group of scientists and
clinicians. The main reason is that pain assessment by facial expression is burdensome.
The coding still consumes many multiples of the real times involved in the behavior.
b) The other limitation include facial coding technique: Anyone who has performed facial
measurement realizes that there are elements of subjectivity in deciding when an action
has occurred or when it meets the minimum requirements for coding. In addition, the
human observer has inherent limitations in his or her capacity to make precise
discriminations at the margins. Alternatives exist or are in development.
c) Some faces are often falsely read as expressing some emotion, even when they are
neutral, because their proportions naturally resemble those another face would
temporarily assume when emoting.
Automatic face detection is a complex problem in image processing. Many methods exist
to solve this problem such as template matching, Fisher Linear Discriminant, Neural
Networks, EIGENFACE, and MRC. Success has been achieved with each method to
varying degrees and complexities.
2.4 Face Detection Approaches
Face detection is the first step of face recognition system. Output of the detection can be
location of face region as a whole, and location of face region with facial features (i.e.
eyes, mouth, eyebrow, nose etc.). Mainly, detection can be classified into two groups as
Knowledge-Based Methods and Image-Based Methods. The methods for detection are
given in Figure 2.2.
Knowledge-Based methods use information about Facial Features, Skin Color or
Template Matching. Facial Features are used to find eyes, mouth, nose or other facial
features to detect the human faces. Skin color is different from other colors and unique,
and its characteristics do not change with respect to changes in pose and occlusion. Skin
color is modeled in each color spaces like RGB (Red-Green-Blue), YCbCr (Luminance35
Blue Difference Chroma-Red Difference Chroma), HSV (Hue- Saturation-Value), YUV
(Luminance-Blue Luminance Difference-Red Luminance Difference), and in statistical
models. Face has a unique pattern to differentiate from other objects and hence a template
can be generated to scan and detect faces.
Figure 2.2 Methods for face detection
2.4.1 Face detection by Template Matching
Once individual candidate face images are separated, template matching is used as not
only a final detection scheme for faces, but also for locating the centroid of the face. The
idea of template matching is to perform cross co-variances with the given image and a
template that is representative of the image. Therefore, in application to face detection,
the template should be a representative face - being either an average face of all the faces
in the training images, or an Eigen face. In our case, both templates were created. The
first step was to crop out the faces from each training image posted on the website. Using
these faces as our set of training images was justified since the same people would be
present in the actual test image- otherwise a larger and more diverse set of training faces
images.
36
2.4.2 Skin color distribution model
In conventional methods, all visible colors are divided into two groups: One is the “skin
color” and the other is not. However, consider two colors near the boundary of the skin
part. Although the difference between them is almost unnoticeable by a human viewer,
one is regarded as “skin color” and the other is not. This is unnatural, and is considered as
one of the reasons of instability in conventional methods for skin color detection.
SCDM is a fuzzy set of skin color. We use a large image set containing to investigate the
distribution of color of the human skin region in order to build the SCDM.
The procedure to build the SCDM is as follows:
1.
Manually select skin regions in each image.
2.
Prepare a table of 92 × 140 entries to record the 2-dimensional chromatic
histogram of skin regions, and initialize all the entries with zero.
3.
Convert the chromaticity value of each pixel in the skin regions UCS, and then
increase the entry of the chromatic histogram corresponding to it by one.
4.
Normalize the table by dividing all entries with the greatest entry in the table.
2.4.3 Face detection by Neural Network
Neural Nets are essentially networks of simple neural processors, arranged and
interconnected in parallel. Neural Networks are based on our current level of knowledge
of the human brain, and attract interest from both engineers, who can use Neural Nets to
solve a wide range of problems, and scientists who can use them to help further our
understanding of the human brain. Since the early stages of development in the 1970’s,
interest in neural networks has spread through many fields, due to the speed of processing
and ability to solve complex problems. As with all techniques though, there are
limitations. They can be slow for complex problems, are often susceptible to noise, and
can be too dependent on the training set used, but these effects can be minimized through
careful design. Neural Nets can be used to construct systems that are able to classify data
into a given set or class, in the case of face detection, a set of images containing one or
37
more face, and a set of images that contains no faces. Neural Networks consist of parallel
interconnections of simple neural processors. Neurons have many weighted inputs, that is
to say each input (p1, p2, p3… pm) has a related weighting (w1, w2, w3… wm)
according to its importance. Each of these inputs is a scalar, representing the data. In the
case of face detection, the shade of GRAY of each pixel could be presented to the neuron
in parallel (thus for a 10x10 pixel image, there would be 100 input lines p1 to p100, with
respective weightings w1 to w100, corresponding to the 100 pixels in the input image).
2.4.4 Face detection by Eigen Face Method
The motivation behind Eigen faces is that it reduces the dimensionality of the training set,
leaving only those features that are critical for face recognition.
Definition
1. The Eigen faces method looks at the face as a whole.
2. In this method, a collection of face images is used to generate a 2-D gray-scale image
to produce the biometric template.
3. Here, first the face images are processed by the face detector. Then we calculate the
Eigen faces from the training set, keeping only the highest Eigen values.
4. Finally we calculate the corresponding location in weight space for each known
individual, by projecting their face images onto the “face space”.
2.5 Image Processing
Image processing is any form of signal processing for which the input is an image, such
as a photograph or video frame; the output of image processing may be either an image
or, a set of characteristics or parameters related to the image. Most image-processing
techniques involve treating the image as a two-dimensional signal and applying standard
signal-processing techniques to it. Image processing usually refers to digital image
processing, but optical and analog image processing also are possible. The acquisition of
images (producing the input image in the first place) is referred to as imaging. Image
processing is a physical process used to convert an image signal into a physical image.
38
The image signal can be either digital or analog. The actual output itself can be an actual
physical image or the characteristics of an image. The most common type of image
processing is photography. In this process, an image is captured using a camera to create
a digital or analog image. In order to produce a physical picture, the image is processed
using the appropriate technology based on the input source type. In digital photography,
the image is stored as a computer file. This file is translated using photographic software
to generate an actual image. The colors, shading, and nuances are all captured at the time
the photograph is taken the software translates this information into an image. When
creating images using analog photography, the image is burned into a film using a
chemical reaction triggered by controlled exposure to light. The image is processed in a
darkroom, using special chemicals to create the actual image. This process is decreasing
in popularity due to the advent of digital photography, which requires less effort and
special training to product images. In addition to photography, there are a wide range of
other image processing operations. The field of digital imaging has created a whole range
of new applications and tools that were previously impossible. Face recognition
software, medical image processing and remote sensing are all possible due to the
development of digital image processing. Specialized computer programs are used to
enhance and correct images. These programs apply algorithms to the actual data and are
able to reduce signal distortion, clarify fuzzy images and add light to an underexposed
image. Animations are series of single images put together into a movie. The images
might be a volume view, a projection, a slice, a time point. The animation is done by just
playing the sequential data set, or by rotating 3D models or volume representations, by
zoom-in & fly through motions, changing of surfaces and transparencies, etc. Today’s
computer allows for calculating and representing animated sequences reasonably fast.
Movie files can be published i.e. in power point or on the web. Also interactive file
formats are possible. If one or more of the images in a data set is taken through a filter
that allows radiation that lies outside the human vision span to pass – i.e. it records
radiation invisible to us - it is of course not possible to make a natural color image. But it
is still possible to make a color image that shows important information about the object.
This type of image is called a representative color image. Normally one would assign
colors to these exposures in chromatic order with blue assigned to the shortest
39
wavelength, and red to the longest. In this way it is possible to make color images from
electromagnetic radiation far from the human vision area, for example x-rays. Most often
it is either infrared or ultraviolet radiation that is used.
2.6 Fusion of Face with Other Biometric Traits
In recent years, biometric authentication has seen considerable improvement in reliability
and accuracy, with some of the biometric traits offering practically good performance
[13] for a comparative survey of up to date biometric authentication technology. On the
other hand, even the best biometrics up to date is at rest facing several problems, some of
them natural to the technology itself. Biometric systems that use a single trait are called
unimodal systems, whereas those that integrate two or more traits are referred to as
multimodal biometric systems. A multimodal biometric system requires an integration
scheme to fuse the information obtained from the individual modalities. The multimodal
biometric systems [14] are found to be extremely useful and exhibit robust performance
over the unimodal biometric systems in terms of several constraints. The aim of any
multimodal system is to acquire multiple sources of information from different modalities
and minimize the error prone effect of mono modal systems. In particular, biometric
authentication systems normally suffer from enrollment problems due to non-universal
biometric traits, vulnerability to biometric spoofing or unsatisfactory accuracy caused by
noisy data acquisition in certain environments. Although some biometrics has gained
more popularity than other biometrics, but each such biometric has its own strengths and
limitations, and no single biometric is expected to meet the desired performance of the
authentication applications. Multi biometrics is relatively a new approach to overcome all
these problems. Driven by lower hardware costs, a multi biometric system uses multiple
sensors for data achievement. In 2002 the research paper “Multi Modal Technology
makes Biometrics Work” states the title of a recent press release from Aurora Defense
[15]. Certainly, multi biometric systems ensure significant improvements over single
biometric systems, for example, higher accuracy and increased resistance to spoofing.
They also claim to be more universal by enabling a user who does not have a particular
biometric identifier to still enroll and authenticate using other biometric traits, thus
eliminating enrollment problems. But can multi biometrics live up to the promotion? At a
40
first quick look, incorporating multiple biometrics into one system appears to be a very
sensitive and understandable concept. There are many different ways to actually merge
various sources of information to make a final authentication conclusion. Information
fusion strategies range from simple Boolean combination to complicated numerical
modeling.
Figure 2.3 Biometric fusion systems
The goal of information fusion, therefore, is to determine the best set of experts in a given
problem domain and devise an appropriate function that can optimally combine the
decisions rendered by the individual experts. The recognition process itself may be
viewed as the reconciliation of evidence pertaining to these multiple modalities. Each
modality on its own cannot always be reliably used to perform recognition. However, the
consolidation of information presented by these multiple experts can result in the accurate
determination or verification of identity. Several approaches can be adopted for
combining the different modalities [16]. Biometric fusion system is shown in figure 2.3.
2.7 Various biometric fusion techniques
In a multi modal biometric system, the information fusion can occur at any of the module
of feature extraction, matcher and decision. Multi biometric systems can be categorized
into five system architectures according to the strategies used for information fusion:

Fusion at image level

Fusion at Feature Extraction Level

Fusion at the Rank Level

Fusion at the Matching score Level
41

Fusion at the Decision Level
That is we classify the systems depending on how early in the authentication process the
information from the different sensors is combined. Biometric authentication is a chain
process, as depicted in Figure 2.4 for a more detailed explanation:
Figure 2.4 Authentication Process Flow
Fusion at the feature extraction level stands for immediate data mixing at the beginning
of the processing chain, while fusion at the decision level represents late integration at the
end of the process. The following section describes each of these architectures in detail:
2.7.1 Fusion at the Image Level
In practice, it is possible to combine only “compatible images”. Therefore, in the context
of facial images, image level combination is used only to combine multiple images of the
same face. Facial images first need to be registered with each other before fusion. In case
facial images from multiple sensors are combined, the sensors must be pre-registered.
The objective of such a fusion is to improve the quality of acquired facial images. The
basis of this improvement is that the useful signal in the facial images captured with
different sensing technologies will be independent because different imaging conditions
capture different surface and sub-surface properties of the skin. Another reason to
conduct the image level fusion of facial images is to increase the acquired fingerprint area
wherein each individual image has captured only a portion of the face.
42
Figure 2.5 Fusion at the Image level
2.7.2 Fusion at the Feature Extraction Level
In this architecture, the information extracted from the different sensors is encoded into a
joint feature vector, which is then compared to an enrollment template which itself is a
joint feature via stored in a database and assigned a matching score as in a single
biometric system.
Figure 2.6 Fusion at the Feature Extraction level
However, feature-level fusion may be preferred because features are more compact than
image, leading to efficient fusion algorithms. But fusion at the feature extraction level is
much less preferable than the other next strategies. Two main problems with this
approach are used to identify the problems:

The feature vectors to be joined might be incompatible due to numerical
problems, or some of them might even be unavailable in cases where the users
possess all biometric identifiers. While the first issue might be resolved by
43
careful system design, leading to a very tightly coupled system, the second one
will cause the enrollment problems we already know from single biometric
systems.

Score generation is problematic: even in a single biometric system, it is difficult
to find a good classifier, i.e. to generate a representative score based on the
matching of feature vector and enrollment template. But for the high-dimensional
joint feature vectors in a multi biometric system, it is even more complicated. As
pointed out in [17], the relationship between the different components of the joint
feature vector may not be linear.
2.7.3 Fusion at the Rank Level
Rank-level fusion is used only in identification systems and is applicable when the
matcher output is a ranking of the “candidates” in the template database. The system is
expected to assign a higher rank to a template that is more similar to the query. Most
identification system actually provides the matching score associated with the candidates.
Therefore, the rank level fusion is widely used in other fields such as pattern recognition
and data mining. This method involves combining identification ranks obtained from
multiple unimodal biometrics. It consolidates a rank that is used for making final
decision. In multimodal biometric system, rank level fusion is the method of
consolidating ranks from different biometric modalities (fingerprints, facial features, iris,
retina etc.) to recognize a user. Ho et al., 1994 describe three methods to combine the
ranks assigned by different matchers. In the highest rank method, each possible identity is
assigned the best (minimum) of all ranks computed by different systems. Ties are broken
randomly to arrive at a strict ranking order and the final decision is made based on the
consolidated ranks. The Borda count method uses the sum of the ranks assigned by the
individual systems to a particular identity in order to calculate the fused rank. The logistic
regression method is a generalization of the Borda count method where a weighted sum
of the individual ranks is used. The weights are determined using logistic regression.
44
2.7.4 Fusion at the Matching Score Level
Each system provides a matching score indicating the proximity of the feature vector
with the template vector. These scores can be combined to assert the veracity of the
claimed identity. Techniques such as logistic regression may be used to combine the
scores reported by the two sensors.
Figure 2.7 Fusion at the Matching Score level
In a multi biometric system built on this architecture, feature vectors are created in
parallel for each sensor and then compare it with the enrollment templates, which are
stored independently for each biometric sample. Based on the closeness of feature vector
and template, each subsystem now computes its own matching score. These users’ scores
are finally combined into a total score, which is handed over to the decision module. The
complete process is shown in figure 2.7.
Score level fusion is widely recognized to offer the best tradeoff between the
effectiveness of fusion and the ease of fusion. While the information contained in
matching scores is not as rich as in images or features, it is much richer than ranks and
decisions. Further, while score-level fusion is not as easy or intuitive as rank-level or
decision-level fusion, it is easier to study and implement than image-level and featurelevel fusion. Scores are typically more accessible and available than images or features
but it does not contain rich information than images or features. Also, fusion at the score
level requires some care. But the main difficulties in the score level fusion may emanate
45
from the non-homogeneity of scores from different matchers, differences in the
distributions of the scores, correlation among the scores, and differences in the accuracies
of different matchers. All scores were mapped to the range [0; 100]. A very welldesigned example for this fusion strategy has recently been presented by Ross and Jain in
two research papers [18]. They include facial scan, iris verification and hand geometry
scan into a common authentication system, and using well known methods for each
identifier (Eigen faces for the facial scan, patterns for the iris system, and commonly used
hand geometry features). A score vector - (x1; x2; x3) - represents the scores of multiple
matchers, with x1, x2 and x3 corresponding to the scores obtained from the 3 modalities.
Matching scores for the three different modalities are then normalized and combined
using one of the following strategies.
1. Sum Rule: The sum rule method of integration takes the weighted average of the
individual score values. This strategy was applied to all possible combinations of the
three modalities. Equal weights were assigned to each modality as the bias of each
matcher was not available.
2. Decision Tree: The C5:0 programs was used to generate a decision tree from the
training set of genuine and impostor score vectors. The training set consisted of 11; 025
impostor score vectors and 225 genuine score vectors. The test set consisted of the same
number of independent impostor and genuine score vectors. This strategy uses a sequence
of threshold comparisons on the different scores to make an authentication decision.
3. Linear Discriminant Function: Linear discriminant analysis of the training set helps
in transforming the 3-dimensional score vectors into a new subspace in which the
separation between the classes of genuine user scores and imposter scores is maximized.
The test set vectors are classified by using the minimum Mahalanobis distance rule (with
the assumption that the two classes have unequal covariance matrices). The optimal
parameters for this transformation are calculated in advance based on a training set. The
output score is defined as the minimum distance to the centroids of the two classes, using
a special metric, the Mahalanobis distance.
46
Based on the experimental results, the authors make the observation that the sum rule
achieves the best performance. Most importantly, they further extend the sum rule using a
really new approach: they suggest applying user-specific weights to the user biometric
samples to be combined as well as using user-specific threshold levels for making the
final authentication decision.
2.7.5 Fusion at the Decision Level
In this fusion strategy, a separate authentication decision is made for each biometric trait.
These decisions are then combined into a final vote, as shown below in figure 2.8.
Decision-level fusion is not as popular as score-level or rank-level fusion. Still, it may be
the only feasible approach if the commercial biometric systems involved in fusion
provide only the final match decision. Similar to the rank-level and score-level fusions,
decision-level fusion is conducted at a high-level, therefore it does not matter which
entities are being combined. Each sensor can capture multiple biometric data and the
resulting feature vectors individually classified into the two classes––accept or reject. A
majority vote scheme, such as that employed in (Zuev and Ivanon, 1996) can be used to
make the final decision.
Figure 2.8 Fusions at the Decision Level
Many different strategies are available to combine the dissimilar decisions into a final
authenticated decision. They range from majority votes to complicated statistical
47
methods. In practice, however, developers seem to prefer the easiest method: Boolean
Conjunctions. Fusion at the decision level occurs at a very late stage of the authentication
process.
2.8 Summary
In this chapter we have discussed all about the face recognition and detection system.
And we also discussed the face detection approaches and also about the image
processing. In this chapter, we compared the fusion of face with other biometric traits.
Multi modal biometric systems integrate the information presented by multiple biometric
indicators. The information can be consolidated at various levels such as image level,
feature extraction level, matching score level, rank level and decision level.
Fusion at the image level: In this level of fusion, we combine only compatible images
because they can be easily combined. Also, we capture the multiple images of the same
biometric traits and then fuse all the images to get a single image of higher quality. This
method is used to increase the acquired facial image area wherein each individual image
has captured only a portion of the face.
Fusion at the feature extraction level: The data obtained from each biometric modality
is used to compute a feature vector. If the features extracted from one biometric indicator
are independent of those extracted from the other, it is reasonable to concatenate the two
vectors into a single new vector, provided the features from different biometric indicators
are in the same type of measurement scale.
Fusion at the rank level: Rank level fusion involves combining identification ranks
obtained from multiple unimodal biometrics. It consolidates a rank that is used for
making final decision. In multi modal biometric system, rank level fusion is the method
of consolidating ranks from different biometric modalities to recognize a user.
Fusion at the matching scores level: Each biometric matcher provides a similarity score
indicating the proximity of the input feature vector with the template feature vector.
These scores can be combined to assert the veracity of the claimed identity.
Fusion at the decision level: each biometric system makes its own recognition decision
based on its own feature vector. A majority vote scheme can be used to make the final
recognition decision.
48
Chapter 3
Literature Survey
For the purpose of this dissertation, the literature survey covers a period of 2000 to 2013.
The dissertation work on “Face and Facial Expression Recognition” is divided into five
main categories:
1. Technology, Application, Challenge and computational Intelligence Solutions.
2. Security issues in biometric Authentication.
3. Emerging methods of Biometric user Authentication.
4. Performance issues in Biometric Authentication Systems.
5. Comparison of various Biometric Authentication Systems.
Anil K. Jain, Arun Ross and Salil Prabhakar in their paper “An Introduction to
Biometrics Recognition” designed a Biometric Recognition system using the four main
modules [19]:
1. Sensor module, which captures the biometric data of an individual.
2. Feature extraction module, in which the acquired biometric data is processed to
extract a set of salient or discriminatory features.
3. Matcher module, in which the features during recognition are compared against the
stored templates to generate matching scores.
4. System database module, which is used store the biometric templates of the enrolled
users.
Weicheng Shen and Tieniu Tan in their paper “Automated Biometrics-based personal
Identification”, proposed a typical automated biometrics-based identification/verification
system that consists of the following major components [20]:
1. Data acquisition component that acquires the biometric data in digital format by
using a sensor.
49
2. Feature extraction component that uses an algorithm to produce a feature vector in
which the components are numerical characterizations of the underlying biometrics.
The feature vectors are designed to characterize the biometrics so that biometric
data collected from one individual, at different times, are “similar”, while those
collected from different individuals are “dissimilar”. In general, the larger the size
of a feature vector (without much redundancy), the higher will be its discrimination
power which is defined as the difference between a pair of feature vectors
representing two different individuals.
3. Matcher component that compares feature vectors obtained from the feature
extraction algorithm to produce a similarity score. This score indicates the degree of
similarity between a pair of biometrics data under consideration.
4. Decision-maker component is the last component of the system that finally
provides/rejects access to the user based on some pre-determined criterion.
Michal Chora in his paper “Emerging Methods of Biometric Human Identification” gives
the idea that Human identification biometrics system may originate from real-life
criminal and forensic applications. Some methods, such as fingerprinting and face
recognition, already proved to be very efficient in computer vision based human
recognition systems. In this paper he focuses on emerging methods of biometrics human
identification also used in the forensic and criminal practice. He says it is our motivation
to design computer vision system based on innovative computing and image processing
that would be used to identify humans on the basis of their ear, palm and lips images
[21]. Those methods have been successfully used by the police and forensic experts and
now gain attention of computer science community. In the article he proposed innovative
methods of ear, lips and palm print biometrics to identify humans.
Qinghan Xiao in his paper “Biometrics Technology, Application, Challenge and
Computational Intelligence Solutions” states that an increasing number of countries have
decided to adopt biometric systems for national security and identity theft prevention
[22]. This trend makes biometrics an important component in security related
applications such as: logical and physical access control, forensic investigation, IT
security, identity fraud detection and terrorist prevention or detection. This paper aims to
50
assist readers as they consider biometric solutions by examining common biometric
technologies, introducing different biometric applications.
Arun Ross, Anil Jain in their paper “Information Fusion in Biometrics” proposed a MultiBiometric system which seeks to alleviate some of drawbacks of single biometric system
by providing multiple evidences of the same identity [23]. These systems help to achieve
an increase in performance that may not be possible using a single biometric indicator.
Multi-Biometric systems provide anti-spoofing measures by making it difficult for an
intruder to spoof multiple biometric traits simultaneously. The authors proposed Fusion
in biometrics with certain levels.
Sanjay Kr. Singh, D. S. Chauhan, Mayank Vatsa, Richa Singh in their paper “A Robust
Skin Color Based Face Detection Algorithm” described a detailed experimental study of
face detection algorithms based on “Skin Color” has been made [24]. The colors of the
human skin represent a special category of colors, because they are distinctive from the
colors of other natural objects. This category is found as a cluster in color spaces, and the
skin color variations between people are mostly due to the intensity differences. Besides,
the face detection based on skin color detection is a faster method as compared to other
techniques. In this paper, the author present a system to track faces by carrying out skin
color detection in four different color spaces: HSI, YCbCr, YES and RGB. Once some
skin color regions have been detected for each color space, author labeled each region
and gets some characteristics such as size and position. The authors were supposing that a
face is located in one the detected regions. After that they compare and employ a polling
strategy between labeled regions to determine the final region where the face effectively
has been detected and located. The authors have compared the algorithms based on these
color spaces and have combined them to get a new skin color based face detection
algorithm which gives higher accuracy. In this paper, the author includes some
experimental results which show that the proposed algorithm is good enough to localize a
human face in an image with an accuracy of 95.18%.
George Chellin, Chandran J and Dr. Rajesh R.S in their paper “Performance Analysis of
Multimodal Biometric System Authentication” states that Traditional identity verification
in computer systems are done which is based on the Knowledge based and Token based
51
identification are leading to fraud. Unfortunately, these may often be forgotten, disclosed
or changed [25]. A reliable and accurate identification/verification technique may be
designed using biometric technologies. Biometric authentication employs unique
combinations of measurable physical characteristics- fingerprint, facial features, iris of
the eye, voice print, hand geometry, vein patterns and so on that cannot be readily
imitated or forged by others. Unimodal biometric systems have variety of problems such
as noisy data, intra-class variations, restricted degree of freedom, non-universality, spoof
attacks and unacceptable error rates. Multimodal biometrics refers the combination of
two or more biometric modalities in a single identification system. The purpose of this
paper is to identify whether the integration of iris and fingerprint biometrics overcome
the hurdles of unimodal biometric system. This paper discuss the various scenarios that
are possible to improve the performance of multimodal biometric systems using the
combined characteristics such as iris and fingerprint, the level of fusion (multimodal
fusion) is applied to that are possible and the integration strategies that can be adopted in
order to increase the overall system performance. Information from multiple sources can
be consolidated in three distinct levels: (i) feature extraction level; (ii) match score level;
(iii) measurement level; and (iv) decision level.
J. D. Woodward in his paper “Biometrics: Privacy’s Foe or Privacy’s Friend?” states that
both the public and private sectors are making extensive use of biometrics for human
recognition [26]. As this technology becomes more economically viable and technically
perfected, and thus more commonplace, the field of biometrics will spark legal and policy
concerns. Critics inevitably compare biometrics to Big Brother and the loss of individual
privacy. The biometric lobby generally stresses the greater security and improved service
that the technology provides. Is biometrics privacy’s friend or privacy’s foe? This paper
explains the various arguments for and against biometrics and contends that while
biometrics may pose legitimate privacy concerns, these issues can be adequately
addressed. In the final analysis, biometrics emerges as privacy’s friend.
L. Hong and A. K. Jain in their paper “Integrating Faces and Fingerprints for Personal
Identification” describes that an automatic personal identification system based only on
the fingerprints or faces is often not able to meet the requirements of the system’s
52
performance i.e. response time requirements as well as the accuracy requirements [27].
Face recognition is fast technology but not extremely reliable, while fingerprint
verification is reliable but inefficient in database retrieval. In this paper, the authors
developed a prototype biometric system which integrates faces and fingerprints. The
system overcomes the limitations of face recognition systems as well as fingerprint
verification systems. The integrated prototype system operates in the identification mode
with an admissible response time. The author describes that the identity established by
the system is more trustworthy than the identity established by a face recognition system.
In addition, the proposed decision fusion scheme makes the performance improvement by
integrating multiple cues with different confidence measures. Experimental results
demonstrate that this system performs very well. It meets the response time as well as the
accuracy requirements. Experimental results demonstrate that the system performs very
well. It meets the response time as well as the accuracy requirements.
Jun Ou, Xiao-Bo Bai Yun Pei ,Liang Ma, Wei Liu [28] in their paper “Automatic Facial
Expression Recognition Using Gabor Filter And Expression Analysis” presents a system
that uses 28 facial feature key-points in images detection and Gabor wavelet filter
provided with 5 frequencies, and 8 orientations. In according to actual demand, the
system can actually extract the feature of low quality facial expression image target, and
is much robust for automatic facial expression recognition. The authors shows some
experimental results in their paper which shows that the performance of the proposed
method achieved excellent average recognition rates, when it is applied to facial
expression recognition system.
Ming Hu in his paper “Application of Rough Sets to Image Pre-processing for Face
Detection” explains that both of face detection and face recognition play a very important
role in the technology of biological character identity [29]. During the processing of face
image, it is indispensable to pre-processing. In this paper, a Pre-processing method for
face color image based on rough sets theory is put forward. In this method, the image is
divided as foreground and background by condition attribution. In order to extract the
skin color regions, foreground is segmented by two kinds of the skin color model, one is
53
YCbCr and the other is HSV color spaces. Then, the opening and closing morphological
operation are used to remove the small regions.
Jiying Wu in his paper “Multi-Scale Preprocessing Model for Face Recognition”
proposed a novel multi-scale preprocessing model (MSPM) for face recognition [30].
MSPM removes lighting effects and enhances the image feature in two scales
simultaneously. It decomposes the original image using Total Variation model. Then the
lighting effects are normalized by self-quotient in the small scale part and equalized in
the large scale part.
Bongjin Jun in his paper “Statistical Face Image Preprocessing and Non-statistical Face
Representation for Practical Face Recognition” states that recognizing face images in real
environment is still a challenging problem since there are many illumination changes
[31]. In this paper, the author proposed a practical face recognition method that combines
statistical global illumination transformation and non-statistical local face representation
method. When a new face image is given, it is transformed into a number of face images
exhibiting different illuminations using a statistical bilinear model-based indirect
illumination transformation. Each illumination transformed image is then represented by
a histogram sequence that concatenates the histograms of the non-statistical multiresolution uniform local Gabor binary patterns (MULGBP) for all the local regions.
Yong Zhang in his paper, “Hand-Drawn Face Sketch Recognition by Humans and a
PCA-Based Algorithm for Forensic Applications” describes that face sketches represent
the original faces in a very concise manner but they are still in a recognizable form,
therefore they play an important role in criminal investigations, human visual perception,
and face biometrics [32]. In this paper, the author compared the performances of humans
and a principle component analysis (PCA)-based algorithm in recognizing face sketches.
A total of 250 sketches of 50 subjects were involved. All of the sketches were drawn
manually by five artists (each artist drew 50 sketches, one for each subject).
Anil K. Jain, Arun Ross, Umut Uludag in their paper “Biometric Template Security:
Challenges and Solutions” explains their ideas that a biometric system is vulnerable to a
variety of attacks aimed at breaking the integrity of the authentication process [33]. These
54
attacks are intended to either circumvent the security afforded by the system or to check
the normal functioning of the system. In this paper, the authors describe the various
threats that can be faced by a biometric system. Therefore, the authors specifically focus
on attacks designed to elicit information about the original biometric data of an individual
from the stored template. Cancelable biometrics may be used to “reset” the biometric
template of a user in the event that the user’s template is compromised. Also, biometric
cryptosystems can also provide contributions to the template security scheme by
supporting biometric matching in secure cryptographic domains.
Jian Yang in his paper “Constructing PCA Baseline Algorithms to Reevaluate ICA-Based
Face-Recognition Performance” states that an independent component analysis (ICA)based face recognition generally evaluates its performance using standard principal
component analysis (PCA) within two architectures, ICA Architecture I and ICA
Architecture II [34]. In this paper, the author analyzes these two ICA architectures and
find that ICA Architecture I involves a vertically centered PCA process (PCA I), while
ICA Architecture II involves a whitened horizontally centered PCA process (PCA II).
Thus, it makes sense to use these two PCA versions as baselines to re-evaluate the
performance of ICA-based face-recognition systems.
Walid Riad Boukabou in his paper “An Improved LDA Approach with DFB Processing
for Face Recognition” states that recently many face recognition systems were developed
[35]. However changing in pose and illumination conditions remains largely unsolved.
Therefore in this paper, the author presents a new method for face recognition that was
based on Linear Discriminant Analysis with a directional filter bank pre-processing. In
this paper, the author shows some experimental results which shows that the DFB preprocessing step improves the original LDA method and a very good recognition rate is
obtained which leads to performance improvement.
Bruce A. Draper in his paper, “Recognizing Faces with PCA and ICA” compares the
principal component analysis (PCA) and independent component analysis (ICA) in the
context of a baseline face recognition system, and a comparison is motivated by
contradictory claims in the literature [36]. This paper shows how the relative performance
of PCA and ICA depends on the task statement, the ICA architecture, the ICA algorithm,
55
and (for PCA) the subspace distance metric. It then also explains the space of PCA/ICA
comparisons by systematically testing two ICA algorithms and two ICA architectures
against PCA with four different distance measures on two tasks (facial identity and facial
expression). In the process, this paper verifies the results of many of the previous
comparisons in the literature, and relates them to each other and to this work. Author are
able to show that the Fast ICA algorithm configured according to ICA architecture II
yields the highest performance for identifying faces, while the Info Max algorithm
configured according to ICA architecture II is better for recognizing facial actions. In
both cases, PCA performs well but not as well as ICA.
Rabia Jafri and Hamid R. Arabnia in their paper “A Survey of Face Recognition
Techniques” stated that Facial recognition system presents a very challenging problem in
the field of the image analysis and computer vision [37], and it also has been received a
great deal of attention over the last few decades because of its many applications in the
various fields. According to them, Face recognition techniques can be broadly divided
into three main categories based on the facial data acquisition methodology. 1) Methods
that operate on intensity images. 2) Those that deal with video sequences, 3) and those
that require other sensory data such as 3D information of images or infra-red imagery.
This paper also defines the need of using the face recognition technology. In this paper,
an overview of some of the well-known methods in each of these categories is provided
and some of the benefits and drawbacks of the schemes mentioned are also examined.
Furthermore, a discussion outlining the benefits for using face recognition, the
applications of this technology, and some of the difficulties plaguing current systems
with regard to this task has also been provided. This paper also discusses some of the
most recent algorithms that was developed for this purpose and tries to give an idea of the
state of the art for the face recognition technology.
Hu Han and Anil K. Jain in their paper “3D Face Texture Modeling from Uncalibrated
Frontal and Profile Images” reveals that 3D face modeling from 2D face images is of
significant importance for facial image analysis, animation and also for recognition [38].
Study tells that previous research on this topic mainly focused on 3D face modeling from
a single 2D face image; however, a single face image can only provide a limited
56
description of a 3D face. In many applications, for example, law enforcement, multi-view
face images are usually captured for a subject during enrollment, which makes it
optimistically build a 3D face texture model, given a pair of frontal and profile facial
images. The author first calculated the relationship between the uncalibrated frontal and
profile face images through facial landmark alignment. An initial 3D face shape is then
reconstructed from the frontal face image, followed by shape refinement utilizing the
depth information provided by the profile image. Finally, face texture is extracted by
mapping the frontal face image on the recovered 3D face shape. In this paper, the
proposed method is utilized for 2D face recognition in two scenarios: one is the
normalization of probe image, and the other one is enhancing the representation
capability of gallery set. Experimental results comparing the proposed method with a
state-of-the-art commercial face matcher and densely sampled LBP on a subset of the
FERET database show the effectiveness of the proposed 3D face texture model.
Hu Han, Charles Otto, and Anil K. Jain in their paper “Age Estimation from Face
Images: Human vs. Machine Performance” states that there has been a growing interest in
automatic age estimation from facial images due to a variety of potential applications in
law enforcement agencies, security control applications, and human computer interaction
(HCI) [39]. However, rather than of these advances in automatic age estimation, it still
remains a challenging problem. This is because the facial aging process is determined not
only by internal factors, e.g. genetic factors, but also some external factors, for e.g.
lifestyle, expression, and environment contributes in the determination process.
Therefore, as a result of it, different people which are of the same age can have quite
different appearances due to different rates of facial aging. In this paper, the authors
proposed a hierarchical approach for automatic age estimation, and provide an analysis of
how aging influences individual facial components. In this paper, the author also shows
some experimental results on the FG-NET, MORPH Album2, and PCSO databases
which show that eyes and nose contains much more information than the other
components and these are also the more stable components than the other facial
components in automatic age estimation. They also studies the ability of humans to
estimate the age using data collected via crowd sourcing or by their ability to remember,
57
and shows that the cumulative score (CS) within 5-year mean absolute error (MAE) of
their method is better than the age estimates provided by human beings.
Brendan F. Klare and Anil K. Jain in their paper “Face Recognition: Impostor-based
Measures of Uniqueness and Quality” presents a framework, named uniqueness-based
non-match estimates (UNE), which explores the ability to improve the face recognition
performance of any face matcher [40]. The first goal of this framework is that a novel
metric that measures the uniqueness of a given user, called the impostor-based
uniqueness measure (IUM). The UNE maps the facial matching scores from any facial
matcher into non-match probability estimates that are conditionally dependent on the
probe image’s IUM. In this paper by using this framework the author justifies that (i) an
improved generalization of matching thresholds, (ii) a score normalization technique that
improves the interoperability for users of different face matchers, and (iii) the predictive
ability of IUM towards the facial recognition accuracy. Studies are conducted on an
operational dataset with 16,000 subjects using three different face matchers in which two
are the commercial, and one is proprietary matcher to illustrate the effectiveness of the
proposed framework.
Zhifeng Li, Unsang Park and Anil K. Jain, in their paper “A Discriminative Model for
Age Invariant Face Recognition” stated that aging variation puts a very challenging
problem to the automatic face recognition systems [41]. Most of the face recognition
studies that have addressed the aging problem are mainly focused on the age estimation
or aging simulation. Designing an appropriate feature representation and an effective
matching framework for age invariant face recognition still remains a big challenging
problem. In this paper, the author proposed a discriminative model to address face
matching in the presence of age variability. In this framework, the author first represent
each face by designing a densely sampled local feature description scheme, in which
scale invariant feature transform (SIFT) and multi-scale local binary patterns (MLBP)
serve as the local descriptors. By densely sampling the two kinds of local descriptors
from the entire facial image, sufficient discriminatory information, including the
distribution of the edge direction in the face image (that is expected to be age invariant)
can be extracted for further analysis. Since both SIFT-based local features and MLBP58
based local features span a high-dimensional feature space, to avoid the over fitting
problem, the author developed an algorithm, called multi-feature discriminant analysis
(MFDA) to process these two local feature spaces in a unified framework. The MFDA is
an extension and improvement of the LDA using multiple features combined with two
different random sampling methods in feature and sample space. By random sampling the
training set as well as the feature space, multiple LDA-based classifiers are constructed
and then combined to generate a robust decision via a fusion rule. The author shows some
experimental results which show that this approach outperforms a state-of-the-art
commercial face recognition engine on two public domain facial aging data sets:
MORPH and FG-NET. They also compared the performance of the proposed
discriminative model with a generative aging model. A fusion of discriminative and
generative models further improves the facial matching accuracy still in the presence of
aging.
Volker Blanz and Thomas Vetter in their paper “Face Recognition Based on Fitting a 3D
Morphable Model” presents a method for face recognition across the many variations in
pose, ranging from frontal to profile views, and across a wide range of illuminations and
changing lighting conditions, including cast shadows and peculiar reflections [42]. To
account for these variations, the algorithm simulates the process of image formation in
3D space, using computer graphics, and it estimates 3D shape and texture of faces from
single images. These estimations are achieved by fitting a statistical, morph able model of
3D faces to images. This model is then trained from using a set of textured 3D scans of
heads. The author describes the construction of the morph able model, an algorithm to fit
the model to images, and a framework for face identification. In this framework, faces are
represented by model parameters for 3D shape and texture. The author also presents the
results obtained with 4,488 images from the publicly available CMU-PIE database and
1,940 images from the FERET database.
Zahid Riaz, Christoph Mayer, Matthias Wimmer, and Bernd Radig in their paper “Model
Based Face Recognition across Facial Expressions” describes a novel idea of face
recognition across facial expression variations using a model-based approach [43]. This
approach is followed in 1) modeling an active appearance model (AAM) for the face
59
image, 2) using optical flow based temporal features for facial expression variability
estimation, 3) and finally applying binary decision trees as a classifier for facial
identification. The novelty lies not only in the generation of appearance models which is
obtained by fitting active shape model (ASM) to the face image using objective but also
using a feature vector which is the combination of shape, texture and temporal parameters
that is robust against facial expression variability. Many experiments have been
performed on Cohn-Kanade facial expression database using 61 subjects of the database
with image sequences consisting of more than 4000 images. The author shows that this
approach has been achieved successful recognition rate up to 91.17% using decision tree
as a classifier in the presence of six different facial expressions.
A.A. Salah, M. Biceg, L. Akarun, E. Grosso, M. Tistarelli in their paper “Hidden Markov
Model-based face recognition using selective attention” describes that the sequential
methods for face recognition relies upon the analysis of local facial features in a
sequential manner, typically with a raster scan [44]. However, the distribution of
discriminative information is not uniform over the facial surface. For example, the nose
and the mouth contain much more information than the cheek or other facial components.
In this paper, the author proposed an extension to the sequential approach, where they
takes into account the local facial features saliency, and replace the raster scan with a
guided scan that mimics the scan path of the human eye. The selective attention
mechanism that guides the human eye operates by coarsely detecting salient locations,
and directing more resources at some interesting or informative parts. The author
simulates this idea by employing a computationally cheap saliency scheme, based on
Gabor wavelet filters. Hidden Markov models are used for classification, and the
observations, i.e. features obtained with the simulation of the scan path, are modeled with
Gaussian distributions at each state of the model. In this paper, the author shows that by
firstly visiting the important locations, this method is able to reach high accuracy with
much shorter feature sequences. The author compared the several features in observation
sequences, among which DCT coefficients results in the highest accuracy.
60
Congyong Su and Li Huang in their paper “Spatio-Temporal Graphical-Model-Based
Multiple Facial Feature Tracking” proposed a spatio-temporal graphical model for
multiple facial feature tracking [45]. Here the graphical model that is described above is
not 2D or 3D facial mesh model. Nonparametric belief propagation is used to infer facial
feature’s interrelationships in a part-based face model, allowing positions and states of
some features in clutter to be recovered. Facial structure is also taken into account,
because facial features have spatial position constraints. The problem is to track multiple
facial features simultaneously when rich expressions are presented on a face. In this
paper, the author proposed a two step solution. In the first step, several independent
condensation-style particle filters are utilized to track each facial feature in the temporal
domain. Particle filters are very effective for visual tracking problems; however multiple
independent trackers ignore the spatial constraints and the natural relationships among
facial features. In the second step, the author use the Bayesian inference belief
propagation to infer each facial feature’s contour in the spatial domain, in which the
author explains the relationships among contours of facial features beforehand with the
help of a large facial expression database. The tracking results in this paper can be used
as motion capture data. Author has planned to use these data to derive a 3D face model
and generate facial animations. The ultimate purpose of multiple facial feature tracking is
for facial animation.
Yan Tong, Wenhui Liao and Qiang Ji in their paper “Facial Action Unit Recognition by
Exploiting their Dynamic and Semantic Relationships” describes that a system could
automatically analyze the facial actions in real time has applications in a wide range of
different fields [46]. However, developing such a system is always challenging due to the
richness, ambiguity, and dynamic nature of facial actions. Although a number of research
groups attempt to recognize facial action units (AUs) by improving either the facial
feature extraction techniques or the AU classification techniques, these methods often
recognize AUs or certain AU combinations individually and statically, ignoring the
semantic relationships among AUs and the dynamics of AUs. Hence, these approaches
cannot always recognize AUs reliably, robustly, and consistently. In this paper, we
propose a novel approach that systematically accounts for the relationships among AUs
and their temporal evolutions for AU recognition. Specifically, we use a dynamic
61
Bayesian network (DBN) to model the relationships among different AUs. The DBN
provides a coherent and unified hierarchical probabilistic framework to represent
probabilistic relationships among various AUs and to account for the temporal changes in
facial action development. Within the system developed by the author, robust computer
vision techniques are used to obtain AU measurements. Such AU measurements are then
applied as evidence to the DBN for inferring various AUs. The author includes some
experiments in his paper which shows that the integration of AU relationships and AU
dynamics with AU measurements yields significant improvement of AU recognition,
especially for spontaneous facial expressions and under more realistic environment
including illumination variation, face pose variation, and occlusion.
Maja Pantic, and Leon J. M. Rothkrantz in their paper “Facial Action Recognition for
Facial Expression Analysis from Static Face Images” discuss the automatic recognition
of facial gestures for e.g., facial muscle activity is rapidly becoming an area of intense
interest in the research field of machine vision [47]. In this paper, the author present an
automated system that was developed to recognize facial gestures in static, frontal- and/or
profile-view color face images. A multi detector approach to facial feature localization is
utilized to spatially sample the profile contour and the contours of the facial components
such as the eyes and the mouth. From the extracted contours of the facial features, we
extract ten profile-contour fiducial points and 19 fiducial points of the contours of the
facial components. Based on these, 32 individual facial muscle actions (AUs) occurring
alone or in combination are recognized using rule-based reasoning. With each scored AU,
the utilized algorithm associates a factor denoting the probability with which the pertinent
AU has been scored and a recognition rate of 86% is achieved.
Michael A. Sayette, Jeffrey F. Cohn, Joan M. Wertz, Michael A. Perrott, and Dominic J.
Parrott in their paper “A Psychometric Evaluation of the Facial Action Coding System
for Assessing Spontaneous Expression” describes that the Facial Action Coding System
(FACS) (Ekman & Friesen, 1978) is a comprehensive and widely used method of
objectively describing the facial activities [48]. Little is known, however, about interobserver reliability in coding the occurrence, intensity, and timing of individual FACS
action units. The present study evaluated the reliability of these measures. Observational
62
data came from three independent laboratory studies designed to elicit a wide range of
spontaneous expressions of emotion. Emotion challenges included olfactory stimulation,
social stress, and cues related to nicotine craving. Facial behavior was video-recorded and
independently scored by two FACS-certified coders. Overall, we found good to excellent
reliability for the occurrence, intensity, and timing of individual action units and for
corresponding measures of more global emotion-specified combinations.
Ying-li Tian, Takeo Kanade and Jeffrey F. Cohn in their paper “Recognizing Action
Units for Facial Expression Analysis” describes that most automatic expression analysis
systems attempt to recognize a small set of prototypic expressions, such as happiness,
anger, disgust, sadness, surprise, and fear [49]. Such prototypic expressions, however,
occur rather infrequently. Human emotions and intentions are more often communicated
by changes in one or a few discrete facial features. In this paper, the authors developed an
Automatic Face Analysis (AFA) system to analyze facial expressions based on both
permanent facial features like brows, eyes, mouth and transient facial features such as
deepening of facial furrows in a nearly frontal-view face image sequence. The AFA
system recognizes fine-grained changes in facial expression into action units (AUs) of the
Facial Action Coding System (FACS), instead of a few prototypic expressions. Multistate face and facial component models are proposed for tracking and modeling the
various facial features, including lips, eyes, brows, cheeks, and furrows. During tracking,
detailed parametric descriptions of the facial features are extracted. With these
parameters as the inputs, a group of action units (neutral expression, 6 upper face AUs,
and 10 lower face AUs) are recognized whether they occur alone or in combinations. The
system has achieved average recognition rates of 96.4% (95.4% if neutral expressions are
excluded) for upper face AUs and 96.7% (95.6% with neutral expressions excluded) for
lower face AUs. The generalizability of the system has been tested by using independent
image databases collected and FACS-coded for ground-truth by different research teams.
Alexander M. Bronstein, Michael M. Bronstein and Ron Kimmel in their paper “ThreeDimensional Face Recognition” presented an expression-invariant 3D face recognition
approach [50]. In this paper the basic assumption was that the facial expressions can be
modeled as isometrics of the facial surface. This allows constructing an expression63
invariant representation of faces using the canonical forms approach. The result is an
efficient and accurate face recognition algorithm, robust to facial expressions that can
distinguish between identical twins. The author explains a prototyping system that was
based on the proposed algorithm and compares its performance to classical face
recognition methods. The numerical methods employed by our approach do not require
the facial surface explicitly. The surface gradients field, or the surface metric, is sufficient
for constructing the expression-invariant representation of any given face. It allows us to
perform the 3D face recognition task while avoiding the surface reconstruction stage.
Claude C. Chibelushi, Fabrice Bourel in their paper “Facial Expression Recognition: A
Brief Tutorial Overview” describes that facial expressions convey non-verbal cues,
which play an important role in interpersonal relationships [51]. Automatic facial
expressions recognitions can be an important component of natural human-machine
interfaces; it may also be used in behavioral science and in clinical practice.
Although humans recognize facial expressions virtually without effort or delay,
reliable expression recognition by machine is still a challenging problem. This
paper presents a high-level overview of automatic expression recognition; it
highlights the main system components and also some research challenges.
Liwei Wang, Yan Zhang, Jufu Feng in their paper “On the Euclidean Distance of
Images” presents a new Euclidean distance for images, which we call IMage Euclidean
Distance (IMED) [52]. Unlike the traditional Euclidean distance, IMED takes into
account the inter-relationships of the pixels. Therefore it is robust to small perturbation of
images. In this paper, the author argues that that IMED is the only intuitively reasonable
Euclidean distance for images. IMED is then applied to image recognition. The key
advantage of this distance measure is that it can be embedded in most image
classification techniques such as SVM, LDA and PCA. The embedding is rather efficient
by involving a transformation referred to as Standardizing Transform (ST). We show that
ST is a transform domain smoothing. Using the Face Recognition Technology (FERET)
database and two state-of-the-art face identification algorithms, the author justifies that a
consistent performance improvement of the algorithms embedded with the new metric
over their original versions.
64
Zhen Leiy, Shengcai Liaoz, Anil K. Jain, and Stan Z. Liy in their paper “Coupled
Discriminant Analysis for Heterogeneous Face Recognition” describes that coupled space
learning is an effective framework for heterogeneous face recognition [53]. In this paper,
the author proposed a novel coupled discriminant analysis method to improve the
heterogeneous face recognition performance. In this paper the author describes the two
main advantages of the proposed method. First, all samples from different modalities are
used to represent the coupled projections, so that sufficient discriminative information
could be extracted. Second, the locality information in kernel space is incorporated into
the coupled discriminant analysis as a constraint to improve the generalization ability. In
particular, two implementations of locality constraint in kernel space (LCKS) based
coupled discriminant analysis methods, namely LCKS- coupled discriminant analysis
(LCKS-CDA) and LCKS- coupled spectral regression (LCKS-CSR) are presented.
Extensive experiments on three cases of heterogeneous face matching (high vs. low
image resolution, digital photo vs. video image and visible light vs. near infrared)
validate the efficacy of the proposed method.
Farrukh Sayeed, M. Hanmandlu, and A.Q. Ansari in their paper “Face Recognition using
Segmental Euclidean Distance” made an attempt to detect the face using the combination
of integral image along with the cascade structured classifier which is built using Adaboost learning algorithm [54]. The detected faces are then passed through a filtering
process for discarding the non face regions. They are individually split up into five
segments consisting of forehead, eyes, nose, mouth and chin. Each segment is considered
as a separate image and Eigen face also called principal component analysis (PCA)
features of each segment is computed. The faces having a slight pose are also aligned for
proper segmentation. The test image is also segmented similarly and its PCA features are
found. The segmental Euclidean distance classifier is used for matching the test image
with the stored one. The success rate comes out to be 88 per cent on the CG database
created from the databases of California Institute and Georgia Institute. However the
performance of this approach on ORL database with the same features is only 70 per
cent. For the sake of comparison, discrete cosine transforms (DCT) and fuzzy features are
tried on CG and ORL databases but using a well known classifier, support vector
machine (SVM). Results of recognition rate with DCT features on SVM classifier are
65
increased by 3 per cent over those due to PCA features and Euclidean distance classifier
on the CG database. The author justifies the results of recognition are improved up to 96
per cent with fuzzy features on ORL database with SVM.
Shalini Gupta, Mia K. Markey, J. K. Aggarwal, Alan C. Bovik in their paper “Three
dimensional face recognition based on geodesic and Euclidean distances” proposed a
novel method to improve the performance of existing three dimensional (3D) human face
recognition algorithms that employ Euclidean distances between facial fiducially points
as features [55]. In this paper, the author further investigated a novel 3D face recognition
algorithm that employs geodesic and Euclidean distances between facial fiducial points.
The author demonstrates that this algorithm was robust enough to changes in facial
expression. Geodesic and Euclidean distances were calculated between pairs of 25 facial
fiducial points. For the proposed algorithm, geodesic distances and “global curvature”
characteristics, defined as the ratio of geodesic to Euclidean distance between a pairs of
points, were employed as features. The most discriminatory features were selected using
stepwise linear discriminant analysis (LDA). These were projected onto 11 LDA
directions, and face models were matched using the Euclidean distance metric. With a
gallery set containing one image each of 105 subjects and a probe set containing 663
images of the same subjects, the algorithm produced EER=1.4% and a rank 1
RR=98.64%. It performed significantly better than the existing algorithms that were
based on the principal component analysis and LDA applied to face range images. This
method’s verification performance for expressive faces was also significantly better than
an algorithm that employed Euclidean distances between facial fiducial points as features.
Haitao Zhao in their paper, “A Novel Incremental Principal Component Analysis and its
Application for Face Recognition” described the major limitations of existing IPCA
methods is that there is no guarantee on the approximation error [56]. In view of this
limitation, this paper proposed a new IPCA method based on the idea of a singular value
decomposition (SVD) updating algorithm, namely an SVD updating-based IPCA
(SVDU-IPCA) algorithm. In the proposed SVDU-IPCA algorithm, Author has
mathematically proved that the approximation error is bounded. A complexity analysis on
the proposed method is also presented. Another characteristic of the proposed SVDU66
IPCA algorithm is that it can be easily extended to a kernel version. The proposed
method has been evaluated using available public databases, namely FERET, AR, and
Yale B, and applied to existing face-recognition algorithms.
Ayan Chakrabarti in their paper, “Super-Resolution of Face Images Using Kernel PCABased Prior” presented a learning-based method to super-resolve face images using a
kernel principal component analysis-based prior model [57]. A prior probability is
formulated based on the energy lying outside the span of principal components identified
in a higher-dimensional feature space. This is used to regularize the reconstruction of the
high-resolution image.
Braun in his paper, “Machine Learning and Recognition of Faces” described a learning
machine has been tested in a closed-loop system in which a teaching machine performs
the function of error detector [58]. The learning curve for a set of ten photographs of
faces is presented. All the faces are recognized correctly after 250 presentations.
Takeo Kanade, Jeffrey F. Cohn and Yingli Tian in their paper “Comprehensive Database
for Facial Expression Analysis” described that within the past decade, a significant effort
has occurred in developing methods of facial expression analysis [59]. Because most
investigators have used relatively limited data sets, the generalizability of these various
methods remains unknown. In this paper, the author described the problem space for
facial expression analysis, which included the level of description, transitions among
expression, eliciting conditions, reliability and validity of training and test data,
individual differences in subjects, head orientation and scene complexity, image
characteristics, and relation to non-verbal behavior. The author then present the CMUPittsburgh AU-Coded Face Expression Image Database, which currently includes 2105
digitized image sequences from 182 adult subjects of varying ethnicity, performing
multiple tokens of most primary FACS action units. This database is the most
comprehensive test-bed to date for comparative studies of facial expression analysis.
Gaurav Mittal in his paper, “Robust Preprocessing Algorithm for Face Recognition”
discussed that the face recognition includes enhancement and segmentation of face
image, detection of face boundary and facial features, matching of extracted features
67
against the features in a database, and finally recognition of the face [60]. The face
detection algorithms are based on either gray level template matching or computation of
geometric relationships among facial features. Though a number of algorithms are
devised for face recognition, the technology is not matured enough to recognize the face
of a person with and without beard to be the same. The author’s research proposed a
robust algorithm for preprocessing the face image with beard for face recognition.
3.1 Security and Privacy Issues in Biometric Authentication
User authentication is fundamental to the protection of information systems. Biometric
authentication comes in play to deal with these difficulties with traditional password
systems. Potentially, biometric systems can be employed in all applications that need
authentication mechanism and so in all applications that today use passwords, PINs, ID
cards or the like [61].
The security concerns related to biometric authentication can be organized into two
categories: concern about the theoretical basis of biometrics and vulnerability of
biometric authentication system as discussed below:
3.1.1 Biometric System Concerns
An evident class of biometric authentication vulnerabilities is those faced by the system
user, which impacts user’s privacy and may show the way to identify theft or system
compromise [62].
Biometrics is not secret: Technology is willingly available to image faces, fingerprints,
and irises and make recording of voice or signature – without subject’s approval or
understanding. From this point of view, Biometrics is not secret. On the other hand, from
a cryptography or privacy’s point of view, biometric data are often considered to be
private or secret.
Biometrics cannot be revoked: A biometric feature is permanently associated with a
user and a compromised biometric model will compromise all applications that make use
of that biometric. But the user cannot change his/her fingerprint or retinal patterns.
68
Biometrics has secondary uses: If a user uses the same biometric feature in multiple
applications, then the user can be tracked if the organization shares biometric data.
Because of the amount and type of information that are also collected along with the
biometrics record, many end users distinguish biometric authentication as an intrusive
process and express concerns about how the information will be used beyond the original
purpose [63].
How reliably unique the biometrics are? Many people like to think of biometrics as
‘unique’, but they are not, at least not with the level of data we can measure [64].
How universal the biometric are? Now all biometric traits are truly universal. The
National Institute of Standards and Technology (NIST) reported that it is not possible to
obtain a good quality fingerprint from approximately two percent of population due to
disabilities, cuts, bruises, wounds etc. [65].
Biometric traits are not always invariant: The biometric data acquired from a user
during verification will not be identical to the data used for generating the user’s template
during enrollment [66]. Even under the same equipment and same environmental
conditions biometric data collected from the same user are likely not to be identical.
Biometric traits may vary due to fatigue, sickness etc., for example, a person’s voice may
change if he/she catches cold, children’s face, and gait change as they grow up.
3.1.2 Vulnerability of Biometric Authentication System
The security of biometric authentication depends on the vulnerability of underlying
biometric system. Since biometric systems are implemented on server computers, they
are at risk to all cryptographic, viruses and other attacks which affect modern computer
systems [67].
To better understand security issues concerned with biometric authentication, it should be
useful to study user components of a typical biometric system, communication channel
among the components, and their vulnerabilities as shown in figure 3.1.
Figure 3.1 shows the major functional components of a standard biometric system, where
major steps in the process of authentication are marked as A, B, C, and so on. Usually,
69
each presented sample (B) is acquired by a sensor (C) processed via segmentation and
feature extraction (D) algorithms.
Figure 3.1 Block diagram of a typical biometric authentication system
If available a sample quality assessment (E) algorithm is used to indicate a need to
reacquire the sample. Biometric features are encoded into template, which is stored (H) in
database or any secure hardware. For biometric encryption systems, a code or token is
combined with the biometric feature in the template. During enrollment, biometric
samples are connected to a claimed identity (A), and during successive verification or
identification, samples are compared with enrolled samples using matching algorithm (I),
and an identity decision (J) is made either automatically, or by a human being reviewing
biometric system outputs. Andy Adler [68] points out the security issues at each of these
components or steps in a typical biometric authentication system, which is discussed
below in short.
Sample presentation (B): The attacker may inject the false biometric sample into the
system. Such attacks are mounted to avoid detection or masquerade/fraud as another
person. The later attack is typically called spoofing.
Sensor (C): Noise can appear in the acquired biometric data due to environmental factors
such as lights, sound, humidity etc. as well as defective or not properly maintained
sensors [68]. Attacks on the biometric sensor may weaken or replace the sensor
hardware. In many cases, an attack on the sensor would take the form of repeat. Unlike
control or knowledge based authentication, accuracy of biometric authentication is much
70
dependent on sensor device used. For example, a computer keyboard does not shows the
hardware specific characteristic in the typed text. However, keyboard characteristics used
for biometric, key-stroke-based authentication significantly affects the resulting sampled
signal due to physical properties like attenuation, pressure sensitivity, and others.
Biometric signals are exposed to distortions based on sensor characteristics. The problem
of unavailability of identical sensor may be applicable for applications in large areas, as
well as long term considerations, where specific hardware may be no longer available
after some time.
Segmentation (D): Biometric segmentation extracts the image or signal of interest from
the background, and a failure means the system does not detect the existence of
appropriate biometric feature. Segmentation attacks may be used to escape observation or
to generate denial of service (DoS) attack.
Feature extraction and Quality assessment (E): Knowledge of feature extraction or
quality assessment algorithms can used in the biometric authentication system that may
be exploited to design some special features in presented biometric samples to cause
wrong features to be considered.
Template creation (G): One of the common claims is that, since template creation is a
one-way function, it is impossible or infeasible to regenerate the image or signal from the
templates. However, recent research has been shown that regeneration of biometric
samples from templates to be feasible.
Data storage (H): For biometric authentication the size of reference data may become
very large, and due to natural inconsistency of biometric information, it is impossible to
apply discrete mathematical techniques like cryptographic hashes to secure the reference
data. Vulnerabilities of template storage concern modifying the storage (adding,
modifying, or removing templates), copying template data for secondary usage (identity
theft), or tampering the identity to which the biometric is assigned. “The biometric
dilemma is that although biometrics can initially improve security, as traditional
biometric databases become widespread, compromises will finally destroy the
biometrics’ value and usefulness for security”.
71
Matching (I): A biometric matcher calculates the matching score related to the
probability that the two biometric samples are from the same individual. For multimodal
or biometric fusion systems, excessive score in one biometric modality may override the
authority of other modalities. Besides the concern of finding methods to increase the
overall accuracy of multi-modal authentication systems, for example, with respect to the
degree of correlation between different modalities, question of finding a meaningful set
of modalities. Biometric system matchers which are based on Fisher discriminant
strategies calculate the global thresholds based on the between cross-covariance relation,
which may be modified by enrolling particularly created biometric samples.
Decision (J): Biometric decisions are often reviewed by human operator. Such operators
are well known to be susceptible to fatigue or monotony. One of the goals of the DoS
attacks is to force operators to discard the biometric system, or not to completely trust its
output by causing it to produce sufficiently large number of errors. All biometric
authentication methods are based on some statistical measurement of matching and
threshold, which makes the process subject to false classification error. Biometric
authentication still is based on probability of matching and so cannot give complete
verification about certain authentication. Many biometric authentication systems allow
the administrator to configure a threshold level that determines the false acceptance rate
and the false denial rate.
Maltoni and et al. [69] classify vulnerability of biometric authentication system as
follows:
Circumvention is an attack which gains access to the protected resources by a technical
measure to undermine the biometric system. Such an attack may undermine the
underlying computer systems (overriding matcher decision, or replacing database
templates) or may involve replay of valid data. This threat can be cast as a privacy attack,
where the attacker accesses the data for that he was not authorized to access (e.g.,
accessing the medical records of another user) or, as a subversive attack, where the
72
attacker manipulates the system (e.g., changing those records, submitting bogus insurance
claims, etc.). It is possible to avoid a biometric system using spoofed traits. For example,
it is possible to construct “gummy fingers” using lifted fingerprint impressions and utilize
them to circumvent a fingerprint based authentication system.
Covert acquisition (contamination) is use of biometric information captured from the
legal users to access a system. Examples include spoofing via capture and playback of
voice password, and lifting latent fingerprints to construct a mold. Further, the biometric
data associated with a specific application can be used in another unintended application.
For e.g., using a fingerprint for accessing medical records instead of the intended use of
office door access control.
Collusion and Coercion are biometric system vulnerabilities from the legal system
users. The main difference is that, in collusion the legal user is will perhaps by bribe,
while the coerced user is forced through threat or blackmail.
Denial of Service (DoS) is an attack which prevents the legal use of the biometric
system. This can take the form of slowing or stopping the system via overload of requests
or by degrading the performance of the system.
Repudiation is the case where the attacker denies to accesses the system. A corrupt user
may deny his/her actions by claiming that his/her biometric data were stolen. For a
biometric authentication system, an online authentication server that processes access
requests via retrieving templates from a database and performing matching with the
transferred biometric data can be sending with many bogus access requests, to a point
where the server’s computational resources cannot handle valid requests to any more.
73
Chapter 4
Conceptual models for face recognition
A complete face recognition system is composed of five parts: image acquisition,
detection of a face, normalization, feature extraction and matching. The image acquisition
step captures the face images. Face detection step is used to find the position of faces in a
given image. Output of the detection can be location of face region as a whole, and
location of face region with facial features (i.e. eyes, mouth, eyebrow, nose etc.). Face
normalization step is used to eliminate non-face objects in the images to save the
computation time. Feature extraction is used for extracting the information that can be
used to distinguished different subjects, creating a template that represents the most
discriminated features of the face, typically it uses texture information. The
corresponding matching stage calculates the distance between image codes and decides
whether it is a match or recognizes the submitted image from the subjects in the data set.
The process of face recognition is usually divided into five steps:
1: Image acquisition: image acquisition step captures the images of a face including the
front profile, left profile or the right profile of the same person. Image acquisition is
usually done using the CCD camera.
2: Detection of a face: this step is used to find the position of faces in a given image.
Face detection performs locating and extracting face image operations for face
recognition system. Output of the detection can be the location of face region as a whole,
and location of face region with facial features (i.e. eyes, mouth, eyebrow, nose etc.). The
goal of face detection is to determine whether or not there are any faces in the image and,
if present, return the image location and extent of each face.
3: Normalization: create the dimensionality consistent representation of the face image
and also eliminate the non- face objects in the images to save the computation time.
74
4: Feature extraction: this stage is used for extracting the information that can be used
to distinguished different subjects, creating a template that represents the most
discriminant features of the face. Some filtering operations are applied to extract feature
candidates and steps are listed below:

Laplacian of Gaussian Filter on Red channel of candidate.

Contrast correction to improve visibility of filter result.

Average filtering to eliminate small noises.
5: Matching: the features vectors are compared using a similarity measure.
Finally, a decision of high confidence level is made to identify whether the user is an
authentic or not. But here, the main problem is that we are not able to distinguish between
a dummy/fake image and a real image. An illustration of the steps for the face recognition
is shown in figure 4.1.
Figure 4.1 Techniques used in face recognition
75
4.1 Spoofing in Biometric System
Attacker can attack any of the above mentioned stage for unauthorized access to the
system; some of those attacks are like biometric sensor attack which happens at the
beginning of the process. In this type of attack fake biometric data like artificial finger, a
mask over a face or a contact lens on an eye etc. may be presented to the sensor and puts
previously stored genuine biometric information into proper place in the processing
chain. Following are some attacks that can be happen on biometric system [70].

Fake biometric data at sensor: In this type of attack fake biometric trait like
fake gummy finger, an iris printout or a face mask is used for unauthorized
access.

Resubmitting previously stored digitized biometric signals (replay attack): A
digitized biometric signal, which has been previously enrolled and stored in the
database, is again send to the system.

Tampering biometric feature: In this attack extracted features of given
biometric trait changed so that it can be accepted by matcher.

Attack
on
enrollment:
An
enrollment
database
used
in
the
verification/identification process can also be altered like by providing fake
biometric traits.
4.1.1 Spoofing in Face Recognition System
Generally speaking, there are three ways to spoof face recognition:
a. Photograph of a valid user
b. Video of a valid user
c. 3D model of a valid user
Photo attack is the cheapest and easiest spoofing approach, since one's facial image is
usually very easily available for the public, for example, downloaded from the web,
captured unknowingly by a camera. The imposter can rotate, shift and bend the photo
before the camera like a live person to fool the authentication system. It is still a
76
challenging task to detect whether an input face image is from a live person or from a
photograph.
Video spoofing is another big threat to face recognition systems, because it is very
similar to live face and can be shot in front of legal user’s face by a needle camera. It has
many physiological clues that photo does not have, such as head movement, facial
expression.
3D model has 3D information of face, however, it is rigid and lack of physiological
information. It is also not very easy to be realistic with live person who the 3D model
imitates. So photo and video are most common spoofing ways to attack face recognition
system.
In general, human is able to distinguish a live face and a photograph without any effort,
since human can very easily recognize many physiological clues of liveness, for example,
facial expression variation, mouth movement, head rotation, eye change. However, the
tasks of computing these clues are often complicated for computer, even impossible for
some clues under the unconstrained environment. These are some common attack mainly
happened on biometric system. Some of attack can be resolved by maintaining some
security like securing database from alteration. But some attacks like fake biometric data
to the sensor cannot be covered by strong security. So, for this type of attacks Liveness
detection technique is used to ensure that the given biometric sample originates from a
living person and is not artificial.
4.2 Face Image Acquisition
The acquisition module captures a series of ocular images; uses a scheme to evaluate
image quality; and selects one image with sufficient facial features information, which
then undergoes additional processing. One of the major challenges of face recognition
system is to capture a high quality image of the overall person. Image acquisition is
considered the most critical step in the project since all subsequent stages highly depends
on the image quality. In order to accomplish this, we use a CCD camera. We set the
resolution to 640×480, the type of the image to jpeg, and the mode to white or black for
77
greater details. Block diagram of face acquisition using CCD camera is shown below in
figure 4.2.
Figure 4.2 Block Diagram of image acquisition using CCD camera
The camera is situated normally between half a meters to one meter from the subject. The
CCD-cameras job is to take the image from the optical system and convert it into the
electronic data. CCD camera then transfers the value of the different pixels out of the
CCD chip. Read out the voltages from the CCD-chip. Thereafter, the signals of each data
are amplified and sent to an ADC (Analog to Digital Converter).
4.3
Proposed
Face
Recognition
Algorithm
for
Image
Acquisition
4.3.1 Architecture of the proposed approach
A block diagram of the proposed face recognition system is shown in figure 4.3.
The architecture of the proposed approach is divided into two steps:
1. Enrollment: User enrollment is a process that is responsible for registering individuals
in the biometric system storage. During the enrollment process, the biometric
characteristics of a person are first captured by a biometric scanner to produce a sample.
Some systems collect multiple samples of a user and then either select the best image or
fuse multiple images or create a composite template. Then the Euclidean distance test is
78
performed to check the Liveness of a person. If the person is live then the enrollment
process takes the enrollment template and stores it in the system storage along with the
demographic information about the user.
2. Authentication: In this process, the subject does not explicitly claim an identity and
the system compares the feature set against the templates of all the subjects in the system
storage; the output is a candidate list that may be empty or contain one or more identifiers
of matching enrollment templates.
Figure 4.3 Architecture of the proposed approach
The modules of the proposed approach are explained as follows:
1. Image Acquisition: An important and complex step of face recognition is image
acquisition of a very high quality of a person. It is difficult to acquire clear images using
the standard CCD camera with ordinary lighting. In this module, a biometric camera is
79
used to capture the user’s face images. Due to distance or illumination changes,
recognition process may produce different recognition rates. A study on illumination
effects on face recognition showed that lighting the face bottom up makes face
recognition a hard task [71]. Also, a study on the direction of illumination [72] showed
the importance of top lighting; it is easier for humans to recognize faces illuminated from
top to bottom than the faces illuminated from bottom to top. May be, the same genuine
user’s facial features may slightly vary at sunlight and twilight. However, the system
confines the Euclidean distance of the same candidate features to the face discriminator
phase. The image acquisition phase should consider three main aspects, that is to say, the
lighting system, the positioning system, and the physical capturing system. The face
recognition system can work both in outdoor and indoor conditions without any hot spot
of lighting intensities.
2. Euclidean Distance Test: This module is used for checking the person’s liveness and
is further described in the section 4.3.2.
3. Face Normalization: Create the dimensionality consistent representation of the face
image and also eliminate the non- face objects in the images to save the computation
time.
4. Feature Extraction: This stage is used for extracting the information that can be used
to distinguish different subjects, creating a template that represents the most
discriminated features of the face. This is done with the help of the PCA technique.
5. Matching: This stage takes a feature set and an enrollment template as inputs and
computes the similarity between them in terms of a matching score. The matching score
is compared to a system threshold to make the final decision; if the match score is higher
than the threshold, the person is recognized, otherwise not.
4.3.2 Euclidean Distance Test
Biometric features may be fake and illegally used. This is a crucial weakness of the
biometric system. This section aims to ensure that an input image actually originates from
a user instead of face photographs or any other artificial sources. In the proposed work,
80
Euclidean distance test is suggested to overcome this problem. So, here we use the
concept of Euclidean distance.
4.3.2.1 Euclidean distance
Euclidean distance between two points in p-dimensional space is a geometrically shortest
distance on the straight line passing through both the points [73].
For a distance between two p-dimensional features x=(x1, x2,………,xp) and y=(y1,
y2,……yp) the Euclidean distance is defined as follows:
d (x, y)=[ ∑𝑝𝑖=1(xi - yi)²]½
So, if the images are of the same person then their Euclidean distance will be minimum
otherwise they come from the different sources.
4.3.2.2 Algorithm of the proposed approach
The algorithm of this method is elaborated as follows:
Step 1: Capture the same person’s face images under the three different profiles- left,
right and front (at least one from each profile) in any random order. For example, LRF or
it may be FLR etc.
Step 2: Measure the Euclidean distance from the captured face images. If these values
are dissimilar then the image is actually coming from a real source (user), otherwise
artificial sources may have been used.
Euclidean distance is given by:
d (x, y)={∑𝑛𝑖=1(xi - yi)²}½
(1)
Using eq’n (1), we can calculate the difference between the Euclidean distances of the
two different images of the same person.
Td= ∑𝑛𝑖=1 | d (xi, yi) – d (xi+1,yi+1) |
EDT= ∫
𝑇𝑟𝑢𝑒,
𝐹𝑎𝑙𝑠𝑒,
(2)
𝑖𝑓 𝑇𝑑 ≠ 0
𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒
(3)
81
Here, EDT is the Euclidean distance test parameter, n is the no of face images taken in
any of the following order i.e., it may be LFR, LRF, FLR, RLF, RFL, FRL whereas L
denotes the left profile and R denotes the right profile and F denotes the front profile of a
person. d (xi, yi) and d(xi+1,yi+1) are also the Euclidean distance of the person under
different profiles. Td is the difference between the Euclidean distances of the two images.
If the Td is not equal to zero that means the person image is real and not coming from a
fake biometric system or any other artificial sources and responding to Euclidean distance
test.
4.3.2.3 Comparison
The proposed work presents a technique for liveness detection which uses the concept of
Euclidean distance. This technique ensures the detection of fake/dummy images and also
verifies that the person is actually alive or not. But the traditional method does not use the
concept of Euclidean distance. Therefore, this method is not able to check that the person
is actually alive or not. May be the person’s image is fake or generated from artificial
sources. The main advantage of the proposed work is that if the person is found fake then
it is detected in the Euclidean distance test module and the further computations are not
done. But in the traditional method, there is no provision for liveness detection and thus it
was not able to detect the liveness of the person.
82
4.4 Summary
This chapter has presented a face recognition system which was tested using two
databases of grayscale face images in order to verify the claimed performance of the face
recognition technology.
The system is to be composed of a number of subsystems, which corresponds to each
stage of a face recognition system. These stages mainly includes image acquisition - to
acquire the image of a face using a biometric camera, face detection - determine whether
or not there are any faces in the image and, if present, return the image location and
extent of each face, normalization – creating a dimensionally consistent representation of
the face region and feature extraction – creating a template containing only the most
discriminated features of the face. The input to the system will be a face image and the
output will be a face template. Spoofing in the biometric system was also discussed in
this chapter. Then, an automatic acquisition algorithm for face image using Euclidean
distances was presented which checks the liveness of the person whether the person is
live or not and we can also check whether the person is fake or real. Finally, a
comparison is also made between the proposed approach and the existing approach to
show the advantages of the proposed approach.
83
Chapter 5
Facial Expression Recognition
5.1 Face Detection/Localization:
The ultimate goal of the face detection/localization is to find an object in an image whose
shape resembles the shape of a face. Many algorithms have been proposed for face
localization and detection, which are based on using shape [74], color information [75],
and also on motion [76]. These methods can be classified into the following four
categories [77]:
A. Knowledge-based methods: These rule-based methods encode human knowledge of
what constitutes a typical face. Usually, the rules capture the relationships between facial
features. These methods were mainly designed for face detection/localization.
B. Feature invariant approaches: These algorithms aim to find structural features that
exist even when the pose, viewpoint, or lighting conditions vary, and then use these to
locate faces. These methods are designed mainly for face detection/localization.
C. Template matching methods: Several standard patterns of a face are stored to
describe the face as a whole or the facial features separately. The correlations between an
input image and the stored patterns are computed for detection. These methods have been
used for both face localization and detection.
D. Appearance-based methods: In contrast to template matching, the models (or
templates) are learned from a set of training images, which should capture the
representative variability of facial appearance. These learned models are then used for
detection. These methods are designed mainly for face detection.
84
5.1.1 Color Models for Skin Color Classification:
The study on skin color classification has gained increasing attention in recent years due
to the active research in content-based image representation. For instance, the ability to
locate image object as a face can be exploited for image coding, editing, indexing or
other user interactivity purposes. Moreover, face localization also provides a good
stepping stone in facial expression studies. It would be fair to say that the most popular
algorithm to face localization is the use of color information, whereby estimating areas
with skin color is often the first vital step of such strategy. Hence, skin color
classification has become an important task. There are four color models namely RGB,
YCbCr, YES and HSI. The same process is applied to each color model to obtain at most
four skin regions. Next, we apply a polling strategy to determine the final region where
the face is located. There are some advantages and disadvantages in each of the
algorithms. Therefore, we have taken the combination of all the four color models to find
the skin region and then from the skin region facial features have been extracted to get
the face from the skin region.
a) RGB Color Space
The RGB color space consists of the three additive primaries: red, green and blue.
Spectral components of these colors combine additively to produce a resultant color. The
RGB model is represented by a 3-dimensional cube with red green and blue at the corners
on each axis (Figure 4.4). Black is at the origin. White is at the opposite end of the cube.
The gray scale follows the line from black to white. In a 24-bit color graphics system
with 8 bits per color channel, red is (255, 0, and 0). On the color cube, it is (1, 0, and 0).
The RGB model simplifies the design of computer graphics systems but is not ideal for
all the applications. The red, green and blue color components are highly correlated. This
makes it difficult to execute some image processing algorithms. Many processing
techniques, such as histogram equalization, work on the intensity component of an image
only. RGB color space is shown in figure 5.1 below:
85
Figure 5.1 RGB color space
b) YCbCr Color Space
YCbCr color space has been defined in response to increasing demands for digital
algorithms in handling video information, and has since become a widely used model in a
digital video. It belongs to the family of television transmission color spaces. The family
includes others such as YUV and YIQ. YCbCr is a digital color system, while YUV and
YIQ are analog spaces for the respective PAL and NTSC systems. These color spaces
separate RGB (Red-Green-Blue) into luminance and chrominance information and are
useful in compression applications however the specification of colors is somewhat
unintuitive. The Recommendation 601 specifies 8 bit (i.e. 0 to 255) coding of YCbCr,
whereby the luminance component Y has an excursion of 219 and an offset of +16.
Figure 5.2 The RGB color space within the YCbCr color space. Colors outside the RGB
cube are not valid.
This coding places the black color at code 16 and places the white color at code 235.
Therefore, by doing so, it reserves the extremes of the range for signal processing foot
86
room and headroom. On the other hand, the chrominance components Cb and Cr have
excursions of +112 and offset of +128, producing a range from 16 to 240 inclusively.
There are some points within the YCbCr color cube that cannot be represented in the
corresponding RGB domain. This causes some difficulty in determining how to correctly
interpret and display some YCbCr signals. To figure out the relationship between the
RGB and YCbCr color space cubes, Fig. 5.2 shows the RGB color space contained
within the YCbCr color space where the color outside the RGB cube are not valid as
mentioned.
c) HSI Color Space
Since hue, saturation and intensity are three properties used to describe color, it seems
logical that there be a corresponding color model, HSI. When using the HSI color space,
you don't need to know what percentage of blue or green is required to produce a color.
You simply adjust the hue to get the color you wish. To change a deep red to pink, adjust
the saturation. To make it darker or lighter, alter the intensity. HIS color space is shown
in figure 5.3 below:
Figure 5.3 HSI color space
87
Many applications use the HSI color model. Machine vision uses HSI color model in
identifying the color of different objects. Image processing applications such as
histogram operations, intensity transformations and convolutions operate only on an
intensity image. These operations are performed with much ease on an image in the HIS
color model.
d) YES color Space
The YES color space is defined as the Y channel is the luminance component, and E, S
channels are the chrominance. The YES color space constitutes a linear transformation
from RGB, free of singularities and provides some computational efficiency. The E and S
channels can be computed from RGB by shifting bits rather than multiplying. However, it
is generally agreed that there does not exist a single color space that is good for all
images [78]. The E and S channels provide a suitable space for recognition of the classes
under consideration based on color in classification applications. The Y channel will be
mainly utilized to model the texture information.
5.1.2 Algorithm for detecting a face
There are four steps for detecting a face in a single image using the color model as
described below:
1. Segmentation: An important task in several vision applications is the image pixel
classification in a discrete number of classes. The objective of segmentation is providing
an effective and real time classification. The first step for segmentation is obtaining a set
of pixel values, which corresponds to the color skin in the images. The set is obtained
manually, selecting a small region directly from a frame where the face is located. Then
this set to define the maximum and minimum skin pixel values for each channel of the
color space. This vector will be employed in subsequent classifications due the skin color
constitutes a regular cluster in the color space. Next, we are classifying the pixels by
applying the thresholding operation to the frame. Thresholding involves the use of the
88
vector values in the corresponding three-dimensional color space. Every vector values
can be considered as a class.
2. Regions: After segmentation is performed, then connect all pixels to produce regular
regions or blobs. This process is exhaustive and can affect the real time performance. The
threshold values have a large influence on the segmentation results. A small threshold
value leads to a large number of small regions while with a large threshold value few
large regions are calculated. Steps of a face detection algorithm are shown in figure 5.4
below:
Figure 5.4 Steps of a face detection algorithm
3. Localization: In this stage firstly detect and locate the face in the frame. Then search
in all labeled regions for the one that satisfies a specific area (an approximately number
of pixels to form a face) and size (high and width considering the region is rectangular).
After the region of interest is located, we obtain the position of the region center.
Dimension and center position of the face region is computed for each color model. In
89
this way, now we get the many regions as color spaces where the detection was
successful for the polling process.
4. Polling: Finally, validate the face detection and position in the frame by polling.
Simply, then examine all the regions, and if there is a region common to at least three
colors spaces, then considered such region as the detected face, in the other way, if the
mentioned condition is not satisfied, the region is discarded and the process is performed
to the next region or the next frame.
5.2 Facial Expression Recognition
Since last three decades of face recognition technology, there exists many commercially
available systems to identify human faces, however face recognition is still an
outstanding challenging problem. This challenge can be attributed to (i) large intrasubject variations such as pose, illumination, expression, and aging is commonly
encountered in face recognition and (ii) large inter-user similarity. Meanwhile this
technology has extended its role to Human-Computer-Interaction (HCI) and HumanRobot- Interaction (HRI). Person identity is one of the key tasks while interacting with
the robots, exploiting the un-attentional system security and authentication of the human
interacting with the system. This problem has been addressed in various scenarios by
researchers resulting in commercially available face recognition systems [79, 80].
However, other higher level applications like facial expression recognition and face
tracking still remain outstanding problem along with person identity. This gives rise to an
idea for generating a framework suitable for solving these issues together.
5.2.1 Facial Action Coding System
A fully automatic facial action coding system was developed using action units which is a
user independent fully automatic system for real time recognition of facial actions.
Ekman and Friesen (1976, 1978) were pioneers in the development of measurement
systems for facial expression. Their system, known as the Facial Action Coding System
90
or FACS, was developed based on a discrete emotions theoretical perspective and is
designed to measure specific facial muscle movements which is shown in figure 4.8. A
second system, EMFACS, is an abbreviated version of FACS that assesses only those
muscle movements believed to be associated with emotional expressions. In developing
these systems, Ekman importantly distinguishes between two different types of
judgments: those made about behavior (measuring sign vehicles) and those that make
inferences about behavior (message judgments). Ekman has argued that measuring
specific facial muscle movements (referred to as action units in FACS) is a descriptive
analysis of behavior, whereas measuring facial expressions such as anger or happiness is
an inferential process whereby assumptions about underlying psychological states are
made.
Figure 5.5 Facial action coding system
In this system following steps are considered to extract the expressions1. Firstly, automatic system is used to detect the face and then the Pain Expression from
an image.
2. Then Gabor filter is applied on the extracted face and Pain Expression. This filter is
used in image processing for edge detection.
3. The selected edges are passed through adaboost (adaptive boosting) to select the
desired features.
4. Then Eigen face (support vector machine) is used to classify the features and arrange
them into action units.
5. The action units are then used to classify different features.
91
The Facial Action Coding System is a human-observer based system designed to detect
subtle changes in facial features. Viewing videotaped facial behavior in slow motion,
trained observers can manually FACS code all possible facial displays, which are referred
to as action units (AU) and may occur individually or in combinations. FACS consists of
44 action units. Thirty are anatomically related to contraction of a specific set of facial
muscles (Table 4.1) [81]. The anatomic basis of the remaining 14 is unspecified (Table
4.2). These 14 are referred to in FACS as miscellaneous actions. Many action units may
be coded as symmetrical or asymmetrical. For action units that vary in intensity, a 5point ordinal scale is used to measure the degree of muscle contraction.
Table 5.1 FACS Action Units
92
FACS itself is purely descriptive and includes no inferential labels. By converting FACS
codes to EMFACS or similar systems, face images may be coded for emotion-specified
expressions (e.g., joy or anger) as well as for more molar categories of positive or
negative emotion [82].
Table 5.2 Miscellaneous Actions Units
5.2.2 Proposed Work
In our proposed work, Face and Facial Expression detection is presented by combining
the Eigen value and Eigen face methods which perform almost as fast as the Eigen face
method but with a significant improved speed. Eigenvectors and Eigen values are
dependent on the concept of orthogonal linear transformation. An Eigenvector is
basically a non-zero vector. The dominant Eigenvector of a matrix is the one
corresponding to the largest Eigen value of that matrix. This dominant Eigenvector is
important for many real world applications.
5.2.2.1 Algorithm of the proposed approach
The proposed work is divided into 3 main stages:
1. Image Preprocessing
2. Training the Database
3. Matching
93
The image preprocessing work mainly consists of four different modules, namely,
histogram equalization, edge detection, thinning, and token generation. The face image is
taken as an input and tokens are produced as output. It includes the conversion of image
to the normalized image as well as to extract the features from the image. The filtration
process includes the adjustment of brightness, contrast, low pass, high pass filtration etc.
The filtration will be done in two phases one for face and other for facial expression. To
enhance the image quality, histogram equalization has been performed.
Figure 5.6 Flow Chart of the proposed approach
It is then followed by the edge detection process. Edge detection plays an important role
in finding out the tokens. It is a terminology in image processing, particularly in the areas
of feature detection and feature extraction, which aims at identifying points in a digital
image at which the image brightness changes sharply or more formally have
discontinuities. After the edge detection, thinning has to be performed. It is applied to
94
reduce the width of an edge to single line. After the thinning process, tokens have been
generated. Tokens divide a data set in to the smallest unit of information used for
subsequent processing. Flow chart of the proposed approach is shown in figure 5.6.
Another task is performed on the available database. The database will be trained by
using Neural Network. The training module stores the token information that comes from
the image pre-processing module. It can be done using any method. Here, neural network
is used. The power of neural networks is realized when a pattern of tokens, during testing,
is given as an input and it identifies the matching pattern it has already learned during
training. The training process will be done only once and after training a feature analysis
based database will be created. The actual match will be performed on this Eigen Face
trained dataset. Again we have to generate two datasets one for the face itself and the
other for the Facial Expression.
Third and the final step are to perform the matching with the database. The match will be
performed on both the face image and the Facial Expression image separately. If both
matches give the results better than the expected value then the person will be
recognized.
5.2.2.2 Eigen Face Method
The motivation behind Eigen faces is that it reduces the dimensionality of the training set,
leaving only those features that are critical for face recognition.
Definition
1. The Eigen faces method looks at the face as a whole.
2. In this method, a collection of face images is used to generate a 2-D gray-scale image
to produce the biometric template.
3. Here, first the face images are processed by the face detector. Then we calculate the
Eigen faces from the training set, keeping only the highest Eigen values.
4. Finally we calculate the corresponding location in weight space for each known
individual, by projecting their face images onto the “face space”.
95
In mathematical terms, the objective is to find the principal components of the
distribution of faces, or the eigenvectors of the covariance matrix of the set of face
images. These eigenvectors can be thought of as a set of features which together
characterize the variation between face images. Each image location contributes more or
less to each eigenvector, so that we can display the eigenvector as a sort of ghostly face
called an Eigen face.
Each face image in the training set can be represented exactly in terms of a linear
combination of the Eigen faces. The number of possible Eigen faces is equal to the
number of face images in the training set. However, the faces can also be approximated
using only the “best” Eigen faces-those that have the largest Eigen values, and which
therefore account for the most variance within the set of face images. The primary reason
for using fewer Eigen faces is computational efficiency. The most meaningful M Eigen
faces span an M-dimensional subspace -“face space”- of all possible images. The Eigen
faces are essentially the basis vectors of the Eigen face decomposition. The idea of using
Eigen faces was motivated by a technique for efficiently representing pictures of faces
using principal component analysis.
The Eigen faces approach for face recognition involves the following initialization
operations:
1. Acquire a set of training images.
2. Calculate the Eigen faces from the training set, keeping only the best M images with
the highest Eigen values. These M images define the “face space”. As new faces are
experienced, the Eigen faces can be updated.
3. Calculate the corresponding distribution in M-dimensional weight space for each
known individual (training image), by projecting their face images onto the face space.
Having initialized the system, the following steps are used to recognize new face images:
1. Given an image to be recognized, calculate a set of weights of the M Eigen faces by
projecting it onto each of the Eigen faces.
96
2. Determine if the image is a face at all by checking to see if the image is sufficiently
close to the face space. If it is a face, classify the weight pattern as either a known person
or as unknown.
3. (Optional) Update the Eigen faces and/or weight patterns.
4. (Optional) Calculate the characteristic weight pattern of the new face image, and
incorporate into the known faces.
5.2.2.3 Calculating Eigen faces
Let a face image (x,y) be a two-dimensional N by N array of intensity values. An image
may also be considered as a vector of dimension N 2 , so that a typical image of size 256
by 256 becomes a vector of dimension 65,536, or equivalently, a point in 65,536dimensional space. An ensemble of images, then, maps to a collection of points in this
huge space.
Images of faces, being similar in overall configuration, will not be randomly distributed
in this huge image space and thus can be described by a relatively low dimensional
subspace. The main idea of the principal component analysis is to find the vector that best
account for the distribution of face images within the entire image space. These vectors
define the subspace of face images, which we call “face space”. Each vector is of length
N 2 , describes an N by N image, and is a linear combination of the original face images.
Because these vectors are the eigenvectors of the covariance matrix corresponding to the
original face images, and because they are face-like in appearance, they are referred to as
“Eigen faces”.
1. The first step is to obtain a set X with M face images. Each image is transformed into a
vector of size N and placed into the set. Therefore, we get a single matrix X of M×N.
2. Read the Input Image I.
3. Normalized the Image Set X as to get it in required format.
4. Calculate the Mean Image from Set S as follows:
µx=1/N ∑𝑁
𝑖=1 𝑋 [m, n]
Where µx is the mean image of the complete matrix set X and m, n are the indices of the
matrix and m=1, 2……..M and n=1, 2……..N.
97
5. Find the Difference between the Mean Image and the Input Image I.
Difference=Mean (µx)-Input Image
6. Calculate the Covariance Matrix for the image.
Cov (x) = ∑𝑁
𝑖=1[µx-Input image]² / N-1
7. Calculate the Eigen Values and Eigen vectors for the image.
8. Calculate the Image Set with lower Euclidean distance.
9. Repeat the Operation on this image set with lower Euclidean distance value to get the
image having the expression closer to the defined image.
10. The image with the lowest Euclidean distance in expression images will be
represented as the resultant expression image.
5.3 Summary
In this chapter, firstly we discussed the existing face detection methods which includes
like Knowledge-based methods, Feature invariant approaches, Template matching
methods, Appearance-based methods. We also discussed the color space for skin color
classification such as RGB color space, YCbCr color space; YES color space, HIS color
space and the algorithm to detect the face region in an image.
Next, a system to recognize the facial expressions was also presented in which the Eigen
face and Principal Component Analysis (PCA) method was used together to find the
principal components of the distribution of faces, or the Eigen vectors of the covariance
matrix of the set of face images. Then the Euclidean distance metric was used as a
decision classifier to recognize the expression of a face like sadness, happy, pain, disgust,
fear etc. And finally, matching is performed on both the face image and the facial
expression image separately. If both the matches give the results better than the expected
value then the person will be recognized.
98
Chapter 6
Face Recognition Applications
There are numerous applications for the face recognition technology. Many applications
for face recognition have been envisaged. Commercial applications have so far only
scratched the surface of the potential. The further sections describe the applications
where the technology is currently being deployed and where it shows some future
potential and these applications are divided into two sections as described below:
6.1 Government Use applications
6.1.1 Law Enforcement: Minimizing victim trauma by narrowing mug-shot searches,
verifying identify for court records, and comparing school surveillance camera images to
known child molesters.
a) Identification Systems
Two US States (Massachusetts and Connecticut [83]) are testing face recognition for the
policing of Welfare benefits. This is an identification task, where any new applicant
being enrolled must be compared against the entire database of previously enrolled
claimants, to ensure that they are not claiming under more than one identity.
Unfortunately face recognition is not currently able to reliably identify one person among
the millions enrolled in a single state’s database, so demographic information such as zip
code, age, name, gender etc. are used to narrow the search (thus limiting its
effectiveness), and human intervention is required to review the false alarms that such a
system will produce. Here a more accurate system such as fingerprint or iris-based person
recognition is more technologically appropriate, but face recognition is chosen because it
is more acceptable and less intrusive. In Connecticut, face recognition is the secondary
biometric added to an existing fingerprint identification system. Several US States,
including Illinois, have also instituted face recognition for ensuring that people do not
obtain multiple driving licenses.
99
6.1.2 Security/Counterterrorism: Access control, comparing surveillance images to
known terrorists. To identify and verify terrorists at airports, railway stations and malls
the face recognition technology will be the best choice in India as compared with other
biometric technologies since other technologies cannot be helpful in crowdy places.
a) Access Control
Face verification, matching a face against a single enrolled template, is well within the
capabilities of current Personal Computer hardware. Since PC cameras have become
widespread, their use for face-based PC logon has become feasible, though take-up seems
to be very limited. Increased ease-of-use over password protection is hard to argue with
today’s somewhat unreliable and unpredictable systems, and for few domains there are
motivations to progress beyond the combinations of password and physical security that
protect most enterprise computers. As biometric systems tend to be third party, software
add-ons the systems do not yet have full access to the greater hardware security
guarantees afforded by boot-time and hard disk passwords. Visionics’ face-based screen
lock is one example, bundled with PC cameras. Naturally such PC-based verification
systems can be extended to control authorization for single-sign-on to multiple networked
services, for access to encrypted documents and transaction authorization, though again
uptake of the technology has been slow. Face verification is being used in kiosk
applications, notably in Mr. Payroll’s (now Innoventry) check-cashing kiosk with no
human supervision. Innoventry claims to have one million enrolled customers. Physical
access control is another domain where face recognition is attractive (e.g. Cognitec’s
Face VACS, Miros’ True Face) and here it can even be used in combination with other
biometrics. Bioid [84] is a system which combines face recognition with speaker
identification and lip motion.
b) Surveillance
The application domain where most interest in face recognition is being shown is
probably surveillance. Video is the medium of choice for surveillance because of the
richness and type of information that it contains and naturally, for applications that
require identification, face recognition is the best biometric for video data. Although gait
or lip motion recognition has some potential. Face recognition can be applied without the
100
subject’s active participation, and indeed without the subject’s knowledge. Automated
face recognition can be applied ‘live’ to search for a watch-list of ‘interesting’ people, or
after the fact using surveillance footage of a crime to search through a database of
suspects. The deployment of face-recognition surveillance systems has already begun,
though the technology is not accurate enough yet [85]. The US government is investing
in improving this technology [86] and while useful levels of recognition accuracy may
take some time to achieve, technologies such as multiple steerable zoom cameras, nonvisible wavelengths and advanced signal processing are likely to bring about superhuman perception in the data gathering side of surveillance systems.
6.1.3 Immigration
The U.S. government has recently begun a program called US-VISIT (United States
Visitor and Immigrant Status Indicator Technology), aimed at foreign travelers gaining
entry to the United States. When a foreign traveler receives his visa, he will submit
fingerprints and have his photograph taken. The fingerprints and photograph are checked
against a database of known criminals and suspected terrorists. When the traveler arrives
in the United States at the port of entry, those same fingerprints and photographs will be
used to verify that the person who received the visa is the same person attempting to gain
entry.
6.1.4 Correctional institutions/prisons
It includes the inmate tracking, and employee access. Keeping track of inmates within a
prison or jail is a constant challenge, especially as they move from one part of the facility
to another. Monitoring their movements requires corrections officers to accurately
identify individual prisoners by sight as they pass through security posts. It also requires
frequent telephone and radio communications between officers at two or more security
posts, paper passes authorizing inmates’ movements, and dry-erase or clip boards with
handwritten records to note when prisoners left one area and entered another.
6.1.5 Legislature
It includes the verification of the identity of congressman prior to vote.
101
6.2 Commercial Use
6.2.1 Banking
It includes the minimizing the fraud by verifying the identity of the person. Banks have
been very conservative in deploying biometrics as they risk losing far more through
customers disaffected by being falsely rejected than they might gain in fraud prevention.
Customers themselves are reluctant to incur burdensome additional security measures
when their personal liability is already limited by law. For better acceptance, robust
passive acquisition systems with very low false rejection probabilities are necessary.
Automated Teller Machines, already often equipped with a camera, have also been an
obvious candidate for face recognition systems for e.g. Face PIN, but development seems
not to have got beyond pilot schemes. In order to prevent the frauds of ATM in India, it is
recommended to prepare the database of all ATM customers with the banks in India &
deployment of high resolution camera and face recognition software at all ATMs. So,
whenever user will enter in ATM his photograph will be taken to permit the access after
it is being matched with stored photo from the database.
6.2.2 Pervasive Computing
Another domain where face recognition is expected to become very important, although
it is not yet commercially feasible, is in the area of pervasive or ubiquitous computing.
Many people are envisaging the pervasive deployment of information devices.
Computing devices, many already equipped with sensors, are already found throughout
our cars and in many appliances in our homes, though they will become ever more
widespread. All of these devices are just now beginning to be networked together. We
can envisage a future where many everyday objects have some computational power,
allowing them to adapt their behavior —to time, user, user control and a host of other
factors. The communications infrastructures permitting such devices to communicate to
one another are being defined and developed for e.g. Bluetooth, IEEE 802.11. So while it
is easy to see that the devices will be able to have a well-understood picture of the virtual
world with information being shared among many devices, it is less clear what kind of
information these devices will have about the real physical world. Most devices today
have a simple user interface with inputs controlled only by active commands on the part
102
of the user. Some simple devices can sense the environment, but it will be increasingly
important for such pervasive, networked computing devices to know about the physical
world and the people within their region of interest. Only by making the pervasive
infrastructure ‘human aware’ can we really reap the benefits of productivity, control and
ease-of-use that pervasive computing promises. One of the most important parts of
human awareness is to know the identity of the users close to a device, and while there
are other biometrics that can contribute to such knowledge, face recognition is the most
appropriate because of its passive nature. There are many examples of pervasive face
recognition tasks: Some devices such as Personal Digital Assistants (PDAs) may already
contain cameras for other purposes, and in good illumination conditions will be able to
identify their users. A domestic message centre may have user personalization that
depends on identification driven by a built-in camera. Some pervasive computing
environments may need to know about users when not directly interacting with a device,
and may be made ‘human aware’ by a network of cameras able to track the people in the
space and identify each person, as well as have some understanding of the person’s
activities. Thus a video conference room could steer the camera and generate a labeled
transcript of the conference; an automatic lobby might inform workers of specific
visitors; and mobile workers could be located and kept in touch by a system that could
identify them and redirect phone calls.
6.2.3 Voter verification
It includes the minimizing the fraud by verifying the identity of a person. Duplicate voter
are being reported in India. To prevent this, a database of all voters, of course, of all
constituencies, is recommended to be prepared. Then at the time of voting the resolution
camera and face recognition equipped of voting site will accept a subject face 100% and
generates the recognition for voting if match is found. Some government agencies have
also been using the systems for security and to eliminate voter fraud. The registration
process requires the scanning of all 10 fingerprints and capturing of pictures of
applicants. The voter list includes pictures of all voters. Stakeholders have requested the,
in addition to the biometric voter registration, the commission should procure biometric
voter verification equipment for the Election Day. It will:

Prevent multiple voting.
103

Prevent voter impersonation.

Prevent ballet stuffing.
6.2.4 Healthcare
Minimize fraud by verifying identity. A New Jersey health care provider in a high-crime
area uses facial recognition technology at the entrance to its emergency room to
determine if a person entering the building has a criminal history and may pose a danger
to staff. It pays a subscription fee to access several national databases to verify the
identities of people in its database. Additionally, it reportedly uses the same technology
to screen vendors coming onto the premises.
6.2.5 Day Care
It includes the verification of the identity of the individuals picking up the children. It is
mainly used in schools in which the some person has to pick the children to bring them
back at home. Therefore, in the database the children parents photograph is maintained to
verify the identity of the person.
6.2.6 Missing Children/Runaways
Search surveillance images and the internet for missing children and runaways. The
photograph a children is given to the police station and to be published in the newspapers
to verify the identity of the person.
6.2.7 Residential Security
This application alerts the homeowners of about approaching the personnel at their home.
In includes the CCTV cameras which are used worldwide like in the malls,
colleges/university campus to provide the security and to prevent the thefts and many
other crimes.
6.2.8 Internet, E-commerce
This application verifies the identity of a person for Internet purchases. There are many
sites which includes the online shopping in which the person has to pay the money with
internet banking. Therefore, to secure the transaction that the person made, face
recognition technology is used to verify the identity of the person.
104
6.2.9 Benefit payment
It minimizes the fraud by verifying the identity of the person. Attackers may attack on
someone’s privacy things like their PIN no and credit card no and then withdraw the
money from their bank account and bankrupt the person whose PIN no is stolen.
6.3 Other areas where face recognition used

Passport and Visa verification

Driving License verification

To identify and verify terrorists at airports, railway stations and malls

In defense ministry and all other important places

Credit card authentication

Entitlements and benefits authorization

Automobile ignition and unlocking; anti-theft devices

Secure financial transactions

Internet security, control of access to privileged information

Ticketless travels, authentication of rights to services

Premises access control (home, office, laboratory, etc.)

In various important examinations such as SSC, HSC, Medical, Engineering,
MCA, MBA, B- Pharmacy, Nursing courses etc. The examinee can be identified
and verified using Face Recognition Technique.

In police station to identify and verify the criminals.

Vaults and lockers in banks for access control verification and identification of
authentic users.

Present bar code system could be completely replaced with the face recognition
technology as it is a better choice for access & security since the barcode could be
stolen by anybody else.
105
Chapter 7
Conclusion and Future Scope
Face recognition has proven to be a very useful and multipurpose security measure. It is a
quick and accurate way of identifying an individual with no scope for human error. Face
recognition is widely used in banking and in law enforcement applications and also have
many applications in other fields where security is necessary. In face recognition field,
there are many problems have been solved, but, there are also many problems need us to
solve. A major difficulty is the design of face image acquisition system to ensure that
whether a person is actually live or not. Therefore, in the proposed work, a Euclidean
distance test (EDT) is generated. Euclidean distance test (EDT) is used for checking a
person’s aliveness which ensures the detection of fake/dummy images. The proposed
liveness detection approach not only verifies that the given biometric sample is from an
authorized person but also verifies that the given biometric sample is real or fake.
Initially, it checks the liveness of the person. If the person is found alive, only then the
further calculations are performed. This proposed approach is suitable for any real time
applications such as e-voting, employee management, terrorist identification, passport
verification etc. EDT opens a new path in face recognition research work. In further
development, this system can be improved to identify a person at a few meters distance
apart. Next, in this dissertation we presented a model for facial expression recognition. In
this model, the presented system is a multimodal architecture in which the detection of a
person is performed on the basis of face and the Facial Expression values. We have
maintained a dataset of face and Facial Expression images. The user will pass the input in
the form of face image and the comparison will be performed on both the face and the
Facial Expression images. To perform the recognition process the PCA and the Eigen
face approach is used collectively. The basic steps of recognition are same as some
existing approach. Here we have providing an approach to use combination of
approaches to identify the face and Facial Expression images. To do this, we used the
106
concept of Euclidean distance. On the basis of Euclidean distance we recognize the
person’s expression. In the proposed work, we compare the Euclidean distance between
the actual image of a person and the expression image of a person. Therefore, on the basis
of the range of Euclidean distance we recognize the person’s expression. The present
work can be extended by researchers in different direction. The foremost process is to
improve the identification approach by using some neural or the fuzzy based analysis.
Another improvement can be done to replace the face recognition using facial feature
recognition and facial Expression detection by radial curvature detection. The proposed
work is about to detect the face from the still image and to recognize it we can future
enhance this work in different cases. First of all we can include the concept of face
recognition from the multiple person images. It means its work will be to detect the
person from the set of face on single images and to recognize his expression. Another
extension can be done in area to include the moving image i.e. to capture the image from
the webcam and to perform the recognition.
Based on the performance, applications, reliability, ease of use, software and hardware
devices that currently support it, face recognition technology has potential for widespread
use. Face recognition technology cost comparably much better with many other biometric
traits.
107
Bibliography:
[1] Jain, A.K., Bolle, R. and Pankanti, S., Biometrics: Personal Identification in a
networked Society, Kluwer Academic Publishers, p.37, (2000).
[2] G. Hua, M.-H. Yang, E. Learned-Miller, Y. Ma, M. Turk, D. J. Kriegman, and T. S.
Huang, “Introduction to the special section on real-world face recognition,” IEEE
Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 10, pp. 1921–
1924, 2011.
[3] N. A. Spaun, “Forensic biometrics from images and video at the Federal Bureau of
Investigation,” in Proc. BTAS, 2007, pp. 1–3.
[4] Chander Kant, Sheetal Verma “Biometric Recognition System: An introduction”
published in National level seminar on Convergence of IT and Management on 24 Nov.
2007 at TIMT, Yamunanagar, (2007).
[5] Arun Ross, “Iris recognition: the path forward”, IEEE Computer Society, (2010).
[6] Makram Nabti, Ahmed Bouridane, “An effective and fast iris recognition system
based on a combined multiscale feature extraction technique”, Pattern Recognition, 41,
pp.868-879, (2008).
[7] rhien-Lien Hsu, Mohamed Abdel-Mottabel, Anil K. Jain “Face Detection in color
Images”, IEEE trans. Pattern Analysis and machine Intelligence, Vol. 24, no. 5, pp.696706, (May2002).
[8] N. A. Spaun, “Facial comparisons by subject matter experts: Their role in biometrics
and their training,” in Proc. ICB, 2009, pp. 161–168.
[9]
Wang, C., and Brandstein, M.S., “A hybrid real-time face tracking system” Proc.
ICASSP 1998, Seattle, WA, May, 1998, pp. 3636-3740.
[10] A. King, “A survey of methods for face detection,” Technical Report. McGill Unv.,
Mar. 2003.
108
[11] M. Yang, D.J. Kriegman and N. Ahuja, “Detecting faces in images: a survey,” IEEE
Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 1, pp. 34-58, Jan
2002. [12] N. A. Spaun, “Forensic Biometrics from Images and Video at the Federal
Bureau of Investigation,” in Proc. BTAS, 2007, pp. 1–3.
[13] Anil K. Jain, Arun Ross and SalilPrabhakar2, “An Introduction to Biometrics
Recognition”, IEEE Transactions on Circuits and Systems for Video Technology, Special
issue on Image-and-Video-Based Biometrics, Vol. 14, No. 1, January 2004.
[14] Bubeck, U. M. and Sanchez D., “Biometric Authentication: Technology and
Evaluation”, 2003.
[15] Ross, A., Nandakumar, K., Jain, A.K.: Handbook of Multibiometrics. Springer
Verlag (2006).
[16] Aurora Defense “Multi Modal Technology makes Biometrics Work”. PR Web Press
Release, LLC, (2002).
[17] R. A. Wasniowski, “Data Fusion for Biometrics Authentication”. RAW99-SR-320,
(2002).
[18] Prabhakar, S. and Jain, A. K. “Decision-level Fusion in Biometric Verification” to
appear in Pattern Recognition Vol. 35, No. 4, (2002).
[19] Ross, A. and Jain, A. K., “Information Fusion in Biometrics” to appear in Pattern
Recognition Letters, (2003).
[20] Weicheng Shen and Tieniu Tan, “Automated Biomertics-based personal
Identification”, Arnold and Mabel Beckman Center of the National Academics of
Sciences and Engineering in Irvine, CA, August 1998.
[21] Michal Chora, “Emerging Methods of Biometrics Human Identification” Image
Processing Group, Institute of Telecommunications, 695-882, IEEE, (2007).
[22] Qinghan Xiao, “Biometrics Technology, Application, Challenge, and Computational
Intelligence Solutions” IEEE Computational Intelligence Magazine, (May 2007).
109
[23] Arun Ross, Anil K. Jain, “Information Fusion in Biometrics”, Department of
Computer Science and Engineering, 2002.
[24] Sanjay Kr. Singh, D. S. Chauhan, Mayank Vatsa, and Richa Singh, “A Robust Skin
Color Based Face Detection Algorithm” Tamkang Journal of Science and Engineering,
Vol. 6, No. 4, pp. 227-234 (2003).
[25] George Chellin Chandran, J. and Dr. Rajesh, R.S. “Performance Analysis of
Multimodal Biometric System Authentication” IJCSNS International Journal of
Computer Science and Network Security, Vol. 9, No. 3, (March 2009).
[26] J. D. Woodward, “Biometrics: Privacy’s foe or privacy’s friend?” in the Proceedings
of the IEEE (Special Issue on Automated Biometrics), vol. 85, pp. 1480–1492,
September 1997.
[27] L. Hong and A. K. Jain, “Integrating faces and fingerprints for personal
identification,”
IEEE Trans. Pattern Anal. Machine Intel, vol. 20, pp. 1295–1307, Dec.
1998.
[28] Jun Ou, Xiao-Bo Bai, “Automatic Facial Expression Recognition Using Gabor Filter
and Expression Analysis”, 2010 Second International Conference on Computer Modeling
and Simulation 978-0-7695-3941-6/10© 2010 IEEE.
[29] Ming Hu, “Application of Rough Sets to Image Pre-processing for Face Detection”,
Proceedings of the 2008 IEEE International Conference on Information and Automation
978-1-4244-2184-8/08© 2008 IEEE.
[30] Jiying Wu, “Multi-Scale Preprocessing Model for Face Recognition”, in the ICSP
2008 Proceedings 978-1-4244-2179-4/08©2008 IEEE.
[31] Bongjin Jun, “Statistical Face Image Preprocessing and Non-statistical Face
Representation for Practical Face Recognition”, 978-1-4244-5950-6/09©2009 IEEE.
[32]Yong Zhang, “Hand-Drawn Face Sketch Recognition by Humans and a PCA-Based
Algorithm for Forensic Applications”, IEEE Transactions on Systems, Man, and
Cybernetics—PART A: SYSTEMS AND HUMANS 1083-4427© 2010 IEEE.
110
[33] Jain A.K, Ross Arun and Uludag.U. “Biometrics Template security: Challenges and
solutions” in Proceedings of European Signal Processing Conference September 2005.
[34] Jian Yang,” Constructing PCA Baseline Algorithms to Reevaluate ICA-Based FaceRecognition Performance”, IEEE Transactions on Systems, Man, and Cybernetics—
PART B: CYBERNETICS 1083-4419© 2007 IEEE.
[35] Walid Riad Boukabou, “An Improved LDA Approach With DFB Preprocessing for
Face Recognition”, 2007 ECSIS Symposium on Bio-inspired, Learning, and Intelligent
Systems for Security 0-7695-2919-4/07© 2007 IEEE.
[36] Bruce A. Draper, “Recognizing Faces with PCA and ICA”.
[37] Rabia Jafri and Hamid R. Arabnia, “A Survey of Face Recognition Techniques”
Journal of Information Processing Systems, Vol.5, No.2, June 2009.
[38] Hu Han and Anil K. Jain, “3D Face Texture Modeling from Uncalibrated Frontal
and Profile Images” published in the 6th IAPR International Conference on Biometrics
(ICB), June 4 - 7, 2013, Madrid, Spain.
[39] Hu Han, Charles Otto, and Anil K. Jain, “Age Estimation from Face Images: Human
vs. Machine Performance” published in the 6th IAPR International Conference on
Biometrics (ICB), June 4 - 7, 2013, Madrid, Spain.
[40] Brendan F. Klare and Anil K. Jain, “Face Recognition: Impostor-based Measures of
Uniqueness and Quality” in the Proceedings of the IEEE Conference on Biometrics:
Theory, Applications, and Systems, 2012.
[41] Zhifeng Li, Unsang Park and Anil K. Jain, “A Discriminative Model for Age
Invariant Face Recognition” in the proceedings of IEEE Transactions on Information
Forensics and Security.
[42] Volker Blanz and Thomas Vetter, “Face Recognition Based on fitting a 3D Morph
able Model” published in the IEEE Transactions on Pattern Analysis and Machine
Intelligence, Vol. 25, No. 9, September 2003.
111
[43] Zahid Riaz, Christoph Mayer, Matthias Wimmer, and Bernd Radig “Model Based
Face Recognition across Facial Expressions” in the proceedings of IEEE transactions.
[44] A.A. Salah, M. Biceg, L. Akarun, E. Grosso, M. Tistarelli, “Hidden Markov Modelbased face recognition using selective attention”.
[45] Congyong Su and Li Huang, “Spatio-Temporal Graphical-Model-Based Multiple
Facial Feature Tracking”.
[46] Yan Tong, Wenhui Liao and Qiang Ji, “Facial Action Unit Recognition by
Exploiting their Dynamic and Semantic Relationships” published in IEEE Transactions
on Pattern Analysis and Machine Intelligence, Vol. 29, No. 10, October 2007.
[47] Maja Pantic, and Leon J. M. Rothkrantz “Facial Action Recognition for Facial
Expression Analysis from Static Face Images” published in the IEEE Transactions
Systems, Man, and Cybernetics—Part B: Cybernetics, Vol. 34, No. 3, June 2004.
[48] Michael A. Sayette, Jeffrey F. Cohn, Joan M. Wertz, Michael A. Perrott, and
Dominic J. Parrot “A Psychometric Evaluation of the Facial Action Coding System for
Assessing Spontaneous Expression” published in the Journal of Nonverbal Behavior
25(3), 2001 Human Sciences Press, Inc.
[49] Ying-li Tian, Takeo Kanade and Jeffrey F. Cohn, “Recognizing Action Units for
Facial Expression Analysis”.
[50] Alexander M. Bronstein, Michael M. Bronstein and Ron Kimmel “ThreeDimensional Face Recognition” published in the International Journal of IJCV - second
review - 3".tex; 14 December, 2004; 20:14; p.1-44.
[51] Claude C. Chibelushi, Fabrice Bourel “Facial Expression Recognition: A Brief
Tutorial Overview”.
[52] Liwei Wang, Yan Zhang, Jufu Feng, “On the Euclidean Distance of Images”.
112
[53] Zhen Leiy, Shengcai Liaoz, Anil K. Jainz, and Stan Z. Liy, “Coupled Discriminant
Analysis for Heterogeneous Face Recognition” in the proceedings of IEEE.
[54] Farrukh Sayeed, M. Hanmandlu, and A.Q. Ansari “Face Recognition using
Segmental Euclidean Distance” Defence Science Journal, Vol. 61, No. 5, September
2011, pp. 431-442 @ 2011, DESIDOC.
[55] Shalini Gupta, Mia K. Markey, J. K. Aggarwal, Alan C. Bovik “Three dimensional
face recognition based on geodesic and Euclidean distances” in the Proceedings of
SPIEIS& T Electronic Imaging, SPIE Vol. 6499, 64990D, © 2007 SPIEIS& T ∙
0277786X/ 07/$15.
[56] Haitao Zhao, “A Novel Incremental Principal Component Analysis and its
Application for Face Recognition” published in the IEEE Transactions on Systems, Man,
and Cybernetics—Part B: Cybernetics 1083-4419© 2006 IEEE.
[57] Ayan Chakrabarti,” Super-Resolution of Face Images Using Kernel PCA-Based
Prior”, IEEE Transactions on Multimedia 1520-9210© 2007 IEEE.
[58] J.Braun, “Machine Learning and Recognition of Faces”.
[59] Takeo Kanade, Jeffrey F. Cohn and Yingli Tian, “Comprehensive Database for
Facial Expression Analysis”.
[60] Gaurav Mittal, “Robust Preprocessing Algorithm for Face Recognition” in the
Proceedings of the 3rd Canadian Conference on Computer and Robot Vision (CRV’06)
0-7695-2542-3/06© 2006 IEEE.
[61] R. D. Madalina Baltatu and R. D’Amico, “Towards ubiquitous Acceptance of
Biometric Authentication: Template Protection Techniques, vol. 3087/2004 Springer
Berlin / Heidelberg, (2004).
[62] R. M. B. N. K. Ratha, J. H. Connel and S. Chikkerur, “Cancelable biometrics: A
case study in fingerprints” published in the IEEE International Conference on Pattern
Recognition, volume 4, pages 370–373, Hong Kong, China, 2006 IEEE.
113
[63] M. Boatwright and X. Luo, “What do we know about biometrics authentication?” In
InfoSecCD ’07: Proceedings of the 4th annual conference on Information security
curriculum development, pages 1–5, New York, NY, USA, 2007.
[64] T. E. Boult and R.Woodworth, “Privacy and Security Enhancements in Biometrics”
published in Springer, US, 2005.
[65] A. K. Jain, “Biometric Recognition: How Do I Know Who You Are?” published in
Springer, US, 2005.
[66] E. Kindt, “Biometric applications and the data protection legislation” published in
Springer, US, 2007.
[67] A. Adler, “Biometric System Security” published in Springer, US, 2007.
[68] C. Vielhauer, “Biometric User Authentication for IT Security: from Fundamentals to
Handwriting” published in Springer, US, 2005.
[69] A. K., J. D. Maltoni, D. Miao and S. Prabhakar, “Handbook of Fingerprint
Recognition” Springer-Verlag, 2003.
[70] Galbally, J., Fierrez, J., Ortega-Garcia, J.: Vulnerabilities in biometric systems:
Attacks and Recent Advances in Liveness Detection, Biometrics Recognition Group,
Madrid, Spain, 2007, p.8.
[71] W. Zhao, R. Chellappa, A. Rosenfeld, and P.J. Phillips, “Face Recognition:
AliteratureSurvey”, Technical Report, Univ. of Maryland, 2000.
[72] V. Bruce, “Identification of Human Faces”, pp. 615-619, Image Processing and Its
Applications, Conference Publication No. 465, IEEE, 1999.
[73] ] J. Li, G. Chen, and Z. Chi, “A Fuzzy Image Metric with Application to Fractal
Coding,” IEEE Trans. Image Processing, vol. 11, no. 6, pp. 636-643, June 2002.
[74] N. Ahuja, “A Transform for Multiscale Image Segmentation by Integrated Edge and
Region Detection,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 18, no.
9, pp. 1211-1235, Sept. 1996.
114
[75] Y. Amit, D. Geman, and B. Jedynak, “Efficient Focusing and Face Detection,” Face
Recognition: From Theory to Applications,H. Wechsler, P.J. Phillips, V. Bruce, F.
Fogelman-Soulie, and T.S. Huang, eds., vol. 163, pp. 124-156, 1998.
[76] R. Chellappa, C.L. Wilson, and S. Sirohey, “Human and Machine Recognition of
Faces: A Survey,” Proc. IEEE, vol. 83, no. 5, pp. 705-740, 1995.
[77] M.Yang, D.J. Kriegman, and N.Ahuja, Detecting Faces in Images: A Survey, IEEE
Trans. on Pattern Analysis and Machine Intelligence, VOL. 24, NO. 1, Jan. 2002.
[78] J. Liu and Y. H. Yang, “Multiresolution color image segmentation”, IEEE Trans. on
PAMI, vol. 16, pp. 689-700, July 1994.
[79] W. Zhao, R. Chellapa, A. Rosenfeld and P.J. Philips, “Face Recognition: A
Literature Survey”, UMD CFAR Technical Report CAR-TR-948, 2000.
[80] William A. Barrett, “A Survey of Face Recognition Algorithms and Testing
Results”, Proceeding of IEEE, 1998.
[81] A.J. Zlochower, Deciphering emotion from the face: An evaluation of FACS, EMG,
and computer-vision based approaches to facial expression analysis. Unpublished
manuscript, University of Pittsburgh, 1997.
[82] R. Matias, J.F. Cohn, S. Ross, A comparison of two systems to code infants'
affective expression. Developmental Psychology, 25: 483-489, 1989.
[83]
Biometrics
in
Human
Services
User
Group.
URL:
http://www.dss.state.ct.us/digital.htm.
[84] Ana Orubeondo, “A New Face for Security” in InfoWorld.com, May 2001.
[85] Lee Gomes. Can Facial Recognition Help Snag Terrorists? The Wall Street Journal,
September 21 2001.
[86] Defense Advanced Research Projects Agency, “Human Identification at a Distance”,
BAA00-29 edition, Feb 2000.
115
116
117
118
Download