Professional Documents
Culture Documents
Faculty of Computer
& Information Sciences
Computer Science Department
By:
Supervised By:
I would like to thank Prof. Dr. Taymoor Nazmy - the vice dean of the faculty of
Computer and Information Sciences, Ain Shams University- for his support. My
great thanks to Prof. Dr. Saied ELGhonaimy, Computer Systems Dept. – Faculty
of Computer and Information Sciences, Ain Shams University, for his
encouragement and support. Special thanks to Prof. Dr. Mohammed Hashem,
Head of Information Systems Dept. – Faculty of Computer and Information
Sciences, Ain Shams University, for his advices and support.
I would like to express my deep appreciation and thanks to all who supervise me:
Prof. Dr. Mostafa Seyam (god bless him), Prof. Dr. Taha I. El-Areif, Dr.
Khaled A. Nagaty and Dr. Haitham ELMessairy, for their great help and
encouragement during the execution of this work. Special thanks to Dr. Khaled
A. Nagaty for his follow, care, patience and support for me. Also, I would like to
thank my colleagues Mona Wagdy, Amr EL-Desoky, Kareem Emara,
Mahmoud Hossam and Mohammed Hamdy for their valuable help and
cooperation.
I would like to thank the team of “Face Recognition using Eigenface, NN and
Mosaicing Techniques” graduation project, Faculty of Computer and Information
Sciences, Ain Shams University, 2005, for their project which is used in
experiments of this work.
II
Publications
The work presented in this thesis has been published in the following conferences:
1. T.I. El-Arief, K.A. Nagaty, and A.S. El-Sayed, “Eigenface vs.
Spectroface: A Comparison on the Face Recognition Problems”,
IASTED Signal Processing, Pattern Recognition, and Applications
(SPPRA’07), Austria, 2007.
2. S. El-Sayed, K. A. Nagaty, T. I. El-Arief, “An Enhanced Histogram
Matching Approach using the Retinal Filter’s Compression Function
for Illumination Normalization in Face Recognition”, ICIAR’08,
Springer-Verlag LNCS 5112, pp. 873–883, Portugal, 2008.
III
Abstract
Although many face recognition techniques and systems have been proposed,
evaluations of the state-of-the-art techniques and systems have shown that
recognition performance of most current technologies degrades due to the
variations of illumination. In the last face recognition vendor test FRVT 2006,
they conclude that relaxing the illumination condition has a dramatic effect on the
performance. Moreover, It has been proven both experimentally and theoretically
that the variations between the images of the same face due to illumination are
almost always larger than image variations due to change in face identity.
There has been much work dealing with illumination variations in face
recognition. Although most of these approaches can cope with illumination
variation well, some may bring negative influence on images without illumination
variation. In addition, some approaches show great difference on performance
when combined with different recognition methods. Some other approaches
require perfect alignment of face within the image which is difficult to achieve in
practical/real-life systems.
To verify both the flexibility to face recognition approaches and robustness to non-
aligning of faces, the proposed approach is tested over two face recognition
methods representing the two broad categories of the holistic-based approach –
IV
namely Standard Eigenface method from the Eigenspace-based category and
Spectroface from the frequency-based category. In each method, the testing is
done using both aligned and non-aligned versions of the Yale B database.
In order to compare the proposed approach with other approaches, we select four
best-of-literature illumination normalization approaches among 38 different
approaches based on surveying nine different comparative studies. All five
approaches are compared using Eigenface and Spectroface methods on images
with illumination variations and images with other facial and geometrical
variations.
In illumination variation, the proposed approach gives the best results on the
Eigenface and the second best results on the Spectroface, when images are not
perfectly aligned. Moreover, the proposed approach is the minimum affected
approach (i.e. most robust) due to the non-aligning of faces on both methods.
In other facial and geometrical variations, the proposed approach has the minimum
negative influence on each of the two methods among the four other approaches.
In this work, all illumination normalization approaches are tested on two face
recognition methods representing the two broad categories of the holistic-based
approach. It’s important to extend this work to include local-based face
recognition methods in testing these approaches as they may introduce great
difference in performance when combined with such methods.
Moreover, this work introduces a technology evaluation for the proposed approach
and the other best-of-literature approaches. In order to complete the thorough
evaluation cycle, both scenario and operational evaluations need to be performed
for these approaches.
V
VI
Table of Contents
ACKNOWLEDGEMENT .......................................................................................................................... II
PUBLICATIONS........................................................................................................................................ III
ABSTRACT ................................................................................................................................................ IV
TABLE OF CONTENTS ..........................................................................................................................VII
LIST OF FIGURES.................................................................................................................................... IX
LIST OF TABLES.....................................................................................................................................XII
CHAPTER 1: INTRODUCTION ............................................................................................................... 1
1.1 BIOMETRICS AND FACE RECOGNITION ................................................................................................. 1
1.2 PROBLEM DEFINITION........................................................................................................................... 3
1.3 METHODS CATEGORIZATION ................................................................................................................ 3
1.4 VARIATIONS CATEGORIZATION ............................................................................................................ 3
1.5 SUCCESSFUL SCENARIOS ...................................................................................................................... 4
1.6 COMMERCIAL SYSTEMS ........................................................................................................................ 5
1.7 RECENT EVALUATIONS ......................................................................................................................... 6
1.8 THESIS OBJECTIVES AND ORGANIZATION ............................................................................................. 7
CHAPTER 2: FACE RECOGNITION APPROACHES .......................................................................... 9
2.1 INTRODUCTION ..................................................................................................................................... 9
2.2 LOCAL-BASED APPROACHES .............................................................................................................. 10
2.3 HOLISTIC-BASED APPROACHES .......................................................................................................... 15
2.3.1 Eigenspace-based Category ....................................................................................................... 16
2.3.2 Frequency-based Category ........................................................................................................ 21
2.3.3 Other Holistic-Based Approaches.............................................................................................. 24
2.4 HYBRID APPROACHES......................................................................................................................... 27
2.5 PERFORMANCE EVALUATIONS AND COMPARATIVE STUDIES.............................................................. 28
2.5.1 Performance Evaluation ............................................................................................................ 28
2.5.2 Comparative Studies .................................................................................................................. 30
CHAPTER 3: ILLUMINATION NORMALIZATION APPROACHES .............................................. 31
3.1 INTRODUCTION ................................................................................................................................... 31
3.2 MODEL-BASED APPROACHES ............................................................................................................. 32
3.3 IMAGE-PROCESSING-BASED APPROACHES ......................................................................................... 40
3.3.1 Global Approaches .................................................................................................................... 40
3.3.2 Local Approaches ...................................................................................................................... 45
3.4 COMPARATIVE STUDIES & BEST-OF-LITERATURE APPROACHES ........................................................ 53
CHAPTER 4: SETUP THE ENVIRONMENT ....................................................................................... 62
4.1 INTRODUCTION ................................................................................................................................... 62
4.2 METHODS DESCRIPTIONS ................................................................................................................... 63
4.2.1 Standard Eigenface Method ....................................................................................................... 63
4.2.2 Spectroface Method ................................................................................................................... 63
4.3 DATABASES DESCRIPTIONS ................................................................................................................ 65
4.3.1 UMIST database ........................................................................................................................ 65
4.3.2 Yale B database.......................................................................................................................... 65
4.3.3 Grimace database ...................................................................................................................... 66
4.3.4 JAFFE database......................................................................................................................... 67
4.3.5 Nott-faces database .................................................................................................................... 68
4.3.6 Yale database ............................................................................................................................. 68
4.3.7 Face 94 database ....................................................................................................................... 68
VII
4.4 EXPERIMENTAL RESULTS ................................................................................................................... 69
4.4.1 Pose Variation ........................................................................................................................... 69
4.4.2 Facial Expressions Variation ..................................................................................................... 70
4.4.3 Non-Uniform Illumination Variation ......................................................................................... 72
4.4.4 Translation Variation ................................................................................................................. 73
4.4.5 Scaling Variation ....................................................................................................................... 75
4.5 SUMMARY .......................................................................................................................................... 76
CHAPTER 5: THE PROPOSED ILLUMINATION NORMALIZATION APPROACH................... 77
5.1 INTRODUCTION ................................................................................................................................... 77
5.2 IDEA OF THE PROPOSED APPROACH .................................................................................................... 77
5.3 HISTOGRAM MATCHING ALGORITHM ................................................................................................. 79
5.4 IMAGE ENHANCEMENT METHODS ...................................................................................................... 81
5.4.1 Histogram Equalization (HE) .................................................................................................... 81
5.4.2 Log Transformation (LOG) ........................................................................................................ 81
5.4.3 Gamma Correction (GAMMA) .................................................................................................. 81
5.4.4 Compression Function of the Retinal Filter (COMP) ................................................................ 82
5.5 THE ENHANCED HM APPROACHES ..................................................................................................... 83
5.5.1 Enhancement After HM .............................................................................................................. 83
5.5.2 Enhancement Before HM ........................................................................................................... 84
5.5.3 Further Enhancement ................................................................................................................ 85
5.6 VERIFICATION OF THE SELECTION CONDITIONS ................................................................................. 85
5.7 EXPERIMENTAL RESULTS ................................................................................................................... 87
5.8 SUMMARY .......................................................................................................................................... 97
CHAPTER 6: EVALUATE THE PROPOSED APPROACH................................................................ 99
6.1 INTRODUCTION ................................................................................................................................... 99
6.2 IMPLEMENTATION OF THE COMPARED APPROACHES .......................................................................... 99
6.2.1 Preprocessing Chain Approach (CHAIN).................................................................................100
6.2.2 Local Normal Distribution (LNORM) .......................................................................................101
6.2.3 Single Scale Retinex with Histogram Matching (SSR-HM) ......................................................101
6.2.4 Local Binary Patterns (LBP) ....................................................................................................102
6.2.5 Proposed Approach (GAMMA-HM-COMP) ............................................................................102
6.3 COMPARISON ON ILLUMINATION VARIATIONS ..................................................................................103
6.3.1 Aligned Faces............................................................................................................................103
6.3.2 Non-Aligned Faces....................................................................................................................103
6.4 COMPARISON ON OTHER VARIATIONS ...............................................................................................106
6.4.1 Pose Variations .........................................................................................................................107
6.4.2 Facial Expressions Variations ..................................................................................................108
6.4.3 Translation Variations ..............................................................................................................111
6.4.4 Scaling Variations .....................................................................................................................115
6.5 SUMMARY .........................................................................................................................................116
CHAPTER 7: CONCLUSIONS AND FUTURE WORKS ....................................................................118
7.1 CONCLUSIONS....................................................................................................................................118
7.2 FUTURE WORKS ................................................................................................................................120
REFERENCES ..........................................................................................................................................121
VIII
List of Figures
Figure 1.1: Distribution of some biometrics over the market 1
Figure 1.2: Number of published items and citations on face recognition between 1991 and 2006 2
Figure 1.3: Easy scenarios in face recognition 4
Figure 1.4: Easy scenarios in face recognition 4
Figure 1.5: Difficult scenarios for face recognition 5
Figure 2.1: Face bunch graph (FBG) serves as a general representation of faces. It is designed to
cover all possible variations in the appearance of faces. The FBG combines information from a
number of face graphs. Its nodes are labeled with set of jets, called bunches, and its edges are
labeled averages of distance vectors. During comparison to an image, the best fitting jet in each
bunch, indicated by gray shading, is selected independently. 11
Figure 2.2: A visualized example for the steps of automatically localizing features. In (e), the black
cross on a white background indicates an extracted and stored feature vector at this location
while white cross on black background indicates an ignored feature vector. 12
Figure 2.3: System overview of the component-based face detector using four components 13
Figure 2.4: Sample of the normalized whole face image and the three regions that are used for the
local analysis 14
Figure 2.5: Example for illustrating the basic LBP operator 15
Figure 2.6: Examples of circular neighborhoods with number of sampled points P and radius R (P,R)
15
Figure 2.7: Block diagram of the standard Eigenface method 16
Figure 2.8: The subspace LDA face recognition system 17
Figure 2.9: Flowchart for the Face recognition using evolutionary pursuit (EP) method 18
Figure 2.10: Image synthesis model for Architecture 1. To find a set of IC images, the images in X are
considered to be a linear combination of statistically independent basis images, S, where A is an
unknown mixing matrix. The basis images are estimated as the learned ICA output U. 19
Figure 2.11: Example of the projection map and the projection-combined image 19
Figure 2.12: 3-level wavelet decomposition 22
Figure 2.13: (a) input image, (b) the log-magnitude of its DCT, (c) the scanning strategy of
coefficients 22
Figure 2.14: Most variant frequencies: a) real, b) imaginary and c) selected numbering 23
Figure 2.15: Spectroface representation steps 24
Figure 2.16: Examples of FBT of (A) an 8 radial cycles image, (B) a 4 angular cycles image and (C)
an image of the average of these images. The magnitude of the FBT coefficients is presented in
colored levels (red indicates the highest value) 24
Figure 2.17: Block diagram for face recognition based on moments 25
Figure 2.18: Examples of the Trace transform on (a) full image (b) masked with rectangular shape
and (c) masked with elliptical shape. 26
Figure 2.19: Training and recognition stages of Face Recognition Using Local and Global Features
approach 27
Figure 2.20 (a) A gray-scale face image, (b) it’s edginess image, and (c) the cropped eyes. 28
Figure 3.1: The same individual imaged with the same camera and the same facial expression may
appear dramatically different with changes in the lighting conditions. 31
Figure 3.2: Effect of applying QIR on an illuminated face image from Yale B database 34
Figure 3.3: Effect of applying SQI approach to illuminated face images from Yale B and CMU PIE
databases. 34
Figure 3.4: The effect of the scale, σ, on processing an illuminated facial image using the SSR. 37
Figure 3.5: Histogram fitted version of SSR with σ = 6 38
IX
Figure 3.6: Discretization lattice for the PDE in equation 3.23 39
Figure 3.7: Effect of applying GROSS approach on some illuminated face images from Yale B
database 40
Figure 3.8: Effect of applying histogram equalization on an illuminated image 41
Figure 3.9: Histogram matching process to an illuminated image 42
Figure 3.10: Transformation functions of LOG and GAMMA (L: number of gray levels) 43
Figure 3.11: Effect of applying LOG approach to an illuminated face image. 44
Figure 3.12: Effect of applying GIC to an illuminated face image 45
Figure 3.13: Effect of applying NORM approach to an illuminated image. (Note that the gray-level of
the resulting image is stretched to [0,255] for displaying purpose only) 45
Figure 3.14: Effects of applying the three local normalization methods to an illuminated face image46
Figure 3.15: An example of ideal region partition 46
Figure 3.16: The four regions for illumination normalization 47
Figure 3.17: The effects of applying region-based strategy of HE and GIC over the four face regions
47
Figure 3.18: Block histogram matching. In each image pair, the left one is the input image while the
right one is the reference image. 47
Figure 3.19: The windowing filter H used in the Block HM method 48
Figure 3.20: Images before and after intensity normalization with BHM. (a) Input images, (b)
corresponding output images after applying BHM 49
Figure 3.21: The LBP operator 49
Figure 3.22: The extended LBP operator with (8,2) neighborhood. Pixel values are interpolated for
points which are not in the center of a pixel. 50
Figure 3.23: Original image (left) processed by the LBP operator (right). 50
Figure 3.24: Effects of applying the image processing steps proposed by [91] 51
Figure 3.25: Examples of images of one person from the Extended Yale-B frontal database. The
columns respectively give images from subsets 1 to 5. 53
Figure 3.26: Summarization for the first four comparative studies. For each study, it shows the
normalization approach to be compared, the face databases and the face recognition
approaches in addition to the best normalization approaches from each study (grayed boxes) 55
Figure 3.27: Summarization for the nine comparative studies showing some relations between these
studies in addition to the final best normalization approaches from all studies (dark grayed
boxes). For each study, it shows the normalization approach to be compared, the face
databases and the face recognition approaches in addition to the best normalization approaches
from each study (light grayed boxes) 60
Figure 4.1: Standard Eigenface block diagram 63
Figure 4.2: Spectroface block diagram 64
Figure 4.3: UMIST: selected images for one subject in both training and testing sets 66
Figure 4.4: Yale B: Training images for one subject in the four subsets with the light angle of each
image 66
Figure 4.5: Selected images for one subject from each database used for studying the facial
expression variation 67
Figure 4.6: Example images from JAFFE database. The images in the database have been rated by
60 Japanese female subjects on a 5-point scale for each of the six adjectives. The majority vote
is shown underneath each image (with natural being defined through the absence of a clear
majority) 67
Figure 4.7: Face 94: 15 images for each subject in both training and testing sets 69
Figure 4.8: Translation Variation: example for translating with and without circulation 74
Figure 5.1: Histogram matching process to an illuminated image 80
X
Figure 5.2: Transformation functions of LOG and GAMMA (L: number of gray levels) 82
Figure 5.3: Effect of the four enhancement methods on an illuminated face 83
Figure 5.4: Block diagram of applying the image enhancement method after the HM 83
Figure 5.5: Effects of applying the image enhancement methods after applying the HM 84
Figure 5.6: Block diagram of applying the image enhancement method before the HM 84
Figure 5.7: Effects of applying the image enhancement methods before applying the HM 84
Figure 5.8: Block diagram showing the further enhancement of combinations in 5.5.1 and 5.5.2 85
Figure 5.9: Effects of further enhancement on both HM-GAMMA and GAMMA-HM combinations
using each of the four enhancement methods 85
Figure 5.10: Sample faces from Yale B database – automatically and manually cropped 86
Figure 5.11: Eigenface method over YALE B-AUTO: Effects of further enhancement over the eight
single enhancements using (a) HE, (b) LOG, (c) GAMMA and (d) COMP 89
Figure 5.12: Eigenface method over YALE B-MANU: Effects of further enhancement over the eight
single enhancements using (a) HE, (b) LOG, (c) GAMMA and (d) COMP 91
Figure 5.13: Spectroface method over YALE B-AUTO: Effects of further enhancement over the eight
single enhancements using (a) HE, (b) LOG, (c) GAMMA and (d) COMP 92
Figure 5.14: Spectroface method over YALE B-MANU: Effects of further enhancement over the
eight single enhancements using (a) HE, (b) LOG, (c) GAMMA and (d) COMP 93
Figure 5.15: Effects of the five enhancement combinations that satisfy the three conditions 95
Figure 6.1: Average increasing/decreasing in recognition rates after applying each of the five
illumination normalization approaches on YALE B-MANU version 105
Figure 6.2: Average increasing/decreasing in recognition rates after applying each of the five
illumination normalization approaches on YALE B-AUTO version 106
Figure 6.3: Performance decreasing of each normalization approach due to the non-aligning of faces
(i.e. subtracting the performance on YALE B-AUTO from the performance on YALE B-
MANU) 106
Figure 6.4: Average difference in recognition rates after applying each of the five illumination
normalization approaches on UMIST database 108
Figure 6.5: Average difference in recognition rates after applying each of the five illumination
normalization approaches on Yale database 109
Figure 6.6: Average difference in recognition rates after applying each of the five illumination
normalization approaches on Grimace database 110
Figure 6.7: Average difference in recognition rates after applying each of the five illumination
normalization approaches on JAFFE database 110
Figure 6.8: Average difference in recognition rates after applying each of the five illumination
normalization approaches on Nott-faces database 110
Figure 6.9: Average decreasing in recognition rates after translating with circulation 114
Figure 6.10: Average decreasing in recognition rates after translating without circulation 115
Figure 6.11: Average decreasing in recognition rates when applying each of the five illumination
normalization approaches before and after scaling the Face 94 database 116
XI
List of Tables
Table 1.1: Different applications of face recognition 2
Table 2.1: A brief comparison between holistic-based and local-feature-based approaches 9
Table 3.1: Default parameter settings for CHAIN approach 52
Table 3.2: List for 24 illumination normalization approaches that LNORM perform better than them
57
Table 3.3: The 38 different illumination normalization approaches appear in the above nine
comparative studies together with the corresponding studies numbers. (Note that the cited
approaches, from 29 to 38, are not described in details in their corresponding comparative
studies) 61
Table 4.1: Comparison between results in Lai et al. [43] and in our implementation (better rates are
italic) 65
Table 4.2: Pose Variation: recognition rates over 12 training cases (top four rates in each method are
italic) 70
Table 4.3: Expressions Variation: recognition rates over four databases with two Eigenface tests 71
Table 4.4: Illumination Variation: recognition rates over 25 training cases (top three rates in each
method are italic) 73
Table 4.5: Translation Variation: chosen cases from the six databases and their recognition rates 73
Table 4.6: Translation Variation: average decreasing in the recognition rates of both methods after
translating with circulation in the four directions 74
Table 4.7: Translation Variation: average decreasing in the recognition rates of both methods after
translating without circulation in the four directions 75
Table 4.8: Scaling Variation: description of the training cases 75
Table 4.9: Scaling Variation: decreasing in recognition rates after scaling all images in the testing set
76
Table 5.1: The 25 different training cases used in testing 87
Table 5.2: The number of combinations that lead to increase the recognition rates after using each of
the enhancement methods for further enhancement 94
Table 5.3: Results of using the best five combinations with the Eigenface method over the two
versions of the database. Average recognition rate is calculated over the 25 different training
cases. (The best average differences are italic) 96
Table 5.4: Results of using the best five combinations with the Spectroface method over the two
versions of the database. Average recognition rate is calculated over the 25 different training
cases. (The best average differences are italic) 97
Table 6.1: Default parameter settings for CHAIN approach 100
Table 6.2: Results of applying CHAIN with and without sliding on Spectroface method on both
versions of the YALE B database 100
Table 6.3: Difference between our implementation of the LNORM and the original one 101
Table 6.4: Results of applying LNORM with and without sliding on Spectroface method on both
versions of the YALE B database 101
Table 6.5: Difference between our implementation of the SSR-HM and the original one 102
Table 6.6: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over YALE B-MANU version. Average recognition rate is
calculated over the 25 different training cases. 104
Table 6.7: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over YALE B-AUTO version. Average recognition rate is
calculated over the 25 different training cases. (0: NONE, 1: LNORM, 2: LBP, 3: CHAIN, 4:
SSR-HM, 5: GAMMA-HM-COMP, nor: normal, ver: vertical, hor: horizontal) 105
XII
Table 6.8: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over UMIST database. Average recognition rate is
calculated over all training cases. 107
Table 6.9: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over Grimace, Yale, JAFEE, and Nott-faces databases.
Average recognition rate is calculated over all training cases. 109
Table 6.10: Average decreasing in the recognition rates of both methods after translating with
circulation in the four directions while applying (a) LNORM, (b) LBP, (c) CHAIN, (d) SSR-
HM and (e) GAMMA-HM-COMP approaches as preprocessing step. 111
Table 6.11: Average decreasing in the recognition rates of both methods after translating without
circulation in the four directions while applying (a) LNORM, (b) LBP, (c) CHAIN, (d) SSR-
HM and (e) GAMMA-HM-COMP approaches as preprocessing step. 113
Table 6.12: Decreasing in recognition rates after applying each of the five illumination normalization
approaches with both Eigenface and Spectroface methods over Face 94 database. Average
decreasing in recognition rate is calculated over all training cases. 115
XIII
XIV
CHAPTER 1: Introduction
1.1 Biometrics and Face Recognition
Biometric recognition [1] refers to the use of distinctive physiological (e.g., fingerprints,
face, retina, iris) and behavioral (e.g., gait, signature) characteristics, called biometric
identifiers, for automatically recognizing individuals. Because biometric identifiers
cannot be easily misplaced, forged, or shared, they are considered more reliable for
person recognition than traditional token or knowledge-based methods. Others typical
objectives of biometric recognition are user convenience (e.g., service access without a
Personal Identification Number), better security (e.g., difficult to forge access). Fig.1.1
shows the distribution of some biometrics over the market.
1
Figure 1.2: Number of published items and citations on face recognition between 1991 and 2006
The strong demand for user friendly systems which can secure our assets and protect our
privacy without losing our identity in a sea of numbers is obvious. At present, one needs
a PIN to get cash from an ATM, a password for a computer, a dozen others to access the
internet, and so on. Although extremely reliable methods of biometric personal
identification exist, e.g., fingerprint analysis and retinal or iris scans, these methods have
yet to gain acceptance by the general population due to their needs to the cooperation of
the participants. A personal identification system based on analysis of frontal or profile
images of the face is non intrusive and therefore user friendly. Moreover, personal
identity can often be ascertained without the participant’s cooperation or knowledge. In
addition, the need for applying FRT has been boosted by recent advances in multimedia
processing along with others such as IP (Internet Protocol) technologies. Table-1.1 lists
some of the applications of face recognition [2]:
Table 1.1: Different applications of face recognition
Area Specific Applications
Drivers’ Licenses, Entitlement Programs
Biometrics Immigration, National ID, Passports, Voter Registration
Welfare Fraud
Desktop Logon (Windows XP, Windows Vista)
Application Security, Database Security, File Encryption
Information Security
Intranet Security, Internet Access, Medical Records
Secure Trading Terminals
Advanced Video Surveillance, CCTV Control
Law Enforcement and
Portal Control, Post-Event Analysis
Surveillance
Shoplifting and Suspect Tracking and Investigation
Smart Cards Stored Value Security, User Authentication
Access Control Facility Access, Vehicular Access
2
1.2 Problem Definition
A general statement of face recognition problem can be formulated as follows [2]: Given
still or video images of a scene, identify or verify one or more persons in the scene using
a stored database of faces. Available collateral information such as race, age, gender,
facial expression and speech may be used in narrowing the search (enhancing
recognition). The solution of the problem involves segmentation of faces (face detection)
from cluttered scenes, feature extraction from the face region, identification or
verification. In identification problems, the input to the system is an unknown face, and
the system reports back the decided identity from a database of known individuals,
whereas in verification problems, the system needs to confirm or reject the claimed
identity of the input face.
3
and other objects. Viewing geometry includes pose changes, either by the observer or the
object to be recognized; illumination changes include shading, color, self-shadowing, and
specular highlights; imaging process variations include resolution, focus, imaging noise,
sampling technique, and perspective distortion effects; variation from other objects
include occlusions, shadowing, and indirect illumination. These sources of variation may
or may not hinder the recognition process depending on which algorithm is used. It is
possible that the variation due to factors such as facial expression, lighting, occlusions,
and pose is larger than the variation due to identity [6], [7]. That makes identification
under such varying environments a difficult task. However, human proficiency at face
recognition [8] has motivated enormous research in this area despite these challenges.
(The ability of humans to recognize faces is also an actively researched field with widely
varying results depending on numerous factors. Additional information on this topic can
be found predominantly in the psychology literature [9], [10].)
4
Figure 1.5: Difficult scenarios for face recognition
5
1.7 Recent Evaluations
Since 1993, a series of six face recognition technology evaluations sponsored by U.S.
Government have been held. In thirteen years, performance has improved by two orders
of magnitude and there exist numerous companies selling face recognition systems [12].
The evaluations provided regular assessments of the state of the technology and helped to
identify the most promising approaches. The challenge problems also nurtured research
efforts by providing large datasets for use in developing new algorithms. The Face
Recognition Technology (FERET) program, Face Recognition Grand Challenge (FRGC)
and Face Recognition Vendor Test (FRVT) evaluations and challenge problems were
instrumental in advancing face recognition technology, and they show the potential for
the evaluation and challenge problem paradigm to advance biometric, pattern recognition,
and computer vision technologies.
One of the main conclusions of the face recognition vendor test FRVT 2002 is that face
recognition from outdoor imagery remains a research challenge area.
Moreover, the primary goal of the last technology evaluation in the series (FRVT 2006)
is to look at recognition from high-resolution still images and three-dimensional (3D)
face images, and measures performance for still images taken under controlled and
uncontrolled illumination. Following are some conclusions from this evaluation, [12]:
• The FRVT 2006 results show that relaxing the illumination condition still has a
dramatic effect on performance.
• Face recognition performance on still frontal images taken under controlled
illumination has improved by an order of magnitude since the FRVT 2002. There are
three primary components to the improvement in algorithm performance since the
FRVT 2002:
1. The recognition technology,
2. Higher resolution imagery,
3. Improved quality due to greater consistency of lighting.
• Since performance was measured on the low-resolution dataset in both the FRVT
2002 and the FRVT 2006, it is possible to estimate the improvement in performance
due to algorithm design. The improvement in algorithm design resulted in an increase
in performance by a factor of between four and six depending on the algorithm. For
the results on the high and very-high resolution datasets, the improvement in
performance comes from a combination of algorithm design and image size and
quality. This is because new recognition techniques have been developed to take
advantage of the larger high quality face images.
• The FRVT 2006 and the Iris Challenge Evaluation (ICE 2006) compared recognition
performance from very-high resolution still face images, 3D face images, and single-
iris images. On the FRVT 2006 and the ICE 2006 datasets, recognition performance
6
of all three biometrics is comparable when all three biometrics are acquired under
controlled illumination.
Since the human visual system contains a very robust face recognition capability that is
excellent at recognizing familiar faces [5]. However, human face recognition capabilities
on unfamiliar faces fall far short of the capability for recognizing familiar faces. The
FRVT 2006, for the first time, integrated measuring human face recognition capability
into an evaluation. Performance of humans and computers was compared on the same set
of images. The FRVT 2006 human and computer experiment measures the ability to
recognize faces across illumination changes. This experiment found that algorithms are
capable of human performance levels, and that at false accept rates in the range of 0.05,
machines can out-perform humans.
7
each approach. In addition, the types of performance evaluations and literature
comparative studies are introduced in this chapter.
Chapter 3 surveys the two main illumination normalization approaches for face
recognition, namely model-based and image-processing-based approaches. It introduces
brief descriptions about some methods under each approach. In addition, nine
comparative studies are introduced at the end of the chapter to select the best-of-literature
approaches.
Chapter 4 introduces the detailed descriptions about the environment that we build up in
order to use it for testing our proposed illumination normalization approach and the other
approaches. The chapter includes descriptions about the selected face recognition
methods and the selected databases that cover five different face recognition variations.
The experimental results of the selected methods over each database are also introduced
in this chapter. All experiments are done without applying any illumination normalization
approach. This allows us to study the effects of any illumination normalization approach
on the selected methods over each variation separately.
Chapter 5 proposes an illumination normalization approach based on enhancing the
image resulting from histogram matching. Four different image enhancement methods
are experimentally tried in two different approaches, 1) After HM; on the resulting image
from HM, 2) Before HM; on the reference image before matching the input image on it.
The best approach is chosen such that it proves flexibility to the two selected face
recognition methods and independency to face alignment.
Chapter 6 evaluates the proposed illumination normalization approach and other best-of-
literature approaches over images with illumination variation and images with other
facial and geometrical variations using the two selected face recognition methods.
Chapter 7 contains the final conclusions of this work in addition to suggestions for
future works.
As a start for this work, the following chapter surveys the three main face recognition
approaches which are local-based, holistic-based and hybrid approaches with brief
descriptions about some methods under each approach. In addition, the types of
performance evaluations and literature comparative studies are introduced in this chapter.
8
CHAPTER 2: Face Recognition Approaches
2.1 Introduction
A number of intensity-image face recognition methods have been proposed and
implemented in commercial systems. Basically, they can be divided into holistic-based,
local-feature-based, and hybrid approaches. Even though approaches of all these types
have been successfully applied to the task of face recognition, they do have certain
advantages and disadvantages. Thus an appropriate approach should be chosen based on
the specific requirements of a given task.
Local-feature-based methods rely on the identification of certain fiducial points on the
face such as the eyes, the nose, the mouth, etc. The location of those points can be
determined and used to compute geometrical relationships between the points as well as
to analyze the surrounding region locally. Thus, independent processing of the eyes, the
nose, and other fiducial points is performed and then combined to produce recognition of
the face.
Holistic-based methods treat the image data simultaneously without attempting to
localize individual points. The face is recognized as one entity without explicitly isolating
different regions in the face. Holistic techniques utilize statistical analysis, neural
networks, and transformations. They usually require large samples of training data. The
advantage of holistic-based methods is that they utilize the face as a whole and do not
destroy any information by exclusively processing only certain fiducial points. Thus, they
generally provide more accurate recognition results. However, such techniques are
sensitive to variations in position, scale, etc., which restrict their use to standard, frontal
mug-shot images. Table-2.1 shows a brief comparison between both approaches.
Table 2.1: A brief comparison between holistic-based and local-feature-based approaches
Holistic-based Local-feature-based
Extract Feature Vector from the whole face Extract Feature Vector at certain locations
Sensitive to pose and illumination changes Robust to pose and illumination changes
Depend on accurate feature detection, which is
Feature detection is not required
not simple
Computationally less expensive Computationally more expensive
The rest of this chapter is organized as follows: section 2 contains brief description about
some local-based approaches. Section 3 describes some holistic-based approaches.
Examples of hybrid approaches appear in section 4. Section 5 describes the types of
performance evaluation and introduces results from existing comparative studies.
9
2.2 Local-Based Approaches
10
the 17 facial features is localized individually, without taking its relative positions to
other facial features into consideration. In graph adjusting stage, relative positions
between facial features are utilized to localize those misplaced facial features from the
second stage.
After facial feature detection, 17 basic facial features, each labeled with a 40-dimensional
Gabor-based complex vector, are detected for each new face image. Face recognition is
then executed on the basis of these complex vectors, which represent local features of the
areas around the multiple facial features. Two face recognition approaches, named Two-
Layer Nearest Neighbor (TLNN) and Modular Nearest Feature Line (MNFL)
respectively, are proposed.
Figure 2.1: Face bunch graph (FBG) serves as a general representation of faces. It is designed to
cover all possible variations in the appearance of faces. The FBG combines information from a
number of face graphs. Its nodes are labeled with set of jets, called bunches, and its edges are labeled
averages of distance vectors. During comparison to an image, the best fitting jet in each bunch,
indicated by gray shading, is selected independently.
11
Figure 2.2: A visualized example for the steps of automatically localizing features. In (e), the black
cross on a white background indicates an extracted and stored feature vector at this location while
white cross on black background indicates an ignored feature vector.
12
the training data for the geometrical classifier. The geometrical configuration classifier,
which is again a linear SVM, is trained on the X–Y locations and continuous outputs of
the 14 component classifiers.
Figure 2.3: System overview of the component-based face detector using four components
In training stage, the component-based face detector is first run over each image in the
training set to extract the local facial components. Only 10 out of the 14 local facial
components are kept for face recognition, removing those that either contained few gray
value structures (e.g. cheeks) or strongly overlapped with other components. Each of the
10 components is then normalized in size and their gray values are combined into a single
feature vector. These feature vectors are used to train a one-vs-all linear SVM for every
person in the database. In testing stage, the feature vector of the probe image is first
extracted and then provided to the one-vs-all linear SVM for every person in order to
recognize it. The matched person is the one that its corresponding SVM gives the
maximum value for the probe feature vector.
13
embedded in a dissimilarity space, where each image is represented by its distance to all
the other images, and a Pseudo-Fisher discriminator is built.
Figure 2.4: Sample of the normalized whole face image and the three regions that are used for the
local analysis
Verification test results on the FERET database show that the local-based algorithm
outperforms the global-FBT version. The local-FBT algorithm performed as state-of-the-
art methods under different testing conditions, indicating that the proposed system is
highly robust for expression, age, and illumination variations. In addition, the
performance of the proposed system is also evaluated under strong occlusion conditions
and found that it is highly robust for up to 50% of face occlusion. However, when the
verification system is automated completely by implementing face and eye detection
algorithms, the performance of the local approach is reduced and becomes only slightly
superior to the global approach.
14
the center pixel value and considering the result as a binary number. Then the histogram
of the labels can be used as a texture descriptor. See Fig.2.5 for an illustration of the basic
LBP operator.
15
2.3.1 Eigenspace-based Category
In Eigenspace-based category, Principal Component Analysis (PCA) – usually called
Eigenface – plays a key role in many holistic methods. Sirovich and Kirby [23] propose a
method that uses Karhunen-Loève transform to represent human faces. In 1991, Turk and
Pentland [24] develop a face recognition system using PCA (K-L expansion). Along this
direction, many Eigenspace based recognition systems have been developed, they differ
mostly in the kind of projection approaches (standard-, differential- or kernel-
Eigenspace), in the projection algorithm employed (PCA, ICA and FLD), in the use of
simple or differential images before/after projection, and in the similarity matching
criterion or classification method employed (Euclidean, Mahalanobis, Cosine distances
and SOM-Clustering, RBF, LDA, and SVM). Many comparative studies between
different Eigenspace-based methods have been established. Following are some brief
descriptions about existing Eigenspace-based methods in addition to some Eigenspace-
based comparative studies.
16
2. Subspace Linear Discriminant Analysis (LDA) 1999
The method in [25] consists of two steps: first, the face image is projected from the
original vector space to a face subspace via PCA where the subspace dimension is
carefully chosen, and then use Linear Discriminant Analysis (LDA) to obtain a linear
classifier in the subspace. The criterion that is used to choose the subspace dimension
enables the system to generate class-separable features via LDA. Fig.2.8 shows the main
steps of this method.
17
Figure 2.9: Flowchart for the Face recognition using evolutionary pursuit (EP) method
18
Figure 2.10: Image synthesis model for Architecture 1. To find a set of IC images, the images in X are
considered to be a linear combination of statistically independent basis images, S, where A is an
unknown mixing matrix. The basis images are estimated as the learned ICA output U.
Figure 2.11: Example of the projection map and the projection-combined image
19
analysis on each. All images are then projected on both intrapersonal and extrapersonal
spaces. In testing, the probe image is projected on both spaces and then the Euclidean
distances are computed between the interior projected vectors and the exterior projected
vectors of both the input image and training images in order to get the Bayesian similarity
score which used for recognition.
20
• Considering recognition rates, generalization ability as well as processing time, the
best results are obtained with the post-differential approach, using either a Bayesian
Classifier or SVM.
• In the specific case of the Yale Face Database, where the requirements are not very
high, any of the compared approaches gives rather similar results. Thanks to their
simplicity, Eigenfaces or Fisherfaces are probably the best alternatives.
• Although kernel methods obtain the best recognition rates, they suffer from problems
such as low processing speed and the difficulty to adjust the kernel parameters.
21
Figure 2.12: 3-level wavelet decomposition
Figure 2.13: (a) input image, (b) the log-magnitude of its DCT, (c) the scanning strategy of
coefficients
22
4. Face Recognition in Fourier Space 2000
The work in [42] describes a simple face recognition system based on an analysis of faces
via their Fourier spectra. The feature vectors are constructed by taking the Fourier
coefficients at selected frequencies as shown in Fig.2.14. Recognition is done by finding
the closest match between feature vectors using the Euclidian distance classifier.
Figure 2.14: Most variant frequencies: a) real, b) imaginary and c) selected numbering
23
Figure 2.15: Spectroface representation steps
Figure 2.16: Examples of FBT of (A) an 8 radial cycles image, (B) a 4 angular cycles image and (C)
an image of the average of these images. The magnitude of the FBT coefficients is presented in
colored levels (red indicates the highest value)
24
Figure 2.17: Block diagram for face recognition based on moments
25
4. Face Recognition using a New Texture Representation of Face Images
2003
Authors of [51] present a new texture representation of face image using a robust feature
from the Trace transform. The masked Trace transform (MTT) offers texture information
for face representation which is used to reduce the within-class variance by masking out
the background and non pure-face information. Fig.2.18 shows an example of the trace
transform on a full face image and its masked version. The method starts with
transforming the image space to the Trace transform space to produce the MTT.
Weighted Trace transform (WTT) is then calculated which identifies the tracing lines of
the MTT that produce similar values irrespective of intra-class variations. Finally, a new
distance measure is proposed by incorporating the WTT for measuring the dissimilarity
between reference and test images.
Figure 2.18: Examples of the Trace transform on (a) full image (b) masked with rectangular shape
and (c) masked with elliptical shape.
26
2.4 Hybrid Approaches
Figure 2.19: Training and recognition stages of Face Recognition Using Local and Global Features
approach
27
to be used as a facial feature that is robust to facial expressions and occlusions (especially
when the lower part of the face is fully occluded).
Figure 2.20 (a) A gray-scale face image, (b) it’s edginess image, and (c) the cropped eyes.
Next, the facial features are encoded to lower-dimensional feature spaces using the
principal component analysis (PCA) in conjunction with Fisher’s Linear Discriminant
(FLD). Three individual spaces are constructed corresponding to the three facial features.
The distance-in-feature-space (DIFS) values are calculated for all the images in the
training set and in each of the feature spaces. These values are used to compute the
distributions of the DIFS values. Given a new test image, the three facial features are first
extracted and their DIFS values are computed in each feature space. Each feature
provides an opinion on the claim in terms of a confidence value. The confidence values
of all the three features are fused for final recognition. The identity established by the
proposed fusion technique is more reliable compared to the case when features are used
individually.
28
evaluations. This evaluation class includes the FERET (face recognition technology)
series of face recognition evaluations and the FRVT (face recognition vendor test) series
[114].
Scenario evaluations aim to evaluate the overall capabilities of the entire system for a
specific application. In face recognition, a technology evaluation would study the face
recognition algorithms only but the scenario evaluation studies the entire system,
including camera and camera-algorithm interface, for a specific application. An example
is face recognition systems that verify the identity of a person entering a secure room.
Each tested system would normally have its own acquisition sensor and would thus
receive slightly different data. Scenario evaluations are not always completely repeatable
for this reason, but the approach used can always be completely repeatable. Scenario
evaluations typically take a few weeks to complete because multiple trials, and for some
scenario evaluations, multiple trials of multiple subjects/areas, must be completed.
Results from a scenario evaluation typically show areas that require additional system
integration, as well as provide performance data on systems for a specific application. An
example of the scenario evaluation is the UK Biometric Product Testing [56].
At first glance, an operational evaluation appears very similar to a scenario evaluation,
except that the test is at the actual site and uses actual subjects. Rather than testing for
performance, however, operational evaluations aim to study the workflow impact of
specific systems installed for a specific purpose. Operational evaluations are not very
repeatable unless the actual operational environment naturally creates repeatable data.
Operational evaluations typically last from several weeks to several months. The
evaluation team must first examine workflow performance prior to technology insertion,
and again after users are familiar with the technology. Accurate analysis of the benefit of
the new technology requires a comparison of the workflow performance before and after
the technology insertion.
In an ideal three-step evaluation process, technology evaluations are performed on all
applicable technologies that could conceivably meet requirements. The technical
community will use the results to plan future research and development (R&D) activities,
while potential end-users will use the results to select promising systems for application-
specific scenario evaluations. Results from the scenario evaluation will enable end-users
to find the best system for their specific application and have a good understanding of
how it will operate at the proposed location. This performance data, combined with
workflow impact data from subsequent operational evaluations, will enable decision
makers to develop a solid business case for a large-scale installation.
29
2.5.2 Comparative Studies
Many comparative studies have been established in the last 10 years due to increasing in
the number of available algorithms and techniques. These comparative studies usually try
to evaluate two or more face recognition algorithms using one or more small to medium
size databases. They can be classified into two categories according to the nature of
databases they work on:
1. Comparative studies using global database(s) usually contain more than one
variation to determine which algorithm is better over these databases, as in [40],
[57-60].
2. Comparative studies for specific variation(s) using suitable database(s) each with
one variation only to study the algorithms over each variation separately, as in
[21], [39], [61-65].
Under the first category, [57] evaluate three holistic-based algorithms and a local one
using seven different classifiers over FERET database with variations in illumination and
aging. In [40], seven Eigenspace-based algorithms with five similarity matching criteria
are compared over two databases, Yale with variations in illumination and expressions
and FERET. Five algorithms based on local binary patterns are compared with a holistic
and a local algorithm in [59] over two databases, BANCA with complex background and
difficult lightning conditions and XM2VTS with uniform background. In [60], two
holistic algorithms and a local one are compared over four different databases, each with
two or more variations.
In the second category, only the pose variation is considered in [61] and [62]. In [61], two
holistic-based algorithms are compared using two databases (ALAN and UMIST). In
[62], five holistic-based algorithms are compared using one database (FERET). In [39]
and [21], expressions, illumination and aging variations are tested separately using
FERET database over three holistic-based algorithms in [39] and two holistic-based and
two local-based algorithms in [21]. Non-uniform illumination variation is considered in
[63] in which five holistic-based algorithms are compared over two different databases,
CMU-PIE and YALE B each with illumination variations only.
As the main aim of this thesis is to propose an illumination normalization approach, the
next chapter will discuss the different illumination normalization approaches in the
literature. Also, nine comparative studies are introduced at the end of the chapter to select
the best-of-literature approaches.
30
CHAPTER 3: Illumination Normalization Approaches
3.1 Introduction
Although many face recognition techniques and systems have been proposed in the last
20 years, evaluations of the state-of-the-art techniques and systems have shown that
recognition performance of most current technologies degrades due to the variations of
illumination [78], [12]. In the last face recognition vendor test FRVT 2006 [12], they
conclude that relaxing the illumination condition has a dramatic effect on the
performance. Moreover, It has been proven both experimentally [84] and theoretically
[14] that the variations between the images of the same face due to illumination are
almost always larger than image variations due to change in face identity. As is evident in
Fig.3.1, the same subject, with the same facial expression, can appear strikingly different
when light source direction and viewpoint vary [79].
Figure 3.1: The same individual imaged with the same camera and the same facial expression may
appear dramatically different with changes in the lighting conditions.
There has been much work dealing with illumination variations in face recognition.
Generally, these approaches can be classified into two categories: model-based and
image-processing-based approaches.
Model-based approaches derive a model of an individual face, which will account for
variation in lighting conditions. Examples of this approach include spherical harmonics
representation [81], Eigen Light-Fields [82], illumination cone [64], Quotient Image
[100], Self Quotient Image [107] and Retinex algorithms [89], [90], [66]. Though the
model-based approaches are perfect in theory, they require a training set with several
different lighting conditions for the same subject, which can be considered as a weakness
for realistic applications. Although some work has been done to enlarge a small learning
set by virtually re-imaging the input face image as in [83], [96], the requirements of
additional constraints or assumptions in addition to the highly computational cost make
the model-based approaches unsuitable for realistic applications [80], [71].
Image-processing-based approaches attempt to normalize the variation in appearance due
to illumination, either by image transformations or by synthesizing a new image from the
given image in some normalized form. Recognition is then performed using this
31
canonical form. Examples of this approach include histogram equalization/matching
HE/HM [71], gamma intensity correction GIC [84], local binary patterns LBP [80] and
local normal distribution LNORM [72]. Compared to the model-based approach,
preprocessing has two main advantages: it is completely stand-alone and thus can be used
with any classifier. Moreover, it transforms images directly without any training images,
assumptions or prior knowledge. Therefore, they are more commonly used in practical
systems for their simplicity and efficiency.
The rest of this chapter is organized as follows: sections 2 and 3 contain the description
of some model-based and image-processing-based approaches, respectively. Section 4
contains the results of existing comparative studies among different illumination
normalization approaches focusing on the best approaches of each comparison and then
concludes the best-of-literature approaches from these studies.
32
Where s0 (point light source) be a pre-defined canonical lighting condition.
Obviously, the Quotient Illumination is completely independent of the surface
reflectance (albedo), and depends only on the variance of the lighting condition
from the pre-defined canonical lighting one (considering all the shapes are
assumed to be the same). Thus, Quotient Illumination can be computed easily by
calculating the quotient between the images of the object i of the ideal class of
objects as follows:
ρ i ( x, y )n( x, y )T .s j I ij ( x, y )
R j ( x, y ) = = (3.4)
ρ i ( x, y )n( x, y ) T .s0 I i 0 ( x, y )
where Iij is the image of object i captured under the j-th lighting condition, and Ii0
is the image of the same object i captured under the canonical lighting condition.
3. Quotient illumination bootstrap set: since faces are not strictly ideal class of
objects as the 3D shapes of faces are different despite their approximate
similarity. Therefore, a Quotient illumination bootstrap set needs to be
constructed. It consists of a set of pairs of face images captured under some non-
canonical lighting condition and under the pre-defined canonical lighting
condition, as follows:
{(I ij
, I i 0 ) | i = 1,2,..., N ; j = 1,2,..., L} (3.5)
where N is number of persons and L is number of non-canonical lighting
conditions in the system. Given such a bootstrap set, Quotient Illumination can be
statistically modeled or computed simply as the mean over all the faces in the set,
for instance:
1 N I ij ( x, y )
R j ( x, y ) =
N
∑I
i =1 ( x, y )
, j = 1,2,..., L (3.6)
i0
33
(a) Illuminated Image (b) QIR image
Figure 3.2: Effect of applying QIR on an illuminated face image from Yale B database
a) Yale B
b) CMU PIE
Figure 3.3: Effect of applying SQI approach to illuminated face images from Yale B and CMU PIE
databases.
The only processing needed for SQI is smoothing filtering. A weighed Gaussian filter is
designed for anisotropic smoothing according to the following equation:
F = G *W (3.9)
34
where W is the weight and G is the Gaussian kernel, and N is the normalization factor for
which:
1
N Ω
∑WG = 1 (3.10)
where Ω is the convolution kernel size. The convolution region is divided into two sub-
regions M1 and M2 with respect to a threshold τ. Where M1 has more pixels than M2, τ is
calculated by:
τ = Mean ( I Ω ) (3.11)
For the two sub-regions, W has corresponding value:
⎧0 I (i, j ) ∈ M 2
W (i, j ) = ⎨ (3.12)
⎩1 I (i, j ) ∈ M 1
If the convolution image region is smooth, i.e. little gray value variation (non-edge
region), there is also little difference between the smoothing the whole region and part of
the region. If there is large gray value variation in convolution region, i.e. edge region,
the threshold can divide the convolution region into two parts M1 and M2 along the edge
and the filter kernel will convolute only with the large part M1, which contains more
pixels. Therefore the halo effects can be significantly reduced by the weighted Gaussian
kernel.
The essence of this anisotropic filter is that it smoothes only the main part of convolution
region (i.e. only one side of edge region in case of step edge region).
The division operation in the SQI may magnify high frequent noise especially in low
signal noise ratio regions, such as in shadows. To reduce noise in Q, a nonlinear
transformation function is used to transform Q into D,
D = T (Q) (3.13)
where T is a nonlinear transform which may be Log, Arctangent or Sigmoid nonlinear
function.
The implementation of SQI approach is summarized below:
1. Select several smoothing kernel G1, G2, …, Gn and calculate corresponding
weights W1, W2, …, Wn according to image I, and then smooth I by each weighted
anisotropic filter WGi.
1
Iˆk = I * WGk , k = 1,2,..., n (3.14)
N
Calculate self-quotient image (SQI) between input image I and its smoothing
version
I
Qk = , k = 1,2,..., n (3.15)
ˆI
k
35
2. Transfer self-quotient image (SQI) with nonlinear function
Dk = T (Qk ), k = 1,2,..., n (3.16)
3. Summarize nonlinear transferred results
n
Q = ∑ mk Dk , k = 1,2,..., n (3.17)
k =1
The m1, m2,… mn are the weights for each scale of filter and are set to one in
experiments of [107].
36
where σ is the standard deviation of the filter and controls the amount of spatial detail that
is retained, and κ is a normalization factor that keeps the area under the Gaussian curve
equal to one.
The standard deviation of the Gaussian σ is referred to as the scale of the SSR. A small
value of σ provides very good dynamic range compression but at the cost of poorer color
rendition, causing graying of the image in uniform areas of color. Conversely, a large
scale provides better color rendition but at the cost of dynamic range compression [89].
Since face recognition is conventionally performed on grey-scale images, the loss of
color is out of concern here. Moreover, the dynamic range compression gained by small
scales is the essence of the illumination normalization process proposed in [66]. All the
shadowed regions are grayed out to a uniform color, eliminating soft shadows and
specularities and hence creating an illumination invariant signature of the original image.
Fig.3.4 illustrates the effect of Retinex processing on a facial image, I, for different
values of σ. As σ increases, the normalized image IN, contains reduced graying and lesser
loss of intensity values, as seen in Fig.3.4 (c) and (d). However, for larger values of σ, the
shadow is still visible. On the other hand, with σ = 6 in Fig.3.4 (b), the resulting image
has grayed out the shadow region to blend in with the rest of the face.
4. GROSS Method
The work in [85] proposes an illumination normalization approach, we call it GROSS
approach, for illumination invariant face recognition. Same as Retinex algorithms,
GROSS is based on the reflectance illumination model rather than the Lambertian model,
see equation 3.18.
The GROSS approach is motivated by two widely accepted assumptions about human
vision:
1. Human vision is mostly sensitive to scene reflectance and mostly insensitive to
the illumination conditions.
37
2. Human vision responds to local changes in contrast rather than to global
brightness levels.
Having these assumptions, the goal is to find an estimate of L(x, y) such that when it
divides I(x, y) it produces R(x, y) in which the local contrast is appropriately enhanced.
(a) Illuminated Image, I (b) IR with σ = 6 (d) Well-lit Image, Î (e) ÎR with σ = 6
Ω Ω
(3.22)
where the first term drives the solution to follow the perception gain model, while the
second term imposes a smoothness constraint. Here Ω refers to the image. The parameter
38
λ controls the relative importance of the two terms. The space varying permeability
weight ρ(x, y) controls the anisotropic nature of the smoothing constraint.
The Euler-Lagrange equation for this calculus of variation problem yields:
λ
L + (Lxx + L yy ) = I (3.23)
ρ
Discretized on a rectangular lattice, this linear partial differential equation (PDE)
becomes:
⎡ ⎤
⎢
Li , j + λ ⎢
1
(Li , j − Li , j−1 ) + 1
(Li , j − Li , j+1 ) + 1
(Li , j − Li−1, j ) + 1
(Li , j − Li+1, j )⎥⎥ = I (3.24)
hρ 1 hρ 1 hρ 1 hρ 1
⎢⎣ i , j − 2 i, j+
2
i− , j
2
i+ , j
2
⎥⎦
where h is the pixel grid size and the value of each ρ is taken in the middle of the edge
between the center pixel and each of the corresponding neighbors (see Fig.3.6). In this
formulation, ρ controls the anisotropic nature of the smoothing by modulating
permeability between pixel neighbors. Equation 3.24 can be solved numerically using
multi-grid methods for boundary value problems [108] in order O(N).
Fig.3.7 shows the effect of applying GROSS approach on some illuminated face images
from Yale B database.
39
Note that the GROSS approach does not require any training steps, knowledge of 3D face
models or reflective surface models. Only a single parameter, which meaning is intuitive
and simple to understand, needs to be adjusted by the user.
Figure 3.7: Effect of applying GROSS approach on some illuminated face images from Yale B
database
40
size M × N with G gray levels and cumulative histogram H(g), the transfer function at
certain level T(g) is given as follows:
H ( g ) × (G − 1)
T (g) = (3.26)
M ×N
Fig.3.8 shows the effect of applying histogram equalization to an illuminated image. The
illuminated image and its histogram are shown in Fig.3.8 (a), (b) while the equalized
image and its corresponding histogram are shown in Fig.3.8 (c), (d).
Illuminated
Image
(a) (b)
Equalized
Image
(c) (d)
Figure 3.8: Effect of applying histogram equalization on an illuminated image
41
Illuminated
Image
(a) (b)
Well-lit
Image
(c) (d)
Resulting
Image
(e) (f)
Figure 3.9: Histogram matching process to an illuminated image
To explain the algorithm, Let H(i) be the histogram function of an illuminated image X,
and G(i) be the desired histogram of the well-lit image Y, we wish to map H(i) to G(i) via
a transformation FHÆG(i). We first compute a transformation function for both H(i) and
G(i) that will map the histogram to a uniform distribution, U(i). These functions are
FHÆU(i) and FGÆU(i), respectively. Equations 3.27 and 3.28 depict the mapping to a
uniform distribution, which is also known as histogram equalization [74].
i
∑ H ( j)
j =0
FH → U(i) = N −1 (3.27)
∑ H ( j)
j =0
∑ G( j)
j =0
FG → U(i) N −1 (3.28)
∑ G( j)
j =0
Where N is the number of discrete intensity levels. N = 256 for 8-bit grayscale images.
42
To find the mapping function, FHÆG(i), the function FGÆU(i) is inverted to obtain FUÆG(i).
Since the domain and the range of the functions of this form are identical, the inverse
mapping is trivial and is found by cycling through all values of the function. However,
due to the discrete nature of these functions, inverting can yield a function which is
undefined for certain values. Thus, linear interpolation is used and assumes smoothness
to fill undefined points of the inverse function according to the values of well-defined
points in the function. As a result, a fully defined mapping FUÆG(i) is generated which
transforms a uniform histogram distribution to the distribution found in histogram G(i).
The mapping FHÆG(i) can then be defined as in equation 3.29, [66].
FH →G ( i ) = FU →G ( FH →U ( i ) ) (3.29)
It’s common in literature to match all images, in both training and testing sets, with a
single histogram of either a fixed well-lit image as in [71], [67] or an average image as in
[72].
Figure 3.10: Transformation functions of LOG and GAMMA (L: number of gray levels)
43
LOG
Illuminated Resulting
Image Image
γ (3.32)
x, y
G (I xy ; γ ) = c.I xy
γ (3.33)
is the Gamma transform; c is a gray stretch parameter used to linearly scaling the result to
be in the range of [0, 255], and γ is the Gamma coefficient.
From equations 3.32 and 3.33, intuitively, the GIC is expected to make the overall
brightness of the input images best fit that of the pre-defined normal face images. Thus,
its intuitive effect is that the overall brightness of all the processed face images is
44
adjusted to the same level as that of the common normal face I0. See Fig.3.12 for its
intuitive effect.
GIC
NORM
Illuminated Resulting
Image Image
Figure 3.13: Effect of applying NORM approach to an illuminated image. (Note that the gray-level of
the resulting image is stretched to [0,255] for displaying purpose only)
45
There are three local normalization methods proposed by [72] which are: Local
Histogram Equalization (LHE), Local Histogram Matching (LHM) and Local Normal
Distribution (LNORM). They are the same as their global counterparts described in
section 3.3.1 but applied locally. Applying a function locally mean the following; take a
window from the image, starting in the up left corner, with a window size considerably
smaller than the image size. The global normalization function is applied to the
windowed image. This process is repeated by moving the window pixel by pixel all over
the image and for each one applying the normalization function. Because the windows
overlap, the final pixel value is the average of all the results for that particular pixel.
Fig.3.14 shows the effect of applying the three local normalization methods, LHE, LHM
and LNORM to an illuminated face image.
Figure 3.14: Effects of applying the three local normalization methods to an illuminated face image
46
Figure 3.16: The four regions for illumination normalization
After the coarse partition of the face regions, HE or GIC can be conducted in the four
regions separately. Hereafter, the region-based HE is abbreviated to RHE, and the region-
based GIC to RGIC. The effects of the RHE and RGIC can be seen from Fig.3.17.
(a) Illuminated Image (b) RHE (c) RGIC (d) RGIC + RHE
Figure 3.17: The effects of applying region-based strategy of HE and GIC over the four face regions
Figure 3.18: Block histogram matching. In each image pair, the left one is the input image while the
right one is the reference image.
47
The histogram-matching block image intensity values are scaled with a windowing filter
H which is as defined below and is pictorially shown in Fig.3.19:
BO (n, m ) = BO (n, m )H (n, m ),1 ≤ n, m ≤ M (3.35)
where
⎧ 4nm M
⎪ 2
, 1 ≤ n, m ≤ ,
M 2
⎪ 4m(M − n + 1) M M
⎪ , < n ≤ M ,1 ≤ m ≤ ,
⎪
H (n, m ) = ⎨
2
M 2 2
4n(M − m + 1) M M (3.36)
⎪ , 1≤ n ≤ , < m ≤ M,
⎪
2
M 2 2
⎪ 4(M − n + 1)(M − m + 1) M
⎪⎩ 2
, < n, m ≤ M .
M 2
48
(a) (b)
Figure 3.20: Images before and after intensity normalization with BHM. (a) Input images, (b)
corresponding output images after applying BHM
The decimal form of the resulting 8-bit word (LBP code) can be expressed as follows:
7
LBP(xc , yc ) = ∑ s(in − ic )2 n (3.37)
n =0
where ic corresponds to the grey value of the center pixel (xc, yc), in to the grey values of
the 8 surrounding pixels, and function s(x) is defined as:
⎧1, x ≥ 0,
s(x ) = ⎨ (3.38)
⎩0, x < 0.
By definition, the LBP operator is unaffected by any monotonic gray-scale
transformation which preserves the pixel intensity order in a local neighborhood. Note
that each bit of the LBP code has the same significance level and that two successive bit
49
values may have a totally different meaning. Actually, The LBP code may be interpreted
as a kernel structure index.
Later, Ojala et al. [87] extended their original LBP operator to a circular neighborhood of
different radius size. Their LBPP,R notation refers to P equally spaced pixels on a circle of
radius R. In [80], the LBP8;2 operator illustrated in Fig.3.22 is used to preprocess the input
image before providing it to the face authentication algorithms: the face is represented
with its texture patterns given by the LBP operator at each pixel location as shown in
Fig.3.23.
Figure 3.22: The extended LBP operator with (8,2) neighborhood. Pixel values are interpolated for
points which are not in the center of a pixel.
LBP operator
Figure 3.23: Original image (left) processed by the LBP operator (right).
The conducted experiments in [80] show that the LBP operator provides a texture
representation of the face which improves the performances of two different face
authentication classifiers (PCA-LDA and HMM) as compared to histogram equalization.
Moreover, obtained results are comparable when using the GROSS algorithm proposed in
[85] and described previously in section 3.2, on the same databases [88] and removes the
need for parameter selection.
5. Laplacian of Gaussian
In order to remove the influence caused by illumination variations, [91] apply image
processing which is a combination of histogram equalization, Laplacian of Gaussian filter
and contrast adjustment. Histogram equalization is applied first to enhance biased
contrast image in which some pixels are concentrated on a narrow range of the pixel
intensity.
In order to remove the information of pixel intensity while reserving local features that
are useful to recognition, [91] apply Laplacian of Gaussian (LoG) filter following the
histogram equalization. The equation of the 2D LoG function centered on zero and with
Gaussian standard deviation σ has the form:
50
1 ⎡ x 2 + y 2 ⎤ − x2σ+ y
2 2
LoG (x, y ) = − 4 ⎢1 −
2σ 2 ⎥⎦ e
2
(3.39)
πσ ⎣
Fig.3.24 shows the effects of the three steps of this approach. As shown in Fig.3.24 (c),
local features of each face are reserved and the influence of lighting is almost removed.
However, contrast of processed image is biased to a certain range. In order to improve
this contrast and emphasize the local features, contrast adjustment is applied. The final
processed image is shown in Fig.3.24 (d).
Figure 3.24: Effects of applying the image processing steps proposed by [91]
51
≈ 2 typically gives the best results, but values up to about 4 are not too damaging and
may be preferable for datasets with less extreme lighting variations.
Masking. If a mask is needed to suppress facial regions that are felt to be irrelevant or
too variable, it should be applied at this point. Otherwise, either strong artificial gray-
level edges are introduced into the convolution, or invisible regions are taken into
account during contrast equalization.
Contrast Equalization. The final step of the CHAIN approach is to globally rescale the
image intensities to standardize a robust measure of overall contrast or intensity variation.
It is important to use a robust estimator because the signal typically still contains a small
admixture of extreme values produced by highlights, garbage at the image borders and
small dark regions such as nostrils. A simple and rapid approximation based on a two
stage process is applied to accomplish this:
I ( x, y )
I ( x, y ) =
[ ((
mean I (x′, y ′)
a
1
a
) )] (3.40)
I ( x, y )
I ( x, y ) =
[mean(min(τ , I (x′, y′) ) )] a
1
a (3.41)
Here, a is a strongly compressive exponent that reduces the influence of large values, τ is
a threshold used to truncate large values after the first phase of normalization, and the
mean is over the whole (unmasked part of the) image.
The resulting image is now well scaled but it can still contain extreme values. To reduce
their influence on subsequent stages of processing, a nonlinear function is finally applied
to compress over-large values. In [109], the hyperbolic tangent I(x, y) = τ tanh(I(x, y)/τ),
is used thus limiting I to the range (−τ, τ).
In [109], the default settings of the various parameters of CHAIN approach are
summarized in Table-3.1. Moreover, it’s found that the CHAIN approach gives similar
results over a broad range of parameter settings, which greatly facilitates the selection of
parameters. Fig.3.25 shows the effect of applying CHAIN approach on various
illuminated faces from Extended Yale B database.
Table 3.1: Default parameter settings for CHAIN approach
Procedure Parameter Value
Gamma Correction γ 0.2
σ0 1
DoG Filtering
σ1 2
Contrast α 0.1
Equalization τ 10
52
(a) Illuminated face images (b) After applying CHAIN approach
Figure 3.25: Examples of images of one person from the Extended Yale-B frontal database. The
columns respectively give images from subsets 1 to 5.
1. Study 1
The study in [71] empirically compares five image-processing-based approaches for
illumination insensitive face recognition, which are:
1. Histogram Equalization (HE)
2. Histogram Matching (HM)
3. Logarithmic Transform (LOG)
4. Gamma intensity correction (GIC)
5. Self-Quotient Image (SQI)
These approaches are compared on the CMU-PIE database [102], the FERET database
[120] and the CAS-PEAL database [103]. The PCA followed by LDA approach is used
for face recognition.
The results on the lighting subsets of the three databases show that HM gives the best
results among the four other approaches over FERET and CAS-PEAL, while comes after
GIC over CMU-PIE.
2. Study 2
The work in [84] proposes the following three illumination normalization approaches:
1. Gamma Intensity Correction (GIC) method.
2. Region-based strategy combining GIC and the Histogram Equalization (HE).
53
3. Quotient Illumination Relighting (QIR) method.
Experiments are then conducted to compare the following approaches:
1. HE: Histogram equalization globally over the images.
2. RHE: Region-based Histogram equalization.
3. GIC: Gamma Intensity Correction globally.
4. RGIC: Region-based GIC.
5. GIC+RHE: perform RHE after GIC.
6. RGIC+RHE: perform RHE after RGIC.
7. RHE+RGIC: perform RGIC after RHE.
8. HE+RGIC: perform RGIC after HE.
9. QIR: Quotient Illumination Relighting.
The above approaches are empirically compared on the Yale B database [64] and
Harvard database [106]. The simplest normalized correlation, i.e., cosine of the angle
between two image vectors, is exploited as the distance measurement. And for all
experiments, classification is performed using the nearest neighbor classifier.
The results show that the proposed QIR approach gives the best results on both databases.
However, the terrific performance of QIR is based on the assumption that the lighting
modes of the images are known or can be estimated. This is a strong constraint in a
practical application system. In contrast, the RHE combined with RGIC methods are
more general and practical to be exploited in a recognition system efficiently, since they
need not the illumination estimation procedure.
3. Study 3
The work in [96] empirically compares the following 12 approaches for illumination
insensitive face recognition:
1. Correlation method [93].
2. Eigenface method [24].
3. Eigenface method without the first three principle components [32].
4. Nearest Neighbor using 9 training images per subject [93].
5. Linear Subspace [94].
6. Gradient angles [97]
7. Cones-attached [95].
8. Cones-cast [95].
9. Harmonic images (no cast shadow) [96].
10. Harmonic images-cast (with cast shadows) [96].
11. Nine point of lights (9PL) using simulated images [96].
12. Nine point of lights (9PL) using real images [96].
The above approaches are empirically compared on the Yale B database [64]. In all the
experiments on the last four methods, the actual recognition algorithm is straightforward.
54
For each test image, the usual L2 (Euclidean distance) is computed between the image
and all the subspaces. The identity associated with the subspace that gives the minimal
distance to the image is declared to be its identity.
The results show that only two out of the 12 approaches give 100% recognition rate, they
are the Cones-cast and the Nine point of lights (9PL) using real images.
4. Study 4
The work in [85] introduces a simple and automatic image-processing-based approach for
illumination normalization in face recognition, we call it GROSS approach. It empirically
compares the proposed approach with two other approaches which are Histogram
Equalization (HE) and Gamma Correction (GAMMA).
These approaches are compared on the Yale B database [64] and the CMU PIE database
[102]. In all experiments, the recognition accuracies are reported for two algorithms:
Eigenfaces (Principal Component Analysis (PCA)) and FaceIt [104], a commercial face
recognition system from Identix.
The results show the superiority of the proposed approach (GROSS) over the HE and
GAMMA.
Fig.3.26 gives a summarization about the above four studies showing the best approaches
from each study.
Figure 3.26: Summarization for the first four comparative studies. For each study, it shows the
normalization approach to be compared, the face databases and the face recognition approaches in
addition to the best normalization approaches from each study (grayed boxes)
55
5. Study 5
The work in [66] proposes an illumination normalization approach based on applying the
Single Scale Retinex (SSR) followed by histogram matching (HM) to bring all the
images to the same dynamic range of intensity, we notice it by SSRÆHM. It then
compares the proposed approach with the histogram matching (HM) approach for
illumination insensitive face recognition.
The Yale B face database [64] is used for face recognition experiments. Support Vector
Machine (SVM) is used as the learning scheme [92] for the face recognition experiments.
The results show that using the proposed approach, SSRÆHM, as an illumination
normalization approach gives better recognition rates than using the HM alone.
6. Study 6
The study in [72] empirically compares the following seven image-processing-based
approaches, four global and three local, for illumination insensitive face recognition:
1. Gamma Intensity Correction (GIC).
2. Histogram Equalization (HE).
3. Histogram Matching (HM).
4. Normal Distribution (NORM).
5. Local Histogram Equalization (LHE).
6. Local Histogram Matching (LHM).
7. Local Normal Distribution (LNORM).
These approaches are compared on the Extended Yale B database [96] and the Yale B
database [64]. In all experiments, the nearest neighbor classifier using the Euclidean
distance between the images is used for recognition.
The results of the first experiment on the Extended Yale B database show that LNORM
gives the best results among the six other approaches.
The second experiment on the Yale B database is performed to be able to compare the
results of the LNORM with results found in the literature. When using LNORM with
window size 5 × 5 and training with five randomly images from Subset 1, the results in
[72] show that:
1. LNORM outperforms all the nine illumination normalization approaches appear
in Study 2 [84].
2. LNORM performs better than 10 approaches from Study 3 [96], while it gives
comparable results to the Study 3’s best two approaches which are Cones-cast and
the Nine point of lights (9PL) using real images.
3. LNORM perform better than Harmonic Image Exemplars approach [98].
As a result, we can find that LNORM is better than a total of 24 approaches. These
approaches are listed below in Table-3.2 in conjunction with the corresponding Study of
each approach.
56
Table 3.2: List for 24 illumination normalization approaches that LNORM perform better than them
I Approach Study No.
1 HE: Histogram equalization globally over the images Study 2
2 RHE: Region-based Histogram equalization Study 2
3 GIC: Gamma Intensity Correction globally Study 2
4 RGIC: Region-based GIC Study 2
5 GIC+RHE: perform RHE after GIC Study 2
6 RGIC+RHE: perform RHE after RGIC Study 2
7 RHE+RGIC: perform RGIC after RHE Study 2
8 HE+RGIC: perform RGIC after HE Study 2
9 QIR: Quotient Illumination Relighting Study 2
10 Correlation method [93] Study 3
11 Eigenface method [24] Study 3
12 Eigenface without the first three principle components [32] Study 3
13 Nearest Neighbor using 9 training images per subject [93] Study 3
14 Linear Subspace [94] Study 3
15 Gradient angles [97] Study 3
16 Cones-attached [95] Study 3
17 Harmonic images (no cast shadow) [96] Study 3
18 Harmonic images-cast (with cast shadows) [96] Study 3
19 Nine point of lights (9PL) using simulated images [96] Study 3
20 Harmonic Image Exemplars [98] Study 6
21 Histogram Matching (HM) Study 6
22 Normal Distribution (NORM) Study 6
23 Local Histogram Equalization (LHE) Study 6
24 Local Histogram Matching (LHM) Study 6
7. Study 7
The work in [105] introduces the Logarithmic Total Variation (LTV) model as a
preprocessing technique for face recognition under varying illumination. The proposed
approach is empirically compared with the following four approaches:
1. Quotient Image (QI) [100].
2. Quotient Illumination Relighting (QIR) proposed in Study 2 [84].
3. Self Quotient Image (SQI) [101].
4. Histogram Equalization (HE).
These approaches are compared on Yale B face database [64] and the CMU PIE database
[102]. Then an outdoor database [105] is used for evaluating the performance under
natural lighting condition. Two different face recognition approaches are used for
evaluation, template matching and PCA.
The results show that the proposed approach (LTV) always gives the best results among
the four other normalization approaches.
57
Moreover, the results on Yale B database show that the LTV gives 100% recognition rate
over all the five subsets which is better than the Harmonic Image Exemplar [98] and
comparable (or may better) to the 9PL (real images) and Cones-cast from Study 3 [96]
that give the same recognition rate but on four subsets only (no results are reported
wherever, for the tests on subset 5).
In addition, the proposed approach (LTV) reached similar results to the ones obtained by
Corefaces [99] based on PCA recognition on Yale B and CMU PIE database.
As a result, we can find that LTV is better than the following four normalization
approaches:
1. Quotient Image (QI) [100].
2. Quotient Illumination Relighting (QIR) proposed in Study 2 [84].
3. Self Quotient Image (SQI) [101].
4. Histogram Equalization (HE).
While gives comparable (or may better) results to the following three normalization
approaches:
1. Nine Point of Lights (9PL) form Study 3 [96].
2. Cones-cast from Study 3 [96].
3. Corefaces [99].
8. Study 8
The work in [80] proposes an original preprocessing technique based on Local Binary
Pattern (LBP) for illumination robust face authentication. It empirically compares the
proposed approach with two other approaches which are GROSS approach from Study 4
[85] and Histogram Equalization (HE).
The efficiency of the proposed approach is empirically demonstrated using both an
appearance-based (LDA) and a feature-based (HMM) face authentication systems on two
databases: BANCA and XM2VTS (with its darkened set).
Conducted experiments show that the proposed preprocessing approach (LBP) is suitable
for face authentication: results are comparable with or even better than those obtained
using the GROSS approach proposed in Study 4 [85], while it removes the need for
parameter selection.
9. Study 9
The work in [109] presents a simple and efficient preprocessing chain (CHAIN),
described previously in section 3.3.2, that eliminates most of the effects of changing
illumination while still preserving the essential appearance details that are needed for
recognition. It empirically compares the proposed approach with the following
approaches:
58
1. Histogram Equalization (HE).
2. Multi Scale Retinex (MSR) [89].
3. Self Quotient Image (SQI) [101].
4. Logarithmic Total Variation (LTV) proposed in Study 7 [105].
5. GROSS approach proposed in Study 4 [85].
These approaches are empirically compared over three different databases, namely Face
Recognition Grand Challenge version 1 experiment 4 (FRGC-104) [110], Extended Yale
B [96], and CMU PIE [102]. The efficiency of the proposed approach is empirically
demonstrated using Local Ternary Patterns (LTP) [109], a generalization of the Local
Binary Pattern (LBP), combined with a local distance transform (DT) based similarity
metric as a classifier [111].
The results show that the proposed CHAIN approach outperforms all the five approaches
over the three databases. However, the LTV approach gives recognition rates marginally
less than the CHAIN on Extended Yale B database and equal to it on CMU PIE database
but about 300 times slower than CHAIN approach [109].
Fig.3.27 completes the summarization appear previously in Fig.3.26 about all the above
nine comparative studies showing some relations between these studies in addition to the
best approaches resulting from all studies.
We can conclude from these nine comparative studies that among 38 different
illumination-normalization approaches appear in Table-3.3, the following 7 approaches
are the best for face recognition:
1. Single Scale Retinex followed by Histogram Matching (SSRÆHM) [66].
2. Local Normal Distribution (LNORM) [72].
3. Preprocessing Chain Approach (CHAIN) [109].
4. Nine Point of Lights with real images (9PL) [96].
5. Cones-cast [95].
6. Corefaces [99].
7. Local Binary Patterns (LBP) [80].
Observe that the second approach (LNORM) is proven to be better than 24 out of these
38 different illumination-normalization approaches as shown previously in Table-3.2.
Four out of the above seven best-of-literature approaches will be compared later in
chapter 6 together with the proposed approach in chapter 5. The comparisons are done on
illuminated and non-illuminated face images in order to study both the positive and
negative effects of each approach.
59
Study 1 Study 2 Study 3 Study 4
Approaches: Approaches: Approaches: Correlation, Approaches:
HE, HM, GIC, LOG, SQI HE, RHE, GIC, RGIC, Eigenface, Eigenface w/o 1st GROSS, HE, GAMMA
Databases: GIC+RHE, RGIC+RHE,
RHE+RGIC, HE+RGIC,
3 PCs, NN (9 images), LS, Databases:
CMU-PIE, FERET, CAS-
GA, Cones-attached, Cones- Yale B, CMU PIE
PEAL QIR
Databases: cast, Harmonic (no cast), Recog. Methods:
Recog. Method:
Harmonic-cast, 9PL Eigenfaces, FaceIt
PCAÆLDA Yale B, Harvard
Recog. Method: (simulated), 9PL (real)
CosineÆNN Database: Yale B
Recog. Method:
Euclidean Distance
HE: Histogram Equalization LS: Linear Subspace LTV: Logarithmic Total Variation
RHE: Region-based HE GA: Gradient Angles HIE: Harmonic Image Exemplar
HM: Histogram Matching NN: Nearest Neighbor QI: Quotient Image
GIC: Gamma Intensity Correction QIR: Quotient Illumination Relighting 9PL: Nine Point of Lights
RGIC: Region-based GIC LOG: Logarithmic Function LBP: Local Binary Pattern
QIR: Quotient Illumination Relighting SQI: Self Quotient Image SSR: Single Scale Retinex
LHE: Local Histogram Equalization LHM: Local Histogram Matching MSR: Multi Scale Retinex
LNORM: Local Normal Distribution NORM: Normal Distribution CHAIN: Preprocessing Chain Appr.
Figure 3.27: Summarization for the nine comparative studies showing some relations between these
studies in addition to the final best normalization approaches from all studies (dark grayed boxes).
For each study, it shows the normalization approach to be compared, the face databases and the face
recognition approaches in addition to the best normalization approaches from each study (light
grayed boxes)
60
Table 3.3: The 38 different illumination normalization approaches appear in the above nine
comparative studies together with the corresponding studies numbers. (Note that the cited
approaches, from 29 to 38, are not described in details in their corresponding comparative studies)
Illumination Normalization Study Illumination Normalization Study
I I
Approach No. Approach No.
Histogram Equalization (HE) 1,2,4, Histogram Matching (HM)
1 2 1,5,6
6,7,8
3 Gamma Intensity Correction (GIC) 1,2,6 4 Logarithmic Function 1
5 Normal Distribution 6 6 Gamma correction 4
7 Local HE 6 8 Local HM 6
9 Local Normal Distribution 6 10 Local Binary Pattern 8
11 Region-based HE 2,6 12 Region-based GIC 2,6
GIC followed by Region-based Region-based GIC followed by
13 2,6 14 2,6
HE Region-based HE
Region-based HE followed by HE followed by Region-based GIC
15 2,6 16 2,6
Region-based GIC
Single Scale Retinex followed by Quotient Illumination Relighting
17 5 18 2,6,7
HM
19 Self Quotient Image 1,7 20 Quotient Image 7
21 Logarithmic Total Variation 7 22 GROSS method 4,8
Harmonic images (no cast Harmonic images-cast (with cast
23 3 24 3
shadow) shadows)
Nine point of lights (9PL) using Nine point of lights (9PL) using
25 3,6,7 26 3
simulated images real images
27 Multi Scale Retinex 9 28 Preprocessing Chain approach 9
29 Harmonic Image Exemplar [98] 6,7 30 Corefaces [99] 7
31 Linear Subspace [94] 3 32 Gradient Angles [97] 3
33 Cones-attached [95] 3 34 Cones-cast [95] 3
35 Correlation method [93] 3 36 Eigenface method [24] 3
Eigenface without the first three Nearest Neighbor using 9 training
37 3 38 3
principle components [32] images per subject [93]
After surveying the different face recognition approaches in chapter 2 and the different
illumination normalization approaches in chapter 3, we will introduce in the following
chapter the detailed descriptions about the environment that we build up in order to use it
for testing our proposed illumination normalization approach and the other approaches.
The chapter includes descriptions about the selected face recognition methods and the
selected databases that cover five different face recognition variations. The experimental
results of the selected methods over each database are also introduced in this chapter. All
experiments are done without applying any illumination normalization approach. These
results are considered as a baseline and allow us to study the effects of any illumination
normalization approach on the selected methods over each variation separately.
61
CHAPTER 4: Setup the Environment
4.1 Introduction
This chapter introduces the detailed descriptions about the environment that we build up
in order to use it for testing our proposed illumination normalization approach and the
other approaches. The chapter includes descriptions about the selected face recognition
methods and the selected databases in addition to the reasons behind these selections. The
experimental results of the selected methods over each database are also introduced in
this chapter.
In Chapter 2, we revise the main face recognition approaches and we found that many
face recognition methods fall into the holistic-based approach. One possible reason is that
these methods usually utilize the face as a whole and do not destroy any information by
exclusively processing only certain fiducial points which make them generally provide
more accurate recognition results. Moreover, most of these holistic-based approaches are
fall into two broad categories, Eigenspace-based category and frequency-based category.
To be able to widely studying the effects of any preprocessing/illumination normalization
approach on the two broad holistic-based categories, we chose one method under each
category representing the main characteristics of this category. The two chosen methods
are the Standard Eigenface method [24] which considered the core of many Eigenspace-
based methods, and the Holistic Fourier Invariant Features (Spectroface) method [43] that
represents the main characteristics of the frequency-based category.
These two methods are compared over five different face recognition variations using
suitable database(s) for each variation. These variations are divided into two-geometrical,
which are translation and scaling, and three-facial, which are 3D pose, facial expressions,
and non-uniform illumination. No preprocessing/illumination normalization approach is
applied to both methods during the comparison.
The aim of these comparisons is to establish a base that can be used for further studying
the effects of any preprocessing/illumination normalization approach on each of the five
variations separately using two methods representing the two broad holistic-based
categories, Eigenspace-based and frequency-based.
The rest of this chapter is organized as follows; section 2 describes the two face
recognition methods. Section 3 describes the face databases used in comparisons and how
they are prepared and configured for training and testing. Section 4 contains the
comparison results over each of the five variations in addition to some observations about
these results. Finally, the chapter summary is presented in section 5.
62
4.2 Methods Descriptions
4.2.1 Standard Eigenface Method
The standard Eigenface method [24] approximates the face images by lower dimensional
feature vectors. In training phase, the projection matrix (W ∈ RN × M) – which achieves the
dimensional reduction – is obtained using all the database face images, where N and M
denote for the dimension of image and feature vector respectively. Eigenvectors and
eigenvalues are computed on the covariance matrix of the training images. The M highest
eigenvectors are kept – which form the projection matrix W. Finally, the known
individuals are projected into the face space (pk), where p denotes for the feature vector
and k denotes for person number. These feature vectors are stored in addition to the mean
face.
The recognition process works as in Fig.4.1: a preprocessing module transforms the face
image into a unitary vector using a normalization module [24]. Then, subtract the mean
face from this unitary vector. The resulting vector I is projected using the projection
matrix W. This projection corresponds to a dimensional reduction of the input, starting
with vector I in RN (where N is the dimension of the image vector) and obtaining the
projected vector q in RM, with M<<N. Then, the similarity of q with each of the reduced
vectors (pk) is computed using Euclidean distance. The class of the most similar vector is
the result of the recognition process, i.e. the identity of the face.
63
in turn, reduces the computation of the recognition system. After decomposing the face
image, the holistic Fourier invariant features (Spectroface) are extracted from the low
frequency subband image by applying Fourier transform twice. The first FFT is applied
to the low frequency subband to make it invariant to the spatial translation. Then the
second Fourier transform is applied to the polar transformation of the result to make it
invariant to scale and on-the-plane rotation. The block diagram of the Spectroface
representation is shown in Fig.4.2.
In the recognition stage, the probe image is translated into Spectroface representation,
and then matches it – using Euclidean distance – with those referent images stored in the
gallery to identify the face image.
Our implementation for this method is done using C++ language. Different from
implementation in [43], we do not use the two preprocessing steps, namely histogram
equalization and intensity normalization, as when we use them over the ORL [112]
database, the recognition rates are always decreased. One possible reason for this
decreasing is that there are no illumination changes in the ORL database. Table-4.1
shows the recognition rates obtained by our implementation against those mentioned in
[43]. It is shown that our implementation gives approximately the same results over the
ORL database. However, in Yale database, the results of our implementation are
approximately the same as those of [43] except when the testing includes the two non-
uniform-illuminated images, the results are decreased significantly. The possible reason
for this is that we don’t apply the two preprocessing steps mentioned in [43].
64
Table 4.1: Comparison between results in Lai et al. [43] and in our implementation (better rates are
italic)
ORL DB (rank 1) Yale DB –not cropped (rank 1)
1 3 1 training image 2 training images
training training without 2 without 2
with 2 illum with 2 illum
image images illum illum
Method in [43] 76.38% 94.64% 95.0% 91.33% 99.05% 95.56%
65
normal 10 -15° pose 30° pose 45° pose 75-80° pose
Figure 4.3: UMIST: selected images for one subject in both training and testing sets
We use only these four subsets. All images are cropped to include only the head portion.
Subject’s images on each subset are divided into training and testing as follows: subset 1
is divided into 3 training images and 5 testing images; each of subset 2, 3, and 4 is
divided into 4 training images and 8, 8 and 10 testing images, respectively. As a result,
the training set consists of 15 images × 10 subjects while the testing set consists of the
remaining 31 images × 10 subjects. Fig.4.4 shows the training images, randomly selected,
in each subset and the light angle of each image.
Figure 4.4: Yale B: Training images for one subject in the four subsets with the light angle of each
image
66
4.3.4 JAFFE database
This database [65] is used for studying the facial expression variation. It contains 213
images of 10 Japanese female models obtained in front of a semi-reflective mirror. Each
subject was recorded three or four times while displaying the six basic emotions and a
neutral face. The camera trigger was controlled by the subjects. The resulting images
have been rated by 60 Japanese women on a 5-point scale for each of the six adjectives.
The rating results are distributed along with the images. Fig.4.6 shows example images
for one subject along with the majority rating. The images are originally printed in
monochrome and then digitized using a flatbed scanner.
In our comparisons, all images are cropped using the face detection function in the Intel
OpenCV library [116] to contain only the head portion, examples are shown in Fig.4.5.
(a) Notterdam
5 images × 70
subjects
(b) Yale
8 images × 15
subjects
(c) JAFFE
213 images
10 subjects
(d) Grimace
20 images × 18
subjects
Figure 4.5: Selected images for one subject from each database used for studying the facial expression variation
Figure 4.6: Example images from JAFFE database. The images in the database have been rated by
60 Japanese female subjects on a 5-point scale for each of the six adjectives. The majority vote is
shown underneath each image (with natural being defined through the absence of a clear majority)
67
4.3.5 Nott-faces database
This database [117] is used for studying the facial expression variation. It consists of 70
males each with 7 images, 4 frontal images with facial expressions, 1with bathing cap
and 2 in 3/4 profile. The images are with non-fixed background. There are some
translation of face in image, very small head scale variation and small linear uniform
illumination.
We exclude the 2 images with 3/4 profile from our test and work only on the 5 frontal
images which form a total of 5 images × 70 subjects. All images are cropped using the
face detection function in the Intel OpenCV library [116] to contain only the head
portion. Fig.4.5 shows sample images from this database.
68
0% +8% +17% -8% -17%
Scale
Up
Scale
Down
Figure 4.7: Face 94: 15 images for each subject in both training and testing sets
69
rotation in the chosen images. Note that the normal images are common in all cases. The
cases are chosen to cover all possible combinations for training by both four and six
images/subject. This is in addition to training by the two normal images per subject and
by all the training images up to 75˚ (10 images per subject). The testing is done using all
the 400 images of the testing set.
2. Results
We observe the following results from Table-4.2:
First, concerning the comparison between both methods, the Spectroface method gives
better recognition rates than Eigenface method in 10 out of 12 training cases with average
difference of 4.6%. This means that for the 3D pose variation, the Spectroface method
outperforms the Eigenface method.
Second, concerning the best training case, the top four rates in both methods show that
there is no significant difference between training by all the five angles – namely 0˚, 10˚,
30˚, 45˚, and 75˚ (10 images/subject) and training by three angles, 0˚, 75˚, and an in-
between angle [10˚, 45˚], (six images/subject). This means that to achieve the best
recognition rate over poses up to 75-80˚, the system should be trained by normal image +
75˚ pose + an in-between angle [10˚, 45˚].
Table 4.2: Pose Variation: recognition rates over 12 training cases (top four rates in each method are
italic)
Training Case # train/subject Eigenface Spectroface
normal only 2 64.0 48.0
normal + 10˚ 4 68.5 67.3
normal + 30˚ 4 75.0 76.0
normal + 45˚ 4 87.0 89.0
normal + 75˚ 4 85.5 89.0
normal + 10˚ + 30˚ 6 74.0 74.5
normal + 10˚ + 45˚ 6 84.5 90.0
normal + 10˚ + 75˚ 6 87.5 94.5
normal + 30˚ + 45˚ 6 85.0 90.3
normal + 30˚ + 75˚ 6 88.5 94.8
normal + 45˚ + 75˚ 6 87.5 95.0
normal+10˚+30˚+45˚+75˚ 10 88.0 95.0
70
training cases for each database. In each case, the testing is done using all other images
that are not included in the training. In Nott-faces, the first two cases are tested two times,
one without the capped images and other with them.
In Eigenface method, we try to improve its results in facial expressions variation by first
applying the wavelet transform on the original image, and then compute the Eigenface
from the resulting low subband. This low subband contains less information about facial
expressions which usually founded in LH and HL subbands. All previous training &
testing cases are applied to this low subband to be able to study the effect of it on the
recognition rates.
2. Results
We observe the following results from Table-4.3:
First, concerning the comparison between both methods, the Spectroface method gives
better recognition rates than the Eigenface on Original images in all the 14 training cases
with average difference of 5.4%.
Table 4.3: Expressions Variation: recognition rates over four databases with two Eigenface tests
Eigenface
Training Case # train × # test × Spectro-
Database on on
(images/subject) #subject #subject face
Original Wavelet
normal only 1 × 15 7 × 15 92.4 81.9 81.9
Yale normal + 2 expressions1 3 × 15 5 × 15 98.7 96.0 96.0
normal + 3 expressions 4 × 15 4 × 15 98.3 96.7 98.3
normal only 1 × 18 19 × 18 100 96.2 96.8
Grimace normal + 2 expressions 3 × 18 17 × 18 100 97.1 97.1
normal + 4 expressions 5 × 18 15 × 18 100 96.7 96.7
normal only 1 × 10 2032 93.1 84.2 86.2
JAFFE normal + 2 expressions 3 × 10 183 97.8 90.2 90.7
normal + 4 expressions 5 × 10 163 97.6 89.6 89.6
normal only 1 × 70 4 × 70 59.6 55.4 55.7
with
Nott-faces
Second, in Eigenface method, the last two columns show that computing the Eigenface
from the low subband of the wavelet transform gives better results than computing it
from the original image directly. The results are better in 9 training cases with average
1
Normal + N expression(s):- means that we train with normal image + N images each contains single
expression – randomly selected.
2
Number of testing images for all subjects, since in JAFFE database, each subject has different number of
images.
71
difference of 1.2% and are equal in the remaining cases. However, the Spectroface
method gives better recognition rates than the Eigenface on Wavelet also in all the 14
training cases with average difference of 4.7%.
Thus, in facial expression variation, it is better to use the wavelet low subband as it
contains less information about facial expressions which usually founded in LH and HL
subbands. However, applying the frequency-based method on the low subband of the
wavelet transform is much better than applying the PCA-based method on it. One
possible reason is that the wavelet low subband contains information about the
frequencies’ positions, and since the changing in facial expression causes changes in
pixels’ positions within the face, thus any information about pixels’ positions will contain
also information about facial expressions. As a result, applying the PCA (Eigenface) on
the wavelet low subband directly will still affected by some information about facial
expressions. In contrast, applying the Fourier transform on the wavelet low subband will
eliminate the information about pixels’ positions contained in this subband which leads to
reduce the information about facial expressions.
Briefly, for the facial expressions variation, the Spectroface method outperforms both
Eigenface on original image and Eigenface on the low subband of the wavelet transform.
However, applying Eigenface on the low subband is better than applying it on the
original image.
2. Results
Concerning the comparison between both methods, Table-4.4 shows that the Spectroface
method gives better recognition rates than the Eigenface method in all the 25 training
72
cases with average difference of 9.0%. This means that for non-uniform illumination
variation, the Spectroface method outperforms the Eigenface method.
Table 4.4: Illumination Variation: recognition rates over 25 training cases (top three rates in each
method are italic)
73
translation are filled by fixed color (gray in our case). In both times, each test image is
translated with 2, 4, 6, and 8 pixels in each of the four directions which give 16 new
recognition rates for each training case. Then, the decreasing in recognition rates is
calculated. Finally, the average decreasing for each translating value is calculated over
the four directions. Examples for translating are shown in Fig.4.8.
2. Results
First, in translating with circulation, Table-4.6 shows that the recognition rates of
Eigenface are decreased significantly. On other hand, the recognition rates of Spectroface
are not affected by these translations, as the maximum decrease value in all the 24 testing
cases is 0.8%.
Original Right by 2 Right by 4 Right by 6 Right by 8
(a) Translating
with circulation
(b) Translating
without circulation
Figure 4.8: Translation Variation: example for translating with and without circulation
Table 4.6: Translation Variation: average decreasing in the recognition rates of both methods after
translating with circulation in the four directions
(a) Eigenface Method (b) Spectroface Method
Translation Value Translation Value
Database Database
2 4 6 8 2 4 6 8
UMIST 1.7 4.6 10.5 20.7 UMIST 0.0 0.0 0.0 0.0
Grimace 0.1 1.3 11.7 29.9 Grimace 0.0 0.0 0.0 0.0
Yale 2.1 10.5 19.6 31.7 Yale 0.0 0.0 0.0 0.0
JAFFE 2.9 11.9 20.4 31.5 JAFFE 0.8 0.0 0.8 0.0
Nott-faces 4.1 14.3 27.7 39.6 Nott-faces 0.0 0.0 0.0 0.0
Yale B 2 6.1 13.6 18.8 Yale B 0.0 0.0 0.0 0.0
Average 2.2 8.1 17.3 28.7 Average 0.1 0.0 0.1 0.0
Second, in translating without circulation, Table-4.7 shows that the recognition rates in
both methods are decreased. However, the decreasing in Eigenface is much more
significant than in Spectroface – see the average row.
As a result, it is clear that the Spectroface method is more robust against the translation
variation than the Eigenface method.
74
Table 4.7: Translation Variation: average decreasing in the recognition rates of both methods after
translating without circulation in the four directions
(a) Eigenface Method (b) Spectroface Method
Translation Value Translation Value
Database Database
2 4 6 8 2 4 6 8
UMIST 1.5 5.5 10.9 22.2 UMIST 0.6 1.8 4.5 9.2
Grimace 0.2 1.4 12.8 37.7 Grimace 0.1 2.1 7.6 12
Yale 2.1 10.9 20.4 31.3 Yale 0 0 0.4 1.2
JAFFE 2.5 12.6 22.9 34.3 JAFFE 1.2 1.4 3.5 5.3
Nott-faces 3.6 15.1 26.8 38.7 Nott-faces 0 1.8 6.6 13.2
Yale B 2.4 8.5 16.5 22.8 Yale B 0.1 2.3 7.5 13.3
Average 2.1 9 18.4 31.2 Average 0.3 1.6 5 9
The testing is done using all the images in the testing set. For each training case, the
testing is done two times, before and after scaling, in order to record the decreasing in
recognition rates after scaling all testing images, see Table-4.9.
2. Results
Concerning the comparison between both methods, Table-4.9 shows that the Eigenface
method gives better results (less decreasing in recognition rates) than the Spectroface
method in six out of the seven used training cases with average difference 2.7%. This
means that for the scaling variation, the Eigenface method outperforms the Eigenface
method.
75
Table 4.9: Scaling Variation: decreasing in recognition rates after scaling all images in the testing set
Training Case Eigenface Spectroface
normal only 14.6 19.1
normal + up8 6.6 7.7
normal + down8 9.8 12.8
normal + up8 + down8 0.7 0.7
normal + up17 5.8 8.0
normal + down17 8.7 13.1
normal + up17 + down17 0 0.9
4.5 Summary
In this chapter, we introduce a comparison between two holistic-based face recognition
methods chosen to represent the two broad categories of the holistic-based approach –
namely Standard Eigenface method from the PCA-based category and Spectroface from
the frequency-based category. Seven databases, ranged from small to medium size, are
used to compare the two methods against five main variations separately using suitable
database(s) for each variation. All comparisons are applied without using any
preprocessing/illumination normalization approaches.
The aim of these comparisons is to establish a base that can be used for further studying
the effects of any preprocessing/illumination normalization approach on each of the five
variations separately using two methods representing the two broad holistic-based
categories, Eigenspace-based and frequency-based categories.
Moreover, the comparison results show that the Spectroface method outperforms the
Eigenface method in four out of the five variations – namely the 3D pose, facial
expressions, non-uniform illumination, and translation variations while the Eigenface
method is better in the scaling variation. Also, in facial expressions variation, applying
the frequency-based method on the low subband of the wavelet transform is much better
than applying the PCA-based method on it. One possible reason is that applying the PCA
(Eigenface) on the wavelet low subband directly will still be affected by some
information about pixels’ positions which represent information about facial expressions.
In contrast, applying the Fourier transform on this subband will eliminate this
information about pixels’ positions which leads to reduce the information about facial
expressions.
In the next chapter, we will describe the proposed illumination normalization approach
together with its results on illuminated databases. Moreover, the comparisons of the
proposed approach with other best-of-literature approaches will be discussed in chapter 6.
We’ll use the results of this chapter as a baseline for those of the next two chapters to see
the effect of the normalization approaches on different databases.
76
CHAPTER 5: The Proposed Illumination Normalization
Approach
5.1 Introduction
As we stated previously in Chapter 3, illumination normalization approaches can be
classified into two categories: model-based and image-processing based approaches.
Although the model-based approaches are perfect in theory, the requirement of
assumptions and constraints in addition to their highly computational cost make these
approaches unsuitable for realistic applications. On the other-hand, the image-processing
based approaches are more commonly used in practical systems for their simplicity and
efficiency.
Although most of the illumination normalization approaches can cope with illumination
variation well, some may bring negative influence on images without illumination
variation. In addition, some approaches show great difference on performance when
combined with different face recognition approaches. Some other approaches require
perfect alignment of face within the image which is difficult to achieve in practical/real-
life systems.
So, in this chapter, we aim to propose an image-processing-based illumination
normalization approach that proves flexibility to different face recognition approaches
and independency to face alignment. These make it suitable for practical/real-life systems
as it can be used with different face recognition approaches and doesn’t need any pre-
assumptions or constraints concerning the face alignment.
This chapter is organized as follows: section 2 describes the idea behind the proposed
illumination normalization approach. Sections 3 to 6 contain the detailed descriptions of
the proposed approach. Experiments appear in section 7. Finally, chapter summary is
presented in section 8.
77
among three globally-applied approaches, which are normal distribution (NORM), HE
and GIC.
Moreover, Histogram matching has the following two main advantages:
1. It is a preprocessing step that can be applied with any face recognition approach.
2. It is insensitive to geometrical effects on the image as it’s applied globally and
thus no additional alignment steps are required.
Although enhancing the image resulting from the HM can lead to increase the recognition
rates over using the HM alone, no attempts have been made to combine the HM with
other image enhancement methods for illumination normalization. Also, the compression
function of the Retinal filter [73] as an image enhancement method has not been used in
the literature. So, it’s very interesting to combine the HM with other image enhancement
methods as illumination normalization for face recognition.
As a result, we introduce a new illumination normalization approach based on enhancing
the image resulting from HM. Four different image enhancement methods are used in this
study – three of them are common in literature, namely histogram equalization, log
transformation and gamma correction [74], while the fourth method is newly suggested to
be used as an image enhancement method in this study, which is the compression
function of the Retinal filter [73]. These four image enhancement methods are applied in
two different approaches through this study:
1. After histogram matching; on the resulting image from HM.
2. Before histogram matching; on the reference image before matching the input
image on it.
In addition, for each approach, we try to further enhancing the results by applying one of
these four methods again. Finally, the proposed approach is chosen from these
combinations based on the increase in recognition rates against using the HM alone
regardless of the following conditions:
1. Face recognition approach that the normalization approach is applied with it,
2. Face alignment within the image,
3. Number of training images, and the degree of illumination within these images.
This ensures both the flexibility of the proposed approach among different face
recognition approaches and the ability to apply it on practical/real-life systems in which
perfect alignment of faces is difficult to achieve. The verifications of these conditions are
described in detail later in this chapter.
All previous combinations are empirically demonstrated and compared over Yale B
database [64] using the two holistic-based face recognition approaches introduced
previously in Chapter 4, namely, standard Eigenface [24] and Spectroface [43]. These
two approaches are chosen to represent the two broad holistic-based categories,
Eigenspace-based and Frequency-based respectively [22].
78
The rest of this chapter is as follows: section 3 contains the description of the histogram
matching algorithm. Section 4 contains the description of the four image enhancement
methods. In section 5, the different approaches of applying these four methods to enhance
the image resulting from HM are introduced. Section 6 is dedicated to describe the
verification of the selection conditions using the Yale B database. Experimental results
showing the best combinations of HM with different image enhancement methods are
presented in section 7. Finally, chapter summary is presented in section 8.
∑ H ( j)
j =0
FH → U(i) = N −1 (5.1)
∑ H ( j)
j =0
∑ G( j)
j =0
FG → U(i) N −1 (5.2)
∑ G( j)
j =0
Where N is the number of discrete intensity levels. N = 256 for 8-bit grayscale images.
To find the mapping function, FHÆG(i), we invert the function FGÆU(i) to obtain FUÆG(i).
Since the domain and the range of the functions of this form are identical, the inverse
mapping is trivial and is found by cycling through all values of the function. However,
due to the discrete nature of these functions, inverting can yield a function which is
undefined for certain values. Thus, we use linear interpolation and assume smoothness to
fill undefined points of the inverse function according to the values of well-defined points
79
in the function. As a result, we generate a fully defined mapping FUÆG(i) which transforms
a uniform histogram distribution to the distribution found in histogram G(i). The mapping
FHÆG(i) can then be defined as in equation 5.3, [66].
FH →G ( i ) = FU →G ( FH →U ( i ) ) (5.3)
Illuminated
Image
(a) (b)
Well-lit
Image
(c) (d)
Resulting
Image
(e) (f)
Figure 5.1: Histogram matching process to an illuminated image
It’s common in literature to match all images, in both training and testing sets, with a
single histogram of either a fixed well-lit image as in [71], [67] or an average image as in
[72]. In this work, the reference image for HM is constructed by calculating the average
image of a set of well-lit images – one for each subject which gives, by our experiments,
better results than using a single well-lit image for the whole image set.
The complexity to match the histogram of the input image to the one of the reference
image is O(L), where L is the number of pins in the histogram (equal to 255 in gray-scale
image). While the complexity of applying the new histogram to the input image takes
order O(N × M), where N, M represent the image resolution. This makes the complexity
of the whole HM process is O(N × M).
80
5.4 Image Enhancement Methods
The principal objective of image enhancement is to process the original image to be more
suitable for the recognition process. Many image enhancement methods are available in
the literature. Usually, a certain number of trial and error experiments are required before
a particular image enhancement method is selected [74]. In this study, four image
enhancement methods are chosen. Three of them are common in literature, namely
histogram equalization, log transformation and gamma correction, while the fourth
method which called the compression function of the Retinal filter [73] is newly
suggested to be used as an image enhancement method in this study.
s = c log(1 + r ) (5.5)
Where r and s are the old and new intensity value, respectively and c is a gray stretch
parameter used to linearly scaling the result to be in the range of [0, 255]. The shape of
the log curve in Fig.5.2 shows that this transformation maps a narrow range of dark input
gray-levels (shadows) into a wider range of output gray levels. The opposite is true for
the higher values of the input gray-levels.
s = cr 1 γ (5.6)
81
where r and s are the old and new intensity value, respectively, c is a gray stretch
parameter used to linearly scaling the result to be in the range of [0, 255] and γ is a
positive constant. In our case, the γ is chosen to be greater than 1 (empirically, it’s chosen
to be four) in order to map a narrow range of dark input values (shadows) into a wider
range of output values, with the opposite being true for higher values of input levels as
shown in Fig.5.2. Unlike the log transformation, the gamma correction has a family of
possible transformation curves obtained simply by varying γ values.
Figure 5.2: Transformation functions of LOG and GAMMA (L: number of gray levels)
82
The definition of the compression function C is based on X0:
(255 + X0) Iin
C= (5.8)
Iin + X0
Fig.5.3 shows the result of applying each of the four enhancement methods on a face with
non-uniform illumination.
Illuminated HE LOG GAMMA COMP
Average well-lit
Reference Image
Figure 5.4: Block diagram of applying the image enhancement method after the HM
83
Illuminated HM
Figure 5.5: Effects of applying the image enhancement methods after applying the HM
Figure 5.6: Block diagram of applying the image enhancement method before the HM
Illuminated HM
Figure 5.7: Effects of applying the image enhancement methods before applying the HM
84
5.5.3 Further Enhancement
Here, we further enhancing the result of each combination using each of the four
enhancement methods which give us 8 × 4 = 32 additional combinations. Fig.5.8 shows
block diagrams for such enhancements. The effects of further enhancement on both the
HM-GAMMA and GAMMA-HM combinations using each of the four enhancement
methods are illustrated in Fig.5.9.
Figure 5.8: Block diagram showing the further enhancement of combinations in 5.5.1 and 5.5.2
Illuminated HM
Figure 5.9: Effects of further enhancement on both HM-GAMMA and GAMMA-HM combinations
using each of the four enhancement methods
85
1. Face recognition approach that the normalization approach is applied with it,
2. Face alignment within the image,
3. Number of training images, and the degree of illumination within these images.
This ensures both the flexibility of the proposed approach among different face
recognition approaches and the ability to apply it on practical/real-life systems in which
perfect alignment of faces is difficult to achieve.
We use the Yale B database [64] – frontal images only – as described in section 4.3.2 for
studying and comparing the 40 enhancement combinations.
In order to verify the first condition, each of the 40 enhancement combinations is applied
with the two face recognition methods, Eigenface and Spectroface, representing the two
broad holistic-based categories, Eigenspace-based and Frequency-based respectively. The
better enhancement combination the one that always enhances the recognition results in
both methods.
To verify the second condition, all images are cropped in two different ways to include
only the head portion:
1. Automatic cropping using the face detection function in Intel OpenCV library
[116] to produce non-aligned version of the database; we call it YALE B-AUTO.
2. Manual cropping using the landmarks’ coordinates available on the Yale B
website [119] to produce an aligned version of it; we call it YALE B-MANU.
These two versions, shown in Fig.5.10, allow us to test the robustness of each
enhancement combination against the geometrical changes in faces within the images.
The better enhancement combination, the one that always enhances the recognition
results either with or without aligning the faces inside images.
YALE
B-
AUTO
YALE
B-
MANU
Figure 5.10: Sample faces from Yale B database – automatically and manually cropped
To verify the third condition, all the 25 different training cases, described in section 4.4.3,
are used in the testing with this database, as shown in Table-5.1, in which the normal
image is common in all cases. These training cases are chosen to cover both the training
with each elementary subset – namely subset 1, 2, 3, and 4, and the training with the
seven combinations of these subsets where subset 1 is essential in all of them as it
86
contains the lowest illumination. Each elementary subset is composed of training by the
normal image and either the vertical, horizontal or both lighting. While each combination
is composed of training by the normal image and either vertical lighting or vertical and
horizontal lighting.
These training varieties help us to test the robustness of each enhancement combination
against the number of training images and the changes in illumination direction of these
images. The better enhancement combination, the one that always increases the
recognition rates regardless of the training case.
Table 5.1: The 25 different training cases used in testing
Elementary Subsets Seven Combinations
Training Case(train. Training Case(train.
Subsets Subsets
images/subject) images/subject)
nor only nor + 4 ver
1 1, 2
nor + 2 ver nor + 4 ver + 2 hor
nor + 2 ver nor + 4 ver
1, 3
2 nor + 2 hor nor + 4 ver + 2 hor
nor + 2 ver + 2 hor nor + 4 ver
1, 4
nor + 2 ver nor + 4 ver + 2 hor
3 nor + 2 hor nor + 6 ver
1, 2, 3
nor + 2 ver + 2 hor nor + 6 ver + 4 hor
nor + 2 ver nor + 6 ver
1, 2, 4
4 nor + 2 hor nor + 6 ver + 4 hor
nor + 2 ver + 2 hor nor + 6 ver
1, 3, 4
nor + 6 ver + 4 hor
nor: normal 1, 2, 3, nor + 8 ver
ver: vertical hor: horizontal 4 nor + 8 ver + 6 hor
87
by ensuring the increasing of recognition rates in all the 25 training cases, it proves that
the chosen combination is not affected by either the number of training images or the
changes in illumination direction of these images.
As described in section 5.5, 32 out of the 40 enhancement combinations are for further
enhancement. So, to see if further enhancement the image leads to further increasing the
recognition rates or not, we plot the average difference in recognition rates from applying
the HM alone for each of the eight single enhancement combinations together with the
ones achieved by further enhancing them using either HE, GAMMA, LOG or COMP.
Fig.5.11 shows such plotting for the Eigenface method over the YALE B-AUTO
database. Fig.5.12 is dedicated for the Eigenface method over YALE B-MANU while
Figures 5.13, 5.14 are dedicated for the Spectroface method over the YALE B-AUTO
and YALE B-MANU, respectively.
We summarize the results of the further enhancement combinations in Table-5.2 as
follows:
For each of the four further enhancement combinations, corresponding to applying either
HE, GAMMA, LOG or COMP after the eight single enhancement combinations, we
count how many times the further enhancement combination lead to increase the average
difference in recognition rates over the eight single enhancement combinations.
88
(b) Further enhancement using LOG
89
(a) Further enhancement using HE
90
(d) Further enhancement using COMP
Figure 5.12: Eigenface method over YALE B-MANU: Effects of further enhancement over the eight
single enhancements using (a) HE, (b) LOG, (c) GAMMA and (d) COMP
91
(c) Further enhancement using GAMMA
92
(b) Further enhancement using LOG
93
Table 5.2: The number of combinations that lead to increase the recognition rates after using each of
the enhancement methods for further enhancement
Face Further Enhancement Combination Using:
Recognition Database HE GAMMA LOG COMP
Method (8 combinations) (8 combinations) (8 combinations) (8 combinations)
YALE B-
AUTO 0 5 5 8
Eigenface
YALE B-
MANU 0 0 2 8
YALE B-
AUTO 0 1 0 5
Spectroface
YALE B-
MANU 1 0 0 5
It’s clear from Table-5.2 that further enhancement the image using any of the three
traditional enhancement methods – namely HE, GAMMA and LOG, doesn’t lead to
further enhancement in recognition rates of both the Eigenface and Spectroface methods
especially in the YALE B-MANU version, see the second and fourth rows. Only COMP
that’s lead to further enhancement in recognition rates of both face recognition methods
over the two database’s versions. For clarification, in Spectroface method over the YALE
B-MANU (last row in Table-5.2), we can see that when applying HE as further
enhancement after each of the eight single combinations, ONLY one of these
combinations get further increasing in its average recognition rate due to further
enhancement it with HE. When applying either GAMMA or LOG as further
enhancement, NONE of the eight single combinations get further increasing in its
average recognition rate due to further enhancement it with HE. On other hand, when
applying COMP as further enhancement, five out of the eight single combinations get
further increasing in their average recognition rates after applying the COMP over those
accomplished before applying it.
As a result, only five out of 40 enhancement combinations are satisfying the three
previously mentioned conditions, their effect is shown in Fig.5.15:
1. GAMMA-HM, where gamma is applied before HM.
2. GAMMA-HM-COMP, where gamma is applied before HM, then the result is
further enhanced by applying the compression function.
3. HE-HM-COMP, where equalization is applied before HM, then the result is
further enhanced by applying the compression function.
4. COMP-HM-COMP, where compression function is applied before HM, then the
result is further enhanced by applying it again.
5. HM-HE-COMP, where equalization is applied after HM, then the result is further
enhanced by applying the compression function.
94
Illuminated
GAMMA-HM- COMP-HM-
GAMMA-HM HE-HM-COMP HM-HE-COMP
COMP COMP
Figure 5.15: Effects of the five enhancement combinations that satisfy the three conditions
Table-5.3 and Table-5.4 show the results of using these combinations with the Eigenface
and the Spectroface methods, respectively, over both versions of the Yale B database. In
addition to the results over the 25 training cases, both the average recognition rate of each
combination over these training cases and the difference between it and the average
recognition rate of applying the HM alone are shown in the last two rows, respectively.
It appears from Table-5.3 that using the second enhancement combination, namely
GAMMA-HM-COMP, with the Eigenface method gives the best average difference from
HM alone (see last row) over the four other combinations in both database’s versions.
While in the Spectroface method, Table-5.4 shows that there are no significant
differences between using any of the five combinations in each of the database’s
versions, the enhancement in recognition rates ranged from 3.7% to 4.2% in YALE B-
AUTO and from 6.6% to 7.4% in YALE B-MANU.
As a result, we can choose the GAMMA-HM-COMP combination as the best
enhancement combination over the 40 different combinations according to the criteria
stated above.
95
Table 5.3: Results of using the best five combinations with the Eigenface method over the two
versions of the database. Average recognition rate is calculated over the 25 different training cases.
(The best average differences are italic)
(1: GAMMA-HM, 2: GAMMA-HM-COMP, 3: HE-HM-COMP, 4: COMP-HM-COMP, 5: HM-HE-
COMP, nor: normal, ver: vertical, hor: horizontal)
96
Table 5.4: Results of using the best five combinations with the Spectroface method over the two
versions of the database. Average recognition rate is calculated over the 25 different training cases.
(The best average differences are italic)
(1: GAMMA-HM, 2: GAMMA-HM-COMP, 3: HE-HM-COMP, 4: COMP-HM-COMP, 5: HM-HE-
COMP, nor: normal, ver: vertical, hor: horizontal)
5.8 Summary
Many illumination normalization approaches have been proposed in literature and can be
classified into two categories: model-based and image-processing based approaches. The
image-processing based approaches are more commonly used in practical systems for
their simplicity and efficiency.
97
Although most of the illumination normalization approaches can cope with illumination
variation well, some may bring negative influence on images without illumination
variation. In addition, some approaches show great difference on performance when
combined with different face recognition approaches. Some other approaches require
perfect alignment of face within the image which is difficult to achieve in practical/real-
life systems.
This chapter introduces a new image-processing based illumination normalization
approach based on enhancing the image resulting from histogram matching using the
gamma correction and the Retinal filter’s compression function, which we called
GAMMA-HM-COMP approach. It is based on three consecutive steps:
1. Applying the gamma correction on the reference average well-lit image,
2. Histogram matching the input image to the result from 1,
3. Applying the Retinal filter’s compression function to further enhancing the result
of 2.
Among 40 different enhancement combinations, GAMMA-HM-COMP approach proves
its flexibility among different face recognition approaches and independency to face
alignment. These make it suitable for practical/real-life systems as it can be used with
different face recognition approaches and doesn’t need any pre-assumptions or
constraints concerning the face alignment. The results show that the GAMMA-HM-
COMP leads to average increasing in recognition rates over HM alone ranges from 4~7%
in Eigenface and Spectroface methods using aligned and non-aligned versions of the Yale
B database.
Moreover, in this study, the compression function of the Retinal filter is newly applied as
an image enhancement method. It proves its suitability for further enhancement rather
than the other three traditional enhancement methods which are the histogram
equalization, gamma correction and log transformation.
98
CHAPTER 6: Evaluate the Proposed Approach
6.1 Introduction
The aim of this chapter is to establish comparative studies between the proposed
illumination normalization approach and best-of-literature approaches over images with
illumination variation and images with other facial and geometrical variations using the
two selected face recognition methods. This allows us to test which of these approaches
is flexible to different face recognition approaches, which is independent on face
alignment and which has less side-effects over variations other than illumination.
As chapter 3 introduced previously, there are seven best-of-literature approaches selected
among 38 different illumination normalization approaches based on surveying nine
different comparative studies. Here we chose four out of these seven approaches to
compare with the proposed approach. The chosen approaches are:
1. Single Scale Retinex with Histogram Matching (SSR-HM).
2. Local Normal Distribution (LNORM).
3. Local Binary Patterns (LBP).
4. Preprocessing Chain Approach (CHAIN).
The detailed descriptions of these approaches are introduced previously in chapter 3.
The rest of this chapter is organized as follows: section 2 describes the implementation
parameters of the four approaches and the proposed one in addition to the difference in
results between our implementation of some of these approaches and the published ones.
The comparison between the four approaches and the proposed one over images with
illumination variations and images with other facial and geometrical variations are
introduced in sections 3 and 4, respectively. Finally, chapter summary is introduced in
section 5.
99
6.2.1 Preprocessing Chain Approach (CHAIN)
As described previously in chapter 3, the CHAIN approach consists of four consecutive
steps, which are:
1. Gamma Correction.
2. Difference of Gaussian (DoG).
3. Masking.
4. Contrast Equalization.
Here we use the original implementation of the CHAIN approach used by the authors of
[109] without the masking step. The original implementation can be found in [121]. Also,
we use the same default settings of the various parameters of the CHAIN approach that
are summarized in Table-6.1. Moreover, it’s found by authors of [109] that the CHAIN
approach gives similar results over a broad range of parameter settings, which greatly
facilitates the selection of parameters.
Table 6.1: Default parameter settings for CHAIN approach
Procedure Parameter Value
Gamma Correction γ 0.2
σ0 1
DoG Filtering
σ1 2
Contrast α 0.1
Equalization τ 10
We use the output of applying CHAIN approach as it is without normalizing it to [0-255].
However, in Spectroface method, we bring to positive the whole grayscale range of the
resulting image by adding fixed value (equal 15) to all pixels. This gives much better
results on YALE B database with its two versions (YALE B-AUTO & YALE B-MANU)
as shown in Table-6.2.
Table 6.2: Results of applying CHAIN with and without sliding on Spectroface method on both
versions of the YALE B database
CHAIN Preprocessing
Database
without sliding with sliding
YALE B-AUTO 31.4% 72.2%
YALE B-MANU 59.1% 96.5%
The possible reason behind this is when we bring the whole range to positive; the DC
component of the FFT will have the maximum magnitude over all other components.
Thus, when we normalize the FFT magnitudes by dividing on DC to remove the scaling
factor (refer to section 4.2.2), the consistency between the FFT magnitudes remains the
same on the image itself and on all other images (since the DC always the max in all
images). On the other hand, if we don’t bring the whole range to positive, the max
magnitude will appear at different locations for different images, even for the images of
the same person. So, when we divide all FFT magnitudes on it for normalization, the
consistency between the FFT magnitudes remains the same on the image itself but differ
100
on other images. This leads to misclassification even between the images of the same
person due to the different location of the max value after normalization between the
images.
101
of a set of well-lit images – one for each subject rather than using a single well-lit image
for the whole image set.
To test if there is a difference between our implementation and the original one, we re-
implement the original experiments of the SSR-HM published in [66] using the best
sigma in the published work (σ = 2) but with the Eigenface as recognition approach rather
than the SVM that is used in [66]. The results in Table-6.5 show that there are no
significant difference between our implementation and the original one except in one
experiment. These differences may be due to the randomization of the training images
and/or the different recognition approach used by our experiments (Eigenface rather than
SVM).
Table 6.5: Difference between our implementation of the SSR-HM and the original one
Original Results Our Results
Normaliz. Differ-
Database Description Recog. Recog.
Approach % % ence
Approach Approach
train: 1 from S1
Yale B – head SVM 99.0 Eigenface 100 +1.0%
test: remain 63
SSRÆHM Yale B – head train: 1 from any subset
SVM 90.2 Eigenface 98.2 +8.0%
(σ = 2) (average of 20 test: remain 63
random from the train: 2 from any subset
SVM 99.8 Eigenface 99.7 -0.1%
whole database) test: remain 62
102
each subject rather than using a single well-lit image for the whole image set. Finally, the
parameters of the Gaussian filter used in the compression function of the Retinal filter is
originally taken from [73] which uses a Gaussian filter of size 15 × 15 with standard
deviation σ = 2.
103
This means that these approaches require the images to be perfectly aligned which is
difficult to achieve in practical/real-life systems.
Fig.6.3 (a) and (b) shows the decreasing in the performance of each approach due to the
non-aligning of faces on Eigenface and Spectroface respectively. (i.e. the difference
between the performance of each approach on the YALE B-MANU and YALE B-
AUTO). It’s clear that the minimum affected approach due to the non-aligning of faces
on both methods is the proposed approach.
Table 6.6: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over YALE B-MANU version. Average recognition rate is
calculated over the 25 different training cases.
(0: NONE, 1: LNORM, 2: LBP, 3: CHAIN, 4: SSR-HM, 5: GAMMA-HM-COMP, nor: normal, ver:
vertical, hor: horizontal)
104
Eigenface on YALE B-MANU
(NONE = 66.0%)
34
32
31
30
29
28
LNORM
CHAIN S1
LBP
SSR-HM
Normalization Approach Proposed
105
Eigenface on YALE B-AUTO
(NONE = 53.3%)
25
15
10
-5
LNORM
CHAIN S1
LBP
SSR-HM
Normalization Approach Proposed
The final conclusion from comparison on illuminated images is to use the SSR-HM
approach when the illuminated images are aligned properly while use the proposed
approach when the illuminated images are not aligned.
106
In the two-facial variations, each approach is applied on all training cases of each
database, then, its average recognition rate over these training cases is calculated and
compared with the corresponding baseline average rate that is calculated before in
chapter 4. In the two-geometrical variations, the average recognition rate for each
approach is calculated before and after each variation, then the difference between the
two averages is calculated and compared with the corresponding baseline difference that
is calculated before in chapter 4. Following are the comparison results on each face
recognition variation using both Eigenface and Spectroface methods.
Table 6.8: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over UMIST database. Average recognition rate is calculated
over all training cases.
(0: NONE, 1: LNORM, 2: LBP, 3: CHAIN, 4: SSR-HM, 5: GAMMA-HM-COMP)
Eigenface Spectroface
Training Case
0 1 2 3 4 5 0 1 2 3 4 5
normal only 64 25 35.5 23.5 28.5 44.5 48 41.3 48.8 36.8 36.3 40
normal + 10˚ 68.5 30.5 47.3 40.5 54 60 67.3 52.8 63 49 55 55.5
normal + 30˚ 75 37.5 57 43.5 56 71 76 57 74 58.5 54 65.8
normal + 45˚ 87 46.5 64.5 43 65 78 89 71.5 82.5 66.8 67.3 78
normal + 75˚ 85.5 52 67.5 50 61 80.5 89 76.8 84.8 67 73 83.3
normal + 10˚ + 30˚ 74 41.5 57.5 46 56.5 68.5 74.5 58.5 72.5 58 57.5 65
normal + 10˚ + 45˚ 84.5 55.5 69 52.5 73 79 90 76.8 85 69.3 72.3 80.5
normal + 10˚ + 75˚ 87.5 61 75.8 63 74.5 87.5 94.5 86.5 90.8 79.5 83.3 89.8
normal + 30˚ + 45˚ 85 49 71 50 67.5 80 90.3 76.3 85.5 70.8 70.5 80
normal + 30˚ + 75˚ 88.5 62.5 78 57.5 74 88 94.8 88 91 82.5 82.3 90.3
normal + 45˚ + 75˚ 87.5 63 79.3 54 70.5 87.5 95 83.5 89.3 79.5 80.5 87.8
normal+10˚+30˚+45˚+75˚ 88 68.5 83.3 66 81 90 95 90.5 92.5 86 85.8 90.3
Average Recognition
81.3 49.4 65.5 49.1 63.5 76.2 83.6 71.6 80 67 68.2 75.5
Rate
107
(a) Eigenface (b) Spectroface
Figure 6.4: Average difference in recognition rates after applying each of the five illumination
normalization approaches on UMIST database
Although all approaches lead to decrease the recognition rates as shown in Fig.6.3, the
proposed approach has the least side-effect due to 3D pose variation on Eigenface
method while the LBP has the least side-effect on the Spectroface method.
108
Table 6.9: Results of applying each of the five illumination normalization approaches with both
Eigenface and Spectroface methods over Grimace, Yale, JAFEE, and Nott-faces databases. Average
recognition rate is calculated over all training cases.
(0: NONE, 1: LNORM, 2: LBP, 3: CHAIN, 4: SSR-HM, 5: GAMMA-HM-COMP, expr: expression(s))
Eigenface Spectroface
DB Training Case
0 1 2 3 4 5 0 1 2 3 4 5
normal only 81.9 75.2 90.5 66.7 81 88.6 92.4 83.8 91.4 73.3 84.8 93.3
Yale
3
normal + 2 expr 96 88 93.3 84 94.7 94.7 98.7 94.7 96 86.7 97.3 97.3
normal + 3 expr 96.7 85 91.7 81.7 93.3 93.3 98.3 95 98.3 86.7 98.3 98.3
normal only 96.2 76.6 80.1 82.2 87.1 93.6 100 84.8 93 93.3 95 100
Grimac
normal + 2 expr 97.1 88.9 93.1 89.2 92.8 96.7 100 94.8 98.7 99 98 99
e
normal + 4 expr 96.7 91.5 94.1 88.5 91.5 96.7 100 94.4 99.3 99.3 97.8 100
normal only 84.2 65.5 65 55.7 73.4 79.8 93.1 82.3 81.8 69 84.2 87.7
JAFFE
normal + 2 expr 90.2 71 82 71.6 78.1 85.2 97.8 89.1 87.4 85.8 91.8 94.5
normal + 4 expr 89.6 77.9 78.5 75.5 77.9 89.6 97.6 90.8 93.9 84.7 95.7 97.6
normal only 55.4 43.9 52.5 36.1 42.5 59.6 57.1 51.4 46.1 49.3 62.9 66.1
with cap.
Nott-faces
normal + 1 expr 55.7 46.7 51 35.7 45.2 60 62.9 51.9 54.8 54.3 62.4 63.3
normal + 2 expr 47.9 43.6 43.6 33.6 42.1 50.7 56.4 47.9 47.1 51.4 52.9 52.9
normal only 69 51 62.9 39.5 48.1 76.2 63.8 65.2 55.2 57.6 80 83.3
w/out
cap
normal + 1 expr 76.4 55.7 65 40 52.9 84.3 77.1 71.4 70 67.1 87.9 87.9
Average Recognition
80.9 68.6 74.5 62.9 71.5 79.4 86.3 80.5 80.5 75.1 80.2 86.7
Rate
Eigenface on YALE
(NONE = 91.5%)
2
0
Average difference from NONE
-2
-4
-6
-8
-10
-12
-14
-16
LNORM
CHAIN S1
LBP
SSR-HM
Normalization Approach Proposed
3
normal + N expr:- means that we train with normal image + N images each contains single expression –
randomly selected.
109
Eigenface on Grimace
(NONE = 96.7%)
0
-4
-6
-8
-10
-12
LNORM
CHAIN S1
LBP
SSR-HM
Normalization Approach Proposed
-5
-10
-15
-20
-25
LNORM
CHAIN S1
LBP
SSR-HM
Normalization Approach Proposed
-5
-10
-15
-20
-25
LNORM
CHAIN S1
LBP
SSR-HM
Normalization Approach Proposed
110
6.4.3 Translation Variations
Here we use the same training and testing methodologies described in chapter 4 for
testing each of the five illumination normalization approaches. As described previously in
chapter 4, the testing of translation is applied two different times; first, by translating with
circulation in which the output pixels after translation are circulated to fill the empty
pixels in the opposite direction. Second, translating without circulation in which the
empty pixels after translation are filled by fixed color (gray in our case).
For translating with circulation case, Table-6.10 (a)-(e) shows the average decreasing in
recognition rates4 per database on both Eigenface and Spectroface methods for each of
the five approaches. Table-6.11 (a)-(e) is dedicated for translating without circulation
case. The average decreasing in recognition rates over all databases are shown in the last
rows of each table.
Table 6.10: Average decreasing in the recognition rates of both methods after translating with
circulation in the four directions while applying (a) LNORM, (b) LBP, (c) CHAIN, (d) SSR-HM and
(e) GAMMA-HM-COMP approaches as preprocessing step.
(a) LNORM Approach
Eigenface Method Spectroface Method
Translation Value Translation Value
Database Database
2 4 6 8 2 4 6 8
UMIST 18.3 34.9 41.4 42.8 UMIST 1.2 0.3 1.6 0.4
Grimace 8.5 38.6 58 64 Grimace 0 0.1 0 0.2
Yale 13.3 51.7 70.8 73.7 Yale 0.4 3.3 0.4 2.9
JAFFE 11.8 39.3 53.4 53.8 JAFFE 4.2 2.9 3.8 2.6
Nott-faces 18.2 40.3 51.6 53.6 Nott-faces 4.1 3.2 4.6 2
Yale B 7 24.1 37.5 41 Yale B 1 0 0.6 0.6
Average 12.9 38.2 52.1 54.8 Average 1.8 1.6 1.8 1.5
4
There are some cases in which the recognition rates are marginally increased, we consider it as noise and
set the decreasing value to 0 to indicate that there’s no decreasing due to the translation.
111
(c) CHAIN Approach
Eigenface Method Spectroface Method
Translation Value Translation Value
Database Database
2 4 6 8 2 4 6 8
UMIST 7.6 29.4 41.4 42.3 UMIST 6.3 6.9 8.2 8
Grimace 4 38.5 61.9 73.1 Grimace 0.5 0.7 1.3 0.8
Yale 12.5 44.2 63 69.6 Yale 7.1 11.7 6.7 12.1
JAFFE 12.4 39.2 53.8 55.6 JAFFE 4.2 4.6 4.9 4.6
Nott-faces 8 22.3 31.1 34.3 Nott-faces 7.3 8.8 9.5 7.5
Yale B 6.9 24.8 35.6 38.9 Yale B 2.3 4.1 3.8 3.6
Average 8.6 33.1 47.8 52.3 Average 4.6 6.1 5.7 6.1
112
Table 6.11: Average decreasing in the recognition rates of both methods after translating without
circulation in the four directions while applying (a) LNORM, (b) LBP, (c) CHAIN, (d) SSR-HM and
(e) GAMMA-HM-COMP approaches as preprocessing step.
(a) LNORM Approach
Eigenface Method Spectroface Method
Translation Value Translation Value
Database Database
2 4 6 8 2 4 6 8
UMIST 17.9 34 41 43.4 UMIST 1.7 1.2 1.7 2.6
Grimace 8.6 38.7 58.9 63.3 Grimace 0.1 0 0 0.5
Yale 15 52.9 70.4 73.8 Yale 1.7 2.9 1.3 4.2
JAFFE 10.8 38.3 54.2 55 JAFFE 4 3.4 4.4 3.4
Nott-faces 17.8 40.2 51.2 52.3 Nott-faces 4.3 5.9 7.1 10.2
Yale B 7.4 25.9 38.1 40.1 Yale B 1.5 1.9 2.6 3.9
Average 12.9 38.3 52.3 54.7 Average 2.2 2.6 2.9 4.1
113
(d) SSR-HM Approach
Eigenface Method Spectroface Method
Translation Value Translation Value
Database Database
2 4 6 8 2 4 6 8
UMIST 14.6 39.4 52.4 55.8 UMIST 8.4 23.9 31.4 32.4
Grimace 2.4 22.8 54 69.4 Grimace 0.7 4.6 9.6 5.1
Yale 7.5 31.6 52.9 64.1 Yale 1.7 5.8 7.1 12.5
JAFFE 4.7 23.6 38.9 46.1 JAFFE 1.2 4.9 6.7 6.3
Nott-faces 9.7 26.8 37 42.7 Nott-faces 7.1 16.2 22.5 24.3
Yale B 4.4 23.1 40.5 45 Yale B 12.3 30.6 38.5 33.4
Average 7.2 27.9 46 53.9 Average 5.2 14.3 19.3 19
Fig.6.9 and Fig.6.10 show the average decreasing curves after translating with and
without circulation, respectively, on both (a) Eigenface and (b) Spectroface methods.
Observe that in Figures 6.9 and 6.10, in both Eigenface and Spectroface, the best two
curves with minimum average decreasing in recognition rates are the NONE curve and
the proposed approach curve. This means that the proposed approach has the least side-
effect due to translation variation on both Eigenface and Spectroface methods among the
four other approaches. Note that the performance of the SSR-HM approach is
dramatically affected by the translation variation over both recognition methods.
114
(a) Eigenface (b) Spectroface
Figure 6.10: Average decreasing in recognition rates after translating without circulation
Eigenface Spectroface
Training Case
0 1 2 3 4 5 0 1 2 3 4 5
normal only 14.6 51.4 46.3 38.9 40.9 19.7 19.1 62.4 47.1 62.7 49.9 24
normal + up8 6.6 33.1 30.1 24.9 27.1 9.7 7.7 42.9 30.9 43.6 33.2 9.2
normal + down8 9.8 34.3 30.2 23.9 27.2 15.8 12.8 41.3 27.4 41.9 31.5 17.3
normal + up8 + down8 0.7 19.2 14.7 9.8 11.5 3.2 0.7 21.9 10.5 21.4 14.3 2.1
normal + up17 5.8 40.7 31.8 31.9 32.3 8.9 8 48.7 31 45.2 35.7 9.5
normal + down17 8.7 38.2 33.3 25.1 26.7 14.1 13.1 48.3 29.4 49.8 32.2 17.7
normal + up17 + down17 0 31.7 21.2 19.2 18.1 0.5 0.9 33.9 12.9 32.6 17.8 2.5
Average Decreasing
6.6 35.5 29.7 24.8 26.3 10.3 8.9 42.8 27 42.5 30.7 11.8
in Recognition Rate
Fig.6.11 (a) and (b) shows the average decreasing in recognition rates after applying each
of the five approaches on Eigenface and Spectroface respectively.
115
(a) Eigenface (b) Spectroface
Figure 6.11: Average decreasing in recognition rates when applying each of the five illumination
normalization approaches before and after scaling the Face 94 database
It’s clear from Fig.6.10 that the proposed approach has the least side-effect due to
scaling variation on both Eigenface and Spectroface methods among the four other
approaches, while has side-effect slightly more than that in the NONE case (i.e. without
applying any approach). Note that the performance of the other four approaches is
dramatically affected by the scaling variation over both recognition methods.
6.5 Summary
In this chapter, we establish comparative studies between the proposed illumination
normalization approach and four best-of-literature approaches over images with
illumination variation and images with other facial and geometrical variations using the
two selected face recognition methods.
When dealing with illuminated images, the results show that the proposed approach is
the best one in the Eigenface method and the second best after SSR-HM in the
Spectroface method when the images are not perfectly aligned. However, the SSR-HM
is the best one in both methods when the images are perfectly aligned. In addition, the
proposed approach is the minimum affected approach (i.e. most robust) due to the non-
aligning of faces on both methods.
When dealing with non-illuminated images, the proposed approach brings the least
side-effects, among the four other approaches, in both methods for each of the two-facial
variations and the two-geometrical variations (except in the Spectroface method over
pose variation, it comes in second place after LBP). Moreover, the performance of the
SSR-HM is dramatically affected by the two-geometrical variations, translation and
scaling, on both recognition methods.
Thus, we can conclude the following about the proposed approach:
1. It’s flexible to different face recognition approaches, as it usually gives the best
results on both methods either on illumination variations or other variations.
116
2. It’s robust to the non-aligning of faces rather than the other four approaches
which show great difference in performance when applied to non-aligned face
images.
3. It has the least side-effects, among the four other approaches, over both facial
variations and geometrical variations.
The following chapter of this thesis will summarize the conclusion of the work together
with the suggestions for the future works.
117
CHAPTER 7: Conclusions and Future Works
7.1 Conclusions
Although many face recognition techniques and systems have been proposed, evaluations
of the state-of-the-art techniques and systems have shown that recognition performance
of most current technologies degrades due to the variations of illumination. As prove for
this, the last face recognition vendor test FRVT 2006 concludes that relaxing the
illumination condition has a dramatic effect on the performance. Moreover, It has been
proven both experimentally and theoretically that the variations between the images of
the same face due to illumination are almost always larger than image variations due to
change in face identity.
There has been much work dealing with illumination variation in face recognition.
Although most of these approaches can cope with illumination variation well, some may
bring negative influence on images without illumination variation. In addition, some
approaches show great difference on performance when combined with different
recognition methods. Some other approaches require perfect alignment of face within the
image which is difficult to achieve in practical/real-life systems.
The results show that the proposed approach is the best one in the Eigenface method and
the second best in the Spectroface method when dealing with illuminated images that are
not perfectly aligned. Moreover, the performance of the other approaches is significantly
affected by the aligning of the faces inside images, opposite to the proposed approach
which not significantly affected by the aligning condition. In addition, the proposed
approach brings the least side-effects, among the four other approaches, in each of the
two methods when dealing with either facial or geometrical variations.
118
These results lead to conclude that the proposed approach:
1. is flexible to different face recognition approaches,
2. is robust to the non-aligning of faces, and
3. brings the least side-effects on images with either facial or geometrical variations.
In addition, this thesis establishes an environment that can be used for further studying
the effects of any preprocessing/illumination normalization approach. The environment
consists of:
1. Two face recognition methods representing the two broad categories of the
holistic-based face recognition approach – namely Standard Eigenface method
from the Eigenspace-based category and Spectroface from the Frequency-based
category.
2. Seven databases representing five different face recognition variations with
suitable database(s) for each variation. The variations include three-facial, which
are 3D pose, expressions and non-uniform illumination, and two-geometrical,
which are translation and scaling.
The comparative study between these two face recognition methods over the five
variations show that the Spectroface method outperforms the Eigenface method in each
of 3D pose, facial expressions, non-uniform illumination and translation variations while
the Eigenface method is better in the scaling variation.
Finally, we nominate seven illumination normalization approaches that are considered the
best-of-literature approaches based on surveying nine different comparative studies
containing 38 different approaches. These seven approaches can be considered as
benchmark that can be used for comparing any further new/suggested illumination
normalization approach.
119
7.2 Future Works
As the proposed approach depends mainly on the histogram matching, it’s possible to
apply the HM in block-wise manner rather than on the whole face. This can produce
better results as it will normalize the illumination of each face region separately
according to its illumination condition.
Another possible modification is to apply the region-based GIC to automate the
selection of gamma value over each region separately rather than using a single, fixed,
gamma value over the whole face.
In this work, all illumination normalization approaches are tested on two face recognition
methods representing the two broad categories of the holistic-based approach. It’s
important to extend this work to include both local-based and hybrid face recognition
methods in testing these approaches as they may introduce difference in performance
when combined with such methods.
Moreover, the environment established in this work can be extended to include additional
databases (specially illuminated ones) and/or other face recognition variations (e.g.
aging).
Finally, this work introduces a technology evaluation for the proposed approach and the
other best-of-literature approaches. In order to complete the thorough evaluation cycle,
both scenario and operational evaluations need to be performed for these approaches.
120
References
[1] Maltoni, D.; Maio D., Jain A.K. & Prabhakar S. Handbook of Fingerprint Recognition, Springer,
New York, 2003.
[2] W. Zhao, R. Chellappa, P. J. Phillips, and A. Rosenfeld, Face recognition: A literature survey, ACM
Computing Surveys (CSUR), 35(4), 399–458, 2003.
[3] Xiaoyang Tan , Songcan Chen , Zhi-Hua Zhou , Fuyan Zhang, Face recognition from a single image
per person: A survey, Pattern Recognition, v.39 n.9, pp.1725-1745, Sep 2006.
[4] Matthew Curtis Hesher, Automated Face Tracking and Recognition, M.Sc. Thesis, the Florida State
University, 2003.
[5] P. J. B. Hancock, V. Bruce, and A. M. Burton, “Recognition of unfamiliar faces,” Trends in
Cognitive Sciences, vol. 4, pp. 330–337, 2000.
[6] Daugman, John. Face and Gesture Recognition: Overview. IEEE Transactions on Pattern Analysis
and Machine Intelligence, vol. 19, issue 7, Jul 1997.
[7] Gong, Shaogang, Stephen J. McKenna, Alexandra Psarrou. Dynamic Vision: From Images to Face
Recognition. Imperial College Press, 2000.
[8] Hochberg, Julian, Ruth Ellen Galper. Recognition of Faces: I. An Exploratory Study. Psychonomic
Science, vol. 9, pp. 619-620, 1967.
[9] Brigham, J. C., A. Maass, L. D. Snyder, K. Spaulding. Accuracy of Eyewitness Identification in a
Field Setting. Journal of Personality and Social Psychology, vol. 42, pp. 673-681, 1982.
[10] Brown, E., K. De_enbacher, W. Sturgill. Memory for Faces and the Circumstances of Encounter.
Journal of Applied Psychology, vol. 62, pp. 311-318, 1977.
[11] L. Torres, Is there any hope for face recognition?, Proc. of the 5th International Workshop on Image
Analysis for Multimedia Interactive Services, WIAMIS 2004, Lisboa, Portugal, Apr 2004.
[12] P. J. Phillips, W. T. Scruggs, A. J. O'Toole, P. J. Flynn, K.W. Bowyer, C. L. Schott, and M. Sharpe,
”FRVT 2006 and ICE 2006 Large-Scale Results”, National Institute of Standards and Technology,
NISTIR 7408, http://face.nist.gov, 2007.
[13] Y. Adini, Y. Moses, and S. Ullman, “Face recognition: The problem of compensating for changes in
illumination direction” IEEE Tran. PAMI, 19(7), 721-732, 1997.
[14] W. Zhao, and R. Chellappa, “Robust face recognition using symmetric shape-from-shading”
Technical report, Center for Automation Research, University of Maryland, 1999.
[15] Laurenz Wiskott, Jean-Marc Fellous, Norbert Kruger, Christoph von der Malsburg, "Face
Recognition by Elastic Bunch Graph Matching," IEEE Transactions on Pattern Analysis and
Machine Intelligence ,vol. 19, no. 7, pp. 775-779, Jul 1997.
[16] Liao, R. and Li, S. Z. 2000. Face Recognition Based on Multiple Facial Features. In Proceedings of
the Fourth IEEE international Conference on Automatic Face and Gesture Recognition 2000. FG.
IEEE Computer Society, Washington, DC, 239, Mar 2000.
[17] E. Hjelmås. Biometric Systems: A Face Recognition Approach. In Proceedings of the Norwegian
Conference on Informatics, pp. 189-197, 2000.
[18] B. Heisele P. Ho and T. Poggio, “Face Recognition with Support Vector Machines: Global versus
Component-Based Approach,” Proc. Int'l Conf. Computer Vision, vol. 2, pp. 688-694, 2001.
[19] ZANA, Y.; CESAR-JR, R. M.; BARBOSA, R. A. Automatic face recognition system based on local
Fourier-Bessel features. In: BRAZILIAN SYMPOSIUM ON COMPUTER GRAPHICS AND
IMAGE PROCESSING, 18. (SIBGRAPI), Brazil, 2005.
[20] R. Paredes and E. Vidal and F. Casacuberta, Local Features for Biometrics-Based Recognition, 2nd
COST 275 Workshop, Biometrics on the Internet Fundamentals, Advances and Applications, 2004.
[21] Timo Ahonen, Abdenour Hadid, Matti Pietikäinen, "Face Description with Local Binary Patterns:
Application to Face Recognition," IEEE Transactions on Pattern Analysis and Machine
Intelligence ,vol. 28, no. 12, pp. 2037-2041, Dec 2006.
[22] T.I. El-Arief, K.A. Nagaty, and A.S. El-Sayed, “Eigenface vs. Spectroface: A Comparison on the
Face Recognition Problems”, IASTED Signal Processing, Pattern Recognition, and Applications
(SPPRA), Austria, 2007.
[23] M. Kirby and L. Sirovich, Application of karhunen-Loeve procedure for characterization of human
faces, IEEE Patt. Anal. Mach. Intell., 12, 103-108, 1990.
121
[24] M. Turk and A. Pentland, Eigenfaces for recognition, Journal of Cognitive Neuroscience, 3(1), 71-
86, 1991.
[25] Zhao .W, Chellappa .R and Phillips .P `Subspace Linear Discriminant Analysis for Face
Recognition', technical report, Center for Automation Research, University of Maryland, College
park, 1999.
[26] Liu, C. and Wechsler, H. 2000. Evolutionary Pursuit and Its Application to Face Recognition. IEEE
Trans. Pattern Anal. Mach. Intell. 22, 6, 570-582, Jun 2000.
[27] Déniz, O., Castrillón, M., and Hernández, M. 2003. Face recognition using independent component
analysis and support vector machines. Pattern Recogn. Lett. 24, 13, 2153-2157, Sep 2003.
[28] Bartlett, M.S., Movellan, J.R., T.J., Sejnowski.: Face Recognition by Independent Component
Analysis. IEEE Transactions on Neural Networks, vol. 13, No. 6, Nov 2002.
[29] Wu, J. and Zhou, Z. 2002. Face recognition with one training image per person. Pattern Recogn. Lett.
23, 14, 1711-1719, Dec 2002.
[30] Moghaddam, B., Jebara, T., and Pentland, A. Bayesian modeling of facial similarity. In Proceedings
of the 1998 Conference on Advances in Neural information Processing Systems II D. A. Cohn, Ed.
MIT Press, Cambridge, MA, 910-916, 1999.
[31] Shaohua Zhou Chellappa, R. Moghaddam, B., Intra-personal kernel space for face recognition,
Sixth IEEE International Conference on Automatic Face and Gesture Recognition, Proceedings.
pp. 235- 240, 2004.
[32] P. Belhumeur, J. Hespanha, and D. Kriegman, Eigenfaces vs. Fisherfaces: Recognition Using Class
Specific Linear Projection, Proc. of the Fourth European Conference on Computer Vision, Vol. 1,
Cambridge, UK, pp. 45-58, Apr 1996.
[33] C. Liu and H. Wechsler, Comparative Assessment of Independent Component Analysis (ICA) for
Face Recognition, Second International Conference on Audio- and Video-based Biometric Person
Authentication, Washington D.C., USA, Mar 1999.
[34] K. Baek, B. Draper, J.R. Beveridge, and K. She, PCA vs. ICA: A Comparison on the FERET Data
Set, Proc. of the Fourth International Conference on Computer Vision, Pattern Recognition and
Image Processing, Durham, NC, USA, pp. 824-827, Mar 2002.
[35] B. Moghaddam, Principal Manifolds and Probabilistic Subspaces for Visual Recognition, IEEE
Trans. on Pattern Analysis and Machine Intelligence, Vol. 24, No. 6, pp. 780-788, Oct 2002.
[36] J.R. Beveridge, K. She, B. Draper, and G.H. Givens, A Nonparametric Statistical Comparison of
Principal Component and Linear Discriminant Subspaces for Face Recognition, Proc. of the IEEE
Conference on Computer Vision and Pattern Recognition, Kaui, HI, USA, pp. 535- 542, Dec 2001.
[37] A. Martinez and A. Kak, PCA versus LDA, IEEE Trans. on Pattern Analysis and Machine
Intelligence, Vol. 23, No. 2, pp. 228-233, Feb 2001.
[38] P. Navarrete and J. Ruiz-del-Solar, Analysis and Comparison of Eigenspace-Based Face Recognition
Approaches, International Journal of Pattern Recognition and Artificial Intelligence, Vol. 16, No. 7,
pp. 817-830, Nov 2002.
[39] Delac, K., Grgic, M., and Grgic, S., Independent comparative study of PCA, ICA, and LDA on the
FERET data set, Int'l Journal of Imaging Systems and Technology, 15(5), 252-260, 2005.
[40] J. Ruiz-del-Solar and P. Navarrete, "Eigenspace-based face recognition: a comparative study of
different approaches," IEEE Transactions on Systems, Man and Cybernetics Part C: Applications and
Reviews, vol. 35, no. 3, pp. 315-325, 2005.
[41] Pan, Z. and H. Bolouri, High speed face recognition based on discrete cosine transforms and neural
networks, Technical Report, Univ. of Hertfordshire, UK, 1999.
[42] Spiess, H. and Ricketts, I., Face recognition in Fourier space, Proc. Vision Interface Conf., Montreal,
Canada, 2000.
[43] J. H. Lai, P. C. Yuen, and G. C. Feng, Face recognition using holistic Fourier invariant features,
Pattern Recognition, 34(1), 95–109, 2001.
[44] Dai, D., Feng, G., Lai, J., and Yuen, P. C. Face Recognition Based on Local Fisher Features. In
Proceedings of the Third international Conference on Advances in Multimodal interfaces. T. Tan, Y.
Shi, and W. Gao, Eds. Lecture Notes In Computer Science, vol. 1948. Springer-Verlag, London,
230-236, 2000.
[45] G. C. Feng, P. C. Yuen, and D. Q. Dai, Human face recognition using PCA on wavelet subband,
Journal of Electronic Imaging, 9(2), 226-233, 2000.
122
[46] D.B. Graham and N.M. Allinson, Characterizing virtual eigen signatures for general purpose face
recognition, Face Recognition: From Theory to Applications, H. Wechsler, P.J. Phillips, V. Bruce, F.
Fogelman-Soulie, and T.S. Huang, eds., 163, 446-456, 1998.
[47] Z. Pan, R. Adams, and H. Bolouri, “Dimensionality reduction of face images using discrete cosine
transforms for recognition." submitted to IEEE Conference on Computer Vision and Pattern
Recognition, 2000.
[48] Zana, Y. and Cesar, R. M. 2006. Face recognition based on polar frequency features. ACM Trans.
Appl. Percept. 3, 1, 62-82, Jan 2006.
[49] Saradha, A. & Annadurai, S. A Hybrid Feature Extraction Approach for Face Recognition Systems.
Institute of Road and Transport Technology, Erode, 2004.
[50] Casasent D et al, “Face Recognition with Pose and Illumination Variations Using New SVRDM
Support-Vector Machine”, Optical Engineering, vol. 43, No. 8, Society of Photo-Optical
Instrumentation Engineers, pp. 1804-1813, Aug 2005.
[51] S. Srisuk and W. Kurutach, Face Recognition using a New Texture Representation of Face Images,
Proceedings of Electrical Engineering Conference, Cha-am, Thailand, pp. 1097-1102, Nov 2003.
[52] A. Ne an, "Embedded Bayesian networks for face recognition," ICME 2002 - IEEE International
Conference on Multimedia and Expo, Lausanne, Switzerland, pp. 133-136, Aug 2002.
[53] Huang, J., Yuen, P. C., Lai, J. H., and Li, C. 2004. Face recognition using local and global features.
EURASIP J. Appl. Signal Process., 530-541, Jan 2004.
[54] P. C. Yuen and J. H. Lai, “Face representation using independent component analysis,” Pattern
Recognition, vol. 35, no. 6, pp. 1247–1257, 2002.
[55] Rao, K. S. and Rajagopalan, A. N. 2005. A probabilistic fusion methodology for face recognition.
EURASIP J. Appl. Signal Process., 2772-2787, Jan 2005.
[56] Tony Mansfield, Gavin Kelly, David Chandler, and Jan Kane, “Biometric Product Testing Final
Report”, CESG/BWG Biometric Test Programme. www.cesg.gov.uk, Mar 2001.
[57] Bolme, D. S., Beveridge, J. R., Teixeira, M. and Draper, B. A., The CSU Face Identification
Evaluation System: Its Purpose, Features, and Structure, ICVS, Graz, Austria, 2003.
[58] V. Jain, Human face classification using neural networks, Project at IITK, Kanpur, India, 1998.
[59] Sebastien M., Yann R., and Guillaume H., On the recent use of local binary patterns for face
authentication, IDIAP, Valais, Swiss, 2006.
[60] J. Zhang, Y. Yan and M. Lades, Face Recognition: Eigenface, Elastic Matching, and Neural Nets,
Proceedings of the IEEE, Vol. 85, No. 9, pp. 1423-1435, 1997.
[61] Alan Brooks and Li Gao, Face recognition: Eigenface and Fisherface performance across pose, Final project
report, Northwestern University, Evanston, IL, 2004.
[62] Y. Zhang, L. Lang, and O. Hamsici, Facial image recognition by subspace learning: a comparative
study, Project at the Ohio State Univ., Ohio, USA, 2004.
[63] Qi Li, Jieping Ye and Chandra Kambhamettu, Linear projection methods in face recognition under
unconstrained illuminations: A comparative study, IEEE Computer Society Conference on Computer
Vision and Pattern Recognition (CVPR'04), Washington, USA, 2004.
[64] Georghiades, A.S., Belhumeur, P.N., and Kriegman, D.J., From few to many: illumination cone
models for face recognition under variable lighting and pose, IEEE Patt. Anal. Mach. Intell., 23(6),
643-660, 2001.
[65] M. J. Lyons, S. Akamatsu, M. Kamachi, and J. Gyoba, Coding facial expressions with gabor
wavelets, Proceedings, Third IEEE Int’l Conf. on Automatic Face and Gesture Recognition, Nara,
Japan, 1998.
[66] Levine, M. D., Gandhi, Maulin R. and Bhattacharyya, Jisnu: “Image Normalization for Illumination
Compensation in Facial Images”. Department of Electrical & Computer Engineering & Center for
Intelligent Machines, McGill University, Montreal, Canada, Unpublished Report (2004)
[67] Yang, J., Chen, X. Kunz, W. and Kundra, H., "Face as an index: Knowing who is who using a PDA",
Inter. Journal of Imaging Systems and Technology, 13(1), pp. 33-41, 2003.
[68] Jebara, T., 3D Pose estimation and normalization for face recognition, Honours Thesis, McGill
University, Canada, 1996.
[69] Dubuisson, S., Davoine, F. and Masson, M., A solution for facial expression representation and
recognition. Signal Process. Image Commun. 17(9). 657-673, 2002.
123
[70] N. Ikizler, J. Vasanth, L. Wong and D. Forsyth, Finding Celebrities in Video, Technical Report,
EECS Department, University of California, Berkeley, USA, 2006.
http://www.eecs.berkeley.edu/Pubs/TechRpts/2006/EECS-2006-77.html
[71] B. Du, S. Shan, L. Qing, W. Gao, “Empirical comparisons of several preprocessing methods for
illumination insensitive face recognition”, Proceedings ICASSP'05, V(2), 2005.
[72] Mauricio Villegas Santamar´_a and Roberto Paredes Palacios, “Comparison of Illumination
Normalization Methods for Face Recognition”, Third COST 275 Workshop Biometrics on the
Internet, Univ. of Hertfordshire, UK, 2005.
[73] Hammal, Z., Eveno, N., Caplier, A., and Coulon, P. 2006. Parametric models for facial features
segmentation, Signal Process., 86 (2), pp. 399-413, Feb 2006.
[74] R. C. Gonzales and R. E. Woods, Digital Image Processing, Addison Wesley Publishing Company,
Inc., New York, 1993.
[75] W. Beaudot, “The neural information processing in the vertebra retina: a melting pot of ideas for
artificial vision”, Phd thesis, tirf laboratory, Grenoble, France, 1994.
[76] Z. Hammal, C. Massot, G. Bedoya, and A. Caplier, "Eyes segmentation applied to gaze direction and
vigilance estimation," Proc. ICAPR '05, 236-246, UK, 2005.
[77] A. S. El-Sayed, K. A. Nagaty, T. I. El-Arief, “An Enhanced Histogram Matching Approach using the
Retinal Filter’s Compression Function for Illumination Normalization in Face Recognition”,
ICIAR’08, Springer-Verlag LNCS 5112, pp. 873–883, Portugal, 2008.
[78] Phillips P. J., Grother P., Micheals R. J, Blackburn D.M., Tabassi E., and Bone J. M. “FRVT 2002:
Evaluation Report”, http://www.frvt.org/DLs/FRVT_2002_Evaluation_Report.pdf, Mar 2003.
[79] Peter Belhumeur. “Ongoing Challenges in Face Recognition”, pp. 5-14, Frontiers of Engineering:
Reports on Leading-Edge Engineering from the 2005 Symposium, ISBN-10: 0-309-10102-6, 2006
[80] Heusch, G., Rodriguez, Y., and Marcel, S., “Local Binary Patterns as an Image Preprocessing for
Face Authentication”, In Proceedings of the 7th international Conference on Automatic Face and
Gesture Recognition (Fgr06) – Vol. 00, Washington, Apr 2006.
[81] R. Basri and D.W. Jacobs, “Lambertian reflectance and linear subspaces”, IEEE Trans. on Pattern
Analysis and Machine Intelligence, 25(2):218–233, 2003.
[82] Gross, R., Matthews, I., and Baker, S., “Eigen Light-Fields and Face Recognition Across Pose”. In
Proceedings of the Fifth IEEE international Conference on Automatic Face and Gesture Recognition,
Washington, USA, 2002.
[83] T.Sim, T.Kanade, Combining Models and Exemplars for Face Recognition: An Illuminating
Example, In Proceedings of Workshop on Models versus Exemplars in Computer Vision, CVPR
2001.
[84] Shan, S., Gao, W., Cao, B., and Zhao, D. “Illumination Normalization for Robust Face Recognition
Against Varying Lighting Conditions”, In Proceedings of the IEEE international Workshop on
Analysis and Modeling of Faces and Gestures AMFG, Washington, Oct 2003.
[85] R. Gross and V. Brajovic. An image preprocessing algorithm for illumination invariant face
recognition. In Audio- and Video-Based Biometric Person Authentication, AVBPA’03, 2003.
[86] T. Ojala, M. Pietik¨ainen, and D. Harwood. A comparative study of texture measures with
classification based on featured distributions. Pattern Recognition, 29(1):51–59, 1996.
[87] T. Ojala, M. Pietik¨ainen, and T. M¨aenp¨a¨a. Multiresolution gray-scale and rotation invariant
texture classification with local binary patterns. IEEE Trans. on Pattern Analysis and Machine
Intelligence, 24(7):971–987, 2002.
[88] J. Short, J. Kittler, and K. Messer. Photometric normalization for face verification. In Audio- and
Video-based Biometric Person Authentication, AVBPA’05, 2005.
[89] D. J. Jobson , Z. Rahman, G. A. Woodell, “A Multiscale Retinex for Bridging the Gap Between
Color Images and the Human Observation of Scenes,” IEEE Transactions on Image Processing,
Volume: 6, No: 3, Page(s): 965-976, Jul 1997.
[90] E. Land, “The Retinex Theory of Color Vision,” Scientific American, Page(s): 108-129, Dec 1977.
[91] H. Ando, N. Fuchigami, M. Sasaki and A. Iwata, "Robust Face Recognition Methods under
Illumination Variations toward Hardware Implementation on 3DCSS", The Fourth Hiroshima
International Workshop on Nanoelectronics for Tera-Bit Information Processing, pp. 139-141, Sep
2005.
[92] V. Vapnik. The Nature of Statistical Learning Theory. Springer, New York, 1995.
124
[93] A. Brunelli and T. Poggio, Face Recognition: Features vs Templates, IEEE Trans. Pattern Anal.
Mach. Intelligence, 15(10):1042-1053, 1993.
[94] P. Belhumeur, D. Kriegman and A. Yuille, The bas-relief ambiguity, In proc. IEEE Conf. on Comp.
Vision and Patt. Recog., p.p 1040-1046, 1997.
[95] A.Georghiades, D. Kriegman, and P. Belhumeur, “From Few to Many: Generative Models for
Recognition under Variable Pose and Illumination,” IEEE Trans. Pattern Analysis and Machine
Intelligence, vol. 40, pp. 643-660, 2001.
[96] Kuang-Chih Lee, Jeffrey Ho, David J. Kriegman, "Acquiring Linear Subspaces for Face Recognition
under Variable Lighting," IEEE Transactions on Pattern Analysis and Machine
Intelligence ,vol. 27, no. 5, pp. 684-698, May 2005.
[97] H. Chen, P. Belhumeur, and D. Jacobs, “In Search of Illumination Invariants,” Proc. IEEE Conf.
Computer Vision and Pattern Recognition, pp. 1-8, 2000.
[98] L. Zhang and D. Samaras. Face recognition under variable lighting using harmonic image exemplars.
In CVPR, volume 1, pages 19–25. IEEE Computer Society, 2003.
[99] M. Savvides, B. V. K. Vijaya Kumar, and P. L. Khosla, “Corefaces - Robust Shift Invariant PCA
based Correlation Filter for Illumination Tolerant Face Recognition” CVPR, 2004.
[100] A.Shashua, and T. Riklin-Raviv, “The quotient image: Class-based re-rendering and recognition with
varying illuminations” IEEE Tran. PAMI, Vol. 23(2), pp. 129-139, 2001.
[101] H. Wang, S. Z. Li, and Y. Wang, “Generalized Quotient Image” CVPR, 2004.
[102] Sim, T., Baker, S., Bsat, M.: The CMU Pose, Illumination, and Expression (PIE) database. In: IEEE
Int. Conf. on Automatic Face and Gesture Recognition, 2002.
[103] W. Gao, B. Cao, S. Shan, D. Zhou, X. Zhang, D. Zhao, “The CAS-PEAL Large-Scale Chinese Face
Database and Baseline Evaluations”, IEEE Transactions on Systems, Man and Cybernetics, Part A,
38(1), pp. 149-161, 2008. http://www.jdl.ac.cn/peal/index.html.
[104] Blackburn, D., Bone, M., Philips, P.: Facial recognition vendor test 2000: evaluation report (2000)
[105] Yin, W. Total Variation Models for Variable Lighting Face Recognition. IEEE Trans. Pattern Anal.
Mach. Intell. 28, 9, 2006.
[106] Peter Halinan, Alan Yeuille, and David Mumford, Harvard face Database.
[107] Haitao Wang, Stan Z Li, Yangsheng Wang, "Face Recognition under Varying Lighting Conditions
Using Self Quotient Image," fg, p. 819, Sixth IEEE International Conference on Automatic Face and
Gesture Recognition (FG'04), 2004.
[108] Press, W., Teukolsky, S., Vetterling, W., Flannery, B.: Numerical Recipes in C. Cambridge
University Press, 1992.
[109] X. Tan and B. Triggs. Enhanced local texture feature sets for face recognition under difficult lighting
conditions. In IEEE Conf. on AMFG, pages 168--182, 2007.
[110] Phillips, P.J., Flynn, P.J., Scruggs, W.T., Bowyer, K.W., Chang, J., Hoffman, K., Marques, J.,Min,
J.,Worek, W.J.: Overview of the face recognition grand challenge. In: Proc. CVPR 2005, San Diego,
CA, pp. 947–954, 2005.
[111] Borgefors, G.: Distance transformations in digital images. Comput. Vision Graph. Image Process.
34(3), 344–371, 1986.
[112] F. Samaria and A. Harter, Parameterization of a stochastic model for human face identification, Proc.
2nd IEEE Workshop on Applications of Computer Vision, Sarasota, USA, 1994.
[113] “Introduction to Biometrics” section of the Biometrics Catalog, http://www.biometricscatalog.org.
[114] “Face Recognition Vendor Test”, www.FRVT.org.
[115] Grimace database, http://cswww.essex.ac.uk/mv/allfaces/grimace.html.
[116] Intel OpenCV Library, http://sourceforge.net/projects/opencvlibrary/.
[117] Nott-faces database, http://pics.psych.stir.ac.uk.
[118] Face 94 database, http://cswww.essex.ac.uk/mv/allfaces/face94.html.
[119] Yale B face database, http://cvc.yale.edu/projects/yalefacesB/yalefacesB.html.
[120] “The Facial Recognition Technology (FERET) Database”,
http://www.itl.nist.gov/iad/humanid/feret/feret_master.html.
[121] Face normalization and descriptor code, http://lear.inrialpes.fr/people/triggs/src/amfg07-demo-
v1.tar.gz.
[122] AcSys Biometrics, http://www.acsysbiometrics.com.
[123] Berninger Software BmbH, http://www.berningersoftware.de.
[124] Biometrica Systems, Inc., http://www.biometrica.com.
125
[125] Cognitec Systems, http://cognitec-systems.de.
[126] Dialog Communication Systems Inc., http://www.bioid.com.tw/.
[127] FaceKey Corp., http://www.facekey.com.
[128] Humanscan GmbH, http://www.bioid.com.
[129] Identix, Inc., http://www.identix.com.
[130] Imagis Technologies, Inc., http://www.imagistechnologies.com.
[131] Keyware Technologies N.V., http://www.keyware.com.
[132] Neurodynamics Limited, http://www.neurodynamics.com.
[133] Viisage Technology, http://www.viisage.com/facetools.htm.
[134] VisionSphere Technologies, http://www.visionspheretech.com.
[135] ZN Vision Technologies, http://www.zn-ag.com.
126
ﻣﻠﺨﺺ اﻟﺮﺳﺎﻟﺔ
ﺑﺎﻟﺮﻏﻢ ﻣﻦ ان اﻟﻌﺪﻳﺪ ﻣﻦ ﻧﻈﻢ وﺗﻘﻨﻴﺎت اﻟﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ ﻗﺪ اﻗﺘﺮﺣﺖ ﺧﻼل اﻟﺴﻨﻮات اﻟﻤﺎﺿﻴﺔ ،إﻻ أن أﺣﺪث ﺗﻘﻴﻴﻢ ﻟﻬﺬﻩ
اﻟﻨﻈﻢ واﻟﺘﻘﻨﻴﺎت ﻗﺪ أﻇﻬﺮ أن أداء ﻣﻌﻈﻤﻬﺎ ﻳﺘﺄﺛﺮ ﺳﻠﺒًﺎ ﻧﻈﺮًا ﻟﺘﻐﻴﺮات اﻹﺿﺎءة .ﻓﻲ ﺁﺧﺮ اﺧﺘﺒﺎر ﻟﻬﺬﻩ اﻟﻨﻈﻢ ﻋﺎم 2006
ﺧﻠﺺ اﻟﻰ ان ﺗﺨﻔﻴﻒ ﺷﺮوط اﻹﺿﺎءة ﻟﻪ ﺗﺄﺛﻴﺮ آﺒﻴﺮ ﻋﻠﻰ ﺁداء هﺬﻩ اﻟﻨﻈﻢ .وﻋﻼوة ﻋﻠﻰ ذﻟﻚ ،ﺛﺒﺖ ﻧﻈﺮﻳًﺎ وﺑﺎﻟﺘﺠﺮﺑﺔ أن
اﻻﺧﺘﻼﻓﺎت ﺑﻴﻦ ﺻﻮر ﻧﻔﺲ اﻟﺸﺨﺺ ﺑﺴﺒﺐ اﻹﺿﺎءة ﻳﻜﻮن ﻏﺎﻟﺒًﺎ أآﺒﺮ ﻣﻦ اﻻﺧﺘﻼﻓﺎت اﻟﺘﻰ ﺗﺤﺪث ﻧﺘﻴﺠﺔ ﻟﻠﺘﻐﻴﻴﺮ ﻓﻲ
هﻮﻳﺔ هﺬا اﻟﺸﺨﺺ.
هﻨﺎك اﻟﻜﺜﻴﺮ ﻣﻦ اﻷﺑﺤﺎث واﻟﻄﺮق اﻟﻤﺨﺘﻠﻔﺔ ﻟﻠﺘﻌﺎﻣﻞ ﻣﻊ ﺗﺄﺛﻴﺮات اﻹﺿﺎءة ﻓﻲ اﻟﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ .ﻋﻠﻰ اﻟﺮﻏﻢ ﻣﻦ ان
ﻣﻌﻈﻢ هﺬﻩ اﻟﻄﺮق ﻳﻤﻜﻨﻬﺎ اﻟﺘﻐﻠﺐ ﻋﻠﻰ اﺧﺘﻼﻓﺎت اﻹﺿﺎءة ﺑﺸﻜﻞ ﻓﻌﺎل ،إﻻ أن ﺑﻌﻀﻬﺎ ﻗﺪ ﻳﻜﻮن ﻟﻪ ﺗﺄﺛﻴﺮ ﺳﻠﺒﻲ ﻋﻠﻰ
اﻟﺼﻮر اﻟﺘﻲ ﻟﻢ ﺗﺘﻌﺮض ﻻﺧﺘﻼﻓﺎت ﺑﺴﺒﺐ اﻹﺿﺎءة .واﻟﺒﻌﺾ اﻵﺧﺮ ﻳﻈﻬﺮ ﻓﺮق آﺒﻴﺮ ﻓﻲ اﻵداء ﻋﻨﺪﻣﺎ ﻳﻘﺘﺮن ﺑﻄﺮق
ﻣﺨﺘﻠﻔﺔ ﻟﻠﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ .وهﻨﺎك اﻟﺒﻌﺾ اﻵﺧﺮ ﻣﻦ هﺬﻩ اﻟﻄﺮق ﻳﺘﻄﻠﺐ ﻣﻮاءﻣﺔ اﻟﻮﺟﻪ ﺿﻤﻦ اﻟﺼﻮرة وهﻮ أﻣﺮ ﻗﺪ
ﻳﺼﻌﺐ ﺗﺤﻘﻴﻘﻪ ﻓﻲ اﻟﻨﻈﻢ اﻟﻌﻤﻠﻴﺔ اﻟﺤﻘﻴﻘﻴﺔ.
ﻟﺬﻟﻚ ﻓﺈﻧﻨﺎ ﻧﻘﺘﺮح ﻓﻲ هﺬﻩ اﻟﺮﺳﺎﻟﺔ ﻃﺮﻳﻘﺔ ﺟﺪﻳﺪة ﻟﻠﺘﻌﺎﻣﻞ ﻣﻊ اﺧﺘﻼﻓﺎت اﻹﺿﺎءة ﺑﺤﻴﺚ ﻻ ﺗﻈﻬﺮ ﻓﺮق آﺒﻴﺮ ﻓﻲ اﻵداء
ﻋﻨﺪﻣﺎ ﺗﻘﺘﺮن ﺑﻄﺮق ﻣﺨﺘﻠﻔﺔ ﻟﻠﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ وآﺬﻟﻚ ﻻ ﺗﺸﺘﺮط ﻣﻮاءﻣﺔ اﻟﻮﺟﻪ ﺿﻤﻦ اﻟﺼﻮرة ﻣﻤﺎ ﻳﺴﻬﻞ اﺳﺘﺨﺪاﻣﻬﺎ
ﻓﻲ اﻟﻨﻈﻢ اﻟﻌﻤﻠﻴﺔ اﻟﺤﻘﻴﻘﻴﺔ.
ﻟﻠﺘﺤﻘﻖ ﻣﻦ أن اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ ﻻ ﺗﻈﻬﺮ ﻓﺮق آﺒﻴﺮ ﻓﻲ اﻵداء ﻋﻨﺪ اﻗﺘﺮاﻧﻬﺎ ﺑﻄﺮق ﻣﺨﺘﻠﻔﺔ ﻟﻠﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ وآﺬﻟﻚ
ﻋﺪم اﺷﺘﺮاﻃﻬﺎ ﻟﻤﻮاءﻣﺔ اﻟﻮﺟﻪ ﺿﻤﻦ اﻟﺼﻮرة ،ﻓﺈﻧﻪ ﻗﺪ ﺗﻢ اﺧﺘﺒﺎرهﺎ ﻋﻠﻰ ﻃﺮﻳﻘﺘﻴﻦ ﻣﺨﺘﻠﻔﺘﻴﻦ ﺗﻤﺜﻼن ﻓﺌﺘﻴﻦ آﺒﻴﺮﺗﻴﻦ ﻣﻦ
ﻼ وﻟﻴﺲ أﺟﺰاء ﻣﻨﻪ .ﻓﻲ آﻞ ﻃﺮﻳﻘﺔ ،ﺗﻢ اﺳﺘﺨﺪام ﻗﺎﻋﺪة ﻃﺮق اﻟﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ واﻟﻤﻌﺘﻤﺪة ﻋﻠﻰ دراﺳﺔ اﻟﻮﺟﻪ آﺎﻣ ً
ﺑﻴﺎﻧﺎت ﻣﻦ اﻟﺼﻮر اﻟﺘﻲ ﺗﺤﺘﻮي ﻋﻠﻰ اﺧﺘﻼﻓﺎت ﺷﺪﻳﺪة ﻓﻲ اﻷﺿﺎءة ﻻﺧﺘﺒﺎر اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ ﻣﺮﺗﻴﻦ ،ﺑﺤﻴﺚ ﻳﻜﻮن
اﻟﻮﺟﻪ ﻣﻮاءﻣًﺎ ﺿﻤﻦ اﻟﺼﻮرة ﻓﻲ اﻟﻤﺮة اﻷوﻟﻰ وﻻ ﻳﻜﻮن ﻣﻮاءﻣًﺎ ﻓﻲ اﻟﻤﺮة اﻟﺜﺎﻧﻴﺔ وذﻟﻚ ﺣﺘﻰ ﻧﺘﻤﻜﻦ ﻣﻦ اﻟﺤﻜﻢ ﻋﻠﻰ
اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ وﻣﺪى اﺣﺘﻴﺎﺟﻬﺎ ﻟﻬﺬا اﻟﺸﺮط ﻣﻦ ﻋﺪﻣﻪ.
ﻣﻦ أﺟﻞ ﻣﻘﺎرﻧﺔ اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ ﻣﻊ ﻃﺮق أﺧﺮى ﻣﻨﺎﻇﺮة ،ﻓﺈﻧﻨﺎ ﻗﻤﻨﺎ ﺑﻌﻤﻞ ﻣﺴﺢ ﻟﺘﺴﻌﺔ ﻣﻦ دراﺳﺎت اﻟﻤﻘﺎرﻧﺔ وﻗﻤﻨﺎ
ﺑﺎﺧﺘﻴﺎر أﻓﻀﻞ أرﺑﻌﺔ ﻃﺮق ﻣﻦ ﺑﻴﻦ 38ﻃﺮﻳﻘﺔ ﻣﺨﺘﻠﻔﺔ ﺷﻤﻠﺘﻬﻢ هﺬﻩ اﻟﺪراﺳﺎت .ﺟﻤﻴﻊ اﻟﻄﺮق اﻟﺨﻤﺲ ﺗﻢ اﺧﺘﺒﺎرهﺎ
وﻣﻘﺎرﻧﺘﻬﺎ ﺑﺎﺳﺘﺨﺪام ﻃﺮﻳﻘﺘﻲ اﻟﺘﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ ،واﻟﺘﻲ ﺗﻢ اﺧﺘﻴﺎرهﻤﺎ ﻣﺴﺒﻘﺎً ،وذﻟﻚ ﻋﻠﻰ ﺻﻮر ﺑﻬﺎ اﺧﺘﻼﻓﺎت ﺑﺴﺒﺐ
اﻹﺿﺎءة وﺻﻮر أﺧﺮى ﺑﻬﺎ اﺧﺘﻼﻓﺎت ﻏﻴﺮ اﻹﺿﺎءة.
أﻇﻬﺮت اﻟﻨﺘﺎﺋﺞ ﻋﻠﻰ اﻟﺼﻮر اﻟﺘﻲ ﺑﻬﺎ اﺧﺘﻼﻓﺎت ﺑﺴﺒﺐ اﻹﺿﺎءة أن اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ آﺎﻧﺖ اﻷﻓﻀﻞ ﻣﻦ ﺑﻴﻦ اﻟﻄﺮق
اﻷﺧﺮى ﻋﻠﻰ ﻃﺮﻳﻘﺔ اﻟﺘﻌﺮف اﻷوﻟﻰ وﺛﺎﻧﻲ أﻓﻀﻞ ﻃﺮﻳﻘﺔ ﻋﻠﻰ ﻃﺮﻳﻘﺔ اﻟﺘﻌﺮف اﻟﺜﺎﻧﻴﺔ وذﻟﻚ ﻋﻨﺪﻣﺎ ﻳﻜﻮن اﻟﻮﺟﻪ ﻏﻴﺮ
ﻣﻮاءﻣًﺎ ﺿﻤﻦ اﻟﺼﻮرة .إﺿﺎﻓﺔ إﻟﻰ ذﻟﻚ ،آﺎﻧﺖ اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ اﻷﻗﻞ ﺗﺄﺛﺮًا ﻓﻲ ﻧﺘﺎﺋﺠﻬﺎ ﻣﻦ ﺑﻴﻦ اﻷرﺑﻊ ﻃﺮق اﻷﺧﺮى
ﻧﺘﻴﺠﺔ ﻟﻌﺪم ﻣﻮاءﻣﺔ اﻟﻮﺟﻪ ﺿﻤﻦ اﻟﺼﻮرة وذﻟﻚ ﻋﻠﻰ آﻠﺘﺎ ﻃﺮﻳﻘﺘﻲ اﻟﺘﻌﺮف اﻟﻤﺴﺘﺨﺪﻣﺘﻴﻦ .أﻣﺎ ﻋﻠﻰ اﻟﺼﻮر اﻟﺘﻲ ﺑﻬﺎ
اﺧﺘﻼﻓﺎت أﺧﺮى ﻏﻴﺮ اﻹﺿﺎءة ،ﻓﻘﺪ أﻇﻬﺮت اﻟﻨﺘﺎﺋﺞ أن اﻟﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ ﻟﻬﺎ أﻗﻞ ﺗﺄﺛﻴﺮ ﺳﻠﺒﻲ ﻣﻦ ﺑﻴﻦ اﻷرﺑﻊ ﻃﺮق
اﻷﺧﺮى وذﻟﻚ أﻳﻀًﺎ ﻋﻠﻰ آﻠﺘﺎ ﻃﺮﻳﻘﺘﻲ اﻟﺘﻌﺮف.
ﺟﻤﻴﻊ ﻃﺮق اﻟﺘﻌﺎﻣﻞ ﻣﻊ اﺧﺘﻼﻓﺎت اﻹﺿﺎءة ﻓﻲ هﺬا اﻟﻌﻤﻞ ﺗﻢ اﺧﺘﺒﺎرهﺎ ﻋﻠﻰ ﻃﺮﻳﻘﺘﻲ ﺗﻌﺮف ﻋﻠﻰ اﻟﻮﺟﻪ ﺗﻤﺜﻼن ﻓﺌﺘﻴﻦ
ﻼ وﻟﻴﺲ أﺟﺰاء ﻣﻨﻪ .ﻟﺬﻟﻚ ﻓﺈﻧﻪ ﻣﻦ اﻟﻤﻬﻢ أن ﻳﺘﻢ ﺗﻮﺳﻴﻊ هﺬا اﻟﻌﻤﻞ
آﺒﻴﺮﺗﻴﻦ ﻣﻦ اﻟﻄﺮق اﻟﻤﻌﺘﻤﺪة ﻋﻠﻰ دراﺳﺔ اﻟﻮﺟﻪ آﺎﻣ ً
127
ﻟﻴﺸﻤﻞ اﻟﻄﺮق اﻟﻤﻌﺘﻤﺪة ﻋﻠﻰ دراﺳﺔ أﺟﺰاء ﻣﻦ اﻟﻮﺟﻪ أﻳﻀًﺎ وذﻟﻚ ﻷﻧﻪ ﻗﺪ ﻳﺤﺪث ﻓﺮق آﺒﻴﺮ ﻓﻲ اﻷداء ﻋﻨﺪﻣﺎ ﺗﻘﺘﺮن هﺬﻩ
اﻟﻄﺮق ﻣﻊ ﻃﺮق اﻟﺘﻌﺎﻣﻞ ﻣﻊ اﺧﺘﻼﻓﺎت اﻹﺿﺎءة.
وﻋﻼوة ﻋﻠﻰ ذﻟﻚ ،ﻳﻘﺪم هﺬا اﻟﻌﻤﻞ ﺗﻘﻴﻴﻤًﺎ ﺗﻜﻨﻮﻟﻮﺟﻴًﺎ ﻟﻠﻄﺮﻳﻘﺔ اﻟﻤﻘﺘﺮﺣﺔ واﻟﻄﺮق اﻷﺧﺮى اﻟﻤﺨﺘﺎرة .ﻟﺬﻟﻚ ﻓﺈﻧﻪ ﻣﻦ اﻟﻤﻬﻢ
ﻋﻤﻞ ﺗﻘﻴﻴﻢ ﺳﻴﻨﺎرﻳﻮ وآﺬﻟﻚ ﺗﻘﻴﻴﻢ ﺗﻨﻔﻴﺬي ﻷداء هﺬﻩ اﻟﻄﺮق ﻣﻦ أﺟﻞ اﻧﺠﺎز ﺗﻘﻴﻴﻢ ﺷﺎﻣﻞ ﻟﻬﺎ.
128
ﺟﺎﻣﻌﺔ ﻋﻴﻦ ﺷﻤﺲ
آﻠﻴﺔ اﻟﺤﺎﺳﺒﺎت و اﻟﻤﻌﻠﻮﻣﺎت
ﻗﺴﻢ ﻋﻠﻮم اﻟﺤﺎﺳﺐ
إﻋـﺪاد
ﺗﺤﺖ اﺷﺮاف
2009
129