Professional Documents
Culture Documents
1. Introduction
Face recognition is an important subfield of pattern recognition, and it is one of the essential person identification methods since it is contactless. During the past two decades, face recognition has attracted much effort of researchers from research fields such as image processing, computer vision and pattern recognition. To date, a lot of approaches have been proposed for face recognition, the prominent method among which is the subspace method. The well-known Eigenfaces [1] and Fisherfaces [2] are some typical examples of this kind of method. In the Eigenfaces method, PCA is used for projecting the image space to a low dimensional feature space. The PCA based methods for face recognition have been developed since then. Recently, Yang and Yang [3] and Yang et al. [4] proposed two-dimensional PCA (2DPCA). Different from PCA, 2DPCA is founded upon 2D image matrix rather than 1D vectors. An image covariance matrix can be constructed directly by using the original image matrices. And a feature matrix for each training face sample can be computed easily. It has been proved that 2DPCA can give a better performance than PCA since the covariance matrix in 2DPCA can be computed more
Corresponding author This work was supported by the National Science Foundation of China Grant No.60573158 and No 10771213..
1
783 777
computing the similarity measure between two face images. The learnt similarity measure adapts to the special face data set. Therefore, the proposed classification measure can reject the influence of undesirable changes in a face pattern to some extent. In this way, we can obtain a better recognition performance. The remainder of this paper is organized as follows. In Section 2, the 2DPCA-based face recognition method, the previous works on similarity measures and the simulated annealing algorithm are introduced briefly. And the idea of using simulated annealing for learning a distance measure is described in Section 3. In Section 4, experimental results are presented and discussed. Finally, conclusions are presented in section 5.
Yi = [ yi1 yi 2 yid ] for each training sample can be computed by yik = X i wk , where k = 1, 2, , d .
Also, we can get a feature matrix
2. Related works
2.1. Review of 2DPCA based face recognition
Given a training face set { X 1 , X 2 , , X N } , 2DPCA uses all training samples to compute the following image total covariance matrix C .
dY (Yi , Y j ) = yik y jk
k =1
(6)
(7)
These similarity measures mentioned above are fixed and can not adapt to the special data set. We will present a new similarity measure based on simulated annealing in section 3.
J ( w) = tr (S w ), (2) where S w is the covariance matrix of the projected vectors of training images, and tr ( S w ) denotes the trace of S w .The covariance matrix S w can be written
as
(3)
(4) the optimal projection vectors w1 , w2 , , wd are the orthonormal eigenvectors of Factually, the image total covariance matrix C corresponding to the first d largest eigenvalues. Thus a feature matrix
J ( w) = wT Cw.
778 784
not an optimal choice to use the sum of all Algorithm 1. Simulated Annealing Begin initialize Tmax , lmax , kmax , T (l ) , x0 , N ( x)
We will learn an optimal similarity measure using the simulated annealing algorithm. Given a group of weights wk , k = 1, , d , we compute the following similarity measure.
d (Yi , Y j ) = wk yik y jk
k =1
/ w. 2 k
k =1
(8)
x ' := N ( x )
2. Compute cost value of x and x ' , and compare them
c = c( x ') c( x)
3. If c < 0 then x := x ' else if exp( c / T ) > rand then x := x ' end if end do
In the simulated annealing algorithm, the training error of the nearest neighbor classifier using (8) as similarity measure is employed as the cost function. During the procedure of annealing, the training error is minimized. And a group of weights is learnt. We linearly combine yik y jk
2
, k = 1, , d accord-
ing to the learnt weights into the optimal similarity measure. The algorithm is presented detailed in Algotithm 2. Algorithm 2. Simulated Annealing for Learning a Similarity Measure Begin initialize Tmax , lmax , kmax , T (l ) , w0 ,
T = T (l )
end do Output x
N ( x) , T = Tmax , w = w0
3. Using simulated annealing for learning a similarity measure
Assume a training face set containing N face samples is given as { X 1 , X 2 , , X N } , where X i is the ith training sample, which is a m n matrix. The feature matrix for X i obtained by 2DPCA is denoted as where w = ( w1 , , wd ) ,and w0 is the randomly selected d dimensional feasible solution. Do for l = 1, , lmax Do for k = 1, , kmax 1. Select a candidate solution randomly from the neighborhood of w , we define the neighborhood as changing one random elements of w initial
T
Yi = [ yi1 yi 2 yid ]
i = 1, 2, , N
where yik , ( k = 1, , d ) is a m dimensional vector. Given the feature matrices Yi and Y j of face matrices
w ':= N (w)
2. Compute cost value of w and w ' , which is the training error of the nearest neighbor classifiers using (8) as similarity measure, and compare them
yik y jk , k = 1, , d
2
carries some information for discriminating face X i and face X j . Yang et al. uses the sum of all
, k = 1, , d to
T = T (l )
end do Output w
end if end do
c = c( w ') c (w) 3. If c < 0 then w := w ' else if exp( c / T ) > rand then w := w '
779 785
(9)
/ w* . 2 k
k =1
(10)
Given the testing feature matrix Yt for each testing face sample and the feature matrix Yi for each training face sample after the transformation by 2DPCA, a nearest neighbor classifier based on the similarity measure (10) can be depicted for the 2DPCA based face recognition as follows.
training, while the remaining five images are used for testing. Fig. 1 shows some images used in the simu-lation. We run four kinds of 2DPCA based methods (Frobenius distance, Yang distance, volume measure, SA similarity measure) on this data set, and Fig. 2 plots the recognition rates according to a different number of projection vectors when each method is applied. The number of projection vectors and recognition rates in the best case are shown in Table 1.
0.99 0.98 Recognition rate 0.97 0.96 0.95 0.94 0.93 2 4 6 8 Number of projection vectors 10 Frobenius distance Yang distance Volume measure SA similarity measure
c = arg min d SA (Y j , Yi )
i * = arg min wk yik y jk i k =1 d 2 * / wk , k =1 d
(11)
where c [1, 2, , N ] , and Y j is classified into the class where Yc belongs to.
4. Experiments
To evaluate the proposed algorithm, we use two well-known face databases, the ORL face database [17] and the YALE face database [18] for experiments.
Fig. 2. Performance of methods with different distance measures using the ORL face database. Table 1. The best recognition rates and the corresponding number of projection vectors on the ORL face database
Fig. 1. Some images of the first individual in the ORL face database. The images above are for training, and the images below are for testing.
The table shows a fact that the maximal number of projection vector usually does not give a best recognition rate. We use the log temperature decreasing function [19] in our experiments, which is shown in Fig. 3. The parameters used in our experiments are shown in Table 2. How the value of cost function changes according to the number of iteration is shown in Fig. 4. The plot shows the simulated annealing algorithm can obtain the global optimal solution after about 600 iterations. From Fig. 2, we can see that the performance of the 2DPCA based methods using SA similarity measure is better than the methods using Frobenius distance measure and Yang distance measure. When the number of projection vectors exceeds two, the 2DPCA based methods using boosted similarity
780 786
measure provides better recognition rates than the method using Volume measure.
500
results are shown in Fig. 6. The number of projection vectors and recognition rates in the best case are shown in Table 3.
400 Temperatures
300
200
100
Fig. 5. Some images of the first individual in the YALE face database. The images above are for training, and the images below are for testing.
0 100 200 300 Number of iterations 400 500
1.1 1.05 1 Recognition rate 0.95 0.9 0.85 0.8 0.75 0.7 2 4 6 8 Number of projection vectors 10 Yang distance Frobenius distance Volume measure SA similarity measure
Fig. 3. The log temperature decreasing function Table 2. The parameters used in our experiments
Tmax
500
lmax
1000
kmax
10
Fig. 6. Performance of methods with different distance measures using the YALE face database. Table3. The best recognition rates and the corresponding number of projection vectors on the Yale face database
Fig. 4. The values of cost function corresponding to the number of iterations, obtained using the simulated annealing algorithm.
The table shows the SA similarity measure outperforms others in the best case. From Fig. 6, we notice that the 2DPCA based methods using SA similarity measure gives better performance than the methods using Frobenius distance measure and Yang distance measure. When the number of projection vectors exceeds five, the 2DPCA based methods using SA similarity measure outperforms the method using Volume measure. Fig. 6 also indicates that generally the performance of the 2DPCA based method using Yang distance measure is better than that using Frobenius distance measure, and the
781 787
5. Conclusion
In this paper we proposed a 2DPCA based method using a new similarity measure learnt by the simulated annealing algorithm for face recognition. The main area of novelty is that it is presented how the simulated annealing algorithm can be used for learning a similarity measure in 2DPCA based method for face recognition. Different from Frobenius distance measure, Yang distance measure and Volume measure, the SA is learnt on the training samples, which minimized the training error. Thus the similarity measure proposed in this paper adapts to the special face data set and can reject the influence of undesirable changes in a face pattern to some extent. The effectiveness of our method is demonstrated by simulation experiments, which shows that our method outperforms the state of the art methods in terms of recognition accuracy. However, this method still has drawbacks on the time complexity. It is an intriguing question to reduce the running time of the method. And the influence of the parameters in the simulated annealing is to be addressed in future. REFERENCES
[1]. M. Turk, A. Pentland, Eigenfaces for recognition, J. Cognitive Neurosci, 3 (1), 1991. [2]. P. Belhumeur, J. Hespanha, and D. Kriegman, Eigenfaces vs. fisherfaces: Class specific linear projection, IEEE Trans. PAMI, 19 (7): 711-720, 1997. [3]. J. Yang, J.Y. Yang, From image vector to matrix: A straightforward image projection technique IMPCA vs. PCA, Pattern Recognition 35, 19971999, 2002. [4]. J. Yang, D. Zhang, and A. Frangi, Two-dimensional PCA: A new approach to appearance-based face representation and recognition, IEEE Trans. Pattern Anal. Machine Intell. 26 (1): 131-137, 2004. [5]. H. Kong, X.C. Li, and K.T. Eam, Generalized 2D principal component analysis, Proc. IEEE Internat.
[10].
[11]. [12].
[13]. [14].
Joint Conf. on Neural Networks, Montreal, Canada, 2005. L. Wang, X. Wang, and X. Zhang, The equivalence of two-dimensional PCA to line-based PCA, Pattern Recognition Lett., 26 (1): 57-60, 2005. M. Li, B. Yuan, 2D-LDA: A statistical linear discriminant analysis for image matrix, Pattern Recognition Lett., 26 (5): 527532 , 2005. J. Yang, D. Zhang, and X. Yong, Two-dimensional discriminant transform for face recognition, Pattern Recognition, 38 (7): 11251129, 2005. J. Ye, R. Janardan, and Q. Li, Two-dimensional linear discriminant analysis, Advances in Neural Information Processing Systems, vol. 17, Vancouver, BC, Canada, 2004. U.D. Chang, Y.G. Kim, and D.W. Kim, 2D Direct LDA for Efficient Face Recognition, Lecture Notes in Control and Information Science, vol. 345, 566 571, 2006. D. Cho, U. Chang, and K. Kim, (2D)2 DLDA for Efficient Face Recognition, Lecture Notes in Computer Science, vol. 4319, 314-321, 2006. W.M. Zuo, D. Zhang, and K. Wang, An assembled matrix distance metric for 2DPCA-based image recognition, Pattern Recognition Lett., 27 (3): 210216, 2006. J. Meng, W. Zhang, Volume measure in 2DPCAbased face recognition, Pattern Recognition Lett., 28 (10): 12031208, 2007. N. Metropolis, A.W. Rosenbluth, and M. N. Rosenbluth, Equations of state calculation by fast computing machines, The Journal of Chemical Physics, 21(3): 1087-1092, 1953. I.O. Bohachevsky, M.E. Johnson, and M.L. Stein, Generalized simulated annealing for function optimization, Technometrics, 28(3): 8, 1986. S. Kirkpatrick, C. D. Gelatt, and M. P. Vecehi, Optimization by simulated annealing, Science, 220(5): 671-680, 1983. ORL, The ORL face database at the AT&T (Olivetti) Research Laboratory, 1992, Available: <http://www.uk.research.att.com/facedatabase.html>. YALE face database, 1997. Available: <http://www.uk.research.att.com/facedatabase.html>. L. Ingber, Simulated annealing: Practice versus theory, Mathematical Computer Modeling, 18: 2957, 1993.
782 788