|
1.IntroductionFace recognition (FR) has recently received extensive attention due to its broad applications. Many techniques have been developed over the past decades for FR. Subspace analysis is one of the most efficient techniques. The most popular subspace techniques are principal component analysis (PCA),1 linear discriminant analysis (LDA),2 and independent component analysis (ICA).3 Besides directly processing image appearance, subspace methods can also be combined with the Gabor feature4 to derive the Gabor-based methods.5, 6, 7 However, the recognition time of these methods depends heavily on the size of the training set because a test image is compared to all training images. Such a recognition scheme is not efficient, especially when the training set is too large. Thus, we propose integrating a novel clustering method, affinity propagation (AP),8 with LDA to form AP-LDA for FR. By using AP on the low-dimensional features obtained from LDA, a representative face image for each subject can be achieved. Thus, AP-LDA needs to use only representative face images rather than all training face images for recognition. We combine AP with LDA for the following reasons: AP can cluster data points into different clusters and detect a representative example for each cluster. By using AP, we intend to achieve a representative face image for each subject. If so, we need to use only representative face images for recognition. However, directly using AP on gray pixel values is computationally expensive and inefficient. For example, the difference between two pixels on the same positions in two images of the same subject is obvious when the illumination is different. However, they should be considered to be similar to each other in fact. Thus, more discriminating and efficient low-dimensional features should be extracted before using AP for clustering. Therefore, LDA is adopted, as it can not only be used for dimensionality reduction but also extract discriminative features. 2.Review of the AP Clustering Method and the LDA Method2.1.AP Clustering MethodAP8 first builds a similarity matrix , in which between data points and is their negative Euclidean distance . Before clustering, each data point also needs to be assigned a number , which describes the a priori knowledge of how good point is as a representative example. The data points with larger values of are more likely to be chosen as representative examples. These values are referred to as preferences. In fact, the probability of each point being the representative example is the same; thus, the preferences should be set to the same value, which can be varied to produce different numbers of clusters. Generally, such a value takes the median of the . After the construction of a similarity matrix and the setting of preferences, two kinds of messages (responsibility and availability) are passed between data points. The responsibility , sent from to candidate representative example , reflects the accumulated evidence for how proper it would be for to serve as the representative example for . It is updated using the rule: The availability , sent from candidate representative example to , reflects the accumulated evidence for how well-suited is to choose as its representative example. It is computed by the rule: It is clear that availabilities and responsibilities can be combined to recognize representative examples at any time. For , the that maximizes indicates that serves as the representative example for . 2.2.LDA MethodSuppose there is a set of -dimensional samples belonging to classes . LDA2 aims to find that maximizes the ratio of the between-class scatter matrix to the within-class scatter matrix, i.e., where and are the between-class matrix and the within-class scatter matrix: where is the mean of class , is the total sample mean, and is the size of the data points in . are generalized eigenvectors of and corresponding to the largest generalized eigenvalues , i.e.,Note that there are at most nonzero generalized eigenvalues; thus, the dimension of the reduced space is . At the same time, to make the within-class scatter matrix nonsingular, PCA is first used to reduce the dimension of the feature to , and then LDA is applied to reduce the dimension to . 3.Summary of AP-LDASuppose that there is a face data set of face images belonging to different subjects. First, face images per person (hence, in total) are selected for training using LDA, obtaining corresponding dimensional features. Second, AP is used to cluster these features into different clusters and obtain a representative feature for each cluster. Note that the number of clusters obtained by using AP may not equal the number of subjects, as the former is influenced by the values of preference,8 while the latter is fixed. To solve this problem, we repeatedly vary the value of the preferences until two such numbers equal to each other. Last, each test image is converted to a low-dimensional feature, which is then compared to representative features and identified using a nearest-neighbor classifier. Since AP-LDA uses only representative features for recognition, the recognition time of AP-LDA depends mainly on the number of subjects. It increases linearly with the , while the recognition time of the Fisherface method increases with the training image size . Obviously, our AP-LDA method is more computationally efficient than the Fisherface method. 4.Experiments and DiscussionIn this section, experiments on three benchmark face data sets [Yale9; extended Yale10; and Pose, Illumination, and Expression (PIE)11] are carried out to show the effectiveness of our AP-LDA method and also to compare it to Fisherface. All face images are cropped based on the centers of eyes such that facial areas contain only the face. All cropped images are then normalized to the size of , with 256 gray levels per pixel. In the following experiments, different numbers of images per subject were randomly selected for training, and the rest were used for testing. To minimize the possible misleading results, the final recognition rates were obtained by averaging the results over five random splits. The Yale database contains 165 images with 11 different images for each of the 15 distinct subjects. Comparative results are summarized in Table 1. It is clear that AP-LDA obviously outperforms Fisherface when the number of training images per person is more than 3. Table 1Comparative results on Yale database.
The Extended Yale database is extended from the Yale database. It contains approximately 64 near-frontal images for each of 38 distinct subjects. We randomly selected 20 images per person for our experiments. Table 2 presents the results. We can see that AP-LDA remarkably outperforms Fisherface in all cases. Table 2Comparative results on subset from Extended Yale database.
The CMU Pose, Illumination, and Expression (PIE) database contains 41,368 facial images of 68 individuals. We randomly chose 30 images per person for our experiments. The results are shown in Table 3. Again, AP-LDA consistently outperforms Fisherface. Table 3Comparative results on subset from CMU PIE database.
Comparative experimental results show that:
5.ConclusionsWe have introduced in this letter a novel AP-LDA method for face recognition. Unlike Fisherface, which uses all training face images for recognition, our AP-LDA method uses only representative face images. This makes AP-LDA more efficient. Experiments also indicate that AP-LDA outperforms Fisherface in terms of recognition rate. AcknowledgmentsThe authors would like to thank the anonymous reviewers for their critical and constructive comments and suggestions. This research has been supported by the National Natural Science Foundation of China (Grant Nos. 60675023 and 60602012). ReferencesM. A. Turk and
A. P. Pentland,
“Face recognition using eigenfaces,”
586
–591
(1991). Google Scholar
P. N. Belhumeur,
J. P. Hespanha, and
D. J. Kriegman,
“Eigenfaces versus fisherfaces: recognition using class specific linear projection,”
IEEE Trans. Pattern Anal. Mach. Intell., 19
(7), 711
–720
(1997). https://doi.org/10.1109/34.598228 0162-8828 Google Scholar
M. S. Bartlett,
J. R. Movellan, and
T. J. Sejnowski,
“Face recognition by independent component analysis,”
IEEE Trans. Neural Netw., 13
(6), 1450
–1464
(2002). https://doi.org/10.1109/TNN.2002.804287 1045-9227 Google Scholar
L. Wiskott,
J. M. Fellous,
N. Kuiger, and
C. von der Malsburg,
“Face recognition by elastic bunch graph matching,”
IEEE Trans. Pattern Anal. Mach. Intell., 19
(7), 775
–779
(1997). https://doi.org/10.1109/34.598235 0162-8828 Google Scholar
C. Liu,
“Gabor-based kernel PCA with fractional power polynomial models for face recognition,”
IEEE Trans. Pattern Anal. Mach. Intell., 26
(5), 572
–581
(2004). https://doi.org/10.1109/TPAMI.2004.1273927 0162-8828 Google Scholar
C. Liu and
H. Wechsler,
“Independent component analysis of Gabor features for face recognition,”
IEEE Trans. Neural Netw., 14
(4), 919
–928
(2003). https://doi.org/10.1109/TNN.2003.813829 1045-9227 Google Scholar
C. Liu and
H. Wechsler,
“Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition,”
IEEE Trans. Image Process., 11
(4), 467
–476
(2002). https://doi.org/10.1109/TIP.2002.999679 1057-7149 Google Scholar
B. J. Frey and
D. Dueck,
“Clustering by passing messages between data points,”
Science, 315
(5814), 972
–976
(2007). https://doi.org/10.1126/science.1136800 0036-8075 Google Scholar
P. N. Belhumeur and
D. J. Kriegman,
“The Yale face database,”
(1997) http://cvc.yale.edu/projects/yalefaces/yalefaces.html Google Scholar
K. C. Lee,
J. Ho, and
D. J. Kriegman,
“Acquiring linear subspaces for face recognition under variable lighting,”
IEEE Trans. Pattern Anal. Mach. Intell., 27
(5), 684
–698
(2005). https://doi.org/10.1109/TPAMI.2005.92 0162-8828 Google Scholar
T. Sim,
S. Baker, and
M. Bsat,
“The CMU pose, illumination, and expression (PIE) database,”
46
–51
(2002). Google Scholar
|