1 Dual Attributes for Face Verification Robust to Facial Cosmetics Lingyun Wen and Guodong Guo*, Senior Member, IEEE Abstract—Recent studies have shown that facial cosmetics have an influence on face recognition. Then a question is asked: Can we develop a face recognition system that is invariant to facial cosmetics? To address this problem, we propose a method called dual-attributes for face verification, which is robust to facial appearance changes caused by cosmetics or makeup. Attribute-based methods have shown successful applications in a couple of computer vision problems, e.g., object recognition and face verification. However, no previous approach has specifically addressed the problem of facial cosmetics using attributes. Our key idea is that the dual-attributes can be learned from faces with and without cosmetics, separately. Then the shared attributes can be used to measure facial similarity irrespective of cosmetic changes. In essence, dual-attributes are capable of matching faces with and without cosmetics in a semantic level, rather than a direct matching with low-level features. To validate the idea, we ensemble a database containing about 500 individuals with and without cosmetics. Experimental results show that our dualattributes based approach is quite robust for face verification. Moreover, the dual attributes are very useful to discover the makeup effect on facial identities in a semantic level. Index Terms—Dual attributes, face authentication, face verification, facial cosmetics, robust system, semantic-level matching. I. I NTRODUCTION It is quite common for women to wear cosmetics to hide facial flaws and appear more attractive. Archaeological evidence of cosmetics dates at least back to the ancient Egypt and Greece [1], [5]. The improved attractiveness using cosmetics has been studied in [13], [8]. Facial cosmetics or makeup can change the perceived appearance of faces [23], [27], [7]. In human perception and psychology studies [23], [27], it is revealed that light makeup slightly helps recognition, while heavy makeup significantly decreases human ability to recognize faces. In a computational approach, the impact of facial makeup on face recognition has been presented very recently [7]. There are eight aspects of facial makeup to impact the perceived appearance of a female face [7]: facial shape, nose shape and size, mouth size and contrast, form, color and location of eyebrows, shape, size and contrast of the eyes, dark circles underneath the eyes, and skin quality and color. It was also shown that facial makeup can significantly change the facial appearance, both locally and globally [7]. Existing face matching methods based on contrast and texture information can be impacted by the application of facial makeup [7]. Is it possible to develop a face recognition system that is invariant or insensitive to facial cosmetics? To answer this Manuscript received January 25, 2013. This work was supported in part by a Center for Identification Technology Research (CITeR) grant. L. Wen and G. Guo are with the Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV, 26506, USA. E-mail: lwen@mix.wvu.edu, guodong.guo@mail.wvu.edu. (Corresponding author, G. Guo, phone: 304-293-9143). question, we study how to develop a method that is robust to facial changes caused by facial cosmetics or makeup. In previous face recognition research, there are a number of studies to address various variations in face recognition, such as pose [18], [24], illumination [12], expression [19], [4], resolution [24], etc., however, few of previous works are to deal with the influence of cosmetics on face recognition. In this paper, we propose a method called dual-attributes for face verification, which is robust to facial appearance changes caused by cosmetics or makeup. Our key idea is that the dual-attributes can be learned from faces with and without cosmetics, separately. Then the shared attributes can be used to measure facial similarity irrespective of cosmetic changes. In essence, dual-attributes are capable of matching faces with and without cosmetics in a semantic level, rather than a direct matching with low-level features. Attribute-based methods have shown successful applications in a couple of computer vision problems, e.g., object recognition and face verification [17], [11], [15]. However, no previous approach has specifically addressed the problem of facial cosmetics using attributes. Attributes are a semantic level description of visual traits [17], [11]. For example, a horse can be described as fourlegged, mammal, can run, can jump, etc. A nice property of using attributes for object recognition is that the basic attributes might be learned from other objects, and shared among different categories of objects [10]. Facial attributes [15] are a semantic level description of visual traits in faces, such as big eyes, or a pointed chin. Kumar et al. [15], [16] showed that a robust face verification can be achieved using facial attributes, even if the face images are collected from uncontrolled environments on the Internet. A set of binary attribute classifiers are learned to recognize the presence or absence of each attribute. Then the set of attributes can be combined together to measure the similarity between a pair of faces for verification [15]. Parikh et al. proposed relative attributes [22], and emphasized the strength of an attribute in an image with respect to other images. For instance, the approach in [22] can predict person C is smiling more than person A but less than person B. The method can define richer textual descriptions for a new image. Facial attributes can also be used for face image ranking and search [14], [25]. The query can be given by semantic facial attributes, e.g., a man with moustache, rather than just a query image. Here we propose a method called dual-attributes to deal with face verification under the presence of facial cosmetics or makeup. Previous facial attributes [14], [25], [15], [16], [22] have not addressed the problem of facial makeup in face recognition. Our main contributions include: (1) a method called dual- 2 Fig. 1. Illustration of our dual-attributes approach to face verification. Two sets of attributes are learned in face images with and without cosmetics, separately. In testing, the pair of query faces undergo two different groups of attribute classifiers, and the shared attributes are used to measure the similarity between the pair of faces in a semantic level, rather than a direct match with low-level features. attributes is proposed to reduce the influence of facial cosmetics in face verification; and (2) a database of about 500 female individuals is collected to facility the study of facial makeup on face verification. Our approach is illustrated in Figure 1. In the remaining, we present our dual-attributes method in Section II. The database description, experimental setup and results are given in Section III. Finally, we draw conclusions. II. D UAL ATTRIBUTES FOR FACE M ATCHING Given the recent studies that showed that facial cosmetics has an influence on face recognition [23], [27], [7], it will be helpful to develop a face recognition system that is robust to facial cosmetics. In this paper, we use dual-attributes for face verification that focuses on reducing the influence of facial makeup on face matching. Attributes are visual properties of an object or face which can be described semantically [17], [11], [15], [22]. For face verification, Kumar et al. [15] proposed a set of facial attributes for face verification, and found that the attributesbased method can work successfully for face verification even with significant changes of pose, illumination, expression, and other imaging conditions. Built on Kumar et al.’s work, we want to study the influence of cosmetics on attributes, and develop a new method called dual-attributes for robust face verification. In this section, we first introduce the attributes that are used to describe the visual appearance of female faces. Then we present the details on how to compute the dual-attributes. Finally, we describe face verification using dual-attributes to deal with facial cosmetics in face matching. A. Attributes We use 28 attributes to describe female faces with and without cosmetics. Since our focus is on reducing the facial makeup effect on face recognition, we do not consider some groups of people, such as children or men. So there is no attribute related to children or males. The list of the 28 attributes is given in Table I. Some examples of these attributes are shown in Figure 2. The attributes used in our approach are learned from face images with and without cosmetics, separately. As shown in Figure 2, we can visually check the facial appearance differences between faces with and without cosmetics. To better describe the female faces with and without cosmetics, we used some detailed attribute features, compared to the attributes used in [15], [14], [25]. There is no specific consideration of facial cosmetics for face recognition in [15], [14], [25]. For example, in [25] and [14], only the attribute of eye glasses is used to describe the characteristics of eyes, while only the attributes of eye width, eyes open, and glasses are used in [15]. In our approach, more detailed attributes are used, such as blue eyes, black eyes, double-fold eyelids, slanty eyes, and eye crows feet, which include more details related to female beauty and facial cosmetics. From Figure 2, one can get an intuitive observation of the appearance variations of the same attributes between nonmakeup and makeup faces. To quantify the differences, it might be necessary to learn the attribute classifiers separately, for the makeup or non-makeup face images. Therefore, we propose the dual-attributes classifiers, which will be presented in the next. TABLE I L IST OF THE 28 ATTRIBUTES USED IN OUR APPROACH . Asian Attractive Big eyes Big nose Black eyes Blue eyes Double-fold eyelids Eye bage Eye crows feet Eyebrow long Eyebrow thickness Eyebrow wave Full bang Half bang High nose Lip line Nose middle heave No bang Open nose Philtrum Pointed chin Red cheek Small eye distance Slanty eyes Square jaw Thick mouth Wide cheek White B. Dual-Attributes Classifiers To address the visual appearance differences between faces with and without makeup in attribute learning, two classifiers are learned for each attribute, one is on faces with makeup, and the other is on faces with non-makeup. They are called dualattributes classifiers. For the i-th attribute, the two classifiers 3 Fig. 2. Example face images for some attributes. Faces in the same row are the positive and negative examples for a given attribute label. From left to right: positive non-makeup face, positive makeup face, negative non-makeup face, and negative makeup face. Fig. 3. Local patches in faces for dual-attributes learning. are denoted as N M Ci and M Ci , for non-makeup and makeup faces, respectively. In Section II-A, we introduced 28 attributes. Three attributes among the 28 are related to color: blue eye, black eye, and red cheek. The R, G, and B channels of an image (using local patches) are combined together to form the color feature. For other attributes, visual features that can characterize the local shapes or texture will be used, e.g., the histogram of oriented gradient (HOG) [6] features can be used to represent the local patches as shown in Figure 3. We can compare between the use of dual attributes and direct feature matching using the same low-level features, to show the benefit of using the dual attributes. Note that our emphasis is the concept of dual attributes rather than low-level features. In Figure 3, 12 local patches are illustrated, which are used to learn the attributes in faces. For example, attributes related to eyes use features mainly extracted from the eye patches. Each attribute classifier is trained using combined color and other features. We emphasize the separation between makeup and nonmakeup faces in learning the attribute classifiers. So, the attribute vector for the group of makeup face is M C(I) = hM C1 , M C2 , · · · , M Cn i, while the attribute vector for the group of non-makeup faces is N M C(I) = hN M C1 , N M C2 , · · · , N M Cn i, where n is the number of attributes, I is the index of the I-th individual. To develop an automated method for dual-attributes learning, we use an automated detector for face and facial feature detection [29], [9]. Nine facial points are detected automatically. Then all face images are normalized according to the positions of two eyes and the whole face is divided into patches according to these facial point locations. Different features can be extracted automatically from the local patches, as shown in Figure 3. In our dual-attributes classifier learning, the support vector machines (SVMs) with the RBF kernel [28] are used. C. Face Verification with Dual-Attributes The above learned dual-attributes can be combined together for face matching. Note that this matching is on the semantic level represented by the attributes, rather than using lowlevel facial features for direct matching. In our dual-attributes approach, the attributes for a makeup face will be detected using the attribute classifiers learned from makeup faces, while the attributes for a non-makeup face will be obtained by using the classifiers learned from non-makeup faces. For face verification, all available attributes extracted from a pair of faces form a vector of values, denoted by F (I) = hM C(I), N M C(I)i for individual I. A SVM face verification classifier will be trained to distinguish whether two face images are the same woman or not. In [15], the attribute vector is binary with values either 0 and 1, where 0 means the face does not have the attribute, while 1 means the face has the attribute. In our experiments, we use distances from the test sample to the SVM hyperplane instead of the binary values, and found the face verification accuracy can be improved by about two percent over the binary values. III. E XPERIMENTS We evaluate our dual-attributes approach to face verification experimentally. The face images are detected and normalized, and the attributes are learned from the makeup and nonmakeup faces separately. To verify the influence of facial cosmetics on face verification, we also perform a cross-makeup attributes classification. This further proves the influence of facial makeup on face analysis, and the necessity of using dualattributes to deal with facial cosmetics for face verification. We introduce the database that we assembled. Then we compare the performance of attribute classifications within the same group of faces or across groups (i.e., from makeup to non-makeup, or vice versa). Finally, face verification is performed using our dual-attributes. 4 A. Database There is no public available face database containing both makeup and non-makeup faces, to the best of our knowledge. To facilitate this study, we assembled a face database containing 501 pairs of female individuals. Each pair has two face images, one with makeup and the other without. All face images are collected from the Internet with text information about makeup or non-makeup. So the labels of makeup and non-makeup, and the facial identities are collected together with the face images from the Internet, rather than labeled afterward. To the best of our knowledge, this type of database is collected for the first time of its kind. Some example face images from our database are shown in Figure 4. From these examples, we can see that there are also some other variations in addition to makeup, e.g., expression, pose, occlusion, lighting change, etc. In this study, we focus on the cosmetics for face verification. Fig. 4. Some examples of face pairs without (left) and with (right) cosmetics in the database. We also asked five human subjects to manually label the facial attributes. The participants discussed with each other for each attribute label to find if there is an inconsistency. If an attribute label cannot be determined by the participants, or cannot come up with an agreement, it is just discarded. Finally, we got 28 attributes agreed among the participants, which are used as the ground truth attribute labels for our experiments. B. Within-Group and Cross-Group Attribute Classifications We evaluate the dual-attributes classification results and then combine the attributes for face verification. In examining the dual-attributes, we also perform a cross-group attributes classification with two purposes: (1) investigating the influence of facial cosmetics on attribute classification; and (2) justifying the necessity of using dual-attributes. Here we have two groups, one contains makeup faces, and the other contains non-makeup faces. In other words, each individual has two face images in our database, one is in the makeup group, and the other is in the non-makeup group. When both learning and testing are in the same group for attribute classifications, it is called within-group classification, while the training and testing are in two different groups, e.g., from makeup to nonmakeup, or vice versa, it is called cross-group or cross-makeup attribute classification. We randomly divide the individuals into training and testing sets for attribute classification. Each pair of faces for an individual are either in the training or testing set. There are about 4/5 individuals or face pairs are in the training set, and the remaining 1/5 are used for testing the attribute classification. The testing set is also used for the face verification experiments. For each attribute, two SVM classifiers are trained, using two groups of faces, makeup and non-makeup, separately. The attribute classification results are shown in Table II. The first column is the attribute classification accuracies using nonmakeup faces for both training and testing. Please note there is no overlap of individuals between the training and testing sets. The second column shows the classification results using the same non-makeup faces for testing, but with different attribute classifiers trained by using makeup faces. The third column ‘decrease1’ shows the decrease of classification accuracy when the second column is compared to the first column in each row. If the decrease is negative, it means the accuracy is increased. The remaining three columns have the similar meanings. To better understand the attribute classification results in various cases, we perform an accuracy change analysis (ACA) for the attribute classification results. Specifically, we categorize the accuracy changes into four cases: “Small change,” “Mpreferred,” “NM-preferred,” and “Big change,” where “M” is for makeup, and “NM” is for non-makeup. When the makeup group is used for training, while the cross-group accuracy is higher than the within-group, it is categorized as “Mpreferred.” Similarly, we may have “NM-preferred.” From Table II, one can observe the four cases. Firstly, the attribute classification results belonging to the case of “Small change” include: ‘Big nose’, ‘Nose middle heave’, ‘Philtrum’, ‘Wide Cheek’, and ‘White’. Our interpretation is that the facial cosmetics does not change those attributes (mainly on shape) too much. In this case, the attribute classification accuracy changes are small when the cross-group accuracies are compared to the within-group. To tolerate the small estimation errors caused by other variations (pose, expression, illumination, etc.) and consider the fact that the training database is not big, we use threshold values between -2% and 2% to determine this case. In other words, if the accuracy changes are within the range of -2% to 2%, it is termed as a small change. The second case is called “M-preferred,” which means the cross-group accuracy is increased when the makeup group is used for training attribute classifiers. In this case, the attribute class includes: ‘High nose’, ‘Lip line’, and ‘Asian’. Our interpretation is that the facial makeup will enhance the classification accuracy for those attribute classes. 5 TABLE II ATTRIBUTE CLASSIFICATION RESULTS FOR BOTH WITHIN - GROUP AND CROSS - GROUP CASES . W E USE “A → B” TO INDICATE THAT ‘A’ IS FOR TRAINING , WHILE ‘B’ IS FOR TESTING . T HE TWO GROUPS ARE ‘ MAKEUP ’ AND ‘ NON - MAKEUP ’ FACES . T HE COLUMN ’ DECREASE 1’ SHOWS ACCURACY DECREASE WHEN THE TRAINING GROUP IS CHANGED TO MAKEUP. T HE COLUMN ‘ DECREASE 2’ SHOWS ACCURACY DECREASE WHEN THE TRAINING GROUP IS CHANGED TO NON - MAKEUP FACES . Big nose Nose middle heave Philtrum Wide cheek White Half bang High nose Lip line Asian Double-fold eyelids Small eye distance Open nose Pointed chin Thick lips Eyebrow long Eyebrow thickness Eyebrow wave Blue eyes Black eyes Eye bag Slanty eyes Eye crows feet Big eye Square jaw Red cheek Attractive Full bang No bang non-makeup → non-makeup 74 % 73 % 61 % 85 % 88 % 85 % 70 % 63 % 84 % 66 % 74 % 72 % 90 % 63 % 68 % 64 % 68 % 75 % 85 % 63 % 67 % 71 % 64 % 64 % 67 % 68 % 92 % 95 % makeup → non-makeup 74 % 73 % 60 % 87 % 88 % 84 % 73 % 66 % 87 % 62 % 69 % 68 % 83 % 61 % 58 % 32 % 68 % 74 % 81 % 65 % 54 % 65 % 59 % 62 % 61 % 64 % 92 % 95 % Similarly, the third case is called “NM-preferred,” which means the cross-group accuracy is increased when the nonmakeup group is used for training attribute classifiers. In this case, the attribute class includes only the ‘Double-fold eyelids.’ Our interpretation is that this attribute may be easier to be learned from non-makeup face images. The last case is the “Big change”, which means that crossgroup classifications result in significant accuracy decreases. There are 18 attributes belonging to this case, as can be seen in Table II. Our decision of the four cases are based on the rules as below: Small change, d1 and d2 ∈ [−2%, 2%], M -pref erred, d1 < −2% and d2 ≥ 0, ACA = N M -pref erred, d1 ≥ 0 and d2 < −2%, Big change, otherwise where ACA is for ‘Accuracy change analysis’, d1 is for ‘decrease1,’ d2 is for ‘decrease2.’ M-preferred means ‘Makeup preferred’, and NM-preferred means ‘Non-makeup preferred.’ The summary of the four cases is shown in Table III for a clearer view. From Table III, we can understand that most of these attribute changes or no changes are close to our common knowledge. Women like to wear makeup near eyes to appear decrease1 0% 0% 1% -2 % 0% 1% -3 % -3 % -3 % 4% 5% 4% 7% 2% 10 % 32 % 0% 1% 4% -2 % 13 % 6% 5% 2% 6% 4% 0% 0% makeup → makeup 64 % 65 % 71 % 84 % 84 % 81 % 78 % 69 % 88 % 62 % 67 % 74 % 85 % 71 % 64 % 84 % 62 % 73 % 78 % 71 % 59 % 69 % 67 % 71 % 71 % 78 % 96 % 85 % non-makeup → makeup 64 % 65 % 70 % 82 % 85 % 80 % 73 % 65 % 84 % 68 % 67 % 67 % 75 % 68 % 58 % 48 % 53 % 70 % 76 % 66 % 58 % 67 % 60 % 60 % 67 % 78 % 89 % 80 % decrease2 0% 0% 1% 2% -1 % 1% 4% 4% 4% -4 % 0% 7% 10 % 3% 6% 36 % 9% 3% 2% 5% 1% 2% 7% 11 % 4% 0% 7% 5% TABLE III ATTRIBUTE CLASSIFICATION ACCURACY CHANGE ANALYSIS (ACA). T HE CHANGE ANALYSIS RESULTS IN FOUR CASES : S MALL CHANGE , M AKEUP PREFERRED , N ON - MAKEUP PREFERRED , AND B IG CHANGE . Attribute Classification Small change Makeup preferred Non-Makeup preferred Big change Attributes Big nose, Nose middle heave, Philtrum, Wide Cheek, White, Half bang, High nose, Lip line, Asian, Double-fold eyelids, Small eye distance, Open nose, Pointed chin, Thick lips, Eyebrow thickness, Eyebrow long, Eyebrow wave, Blue eyes, Black eyes, Slanty eyes, Eye crows feet, Big eyes, Square jaw, Red cheek, Attractive, Full bang, Forehead visible, more attractive. Usually there is less makeup designed to change the appearance of nose. That is why 2 out of 4 attributes related to nose have small changes, while most eye-related attributes (8 out of 9) have big changes. That is the makeup effect. One interesting attribute is ‘double-fold 6 eyelids.’ Usually we can observe whether a female is doublefold eyelids or not when there is no makeup around the eyes. But with makeup, it may bring some difficulty to determine if the woman has double-fold eyelids or not, especially using faked eyelashes for makeup. As a result, we categorized the ‘double-fold eyelids’ as non-makeup preferred attribute. Our dual-attributes-based analysis makes it easier to understand and interpret the computational approaches in terms of the effect of facial cosmetics on facial image analysis. This is a nice property using the attributes, and is different from previous approaches, where either purely human perception is used [23], [27] or just low-level features are computed [7]. On the other hand, we can see that the accuracies of many attribute classifiers are decreased, some may be decreased by more than 30%. This indicates the necessity of learning dual-attributes in facial image analysis for matching between makeup and non-makeup faces. Next, we will show the use of dual-attributes for face verification. C. Face Verification Results For face verification, we use the same test set as that used in our attribute classification experiments. We have about 100 pairs of positive faces for testing. There is no overlap between training and test sets. To have some negative pairs, we randomly pick up one face from the testing set other than the individual under consideration. The random selection keeps the makeup and non-makeup group information, i.e., if the face is non-makeup under consideration, the randomly selected negative face will be from the makeup group. So our face matching is always between a pair of makeup and non-makeup faces, no matter whether they are positive or negative pairs. The random pairing process was repeated for each individual. As a result, we have 100 negative pairs of face images. In total, we have 200 pairs of faces for testing. We combine all dual-attributes, as introduced in Section II-C, for face verification experiments. Some results are shown in Table IV. When the HOG feature is used to learn the dualattributes, the face verification accuracy is 71.0%, which is higher than the 66.5% accuracy when the HOG feature is used to match face images directly. Since only one face image exists for each individual in the gallery, it is not appropriate to learn a classifier for each individual. We used the cosine distance measure [21] for direct matching of two faces, which is much better than the Euclidean distance. The LBP (local binary pattern) feature is often used for face recognition [2]. We also compared the dual-attributes and direct matching based on the LBP feature. As shown in Table IV, the accuracy is 66.0% when the LBP feature is used for direct matching. When the LBP feature is used for dual attributes learning, the accuracy is 67.5%, slightly higher than the direct matching. This demonstrates two things: (1) facial makeup has a significant influence on facial texture, and the classical LBP feature cannot work well for makeup faces, and (2) the LBP feature is not good to learn the dual-attributes for makeup face verification. The receiver operating characteristic (ROC) curves are shown in Figure 5 with the area under the curve (AUC) values. We can observe that the verification performance based TABLE IV FACE VERIFICATION ACCURACY OF THE DUAL - ATTRIBUTES METHOD . Dual-Attr. (HOG) 71.0% HOG 66.5% Dual-Attr. (LBP) 67.5% LBP 66.0% on dual-attributes using HOG features is consistently higher than the direct matching using the HOG feature (or the LBP feature). The performance of dual-attributes learned by the LBP features is not consistent. For some false positive rate (FPR), the true positive rate (TPR) is higher than the LBP features, but for some other FPR, the TPR is lower than the direct matching with the LBP feature. As a result, the HOG feature is appropriate to learn the dual-attributes in our current experiments, while the LBP cannot work very well. We also compare the accuracies of direct matching using some other low-level features. When the PCA feature [26] is used, the accuracy is 65.0%; when the Scale-Invariant Feature Transform (SIFT) [20] feature is used, the accuracy is 65.5%; when the Gabor filters [30] are used for feature extraction, the accuracy is 64.5%. The linear discriminate analysis (LDA) [3] is not proper to use in our case, since there is only one example exists for each individual in the gallery. All these accuracies are lower than the 71.0% based on our dualattributes learned by the HOG feature. These comparisons demonstrate that learning the dual-attributes is useful to deal with facial cosmetics in face recognition. The PCA feature is not proper to learn the dual-attributes. We also used the SIFT feature to learn the dual-attributes. Experimentally, we found that the SIFT based dual-attributes cannot work well for low FPR in the ROC curve, although it can perform well when the FPR is higher (the curve is not shown here). So we do not recommend to use the SIFT feature for dual-attribute learning. Although attributes have been used for face verification in previous works, e.g., [15], there are several differences between our work and that in [15]: (1) We study facial makeup on face verification while there is no specific study on facial makeup in [15]; (2) The main focus in [15] is about face verification with pose, illumination, and expression (PIE) changes, which also appears in our database, but we emphasize the makeup influence and study face verification across facial makeup specifically; (3) In our database, there are faces with large head pose variations, while the face images in [15] were filtered by a frontal-view face detector and thus the pose variations are not very large; (4) The population contains women adults only in our database, while there are both men and women, young and adult in [15], where many attributes such as the age group, gender, beards, etc., can be used in [15], but those attributes cannot be used to help face authentication in our study. More importantly, we propose the new concept called dual-attributes, which has not been presented in any previous attribute-base approaches, to the best of our knowledge. To have a more intuitive understanding of the face verification results with facial cosmetics, we show some real examples of the verification results in Figure 6. There are four different results: true positive, false negative, true negative, and false 7 positive. One can see that for false negative results, it is quite challenging to match those faces. For false positive results, the face pairs do have some overall similarity and similar attributes. It is quite challenging to separate them into different individuals. 1 0.9 0.8 0.7 TPR 0.6 0.5 0.4 0.3 Dual−attr. HOG (AUC: 0.75) Dual−attr. LBP (AUC: 0.71) HOG match (AUC: 0.66) LBP match (AUC: 0.69) 0.2 0.1 0 0 0.2 0.4 0.6 0.8 1 FPR Fig. 5. ROC curves of the dual-attributes methods for face verification, compared to using the same features but without attributes learning. The LBP and HOG features are used here. The area under the curve (AUC) values are also computed for each curve. IV. C ONCLUDING R EMARKS We have proposed a method called dual-attributes for robust face verification with respect to facial cosmetics. A database of about 500 pairs of face images (with and without facial makeup) has been collected to facilitate our study. The empirical study has shown that the dual-attributes method is quite robust for face verification, matching between nonmakeup and makeup faces. In addition to face verification, the dual-attributes can be used to understand and interpret the influences of facial cosmetics on facial identities in a semantic level, which is difficult for previous approaches that are purely based on a direct matching of low-level features. Our preliminary result is promising. In the future, we will enhance the authentication system by learning the dual-attributes on a larger database, and explore other learning methods for dualattributes computation. R EFERENCES [1] L. Adkins and R. Adkins. Handbook to life in ancient Greece. Oxford University Press, 1998. [2] T. Ahonen, A. Hadid, and M. Pietikainen. Face recognition with local binary patterns. In Eur. Conf. on Comput. Vision, pages 469–481, 2004. [3] P. Belhumeur, J. Hespanha, and D. Kriegman. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE Trans. on Pattern Analysis and Machine Intelligence, 19(7):711–720, 1997. [4] A. Bronstein, M. Bronstein, and R. Kimmel. Expression-invariant 3d face recognition. In Audio-and Video-Based Biometric Person Authentication, pages 62–70. Springer, 2003. [5] B. Burlando, L. Verotta, L. Cornara, and E. Bottini-Massa. Herbal Principles in Cosmetics: Properties and Mechanisms of Action, volume 8. CRC PressI Llc, 2010. [6] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In IEEE Conf. on CVPR, pages 886–893, 2005. Fig. 6. Some examples of our face verification results based on the dualattributes method. (a) true positive, (b) false negative, (c) true negative, and (d) false positive. Here “positive” means (the ground truth of) the pair is genuine, while “negative” means the pair is impostor. If the method returns a correct result, it is called “true,” otherwise, it is called “false.” [7] A. Dantcheva, C. Chen, and A. Ross. Can facial cosmetics affect the matching accuracy of face recognition systems? In IEEE Conf. on Biometrics: Theory, Applications and Systems. IEEE, 2012. [8] A. Dantcheva and J. Dugelay. Female facial aesthetics based on soft biometrics and photo-quality. In IEEE Conf. on Institute for Computational and Mathematical Engineering. IEEE, 2011. [9] M. Everingham, J. Sivic, and A. Zisserman. “Hello! My name is... Buffy” – automatic naming of characters in TV video. In Proceedings of the British Machine Vision Conference, 2006. [10] A. Farhadi, I. Endres, and D. Hoiem. Attribute-centric recognition for cross-category generalization. In IEEE Conf. on CVPR, pages 2352– 2359, 2010. [11] A. Farhadi, I. Endres, D. Hoiem, and D. Forsyth. Describing objects by their attributes. In Computer Vision and Pattern Recognition, IEEE Conference on, pages 1778–1785, 2009. [12] A. Georghiades, P. Belhumeur, and D. Kriegman. From few to many: Illumination cone models for face recognition under variable lighting and pose. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 23(6):643–660, 2001. [13] N. Guéguen. The effects of womens cosmetics on mens courtship behavior. North American Journal of Psychology, 10(1):221–228, 2008. [14] N. Kumar, P. Belhumeur, and S. Nayar. Facetracer: A search engine for large collections of images with faces. Computer Vision–ECCV 2008, pages 340–353, 2008. [15] N. Kumar, A. Berg, P. Belhumeur, and S. Nayar. Attribute and simile classifiers for face verification. In Computer Vision, 2009 IEEE 12th International Conference on, pages 365–372. IEEE, 2009. [16] N. Kumar, A. C. Berg, P. N. Belhumeur, and S. K. Nayar. Describable visual attributes for face verification and image search. In IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), October 2011. 8 [17] C. Lampert, H. Nickisch, and S. Harmeling. Learning to detect unseen object classes by between-class attribute transfer. In Computer Vision and Pattern Recognition, IEEE Conference on, pages 951–958, 2009. [18] A. Li, S. Shan, X. Chen, and W. Gao. Maximizing intra-individual correlations for face recognition across pose differences. In Computer Vision and Pattern Recognition, 2009. IEEE Conference on, pages 605– 611. IEEE, 2009. [19] Y. Liu, K. Schmidt, J. Cohn, and S. Mitra. Facial asymmetry quantification for expression invariant human identification. Computer Vision and Image Understanding, 91(1):138–159, 2003. [20] D. Lowe. Distinctive image features from scale-invariant keypoints. International journal of computer vision, 60(2):91–110, 2004. [21] T. Pang-Ning, M. Steinbach, and V. Kumar. Introduction to data mining. WP Co, 2006. [22] D. Parikh and K. Grauman. Relative attributes. In Computer Vision (ICCV), 2011 IEEE International Conference on, pages 503–510. IEEE, 2011. [23] G. Rhodes, A. Sumich, and G. Byatt. Are average facial configurations attractive only because of their symmetry? Psychological Science, 10(1):52–58, 1999. [24] A. Sharma and D. Jacobs. Bypassing synthesis: Pls for face recognition with pose, low-resolution and sketch. In CVPR, pages 593–600. IEEE, 2011. [25] B. Siddiquie, R. Feris, and L. Davis. Image ranking and retrieval based on multi-attribute queries. In Computer Vision and Pattern Recognition, 2011 IEEE Conference on, pages 801–808. IEEE, 2011. [26] M. Turk and A. Pentland. Eigenfaces for recognition. J. of Cognitive Neuroscience, 3(1):71–86, 1991. [27] S. Ueda and T. Koyama. Influence of make-up on facial recognition. Perception, 39(2):260, 2010. [28] V. N. Vapnik. Statistical Learning Theory. John Wiley, New York, 1998. [29] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple. In Proc. IEEE CVPR, 2001. [30] L. Wiskott, J. Fellous, N. Kruger, and C. von der Malsburg. Face recognition by elastic bunch graph matching. IEEE Trans. Pattern Anal. Mach. Intell., 19(7):775–779, 1997. Lingyun Wen received her B.S. degree in Computer Science from Shandong Normal University, Jinan, China, in 2006. In 2009, she received her M.S. degree in Computer Science from University of Science and Technology of China (USTC), Hefei, China. She worked as a research engineer in Beijing, China, from 2009 to 2010. She is currently a graduate student in the Lane Department of Computer Science and Electrical Engineering at West Virginia University. Her research area includes computer vision and machine learning. Guodong Guo (M’07-SM’07) received his B.E. degree in Automation from Tsinghua University, Beijing, China, in 1991, the Ph.D. degree in Pattern Recognition and Intelligent Control from Chinese Academy of Sciences, in 1998, and the Ph.D. degree in computer science from the University of Wisconsin-Madison, in 2006. He is currently an Assistant Professor in the Lane Department of Computer Science and Electrical Engineering, West Virginia University. In the past, he visited and worked in several places, including INRIA, Sophia Antipolis, France, Ritsumeikan University, Japan, Microsoft Research, China, and North Carolina Central University. He won the North Carolina State Award for Excellence in Innovation in 2008, and Outstanding New Researcher of the Year (2010-2011) at CEMR, WVU. His research areas include computer vision, machine learning, and multimedia. He has authored a book, Face, Expression, and Iris Recognition Using Learning-based Approaches (2008), published over 40 technical papers in face, iris, expression, and gender recognition, age estimation, multimedia information retrieval, and image analysis, and filed three patents on iris and texture image analysis.