J Sign Process Syst (2014) 74:405–416 DOI 10.1007/s11265-012-0728-9 Uniform Local Derivative Patterns and Their Application in Face Recognition Huorong Ren & Jianwei Sun & Yanhong Hao & Xinxin Yan & Yang Liu Received: 1 March 2012 / Revised: 11 September 2012 / Accepted: 18 December 2012 / Published online: 3 January 2013 # Springer Science+Business Media New York 2012 Abstract In recent years, local feature descriptors have received more and more attention due to their effectiveness in the field of face recognition. Local Derivative Patterns (LDPs) for local feature descriptions attract researchers’ great interest. However, an LDP produces 2p different patterns for p neighbors through the transition of LDPs for an image, which lead to high dimension features for image analysis. In this paper, LDPs are expanded to Uniform Local Derivative Patterns (ULDPs) that have the same binary encoding way as LDPs but different transition patterns by introducing uniform patterns. A uniform pattern is the one that contains at most two bitwise transitions from 0 to 1 or vice versa when the binary bit is circular. Then, the number of the transition patterns is reduced from 2p to p(p −1)+3 for p neighbors, e.g., 256 to 59 for p=8. For face recognition, the histogram features are combined together in four directions, and both non-preprocessed and preprocessed images are used to evaluate the performance of the proposed ULDPs method. Extensive experimental results on three publicly available face databases show that the proposed ULDPs approach has better recognition performance than that obtained by using the LDPs method. Keywords Face recognition . Local derivative patterns (LDPs) . Uniform pattern . Uniform local derivative patterns (ULDPs). 1 Introduction In the past decades, a great deal of attention from the academic and industrial communities has been focused on H. Ren (*) : J. Sun : Y. Hao : X. Yan : Y. Liu School of Electro-Mechanical Engineering, Xidian University, No.2, South TaiBai Road, Xi’an 710071, China e-mail: hrren001@163.com the face recognition due to its wide applications to information security, network security, and personal identification. There are many well-known methods for face recognition, including Fisherface [1], Eigenface [2, 3], Linear Discriminant Analysis (LDA) [4], and kernel methods [5], which have been extensively studied and proved successful. However, these methods require a great large number of trained images, and actually the images available for training are few or even just one. In addition, they only extract holistic features of the face images since their local features are neglected, while the local features are essential information for face recognition. Recently, non-statistical local feature descriptors called Local Binary Patterns (LBPs) have become more and more popular in the field of face recognition. Initially, LBPs are designed by Ojala et al. [6] for texture feature classification due to its invariance with monotonic gray level changes. Ahonen et al. [7, 8] first apply LBPs to the area of face recognition. After that, LBP-based face recognition methods gain much more attention due to their robustness in illumination and expression. Naturally, the methods based on LBPs are developed with several variations for improved performance in face recognition. Since the threshold of LBPs is defined exactly as the value of the central pixel, LBPs are sensitive to noise in uniform regions. To address this problem, Tan and Triggs [9] extend LBPs to Local Ternary Patterns (LTPs) with a 3value code, in which the threshold is set to be slightly larger than zero. Ahonen et al. [10] propose Soft Histogram for Local Binary Patterns (SLBPs) that contribute one pixel typically to more than one bin meanwhile the sum of contributions of the pixel to all bins is always 1. An SLBP also suffers from the same problem that it is difficult to set a suitable threshold, as the case in LTPs. Furthermore, Heikkila et al. [11] propose Center-Symmetric Local Binary Patterns (CSLBPs) by comparing center-symmetric pairs of pixels around a central pixel instead of comparing each pixel 406 with the central pixel. Liu et al. [12] present a new face recognition method, three-level face features, by using CSLBPs three times. More ways to improve LBPs are to combine LBPs with other approaches [13, 14]. For example, the Multi-resolution Histograms of Local Variation Patterns (MHLVPs) proposed by Zhang et al. [15] represent face images as the concatenation of the local spatial histogram of local variation patterns computed from the multi-resolution Gabor features. Zhang et al. [16] also report a face recognition method, based on Local Gabor Binary Pattern Histogram Sequence (LGBPHS), which extracts Gabor and LBPs features of face images. The methods mentioned above suffer from a common problem: they capture non-directional first-order features only. To find more detailed information, Zhang et al. [17] expand LBPs to Local Derivative Patterns (LDPs) and high-order LDPs that encode directional pattern features at four local derivative variations. Thus, they contain more spatial information than traditional LBPs. Their work also designs the Gabor Local Derivative Patterns (GLDPs) that are proved successful in illumination variation face recognition by combining Gabor Wavelet with LDPs. All the methods mentioned above usually cause a high dimensionality that to a large extent limits their applications. This paper expands LDPs to Uniform Local Derivative Patterns (ULDPs) for face recognition under different appearance variations. In this framework, transition patterns of LDPs are divided into two categories: uniform and nonuniform ones. Therefore, the proposed method produces p (p−1)+3 patterns for p neighbors only, much less than 2p patterns in normal LDPs such that the size of the features dimensionality and the computational costs are greatly reduced. Different from the existing statistical methods, a ULDP directly extracts local features from the input images or feature images, as done in an LDP, which is why a ULDP has more robustness to unpredictable distribution images than the statistical methods. Similar to other non-statistical methods, a ULDP requires one image for per person for training only. Moreover, preprocessing methods are applied to combine with ULDPs for face recognition in this paper, which effectively enhances the performance of ULDPs in illumination variations. The remainder of the paper is organized as follows. Section 2 briefly reviews LDPs method and discusses ULDPs in detail. Section 3 proposes the application of the proposed ULDPs method in face recognition. Furthermore, preprocessing methods with γ correction and the Different of Gaussian (DoG) filter are described in this section. Section 4 reports experimental results to evaluate the performance of the methods on face recognition. Finally, Section 5 concludes this paper. J Sign Process Syst (2014) 74:405–416 2 Uniform Local Derivative Patterns In this section, we first briefly review the basic idea of Local Derivative Patterns, and then discuss in detail the proposed method, i.e., Uniform Local Derivative Patterns, in which the number of different patterns is reduced from 2p to p(p−1)+3 for p neighbors. 2.1 Local Derivative Patterns LDPs are proposed for face recognition through a high order local pattern descriptor. It is a general framework to encode directive pattern features from local derivative variation. The most notable characteristic of this method is that an LDP captures local features in four directions: 0°, 45°, 90°, and 135°, and concatenates the transition result as a 32-bit binary string. The first-order derivative for a pixel in 0°, 45°, 0 90°, or 135° direction, denoted as Ia ðZÞ, is calculated in a 3× 3 area around it, where α=0°, 45°, 90°, and 135°. Let Z0 be the central point of a local region I (Z), and Zi’s are the neighbors of Z0, where i=1, 2, ..., 8, as shown in Fig. 1. Hence the first-order derivatives at Z = Z0 are as follows [17]: 0 I0 ðZ0 Þ ¼ I ðZ0 Þ I ðZ4 Þ ð1Þ 0 I ðZ3 Þ ð2Þ 0 I ðZ2 Þ ð3Þ I45 ðZ0 Þ ¼ I ðZ0 Þ I90 ðZ0 Þ ¼ I ðZ0 Þ 0 I135 ðZ0 Þ ¼ I ðZ0 Þ I ðZ1 Þ ð4Þ Z9 Z10 Z11 Z12 Z13 Z24 Z1 Z2 Z3 Z14 Z23 Z8 Z0 Z4 Z15 Z22 Z7 Z6 Z5 Z16 Z21 Z20 Z19 Z18 Z17 Figure 1 Example of LDPs representation. J Sign Process Syst (2014) 74:405–416 407 The second-order directional LDPs derived from the firstorder derivative, in α direction at Z = Z0, can be defined as: n 0 0 0 o 0 0 0 LDPa2 ðZ0 Þ ¼ f Ia ðZ0 Þ; Ia ðZ1 Þ ; f Ia ðZ0 Þ; Ia ðZ2 Þ ; ; f Ia ðZ0 Þ; Ia ðZ8 Þ where f( , ) is a binary coding function, determining the types of local pattern transitions. It encodes the cooccurrence of two derivative directions at different neighboring pixels, as shown in [17]: 0 0 f Ia ðZi Þ; Ia Zj ¼ 2.2 The Theory of ULDPs An LDP directly encodes the binary result of the secondorder derivatives in four directions. For eight neighbors, 256 (28) different local derivative patterns are produced in each direction. It is most remarkable that some local derivative patterns contain the vast majority of the image features, sometimes over 90% in LDPs, and it is likely to LBPs [18]. These patterns, called uniform ones, have the same specialty in common that contains little spatial transitions. We design uniform local derivative patterns (ULDPs) by introducing a uniform pattern that reduces the number of the transition patterns of LDPs. ULDPs have the same binary encoding way as LDPs, while the final ULDPs transition patterns have a new encoding way that is different from LDPs. To formally define ULDPs, the transition patterns of LDPs are first divided into two categories: uniform patterns and non-uniform patterns, according to the number of spatial transitions (bitwise 0/1 change) of their patterns. An LDP pattern is said to be uniform if its binary pattern contains at most two bitwise transitions from 0 to 1 or vice versa when the bit is circular, while all the other binary patterns are attributed to non-uniform patterns. For example, 000000002 (0 transitions) and 011000002 (2 transitions) are uniform whereas 010100002 (4 transitions) and 100000102 (4 transitions) are non-uniform. For eight neighbors, there are nine types of patterns which are uniformed as shown in Fig. 3, where patterns 111111112 and 000000002 have zero time transitions, while the others have exactly 2 times transitions when the bit is circular. Each of the nine types of the patterns except for 111111112 and 000000002, has eight kinds of rotation forms when the bit is circular. Therefore, there are totally 7×8+2=58 different uniform patterns for 0 0 0; if Ia ðZi Þ Ia Zj > 0 0 0 1; if Ia ðZi Þ Ia Zj 0 a ¼ 0 ; 45 ; 90 ; and 135 ; i ¼ 1; 2; ; 8: ð6Þ The second-order directional LDPs can also be defined as: LDPa2 ðZ0 Þ ¼ 8 0 X 0 f Ia ðZ0 Þ; Ia ðZn Þ 2n ð7Þ 1 n¼1 Finally the second-order LDPs, denoted by LDP2(Z), is defined as the concatenation of the four second-order directional LDPs, complied with Eq. (8): n o 0 LDP2 ðZÞ ¼ LDPa ðZÞ a ¼ 0 ; 45 ; 90 ; and 135 ð8Þ A computational example of LDPs is illustrated in Fig. 2. To compute an LDP in α=0° direction at the central pixels, we calculate the first-order derivatives by Eq. (1). Then, we obtain the second-order directional LDPs in 0° direction by Eq. (5) or Eq. (7), i.e., LDP02 ¼ 11000000, an 8-bit string, which is denoted by transition pattern 3 (the left-most bit is the low bit). Similarly, the calculated second-order LDPs in 45°, 90°, and 135° are “00001100” (transition pattern 48), “00011110” (transition pattern 120), and “00001100” (transition pattern 48), respectively. Finally, all the second-LDPs binary results are concatenated to gain the LDP 2 (Z), “110000000000110000011111000001100”, a 32-bit string. Figure 2 Example to obtain the second-order LDPs in α=0° direction. ð5Þ (a) (b) (c) 81 75 72 72 89 65 64 58 54 55 6 4 -1 1 53 52 53 54 56 -1 -1 -2 0 51 52 59 67 75 -7 -8 -8 0 58 57 59 69 81 Original sub-region First-order derivative 1 0 0 0 0 Second-order LDPs 408 J Sign Process Syst (2014) 74:405–416 Figure 3 Nine uniform binary patterns that can occur in the circularly symmetric neighbor set, where black and white circles correspond to bit values of 0 and 1 in the 8-bit output of the operator, respectively. eight neighbors. Then ULDPs in α direction at Z = Z0 for 8 neighbors can be calculated by the following equations: 8" # 8 > p 1 P 0 0 > > f Ia ðZ0 Þ; Ia Zp 2 > > > < p¼1" # ULDP ¼ 8 p 1 P 0 0 > > N f Ia ðZ0 Þ; Ia Zp 2 ; if g ðLDPÞ 2 > > > p¼1 > : 58 ; otherwise ð9Þ 0 0 0 0 f Ia ðZ0 Þ; Ia ðZ1 Þ gðLDPÞ ¼ f Ia ðZ0 Þ; Ia ðZ8 Þ 7 P 0 0 0 0 f Ia ðZ0 Þ; Ia Zpþ1 f Ia ðZ0 Þ; Ia Zp þ p¼1 ð10Þ where N(x) is a function to count the total number of the non-uniform patterns, which is less than x, and g(LDP) is used to calculate the number of spatial transitions in which bitwise 0/1 changes of LDPs patterns occur. Equation (10) implies an LDP pattern (an 8-bit string) first moves one bit, and then subtracts the original binary value of LDPs pattern from the resulting pattern after removal. Finally, the number of spatial transitions is obtained by calculating the sum of absolute values of the difference. It is show in Eq. (9) that a pattern is uniform only if its number of spatial transitions is not more than 2, denoted by g (LDP) ≤2. By the previous descriptions, it is clear that there are exactly p(p-1)+2 uniform patterns numbered from 0 to p (p-1)+1 for p neighbors (i.e., from 0 to 57 for 8 neighbors) and we define all the non-uniform patterns as pattern p(p-1)+2 that is exactly 58 for p=8. An example of ULDPs computation for eight neighbors is illustrated in Fig. 2. Noted that the 8-bit binary result in Fig. 2, i.e., “11000000”, contains two times bitwise transitions. Consequently, it is a uniform pattern and its value of ULDPs pattern is 3 by Eq. (9). The three results, “00001100”, “000111110”, and “00001100”, are uniform patterns, corresponding to the values of ULDPs patterns: 17, 25, and 17, respectively, which are different from LDPs patterns calculated previously. A ULDP produces p(p-1)+3 different patterns that consist of p(p-1)+2 uniform patterns and one non-uniform pattern. Compared with 2p different patterns produced by LDPs, the quantity is reduced greatly. As well known, the final image features used for image analysis are the histogram features accumulated by feature descriptors of an input image. The reason that why ULDPs histogram features provide better discrimination compared against the LDPs histogram features comes down to the difference in their statistical properties. The relative proportion of the nonuniform patterns of all patterns accumulated into a histogram is so small that their probabilities cannot be estimated reliably. In addition, the uniform patterns reduce the influence of the high frequency noise that can be defined as non-uniform patterns. This means that it can extract representative features of the face images with the feature vectors of the uniform and nonuniform patterns. As shown in Fig. 4, the ULDPs produce a 59-dimensional feature vector less than the 256dimensional feature vector produced by LDPs. Furthermore, we can also see that there are a large number of zeros between patterns 100 and 125 in LDPs histogram, and no such sparse problem exists in ULDPs histogram. 3 ULDPs for Face Recognition This section discusses the application of ULDPs to face recognition. The overall framework of the proposed method used for histogram feature extraction is illustrated in Fig. 5. In this method, the ULDPs-based face recognition is modeled by the following procedures: (1) normalizing an input face image according to the location of the eyes and preprocessing it by using γ correction and the Different of Gaussian (DoG) filter; (2) labeling the image by the ULDPs descriptor in four directions; (3) dividing the transition results of each image into several rectangular sub-regions and calculating histogram features in each region; and (4) establishing the face model with the concatenation of all the histograms sequence and recognizing the face image by the histogram intersection and the nearest neighbor classifier. In what follows, we discuss these procedures in detail. 3.1 Image Normalization and Preprocessing For face recognition, images from a database are in a generally normalized size by cropping images according to the location of eyes to eliminate the impact of the background. Moreover, preprocessing methods are often used to decrease noise factors, such as γ correction and the Different of Gaussian (DoG) filter, which are believed to be the preferable methods for this problem. J Sign Process Syst (2014) 74:405–416 409 250 250 225 225 200 200 175 175 150 150 125 125 100 100 75 75 50 50 25 25 0 0 0 25 50 75 100 125 150 175 200 225 250 0 5 10 15 20 Histogram of LDPs 25 30 35 40 45 50 55 58 Histogram of ULDPs Figure 4 Differences of the histogram between LDPs (middle) and ULDPs (right). γ correction is a nonlinear gray-level transformation that replaces gray-level I with Iγ (for γ >0) or log(I) (for γ=0), where γ is a user-defined parameter ranging from 0 to 1, reflecting the performance of nonlinear transformation. γ correction increases the contrast of images by enhancing the local dynamic ranges in the shadowed regions of the image while compressing them in bright regions. It is clear that the only factor in γ correction is the value of γ that is widely used in the range [0, 0.5]. To correct the uneven illumination, an image is first changed into the frequency domain by two-dimensional Fourier transform. Then it is filtered by the Different of Gaussian (DoG) filter to remove redundant information, while preserving the useful low-frequency information. The DoG filter is a common filter in the field of computer vision and image analysis, which consists of two Gaussian functions with different standard deviations [19]: GðsÞ ¼ A1 e s2 2σ2 1 A2 e s2 2σ2 2 A1 A2 > 0; σ1 > σ2 ð11Þ It is obvious that the performance of the DoG filter is principally decided by two parameters: σ1 and σ2. By a series of experiments, we find that when γ=0.2, A1 =A2 =1, σ1 =2 and σ2 =1, the proposed ULDPs can achieve the best performance. Thus, these values are determined as the default setting. 3.2 Histogram of ULDPs The purposes to develop the ULDPs-based method in the previous sections are used for face recognition. The images collected under various conditions such as illumination, accessory, and expression are really challenges to face recognition. We model the distribution of ULDPs by spatial histograms [8, 17, 20] because of their robustness to the various conditions. A ULDP captures histogram features in four invariant directions: 0°, 45°, 90°, and 135°. To avoid the loss of too much Micro-local information after ULDPs encoding, an image is divided into several sub-regions. In α direction, ULDPsα are spatially divided into rectangular sub-regions represented by S1, S2, ..., and SN, and then the spatial histograms of ULDPs, denoted as HULDP [17], can be extracted as follow: HULDPði; aÞ ¼ fHULDPa ðSi Þji ¼ 1; ; N ; Original image Histogram Feature Preprocessed image ULDP at four directions Figure 5 The framework of ULDPs-based histograme feature extraction. ð12Þ ULDP Normalized image a ¼ 0 ; 45 ; 90 ; and 135 g 410 J Sign Process Syst (2014) 74:405–416 where HULDPa ðSi Þ is the ULDPs histogram features extracted from the local sub-region Si. For histogram matching, the histogram intersection is low in computational complexity and simple in operations among existing similarity measurements. Thus, it is used for the matching measurement of the two histograms. Histogram intersection is defined as follows: DHI ðH; S Þ ¼ min SHI B P i¼1 Hi ; B P ð13Þ Si i¼1 Normalized image B X minðHi ; Si Þ ð14Þ i¼1 where DHI ðH; S Þ is the histogram intersection statistic result with H = (H1,H2, ...,HB)T and S = (S1,S2, ...,SB)T, and B is the total number of the histogram features. The similarity of the two histograms calculated by Eqs. (13) and (14) is used for the classification with the nearest neighbor classifier. 4 Experimental Results and Analysis The preceding sections deal with theoretical aspects of the proposed framework. To empirically test the main premises of our work and the effectiveness of the described methodology, we conducted a series of experiments to illustrate the performance of the proposed ULDPs for face recognition that covers various conditions including illumination, expression, pose, and accessory. An extensive set of publicly available face databases, FERET [21], CMU PIE [22], Extended Yale B [23, 24], and CAS-PEAL-R1 [25], were used in the experiments. In what follows, the first subsection conducts experiments on a subset of the widely used the FERET database to test the robustness of ULDPs. Then, we conducted the comparative experiments to test the performance of ULDPs on the CAS-PEAL-R1 database to test ULDPs in expression and accessory. Furthermore, the CMU PIE database with a variety of poses and illuminations was utilized to evaluate the performance of the proposed method. The experiment result on the Extended Yale B database with severe illumination variations was also reported in this work. Before proceeding to the experiments, we take a quick look at the statistical data of the appearing uniform patterns in ULDPs. In ULDPs, there are 58 uniform patterns among the 256 possible patterns. For the images in the FERET database, the 58 uniform patterns contribute 64.99 %, 73.01 %, 69.72 %, and 73.22 % in the four directions, 0°, 45°, 90°, and 135°, respectively. In the CAS-PEAL-R1 database, they are 81.67 %, 86.53 %, 83.7 %, and Original image Figure 6 Example of a normalized FERET image. 84.65 %, respectively. In the CMU PIE database, they are 73.41 %,79.64 %,69.98 %, and 79.57 %, respectively. In the Extended Yale B database, the proportions of uniform patterns are relatively low: 62.45 %, 68.75 %, 60.87 %, and 68.86 % in the four directions, respectively. It indicates that the uniform patterns occupy the majority of the information in an image, which represents the essential lower frequency information. More to the point, pattern 000000002 representing the flat region is the most frequent pattern in the all images. 4.1 Experimental Comparisons on the FERET Database The comparative experiments between LDPs and ULDPs were first conducted on the FERET face database that is widely used to evaluate face recognition algorithms. The FERET database consists of a total of 14,051 gray-scale images representing 1,199 individuals. The images contain variations in lighting, facial expressions, pose angele, etc. In this work, only frontal faces are considered. These facial images can be divided into five sets as follows: Recognition rate (%) SHI ðH; S Þ ¼ 100 90 80 70 60 50 40 30 20 10 0 4×4 fb fc dup I dup II Average 5×5 6×6 7×7 8×8 9×9 Sub-region number Figure 7 Comparative face recognition rates of ULDPs on the FERET database with different sub-regions. J Sign Process Syst (2014) 74:405–416 (a) 411 95 Recognition rate (%) 85 fb fc dup I dup II Average 75 65 55 45 35 25 15 LBPs ULBPs LDPs ULDPs Results on the non-preprocessed images. (b) 100 Recognition rate (%) 90 80 70 60 fb fc dup I dup II Average 50 40 30 20 PLBPs PULBPs PLDPs PULDPs Results on the preprocessed images. Figure 8 Comparative identification accuracies on the FERET database. a Results on the non-preprocessed images. b Results on the preprocessed images. & & & & & fa set, used as a gallery set, contains frontal images of 1,196 people. fb set (1,195 images). The subjects were asked for an alternative facial expression than in the fa photograph. fc set (194 images). The photos were taken under different lighting conditions. dup I set (722 images). The photos were taken later in time between 1 min to 1031 days. dup II set (234 images). This is a subset of the dup I set containing those images that were taken at least after 18 months after the corresponding gallery image. Figure 9 Samples of the frontal image of one subject from the CAS-PEAL-R1 database. To ensure the reliabilty of the test, the publicly available CSU face identification evaluation system [26] was utilized to normalize the images. All the images were normalized and cropped to 150×130 pixels by the coordinates of the eyes. Figure 6 shows an example of a normalized FERET image. In this work, fa containing 1,196 frontal images of 1,196 subjects was used as a gellery set, while fb, fc, dup I, and dup II were used as probe sets. The first experiment was designed to explore the influence of the sub-region number on recognition accuracy of the normalized images. In this experiment, images were divided into a variable number of regions, from 4×4 to 9× 9, and the obtained histograms from each region were concatenated to obtain the descriptor. Figure 7 shows that recognition accuracy on all four subsets of the FERET database is improved when the sub-region number increases from 4×4 to 8×8, while fb and dup II subsets decrease when the number reaches to 9×9. Moreover, the average curve of recognition accuracy becomes flatter when the sub-region number increases from 8×8 to 9×9. Therefore, we chose 8× 8 as the default size in the experiment on the FERET database. To observe the performance of ULDPs under different conditions, the next experiments were conducted on the individual probe sets. Experimental results in Fig. 8(a) demonstrate that ULDPs work better than the other methods: LBPs, ULBPs, and LDPs in average. It is clear that ULDPs achieve much higher recognition accuracy compared against LDPs. The experiments were also designed to evaluate the effectiveness of ULDPs on the preprocessed images with the DoG filter and γ correction. The recognition rates obtained with different descriptors on the preprocessed images are shown in Fig. 8(b). It should be noted that PULDPs (ULDPs with the preprocessing method) perform better than PLBPs (LBPs with the preprocessing method), PULBPs (ULBPs with the preprocessing method) and PLDPs (LDPs with the preprocessing method) in average. Specially, on the fc subset, when gallery and probe images were taken under different illumination, PULDPs give the recognition accuracy as high as 94.85 %. The results on the fc subset show that 412 J Sign Process Syst (2014) 74:405–416 (a) 75 95 Recognition rate (%) Recognition rate (%) 90 85 80 75 Expression Accessory Average 65 45 non-preprocessed method ULBPs LDPs with preprocessed method 25 LBPs ULDPs Result on the non-preprocessed images. ULBPs LDPs ULDPs Figure 12 Comparative face recognition rates of ULDPs, LDPs, ULBPs, and LBPs on the CMU PIE databse. over LBPs in dup I and dup II subsets, in which LBPs give the best result with the original images. It means that PULDPs are more robust to aging than that of ULDPs. (b) 90 Recognition rate (%) 55 35 70 60 LBPs 65 86 4.2 CAS-PEAL-R1 Database - Expression and Accessory Variant 82 78 Expression Accessory Average 74 70 PLBPs PULBPs PLDPs PULDPs Result on the preprocessed images. Figure 10 Comparative the face recognition rates of ULDPs, LDPs, ULBPs, and LBPs on the CAS-PEAL-R1 database. a Result on the non-preprocessed images. b Result on the preprocessed images. the preprocessing method is powerful for ULDPs in illumination variation. It is worth mentioning that PULDPs achieve 10.66 % and 13.51 % performance improvements Pose 09 Pose 27 Pose 29 The CAS-PEAL-R1 face database contains 99,594 images among 1,040 individuals (595 males and 445 females) with different sources of variation, especially Pose, Expression, Accessory, and Lighting (PEAL). We choose the first 300 subjects (one image per person) from the gallery set of CASPEAL-R1 as our gallery set. Our probe set contains 1,939 frontal face pictures, including expression (878 pictures) and accessory (1061 pictures), which are chosen from the probe set of CAS-PEAL-R1. Some sample images are shown in Fig. 9. In the following experiments, all the employed images were normalized to 150×130 pixels by the CSU face identification evaluation system to eliminate the influence of the background. To observe the performance of ULDPs under different conditions in expression and accessory, the experiments were conducted on the CAS-PEAL-R1 database that includes variations in facial expressions and accessories. The experimental results as shown in Fig. 10(a) demonstrate that ULPDs obtain better recognition accuracy on expression subset than LDPs but work worse than LDPs on the accessory subset. Since the patterns distribution in histogram introduced by the accessories is different from the face natural variations, ULDPs are of plain robustness to accessories and have slightly better recognition accuracy to expression. Pose 05 Pose 07 Figure 11 Samples of the frontal image of one subject from the CMU PIE database. Table 1 Experiment using the CMU PIE database with different poses. Method Pose 5 Pose 7 Pose 9 Pose 27 Pose 29 ULDPs PULDPs 0.52 0.74 0.60 0.79 0.54 0.78 0.68 0.88 0.55 0.68 J Sign Process Syst (2014) 74:405–416 413 Figure 13 Samples of the frontal image of one subject from the Extended Yale B database. We also conducted the experiments to evaluate the effectiveness of the PULDPs on the CAS-PEAL-R1 database. It can be maintained from Fig. 10(b) that PULDPs perform better than PLBPs, PULBPs and PLDPs in both expression and accessory subsets. However, it should be noted from Fig. 10 that PULDPs have some lower accuracy than that of ULDPs. A reasonable explanation is that the preprocessing method smoothes out the minute variations of expression on the expression subset. 4.3 CMU PIE Database - Pose and Illumination Variant We further explored the impact of the preprocessing on the performance of ULDPs under various poses and illuminations using the CMU PIE database. The database contains totally 41,368 images of 68 individuals, including a variety of poses, illuminations, and expressions. The images with the variations in both pose and illumination are used in the experiment. These images include seven different poses, labeled by Poses 05, 07, 09, 11, 27, 29, and 37, respectively, and each pose has 21 different illumination conditions labeled by Flashes 2 to 22. Here we choose five poses for our experiments, Pose 05, 07, 09, 27, and 29, since the five sets have slight pose variations only. In this experiment, all the images were cropped to 64×64 pixels according to the location of eyes, aiming to obtain the whole facial image and eliminate the complex background. Figure 11 shows an example of the normalized images of a person. The frontal images (Pose 27) with the frontal illumination condition (Flash 08) of a person are chosen to build the gallery set. The remaining images are used as the probe set. In this experiment, we use 12×12 sized sub-region with 128 histogram bins for both LBPs and LDPs. Figure 12 depicts the recognition rates of LBPs, ULBPs, LDPs, and ULDPs on both non-preprocessed and preprocessed images by using the probe set. It can be observed from Fig. 12 that the ULDPs have a better recognition rate than the others on both non-preprocessed and preprocessed images. Compared with LBPs(39.79 %), ULBPs(40.07 %), and LDPs (55.45 %), ULDPs(59.37 %) obtain the highest recognition rate on non-preprocessed images. Moreover, the PULDPs (77.37 %) are also higher than PLDPs (74.1 %) and PLBPs (40.07 %). The results show that ULDPs can obtain more discriminative features than LBPs and LDPs. 100 R ecognition rate (% ) 95 90 Table 2 Computational time (CPU seconds) of LDPs and ULDPs on different databases. 85 non-preprocessed method 80 with preprocessed method 75 FERET CAS-PEAL-R1 CMU PIE Extended Yale B 10,301 3,034 10,424 3,073 1,783 626 1,751 629 1,470 562 1,422 577 1,520 569 1,537 593 70 65 60 LBPs ULBPs LDPs Figure 14 Experiment results on the Extended Yale B databse. ULDPs LDPs ULDPs PLDPs PULDPs 414 We aslo conducted an experiment to explore the robustness of the proposed ULDPs method to poses. Tab. 1 shows the results of this experiment. The experiment results in Table 1 show that the PULDPs have a better robustness to poses than ULDPs at all the five poses. Compared with other poses, ULDPs(68 %) and PULDPs(88 %) obtain the best recognition accuracy at Pose 27. As stated previously, the images at Pose 27 are the fontal ones. Then, a conclusion can be reached from the results that both ULDPs and PULDPs have poor robutness to pose variations. Although the performance of ULDPs decreases with the pose variations of the images in this database, it still maintains superiority in recognition accuracy over both LBPs and LDPs. 4.4 Extended Yale B Database - Illumination Variant The Extended Yale B database was used for the comparative studies among ULDPs, LDPs, and LBPs under different illumination variations. The database contains totally 2,432 frontal images of 38 individuals (each subject is imaged under 64 different illumination conditions). All the images have already been manually aligned and cropped to 168 × 192 pixels for public downloading from the database website. In our framework, these images were normalized to 84 × 96. Figure 13 shows the samples of the frontal images. The frontal face images with lighting direction of 0o azimuth (“A+000”) and 0o elevation (“E + 00”) are used to build the gallery set. All the remaining images belong to the probe set. Here, 12×13 pixel windows were selected since they are good trade-off between recognition performance and feature vector length. For the Extended Yale B database, the recognition rates on the both non-preprocessed and preprocessed probe sets are illustrated in Fig. 14. It indicates that ULDPs obtain a 3.13 % performance improvement over LDPs. Meanwhile the recognition accuracy of PULDPs reaches its maximum of 99.04 %, and obtains 31.5 % performance improvement than that of ULDPs. It means that PULDPs are more robust against illumination variation. We can also learn from Fig. 14 that both PULPDs and PLDPs get much better recognition performance than LBPs and ULBPs. In addition, the computational time of the experiments was also considered, as shown in Table 2. Compared with LDPs, the time for ULDPs decreases 70.54 %, 64.89 %, 61.77 % and 62.57 %, on FERET, CAS-PEAL-R1, CMU PIE and Extended Yale B databases, respectively. All the experiments were carried out on a computer running Windows XP with an AMD Athlon II CPU processor (X4 640 @ 3.0 GHz), a 2-GB RAM. For all methods, we used the original MATLAB implementation by using MATLAB 7.10.0 (R2010a). J Sign Process Syst (2014) 74:405–416 5 Conclusions Local Derivative Patterns (LDPs) are extended to Uniform Local Derivative Patterns (ULDPs) in this paper. Experimental results, on a set of publicly available face databases, demonstrate that the proposed ULDPs have better robustness with the respect to facial expression, illumination and aging than LDPs. Compared with LDPs, the computational time for ULDPs decreases over 60 % in all four databases used in this paper, although ULDPs get worse in some subset with accessories. We could say that the ULDPs method is an excellent choice if the computational time is a major concern with good recognition rates. In addition, PULDPs demonstrate the validity and efficiency of a combination of ULDPS and the preprocessing method for variable illumination in face recognition. The recognition accuracy of PULDPs reaches its maximum of 99.04 % on the Extended Yale B database. However, it should be noted that the performance of ULDPs is still relatively low when face images contain various accessory. Future work includes improving their robustness with accessory variations. Another important topic is looking for an effective feature selection method to reduce the length of the feature vector. Acknowledgments This work is partially supported by “the Fundamental Research Funds for the Central Universities, K50510040013”. References 1. Belhumeur, P. N., Hespanha, J. P., & Kriegman, D. J. (1997). Eigenfaces vs. Fisherfaces: recognition using class specific linear projection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(7), 711–720. 2. Sirovich, L., & Kirby, M. (1987). Low dimensional procedure for the characterization of human face. Journal of the Optical Society of America, 4(3), 519–524. 3. Turk, M., & Pentland, A. (1991). Eigenfaces for recognition. Journal of Cognitive Neuroscience, 3(1), 71–86. 4. Etemad, K., & Chellappa, R. (1997). Discriminant analysis for recognition of human face images. Journal of the Optical Society of America, 14(8), 1724–1733. 5. Yang, J., Frangi, A. F., Yang, J. Y., Zhang, D., & Jin, Z. (1997). KPCA plus LDA: a complete kernel fisher discriminant framework for feature extraction and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(2), 230–244. 6. Ojala, T., Pietikainen, M., & Harwood, D. (1996). A comparative study of texture measures with classification based on featured distribution. Pattern Recognition, 29(1), 51–59. 7. Ahonen, T., Hadid, A., and Pietikainen, M. (2004). Face recognition with local binary patterns, in Proc. 8th European Conference on Computer Vision (pp. 469–481). Prague, Czech Republic. 8. Ahonen, T., Hadid, A., & Pietikainen, M. (2006). Face description with local binary patterns: application to face recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(12), 2037–2041. J Sign Process Syst (2014) 74:405–416 9. Tan, X., & Triggs, B. (2010). Enhanced local texture feature sets for face recognition under difficult lighting conditions. IEEE Transactions on Image Processing, 19(6), 1635–1650. 10. Ahonen T., and Pietikainen M. (2007). Soft histograms for local binary patterns, in Proc. Fin. Signal Process. Symp. (pp. 1–4). Oulu, Finland. 11. Heikkila, M., Pietikainen, M., & Schmid, C. (2009). Description of interest regions with local binary patterns. Pattern Recognition, 42 (3), 425–436. 12. Liu, C., Jianyun, L., & Li, L. (2011). Three-level face features for face recognition based on center-symmetric local binary pattern. Proceedings IEEE International Conference on Computer Science and Automation Engineering, 4, 394–398. 13. Pang, Y., Yuan, Y., & Li, X. (2008). IEEE Transactions on Circuits and Systems for Video Technol, 18(7), 989–993. 14. Zhang, B., Shan, S., Chen, X., & Gao, W. (2007). Histogram of gabor phase patterns (HGPP): a novel objects representation approach for face recognition. IEEE Transactions on Image Processing, 16(1), 57–68. 15. Zhang, W., Shan, S., Zhang, H., Gao, W., and Chen, X. (2005). Multi-resolution histograms of local variation patterns (MHLVP) for robust face recognition, in Proc. Audio-Video-Based Biometric Person Authentication (pp. 937–944). New York, USA. 16. Zhang, W., Shan, S., Zhang, H., Gao, W., and Chen, X. (2005). Local gabor binary pattern histogram sequence (LGBPHS): A novel non-statistical model for face representation and recognition, in Proc. IEEE Int. Conf. Computer Vision vo.1, (pp. 786–791). Beijing, China. 17. Zhang, B., Gao, Y., Zhao, S., & Liu, J. (2010). Local derivative pattern versus local binary pattern: face recognition with high order local pattern descriptor. IEEE Transactions on Image Processing, 19(2), 533–544. 18. Ojala, T., Pietikainen, M., & Maenpaa, T. (2002). Multiresolution gray-scale and rotation variant texture classification with local binary patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(7), 971–987. 19. Castleman, K. R. (1996). Digital image processing [M] (pp. 212– 213). Engle wood Cliffs: Prentice Hall. 20. Zhang, H., Gao, W., Chen, X., & Zhao, D. (2005). Learning informative features for spatial histogram-based object detection. Proceedings IEEE International Joint Conference Neural Networks, 3, 1806–1811. 21. Phillips, P. J., Moon, H., Rizvi, S. A., & Rauss, P. J. (2000). The FERET evaluation methodology for face-recognition algorithms. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(10), 1090–1104. 22. Sim, T., Baker, S., & Bsat, M. (2003). The CMU pose, illumination, and expression database. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(12), 1615–1618. 23. Georghiades, A. S., Belhumeur, P. N., & Kriegman, D. J. (2001). From few to many: illumination cone models for face recognition under variable lighting and pose. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(6), 643–660. 24. Lee, K. C., Ho, J., & Kriegman, D. J. (2005). Acquiring linear subspaces for face recognition under variable lighting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27 (5), 684–698. 25. Gao, W., Cao, B., Shan, S., Chen, X., Zhou, D., Zhang, X., & Zhao, D. (2008). The CAS-PEAL large-scale Chinese face database and baseline evaluations. IEEE Transactions on Systems, Man, Cybernetics - Part A: Systems and Humans, 38(1), 149–161. 26. Beveridge, J. R., Bolme, D., Draper, B. A., & Teixeira, M. (2005). The CSU face identification evaluation system: its purpose, features, and structure. Machine Vision and Applications, 16(2), 128– 138. 415 Huorong Ren received the B.Sc., M.Sc., and Ph.D. degrees in electronic engineering from Xidian University, China, in 1994, 1999, and 2004, respectively. Currently, he is an Associate Professor with the School of Electro-Mechanical Engineering, Xidian University, China. His research interests include face recognition, pattern recognition, computer vision, and bio-security. Jianwei Sun received the B.S. degree from Xidian University, Xian, China, in 2010. He is currently a postgraduate student at School of Electro-Mechanical Engineering, Xidian University, China. His research interests include computer vision, machine learning and pattern recognition. Yanhong Hao received the B.S. degree and M.S. degree from Xidian University, Xian, China, in 2001 and 2005, respectively. She is currently a PhD student with the Measuring and Testing Technologies and Instruments, Xidian University, China. Her research interests include signal processing, and pattern recognition. 416 Xinxin Yan received the B.S. degree from Xi’an University of Technology, China, in 2010. Currently, he is a postgraduate student with the School of Electro-Mechanical Engineering, Xidian University, China. His research interests include face recognition and pattern recognition. J Sign Process Syst (2014) 74:405–416 Yang Liu is a postgraduate student at School of Electro-Mechanical Engineering, Xidian University, China. He received his B. S. degree at the Xidian University, Xian, China, in 2010. His research interests include computer vision, machine learning and pattern recognition.