Intelligent Control and Automation, 2008. WCICA 2008. • NMF considers factorizations of the form: π ≈ ππ» Where π ∈ π + πΉ×πΏ , π ∈ π + πΉ×π , π» ∈ π + π×πΏ , π βͺ πΉ • To measure the cost of the decomposition, one popular approach is to use the Kullback-Leibler (KL) divergence metric, the cost for factorizing X into ZH is evaluated as: πΏ π·πππΉ (π| ππ» = πΏ π=1 πΉ = π₯π,π ln π=1 π=1 πΎπΏ(π₯π | πβπ π₯π,π + π§ β π π,π π,π π§π,π βπ,π − π₯π,π π • Using the Expectation Maximization (EM) algorithm and an appropriately designed auxiliary function, it has been shown in “Algorithms for non-negative matrix factorization,” the update rule for the π‘-th iteration for βπ,π (π‘) is given by: π₯π,π (π‘−1) π π§π,π (π‘−1) β (π‘−1) π§ π,π π π,π βπ,π (π‘) = βπ,π (π‘−1) (π‘−1) π§ π,π π • while for π§π,π (π‘) the update rule is given by: π§′ π,π (π‘) π βπ,π = π§π,π (π‘−1) (π‘−1) π₯π,π (π‘−1) (π‘−1) π§β β π,π π,π π (π‘−1) β π π,π • Finally, the basis images matrix π is normalized so that its column vectors elements sum up to one: ′ (π‘) π§ π,π π§π,π (π‘) = ′ (π‘) π§ π π,π Cluster 1 … Class 1 Cluster πΆ1 Cluster 1 … Cluster πΆπ Image π … … … Database π … … Class π Image 1 Cluster πΆπ Image π(π)(π) Cluster 1 Class π … Cluster πΆπ Image 1 Database π Image π Class π Image π(π)(π) Cluster πΆπ NMF Dimensionality reduction Feature Vector ππ (π)(π) = ππ,1 (π)(π) β― ππ,π (π)(π) Mean of Feature Vector for the π-π‘β cluster of the π-π‘β classοΌ π (π)(π) = 1 π(π)(π) π(π)(π) ππ (π)(π) = π1 (π)(π) β― ππ (π)(π) π=1 π π • Using the above notations we can define the within cluster scatter matrix ππ€ asοΌ π πΆπ π(π)(π) ππ€ = ππ (π)(π) − π (π)(π) ηπ (π)(π) − π (π)(π) π π=1 π=1 ρ=1 • and the between cluster scatter matrix ππ asοΌ π π πΆπ πΆπ π (π)(π) ππ = − π (π)(π) π (π)(π) π=1 π≠1 π=1 π=1 • Our GoalοΌ π‘π ππ€ ↓ πππ π‘π[ππ ] ↑ − π (π)(π) π Since we desire the trace of matrix ππ€ to be as small as possible and at the same time the trace of ππ to be as large as possible, the new cost function is formulated asοΌ πΌ π½ π·ππ·πππΉ π||ππ» = π·πππΉ π||ππ» + π‘π ππ€ − π‘π ππ 2 2 1 where πΌ and π½ are positive constants, while is used to simplify 2 subsequent derivations. Consequently, the new minimization problem is formulated as: min π·ππ·πππΉ π||ππ» π,π» subject to : π§π,π ≥ 0, βπ,π ≥ 0, π ππ,π = 1, ∀π, π, π. • The constrained optimization problem is solved by introducing Lagrangian multipliers βΆ β πΌ π½ = π·πππΉ π||ππ» + π‘π ππ€ − π‘π ππ + ∅π,π π§π,π 2 2 π π + π π Ρ±π,π βπ,π πΌ π½ β = π·πππΉ π||ππ» + π‘π ππ€ − π‘π ππ + π‘π ∅π π + π‘π Ρ±π» π 2 2 • Consequently, the optimization problem is equivalent to the minimization of the Lagrangian πππ min β π,π» • To minimize β, we first obtain its partial derivatives with respect to π§π,π and βπ,π and set them equal to zero : πβ =− πβπ,π πβ =− ππ§π,π π π π₯π,π π§π,π + π§ β π π,π π,π π₯π,π π§π,π + π§ β π π,π π,π π π πΌ ππ‘π ππ€ π½ ππ‘π ππ π§π,π + Ρ±π,π + + 2 πβπ,π 2 ππ§π,π πΌ ππ‘π ππ€ π½ ππ‘π ππ π§π,π + Ρ±π,π + + 2 πβπ,π 2 ππ§π,π • DNMF combines Fisher’s criterion in the NMF decomposition • and achieves a more efficient decomposition of the • provided data to its discriminant parts, thus enhancing separability • between classes compared with conventional NMF π = π£1 , π£2 , … , π£π ∈ π 1×π ↓ Dimensionality reduction π = π’1 , π’2 , … , π’π ∈ π 1×π Where π βͺ π π∗π΅ =π Where π΅ ∈ π π×π π∗π΅ =π Where π ∈ π 1×2 π΅ ∈ π 2×1 π ∈ π 1×1 • • • • • • Introduction Principal Component Analysis (PCA) Method Non-negative Matrix Factorization (NMF) Method PCA-NMF Method Experiments Result and Analysis Conclusion • In this paper, we have detailed PCA and NMF, and applied them to feature extraction of facial expression images. • We also try to process basic image matrix and weight matrix of PCA and make them as the initialization of NMF. • The experiments demonstrate that the method has got a better recognition rate than PCA and NMF. Let’s suppose that m expression images are selected to take part in training, the training set X is defined by π = π₯1, π₯2 … π₯π ∈ π π×π (1) Covariance matrix corresponding to all training samples is obtained as π πΆ= π₯π − π’ π₯π − π’ π (2) π=1 u, average face, is defined by π 1 π’= π₯π π π=1 (3) π = πππ€ × πππ # of training data π=π πππ average face π + + πππ€ π π₯1 π₯2 π₯3 π₯4 Training Data Set π ∈ π π×π π’ + Let A = π₯1 − π’, π₯2 − π’ … π₯π − π’ Then (2) becomes πΆ = π΄π΄π π΄ ∈ π π×π πΆ ∈ π π×π (4) (5) Matrix πΆ has π eigenvectors and eigenvalues. Image 50x50 π = 50 × 50 = 2500 It is difficult to get 2500 eigenvectors and eigenvalues. Therefore, we get eigenvectors and eigenvalues of π΄π΄π by solving eigenvectors and eigenvalues of π΄π π΄. π΄π π΄ ∈ π π×π The vectors π£π (π = 1,2 … π)and scalars ππ (π = 1,2 … π)are the eigenvectors and eigenvalues of covariance matrix π΄π π΄. Then, eigenvectors π’π of π΄π΄π are defined by 1 π’π = π΄π£π π = 1,2 … π 6 ππ Sorting ππ by size: π1 ≥ π2 ≥ β― ≥ ππ > β― > 0 Generally, the scale is capacity that π eigenvalues occupied: π π=1 ππ ≥ πΌοΌusally, πΌ is 0.9~0.99. (7) π π=1 ππ Set π is a projection matrix, π = π1 , π2 … ππ . And then, every facial expression image feature can be denoted by following equation ππ = π π π₯π − π’ (8) PCA basic images Given a non-negative matrix π, the NMF algorithms seek to find non-negative factors π΅ and π» of π ,such that : ππ×π ≈ π΅π×π π»π×π (9) where π is the number of feature vector satisfies π×π π< π+π (10) Iterative update formulae are given as follow: π΅π π ππ π»ππ ← π»ππ π π΅ π΅π» ππ ππ» π ππ π΅ππ ← π΅ππ π΅π»π» π ππ set π΄ = π΅π», then define objective function min π − π΄ 2 (11) (12) (13) And then, every facial expression image feature can be denoted by following equation ππ = π −1 π₯π (14) NMF basic images First, get projective matrix π and weight matrix πΊ by PCA method. Initialization is performed for matrices π΅ and π» by following π΅ = min 1, abs πππ (15) π» = min 1, abs πΊππ (16) NMF basic images PCA-NMF basic images anger anger disgust disgust fear fear happy neutral happy neutral sad sad surprise surprise The comparison of recognition rate for every expression ( The training set comprises 70 images and the test set of 70 images) The comparison of recognition rate for every expression ( The training set comprises 70 images and the test set of 143 images) The comparison of recognition rate for every expression ( The training set comprises 140 images and the test set of 73 images) The discussion or r The results of experiments demonstrate that NMF and PCA-NMF can outperform PCA. The best recognition rate of facial expression image is 93.72%. On the whole, our approach provides good recognition rates.