Face Recognition by Support Vector Machines

advertisement
Face Recognition by Support Vector
Machines
Guodong Guo, Stan Z. Li, and Kapluk Chan
School of Electrical and Electronic Engineering
Proceedings of the Fourth IEEE International Conference on Automatic
Face and Gesture Recognition 2000
Nanyang Technological University, Singapore 639798
指導教授:王啟州 教授
學生:陳桂華
Outline







Abstract
Introduction
Basic Theory of Support Vector Machines
Multi-class Recognition
Experimental Results-Face Recognition on the ORL Face Database
Experimental Results-Face Recognition on a Larger Compound Database
Conclusions
Abstract


In this paper, the SVMs with a binary tree recognition
strategy are used to tackle the face recognition problem.
We compare the SVMs based recognition with the standard
eigenface approach using the Nearest Center Classification
(NCC) criterion.
Introduction


Face recognition technology can be used in wide range of
applications such as identity authentication, access control,
and surveillance.
We propose to construct a binary tree structure to recognize
the testing samples.
Basic Theory of Support Vector
Machines(1/5)
Fig 1. Classification between two classes using hyperplanes
Basic Theory of Support Vector
Machines(2/5)

1
點集合 {xi , yi }, i  1,..., n
xi  R d , yi {1,1}

直線 f ( x)  wT x  b

使所有
(2)
(3)
的點落在 yi  1 這一邊,而所有
f ( x)  0 的點落在 yi  1 這一邊,就可以根據 f (x) 的
正負號來區分點是屬於兩個集合的哪一邊。
f ( x)  0
Basic Theory of Support Vector
Machines(3/5)

那兩條虛線是支援超平面(Support Hyperplane),式子
如下: T
w x  b 
(4)
wT x  b  
(5)

中間那條線為最佳支援超平面與margin距離為d
margin  2d  2 / w  w 越小則margin越大

將限制條件寫成下面兩個式子:
wT xi  b  1
yi  1
(6)
wT xi  b  1
yi  1
( 7)
Basic Theory of Support Vector
Machines(4/5)

兩個限制式可以寫成一個限制式,如下:


yi wT xi  b  1  0


(8)
總合上式,得以下的目標函式:
1 2
minimize 2 w
i
subject to yi ( wT xi  b)  1  0
可以利用 Lagrange Multiplier Method將上面的式子轉成
一個二次方程式,找出可以使L為最小值的w, b,  i。


1 2 N
Lw, b,    w    i yi wT xi  b   1
2
i 1
(9)
Basic Theory of Support Vector
Machines(5/5)

符合條件的極值點會出現在:
T
當 yi w xi -b   1  0 時,  i  0
當 yi wT xi -b   1  0 時, i 必為0
Multi-class Recognition

One is the one-against-all strategy to classify between each
class and all the remaining; The other is the one-againstone strategy to classify between each pair.
Fig 2. The binary tree structure for 8 classes face recognition.
Experimental Results-Face Recognition on the
ORL Face Database(1/5)

The first experiment is performed on the CambridgeORL
face database, which contains 40 distinct persons. Each
person has ten different images, taken at different times.
Fig 3. Four individuals (each in one row) in the ORL face database.
Experimental Results-Face Recognition on the
ORL Face Database(2/5)


Hidden Markov model (HMM) based approach is used,
and the best model resulted in a 13% error rate.
Later, Samaria extends the top-down HMM [14] with
pseudo two-dimensional HMMs [13], and the error rate
reduces to 5%.
Experimental Results-Face Recognition on the
ORL Face Database(3/5)


Lawrence et al [6] takes the convolutional neural network
(CNN) approach for the classification of ORL database,
and the best error rate reported is 3:83%(in the average of
three runs).
Face recognition experiments on the ORL database, we
select 200 samples randomly as the training set.
Experimental Results-Face Recognition on the
ORL Face Database(4/5)


The remaining 200 samples are used as the test set.
The average minimum error rate of SVM in average is
3.0%, while the NCC is 5.25%.
Experimental Results-Face Recognition on the
ORL Face Database(5/5)
Fig 4. Comparison of error rates versus the number of eigenfaces of the
standard NCC and SVM algorithms on the ORL face database.
Experimental Results-Face Recognition on a
Larger Compound Database




The second experiment is performed on a compound data
set of 1079 face images of 137 persons.
It is composed of 544 images.
The remaining 535 images are used as the test set.
In this experiment, the number of classes c = 137, and the
c c ‐1
SVMs based methods are trained for
= 18632 pairs.
2
Experimental Results-Face Recognition on a
Larger Compound Database

To construct the binary trees for testing, we decompose
137 = 32 + 32 + 32 + 32 + 8 + 1.
Experimental Results-Face Recognition on a
Larger Compound Database

The minimum error rate of SVM is 8.79%,which is much
better than the 15.14% of NCC.
Fig 5. Comparison of error rates versus the number of eigenfaces of the
standard NCC and SVM algorithms on the compound face database.
Conclusions


We have presented the face recognition experiments using
linear support vector machines with a binary tree
classification strategy.
The experimental results show that the SVMs are a better
learning algorithm than the nearest center approach for
face recognition.
Thanks for your attendance!
Download