Here

advertisement
[1] Bartlett, M.S., Littlewort, G.C., Lainscsek, C., Fasel, I., Frank, M.G., Movellan, J.R., “Fully automatic facial action
recognition in spontaneous behavior”, In 7th International Conference on Automatic Face and Gesture Recognition, 2006,
p. 223-228.
Bartlett et al. [1] proposed a system that can automatically recognize frontal views of the image from a video stream,
in which 20 AUs are detected for each frame. Here context-independent training has been used. One binary classifier
has been used for each of the 20 AUs. These classifiers have been trained to recognize the occurrence of AUs
regardless of their co-occurrence (a specific AU occurring alone or with others). They also compared the
performance analysis of AdaBoost and linear SVM. Paired t-tests showed a slight advantage of AdaBoost over
linear SVM. One interesting feature that the authors tried to measure is the intensity of specific AUs. They used the
output margin of the system which actually describes the distance to the separating hyperplane as an interpretation
for the intensity of the AU.
[2]Braathen, B., Bartlett, M.S., Littlewort-Ford, G., Smith, E. and Movellan, J.R. (2002). An approach to automatic
recognition of spontaneous facial actions. Fifth International Conference on automatic face and gesture recognition, pg.
231-235.
[3] F. Pighin, J. Hecker, D. Lischinski, R. Szeliski, and D. H. Salesin, “Synthesizing realistic facial expressions from
photographs”, Computer Graphics, 32(Annual ConferenceSeries):75–84, 1998.
Braathan et al. [2] address a natural problem with image collection and shift the paradigm from 2D to 3D facial
images. Older automated facial expression recognition systems have relied on posed images where images clearly
show the frontal view of the face. But this is impractical. Many times the head is in an out-of-image-plane (turned or
nodded) in spontaneous facial images. Braathan et al. tackled three issues sequentially in their project. First the face
geometry was estimated. Thirty images were taken for each subject and the position of eight special features was
identified (ear lobes, lateral and nasal corners of the eyes, nose tip, and base of the center upper teeth). 3D locations
of these eight features were then recovered. These eight 3D location points were fitted in the canonical view of the
subject. Later a scattered data interpolation technique [3] was used to generate and fit the other unknown points in
the face model. Second, a 3D pose estimation technique known as Markov Chain Monte-Carlo method or particle
filtering method was used. This generated a sequence of 3D poses of the head. Canonical face geometry was then
used to warp these images on to a face model and rotate these into a frontal view. Then this image was projected
back to the image plane. Finally, Braatthan et al. used Support Vector Machines (SVM) and Hidden Markov Models
(HMM) for the training and learning of their spontaneous facial expression detection system. Since normally the
head is in a slanted position during severe pain, this system is especially useful for analyzing real time painful
expressions.
[4] Jagdish Lal Raheja, Umesh Kumar, “Human facial expression detection from detected in captured image using back
propagation neural network”, In International Journal of Computer Science & Information Technology (IJCSIT), Vol. 2,
No. 1, Feb 2010,116-123.
[5] Paul Viola, Michael Jones, ”Rapid Object Detection using a Boosted Cascade of Simple features”, Conference on
computer vision and pattern recognition, 2001.
Jagdish and Umesh proposed a simple architecture for facial expression recognition based on token finding and
standard error-based back propagation neural network [48]. After capturing the image from a webcam, they used the
face detection technique devised by Viola and Jones [11]. In order to process the image, histogram equalization was
done to enhance image quality followed by edge detection and thinning. Tokens, which denote the smallest unit of
information, were generated from the resultant image and passed into the neural network. The network was trained
with 100 samples and could classify three expressions. The provided report did not say anything about the number
of nodes in the input layer and hidden layer of the network. It also did not mention about the training samples.
[6] A. B. Ashraf, S. Lucey, J. F. Cohn, T. Chen, K. M. Prkachin, and P. E. Solomon. The painful face II-- Pain expression
recognition using active appearance models. International Journal of Image and Vision Computing, 27(12):1788-1796,
November 2009.
Active Appearance Model (AAM) [6] has been proposed as an innovative way for pain recognition from facial
expressions. AAMs were used to develop the automated machine learning-based system. Use of a Support Vector
Machine (SVM) and leave-one-out procedures lead to a hit rate of 81%. The main advantage of AAMs is the feature
of decoupling appearance and shape parameters from facial images. In AAM, shape s is expressed as a 2D
triangulated mesh. Location of the mesh vertices is related to the original image from which the shape s was derived.
In AAMs, a shape s can be expressed as a combination of a base shape s0 and a set of shape vectors si . Three AAMderived representations are pointed: Similarity Normalized Shape (sn), Similarity Normalized Appearance (an), and
Shape Normalized Appearance (a0). The authors developed their own representations using AAM-derived
representations that they used in painful face detection. Temporal information was not used in recognition of pain
which may increase the accuracy rate.
[7] Md. Maruf Monwar, Siamak Rezaei and Dr. Ken Prkachin, “Eigenimage Based Pain Expression Recognition”, In
IAENG International Journal of Applied Mathematics, 36:2, IJAM_36_2_1. (online vversion available 24 May 2007)
[8] Md. Maruf Monwar, Siamak Rezaei: Appearance-based Pain Recognition from Video Sequences. IJCNN 2006: 24292434
Monwar et al. proposed an automatic pain expression detection system using Eigenimage [7]. Skin color modeling
was used to detect the face from video sequence. Mask image technique was used to extract the appropriate portion
of the face for detecting pain. Each resultant image from masking was projected into a feature space to form an
Eigenspace based on training samples. When a new image arrived, its position in the feature space was compared
with that of the training samples and based on that a decision is drawn (pain or no pain). For this experiment, 38
subjects of different ethnicity, age, and gender were videotaped for two expressions – normal and painful. First the
chromatic color space was used to find the distribution of skin color. Later the Gaussian model was used to represent
this distribution. Then the probability of a pixel being skin was obtained. After segmenting the skin region, a
meaningful portion of the face was detected using mask image and later a bitwise ‘AND’ operation was used with
the mask image and original image to develop the resultant image. These resultant images were used as samples for
training the Eigenfaces method and M number of Eigenfaces with highest Eigenvalues being sorted out. When
detecting a new face, the facial image was projected in the Eigenspace, and the Euclidian distance between the new
face and all the faces in the Eigenspace was measured. The face that represents the closest distance was assumed to
be a match for the new image. Average hit rate was recorded to be 90-92%. Later, the researchers extended their
model [8] to create two more feature spaces – Eigeneyes and Eigenlips. Portions of eyes and lips were used from
facial images for Eigeneyes and Eigenlips methods. All possible combinations of Eigenfaces, Eigeneyes, and
Eigenlips (alone or in combination with others) were used to find pain from images. A combination of all three
together provided the best result in terms of accuracy (92.08%). Here skin pixels have been sorted out using
chromatic color space. However skin color varies a lot based on race and, ethnicity, and detailed elaboration of the
subjects’ skin colors is missing in these studies.
[9] Mayank Agarwal, Nikunj Jain, Manish Kumar, and Himanshu Agrawal, “Face recognition using principle component
analysis,eigneface, and neural network”, In International Conference on Signal Acquisition and Processing, ICSAP, 310314.
Here the authors proposed a methodology for face recognition based on information theory. Principle component
analysis and feed forward back propagation Neural Network has been used for feature extraction and face
recognition respectively. The algorithm has been used on 400 images of 40 different people taken form Olivetti and
Oracle Research Laboratory (ORL) face database. Training dataset is composed of 60% of the images and the rest
are left for test database. The Artificial Neural Network (ANN) had 3 layers – input, output, and one hidden layer.
Though this method achieved 97% accuracy this had two major drawbacks. The method incorporates one ANN for
each person in the face database which is impractical in terms of scalability. Another issue is they have tested only
with different images of the same person whose picture has been used in the training of the ANN which is not
practical for a real life scenario.
[10] Murthy, G. R. S. and Jadon, R. S. (2009). Effectiveness of eigenspaces for facial expression recognition. International
Journal of Computer Theory and Engineering, Vol. 1, No. 5, pp. 638-642.
Here the authors proposed a modified methodology for eigenface based facial expression recognition. The authors
formed a separate eigenspace for each of the six basic emotions from the training images. When a new image comes,
the new image is projected in each of the eigenspaces and a reconstructed image is formed from each eigenspaces. A
classification decision is taken by measuring the mean squeare error between the input image and the reconstructed
image. Using Cohn-Kanade and JAFFE (Japanese Female Facial Expression) database, the method received a
maximum of 83% accuracy for ‘happiness’ and 72% accuracy for ‘disgust’. The paper does not say anything about
facial contour selection. The classification accuracy rate is also quite low.
[11] Singh. S. K., Chauhan D. S., Vatsa M., and Singh R. (2003). A robust skin color based face detection algorithm.
Tamkang Journal of Science and Engineering, Vol. 6, No. 4, pp. 227-234.
Here the authors compared algorithms based on 3 different color spaces RGB, YCbCr, and HSI. Later they
combined them to derive a modified skin color based face detection algorithm. An algorithm based on the venn
diagram of set theory has been used to detect the skin color. A face portion from the original image is taken apart by
detecting 3 points – midpoint of eyes and leaps. This method showed 95% accuracy rate on IITK database. But this
method only detects the facial portion from any image. It does not say anything about expression classification.
[12] Rimé, B., Finkenauera, C., Lumineta, O., Zecha, E., and Philippot, P. 1998. Social Sharing of Emotion: New Evidence
and New Questions. In European Review of Social Psychology, Volume 9.
We have also done an analysis on the mood related applications of Facebook. Rimé et al. [12] argued that everyday
life emotional experiences create an urge for social sharing. Authors also showed that most emotional experiences
are shared with others shortly after they occurred. These research findings show that mood sharing can be an
important area in social networks. At present there are many applications in Facebook which claim to do mood
detection for the user. Here, we have listed top 10 such applications based on the number of active users.
Name
of
Application
My Mood
SpongeBob Mood
The Mood Weather
Report
Name and Mood
Analyzer
Mood Stones
My Friend's Mood
My Mood Today!
How's your mood
today?
Your Mood of the
day
Patrick’s Mood
No. of users
1,822,571
646,426
Mood Categories
56
42
611,132
174
29,803
14,092
15,965
11,224
13
9
6,694
39
4,349
3329
15
Table 1: Mood applications in Facebook
Here user chooses a mood category (symbol) that he/she thinks is the most suitable one. Once the choice is made by
the user, the application gives him the option of sharing the mood with his friends on Facebook. The mood is then
displayed on user’s wall. It is encouraging to see that a large number of users are interested in using the mood
related application. This is the case when users are manually selecting one of the mood categories. If we give them
the mood detection application which works on the principle of image processing and then gives the context aware
results instead of just manually selecting one of the categories, it will really prove revolutionary.
Download