ABSTRACT
We present a novel framework for recognition of facial expressions from a given face image. The framework is based on the assumption that expression information lies in the subspace orthogonal to the subspace representing expression-neutral faces. For deriving the principal subspace of the face images showing no expression, PCA is used as a tool. Then we derive a method to find the orthogonal complement (OC) of the subspace defined by the principal components. It is shown using different tools such as dendrogram and Davies-Bouldin cluster index that the OC of the principal subspace better represents the expressions as compared to the principal subspace in PCA analysis. We have done extensive experiments to validate the recognition capability of the proposed OC space. Two well known publicly available facial expression databases are used for the experiments. We also compare the expression discrimination capability of the OC subspace with some well known features for expression representation. The proposed framework exhibits higher (9.66% on average) recognition capability as compared to the present state-of-the-art works.
- B. Abboud, F. Davoine, and M. Dand. Facial expression recognition and synthesis based on an appearance model. Signal Processing Image Communication, 19(8):723–740, 2004, doi. 10.1016/j.image.2004.05.009.Google ScholarCross Ref
- B. Abboud, F. Davoine, and M. Dang. Expressive face recognition and synthesis. In Computer Vision and Pattern Recognition Workshop, volume 5, June 2003, doi. 10.1109/CVPRW.2003.10056.Google ScholarCross Ref
- S. Agarwal, M. Chatterjee, and D. P. Mukherjee. Synthesis of emotional expressions specific to facial structure. In The Eighth Indian Conference on Vision, Graphics and Image Processing, December 2012, doi. 10.1145/2425333.2425361. Google ScholarDigital Library
- S. Agarwal, M. Chatterjee, and D. P. Mukherjee. Recognizing facial expressions using a novel shape motion descriptor. In The eighth Indian Conference on Vision, Graphics and Image Processing, December 2012, doi. 10.1145/2425333.2425362. Google ScholarDigital Library
- N. Aifanti, C. Papachristou, and A. Delopoulos. The mug facial expression database. In Int. Workshop on Image Analysis for Multimedia Interactive Services, pages 76–84, April 2010.Google Scholar
- M. Bartlett, J. Movellan, and T. Sejnowski. Face recognition by independent component analysis. IEEE Transactions on Neural Networks, 13(6):1450–1464, 2002. Google ScholarDigital Library
- A. J. Calder, A. M. Burton, P. Miller, A. W. Young, and S. Akamatsud. A principal component analysis of facial expressions. SIAM Journal on Scientific and Statistical Computing, 2001, doi: 10.1016/S0042-6989(01)00002-5.Google Scholar
- C. Darwin. The Expression of the Emotions in Man and Animals. IL: Univ. of Chicago Press, Chicago, 1872, 1965.Google ScholarCross Ref
- D. L. Davies and D. W. Bouldin. A cluster separation measure. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1(2):224–227, 1979. Google ScholarDigital Library
- P. Ekman. Facial expressions. in: T. Dalgleish, M. Power(Eds.), Handbook of Cognition and Emotion, Wiley, New York, 1999.Google Scholar
- J. Girard, J. Cohn, M. Mahoor, S. M. Mavadati, and D. Rosenwald. Social risk and depression: Evidence from manual and automatic facial expression analysis. In The 10th IEEE International Conference on Automatic Face and Gesture Recognition, April 2013, doi. 10.1109/FG.2013.6553748.Google ScholarCross Ref
- N. Higham. Computing the polar decompositions-with applications. SIAM Journal on Scientific and Statistical Computing, 7(4):1160–1174, 1986. Google ScholarDigital Library
- B. Jiang, M. F. Valstar, and M. Pantic. Action unit detection fluing sparse appearance descriptors in space-time video volumes. In IEEE International Conference on Automatic Face and Gesture Recognition and Workshops, pages 314–321, March 2011, doi. 10.1109/FG.2011.5771416.Google Scholar
- T. Kanade, Y. Tian, and J. F. Cohn. Comprehensive database for facial expression analysis. Automatic Face and Gesture Recognition, IEEE International Conference on, 0:46, doi. ieeecomputersociety.org/10.1109/AFGR.2000.840611, 2000. Google ScholarDigital Library
- M. Kaur, R. Vashist, and N. Neeru. Recognition of facial expression with principal component analysis and singular value decomposition. International Journal of Computer Applications, 9, Nov 2012.Google Scholar
- P. Lucey, J. F. Cohn, T. Kanade, and J. Saragih. The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In Computer Vision and Pattern Recognition, pages 94–101, June 2010.Google ScholarCross Ref
- D. McDuff, R. E. Kalioubyb, T. Senechal, D. Demirdjian, and R. Picard. Automatic measurement of ad preferences from facial responses gathered over the internet. Image and Vision Computing, 2014, doi: 10.1016/j.imavis.2014.01.004.Google ScholarCross Ref
- D. P. Mukherjee, K. Higashiura, T. Okada, M. Hori, Y.-W. Chen, N. Tomiyama, and Y. Sato. Utilizing disease-specific organ shape components for disease discrimination: Application to discrimination of chronic liver disease from ct data. In MICCAI 2013, Part I, LNCS 8149, pages 235–242, 2013.Google ScholarCross Ref
- Y. Saatci and C. Town. Cascaded classification of gender and facial expression using active appearance models. In In Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition, pages 393–398, April 2006, doi. 10.1109/FGR.2006.29. Google ScholarDigital Library
- C. Shan, S. Gong, and P. W. McOwan. Facial expression recognition based on local binary patterns: A comprehensive study. Image Vision Comput., 27:803–816, May 2009. Google ScholarDigital Library
- K. Sikka, A. Dhall, and M. Bartlett. Weakly supervised pain localization using multiple instance learning. In FG'13, page NA, 2013, doi. 10.1109/FG.2013.6553762.Google ScholarCross Ref
- A. Sánchez, J. V. Ruiz, A. B. Moreno, A. S. Montemayor, J. Hernández, and J. J. Pantrigo. Differential optical flow applied to automatic facial expression recognition. Neurocomputing, 74:1272–1282, 2011. Google ScholarDigital Library
- D. Tao and X. Tang. Orthogonal complement component analysis for positive samples in svm based relevance feedback image retrieval. In Computer Vision and Pattern Recognition, volume 2, pages 586–591, 2004, doi: 10.1109/CVPR.2004.1315217. Google ScholarDigital Library
- U. Tariq, J. Yang, and T. S. Huang. Maximum margin gmm learning for facial expression recognition. In FG'13, page NA, 2013, doi. 10.1109/FG.2013.6553794.Google ScholarCross Ref
- M. Valstar, I. Patras, and M. Pantic. Facial action unit detection using probabilistic actively learned support vector machines on tracked facial point data. In IEEE Conference on Computer Vision and Pattern Recognition, volume 3, pages 76–84, 2005, doi. 10.1109/CVPR.2005.457. Google ScholarDigital Library
- Z. Zhang, M. J. Lyons, M. Schuster, and S. Akamatsu. Comparison between geometry-based and gabor-wavelets-based facial expression recognition using multi-layer perceptron. In IEEE International Conference on Automatic Face and Gesture Recognition, pages 454–459, April 1998. Google ScholarDigital Library
- G. Zhao and M. Pietikäinen. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(6):915–928, 1986. Google ScholarDigital Library
Index Terms
- Recognizing Facial Expressions in the Orthogonal Complement of Principal Subspace
Recommendations
Recognising facial expressions in video sequences
We introduce a system that processes a sequence of images of a front-facing human face and recognises a set of facial expressions. We use an efficient appearance-based face tracker to locate the face in the image sequence and estimate the deformation of ...
Face Recognition Through Different Facial Expressions
Face recognition has become an accessible issue for experts as well as ordinary people as it is a focal non-interfering biometric modality. In this paper, we introduced a new approach to perform face recognition under varying facial expressions. The ...
Biased subspace learning for misalignment-robust facial expression recognition
In this paper, we propose a biased subspace learning approach for misalignment-robust facial expression recognition. While a variety of facial expression recognition methods have been proposed in the literature, most of them only work well when face ...
Comments