Abstract
Human facial expressions are key ingredient to convert an individual’s innate emotion in communication. However, the variation of facial expressions affects the reliable identification of human emotions. In this paper, we present a cloud model to extract facial features for representing human emotion. First, the uncertainties in facial expression are analyzed in the context of cloud model. The feature extraction and representation algorithm is established under cloud generators. With forward cloud generator, facial expression images can be re-generated as many as we like for visually representing the extracted three features, and each feature shows different roles. The effectiveness of the computing model is tested on Japanese Female Facial Expression database. Three common features are extracted from seven facial expression images. Finally, the paper is concluded and remarked.
Similar content being viewed by others
References
Bar M (2009) The proactive brain: memory for predictions. Phil Trans R Soc B 364:1235–1243
Beymer D (1994) Face recognition under varying pose. In: Proceedings of 23rd Image Understanding Workshop, vol 2, pp 837-842
Bindemann M, Scheepers C, Burton AM (2009) Viewpoint and center of gravity affect eye movements to human faces. J Vis 9(2):1–16
Burton AM, Jenkins R, Hancock PJ, White D (2005) Robust representations for face recognition: the power of averages. Cogn Psychol 51(3):256–284
Carbon CC (2008) Famous faces as icons. The illusion of being an expert in the recognition of famous faces. Perception 37(5):801–806
De Bie T, Cristianini N, Rosipal R (2004) Eigenproblems in pattern recognition. In: Bayro-Corrochano E (ed) Handbook of computational geometry for pattern recognition, computer vision, neurocomputing and robotics. Springer-Verlag, Heidelberg
Deng W, Guo J, Hu J, Zhang H (2008) Comment on “100% accuracy in automatic face recognition”. Science 321:912c
Gross R, Shi J, Cohn J (2001) Quo vadis Face Recognition?—the current state of the art in face recognition. Technical Report, Robotics Institute, Carnegie Mellon University, Pittsburgh, 25 pages
Harry W (2007) Reliable face recognition methods: system design, implementation and evaluation. Springer, Berlin
Jain AK, Duin RPW, Mao J (2000) Statistical pattern recognition: a review. IEEE Trans Pattern Anal Mach Intell 22(1):4–37
Jenkins R, Burton AM (2008a) 100% Accuracy in automatic face recognition. Science 319:435
Jenkins R, Burton AM (2008b) Response to comment on “100% accuracy in automatic face recognition”. Science 321:912d
Jenkins R, Burton AM (2011) Stable face representations. Philos Trans R Soc Lond Ser B Biol Sci 366(1571):1671–1683
Johnston RA, Edmonds AJ (2009) Familiar and unfamiliar face recognition: a review. Memory 17(5):577–596
Li DY, Du Y (2007) Artificial intelligence with uncertainty. Chapman and Hall/CRC, London
Li DY, Liu CY, Gan WY (2009) A new cognitive model: cloud model. Int J Intell Syst 24(3):357–375
Li DR, Wang SL, Li DY (2015) Spatial data mining theory and application. Springer, Berlin
Martınez AM (2003) Recognizing expression variant faces from a single sample image per class. In: Proceedings of IEEE Computer Vision and Pattern Recognition, Madison, Vol 1, pp 353
Mihalcea R (2012) Multimodal sentiment analysis. In: Proceedings of the 3rd Workshop in Computational Approaches to Subjectivity and Sentiment Analysis. Association for Computational Linguistics, Stroudsburg, p 1–1
Moghaddam B (2002) Principal manifolds and probabilistic subspaces for visual recognition. IEEE Trans Pattern Anal Mach Intell 24(6):780–788
Newman DJ, Hettich S, Blake CL, Merz CJ (1998) UCI Repository of machine learning databases. University of California, Department of Information and Computer Science, Irvine. http://www.ics.uci.edu/~mlearn/MLRepository.html
Pentland A, Choudhury T (2000) Face recognition for smart environments. IEEE Trans Comput 33(2):50–55
Ralph G, Simon B, Iain M, Takeo K (2004) Face recognition across pose and illumination. In: Li SZ, Jain AK (eds) Handbook of face recognition. Springer-Verlag, Berlin
Ryu JJ, Chaudhuri A (2006) Representations of familiar and unfamiliar faces as revealed by viewpoint-aftereffects. Vis Res 46(23):4059–4063
Severyn A, Moschitti A (2015) Twitter sentiment analysis with deep convolutional neural networks. In: Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, New York, p 959–962
Singh S (2003) Multiresolution estimates of classification complexity. IEEE Trans Pattern Anal Mach Intell 25(12):1534–1539
Wang SL (2002) Data field and cloud model based spatial data mining and knowledge discovery. Ph.D. Thesis (Wuhan University, Wuhan)
Wang SL, Yuan HN (2006) View-angle of spatial data mining. Lect Notes Artif Intell 4093:1065–1076
Yang M-H, Kriegman DJ, Ahuja N (2002) Detecting faces in images: a survey. IEEE Trans Pattern Anal Mach Intell 24(1):34–58
Zhang S, Zhang S, Huang, GW (2016) Multimodal deep convolutional neural network for audio-visual emotion recognition. In: Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval (ICMR ‘16). ACM, New York, p 281–284
Zhao W, Chellappa R, Rosenfeld A, Phillips PJ (2003) Face recognition: a literature survey. ACM Comput Surv:399–458
Acknowledgments
The authors would like to thank Professor Kevin P. Chen and Ching-Chung Li for their proof-reading and comments for this paper. This work was supported in part by a grant from National Natural Science Fund of China (61472039), National Key Research and Development Plan of China (2016YFC0803000, 2016YFB0502604), and Frontier and interdisciplinary innovation program of Beijing Institute of Technology (2016CX11006).
Author information
Authors and Affiliations
Corresponding author
Additional information
Responsible editor: Philippe Garrigues
Rights and permissions
About this article
Cite this article
Wang, S., Chi, H., Yuan, H. et al. Extraction and representation of common feature from uncertain facial expressions with cloud model. Environ Sci Pollut Res 24, 27778–27787 (2017). https://doi.org/10.1007/s11356-017-0237-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11356-017-0237-2