Abstract
Usage of the system for human emotion recognition has been increased in various types of applications of affective computing fields such as human sign language understanding, identification of human mental disorder, and human-computer interaction. Here, we report a video frame-based procedure for the estimation of human emotional behavior. We introduce Circumradius-Incenter-Circumcenter combined geometric signature (CIC) induced from our proposed triangulation method. The method first includes the step of salient landmark identification from face image frames by using the Active Appearance Model (AAM). Here, we fetch geometric features from triangles drawn by landmark points, thereafter core triangles are found based on the CIC feature which plays an important role to get an interpretation of changing information of human emotions. In the end, the extracted core features from core triangles are employed into the Multilayer Perceptron (MLP) classifier to get recognition accuracy. The discrimination power of our proposed system is evaluated on well-known three benchmark video face frame databases, viz., CK+, MMI, and MUG. Moreover, the performance of the proposed procedure is validated by presenting the comparison task with other existing methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mehrabian A, Russell JA (1974) An approach to environmental psychology. MIT Press
Ekman P, Friesen WV (1971) Constants across cultures in the face and emotion. J Pers Soc Psychol 17(2):124
Ghimire D, Lee J, Li Z-N, Jeong S (2017) Recognition of facial expressions based on salient geometric features and support vector machines. Multimed Tools Appl 76(6):7921–7946
Cruz AC, Bhanu B, Thakoor NS (2014) Vision and attention theory based sampling for continuous facial emotion recognition. IEEE Trans Affect Comput 5(4):418–431
Barman A, Dutta P (2020) Human emotion recognition from face images. Springer
Barman A, Dutta P (2019) Facial expression recognition using distance and texture signature relevant features. Appl Soft Comput 77:88–105
Cootes TF, Edwards GJ, Taylor CJ (2001) Active appearance models. IEEE Trans Pattern Anal Mach Intell 23(6):681–685
Cootes TF, Taylor CJ, Cooper DH, Graham J (1995) Active shape models-their training and application. Comput Vis Image Underst 61(1):38–59
Sagonas C, Antonakos E, Tzimiropoulos G, Zafeiriou S, Pantic M (2016) 300 faces in-the-wild challenge: database and results. Image Vis Comput 47:3–18
Lucey P et al (2010) The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE computer society conference on computer vision and pattern recognition-workshops. IEEE
Valstar M, Pantic M (2010) Induced disgust, happiness and surprise: an addition to the MMI facial expression database. In: Proceedings of 3rd international workshop on EMOTION (satellite of LREC): corpora for research on emotion and affect
Aifanti N, Papachristou C, Delopoulos A (2010) The MUG facial expression database. In: 11th international workshop on image analysis for multimedia interactive services WIAMIS 10. IEEE
Yaddaden Y, Adda M, Bouzouane A, Gaboury S, Bouchard B (2017) Facial expression recognition from video using geometric features, pp 4-6
Saeed A, Ayoub A-H, Robert N, Moftah E (2014) Frame-based facial expression recognition using geometrical features. Adv Hum Comput Interact
Wan C, Tian Y, Liu S (2012) Facial expression recognition in video sequences. In: Proceedings of the 10th world congress on intelligent control and automation. IEEE, pp 4766-4770
Shan C, Gong S, McOwan PW (2009) Facial expression recognition based on local binary patterns: a comprehensive study. Image Vis Comput 27(6):803–816
Rahulamathavan Y, Phan RC-W, Chambers JA, Parish DJ (2012) Facial expression recognition in the encrypted domain based on local fisher discriminant analysis. IEEE Trans Affect Comput 4(1):83–92
Acknowledgements
The authors want to state their gratefulness to Prof. Maja Pantic and Dr. A. Delopoulos for making available to use the MMI and MUG databases. The authors also like to express thanks to Department of Science and Technology, Ministry of Science and Technology, Government of India, for supporting with DST-INSPIRE Fellowship (INSPIRE Reg. no. IF160285, Ref. No.: DST/INSPIRE Fellowship/[IF160285]) to carry out research work. The authors are thankful to Department of Computer & System Sciences, Visva-Bharati University for providing infrastructure support.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Nasir, M., Dutta, P., Nandi, A. (2021). Recognition of Transforming Behavior of Human Emotions from Face Video Sequence: A Triangulation-Induced Circumradius-Incenter-Circumcenter Combined Approach. In: Bhattacharyya, S., Dutta, P., Datta, K. (eds) Intelligence Enabled Research. Advances in Intelligent Systems and Computing, vol 1279. Springer, Singapore. https://doi.org/10.1007/978-981-15-9290-4_9
Download citation
DOI: https://doi.org/10.1007/978-981-15-9290-4_9
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-9289-8
Online ISBN: 978-981-15-9290-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)