Abstract
Human emotions play an important role in the interpersonal relations. Emotions are reflected by means of a facial expression. Research and understanding of emotions are very important for human-machine interaction. This article is describing the system for automatic recognition of emotions in a video stream. The main purpose of the work is to develop a method that increases the accuracy of recognizing emotions in the video stream. The separate paragraph describes methods for recognition of the eyes and lips. The article provided the results of comparing the data obtained from the training selection. The recognition accuracy of the developed method is compared with the Artificial Neural Network algorithm. The article considers the main algorithm for obtaining key parameters from a video. The analysis of various methods used in this algorithm is made. To the end of the article annotation and classification of video recordings are described.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Zhao, R., William, G.: Narrowing the semantic gap-improved text-based web document retrieval using visual features. IEEE Trans. Multimedia 4(2), 189–200 (2002)
Tamizharasan, C., Chandrakala, S.: A survey on multimodal content based video retrieval. Int. J. Emerg. Technol. Adv. Eng. Chennai, 3 (2013)
Abhishek, G., Arnab, B.: Emotion recognition from audio and visual data using f-score based fusion. In: Proceedings of the 1st IKDD Conference on Data Sciences, pp. 1–10 (2014)
Nigay, L., Coutaz, J.: A design space for multimodal systems: concurrent processing and data fusion. In: Proceedings of the INTERACT 1993 and CHI 1993 Conference on Human Factors in Computing Systems. New York, pp. 172–178 (1993)
Truong, B.T., Venkatesh, S.: Video abstraction, a systematic review and classification. ACM Trans. Multimedia Comput. Commun. 3 (2007)
Zhang, X.-D., Liu, T.-Y., Lo, K.-T., Feng, J.: Dynamic selection and effective compression of key frames for video abstraction. Pattern Recogn. Lett. 24, 1523–1532 (2003)
Sundaram, H., Chang, S.-F.: Video Scene segmentation using video and audio features. In: 2000 IEEE International Conference on Multimedia Expo, vol. 2, pp. 1145–1148 (2000)
Adcock, J., Girgensohn, A., Cooper, M., Liu, T., Wilcox, L., Rieffel, E.: Fxpal experiments for trecvid 2004. In: Proceedings of the TREC Video Retrieval Evaluation (TRECVID), pp. 70–81 (2004)
Haase, B., Davis, M.E.: Media streams: representing video for retrieval and repurposing. Technical report (1995)
Zhang, T., Xu, C., Zhu, G., Liu, S., Lu, H.: A generic framework for video annotation via semi-supervised learning. IEEE Trans. Multimedia. 1206–1219 (2012)
Alekseev, A.V.: Automatic coloring of grayscale images based on intelligent scene analysis. In: Alekseev, A.V., Rozaliev, V.L., Orlova, Y.A. (eds.) Pattern Recognition and Image Analysis (Advances in Mathematical Theory and Applications). 25(1), pp. 10–21. Pleiades Publishing, Moscow (2015)
Cootes, T., Taylor, C., Cooper, D.: Active shape models-their training and application. Comput. Vis. Image Underst. 61, 38–59 (1995)
Cuiping, Z., Guangda, S.: Human face recognition: a survey. J. Image Graph. 11, 103–111 (2000)
Rozaliev, V.L., Bobkov, A.S., Orlova, Y.A., Zaboleeva-Zotova, A.V., Dmitriev, A.S.: Detailed analysis of postures and gestures for the identification of human emotional reactions. World Appl. Sci. J. (WASJ) 24(24), 151–158 (2013)
Zaboleeva-Zotova, A.V., Orlova, Y.A., Rozaliev, V.L., Bobkov, A.S.: Emotional state recognition based on the motion and posture. In: Operations Research and Data Mining, ORADM 2012: the Workshop, 12–14 March 2012, Cancun Center for Continuous Education of the National Politechnic Institute (IPN). – Cancun, Cancun, Mexico, pp. 161–169 (2012) Eng
Gorodnichy, D.O.: Video-based framework for face recognition in video/D.O. Gorodnichy. In: Second Workshop on Face Processing in Video (FPiV 2005). In: Proceedings of Second Canadian Conference on Computer and Robot Vision CRV 2005, Victoria, BC, Canada, 9–11 May 2005, pp. 330–338
Guo, G.D., Dyer, C.R.: Learning from examples in the small sample case-face expression recognition. IEEE Trans. Syst. 35, 477–488 (2005)
Havran, C., et al.: Independent Component Analysis for face authentication. In: KES 2002 Proceedings - Knowledge-Based Intelligent Information and Engineering Systems, Crema, Italy, pp. 1207–1211 (2002)
Rozaliev, V.L., Orlova, Y.A.: Recognizing and analyzing emotional expressions in movements. In: IsaÃas, P., Spector, J.M., Ifenthaler, D., Sampson, D.G. (eds.) E-Learning Systems, Environments and Approaches, pp. 117–131. Springer, Cham (2015). doi:10.1007/978-3-319-05825-2_9
Alekseev, A.V., Orlova, Y.A., Rozaliev, V.L., Zaboleeva-Zotova, A.V.: Two-stage segmentation method for context-sensitive image analysis. In: Kravets, A., Shcherbakov, M., Kultsova, M., Iijima, T. (eds.) JCKBSE 2014. CCIS, vol. 466, pp. 331–340. Springer, Cham (2014). doi:10.1007/978-3-319-11854-3_28
Shinohara, Y., Otsu, N.: Facial expression recognition using fisher weight maps. In: Proeeedings of IEEE Conference on Automatic Face and Gesture Recognition, Korea-Seoul, pp. 499–504 (2004)
Sujun, Z.: Facial expression recognition algorithm based on active shape model and gabor wavelet. J. Henan Univ. (Nat. Sci.) 9, 40–45 (2010)
Viola, P., Jones, M.: Robust real time object detection. In: 8th IEEE International Conference on Computer Vision. Vancouver, pp. 151–155 (2001)
Xiaofeng, F.: Facial expression recognition based on multi-scale centralized binary pattern. Control Theory Appl. 6, 26–32 (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Rozaliev, V.L., Orlova, Y.A., Guschin, R.I., Verichev, V.V. (2017). General Approach to the Synthesis of Emotional Semantic Information from the Video. In: Kravets, A., Shcherbakov, M., Kultsova, M., Groumpos, P. (eds) Creativity in Intelligent Technologies and Data Science. CIT&DS 2017. Communications in Computer and Information Science, vol 754. Springer, Cham. https://doi.org/10.1007/978-3-319-65551-2_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-65551-2_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-65550-5
Online ISBN: 978-3-319-65551-2
eBook Packages: Computer ScienceComputer Science (R0)