Abstract
Humans use facial expressions as a tool to show emotional states. Facial expression recognition remains an interesting and challenging area of research in computer vision. An improved deep learning approach using a convolution neural network (CNN) is proposed in this paper to predict emotions by analysing facial expressions contained in an image. The model developed in this work consists of one CNN to analyse the primary emotion of the image as being happy or sad and a second CNN to predict the secondary emotion of the image. The proposed model was trained on the FER2013 and Japanese female facial expression (JAFFE) datasets with results suggesting its capability of predicting emotions from facial expressions better than existing state-of-the-art approaches.
Similar content being viewed by others
References
Arnheim, R. (1965). Art and visual perception: A psychology of the creative eye. Berkeley: University of California Press.
Borth, D., Ji, R., Chen, T., Breuel, T., & Chang, S. F. (2013). Large-scale visual sentiment ontology and detectors using adjective noun pairs. In Proceedings of the 21st ACM International Conference on Multimedia (pp. 223–232). ACM.
Boughrara, H., Chtourou, M., Amar, C. B., & Chen, L. (2016). Facial expression recognition based on a mlp neural network using constructive training algorithm. Multimedia Tools and Applications, 75(2), 709–731.
Busso, C., Deng, Z., Yildirim, S., Bulut, M., Lee, C. M., Kazemzadeh, A., Lee, S., Neumann, U., & Narayanan, S. (2004). Analysis of emotion recognition using facial expressions, speech and multimodal information. In Proceedings of the 6th International Conference on Multimodal Interfaces, (pp. 205–211). ACM.
Chen, T., Borth, D., Darrell, T., & Chang, S.F. (2014). Deepsentibank: Visual sentiment concept classification with deep convolutional neural networks. arXiv preprint arXiv:1410.8586.
Cheng, F., Yu, J., & Xiong, H. (2010). Facial expression recognition in Jaffe dataset based on Gaussian process classification. IEEE Transactions on Neural Networks, 21(10), 1685–1690.
Christou, N., & Kanojiya, N. (2019). Human facial expression recognition with convolution neural networks. In Third International Congress on Information and Communication Technology (pp. 539–545). Springer.
Colombo, C., Del Bimbo, A., & Pala, P. (1999). Semantics in visual information retrieval. IEEE Multimedia, 6(3), 38–53.
Dellandrea, E., Liu, N., & Chen, L. (2010). Classification of affective semantics in images based on discrete and dimensional models of emotions. In 2010 International Workshop on Content Based Multimedia Indexing (CBMI) (pp. 1–6). IEEE.
He, X., & Zhang, W. (2018). Emotion recognition by assisted learning with convolutional neural networks. Neurocomputing, 291, 187–194.
Ionescu, R.T., Popescu, M., & Grozea, C. (2013). Local learning to improve bag of visual words model for facial expression recognition. In Workshop on Challenges in Representation Learning, ICML.
Jain, D. K., Shamsolmoali, P., & Sehdev, P. (2019). Extended deep neural network for facial emotion recognition. Pattern Recognition Letters, 120, 69–74.
Jain, N., Kumar, S., Kumar, A., Shamsolmoali, P., & Zareapoor, M. (2018). Hybrid deep neural networks for face emotion recognition. Pattern Recognition Letters, 115, 101–106.
Joshi, D., Datta, R., Fedorovskaya, E., Luong, Q. T., Wang, J. Z., Li, J., et al. (2011). Aesthetics and emotions in images. IEEE Signal Processing Magazine, 28(5), 94–115.
Kahou, S. E., Bouthillier, X., Lamblin, P., Gulcehre, C., Michalski, V., Konda, K., et al. (2016). Emonets: Multimodal deep learning approaches for emotion recognition in video. Journal on Multimodal User Interfaces, 10(2), 99–111.
Kahou, S.E., Pal, C., Bouthillier, X., Froumenty, P., Gu¨l¸cehre, C¸., Memisevic, R., Vincent, P., Courville, A., Bengio, Y., & Ferrari, R.C., et al. (2013). Combining modality specific deep neural networks for emotion recognition in video. In Proceedings of the 15th ACM on International Conference on Multimodal Interaction (pp. 543–550). ACM.
Khorrami, P., Le Paine, T., Brady, K., Dagli, C., & Huang, T.S. (2016). How deep neural networks can improve emotion recognition on video data. In: 2016 IEEE International Conference on Image Processing (ICIP) (pp. 619–623). IEEE.
Lopes, A. T., de Aguiar, E., De Souza, A. F., & Oliveira-Santos, T. (2017). Facial expression recognition with convolutional neural networks: coping with few data and the training sample order. Pattern Recognition, 61, 610–628.
Lu, X., Suryanarayan, P., Adams Jr, R. B., Li, J., Newman, M. G., & Wang, J. Z. (2012). On shape and the computability of emotions. In Proceedings of the 20th ACM International Conference on Multimedia (pp. 229–238). ACM.
Lyons, M. J., Akamatsu, S., Kamachi, M., Gyoba, J., & Budynek, J. (1998). The Japanese female facial expression (Jaffe) database. In Proceedings of Third International Conference on Automatic Face and Gesture Recognition (pp. 14–16).
Machajdik, J., & Hanbury, A. (2010). Affective image classification using features inspired by psychology and art theory. In Proceedings of the 18th ACM International Conference on Multimedia (pp. 83–92). ACM.
Mavani, V., Raman, S., & Miyapuram, K. P. (2017). Facial expression recognition using visual saliency and deep learning. In: Proceedings of the IEEE International Conference on Computer Vision (pp. 2783–2788).
Minaee, S., & Abdolrashidi, A. (2019). Deep-emotion: Facial expression recognition using attentional convolutional network. arXiv preprint arXiv:1902.01019.
Nicolaou, M. A., Gunes, H., & Pantic, M. (2011). A multi-layer hybrid framework for dimensional emotion classification. In: Proceedings of the 19th ACM international conference on Multimedia (pp. 933–936). ACM.
Nie, D., Wang, X. W., Shi, L. C., & Lu, B. L. (2011). Eeg-based emotion recognition during watching movies. In: 2011 5th International IEEE/EMBS Conference on Neural Engineering (pp. 667–670). IEEE.
Shan, K., Guo, J., You, W., Lu, D., & Bie, R. (2017). Automatic facial expression recognition based on a deep convolutional-neural-network structure. In 2017 IEEE 15th International Conference on Software Engineering Research, Management and Applications (SERA) (pp. 123–128). IEEE.
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014). Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research, 15(1), 1929–1958.
Tang, Y. (2013). Challenges in representation learning: Facial expression recognition challenge implementation. Toronto: University of Toronto.
Tzirakis, P., Trigeorgis, G., Nicolaou, M. A., Schuller, B. W., & Zafeiriou, S. (2017). End-to-end multimodal emotion recognition using deep neural networks. IEEE Journal of Selected Topics in Signal Processing, 11(8), 1301–1309.
Valdez, P., & Mehrabian, A. (1994). Effects of color on emotions. Journal of experimental psychology: General, 123(4), 394.
Wei-ning, W., Ying-lin, Y., Sheng-ming, J. (2006). Image retrieval by emotional semantics: A study of emotional space and feature extraction. In 2006 IEEE International Conference on Systems, Man and Cybernetics, vol. 4 (pp. 3534–3539). IEEE.
Xu, C., Cetintas, S., Lee, K., & Li, L. (2014). Visual sentiment prediction with deep convolutional neural networks. arXiv preprint arXiv:1411.5731
Zhao, S., Gao, Y., Jiang, X., Yao, H., Chua, T.S., & Sun, X. (2014) Exploring principles-of-art features for image emotion recognition. In Proceedings of the 22nd ACM international conference on Multimedia (pp. 47–56). ACM.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
We hereby declare that this manuscript is an original work and is not under consideration for publication in any other journal. All authors have approved the submission of the manuscript the journal mentioned above.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Verma, G., Verma, H. Hybrid-Deep Learning Model for Emotion Recognition Using Facial Expressions. Rev Socionetwork Strat 14, 171–180 (2020). https://doi.org/10.1007/s12626-020-00061-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12626-020-00061-6