Skip to main content

Automatic Extraction of Affective Metadata from Videos Through Emotion Recognition Algorithms

  • Conference paper
  • First Online:
New Trends in Databases and Information Systems (ADBIS 2018)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 909))

Included in the following conference series:

Abstract

In recent years, the diffusion of social networks has made available large amounts of user-generated data containing people’s opinions and feelings. Such data are mostly unstructured and hence need to be enriched with a large set of metadata to allow for efficient data indexing and querying. In this work we focus on videos and we extend traditional metadata extraction techniques by taking into account emotional metadata, in order to enable data analysis from an affective perspective. To this purpose, we present a 3-phase methodology for the automatic extraction of emotional metadata from videos through facial expression recognition algorithms. We also propose a simple but versatile model for metadata that takes into account variations in emotions among video chunks. Experiments on a real-world video dataset show that our non-linear classifier reaches a remarkable 72% classification accuracy in facial expression recognition.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://cmusatyalab.github.io/openface/.

  2. 2.

    https://www.ibm.com/watson/.

  3. 3.

    https://azure.microsoft.com/it-it/services/cognitive-services/emotion/.

References

  1. Benitez-Garcia, G., Nakamura, T., Kaneko, M.: Multicultural facial expression recognition based on differences of western-caucasian and east-asian facial expressions of emotions. IEEE Trans. Inf. Syst. 5, 1317–1324 (2018)

    Article  Google Scholar 

  2. Cohen, I., Sebe, N., Garg, A., Chen, L.S., Huang, T.S.: Facial expression recognition from video sequences: temporal and static modeling. Comput. Vis. Image Underst. 91, 160–187 (2003)

    Article  Google Scholar 

  3. Diamantini, C., Mircoli, A., Potena, D., Storti, E.: Semantic disambiguation in a social information discovery system. In: Proceedings of the 2015 International Conference on Collaboration Technologies and Systems (CTS), pp. 326–333 (2015)

    Google Scholar 

  4. Ekman, P.: An argument for basic emotions. Cogn. Emotion 6, 169–200 (1992)

    Article  Google Scholar 

  5. Felbo, B., Mislove, A., Sgaard, A., Rahwan, I., Lehmann, S.: Using millions of Emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1616–1626 (2017)

    Google Scholar 

  6. Fourati, M., Jedidi, A., Gargouri, F.: Generic descriptions for movie document: an experimental study. In: Proceedings of IEEE/ACS International Conference on Computer Systems and Applications, AICCSA 2017, pp. 766–773, October 2018

    Google Scholar 

  7. Huang, J., Yuan, C.: Weighted-PCANet for face recognition. In: Arik, S., Huang, T., Lai, W.K., Liu, Q. (eds.) ICONIP 2015. LNCS, vol. 9492, pp. 246–254. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-26561-2_30

    Chapter  Google Scholar 

  8. Mircoli, A., Cucchiarelli, A., Diamantini, C., Potena, D.: Automatic emotional text annotation using facial expression analysis. In: Proceedings of CEUR Workshop 1848, pp. 188–196 (2017)

    Google Scholar 

  9. Mo, S., Niu, J., Su, Y., Das, S.K.: A novel feature set for video emotion recognition. Neurocomputing 291, 11–20 (2018)

    Article  Google Scholar 

  10. Poria, S., Peng, H., Hussain, A., Howard, N., Cambria, E.: Ensemble application of convolutional neural networks and multiple kernel learning for multimodal sentiment analysis. Neurocomputing 261, 217–230 (2017)

    Article  Google Scholar 

  11. Pramerdofer, C., Kampel, M.: Facial expression recognition using convolutional neural networks: state of the art. arXiv preprint arXiv:1612.02903 (2016)

  12. Sailunaz, K., Dhaliwal, M., Rokne, J., Alhajj, R.: Emotion detection from text and speech: a survey. Soc. Netw. Anal. Min. 8(1) (2018)

    Google Scholar 

  13. Sariyanidi, E., Gunes, H., Cavallaro, A.: Automatic analysis of facial affect: a survey of registration, representation and recognition. IEEE Trans. Patt. Anal. Mach. Intell. 37, 1113–1133 (2015)

    Article  Google Scholar 

  14. Sikos, L.F., Powers, D.M.W.: Knowledge-driven video information retrieval with LOD: Lrom semi-structured to structured video metadata. In: Proceedings of the 2015 Workshop on Exploiting Semantic Annotations in Information Retrieval (ESAIR), pp. 35–37 (2015)

    Google Scholar 

  15. Soltani, M., Zarzour, H., Babahenini, M.C.: Facial emotion detection in massive open online courses. In: Rocha, Á., Adeli, H., Reis, L.P., Costanzo, S. (eds.) WorldCIST’18 2018. AISC, vol. 745, pp. 277–286. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-77703-0_28

    Chapter  Google Scholar 

  16. Sun, W., Zhao, H., Jin, Z.: A complementary facial representation extracting method based on deep learning. Neurocomputing 306, 246–259 (2018)

    Article  Google Scholar 

  17. Sun, Y., Sebe, N., Lew, M.S., Gevers, T.: Authentic emotion detection in real-time video. In: Sebe, N., Lew, M., Huang, T.S. (eds.) CVHCI 2004. LNCS, vol. 3058, pp. 94–104. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-24837-8_10

    Chapter  Google Scholar 

  18. Viola, P., Jones, M.: Robust real-time object detection. Int. J. Comput. Vis., 137–154 (2001)

    Google Scholar 

  19. Yu, Z., Zhang, C.: Image based static facial expression recognition with multiple deep network learning. In: Proceedings of the 2015 International Conference on Multimodal Interaction (ICMI), pp. 435–442 (2015)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Alex Mircoli .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Mircoli, A., Cimini, G. (2018). Automatic Extraction of Affective Metadata from Videos Through Emotion Recognition Algorithms. In: Benczúr, A., et al. New Trends in Databases and Information Systems. ADBIS 2018. Communications in Computer and Information Science, vol 909. Springer, Cham. https://doi.org/10.1007/978-3-030-00063-9_19

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-00063-9_19

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-00062-2

  • Online ISBN: 978-3-030-00063-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics