Skip to main content

Automatic Event Detection in User-Generated Video Content: A Survey

  • Conference paper
  • First Online:
Advances in Electronics, Communication and Computing (ETAEERE 2020)

Abstract

The aim of event detection is to identify interested events in a user-generated content using multiple modalities automatically. However, it is a challenging task particularly when videos are captured in a restricted environment by nonprofessionals. Such videos suffer from poor quality, deprived lighting, blurring, complex camera motion chaotic background clutter, and obstructions. However, with the rise of social media, there is rising popularity of user-generated videos on the Web day-by-day. Each minute, 300 hours of user-generated video are uploaded on you tube due to which people find difficult to search the appropriate content among a large number of videos. Therefore, solutions to this problem are in great demands. In this paper, we study existing technologies for event detection in user-generated videos using multiple modalities. This paper provides key points about feature representations across different modalities, classification techniques.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 299.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 379.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 379.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Lavee, G., Rivlin, E., & Rudzsky, M. (2009). Understanding video events: A survey of methods for automatic interpretation of semantic occurrences in the video. IEEE Transactions on Systems, Man, and Cybernetics, 39(5), 489–504.

    Google Scholar 

  2. Ballan, L., Bertini, M., Del Bimbo, A., Seidenari, L., & Serra, G. (2011). Event detection and recognition for semantic annotation of video. Multimedia Tools and Applications.

    Google Scholar 

  3. Rajendra, S. P., & Keshaveni, N. (2014). A survey of automatic video summarization techniques. Indonesian Journal of Electrical Engineering and Computer Science, 3(1).

    Google Scholar 

  4. Soleymani, M., Garcia, D., Jou, B., Schuller, B., Chang, S.-F., & Pantic, M. (2017). A survey of multimodal sentiment analysis. Image and Vision Computing.

    Google Scholar 

  5. Brezeale, D., & Cook, D. J. (2007). Automatic video classification: A survey of the literature. IEEE Transactions on Systems, Man, and Cybernetics.

    Google Scholar 

  6. Atrey, P. K., Anwar Hossain, M., El Saddik, A., & Kankanhalli, M. S. (2010). Multimodal fusion for multimedia analysis: A survey. Multimedia Systems.

    Google Scholar 

  7. Hassan, E., Gopal, M., Chaudhury, S., & Garg, V. (2011) A hybrid framework for event detection using multi-modal features. In IEEE International Conference on Computer Vision Workshops.

    Google Scholar 

  8. van de Sande, K. E. A., Gevers, T., & Snoek, C. G. M. (2010). Evaluating color descriptors for object and scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(9).

    Google Scholar 

  9. Guo, J., Scott, D., Hopfgartner, F., & Gurrin, C. (2012). Detecting complex events in user-generated video using concept classifiers. In CBMI.

    Google Scholar 

  10. Bay, H., Ess, A., Tuytelaars, T., & Van Gool, L. (2008). SURF: Speeded up robust features. In Computer Vision and Image Understanding.

    Google Scholar 

  11. Jiang, Y.-G., Bhattacharya, S., Chang, S.-F., & Shah, M. (2013). High-level event recognition in unconstrained videos. International Journal of Multimedia Information Retrieval.

    Google Scholar 

  12. Pouyanfar, S., & Chen, S.-C. Automatic video event detection for imbalance data using enhanced ensemble deep learning.

    Google Scholar 

  13. Adams, W., Iyengar, G., Lin, C., Naphade, M., Neti, C., Nock, H., & Smith, J. (2003). Semantic indexing of multimedia content using visual, audio, and text cues. EURASIP Journal on Advances in Signal Processing.

    Google Scholar 

  14. Atrey, P. K., Kankanhalli, M. S., & Jain, R. Information assimilation framework for event detection in multimedia surveillance systems.

    Google Scholar 

  15. Oh, S., McCloskey, S., Kim, E., Vadat, A., Cannons, K. J., Hajimirsadeghi, H., et al. (2014). Multimedia event detection with multimodal feature fusion and temporal concept localization. Machine Vision and Applications.

    Google Scholar 

  16. Cricri1, F., Dabov, K., Curcio, I. D. D., Mate, S., & Gabbouj, M. (2014). Multimodal extraction of events and of information about the recording activity in user-generated videos. Multimedia Tools and Applications.

    Google Scholar 

  17. Koutras, P., Zlatinsi, A., & Maragos, P. (2018). Exploring CNN-based architectures for multimodal salient event detection in videos. In Proceedings of the 13th IEEE Image, Video, and Multidimensional Signal Processing (IVMSP) Workshop.

    Google Scholar 

  18. Güder, M., & Çiçekli, N. K. (2018). Multimodal video event recognition based on association rules and decision fusion. Multimedia Systems.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Sanjay Jain or Mihir Narayan Mohnaty .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Susmitha, A., Jain, S., Mohnaty, M.N. (2021). Automatic Event Detection in User-Generated Video Content: A Survey. In: Mallick, P.K., Bhoi, A.K., Chae, GS., Kalita, K. (eds) Advances in Electronics, Communication and Computing. ETAEERE 2020. Lecture Notes in Electrical Engineering, vol 709. Springer, Singapore. https://doi.org/10.1007/978-981-15-8752-8_29

Download citation

  • DOI: https://doi.org/10.1007/978-981-15-8752-8_29

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-15-8751-1

  • Online ISBN: 978-981-15-8752-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics