Skip to main content

Video Semantics Quality Assessment Using Deep Learning

  • Conference paper
  • First Online:
Intelligent Data Engineering and Automated Learning – IDEAL 2020 (IDEAL 2020)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 12490))

  • 1528 Accesses

Abstract

This work proposes a method to assess the quality of user-generated videos (UGVs) of specific social events. The method is based on matching the semantic information extracted from videos and the information obtained from text news of the same event. Deep learning techniques are used to detect objects in the video scenes. News articles are represented by a set of relevant terms automatically extracted from the news. This paper describes our method and an evaluation of it.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Chen, M., Chen, S., Shyu, M., Wickramaratna, K.: Semantic event detection via multimodal data mining. IEEE Sig. Process. Mag. 23(2), 38–46 (2006). https://doi.org/10.1109/MSP.2006.1621447

    Article  Google Scholar 

  2. Evans, M., Kerlin, L., Larner, O., Campbell, R.: Feels like being there: viewers describe the quality of experience of festival video using their own words. In: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems, pp. LBW029:1-LBW029:6. CHI EA 2018. ACM, New York (2018). https://doi.org/10.1145/3170427.3188507

  3. Jiang, H., Lu, Y., Xue, J.: Automatic soccer video event detection based on a deep neural network combined CNN and RNN. In: 2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI), pp. 490–494, November 2016. https://doi.org/10.1109/ICTAI.2016.0081

  4. Liu, L., et al.: Deep learning for generic object detection: a survey. arXiv abs/1809.02165 (2018)

    Google Scholar 

  5. Sun, L., Yamasaki, T., Aizawa, K.: Photo aesthetic quality estimation using visual complexity features. Multimedia Tools Appl. 77(5), 5189–5213 (2017). https://doi.org/10.1007/s11042-017-4424-4

    Article  Google Scholar 

  6. Pouyanfar, S., Chen, S.: Semantic event detection using ensemble deep learning. In: 2016 IEEE International Symposium on Multimedia (ISM), pp. 203–208, December 2016. https://doi.org/10.1109/ISM.2016.0048

  7. Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6517–6525 (2017)

    Google Scholar 

  8. Seshadrinathan, K., Soundararajan, R., Bovik, A.C., Cormack, L.K.: Study of subjective and objective quality assessment of video. IEEE Trans. Image Process. 19(6), 1427–1441 (2010). https://doi.org/10.1109/TIP.2010.2042111

    Article  MathSciNet  MATH  Google Scholar 

  9. Shahid, M., Khatibi, S., Tuemay, Y.: Popularity index through video semantic quality assessment. In: 2014 IEEE China Summit International Conference on Signal and Information Processing (ChinaSIP), pp. 344–348, July 2014. https://doi.org/10.1109/ChinaSIP.2014.6889261

  10. Shyu, M., Xie, Z., Chen, M., Chen, S.: Video semantic event/concept detection using a subspace-based multimedia data mining framework. IEEE Trans. Multimedia 10(2), 252–259 (2008). https://doi.org/10.1109/TMM.2007.911830

    Article  Google Scholar 

  11. Tapaswi, M.: Story Understanding through semantic analysis and automatic alignment of text and video. Ph.D. thesis, Karlsruhe Institute of Technology (2016)

    Google Scholar 

  12. Webster, A., Jones, C., Pinson, M., Voran, S., Wolf, S.: Objective video quality assessment system based on human perception. In: SPIE’s Symposium on Electronic Imaging: Science and Technology, vol. 1913 (1993). https://doi.org/10.1117/12.152700

  13. Yeh, H., Yang, C., Lee, M., Chen, C.: Video aesthetic quality assessment by temporal integration of photo- and motion-based features. IEEE Trans. Multimedia 15(8), 1944–1957 (2013). https://doi.org/10.1109/TMM.2013.2280250

    Article  Google Scholar 

Download references

Acknowledgments

This work is mainly supported by NOVA LINCS (UIDB/0 4516/2020) with the financial support of FCT - Fundação para a Ciência e a Tecnologia, through national funds. It was also partially supported by Cognitus project.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rui Jesus .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jesus, R., Silveira, B., Correia, N. (2020). Video Semantics Quality Assessment Using Deep Learning. In: Analide, C., Novais, P., Camacho, D., Yin, H. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2020. IDEAL 2020. Lecture Notes in Computer Science(), vol 12490. Springer, Cham. https://doi.org/10.1007/978-3-030-62365-4_16

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-62365-4_16

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-62364-7

  • Online ISBN: 978-3-030-62365-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics