Skip to main content

Application of Bayesian Inference to Automatic Semantic Annotation of Videos

  • Chapter
Model-Based Reasoning in Science, Technology, and Medicine

Part of the book series: Studies in Computational Intelligence ((SCI,volume 64))

Summary. It is an important task to automatically extract semantic annotation of a video shot. This high level semantic information can improve the performance of video retrieval. In this paper, we propose a novel approach to annotate a new video shot automatically with a non-fixed number of concepts. The process is carried out by three steps. Firstly, the semantic importance degree (SID)is introduced and a simple method is proposed to extract the semantic candidate set (SCS) under considering SID of several concepts co-occurring in the same shot. Secondly, a semantic network is constructed using an improved K2 algorithm. Finally, the final annotation set is chosen by Bayesian inference. Experimental results show that the performance of automatically annotating a new video shot is significantly improved using our method, compared with classical classifiers such as Naïve Bayesian and K Nearest Neighbor.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Szummer, M., Picard, R.: Indoor-outdoor image classification. In Syeda-Mahmood, T., ed.: IEEE International Workshop in Content-Based Access to Image and Video Databases, Bombay, India, IEEE Computer Society Press, U.S. (1998) 42-51

    Chapter  Google Scholar 

  2. Vailaya, A., Jain, A., Zhang, H.: On image classification: city vs. landscape. In Li, C.S., ed.: IEEE Workshop on Content-Based Access of Image and Video Libraries, Santa Barbara, CA, USA., IEEE Computer Society Press, U.S. (1998) 3-8

    Chapter  Google Scholar 

  3. Barnard, K., P. Duygulu, D., Forsyth, N., de Freitas, Blei, D., Jordan, M.: Matching words and pictures. Journal of Machine Learning Research 3 (2003) 1107-1135

    Article  MATH  Google Scholar 

  4. Tseng, B.T., Lin, C.Y., Naphade, M.R., Natsev, A., Smith, J.: Normalized classifier fusion for semantic visual concept detection. In Torres, L., Garcia, N., eds.: International Conference on Image Processing, Barcelona, Spain, I.E.E.E. Press (2003) 535-538

    Google Scholar 

  5. Naphade, M.R.: A Probabilistic Framework For Mapping Audio-visual Fea-tures to High-Level Semantics in Terms of Concepts and Context. PhD thesis, University of Illinois at Urbana-Champaign (2001)

    Google Scholar 

  6. Jiménez, A.B.B.: Multimedia Knowledge:Discovery, Classification, Browsing, and Retrieval. PhD thesis, Columbia University (2005)

    Google Scholar 

  7. Jeon, J., Lavrenko, V., Manmatha, R.: Automatic image annotation and retrieval using cross-media relevance models. In Clarke, C., ed.: Proceedings of the 26th International ACM SIGIR Conference on Research and Development in Information Retrieval, Toronto, Canada, ACM Press (2003) 119-126

    Google Scholar 

  8. Lavrenko, V., Manmatha, R., Jeon, J.: A model for learning the semantics of pictures. In Saul, L.K., Weiss, Y., Bottou, L., eds.: Proceedings of the Seven-teenth Annual Conference on Neural Information Processing Systems, Vancou-ver, British Columbia, Canada, MIT Press (2004) 553-560

    Google Scholar 

  9. Feng, S.L., Manmatha, R., Lavrenko, V.: Multiple bernoulli relevance models for image and video annotation. In Davis, L., ed.: IEEE Conference on Computer Vision and Pattern Recognition, Washington DC, USA, IEEE Computer Society (2004) 1002-1009

    Google Scholar 

  10. Cooper, G., Herskovits, E.: A Bayesian method for the induction of probabilistic networks from data. Machine Learning 9(4) (1992) 309-347

    MATH  Google Scholar 

  11. Huang, C.: Inference in belief networks:a procedural guide. International Journal of Approximate Reasoning 11 (1994) 1-158

    Article  MATH  Google Scholar 

  12. W, F., De, X., Weixin, W.: A cluster algorithm of automatic key frame extraction based on adaptive threshold. Journal of Computer Research and Development 42(10) (2005) 1752-1757

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Wang, F., Xu, D., Xu, H., Lu, W., Wu, W. (2007). Application of Bayesian Inference to Automatic Semantic Annotation of Videos. In: Magnani, L., Li, P. (eds) Model-Based Reasoning in Science, Technology, and Medicine. Studies in Computational Intelligence, vol 64. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-71986-1_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-71986-1_26

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-71985-4

  • Online ISBN: 978-3-540-71986-1

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics