Kernel Based Approaches for Context Based Image Annotatıon

  • L. Swati NairEmail author
  • R. Manjusha
  • Latha Parameswaran
Conference paper
Part of the Lecture Notes in Computational Vision and Biomechanics book series (LNCVB, volume 28)


The Exploration of contextual information is very important for any automatic image annotation system. In this work a method based on kernels and keyword propagation technique is proposed. Automatic annotation with a set of keywords for each image is carried out by learning the image semantics. The similarity between the images is calculated by Hellinger’s kernel and Radial Bias Function kernel(RBF)kernel. The images are labelled with multiple keywords using contextual keyword propagation. The results of using the two kernels on the set of features extracted are analysed. The annotation results obtained were validated based on confusion matrix and were found to have a good accuracy. The main advantage of this method is that it can propagate multiple keywords and no definite structure for the annotation keywords has to be considered


Automatic image annotation Hellinger’s kernel RBF kernel Semantics Contextual keyword propagation Gabor features Haralick features 


  1. 1.
    Divvala, S.K., Hoiem, D., Hays, J.H., Efros, A.A., Hebert, M.: An empirical study of context in object detection. In: Proceedings in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1271–1278 (June 2009)Google Scholar
  2. 2.
    Li, J., Wang, J.: Automatic linguistic indexing of pictures by a statistical modelling approach. IEEE Trans. Pattern Anal. Mach. Intell. September 2003; 25(9), 1075–1088 (2003)Google Scholar
  3. 3.
    Li, L.J., Socher, R., Fei-Fei, L.: Towards total scene understanding: classification, annotation and segmentation in an automatic framework. In: IEEE Conference Computer Vision and Pattern Recognition, pp. 2036–2043 (2009)Google Scholar
  4. 4.
    Gao, Y., Fan, J., Xue, X., Jain, R.: Automatic image annotation by incorporating feature hierarchy and boosting to scale up SVM classifiers. In: ACM Multimedia Proceedings, pp. 901–910 (2006)Google Scholar
  5. 5.
    Okabe, T., Kondo, Y., Kitani, K.M., Sato, Y.: Recognizing multiple objects based on co-occurrence of categories. In: The Pacific–Rim Symposium on Image and Video Technology (PSIVT), pp. 497–508 (2009)Google Scholar
  6. 6.
    Jeon, J., Lavrenko, V., Manmatha, R.: Automatic image annotation and retrieval using cross-media relevance models. In: Special Interest Group on Information Retrieval (SIGIR) Proceedings, pp. 119–126 (2003)Google Scholar
  7. 7.
    Zhou, D., Weston, J., Gretton, A., Bousquet,O., Schölkopf, B.: Ranking on data manifolds. Adv. Neural Info. Process Sys. 16, 169–176 (2004)Google Scholar
  8. 8.
    Ladicky, L., Russell, C., Kohli, P., Torr, P.H.S.: Graph cut based inference with co-occurrence statistics. 11th European Conf. Comp. Vision. 6315(5), 239–253 (2010)Google Scholar
  9. 9.
    Yong Jae Lee: Kristen Grauman. Object-graphs for context-aware visual category discovery. IEEE Trans. Pattern Anal. Mach. Intell. 34(2), 346–358 (2012)CrossRefGoogle Scholar
  10. 10.
    Liu, J., Li, M., Ma, W., Liu, Q., Lu, H.: An adaptive graph model for automatic image annotation. In: Proceedings of ACM International Workshop on Multimedia Information Retrieval, pp. 61–70 (2006)Google Scholar
  11. 11.
    Belongie, S., Malik, J., Puzicha, J.: Shape matching and object recognition using shape contexts. IEEE Trans. Pattern Anal. Mach. Intell. 24(4), 509–522 (2002) CrossRefGoogle Scholar
  12. 12.
    Blei, D., Ng., A., Jordan M.: Latent dirichlet allocation. J. Mach. Learn. Res. 3(5), 993–1022 (2003)Google Scholar
  13. 13.
    Hofmann, T.: Unsupervised learning by probabilistic latent semantic analysis. Mach. Learn. 41(2), (2011)Google Scholar
  14. 14.
    Li, J., Zhang, H., Liao, Y.: Image annotation based on bag of visual words and optimized semi-supervised learning method. ICTACT J. Image Video Process. Spec. Iss. Video Process. Multimedia Sys. 5(01), 223–226 (2014)Google Scholar
  15. 15.
    Wu, L., Hoi, S.C.H., Yu, N.: Semantics-preserving bag-of-words models and applications. IEEE Trans. Image Process. 19(7), 1908–1920 (2010)MathSciNetCrossRefzbMATHGoogle Scholar
  16. 16.
    Lu, Z., Horace, H.S.Ip, Peng, Y.: Contextual Kernel and spectral methods for learning the semantics of images. IEEE Trans. Image Process. 20(6), 1739–1750 (2011)Google Scholar
  17. 17.
    Lavrenko, V., Manmatha, R., Jeon, J.: A model for learning the semantics of pictures. Adv. Neural Info. Process. Sys. 16, 553–560 (2004)Google Scholar
  18. 18.
    Feng, S., Manmatha, R., Lavrenko, V.: Multiple bernoulli relevance models for image and video annotation. Proc. IEEE Comp. Soc. Conf. Comp. Vision Pattern Recognit. 2, 1002–1009 (2004)Google Scholar
  19. 19.
    Naphade, M.R., Basu, S., Smith, J.R., Lin, C.Y., Tseng, B.: Statistical modeling approach to content-based video retrieval. IEEE Proc. 16th Int. Conf. 2, 953–956 (2002)Google Scholar
  20. 20.
    Hoffmann, M., Tuytelaars, T., Antanas, L., Frasconi, P., De Raedt, L.: 13 Proceedings of the 2013 IEEE Workshop on Applications of Computer Vision (WACV), pp. 133–139 (2013)Google Scholar

Copyright information

© Springer International Publishing AG  2018

Authors and Affiliations

  • L. Swati Nair
    • 1
    Email author
  • R. Manjusha
    • 1
  • Latha Parameswaran
    • 1
  1. 1.Department of Computer Science EngineeringAmrita School of Engineering, Amrita Vishwa VidyapeethamCoimbatoreIndia

Personalised recommendations