Skip to main content

Spatially Aware Enhancement of BoVW-Based Image Retrieval Exploiting a Saliency Map

  • Conference paper
  • First Online:
Computer Analysis of Images and Patterns (CAIP 2015)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 9257))

Included in the following conference series:

  • 2699 Accesses

Abstract

The Bag-of-Visual-Words (BoVW) scheme is the most popular approach to similar image retrieval. The conventional BoVW models an image as a histogram of local features in which all of the features are uniformly weighed. Recently, researchers have focused on the similarity between the foregrounds of two images, because the foreground properly expresses the semantics of the image. Given an image, these methods approximate the likelihood that a local feature belongs to the foreground with its saliency value derived from the saliency map. The foreground histogram is then constructed in such a way that each local feature is accumulated to the histogram in proportion to its saliency value. Finally, the similarity between images is measured by comparing the foreground histograms of the images. However, the above strategy discounts local features with small saliency values, even if they lie on the genuine foreground. This paper proposes a new technique that does not disregard such local features and examines their spatial surroundings. In particular, a high weight is assigned to a local feature with a low saliency value when the spatial surrounding has a high saliency value, because this local feature is also likely to be a part of the foreground.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Sivic, J., Zisserman, A.: Video google: a text retrieval approach to object matching in videos. In: Proc. of 9th ICCV, pp. 1470–1477 (2003)

    Google Scholar 

  2. Pineda, G.F., Koga, H., Watanabe, T.: Scalable Object Discovery: A Hash-Based Approach to Clustering Co-occurring Visual Words. IEICE Transactions on Information and Systems E94–D(10), 2024–2035 (2011)

    Article  Google Scholar 

  3. Lazebnik, S., Schmid, C., Ponce, J.: Beyond bags of features: spatial pyramid matching for recognizing natural scene categories. In: Proc. of CVPR, pp. 2169–2178 (2006)

    Google Scholar 

  4. Viitaniemi, V., Laaksonen, J.: Spatial extensions to bag of visual words. In: Proc. of CIVR, Article No. 37 (2009)

    Google Scholar 

  5. Yuan, J., Wu, Y., Yang, M.: Discovery of collocation patterns: from visual words to visual phrases. In: Proc. of CVPR 2007

    Google Scholar 

  6. Zhang, Y., Jia, Z., Chen, T.: Image retrieval with geometry-preserving visual phrases. In: Proc. of CVPR, pp. 809–816 (2011)

    Google Scholar 

  7. Yang, Y., Newsam, S.: Spatial pyramid co-occurrence for image classification. In: Proc. of ICCV, pp. 1465–1472 (2011)

    Google Scholar 

  8. de Carvalho Soares, R., da Silva, I.R., Guliato, D.: Spatial locality weighting of features using saliency map with a bag-of-visual-words approach. In: Proc. of IEEE ICTAI, pp. 1070–1075 (2012)

    Google Scholar 

  9. Biagio, S., Bazzani, L., Cristani, M., Murino, V.: Weighted bag of visual words for object recognition. In: Proc. of ICIP, pp. 2734–2738 (2014)

    Google Scholar 

  10. Lowe, D.: Distinctive image features from scale invariant keypoints. IJCV, 91–110 (2004)

    Google Scholar 

  11. Cheng, M.: Global contrast based salient region detection. In: Proc. of CVPR, pp. 409–416 (2011)

    Google Scholar 

  12. Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on PAMI 20, 1254–1259 (1998)

    Article  Google Scholar 

  13. Harel, J., Koch, C., Perona, P.: Graph-based visual saliency. In: Advances in Neural Information Processing Systems, pp. 545–552. MIT Press (2007)

    Google Scholar 

  14. Achanta, R., Hemami, S., Estrada, F., Susstrunk, S.: Frequency-tuned salient region detection. In: Proc. of CVPR, pp. 1597–1604 (2009)

    Google Scholar 

  15. Otsu, N.: A Threshold Selection Method from Gray-Level Histograms. IEEE Transactions on Systems, Man and Cybernetics 9(1), 62–66 (1979)

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hisashi Koga .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Zou, Z., Koga, H. (2015). Spatially Aware Enhancement of BoVW-Based Image Retrieval Exploiting a Saliency Map. In: Azzopardi, G., Petkov, N. (eds) Computer Analysis of Images and Patterns. CAIP 2015. Lecture Notes in Computer Science(), vol 9257. Springer, Cham. https://doi.org/10.1007/978-3-319-23117-4_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-23117-4_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-23116-7

  • Online ISBN: 978-3-319-23117-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics