Advertisement

Towards Optimal Naive Bayes Nearest Neighbor

  • Régis Behmo
  • Paul Marcombes
  • Arnak Dalalyan
  • Véronique Prinet
Part of the Lecture Notes in Computer Science book series (LNCS, volume 6314)

Abstract

Naive Bayes Nearest Neighbor (NBNN) is a feature-based image classifier that achieves impressive degree of accuracy [1] by exploiting ‘Image-to-Class’ distances and by avoiding quantization of local image descriptors. It is based on the hypothesis that each local descriptor is drawn from a class-dependent probability measure. The density of the latter is estimated by the non-parametric kernel estimator, which is further simplified under the assumption that the normalization factor is class-independent. While leading to significant simplification, the assumption underlying the original NBNN is too restrictive and considerably degrades its generalization ability. The goal of this paper is to address this issue.

As we relax the incriminated assumption we are faced with a parameter selection problem that we solve by hinge-loss minimization. We also show that our modified formulation naturally generalizes to optimal combinations of feature types. Experiments conducted on several datasets show that the gain over the original NBNN may attain up to 20 percentage points. We also take advantage of the linearity of optimal NBNN to perform classification by detection through efficient sub-window search [2], with yet another performance gain. As a result, our classifier outperforms — in terms of misclassification error — methods based on support vector machine and bags of quantized features on some datasets.

Keywords

Support Vector Machine Linear Support Vector Machine Locality Sensitive Hashing Hinge Loss Local Image Descriptor 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Boiman, O., Shechtman, E., Irani, M.: In defense of nearest-neighbor based image classification. In: CVPR (2008)Google Scholar
  2. 2.
    Lampert, C., Blaschko, M., Hofmann, T.: Beyond sliding windows: Object localization by efficient subwindow search. In: CVPR (2008)Google Scholar
  3. 3.
    Marszałek, M., Schmid, C., Harzallah, H., van de Weijer, J.: Learning object representations for visual object class recognition. In: Visual Recognition Challange workshop (2007)Google Scholar
  4. 4.
    Lazebnik, S., Schmid, C., Ponce, J.: Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: CVPR (2006)Google Scholar
  5. 5.
    Zhang, H., Berg, A.C., Maire, M., Malik, J.: SVM-KNN: Discriminative nearest neighbor classification for visual category recognition. In: CVPR (2006)Google Scholar
  6. 6.
    Jégou, H., Douze, M., Schmid, C.: Improving bag-of-features for large scale image search. International Journal of Computer Vision 87, 316–336 (2010)CrossRefGoogle Scholar
  7. 7.
    Harzallah, H., Jurie, F., Schmid, C.: Combining efficient object localization and image classification. In: International Conference on Computer Vision (2009)Google Scholar
  8. 8.
    van de Sande, K.E.A., Gevers, T., Snoek, C.G.M.: Evaluating color descriptors for object and scene recognition. In: T-PAMI (2010)Google Scholar
  9. 9.
    Varma, M., Ray, D.: Learning the discriminative power-invariance trade-off. In: ICCV (2007)Google Scholar
  10. 10.
    Bosch, A., Zisserman, A., Munoz, X.: Representing shape with a spatial pyramid kernel. In: International Conference on Image and Video Retrieval, ICIVR (2007)Google Scholar
  11. 11.
    Bosch, A., Zisserman, A., Munoz, X.: Image classification using random forests and ferns. In: ICCV (2007)Google Scholar
  12. 12.
    Yuan, J., Liu, Z., Wu, Y.: Discriminative subvolume search for efficient action detection. In: CVPR (2009)Google Scholar
  13. 13.
    Stone, C.: Optimal uniform rate of convergence for nonparametric estimators of a density function or its derivatives. Recent advances in statistics (1983)Google Scholar
  14. 14.
    Marszałek, M., Schmid, C.: Accurate object localization with shape masks. In: CVPR (2007)Google Scholar
  15. 15.
    Fei-Fei, L., Fergus, R., Perona, P.: One-shot learning of object categories. T-PAMI 28, 594–611 (2006)Google Scholar
  16. 16.
    Fei-Fei, L., Perona, P.: A bayesian hierarchical model for learning natural scene categories. In: CVPR (2005)Google Scholar
  17. 17.
    Yianilos, P.N.: Data structures and algorithms for nearest neighbor search in general metric spaces. In: SODA: ACM-SIAM Symposium on Discrete Algorithms (1993)Google Scholar
  18. 18.
    Dong, W., Wang, Z., Josephson, W., Charikar, M., Li, K.: Modeling LSH for performance tuning. In: CIKM, pp. 669–678. ACM, New York (2008)CrossRefGoogle Scholar
  19. 19.
    Muja, M., Lowe, D.: Fast approximate nearest neighbors with automatic algorithm configuration. In: VISAPP (2009)Google Scholar
  20. 20.
    Jégou, H., Douze, M., Schmid, C.: Product quantization for nearest neighbor search. IEEE Transactions on Pattern Analysis & Machine Intelligence (to appear 2010)Google Scholar
  21. 21.
    Mutch, J., Lowe, D.G.: Object class recognition and localization using sparse features with limited receptive fields. Int. J. Comput. Vision 80, 45–57 (2008)CrossRefGoogle Scholar
  22. 22.
    Lowe, D.: Distinctive image features from scale-invariant keypoints. IJCV (2003)Google Scholar
  23. 23.
    Bosch, A., Zisserman, A., Munoz, X.: Scene classification via pLSA. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3954, pp. 517–530. Springer, Heidelberg (2006)CrossRefGoogle Scholar
  24. 24.
    Ling, H., Soatto, S.: Proximity distribution kernels for geometric context in category recognition. In: ICCV (2007)Google Scholar
  25. 25.
    Opelt, A., Pinz, A., Fussenegger, M., Auer, P.: Generic object recognition with boosting. PAMI 28 (2004/2006)Google Scholar
  26. 26.
    Moosmann, F., Triggs, B., Jurie, F.: Fast discriminative visual codebooks using randomized clustering forests. In: Neural Information Processing Systems, NIPS (2006)Google Scholar
  27. 27.
    Lee, Y., Lin, Y., Wahba, G.: Multicategory support vector machines, theory, and application to the classification of microarray data and satellite radiance data. JASA 99, 67–81 (2004)zbMATHMathSciNetGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2010

Authors and Affiliations

  • Régis Behmo
    • 1
  • Paul Marcombes
    • 1
    • 2
  • Arnak Dalalyan
    • 2
  • Véronique Prinet
    • 1
  1. 1.NLPR / LIAMA, Institute of AutomationChinese Academy of Sciences 
  2. 2.IMAGINE, LIGMUniversité Paris-Est 

Personalised recommendations