SIMBAD: Emergence of Pattern Similarity

  • Joachim M. BuhmannEmail author
Part of the Advances in Computer Vision and Pattern Recognition book series (ACVPR)


A theory of patterns analysis has to suggest criteria how patterns in data can be defined in a meaningful way and how they should be compared. Similarity-based Pattern Analysis and Recognition is expected to adhere to fundamental principles of the scientific process that are expressiveness of models and reproducibility of their inference. Patterns are assumed to be elements of a pattern space or hypothesis class and data provide “information” which of these patterns should be used to interpret the data. The mapping between data and patterns is constructed by an inference algorithm, in particular by a cost minimization process. Fluctuations in the data usually limit the precision that we can achieve to uniquely identify a single pattern as interpretation of the data. We advocate an information-theoretic perspective on pattern analysis to resolve this dilemma where the tradeoff between informativeness of statistical inference and their stability is mirrored in the information-theoretic optimum of high information rate and zero communication error. The inference algorithm is considered as a noisy channel which naturally limits the resolution of the pattern space given the uncertainty of the data.


Cost Function Noisy Channel Approximation Capacity Pattern Space Pattern Recognition Problem 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.



This work has been partially supported by the FP7 EU project SIMBAD and by the SNF project 200021_138117. JB acknowledges very stimulating discussions with A. Busetto, L. Busse, M.H. Chehreghani, M. Frank, M. Mihalák, V. Roth, R. Srámek, W. Szpankowski and P. Widmayer.


  1. 1.
    Alon, N., Ben-David, S., Cesa-Bianchi, N., Haussler, D.: Scale-sensitive dimensions, uniform convergence, and learnability. J. ACM 44(4), 615–631 (1997) MathSciNetCrossRefzbMATHGoogle Scholar
  2. 2.
    Buhmann, J.M.: Information theoretic model validation for clustering. In: International Symposium on Information Theory, Austin Texas. IEEE Press, New York (2010). Google Scholar
  3. 3.
    Buhmann, J.M., Kühnel, H.: Vector quantization with complexity costs. IEEE Trans. Inf. Theory 39(4), 1133–1145 (1993) CrossRefzbMATHGoogle Scholar
  4. 4.
    Buhmann, J.M., Chehreghani, M.H., Frank, M., Streich, A.P.: Information theoretic model selection for pattern analysis. In: Guyon, I., Dror, G., Lemaire, V., Taylor, G., Silver, D. (eds.) ICML 2011 Workshop on “Unsupervised and Transfer Learning”, Bellevue, Washington, vol. 27, pp. 51–65 (2012). Clearwater Beach, Florida, JMLR: W&CP 5 Google Scholar
  5. 5.
    Buhmann, J.M., Mihalák, M., Srámek, R., Widmayer, P.: Robust optimization in the presence of uncertainty. In: Inventions in Theoretical Computer Science 2013, Berkeley. ACM 2013, pp. 505–514 (2012). doi: 10.1145/2422436.2422491 Google Scholar
  6. 6.
    Busse, L.: Information in orderings (learning to order). Ph.D. thesis, # 20600, ETH Zurich, CH-8092 Zurich, Rämistrasse (2012) Google Scholar
  7. 7.
    Busse, L.M., Chehreghani, M.H., Buhmann, J.M.: The information content in sorting algorithms. In: International Symposium on Information Theory, pp. 2746–2750. IEEE Press, Cambridge (2012) Google Scholar
  8. 8.
    Chehreghani, M.H., Giovanni Busetto, A., Buhmann, J.M.: Information theoretic model validation for spectral clustering. In: AISTATS 2012, La Palma. J. Mach. Learn. Res. (W&CP), vol. 22, pp. 495–503 (2012) Google Scholar
  9. 9.
    Cover, T.M., Thomas, J.A.: Elements of Information Theory, 2nd edn. Wiley, New York (1991) CrossRefzbMATHGoogle Scholar
  10. 10.
    Csiczár, I., Körner, J.: Information Theory: Coding Theorems for Discrete Memoryless Systems. Academic Press, New York (1981) Google Scholar
  11. 11.
    Frank, M., Buhmann, J.M.: Selecting the rank of SVD by maximum approximation capacity. In: International Symposium on Information Theory, St. Petersburg, pp. 1036–1040. IEEE Press, New York (2011) Google Scholar
  12. 12.
    Grenander, U.: General Pattern Theory: a Mathematical Study of Regular Structures. Oxford University Press, Oxford (1994) zbMATHGoogle Scholar
  13. 13.
    Grenander, U., Miller, M.I.: Pattern Theory: from Representation to Inference. Oxford University Press, Oxford (2007) Google Scholar
  14. 14.
    Han, L., Rossi, L., Torsello, A., Wilson, R.C., Hancock, E.R.: Information theoretic prototype selection for unattributed graphs. In: Gimel’farb, G.L., Hancock, E.R., Imiya, A., Kuijper, A., Kudo, M., Omachi, S., Windeatt, T., Yamada, K. (eds.) Structural, Syntactic, and Statistical Pattern Recognition. Lecture Notes in Computer Science, vol. 7626, pp. 33–41. Springer, Berlin (2012) CrossRefGoogle Scholar
  15. 15.
    Hofmann, T., Buhmann, J.M.: Pairwise data clustering by deterministic annealing. IEEE Trans. Pattern Anal. Mach. Intell. 19(1), 1–14 (1997) CrossRefGoogle Scholar
  16. 16.
    Lingamneni, A., Krishna Muntimadugu, K., Enz, C., Karp, R.M., Palem, K.V., Piguet, C.: Algorithmic methodologies for ultra-efficient inexact architectures for sustaining technology scaling. In: Proceedings of the 9th Conference on Computing Frontiers, CF’12, pp. 3–12. ACM, New York (2012) Google Scholar
  17. 17.
    Rose, K., Gurewitz, E., Fox, G.: Vector quantization by deterministic annealing. IEEE Trans. Inf. Theory 38(4), 1249–1257 (1992) CrossRefzbMATHGoogle Scholar
  18. 18.
    Vapnik, V.N.: Estimation of Dependencies Based on Empirical Data. Springer, New York (1982) Google Scholar
  19. 19.
    Vapnik, V.N., Chervonenkis, A.Ya.: On the uniform convergence of relative frequencies of events to their probabilities. Theory Probab. Appl. 16, 264–280 (1971) CrossRefzbMATHGoogle Scholar

Copyright information

© Springer-Verlag London 2013

Authors and Affiliations

  1. 1.Swiss Federal Institute of Technology ZurichZurichSwitzerland

Personalised recommendations