Advertisement

Pattern Recognition and Image Analysis

, Volume 28, Issue 4, pp 595–604 | Cite as

On Metric Correction and Conditionality of Raw Featureless Data in Machine Learning

  • S. D. Dvoenko
  • D. O. Pshenichny
Proceedings of the 6th International Workshop

Abstract

Recently, raw experimental data in machine learning often appear as direct comparisons between objects (featureless data). Different ways to evaluate difference or similarity of a pair of objects in image and data mining, image analysis, bioinformatics, etc., are usually used in practice. Nevertheless, such comparisons often are not distances or correlations (scalar products) like a correct function defined on a limited set of elements in machine learning. This problem is denoted as metric violations in ill-posed matrices. Therefore, it needs to recover violated metrics and provide optimal conditionality of corresponding matrices of pairwise comparisons for distances and similarities. This is the correct basis for using of modern machine learning algorithms.

Keywords

metrics similarity dissimilarity distance scalar product condition number determinant principal minor eigenvalue 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    R. A. Horn and C. R. Johnson, Matrix Analysis, 2nd ed. (Cambridge University Press, Cambridge, 2013).Google Scholar
  2. 2.
    R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification (Wiley, New York, 2001).zbMATHGoogle Scholar
  3. 3.
    S. D. Dvoenko, “Clustering and separating of a set of members in terms of mutual distances and similarities,” Trans. Mach. Learn. Data Mining 2 (2), 80–99 (2009).Google Scholar
  4. 4.
    S. D. Dvoenko and D. O. Pshenichny, “A recovering of violated metric in machine learning”, in Proc. 7th Symposium on Information and Communication Technology (SoICT’16) (ACM, New York, 2016), pp. 15–21. DOI: https://doi.org/10.1145/3011077.3011084Google Scholar
  5. 5.
    R. L. Bishop and R. J. Crittenden, Geometry of Manifolds (Academic Press, New York, 1964).zbMATHGoogle Scholar
  6. 6.
    E. Pekalska and R. P. W. Duin, The Dissimilarity Representation for Pattern Recognition. Foundations and Applications (World Scientific, Singapore, 2005).CrossRefzbMATHGoogle Scholar
  7. 7.
    I. Dubchak, I. Muchnik, C. Mayor, I. Dralyuk, and S.–H. Kim, “Recognition of a protein fold in the context of the SCOP classification”, Proteins: Struct. Funct. Genetics 35, 401–407 (1999).CrossRefGoogle Scholar
  8. 8.
    A. A. Goshtasby, Image Registration: Principles, Tools and Methods (Springer, London 2012). DOI: 10.1007/978–1–4471–2458–0CrossRefzbMATHGoogle Scholar
  9. 9.
    B. Schölkopf and A. J. Smola, Learning with Kernels (MIT Press, Cambridge, 2002).zbMATHGoogle Scholar
  10. 10.
    Z.–H. Zhou, F. Roli, and J. Kittler (Eds.), Multiple Classifier Systems, tMCS 2013, Lecture Notes in Computer Science (Springer, Berlin, Heidelberg, 2013), Vol. 7872.Google Scholar
  11. 11.
    R. D. Luce, Individual Choice Behaviour (Wiley, New York, 1959).zbMATHGoogle Scholar
  12. 12.
    D.S. Watkins, Fundamentals of Matrix Computations, 3rd ed. (Wiley, New York, 2010).zbMATHGoogle Scholar
  13. 13.
    A. N. Tikhonov and V. Y. Arsenin, Solutions of Ill–Posed Problems (Winston, New York, 1977).zbMATHGoogle Scholar
  14. 14.
    W. S. Torgerson, Theory and Methods of Scaling (Wiley, New York, 1958).Google Scholar
  15. 15.
    V. Mottl, S. Dvoenko, O. Seredin, C. Kulikowski, and I. Muchnik, “Featureless pattern recognition in an imaginary Hilbert space and its application to protein fold classification”, in Machine Learning and Data Mining in Pattern Recognition, Proc. MLDM 2001, Ed. by P. Perner, Lecture Notes in Computer Science (Springer, Berlin, Heidelberg, 2001), Vol. 2123, pp. 322–336.Google Scholar

Copyright information

© Pleiades Publishing, Ltd. 2018

Authors and Affiliations

  1. 1.Tula State UniversityTulaRussia

Personalised recommendations