Skip to main content

Pattern Recognition and Cluster Analysis

  • Chapter
  • First Online:
Introduction to Artificial Intelligence

Abstract

Let us begin with a terminological remark, which concerns the notion of a pattern. In pattern recognition and cluster analysis various objects, phenomena, processes, structures, etc. can be considered as patterns.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 49.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 99.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    In structural pattern recognition patterns are represented by structural representations, similarly to syntactic pattern recognition. However, their recognition is done with the help of pattern matching methods, not, as in the syntactic approach, by applying formal grammars and automata.

  2. 2.

    For example, patients can be considered as patterns and then pattern recognition can consist of classifying them into one of several disease entities.

  3. 3.

    For example in the area of Business Intelligence we can try to group customers on the basis of their features such as the date of their last purchase, the total value of their purchases for the last two months, etc. into categories which determine a sales strategy (e.g., cross-selling, additional free services/products).

  4. 4.

    A component represents some feature of the pattern.

  5. 5.

    If patterns are images, then noise filtering, smoothing/sharpening, enhancement, and restoration are typical preprocessing operations. Then, features such as edges, characteristic points, etc. are identified. Finally, image segmentation and object identification are performed.

  6. 6.

    This interesting phenomenon is discussed, e.g., in [235].

  7. 7.

    The most popular feature extraction methods include Principal Component Analysis (PCA), Independent Component Analysis, and Linear Discriminant Analysis. The issues related to feature extraction methods are out of the scope of Artificial Intelligence. Therefore, they are not discussed in the book. The reader can find a good introduction to this area in monographs cited at the end of this chapter.

  8. 8.

    This can be done, for example, with the help of the search methods introduced in Chap. 4.

  9. 9.

    Later we equate a pattern with its representation in the form of a feature vector.

  10. 10.

    In our “fish example” a reference pattern corresponds to a fish of the mean length and the mean weight in a given class.

  11. 11.

    In our fish example this means that an unknown fish corresponding to an unknown pattern in Fig. 10.1b is classified as a sprat (\(\omega ^{1}\)), because it resembles the “reference sprat” \(\mathbf {R}^{1}\) more than the “reference eel” \(\mathbf {R}^{2}\). (That is, it is nearer to the “reference sprat” in the feature space.).

  12. 12.

    Evelyn Fix—a professor of statistics of the University of California, Berkeley, a Ph.D. student and then a principal collaborator of the eminent Polish-American mathematician and statistician Jerzy Spława-Neyman, who introduced the notion of a confidence interval (also, the Neyman-Pearson lemma).

  13. 13.

    Joseph Lawson Hodges, Jr—an eminent statistician (Hodges-Lahmann estimator, Hodges’ estimator) at the University of California, Berkeley, a Ph.D. student of Jerzy Spława-Neyman.

  14. 14.

    Thomas M. Cover—a professor at Stanford University, an author of excellent papers concerning models based on statistics and information theory.

  15. 15.

    Peter E. Hart—a professor of the Stanford University, a computer scientists (a co-author of a heuristic search method \(A^{*}\) and the model based on the Hough transform).

  16. 16.

    Vladimir Naumovich Vapnik—a professor at the Institute of Control Sciences, Moscow from 1961 to 1990, then at AT&T Bell Labs and NEC Laboratories, Princeton. His work concerns mainly statistics and Artificial Intelligence (Vapnik–Chervonenkis theory).

  17. 17.

    This is why the method is called Support Vector Machines.

  18. 18.

    Richard O. Duda—a professor of electrical engineering at San Jose State University. His achievements concern pattern recognition. He defined the Hough transform. He is a co-author of the excellent monograph “Pattern Classification and Scene Analysis”.

  19. 19.

    The basic notions of probability theory are introduced in Appendices I.1, B.1 and I.2.

  20. 20.

    Thomas Bayes—an eminent English mathematician and a Presbyterian minister. The “father” of statistics.

  21. 21.

    We may know, for example, that there are four times more patterns belonging to the class \(\omega ^{1}\) than patterns belonging to the class \(\omega ^{2}\) in nature. Then, \(P(\omega ^{1}) = 4/5\) and \(P(\omega ^{2}) = 1/5\).

  22. 22.

    The height of the bar for an interval [ab] should be \(h = p/w\), where p is the number of elements of the learning set which belong to the given class and are in the interval [ab], and w is the number of all elements of the learning set which belong to the given class.

  23. 23.

    John Ross Quinlan—an Australian computer scientist, a researcher at the University of Sydney and the RAND Corporation. His research concerns machine learning, decision theory, and data exploration.

  24. 24.

    Various metrics are introduced in Appendix G.2.

  25. 25.

    Hugo Steinhaus—a Polish mathematician, a professor of Jan Kazimierz University in Lwów (now Lviv, Ukraine) and Wrocław University, a Ph.D. student of David Hilbert, a co-founder of the Lwów School of Mathematics (together with, among others, Stefan Banach and Stanisław Ulam). His work concerns functional analysis (Banach-Steinhaus theorem), geometry, and mathematical logic.

  26. 26.

    James B. MacQueen—a psychologist, a professor of statistics at the University of California, Los Angeles. His work concerns statistics, cluster analysis, and Markov processes.

  27. 27.

    Stephen Curtis Johnson—a researcher at Bell Labs and AT&T, then the president of USENIX. A mathematician and a computer scientist. He has developed cpp—a C language compiler, YACC—a UNIX generator of parsers, int—a C code analyzer, and a MATLAB compiler.

  28. 28.

    In this case we begin with clusters containing single patterns and we can end up with one big cluster containing all patterns of a sample set.

  29. 29.

    In this case we begin with one big cluster containing all patterns of a sample set and we can end up with clusters containing single patterns.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mariusz Flasiński .

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Flasiński, M. (2016). Pattern Recognition and Cluster Analysis. In: Introduction to Artificial Intelligence. Springer, Cham. https://doi.org/10.1007/978-3-319-40022-8_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-40022-8_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-40020-4

  • Online ISBN: 978-3-319-40022-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics