Data Mining — Tools and Techniques

  • P R Limb
  • G J Meggs
Part of the BT Telecommunications Series book series (BTTS, volume 8)


With the advent of powerful desktop computers, organizations are recognizing that data can be more than that. By using the appropriate tools and techniques an experienced analyst can convert voluminous data into valuable information. This can be used to highlight the success (or failure) of marketing campaigns, display processes and be more responsive to customer needs. There are a wide variety of techniques that can be employed for data analysis and increasingly the term ‘data mining’ is used to describe these techniques.


Data Mining Hide Layer Classification Algorithm Multidimensional Space Data Mining Tool 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Anderberg M R: ‘Cluster analysis for applications’, Academic Press (1973).Google Scholar
  2. 2.
    Jain A K A and Dubes R C: ‘Algorithms for clustering data’, Prentice Hall (1988).Google Scholar
  3. 3.
    Taylor C, Michie D and Spiegelhalter D: ‘Machine learning, neural and statistical classification’, Ellis Horwood (1994).Google Scholar
  4. 4.
    Totton K A and Limb P R: ‘Electronic diagnosis using a multilayer perceptron’, BT Technol J, 10, No 3, pp 97–102 (1992).Google Scholar
  5. 5.
    Tattersall G D et al: ‘Feature extraction & visualization of decision support data’, BT Technol J, 10, No 3, pp 110–123 (1992).Google Scholar
  6. 6.
    Tattersall G D: ‘Investigation of data pre-processing for neural net classifiers’, University of East Anglia, DOC: 007, Issue 1 (Nov 1992).Google Scholar
  7. 7.
    Cox S J and Chichlowski K: ‘The application of standard statistical techniques to pattern classification’, University of East Anglia, DOC: 009, Issue 1 (March 1993).Google Scholar
  8. 8.
    Chatfield M G and Collins A J: ‘Introduction to multivariate analysis’, Chapman & Hall (1980).Google Scholar
  9. 9.
    McCulloch W C and Pitts W: ‘A logical calculus of the ideas imminent in nervous activity’, Bulletin of Mathematical Biophysics, 5, pp 115–133 (1943).Google Scholar
  10. 10.
    Rumelhart D E and McClelland J L: ‘Parallel distributed processing, Volume 1, Foundations’, The MIT Press (1987).Google Scholar
  11. 11.
    Caudhill M: ‘Neural networks primer’, Parts I-VIII, AI Expert (Dec 1987 to Aug 1989 ).Google Scholar
  12. 12.
    Lippman R P: ‘An introduction to computing with neural nets’, IEEE ASSP Magazine, 4–23 (Apr 1987).Google Scholar
  13. 13.
    Quinlan J R: ‘Cr.5! Programs for machine learning’, Morgan Kauffman, San Mateo, California (1993).Google Scholar
  14. 14.
    Hunt et al: ‘Experiments in induction’, Academic Press, New York (1966).Google Scholar
  15. 15.
    Quinlan J R: ‘Learning efficient classification procedures and their application to chess end games’ in Michalski R, Carbonnel J and Mitchell T (Eds): ‘Machine learning: an artificial intelligence approach’, Palo Alto! Tioga (1983).Google Scholar
  16. 16.
    Forsyth R and Rada R: ‘Machine intelligence: Applications in expert systems and information retrieval’, p 59–64 (1986).Google Scholar
  17. 17.
    Brieman L et al: ‘Classification and regression trees’, Wadsworth, Monterrey, CA (1984).Google Scholar
  18. 18.
    Boswell R A: ‘Manual for NewID version 2.0’, Technical Report TI/P2154/RAB54/, Turing Institute (Jan 1990).Google Scholar
  19. 19.
    Mueller W and Wysotzki F: ‘Automatic construction of decision trees for classification’, in Moser K and Schader M (Eds): ‘Annals of Operational Research 32’, J C Baltzer A G Science Publishers, Wijdenes, The Netherlands (1994).Google Scholar
  20. 20.
    Nadel B A: ‘Constraint satisfaction algorithms’, Computational Intelligence, 5, Part 4, pp 188–224 (Nov 1989).CrossRefGoogle Scholar
  21. 21.
    Winston P H: ‘Learning structural descriptions from examples’, in Winston P H (Ed): ‘The psychology of computer vision’, McGraw-Hill (1975).Google Scholar
  22. 22.
    Michalski R and Larson J: ‘Incremental generation of VLI hypotheses: The underlying methodology and the description of program QA11’, Urbana: University of Illinois at Urbana-Champaign, Dept of Computer Science Report (ISG 83 - 5 ) (1973).Google Scholar
  23. 23.
    Thornton C J: ‘Techniques in computational learning’, Chapman & Hall (1992).Google Scholar
  24. 24.
    Clark P and Niblett T: ‘Induction in noisy domains’, in Brakto I and Lavrac N (Eds): ‘Progress i machine learning’, Sigma Press (1987).Google Scholar
  25. 25.
    Michaelski R, Mozetic I, Hong J and Lavrac N: ‘The multi-purpose incremental learning system AQ15 and its testing application to three medical domains’, in Proc AAAAI-86, California, Morgan Kauffman (1986).Google Scholar
  26. 26.
    Fisher R A: ‘The use of multiple measurements in taxonomic problems’, Annals of Eugenics, 7, pp 179–177 (1936).Google Scholar
  27. 27.
    Michalski R, Mozetic I, Hong J and Lavrac N: ‘The multi-purpose incremental learning system AQ15 and its testing application to three medical domains’, Proc AAI-86, California, Morgan Kauffman (1986).Google Scholar
  28. 28.
    Clarke W R et al: ‘How non-normality affects the quadratic discriminant function’, Comm Statistics — Theory and Methods IT-16, pp 41–46 (1979).Google Scholar
  29. 29.
    Cox D R: ‘Some procedures associated with the logistic qualitative response curve’ in Dvid F N (Ed): ‘Research papers on statistics: Festschrift for J. Neyman’, pp 57–77, John Wiley, New York (1966).Google Scholar
  30. 30.
    Mahalanobis P C: ‘Historical note on the D3-statistic’, Sankhya 9, p 237 (1948).Google Scholar
  31. 31.
    Fix E and Hodges J L: ‘Discriminatory analysis, nonparametric estimation: consistency properties’ Report 4, Project 21-49-004, USAF School of Aviation Medicine, Randolph Field, Texas (1951).Google Scholar
  32. 32.
    Freidman J H: ‘SMART’s user guide’ Technical Report No 1, Laboratory of Computational Statistics, Department of Statistics, Stanford University (1984).Google Scholar
  33. 33.
    Holland J H: ‘Adaptation in natural and artifical systems’, Ann Arbor: The University of Michigan Press (1975).Google Scholar
  34. 34.
    Kelly J and Davis L: ‘Hybridizing the GA and K-nearest neighbors classification algorithm’, in Proceedings of the Fourth International Conference on Genetic Algorithms, California (Jul 1991).Google Scholar
  35. 35.
    Dodd N: ‘Optimization of network structure using genetic techniques’, AIENG-91: Applications of artificial intelligence in engineering 6: Proceedings of the sixth international conference. Oxford UK, pp 939–944 (July 1991).Google Scholar
  36. 36.
    Goldberg D: ‘Genetic algorithms in search, optimization and machine learning’, (1989).Google Scholar
  37. 37.
    Davis L: ‘Genetic algorithms and simulated annealing’, Pitman, London (1987).zbMATHGoogle Scholar
  38. 38.
    Stone M: ‘Cross-validatory choice and assessment of statistical predictions’, J Roy Statist Soc, 36, pp 111–33 (1974).zbMATHGoogle Scholar
  39. 39.
    Efron B: ‘Estimating the error rate of a prediction rule: improvements on cross- validation’, J Amer Stat Ass, 78, pp 316–331 (1983).MathSciNetzbMATHCrossRefGoogle Scholar
  40. 40.
    Fukunaga K: ‘Introduction to tatistical pattern recognition’, Academic Press (1972).Google Scholar
  41. 41.
    Massy W F: ‘On methods: discriminant analysis of audience characteristics’, J of Advertising Research, 5, pp 39–48 (1965).Google Scholar
  42. 42.
    Wilks S S: ‘Mathematical statistics’, John Wiley & Sons (1963).Google Scholar
  43. 43.
    Weiss S M and Julikowski C A: ‘Computer systems that learn’, Morgan Kauffman (1991).Google Scholar
  44. 44.
    Gordon A D: ‘Classification’, Chapman & Hall (1981).Google Scholar
  45. 45.
    Piatesky-Shapiro G and Frawley W J: ‘Knowledge Discovery in Databases’, AAAI Press (1991).Google Scholar

Copyright information

© British Telecommunications plc 1996

Authors and Affiliations

  • P R Limb
  • G J Meggs

There are no affiliations available

Personalised recommendations