Abstract
To deal with highly uncertain and noisy data, for example, biochemical laboratory examinations, a classifier is required to be able to classify an instance into all possible classes and each class is associated with a degree which shows how possible an instance is in that class. According to these degrees, we can discriminate the more possible classes from the less possible classes. The classifier or an expert can pick the most possible one to be the instance class. However, if their discrimination is not distinguishable, it is better that the classifier should not make any prediction, especially when there is incomplete or inadequate data. A fuzzy classifier is proposed to classify the data with noise and uncertainties. Instead of determining a single class for a given instance, fuzzy classification predicts the degree of possibility for every class.
Adenomatous polyps are widely accepted to be precancerous lesions and will degenerate into cancers ultimately. Therefore, it is important to generate a predictive method that can identify the patients who have obtained polyps and remove the lesions of them. Considering the uncertainties and noise in the biochemical laboratory examination data, fuzzy classification trees, which integrate decision tree techniques and fuzzy classifications, provide the efficient way to classify the data in order to generate the model for polyp screening.
Similar content being viewed by others
References
E. Sato, A. Ouchi, and T. Ishidate, "Polyps and diverticulosis of large bowel in autopsy, population of Akita prefecture, compared with Miyagi: High rate of colorectal cancer in Japan," Cancer, vol. 37, pp. 1316–1321, 1976.
J. Sauar, G. Hoff, and T. Hausken, "Colonoscopic screening ex-amination of relatives of patients with colorectal cancer," Scandi-navian Journal of Gastroenterology,vol. 27, pp. 667–672, 1992.
M. Shieh, I. Chiang, J. Wong, C. Huang, S. Huang, and C. Wang, "Prevalence of colorectal polyps in Tai-wan: 60cm-sigmoidoscopic findings," Biomedical Engineering-Application, Basis, Communication,vol. 7, no. 3, pp. 50–55, 1995.
A.R. Williams, B.A. Balasooriya, and D.W. Day, "Polyps and cancer of the large bowel: A necropsy study in Liverpool," Gut, vol. 23, pp. 835–842, 1982.
Z. Pawlak, "Rough Sets," Kluwer Academic, Dordrecht, 1991.
D. Heath, S. Kasif, and S. Salzberg, "Learning oblique deci-sion trees," in Proceedings of the Thirteenth International Joint Conference on Artificial Intelligence, Chambery, France, 1993, pp. 1002–1007.
S.K. Murthy, "On growing better decision trees from data," PhD dissertation, The Johns Hopkins University, Baltimore, Mary-land, 1995.
S.K. Murthy, S. Kasif, and S. Salzberg, "A system for induc-tion of oblique decision trees," Journal of Artificial Intelligence Research,vol. 2, pp. 1–32, 1994.
S.K. Murthy, S. Kasif, S. Salzberg, and R. Beigel, "OC1: Ran-domized induction of oblique decision trees," in Proceedings of the Eleventh National Conference on Artificial Intelligence, Washington, DC, 1993, pp. 322–327.
J.R. Quinlan, "Decision trees and decision making," IEEE Trans-actions on Systems, Man, and Cybernetics,vol. 20, pp. 339–346, 1990.
J.R. Quinlan, C4 5:Programs for Machine Learning, Morgan Kaufmann: Los Altos, CA, 1993.
P. Clark and T. Niblett, "The CN2induction algorithm," Machine Learning,vol. 3, pp. 261–283, 1989.
R. Rivest, "Learning decision lists," Machine Learning,vol. 2, pp. 229–246, 1987.
M. Sugeno and G.T. Kang, "Structure identification of fuzzy model," Fuzzy Sets and Systems,vol. 28, pp. 15–33, 1988.
J. Schuermann and W. Doster, "A decision theoretic approach to hierarchical classifier design," Pattern Recognition,vol. 17, no. 3, pp. 359–369, 1984.
I. Chiang and J. Hsu, "Integration of fuzzy classifiers with deci-sion trees," in Proceedings of Asian Fuzzy Systems Symposium, Kenting, Taiwan, 1996, pp. 65–78.
J.Y. Hsu and I. Chiang, "Fuzzy classification trees," in Proceed-ings of the Ninth International Symposium on Artificial Intelli-gence, Cancun, Mexico, 1996, pp. 431–438.
I. Chiang and J. Hsu, "Fuzzy classification trees for data anal-ysis," Fuzzy Sets and Systems,vol. 13, no. 1, pp. 87–99, 2002.
R. Duda and P. Hart, Pattern Classification and Scene Analysis, John Wiley and Sons: New York, 1973.
D. Rumelhart, G. Hinton, and R. Williams, "Learning internal representations by error prpagation," in Parallel Distributed Pro-cessing: Explorations in the Microstructure of Cognition, Vol. 1: Foundations, edited by D. Rumelhart and J. McClelland, MIT Press, Cambridge, MA, 1986, pp. 318–362.
M. Pazzani and D. Kibler, "The utility of knowledge in inductive learning," Machine Learning,vol. 9, no. 1, pp. 57–94, 1991.
J.R. Quinlan, "Learning logical definitions from relations," Ma-chine Learning,vol. 5, pp. 239–266, 1990.
J. Dougherty, R. Kohavi, and M. Sahami. "Supervised and unsu-pervised discretization of continuous features," in Proceedings of the Twelfth International Conference on Machine Learning, San Mateo, CA, 1995, pp. 194–202.
U.M. Fayyad and K.B. Irani, "On the handling of continuous-valued attributes in decision tree generation," Machine Learning, vol. 8, pp. 87–102, 1992.
W. Buntine, Myths and legends in learning classification rules, in Proceedings of the Eighth National Conference on Artificial Intelligence, Boston, MA, 1990, pp. 736–742
W. Buntine, "Learning classification trees," Statistics and Com-puting, vol. 2, pp. 63–73, 1992.
A.M. Mood, F.A. Graybill, and D.C. Boes, Introduction to the Theory of Statistics, 3rd edition, McGraw-Hill: NewYork, 1975.
J.R. Quinlan, "Probabilistic decision trees," in Proceedings of the Fourth International Workshop on Machine Learning, edited by P. Langley, Los Altos, CA, 1987.
L. Breiman, J. Friedman, R. Olshen, and C. Stone, Classification and Regression Trees, Chapman and Hall: London, 1984.
M.I. Jordan and R.A. Jacobs, "Supervised learning and divide-and-conquer: A statistical approach," in Proceedings of the Ninth International Conference on Machine Learning, Ab-erdeen, Scotland, 1992, pp. 159–166.
R.G. Casey and G. Nagy, "Decision tree design using a prob-abilistic model," IEEE Transactions on Information Theory, vol. 30, no. 1, pp. 93–99, 1984.
R. Rymon, "An SE-tree based characterization of the induction problem," in Proceedings of the Tenth International Conference on Machine Learning, 1993, pp. 268–275.
X. Boyen and L. Wehenkel, "Automatic induction of fuzzy de-cision trees and its application to power system security assess-ment," Fuzzy Sets and Systems,vol. 102, pp. 3–19, 1999.
K.J. Cios and L.M. Sztandera, "Continuous ID3 algorithm with fuzzy entropy measures," in Proceedings of the International Conference on Fuzzy Systems, San Diego, CA, 1992, pp. 469–476.
A. Su´ arez and J.F. Lutsko, "Globally optimal fuzzy decision trees for classification and regression," IEEE Trans. on Pattern Analysis and Machine Intelligence,vol. 21, no. 12, pp. 1297–1311, 1999.
Y. Yuan and M.J. Shaw, "Induction of fuzzy decision trees," Fuzzy Sets and Systems,vol. 69, pp. 125–139, 1995.
W. Pedrycz and Z.A. Sosnowski, "The design of decision trees in framework of granular data and their application to software quality models," Fuzzy Sets and Systems,vol. 123, pp. 271–290, 2001.
P.M. Murphy and M.J. Pazzani, "Exploring the decision forest: An empirical investigation of Occam's razor in decision tree induction," Journal of Artifical Intelligence Research,vol. 1, pp. 257–275, 1994.
C.Z. Janickow, "Fuzzy decision trees: Issues and methods," IEEE Trans. on System, Man, and Cybernetics B: Cybernetics, vol. 28, no, 1, pp. 1–14, 1998.
P.W. Baim, "A method for attribute selection in inductive learn-ing system," IEEE Transactions on Pattern Analysis and Ma-chine Intelligence,vol. 10, no. 9, pp. 888–896, 1988.
J. Mingers, "An empirical comparison of selection measures for decision-tree induction," Machine Learning,vol. 3, pp. 319–342, 1989.
J.R. Quinlan, "Induction of decision trees," Machine Learning, vol. 1, pp. 81–106, 1986.
C.E. Shannon, "A mathematical theory of communication," The Bell System Technical Journal,vol. 27, pp. 379–423, 623- 656, 1948.
A. De Luca and S. Termini, "A definition of a nonprobabilistic entropy in the setting of fuzzy sets theory," Information and Control,vol. 20, pp. 301–312, 1976.
J.C. Bezdek, Pattern Recognition with Fuzzy Objective Function Algorithms, Plenum Press: New York, 1981.
J. Yerushalmy, "Statistical problems in assessing methods of medical diagnosis, with special reference to X-ray tech-niques," Public Health Reports,vol. 62, pp. 1432–1449, 1947.
P. Murphy and D. Aha, "UCI repository of machine learning databases," Department of Information and Computer Science, University of California at Irvine, 1992.
A.P. White and W.Z. Liu, "Bias in information-based measures in decision tree induction," Machine Learning,vol. 15, pp. 321–329, 1994.
R.S. Michalski, "Learning flexible concepts: Fundamental ideas and method based on two-tiered representation," in Machine Learning: An Artificial Intelligence Approach,vol. III, edited by Y. Kodratoff and R.S. Michalski, Morgan Kaufmann, Los Altos, CA, 1990.
L. Rendell and H. Cho, "Empirical learning as a function of concept character," Machine Learning,vol. 5, no.3, pp. 267–298, 1990.
R. Kerber, "ChiMerge: Discretization of numeric attributes," in Proceedings of the Tenth National Conference on Artificial In-telligence, San Jose, CA, 1992, pp. 123–128.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Chiang, IJ., Shieh, MJ., Hsu, J.Yj. et al. Building a Medical Decision Support System for Colon Polyp Screening by Using Fuzzy Classification Trees. Applied Intelligence 22, 61–75 (2005). https://doi.org/10.1023/B:APIN.0000047384.85823.f6
Issue Date:
DOI: https://doi.org/10.1023/B:APIN.0000047384.85823.f6