Advertisement

Journal of Medical Systems

, 40:239 | Cite as

Hypergraph Based Feature Selection Technique for Medical Diagnosis

  • Nivethitha Somu
  • M. R. Gauthama Raman
  • Kannan Kirthivasan
  • V. S. Shankar SriramEmail author
Systems-Level Quality Improvement
Part of the following topical collections:
  1. Personal Health Systems for Chronic Diseases Monitoring

Abstract

The impact of internet and information systems across various domains have resulted in substantial generation of multidimensional datasets. The use of data mining and knowledge discovery techniques to extract the original information contained in the multidimensional datasets play a significant role in the exploitation of complete benefit provided by them. The presence of large number of features in the high dimensional datasets incurs high computational cost in terms of computing power and time. Hence, feature selection technique has been commonly used to build robust machine learning models to select a subset of relevant features which projects the maximal information content of the original dataset. In this paper, a novel Rough Set based K – Helly feature selection technique (RSKHT) which hybridize Rough Set Theory (RST) and K – Helly property of hypergraph representation had been designed to identify the optimal feature subset or reduct for medical diagnostic applications. Experiments carried out using the medical datasets from the UCI repository proves the dominance of the RSKHT over other feature selection techniques with respect to the reduct size, classification accuracy and time complexity. The performance of the RSKHT had been validated using WEKA tool, which shows that RSKHT had been computationally attractive and flexible over massive datasets.

Keywords

Rough set theory (RST) Hypergraph K – Helly property High dimensional datasets Feature selection Medical diagnosis 

Notes

Acknowledgments

The first and the fourth author thank the Department of Science and Technology, India for INSPIRE Fellowship (Grant No: DST/INSPIRE Fellowship/2013/963) and Fund for Improvement of S&T Infrastructure in Universities and Higher Educational Institutions (SR/FST/ETI-349/2013) for their financial support; The second author thanks the Tata Consultancy Services for their financial support; The third author thanks the Department of Science and Technology - Fund for Improvement of S&T Infrastructure in Universities and Higher Educational Institutions Government of India (SR/FST/MSI-107/2015) for their financial support. We would like to express our gratitude towards the unknown potential reviewers who have agreed to review this paper and provided valuable suggestions to improve the quality of the paper.

References

  1. 1.
    Uzer, M.S., Yilmaz, N., and Inan, O., Feature selection method based on artificial bee colony algorithm and support vector machines for medical datasets classification. Sci. World J. 2013:1–10, 2013. doi: 10.1155/2013/419187.CrossRefGoogle Scholar
  2. 2.
    Li, H., Feature selection for high-risk pattern discovery in medical data. University of Cincinnati, Dissertation, 2012.Google Scholar
  3. 3.
    Fan, Y.J., and Chaovalitwongse, W.A., Optimizing feature selection to improve medical diagnosis. Ann. Oper. Res. 174:169–183, 2010. doi: 10.1007/s10479–008-0506-z.CrossRefGoogle Scholar
  4. 4.
    Pardalos, P.M., Boginski, V.L., and Alkis, V., Data mining in biomedicine. Springer science & business media: New York, 2008.Google Scholar
  5. 5.
    Hammer, P.L., and Bonates, T.O., Logical analysis of data—an overview: from combinatorial optimization to medical applications. Ann. Oper. Res. 148:203–225, 2006. doi: 10.1007/s10479–006-0075-y.CrossRefGoogle Scholar
  6. 6.
    Saastamoinen K, Ketola J (2006) Medical data classification using logical similarity based measures. IEEE Conference on Cybernetics and Intelligent Systems. 1–5. doi: 10.1109/ICCIS.2006.252362
  7. 7.
    Tsirogiannis, G.L., Frossyniotis, D., Stoitsis, J., Golemati, S., Stafylopatis, A., and Nikita, K.S., Classification of medical data with a robust multi-level combination scheme. IEEE International Joint Conference on Neural Networks. 3:2483–2487, 2004. doi: 10.1109/IJCNN.2004.1381020.Google Scholar
  8. 8.
    Huang, M.L., Hung, Y.H., and Chen, W.Y., Neural network classifier with entropy based feature selection on breast cancer diagnosis. J. Med. Syst. 34:865–873, 2010. doi: 10.1007/s10916–009-9301-x.CrossRefPubMedGoogle Scholar
  9. 9.
    Raymer, M.L., Doom, T.E., Kuhn, L.A., and Punch, W.F., Knowledge discovery in medical and biological datasets using a hybrid Bayes classifier/evolutionary algorithm. IEEE Trans. Syst. Man Cybern. 33:802–813, 2003. doi: 10.1109/TSMCB.2003.816922.CrossRefGoogle Scholar
  10. 10.
    Dash, M., and Liu, H., Feature selection for classification. Intelligent Data Analysis. 1:131–156, 1997.CrossRefGoogle Scholar
  11. 11.
    Blum, A.L., and Langley, P., Selection of relevant features and examples in machine learning. Artif. Intell. 97:245–271, 1997.CrossRefGoogle Scholar
  12. 12.
    Pattaraintakorn, P., Cercone, N., and Naruedomkul, K., Rule learning: ordinal prediction based on rough sets and soft-computing. Appl. Math. Lett. 19:1300–1307, 2006. doi: 10.1016/j.aml.2005.08.004.CrossRefGoogle Scholar
  13. 13.
    Pappu, V., Panagopoulos, O.P., Xanthopoulos, P., and Pardalos, P.M., Sparse proximal support vector machines for feature selection in high dimensional datasets. Expert Systems with Applications. 42:9183–9191, 2015.CrossRefGoogle Scholar
  14. 14.
    Cao, B., Shen, D., Sun, J.T., Yang, Q., and Chen, Z., Feature selection in a kernel space. Proceedings of the 24th international conference on. Mach. Learn. 121–128, 2007. doi: 10.1145/1273496.1273512.
  15. 15.
    Zhou, Q., Zhou, H., and Li, T., Cost-sensitive feature selection using random forest: selecting low-cost subsets of informative features. Knowl.-Based Syst. 95:1–11, 2015. doi: 10.1016/j.knosys.2015.11.010.CrossRefGoogle Scholar
  16. 16.
    Prasad M, Sowmya A, Koch I (2004) Efficient feature selection based on independent component analysis. Intelligent Sensors, Sensor Networks and Information Processing Conference. 427–432. doi: 10.1109/ISSNIP.2004.1417499
  17. 17.
    Lu, C., Zhu, Z., and Gu, X., An intelligent system for lung cancer diagnosis using a new genetic algorithm based feature selection method. J. Med. Syst. 38:1–9, 2014. doi: 10.1007/s10916–014-0097-y.CrossRefGoogle Scholar
  18. 18.
    Ozcift, A., and Gulten, A., A robust multi-class feature selection strategy based on rotation forest ensemble algorithm for diagnosis of erythemato-squamous diseases. J. Med. Syst. 36:941–949, 2012. doi: 10.1007/s10916–010–9558-0.CrossRefPubMedGoogle Scholar
  19. 19.
    Nalband, S., Sundar, A., Prince, A.A., and Agarwal, A., Feature selection and classification methodology for the detection of knee-joint disorders. Comput. Methods Prog. Biomed. 127:94–104, 2016. doi: 10.1016/j.cmpb.2016.01.020.CrossRefGoogle Scholar
  20. 20.
    Ozcift, A., Enhanced cancer recognition system based on random forests feature elimination algorithm. J. Med. Syst. 36:2577–2585, 2012. doi: 10.1007/s10916–011–9730-1.CrossRefPubMedGoogle Scholar
  21. 21.
    Hannah, I.H., Bagyamathi, M., and Azar, A.T., A novel hybrid feature selection method based on rough set and improved harmony search. Neural Comput. & Applic. 26:1859–1880, 2015. doi: 10.1007/s00521–015–1840-0.CrossRefGoogle Scholar
  22. 22.
    Hannah, I.H., Azar, A.T., and Jothi, G., Supervised hybrid feature selection based on PSO and rough sets for medical diagnosis. Comput. Methods Prog. Biomed. 113(1):175–185, 2014. doi: 10.1016/j.cmpb.2013.10.007.CrossRefGoogle Scholar
  23. 23.
    Gheyas, I.A., and Smith, L.S., Feature subset selection in large dimensionality domains. Pattern Recogn. 43:5–13, 2010. doi: 10.1016/j.patcog.2009.06.009.CrossRefGoogle Scholar
  24. 24.
    Mitra, P., Murthy, C.A., and Pal, S.K., Unsupervised feature selection using feature similarity. IEEE Trans. Pattern Anal. Mach. Intell. 24:301–312, 2002. doi: 10.1109/34.990133.CrossRefGoogle Scholar
  25. 25.
    Chandrashekar, G., and Ferat, S., A survey on feature selection methods. Comput. Electr. Eng. 40:16–28, 2014. doi: 10.1016/j.compeleceng.2013.11.024.CrossRefGoogle Scholar
  26. 26.
    Wroblewski, J., Finding minimal reducts using genetic algorithms. Second Annual Join Conference on Information Science. 186–189, 1995.Google Scholar
  27. 27.
    Jensen, R., and Shen, Q., Finding rough set reducts with ant colony optimization. UK workshop on computational intelligence. 1, 2003.Google Scholar
  28. 28.
    Sengupta, N., Sen, J., Sil, J., and Saha, M., Designing of on line intrusion detection system using rough set theory and Q-learning algorithm. Neurocomputing. 111:161–168, 2013. doi: 10.1016/j.neucom.2012.12.023.CrossRefGoogle Scholar
  29. 29.
    Swiniarski, R.W., and Skowron, A., Rough set methods in feature selection and recognition. Pattern Recogn. Lett. 24:833–849, 2003. doi: 10.1016/S0167–8655(02)00196–4.CrossRefGoogle Scholar
  30. 30.
    Peters J F, Ramanna S (2008) Feature selection: Near set approach. In: Mining complex data. Springer: Berlin Heidelberg, pp 57–71.Google Scholar
  31. 31.
    Jensen, R., and Shen, Q., Fuzzy-rough sets assisted attribute selection. IEEE Trans. Fuzzy Syst. 15:73–89, 2007. doi: 10.1109/TFUZZ.2006.889761.CrossRefGoogle Scholar
  32. 32.
    Zhang, Z., and Hancock, E.R., Hypergraph based information-theoretic feature selection. Pattern Recogn. Lett. 33:1991–1999, 2012. doi: 10.1016/j.patrec.2012.03.021.CrossRefGoogle Scholar
  33. 33.
    Hu X, Cercone N, Han J (1994) An attribute-oriented rough set approach for knowledge discovery in databases. In: Rough sets, fuzzy sets and knowledge discovery. Springer, pp 90–99.Google Scholar
  34. 34.
    Hu, K., Lu, Y., and Shi, C., Feature ranking in rough sets. AI Commun. 16:41–50, 2003.Google Scholar
  35. 35.
    Mac Parthaláin, N., and Shen, Q., Exploring the boundary region of tolerance rough sets for feature selection. Pattern Recogn. 42:655–667, 2009. doi: 10.1016/j.patcog.2008.08.029.CrossRefGoogle Scholar
  36. 36.
    Slezak, D., Approximate entropy reducts. Fundamenta informaticae. 53:365–390, 2002.Google Scholar
  37. 37.
    Yan, X.Z., Zuopeng, L., and Ru, W.S., Quick attribute reduction algorithm with complexity of max(O (| C|| U|), O (| C|(2)| U/C|)). Comput. J. 29:391–399, 2006. doi: 10.3321/j.issn:0254–4164.2006.03.006.Google Scholar
  38. 38.
    Eesa, A.S., Orman, Z., and Brifcani, A.M.A., A novel feature-selection approach based on the cuttlefish optimization algorithm for intrusion detection systems. Expert Systems with Applications. 42:2670–2679, 2015. doi: 10.1016/j.eswa.2014.11.009.CrossRefGoogle Scholar
  39. 39.
    Lassez J L, Rossi R, Sheel S, Mukkamala S (2008) Signature based intrusion detection using latent semantic analysis. IEEE International Joint Conference on Neural Networks. 1068–1074. doi: 10.1109/IJCNN.2008.4633931
  40. 40.
    Nguyen H, Franke K, Petrović S (2010) Improving effectiveness of intrusion detection by correlation feature selection. ARES’10 International Conference on Availability, Reliability, and Security. 17–24. doi: 10.1109/ARES.2010.70
  41. 41.
    Bakar, A.A., Sulaiman, M.N., Othman, M., and Selamat, M., H (200) Finding minimal reduct with binary integer programming in data mining. Proceedings TENCON. 3:141–146, 2000. doi: 10.1109/TENCON.2000.892239.Google Scholar
  42. 42.
    Wang, X., Yang, J., Teng, X., Xia, W., and Jensen, R., Feature selection based on rough sets and particle swarm optimization. Pattern Recogn. Lett. 28:459–471, 2007. doi: 10.1016/j.patrec.2006.09.003.CrossRefGoogle Scholar
  43. 43.
    Jiang, F., Sui, Y., and Zhou, L., A relative decision entropy-based feature selection approach. Pattern Recogn. 48:2151–2163, 2015. doi: 10.1016/j.patcog.2015.01.023.CrossRefGoogle Scholar
  44. 44.
    Inbarani, H.H., Bagyamathi, M., and Azar, A.T., A novel hybrid feature selection method based on rough set and improved harmony search. Neural Comput. & Applic. 26:1859–1880, 2015. doi: 10.1007/s00521–015–1840-0.CrossRefGoogle Scholar
  45. 45.
    Zhang M, Yao J T (2004) A rough sets based approach to feature selection. In, 2004. IEEE annual meeting of the fuzzy information processing NAFIPS'04. 1: 434–439. doi: 10.1109/NAFIPS.2004.1336322
  46. 46.
    Pawlak, Z., Rough sets. Int. J. Comput. Inform. Sci. 11:341–356, 1982. doi: 10.1007/BF01001956.CrossRefGoogle Scholar
  47. 47.
    Pawlak, Z., Rough sets: theoretical aspects of reasoning about data. Springer science and business media, B V, 2012.Google Scholar
  48. 48.
    Abraham, A., Falc, R., and Bello, R., Rough set theory: a true landmark in data analysis. Springer Verlag: Berlin Heidelberg, 2009.CrossRefGoogle Scholar
  49. 49.
    Deo, N., Graph theory with applications to engineering and computer science. Dover publications: New York, 2016.Google Scholar
  50. 50.
    Berge, C., Graphs and hypergraphs. North-Holland publishing company, Amsterdam, 1973.Google Scholar
  51. 51.
    Kannan, K., Kanna, B.R., and Aravindan, C., Root mean square filter for noisy images based on hyper graph model. Image Vis. Comput. 28:1329–1338, 2010. doi: 10.1016/j.imavis.2010.01.013.CrossRefGoogle Scholar
  52. 52.
    Bretto A, Gillibert L (2005) Hypergraph-based image representation. In: Graph-based representations in pattern recognition. Springer: Berlin Heidelberg, pp 1–11Google Scholar
  53. 53.
    Bretto, A., and Cherifi, H., Aboutajdine D (2002) Hypergraph imaging: an overview. Pattern Recogn. 35:651–658, 2002. doi: 10.1016/S0031–3203(01)00067-X.CrossRefGoogle Scholar
  54. 54.
    Anaraki, J.R., Eftekhari M (2011) Improving fuzzy-rough quick reduct for feature selection. 19th Iranian conference on electrical engineering. 1–6Google Scholar
  55. 55.
    UCI Repository (2016), http://archive.ics.uci.edu/ml/. Accessed 22 Jun 2016.
  56. 56.
    Witten, I.H., and Frank, E., Data Mining: practical machine learning tools and techniques with java implementations. Morgan Kaufmann, San Francisco, 2000.Google Scholar
  57. 57.
    Wang, G.Y., Yu, H., and Yang, D.C., Decision table reduction based on conditional information entropy. Chinese journal of computers - chinese edition. 25:759–766, 2002.Google Scholar

Copyright information

© Springer Science+Business Media New York 2016

Authors and Affiliations

  • Nivethitha Somu
    • 1
  • M. R. Gauthama Raman
    • 1
  • Kannan Kirthivasan
    • 2
  • V. S. Shankar Sriram
    • 1
    Email author
  1. 1.Centre for Information Super Highway (CISH), School of ComputingSASTRA UniversityThanjavurIndia
  2. 2.Department of MathematicsSASTRA UniversityThanjavurIndia

Personalised recommendations