Advertisement

An Improved Weighted ELM with Hierarchical Feature Representation for Imbalanced Biomedical Datasets

  • Liyuan Zhang
  • Jiashi Zhao
  • Huamin Yang
  • Zhengang Jiang
  • Weili Shi
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11061)

Abstract

In medical intelligent diagnosis, most of the real-world datasets have the class-imbalance problem and some strong correlation features. In this paper, a novel classification model with hierarchical feature representation is proposed to tackle small and imbalanced biomedicine datasets. The main idea of the proposed method is to integrate extreme learning machine-autoencoder (ELM-AE) into the weighted ELM (W-ELM) model. ELM-AE with norm optimization is utilized to extract more effective information from raw data, thereby forming a hierarchical and compact feature representation. Afterwards, random projections of learned feature results view as inputs of the W-ELM. An adaptive weighting scheme is designed to reduce the misclassified rate of the minority class by assigning a larger weight to minority samples. The classification performance of the proposed method is evaluated on two biomedical datasets from the UCI repository. The experimental results show that the proposed method cannot only effectively solve the class-imbalanced problem with small biomedical datasets, but also obtain a higher and more stable performance than other state-of-the-art classification methods.

Keywords

Medical intelligent diagnosis Class imbalance data Weighted ELM ELM-AE 

Notes

Acknowledgments

This work is supported by the Science & Technology Development Program of Jilin Province, China (Nos. 20150307030GX, 2015Y059 and 20160204048GX), and by the International Science and Technology Cooperation Program of China under Grant (No. 2015DFA11180), National Key Research and Development Program of China (No. 2017YFC0108303), and Science Foundation for Young Scholars of Changchun University of Science and Technology (No. XQNJJ-2016-08).

References

  1. 1.
    Krawczyk, B.: Learning from imbalanced data: open challenges and future directions. Prog. Artif. Intell. 5(4), 221–232 (2016)CrossRefGoogle Scholar
  2. 2.
    Rahman, M.M., Davis, D.N.: Addressing the class imbalance problem in medical datasets. Int. J. Mach. Learn. Comput. 3(2), 224–228 (2013)CrossRefGoogle Scholar
  3. 3.
    Krawczyk, B., Galar, M., Jelen, Ł., Herrera, F.: Evolutionary undersampling boosting for imbalanced classification of breast cancer malignancy. Appl. Soft Comput. 38, 714–726 (2016)CrossRefGoogle Scholar
  4. 4.
    Ali, S., Majid, A., Javed, S.G., et al.: Can-CSC-GBE: developing cost-sensitive classifier with gentleboost ensemble for breast cancer classification using protein amino acids and imbalanced data. Comput. Biol. Med. 73, 38–46 (2016)CrossRefGoogle Scholar
  5. 5.
    Ren, F., Cao, P., Li, W., et al.: Ensemble based adaptive over-sampling method for imbalanced data learning in computer aided detection of microaneurysm. Comput. Med. Imaging Graph. 55, 54–67 (2017)CrossRefGoogle Scholar
  6. 6.
    Yap, B.W., Rani, K.A., Rahman, H.A.A., Fong, S., Khairudin, Z., Abdullah, N.N.: An application of oversampling, undersampling, bagging and boosting in handling imbalanced datasets. In: Herawan, T., Deris, M.M., Abawajy, J. (eds.) Proceedings of the First International Conference on Advanced Data and Information Engineering (DaEng-2013). LNEE, vol. 285, pp. 13–22. Springer, Singapore (2014).  https://doi.org/10.1007/978-981-4585-18-7_2CrossRefGoogle Scholar
  7. 7.
    Gong, C.L., Gu, L.X.: A novel SMOTE-based classification approach to online data imbalance problem. Math. Probl. Eng., 1–14 (2016)Google Scholar
  8. 8.
    Zong, W., Huang, G.B., Chen, Y.: Weighted extreme learning machine for imbalance learning. Neurocomputing 101, 229–242 (2013)CrossRefGoogle Scholar
  9. 9.
    Sani, S., Massie, S., Wiratunga, N., Cooper, K.: Learning deep and shallow features for human activity recognition. In: Li, G., Ge, Y., Zhang, Z., Jin, Z., Blumenstein, M. (eds.) KSEM 2017. LNCS (LNAI), vol. 10412, pp. 469–482. Springer, Cham (2017).  https://doi.org/10.1007/978-3-319-63558-3_40CrossRefGoogle Scholar
  10. 10.
    Huang, G., Huang, G.B., Song, S., et al.: Trends in extreme learning machines: a review. Neural Netw. 61, 32–48 (2015)CrossRefGoogle Scholar
  11. 11.
    Tang, J.X., Deng, C.W., Huang, G.B.: Extreme learning machine for multilayer perceptron. IEEE Trans. Neural Netw. Learn. Syst. 27(4), 809–821 (2016)MathSciNetCrossRefGoogle Scholar
  12. 12.
    Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2(1), 183–202 (2009)MathSciNetCrossRefGoogle Scholar
  13. 13.
    UCI Machine Learning Repository. http://archive.ics.uci.edu/ml/datasets

Copyright information

© Springer Nature Switzerland AG 2018

Authors and Affiliations

  1. 1.School of Computer Science and TechnologyChangchun University of Science and TechnologyChangchunChina

Personalised recommendations