Skip to main content
Log in

Missing label imputation through inception-based semi-supervised ensemble learning

  • Original Article
  • Published:
Advances in Computational Intelligence Aims and scope Submit manuscript

Abstract

In classification tasks, unlabeled data bring the uncertainty in the learning process, which may result in the degradation of the performance. In this paper, we propose a novel semi-supervised inception neural network ensemble-based architecture to achieve missing label imputation. The main idea of the proposed architecture is to use smaller ensembles within a larger ensemble to involve diverse ways of missing label imputation and internal transformation of feature representation, towards enhancing the prediction accuracy. Following the process of imputing the missing labels of unlabeled data, the human-labeled data and the data with imputed labels are used together as a training set for the credible classifiers learning. Meanwhile, we discuss how this proposed approach is more effective as compared to the traditional ensemble learning approaches. Our proposed approach is evaluated on different well-known benchmark data sets, and the experimental results show the effectiveness of the proposed method. In addition, the approach is validated by statistical analysis using Wilcoxon signed rank test and the results indicate statistical significance of the performance improvement in comparison with other methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  • Abdelgayed TS, Morsi WG, Sidhu TS (2017) Fault detection and classification based on co-training of semisupervised machine learning. IEEE Trans Ind Electron 65(2):1595–1605

    Article  Google Scholar 

  • Cormen TH, Leiserson CE, Rivest RL, Stein C (2009) Introduction to algorithms. MIT Press

  • de Vries S, Thierens D (2021) A reliable ensemble based approach to semi-supervised learning. Knowl Based Syst 215:106738

    Article  Google Scholar 

  • Dong A, Chung F-L, Wang S (2016) Semi-supervised classification method through oversampling and common hidden space. Inf Sci 349:216–228

    Article  Google Scholar 

  • Dua D, Graff C (2017) UCI machine learning repository. http://archive.ics.uci.edu/ml

  • Goldman S, Zhou Y (2000) Enhancing supervised learning with unlabeled data. In: ICML, Citeseer, pp 327–334

  • Gui W, Yue W, Xie Y, Zhang H, Yang C (2018) A review of intelligent optimal manufacturing for aluminum reduction production. Acta Autom Sin 44(11):1957–1970

    Google Scholar 

  • Junior JRB, do Carmo Nicoletti M (2019) An iterative boosting-based ensemble for streaming data classification. Inf Fusion 45:66–78

  • Khan H, Wang X, Liu H (2021) Missing value imputation through shorter interval selection driven by fuzzy c-means clustering. Comput Electr Eng 93:107230

    Article  Google Scholar 

  • Li C, Xie Y, Chen X (2020) Semi-supervised ensemble classification method based on near neighbor and its application. Processes 8(4):415

    Article  Google Scholar 

  • Lin M, Chen Q, Yan S (2013) Network in network. arXiv preprint arXiv:1312.4400

  • Liu Z, Gao Z, Li X (2018) Co-training method based on margin sample addition. Chin J Sci Instrum 39(3):45–53

    Google Scholar 

  • Livieris IE, Kanavos A, Tampakas V, Pintelas P (2018) An ensemble SSL algorithm for efficient chest X-ray image classification. J Imaging 4(7):95

    Article  Google Scholar 

  • Naimi AI, Balzer LB (2018) Stacked generalization: an introduction to super learning. Eur J Epidemiol 33(5):459–464

    Article  Google Scholar 

  • Ng WW, Zhou X, Tian X, Wang X, Yeung DS (2018) Bagging-boosting-based semi-supervised multi-hashing with query-adaptive re-ranking. Neurocomputing 275:916–923

    Article  Google Scholar 

  • Oliver A, Odena A, Raffel C, Cubuk ED, Goodfellow IJ (2018) Realistic evaluation of deep semi-supervised learning algorithms, arXiv preprint arXiv:1804.09170

  • Prakash VJ, Nithya DL (2014) A survey on semi-supervised learning techniques, arXiv preprint arXiv:1402.4645

  • Qiao S, Shen W, Zhang Z, Wang B, Yuille A (2018) Deep co-training for semi-supervised image recognition. In: Proceedings of the European conference on computer vision (ECCV), pp 135–152

  • Ramasamy V, Sidharthan RK, Kannan R, Muralidharan G (2019) Optimal tuning of model predictive controller weights using genetic algorithm with interactive decision tree for industrial cement kiln process. Processes 7(12):938

    Article  Google Scholar 

  • Ren Y, Zhang L, Suganthan PN (2016) Ensemble classification and regression-recent developments, applications and future directions. IEEE Comput Intell Mag 11(1):41–53

    Article  Google Scholar 

  • Sagi O, Rokach L (2018) Ensemble learning: a survey, Wiley Interdisciplinary Reviews. Data Min Knowl Discov 8(4):e1249

    Google Scholar 

  • Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–9

  • Tanha J (2018) Mssboost: a new multiclass boosting to semi-supervised learning. Neurocomputing 314:251–266

    Article  Google Scholar 

  • Van Engelen JE, Hoos HH (2020) A survey on semi-supervised learning. Mach Learn 109(2):373–440

    Article  MathSciNet  Google Scholar 

  • Wang Y, Chen S (2013) Safety-aware semi-supervised classification. IEEE Trans Neural Netw Learn Syst 24(11):1763–1772

    Article  Google Scholar 

  • Wang Y, Li T (2018) Improving semi-supervised co-forest algorithm in evolving data streams. Appl Intell 48(10):3248–3262

    Article  Google Scholar 

  • Wu D, Luo X, Wang G, Shang M, Yuan Y, Yan H (2017) A highly accurate framework for self-labeled semisupervised classification in industrial applications. IEEE Trans Ind Inform 14(3):909–920

    Article  Google Scholar 

  • Yue W, Gui W, Chen X, Zeng Z, Xie Y (2019) Knowledge representation and reasoning using self-learning interval type-2 fuzzy petri nets and extended topsis. Int J Mach Learn Cybern 10(12):3499–3520

    Article  Google Scholar 

  • Zhang K, Lan L, Kwok JT, Vucetic S, Parvin B (2014) Scaling up graph-based semisupervised learning via prototype vector machines. IEEE Trans Neural Netw Learn Syst 26(3):444–457

    Article  MathSciNet  Google Scholar 

  • Zhou Z-H (2009) When semi-supervised learning meets ensemble learning. In: International workshop on multiple classifier systems. Springer, pp 529–538

  • Zhou Z-H, Li M (2005) Tri-training: exploiting unlabeled data using three classifiers. IEEE Trans Knowl Data Eng 17(11):1529–1541

    Article  Google Scholar 

  • Zhu X, Goldberg AB (2009) Introduction to semi-supervised learning. Synth Lect Artif Intell Mach Learn 3(1):1–130

    MATH  Google Scholar 

  • Zuo L, Li L, Chen C (2015) The graph based semi-supervised algorithm with l1-regularizer. Neurocomputing 149:966–974

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China, Guangdong province (No. 2018A 0303130026) and National Natural Science Foundation of China (Grants 61976141 and 61732011).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Han Liu.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Khan, H., Liu, H. & Liu, C. Missing label imputation through inception-based semi-supervised ensemble learning. Adv. in Comp. Int. 2, 10 (2022). https://doi.org/10.1007/s43674-021-00015-7

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s43674-021-00015-7

Keywords

Navigation