Abstract
Random vector functional link (RVFL) is a widely used powerful model for solving real-life problems in classification and regression. However, the RVFL is not able to reduce the impact of noisy data, despite its high generalization capability. This paper presents a new intuitionistic RVFL classifier (IFRVFLC) for binary classification with the goals of improving the overall classification capability of the RVFL network and increasing its classification efficiency on noisy data sets. In IFRVFLC, each training sample is associated with an intuitionistic fuzzy number which consists of membership or non-membership frames. The membership degree of a pattern considers the distance from the respective class centre. The degree of non-membership, on the other hand, is determined by the ratio of the heterogeneous point number to the total number of neighbouring points. To check the efficiency of the proposed IFRVFLC model, its classification performance is compared with the support vector machine (SVM), twin SVM, kernel ridge regression, extreme learning machine , intuitionistic fuzzy SVM, intuitionistic fuzzy twin SVM and RVFL networks. The obtained results show the usability of the proposed IFRVFLC model.
Similar content being viewed by others
References
Asteris PG, Mokos VG (2019) Concrete compressive strength using artificial neural networks. Neural Comput Appl 32(15):11807–11826
Atanasov KT (1986) Intuitionistic fuzzy sets. Fuzzy Set Syst 20:1
Atanassov K (2016) Intuitionistic fuzzy sets. Int J Bioautom 20:1
Atanassov KT (2003) Intuitionistic fuzzy sets: past, present and future. In: EUSFLAT Conf pp 12–19
Balasundaram S, Gupta D (2016) On optimization based extreme learning machine in primal for regression and classification by functional iterative method. Int J Mach Learn Cybern 7(5):707–728
Borah P, Gupta D (2021) Robust twin bounded support vector machines for outliers and imbalanced data. Appl Intell 51(8):5314–5343
Cao F, Ye H, Wang D (2015) A probabilistic learning algorithm for robust modeling using neural networks with random weights. Inf Sci 313:62–78
Chang X, Lin SB, Zhou DX (2017) Distributed semi-supervised learning with kernel ridge regression. J Mach Learn Res 18(1):1493–1514
Chen X, Xiao Y (2020) Geometric projection twin support vector machine for pattern classification. Multim Tools Appl 1–17.
Cichocki A, Unbehauen R, Swiniarski RW (1993) Neural networks for optimization and signal processing, vol 253. Wiley, New York
Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273–297
Dai W, Chen Q, Chu F, Ma X, Chai T (2017) Robust regularized random vector functional link network and its industrial application. IEEE Access 5:16162–16172
Dietterich TG (1998) Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput 10(7):1895–1923
Ding S, Zhang X, Yu J (2016) Twin support vector machines based on fruit fly optimization algorithm. Int J Mach Learn Cybern 7(2):193–203
Drucker H, Wu D, Vapnik VN (1999) Support vector machines for spam categorization. IEEE Trans Neural Netw 10(5):1048–1054
Dua D, Graff C (2019) UCI machine learning repository [http://archive.ics.uci.edu/ml]. University of California, School of Information and Computer Science, Irvine
Gupta D, Gupta U (2021) On robust asymmetric Lagrangian ν-twin support vector regression using pinball loss function. Appl Soft Comput 102:107099
Gupta U, Gupta D (2021) Regularized based implicit Lagrangian twin extreme learning machine in primal for pattern classification. Int J Mach Learn Cybern 12(5):1311–1342
Gupta U, Gupta D (2021) On regularization based twin support vector regression with Huber loss. Neural Process Lett 53:459–515
Gupta D, Hazarika BB, Berlin M (2020) Robust regularized extreme learning machine with asymmetric Huber loss function. Neural Comput Appl 32(16):12971–12998
Ha MH, Huang S, Wang C, Wang XL (2011) Intuitionistic fuzzy support vector machine. J Hebei Univ (Nat Sci Ed) 3:225–229
Ha M, Wang C, Chen J (2013) The support vector machine based on intuitionistic fuzzy number and kernel function. Soft Comput 17(4):635–641
Hazarika BB, Gupta D (2020) Modelling and forecasting of COVID-19 spread using wavelet-coupled random vector functional link networks. Appl Soft Comput 96:106626
Hazarika BB, Gupta D (2021a). Density-weighted support vector machines for binary class imbalance learning. Neural Comput Appl 33(9):4243–4261
Hazarika BB, Gupta D (2021b) Density-weighted twin support vector machines for binary class imbalance learning. Neural Process Lett 54:1091
Hazarika BB, Gupta D, Borah P (2021) An intuitionistic fuzzy kernel ridge regression classifier for binary classification. Appl Soft Comput 112:107816
Hu YH, Tompkins WJ, Urrusti JL, Afonso VX (1993) Applications of artificial neural networks for ECG signal detection and classification. J Electrocardiol 26:66–73
Huang GB, Chen L, Siew CK (2006) Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans Neural Netw 17(4):879–892
Huang GB, Zhu QY, Siew CK (2005) Extreme learning machine: a new learning scheme of feedforward neural networks. Neural Netw 2:985–990
Huang GB, Zhu QY, Siew CK (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501
Jayadeva KR, Chandra S (2007) Twin support vector machines for pattern classification. IEEE Trans Pattern Anal Mach Intell 29(5):905–910
Jayadeva KR, Chandra S (2008) Fuzzy twin support vector machines for pattern classification. Math Programm Game Theory Decis Making. https://doi.org/10.1142/9789812813220_0009
Katuwal R, Suganthan PN (2019) Stacked autoencoder based deep random vector functional link neural network for classification. Appl Soft Comput 85:105854
Kumar MA, Gopal M (2009) Least squares twin support vector machines for pattern classification. Expert Syst Appl 36(4):7535–7543
Lin CF, Wang SD (2002) Fuzzy support vector machines. IEEE Trans Neural Netw 13(2):464–471
Liu Q, He Q, Shi Z (2008) Extreme support vector machine classifier. Pacific-asia conference on knowledge discovery and data mining. Springer, Berlin, pp 222–233
Naik J, Satapathy P, Dash PK (2018) Short-term wind speed and wind power prediction using hybrid empirical mode decomposition and kernel ridge regression. Appl Soft Comput 70:1167–1188
Ngo L, Cha J, Han JH (2019) Deep neural network regression for automated retinal layer segmentation in optical coherence tomography images. IEEE Trans Image Process 29:303–312
Nasiri JA, Mir AM (2020) An enhanced KNN-based twin support vector machine with stable learning rules. Neural Comput Appl 32(16):12949–12969
Oong TH, Isa NAM (2011) Adaptive evolutionary artificial neural networks for pattern classification. IEEE Trans Neural Netw 22(11):1823–1836
Pao YH, Takefuji Y (1992) Functional-link net computing: theory, system architecture, and functionalities. Computer 25(5):76–79
Rakesh K, Suganthan PN (2017) An ensemble of kernel ridge regression for multi-class classification. Proc Comput Sci 108:375–383
Rezvani S, Wang X, Pourpanah F (2019) Intuitionistic fuzzy twin support vector machines. IEEE Trans Fuzzy Syst 27(11):2140–2151
Saunders C, Gammerman A, Vovk V (1998) Ridge regression learning algorithm in dual variables. https://doi.org/10.5555/645527.657464
Scardapane S, Wang D (2017) Randomness in neural networks: an overview. Wiley Interdiscip Rev Data Min Know Discov 7(2):e1200
Sun Y, Li B, Yuan Y, Bi X, Zhao X, Wang G (2019) Big graph classification frameworks based on extreme learning machine. Neurocomputing 330:317–327
Tomar D, Agarwal S (2015) Twin support vector machine: a review from 2007 to 2014. Egypt Inf J 16(1):55–69
Wang LP, Wan CR (2008) Comments on" the extreme learning machine. IEEE Trans Neural Netw 19(8):1494–1495
Wang Y, Luo C (2021) An intelligent quantitative trading system based on intuitionistic-GRU fuzzy neural networks. Appl Soft Comput 108:107471
Wilcoxon F (1992) Individual comparisons by ranking methods. In: Breakthroughs in statistics. Springer, New York, NY, pp 196–202
Wu XH, Zhao PW (2020) Predicting nuclear masses with the kernel ridge regression. Phys Rev C 101(5):051301
Xiao Y, Zijie Z (2020) Infrared image extraction algorithm based on adaptive growth immune field. Neural Process Lett 51:2575–2587
Yu X, Ye X, Zhang S (2022) Floating pollutant image target extraction algorithm based on immune extremum region. Digit Signal Process 123:103442
Zhang L, Suganthan PN (2016) A comprehensive evaluation of random vector functional link networks. Inf Sci 367:1094–1105
Zhang L, Suganthan PN (2017) Benchmarking ensemble classifiers with novel co-trained kernel ridge regression and random vector functional link ensembles [research frontier]. IEEE Comput Intell Mag 12(4):61–72
Zhang W, Wang X, You W, Chen J, Dai P, Zhang P (2019) RESLS: region and edge synergetic level set framework for image segmentation. IEEE Trans Image Process 29:57–71
Zhang Y, Duchi J, Wainwright M (2015) Divide and conquer kernel ridge regression: a distributed algorithm with minimax optimal rates. J Mach Learn Res 16(1):3299–3340
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
All authors have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Mishra, U., Gupta, D. & Hazarika, B.B. An Intuitionistic Fuzzy Random Vector Functional Link Classifier. Neural Process Lett 55, 4325–4346 (2023). https://doi.org/10.1007/s11063-022-11043-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11063-022-11043-w