Skip to main content
Log in

A derived least square extreme learning machine

  • Methodologies and Application
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

Extreme learning machine (ELM) is a single hidden layer feedforward neural network and is proved to be a good machine learning tool. However, the singularity of the ELM activation function results in the poor generalization ability of the systems. This study proposes a least squares ELM with derivative characteristics (DLSELM). The activation function of the network consists of the original and derivative functions due to the introduction of derivative characteristics in the network. All weights and biases of the network are determined by a twice least squares method. Derivative characteristics increase the diversity of activation functions in the network. The regression accuracy of the network and the generalization ability of the system were greatly improved due to the weighs and biases of the DLSELM calculated by twice least methods. DLSELM is applied to different datasets for verifying their performance. Moreover, DLSELM possesses the best regression accuracy, stability, and generalization performance compared with the other networks.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Data availability

The datasets analyzed during this study are available in the University of California machine learning database,Toronto machine learning database, UCI machine learning library. The pumping test data of agricultural irrigation well are not publicly available.

References

  • Abualiga HL, Elaziz MA, Sumari P, Zong WG, Gandomi AH (2021) Reptile search algorithm (rsa): a nature-inspired meta-heuristic optimizer. Expert Syst Appl 191(11):116158

    Google Scholar 

  • Abualigah HL, Yousri D, Elaziz MA (2021) Matlab code of aquila optimizer: a novel meta-heuristic optimization algorithm. Computers Industrial Engineering

  • Cao W, Wang X, Zhong M et al (2018) A review on neural networks with random weights. Neurocomputing 275:278–287

    Article  Google Scholar 

  • Cheng S, Peng P, Lu F (2020) A lightweight ensemble spatiotemporal interpolation model for geospatial data. Int J Geograph Inf Sci 34(9):1849–1872

    Article  Google Scholar 

  • Fanjun L, Junfeu Q, Honggui H (2014) Incremental constructive extreme learning machine. Control Theo Appl 5:6

    MATH  Google Scholar 

  • Fei H, Qinghua L, Deshaung H (2010) An improved approximation approach incorporating particle swarm optimization and a priori information into neural networks. Neural Comput Appl 19(2):255–261

    Article  Google Scholar 

  • Guangbin H, Zhu Q, Siew CK (2003) Extreme learning machine: theory and applications. Neurocomputing 70(1):489–501

    Google Scholar 

  • Guoqiang L, Peifeng N, Xiaolong D, Xiangye Z (2014) Fast learning network: a novel artificial neural network with a fast learning speed. Neural Comput Appl 24:1683–1695

    Article  Google Scholar 

  • Hazir E, Ozcan T, Koc KH (2020) Prediction of adhesion strength using extreme learning machine and support vector regression optimized with genetic algorithm. Arab J Sci Eng 45:6985–7004

    Article  Google Scholar 

  • Hossain M, Mekhilef S, Danesh M, Olatomiwa L, Shamshirband S (2017) Application of extreme learning machine for short term output power forecasting of three grid-connected pv systems. J Clean Prod 167(20):395–405

    Article  Google Scholar 

  • Huang G, Zhu Q, Siew CK et al (2004) Extreme learning machine: a new learning scheme of feedforward neural networks. Neural Netw 2:985–990

    Google Scholar 

  • Huang G, Huang G, Song S, You K (2015) Trends in extreme learning machines: a review. Neural Netw 61:32–48

    Article  Google Scholar 

  • Jeffrey OA, Absalom E (2022) Dwarf mongoose optimization algorithm. Comput Methods Appl Mech Eng 391:114570

    Article  MathSciNet  Google Scholar 

  • Jie Z, Wendong X, Yanjiao L, Sen Z, Zhiqiang Z (2020) Multilayer probability extreme learning machine for device-free localization. Neurocomputing 396:383–393

    Article  Google Scholar 

  • Kumar J, Singh AK, Buyya R (2020) Ensemble learning based predictive framework for virtual machine resource request prediction. Neurocomputing 397:20–30

    Article  Google Scholar 

  • Li Z, Jie Z (2019) Learning from correlation with extreme learning machine. Int J Mach Learn Cybernet 10:3635–3645

    Article  Google Scholar 

  • Lippmann RP (1987) An introduction to computing with neural nets. IEEE Assp Mag 4(2):4–22

    Article  Google Scholar 

  • Liu Z, Li L, Tseng M, Lim MK (2020) Prediction short-term photovoltaic power using improved chicken swarm optimizer—Extreme learning machine model. J Clean Prod 248:119272

    Article  Google Scholar 

  • Li Y, Yang Z (2017) Application of eos-elm with binary jaya-based feature selection to real-time transient stability assessment using pmu data. IEEE Access, 1

  • Malik S, Thome N (2017) On a revisited moore-penrose inverse of a linear operator on hilbert spaces. Filomat 31(7):1927–1931

    Article  MathSciNet  Google Scholar 

  • Oyelade ON, Ezugwu AE (2021) Ebola optimization search algorithm (eosa): a new metaheuristic algorithm based on the propagation model of ebola virus disease

  • Peifeng N, Yunpeng M, Mengning L, Shanshan Y, Guoqiang L (2016) A kind of parameters self-adjusting extreme learning machine. Neural Process Lett 44(3):1–18

    Google Scholar 

  • Renjie H, Ratner E, Stewart D, Bjrk K, Lendasse A (2020) A modified lanczos algorithm for fast regularization of extreme learning machines. Neurocomputing 414:172–181

    Article  Google Scholar 

  • Sand Y, Pravin C (2003) A class+ 1 sigmoidal activation functions for ffanns. J Econ Dyn Control 28(1):183–187

    Article  MathSciNet  Google Scholar 

  • Weipeng C, Jinzhu G, Zhong M, Shubin C, Zhiguang S (2018) Fuzziness-based online sequential extreme learning machine for classification problems. Soft Comput 22(11):3487–3494

    Article  Google Scholar 

  • Weipeng C, Ming Zhong X, Zhiwu ZJ, Qiang W (2019) Online sequential extreme learning machine with dynamic forgetting factor. IEEE Access 7:179746–179757

    Article  Google Scholar 

  • Weipeng C, Zhong M, Xizhao W, Shubin C (2019) Improved bidirectional extreme learning machine based on enhanced random search. Memet Comput 11(1):19–26

    Article  Google Scholar 

  • Wenfeng H, Hong C, Xingsheng G (2019) A novel fault diagnosis technique for wind turbine gearbox. Appl Soft Comput 82:105556

    Article  Google Scholar 

  • Wenjing N, Zhongkai F, Ming Z et al (2019) Forecasting reservoir monthly runoff via ensemble empirical mode decomposition and extreme learning machine optimized by an improved gravitational search algorithm. Appl Soft Comput 82:105589

    Article  Google Scholar 

  • Wu Yu, Yongshan Z, Zhihua LXC, Yaoming C (2018) A multiobjective optimization-based sparse extreme learning machine algorithm. Neurocomputing 317:88–100

    Article  Google Scholar 

  • Xia L, Shaobo L, Jian F, Zongben X (2015) Is extreme learning machine feasible a theoretical assessment (part i). IEEE Trans Neural Netw Learn Syst 26(1):7–20

    Article  MathSciNet  Google Scholar 

  • Xiannong M (2019) Feature selection and enhanced krill herd algorithm for text document clustering. Comput Rev 60(8):318

    Google Scholar 

  • Yan S, Shujing Z, He B et al (2018) Gaussian derivative models and ensemble extreme learning machine for texture image classification. Neurocomputing 277:53–64

    Article  Google Scholar 

  • Yildirim H, Ozkale M (2019) The performance of elm based ridge regression via the regularization parameters. Expert Syst Appl 134:225–233

    Article  Google Scholar 

  • Yingchong L, Thian S, Andrew BJ (2019) Feature fusions for 2.5 d face recognition in random maxout extreme learning machine. Appl Soft Comput 75:358–372

    Article  Google Scholar 

  • Yunkun C, Xinghua S, Yimin W (2016) Convergence of rump’s method for computing the Moore-Penrose inverse. Czechoslovak Math J 66(3):859–879

  • Ze D, Ning M (2019) A novel nonlinear partial least square integrated with error-based extreme learning machine. IEEE Access 7:59903–59912

    Article  Google Scholar 

  • Zhang Yongshan W, Jia ZC, Zhihua C (2017) Instance cloned extreme learning machine. Pattern Recognit 68(68):52–65

    Article  Google Scholar 

Download references

Acknowledgements

The authors did not receive support from any organization for the submitted work. No funding was received to assist with the preparation of this manuscript. No funding was received for conducting this study. No funds, grants, or other support was received. The authors have no relevant financial or non-financial interests to disclose. The authors have no conflicts of interest to declare that are relevant to the content of this article. All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript.The authors have no financial or proprietary interests in any material discussed in this article. Authors are responsible for correctness of the statements provided in the manuscript.

Funding

This project was supported by the National Natural Science Foundation of China (Grant No. 12102273), National Key R& Program of China (Grant No. 2021YFB3900602, No. 2021YFB3900604).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Meiqi Wang.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Human and animal rights

This article does not contain any studies with human participants or animals performed by any of the authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hou, S., Wang, Y., Jia, S. et al. A derived least square extreme learning machine. Soft Comput 26, 11115–11127 (2022). https://doi.org/10.1007/s00500-022-07318-y

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00500-022-07318-y

Keywords

Navigation