Abstract
Stochastic configuration network (SCN) has great potential in developing fast learning model with sound generalization capability and can be easily extended to the distributed computing framework. This paper aims to develop a distributed regularized stochastic configuration network to solve the limitations of traditional centralized learning on the scalability and efficiency in computing and storage resources for massive datasets. The local models are constructed using a classical stochastic configuration network, and the global unified model is built by the alternating direction method of multipliers (ADMM). Elastic net regularization term combining the LASSO and ridge methods is added into loss function of the ADMM optimization to prevent the model from overfitting when the data has high-dimensional collinearity. Each layer of the local regularized SCN model of a node in the topology network is constructed incrementally; its input weights and biases are broadcast to all other nodes under the inequality constraints. Output weights and the Lagrange multipliers of each node are calculated alternately through the decomposition–coordination procedure of the ADMM optimization algorithm until it finally converges to a unified model. A comprehensive study on five benchmark datasets and the ball mill experimental data has been carried out to evaluate the proposed method. The experiment results show that the proposed distributed regularized stochastic configuration network has relative advantages in terms of accuracy and stability compared with the distributed random vector functional link network.
Similar content being viewed by others
References
Gupta D, Rani R (2019) A study of big data evolution and research challenges. J Inf Sci 45(3):322–340
Zhou L, Pan S, Wang J, Vasilakos AV (2017) Machine learning on big data: opportunities and challenges. Neurocomputing 237:350–361
Wu X, Zhu X, Wu G-Q, Ding W (2014) Data mining with big data. IEEE Trans Knowl Data Eng 26(1):97–107
Peteiro-Barral D, Guijarro-Berdiñas B (2013) A survey of methods for distributed machine learning. Prog Artif Intell 2(1):1–11
Galakatos A, Crotty A, Kraska T (2018) Distributed machine learning. Springer New York, New York, pp 1196–1201
Fang Y, Jia Q, Guo L, Wang G (2019) Efficient privacy-preserving machine learning in hierarchical distributed system. IEEE Trans Netw Sci Eng 6(4):599–612
Amiri MM, Gündüz D (2019) Computation scheduling for distributed machine learning with straggling workers. IEEE Trans Signal Process 67(24):6270–6284
Taylor G, Burmeister R, Xu Z, Singh B, Patel A, Goldstein T (2016) Training neural networks without gradients: a scalable admm approach. In: International conference on machine learning, pp 2722–2731
Scardapane S, Wang D, Panella M (2016) A decentralized training algorithm for echo state networks in distributed big data applications. Neural Netw 78:65–74
Boyd S, Parikh N, Chu E, Peleato B, Eckstein J et al (2011) Distributed optimization and statistical learning via the alternating direction method of multipliers. Found Trends Mach Learn 3(1):1–122
Glowinski R, Marrocco A (1975) On the solution of a class of non linear dirichlet problems by a penalty-duality method and finite elements of order one. In: Optimization techniques IFIP technical conference. Springer, pp 327–333
Gabay D, Mercier B (1976) A dual algorithm for the solution of nonlinear variational problems via finite element approximation. Comput Math Appl 2(1):17–40
Mahoney MW et al (2011) Randomized algorithms for matrices and data. Found Trends Mach Learn 3(2):123–224
Scardapane S, Wang D (2017) Randomness in neural networks: an overview. Wiley Interdiscip Rev Data Min Knowl Discov 7(2):e1200
Igelnik B, Pao Y-H (1995) Stochastic choice of basis functions in adaptive function approximation and the functional-link net. IEEE Trans Neural Netw 6(6):1320–1329
Scardapane S, Wang D, Panella M, Uncini A (2015) Distributed learning for random vector functional-link networks. Inf Sci 301:271–284
Li M, Wang D (2017) Insights into randomized algorithms for neural networks: practical issues and common pitfalls. Inf Sci 382:170–178
Wang D, Li M (2017) Stochastic configuration networks: fundamentals and algorithms. IEEE Trans Cybern 47(10):3466–3479
Huang C, Huang Q, Wang D (2019) Stochastic configuration networks based adaptive storage replica management for power big data processing. IEEE Trans Ind Inform 16(1):373–383
Wang D, Li M (2017) Robust stochastic configuration networks with kernel density estimation for uncertain data regression. Inf Sci 412:210–222
Hoerl AE, Kennard RW (2000) Ridge regression: biased estimation for nonorthogonal problems. Technometrics 42(1):80–86
Zou H, Hastie T (2005) Regularization and variable selection via the elastic net. J R Stat Soc Ser B (Stat Methodol) 67(2):301–320
Gilbert EN (1959) Random graphs. Ann Math Stat 30(4):1141–1144
Mierswa I, Morik K (2005) Automatic feature extraction for classifying audio data. Mach Learn 58(2–3):127–149
Acknowledgements
This work was supported by the National Key R&D Program of China under Grant 2018YFB1700200.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that there is no conflict of interest in the research.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhao, L., Zou, S., Huang, M. et al. Distributed regularized stochastic configuration networks via the elastic net. Neural Comput & Applic 33, 3281–3297 (2021). https://doi.org/10.1007/s00521-020-05178-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-020-05178-x