Pinball-Huber boosted extreme learning machine regression: a multiobjective approach to accurate power load forecasting

Power load data frequently display outliers and an uneven distribution of noise. To tackle this issue, we present a forecasting model based on an improved extreme learning machine (ELM). Speciﬁcally, we introduce the novel Pinball-Huber robust loss function as the objective function in training. The loss function enhances the precision by assigning distinct penalties to errors based on their directions. We employ a genetic algorithm, combined with a swift nondominated sorting technique, for multiobjective optimization in the ELM-Pinball-Huber context. This method simultaneously reduces training errors while streamlining model structure. We practically apply the integrated model to forecast power load data in Taixing City, which is situated in the southern part of Jiangsu Province. The empirical ﬁndings conﬁrm the method’s effectiveness.


Introduction
Power load forecasting forms the foundation of power system scheduling and operation [1,2].Ensuring accurate power load forecasting is crucial for the stable operation and economic efficiency of a power system [3,4].Power load data often encompass anomalies and asymmetric noise because of factors such as climate fluctuations and changes in market demands [5].These data elements can impede model training, hence affecting the accuracy of power load forecasting and making the subject of mitigating asymmetric noise a significant concern in the field of power load forecasting.Despite significant advancements in power load forecasting methods, there is a need for further improvement to meet the current demands.Existing loss functions have made substantial progress in enhancing robustness and accuracy.However, their symmetric nature renders them inadequate for effectively addressing the deviations caused by outliers in power load datasets.Thus, there is a need for a comprehensive and robust loss function that can consider both the magnitude and direction of errors within the machine learning framework.Additionally, the number of hidden layer nodes in ELM significantly impacts the complexity of model training and final predictive accuracy.The lack of reliable benchmarks to balance the structural parameters and prediction accuracy also poses a challenge.Therefore, integrating novel and effective optimization techniques is essential for enhancing the ELM model and determining the optimal parameters for precise predictive modeling.
The present paper focuses on two key aspects: asymmetric loss function and multi-objective optimization.To address the aforementioned challenges, the paper contributes the following: (1) An asymmetric Pinball-Huber loss function for more effective data handling is developed.Because of its superior characteristics compared with other loss functions, it has been incorporated into the training objective of the ELM model.(2) The multiobjective optimization algorithm NSGA-II has been used to optimize two critical objectives of the ELM model: training error and output weight.By considering the input weights, hidden layer thresholds, and hidden node numbers as input parameters, the optimized ELM model can achieve minimized training errors and a more streamlined network structure.(3) The superiorty of proposed NSGA-II-ELM-Pinball-Huber model is validated by comparing it with benchmarks (LSTM, GRU, CNN-BiLSTM-Attention) using power load data from Taixing City.This validation has emphasized the effectiveness of the Pinball-Huber loss, highlighting the enhanced performance of the multiobjective optimization algorithm NSGA-II.
The rest of the current paper is organized as follows: Section 2 contains the literature review.Section 3 presents definitions of some terms.Section 4 gives the methodology (method).Section 5 goes over the results.Section 6 discusses the conclusions and future research areas.

Literature review
Over the past few decades, researchers have proposed numerous short-term load forecasting methods [6], which can be broadly categorized into physical, statistical, and intelligent methods [7].Physical methods establish the mathematical relationships between historical data and physical characteristics to achieve power load forecasting.Statistical methods perform mathematical statistics on historical data, establishing the correlations between load and time to make predictions [8].These models typically include linear regression (LR) [9], autoregressive integrated moving average (ARIMA) [10], gray models (GM) [11], and seasonal exponential smoothing(SEs) [7].However, these methods fail to capture the nonlinear characteristics present in load data.
Compared with traditional physical and statistical methods, intelligent methods exhibit greater potential in handling the nonlinear fluctuations and complex relationships within power load data, hence demonstrating higher accuracy in the field of power load forecasting [12,13].Intelligent methods such as artificial neural networks (ANN) [14], support vector regression (SVR) [15], and ELM [16] have found extensive applications in recent power forecasting studies.Among these, ANN is adept at modeling more intricate relationships between the power load and correlated variables compared with other methods, hence leading to its widespread usage in power load forecasting [2,17].ANN, which is akin to the structure of the human brain, can interpret vast amounts of data and transform it into actionable knowledge [18].ELM, an enhanced single-hidden-layer feedforward neural network, has been widely employed in forecasting tasks [19,20].
Unlike traditional artificial neural networks, ELM's input weights and biases in the hidden layer are randomly assigned.ELM derives hidden weights through the least squares method, eliminating the need for adjusting hidden layer weights through iterative backpropagation [21].As a result, the ELM model demonstrates faster learning and more pronounced generalization with minimal preset parameters [22].Numerous ELM-based predictive models have been proposed, showcasing their exceptional regression capabilities in forecasting.Ni et al. [23] employed an ensemble method using ELM and lower upper bound estimation (LUBE) for short-term power prediction.Han et al. [24] developed seasonal multimodels based on ELM by considering the seasonal distribution of power features.The effectiveness of the proposed methods was validated through a comparison with other approaches.Thus, compared with shallow learning systems, ELM exhibits higher efficiency, lower computational costs, and stronger generalization.
The loss function reflects the disparity between the predicted values and actual values during the optimization process, significantly impacting the learning model's generalization and accuracy [25].Chen et al. [26] utilized ELM enhanced with an L2-norm loss function for feature selection.Most neural network methods adopt mean squared error (MSE) or L2 loss function.Unfortunately, MSE loss function relies on Gaussian assumptions, making it sensitive to outliers and challenging to precisely evaluate nonlinear errors.Yang et al. [27] suggested employing the Huber loss function as the model's training objective.The Huber loss treats errors of different magnitudes differently.However, it lacks consideration for the direction of errors.Power load data are nonlinear and often exhibit various asymmetric noise distributions [5], necessitating the development of a new loss function that comprehensively considers both error magnitude and direction.
In conventional ANNs, including ELM, certain parameters are set randomly, leading to a degree of error and variability in the predictive outcomes.Artificial intelligence also exhibits drawbacks such as slow convergence, susceptibility to local optima, and overfitting [11,28].Hence, several intelligent optimization algorithms have been proposed to alleviate these limitations.Optimization algorithms applied to machine learning algorithms have further improved their regression capabilities to some extent [22].For instance, Niu et al.
[29] utilized a cooperative search algorithm that can explore the optimal hyperparameters of support vector machines (SVM), using this algorithm to predict electricity consumption in four Chinese provinces.Niu et al.
[29] optimized BPNN parameters using a genetic algorithm (GA).Shang et al. [30] established a prediction model combining least squares support vector machines (LSSVM) with generalized regression neural networks and optimized the weight coefficients by using the whale optimization algorithm (WOA).Xie et al. [31] proposed a short-term power load forecasting method combining Elman neural network (ENN) and particle swarm optimization (PSO).Differing from traditional random initialization, PSO was employed to search for the optimal learning rate for ENN.Addressing the issue of model parameter determination, arithmetic optimization algorithms (AOA) [32], gene expression programming (GEP) [33], and chimpanzee optimization algorithm (ChOA) [34], among others, have been utilized.Many studies on power load forecasting solely employed single-objective algorithms to optimize a criterion.However, in practical applications, meeting multiple constraints is often necessary [7,35].
The present paper introduces a novel power load forecasting model to address the aforementioned issues.Named NSGA-II-ELM-Pinball-Huber, this model is based on an enhanced Pinball-Huber loss function and multiobjective optimization algorithm NSGA-II.To effectively handle errors and anomalies in power load data, we introduced an asymmetric and robust Pinball-Huber loss function.Within the ELM framework, this loss function is employed as the objective, and the iteratively reweighted least squares (IRLS) method is utilized to determine the output weight vector.The present paper conducts global multiobjective optimization of the ELM model by employing the NSGA-II algorithm to simultaneously optimize training errors and output weights.The experimental results demonstrate that the proposed load forecasting model significantly enhances predictive performanc 3 The preliminaries

Regression loss function
Within various enhanced algorithms, the role of the loss function is to assess the merits and drawbacks of the improved model by computing its minimum value within the improved function.Yet during practical application, because of factors such as the loss function's objective, the nature of the application, data attributes, and the desired level of confidence in the forecasted values, a single loss function cannot be universally applied to all model experiments.Thus, a range of loss functions is required to be explored to optimize the treatment of target-type data and achieve optimal evaluation results [36].

L2-norm loss
L2-norm loss is a smooth function that is derivable in the whole domain and simplifies the calculation.When the error increases, the error is squared because of L2-norm loss, so that the error obtained is amplified.The L2-norm loss function can be described as follows: where r = y − ŷ is the residual, y represents the expected results, and ŷ represents the forecasting results.

L1-norm loss
In the regression problem, L1-norm loss measures the absolute value of the difference between the forecasting value and true value.The L1-norm loss function can be described as follows: The L1-norm loss function is a function commonly used in regression problems.

Huber loss
The Huber loss function was proposed in 1964.It absorbs the advantages of L1-norm and L2-norm loss functions and makes up for their shortcomings.Concerning outliers in the data, Huber loss can perform more robustly.Not only is it more robust to outliers, but Huber is also derivable in the whole domain, greatly simplifying the calculation difficulty.Huber loss function can be described as follows: where parameter δ represents tuning parameters, which control the quadratic and linear range.It is recommended to set the parameter δ to 1.345 [37].

Pinball loss
The Pinball loss function is asymmetric.It not only imposes certain penalties on outliers in data, but it also imposes additional penalties according to different situations of outliers.
In addition, because of the introduction of quantile distance, the Pinball loss function improves the insensitivity to characteristic noise and resampling.The expression of the Pinball loss function is as follows: The parameter τ ∈ [0, 1].When parameter τ = 1, Pinball loss is the same as the L1-norm loss in function, so the Pinball loss function can be considered a generalized L1-norm loss function.In addition, Pinball loss also absorbs the advantages of L1-norm loss and can handle the deviation of outliers.

Biweight loss
Tukey's Biweight loss function is also a non-convex loss function, which can overcome the interference and influence caused by outlier samples and noise samples in regression tasks, hence showing strong robustness in regression tasks [38,39].The Biweight loss function is defined as follows: where c is a tuning constant, which is generally specified as 4.685.At this time, Tukey's Biweight can achieve a regression effect like that of the L2-norm loss function (95% progressive) in minimizing the variance consistent with the normal distribution [40].Tukey's Biweight suppresses the influence of outliers during backpropagation by reducing the gradient size to near zero.Another interesting feature of this loss function is that it imposes a soft constraint between the inner layer and outlier without setting a hard threshold for the residual.

Lncosh loss
Lncosh is a loss function commonly used in regression tasks, with high smoothness.Define it as [41,42]: For the smaller residual r , ln(cosh(r )) is approximately equal to r 2 2 ; for the larger residual r , it is roughly equal to | r | −ln2.This means that the working principle of Lncosh is very similar to the mean square error to a large extent, but it is not greatly affected by the occasional wrong forecasting.It has all the advantages of the Huber loss function, but unlike Huber loss, it is quadratically differentiable everywhere.

Extreme learning machine
Unlike traditional feedforward neural networks, such as BP, ELM is a single hidden layer feedforward neural network that removes the requirement to set an excessive number of node parameters.ELM was introduced by Huang et al. [16] and has demonstrated substantial progress in the realm of artificial intelligence algorithms.
ELM stands out from conventional neural networks because of its distinctive approach.ELM selects its weight matrix between the input and hidden layers randomly, along with hidden layer thresholds, without any further adjustments during algorithm execution.With no need for additional parameter settings, ELM offers simplicity in its usage.As highlighted by Huang et al. [43], ELM commonly employs the Moore-Penrose generalized inverse to determine key node weights.This methodology involves only a single calculation step (linear equation operation) to establish the weight matrix between the hidden and output layers [44].Unlike backpropagation, there's no gradient operation, significantly reducing computational demands and enhancing speed.Furthermore, ELM demonstrates superior generalization compared with alternative algorithms.The structural diagram of ELM is shown in Fig. 1.
A typical ELM network structure consists of an input layer, a hidden layer, and an output layer, with n, L, and m nodes, respectively.For data set x in ] T is the input vector, y i = [y i1 , y i2 , ..., y im ] T is output vector, and the output of ELM can be described as: where w j is the weight from the input layer to the j-th hidden layer node, b j is the threshold of the j-th hidden layer node, β j is the output layer weight connecting the j-th hidden layer node, and G(•) represents the activation function.Equation ( 7) can be simplified as H β = Y .The objective function of the ELM model can be written as follows: Using the least square method to solve the (8), the solution β is the following: where H + is the Moore-Penrose generalized inverse of the hidden layer output matrix H .

Multiobjective optimization
The concept behind the multiobjective optimization algorithm is to identify a collection of optimal Pareto solutions, where each solution fulfills the fundamental criteria of multiple optimization objectives and showcases an optimal state holistically.Within the optimal Pareto solution set, no other solution surpasses itself in all optimization objectives [45].
Achieving this demands that the optimization algorithm extensively explores the solution set, guarantees a global optimization outcome, and prevents being trapped in local optimization.
Derived from biological genetic theory, the genetic algorithm has evolved and found applications across diverse domains [46].By incorporating the genetic algorithm, the drawbacks associated with traditional multiobjective optimization approaches, such as the risk of converging to local optima, are circumvented.This integration ensures that the solutions' diversity is effectively maintained.
The general multiobjective optimization problem can be described as follows: where f i (x) is the optimization objective, x is the solution, and C is the constraint.NSGA-II is an advanced multiobjective optimization algorithm that was improved by Deb et al. [47].NSGA-II introduces the concepts of fast nondominated sorting, crowding-distance sorting, and elitist strategy, which greatly enhance the practical application of NSGA-II.In NSGA-II, we can initialize a certain population P and use the genetic algorithm to select, cross, and mutate the parent population P to produce the offspring population Q.After fast nondominated sorting and crowding distance sorting of the combined population R = P Q, the new population and its Pareto optimal solution set are obtained by using the elitist strategy.The outflow diagram of NSGA-II is shown in Fig. 2. The specific steps are as follows: Step 1: Set population and iteration times and initialize the parent population P.
Step 2: For the parent population P, conduct fast nondominated sorting and crowding distance sorting and assign each individual the rank.
Step 3: Generate the offspring population Q 0 through tournament selection, simulated binary crossover, and polynomial mutation.
Step 4: Combine the parent P t and offspring Q t to get the population R t = P t Q t , where t is the number of iterations; a new parent population P t+1 is selected through elitist strategy.
Step 5: When the iteration reaches the specific number or the termination condition is met, the final population and its Pareto solution set will be obtained; otherwise, let the number of iterations t = t + 1 and go to Step 2.

Proposed Pinball-Huber loss
Section 3.1 provides an overview of six fundamental loss functions: L2-norm, L1-norm, Huber, Pinball, Biweight, and Lncosh.The strengths, weaknesses, and suitable application contexts for each loss function are analyzed.Upon examination and consolidation, it is evident that these loss functions often lack compatibility with robustness and accuracy in diverse evaluation models, measurement approaches, and forecasting experiments.Additionally, they tend to inadequately address standard positive and negative errors and outliers in machine learning challenges.As a result, this may lead to suboptimal evaluation levels and reduced accuracy in forecasting outcomes.
To address the aforementioned challenges, we propose a solution by merging the Pinball loss with the Huber loss.The Pinball loss function offers the ability to adapt to positive and negative errors during forecasting computations, displaying self-adjusting asymmetry.On the other hand, the Huber loss function demonstrates remarkable robustness and effectively handles outliers; however, it treats both positive and negative issues concurrently in the algorithmic process, leading to a reduction in forecasting precision.Our innovation lies in the development of novel loss functions, combining the attributes of Huber and Pinball.This allows for distinct measures to be applied to diverse errors within the training procedure, significantly enhancing the model's performance.The proposed Pinball-Huber loss function is presented as follows: The newly introduced Pinball-Huber loss function comprises two adjustable parameters: δ and τ .Notably, these parameters originate from the Huber and Pinball loss functions and are skillfully merged to leverage their distinct roles.Their combined utilization allows for tailored actions based Fig. 2 The outflow diagram of NSGA-II on the magnitude and direction of training errors.Beyond refining the accuracy of the foundational loss function, the Pinball-Huber approach introduces a novel perspective by categorizing training errors based on their directional attributes.This innovative method presents a fresh approach for addressing outliers.In the context of the power system, where power load data are influenced by variables like weather, season, and market demand, volatility and the presence of outliers and asymmetric noise are common.Our proposed Pinball-Huber loss function addresses these intricacies by meticulously dissecting errors and handling positive and negative scenarios in distinct ways.

NSGA-II-ELM-Pinball-Huber
In the practical application of ELM, there exists a fundamental trade-off between forecasting accuracy and network structure complexity.Achieving higher accuracy demands a network that can personalize its modeling to the data, which often results in an intricate network structure, particularly within the hidden layer, potentially harboring numerous unnecessary nodes.While pursuing a simplified neural network structure, it is not prudent to directly designate a minimal number of nodes and related parameters.Subjectively determining the appropriate count of neurons for the network to accurately capture the input-to-output relationship is not a feasible approach.What is required is a rational and efficient algorithm to assist in identifying the optimal number of nodes for ELM.
Utilizing the minimization of training error and output weight within the ELM-Pinball-Huber model as the dual optimization objectives, we employ the multiobjective optimization algorithm NSGA-II to enhance the ELM model.From the derived set of Pareto front solutions, we carefully choose the most suitable solution to execute the forecasting task.The objective function for this multiobjective optimization endeavor is presented as follows: where r i is the training error and N represents the number of samples; we use the training error based on the new proposed Pinball-Huber loss function as one optimization objective.β is the output weight vector of the output layer in the ELM model, and L is the number of hidden layer nodes; in addition, we take the L1 norm of it as the other optimization objective.
Following the multiobjective optimization process, we arrive at the set of Pareto solutions.By representing the two optimization objectives along the horizontal and vertical axe, respectively, we observe that the solutions within the Pareto set form a U-shaped distribution.This pattern highlights the inherent trade-off between training error and output weight as optimization objectives.The solution situated at the inflection point simultaneously possesses a lower training error and output weight norm, rendering it the optimal choice for our multiobjective optimization.
Furthermore, to validate the effectiveness of the Pinball-Huber loss function, we conducted a separate comparative test.Employing various loss functions in conjunction with ELM and integrating a lasso penalty term, the composite ELM-loss function models were employed for power load forecasting within identical experimental parameters.Elaborate insights into the model's objective function and its solution procedure are provided in Appendix A.

The overall steps
In this section, we provide a combined load forecasting model NSGA-II-ELM-Pinball-Huber. The model is ELM based on the Pinball-Huber loss function and then optimized by the multiobjective optimization algorithm NSGA-II.The steps of the model can be found in pseudocode Algorithm 1.

Experimental setup and evaluation criteria
In this section, alongside the newly proposed Pinball-Huber loss function, various common loss functions are integrated with ELM for comparison, highlighting the performance of the novel loss function.Six loss functions, namely L2-norm, L1-norm, Huber, Biweight, Lncosh, and Pinball-Huber, are employed as the objective functions for ELM, enabling a comparison of their distinct effects.
The specific experiments involve forecasting the 49th observation based on the preceding 48 observations.Multistep experiments, encompassing three-step, five-step, and seven-step forecasts, are conducted under consistent conditions.A choice of 200 is made for the number of hidden layer nodes in ELM, allowing for the demonstration of the compression effect from the lasso penalty within the ELM-loss function model.The entire experimentation is conducted in Matlab, utilizing Matlab2016 to compile the experimental code (Fig. 3).
We use root mean square error (RMSE), mean absolute error (MAE), and mean absolute percentage error (MAPE) to measure the forecasting effect of the model, as follows: and In the three formulas, r i = y i − ŷi (i = 1, 2, ..., N ) is the residual, y i is the actual power load value, and ŷi is the forecasting value, representing the forecasting value of the i-th sample (Table 1).

Taixing power load data
For the Taixing electric power data set, we carry out power load forecasting.From 2018.5.13 to 2021.8.2, the data set records the power load data every other day.In the power data set of Taixing City, there are 1,175 data points.We divide these into the training set and test set by a ratio of 8:2, in which the former includes 940 points and the latter 235 points.The specific characteristics of the data are shown in Table 2.
In Table 1, some of the six loss functions listed need to set parameters.The δ in Huber loss represents the tuning parameters, which determines how to deal with outliers.3. The experimental analysis is as follows:

Comparisons among ELM-Pinball-Huber and ELM with other loss functions
This section presents the comparative experiments conducted on the ELM model using six distinct loss functions that are aimed at substantiating the benefits of the newly introduced Pinball-Huber loss function.Based on the three evaluation metrics-RMSE, MSE, and MAPE-outlined in Table 4, it can be deduced that the forecasting outcomes of ELM utilizing our Pinball-Huber loss function surpass those achieved with other loss functions, both in single-step and multistep forecasting scenarios.Figure 5 provides a visual representation of the ELM's performance in multistep forecasting across the six different loss functions.Hence, adopting the proposed Pinball-Huber loss function as the objective function for ELM can lead to enhanced forecasting capabilities within the power load prediction.Furthermore, an intriguing observation emerged from the comparison among the six loss functions.The L2-norm and L1-norm loss functions exhibited subpar and unstable performance in multistep forecasting.Huber, Biweight, and Lncosh loss functions demonstrated favorable performance, but their stability in multistep experiments displayed notable fluctuations.Conversely, the ELM-Pinball-Huber consistently demonstrated the most optimal forecasting results while maintaining a relatively stable performance throughout the experiments.

Comparisons of ELM with loss functions with/without multiobjective optimization
In the preceding section's comparative experiments, ELM utilizing the Pinball-Huber loss function exhibited consistent advantages in forecasting accuracy.Nonetheless, a noteworthy observation was that achieving higher precision often resulted in elevated output weights within the ELM model.This outcome could lead to intricate network structures and even overfitting issues.To ascertain the enhanced forecasting performance of ELM-Pinball-Huber through NSGA-II optimization, a comparative validation was conducted.Table 5 presents the outcomes of multiobjective optimization for ELM using diverse loss functions.Upon comparison with the results from the pre-multiobjective optimization experiments illustrated in Table 4, the ELM, post-multiobjective optimization not only attains heightened forecasting precision, but it also substantially diminishes the output weight within the ELM model.The distribution of solutions within the Pareto solution set, along with the curve showcasing the alteration in the two optimization objectives with the number of iterations, is depicted in Fig. 6.Notably, as the number of iterations increases, the values of the two optimization objectives consistently decrease.Ultimately, within the figure, a Pareto solution is discernible, maintaining commendable values for both optimization objectives, thereby achieving elevated forecasting accuracy while concurrently   preserving smaller output weights.This simplification considerably reduces the intricacies of the model network.The streamlined ELM necessitates fewer hidden layer nodes, enhancing its generalization capabilities.Moreover, we have also observed that NSGA-II can enhance the performance of various ELM-loss function combinations, indicating its wide applicability for ELM.Notably, the amalgamation of the Pinball-Huber loss function and ELM, following NSGA-II optimization, demonstrates the most optimal performance.The multistep ahead forecasting curves for the NSGA-II-ELM-Pinball-Huber model are displayed in Fig. 7.

Comparisons among NGSA-II-ELM-Pinball-Huber and comparative models
To assess the predictive performance of the NSGA-II-ELM-Pinball-Huber model, we conduct comparative experiments   with three models: LSTM, GRU, and CNN-BiLSTM-Attention. Brief descriptions of these models are as follows: (1) LSTM model: LSTM, an improved variant of traditional RNN, effectively captures the semantic relationships in long sequences, mitigating gradient vanishing or exploding issues.LSTM features a more complex structure [48].
(2) GRU model: Introduced by Cho et al. [49] in 2014, the GRU neural network addresses the gradient vanishing problem in standard recurrent neural networks and shares a similar design philosophy with LSTM.(3) CNN-BiLSTM-Attention model [50]: This model employs complex mathematical operations in the convolutional and pooling layers of the convolutional neural networks (CNN) to extract the spatial features of the input variables.The multi-head attention layer minimizes irrelevant feature impact, enhancing the extracted features.
The BiLSTM layer models trend information in time series, generating a probability model for prediction distribution.
We adjust the hyperparameters of each model to achieve optimal performance, as shown in Table 6.The evaluation metrics for forecasting performance are presented in Table 7.
The predictive performance of LSTM and GRU is similar, showing close values for RMSE and the output weight.Comparing the predictive performance evaluation metrics of LSTM, GRU, and CNN-BiLSTM-Attention with NSGA-II-ELM-Pinball-Huber, the RMSE of the proposed model was always lower than that of the three comparison models.Particularly in single-step forecasting, the prediction error of NSGA-II-ELM-Pinball-Huber (RMSE=84.27)was significantly smaller than that of the three comparative models (RMSE=184.87,184.58, 273.91).These results indicate that the proposed model effectively captured the changing trends in power load data in both the spatial and temporal dimensions.Furthermore, the proposed model maintained a stable structure in multistep forecasting.
Finally, to verify whether the NSGA-II-ELM-Pinball-Huber model significantly improves predictive accuracy in power load forecasting compared with other models, we conducted the Wilcoxon signed-rank test [51].The significance level for the one-tailed test was set at α = 0.05.The original hypothesis posited that there would be no significant difference in the predictive results between our model and the comparative models in power load forecasting.If the p-value is less than 0.05, the original hypothesis will be rejected (h=1).The predicted values of the proposed model and three comparison models were turned into Wilcoxon signed-rank tests separately in multistep forecasting from 1 to 7 steps.The results are shown in Table 8.

Conclusion and future work
In the current paper, we have introduced a robust Pinball-Huber loss function that demonstrates remarkable resistance to outliers and substantially reduces the likelihood of overfitting.This loss function effectively manages outliers and asymmetrical noise within the dataset, serving as the objective function for training the ELM model.Given the ELM's susceptibility to preset parameters' influence, and aiming to ensure forecasting accuracy while maximizing and simplifying the ELM network structure, as well as preventing the squandering of training time and the emergence of overfitting because of an excessive number of hidden layer nodes, we employed the NSGA-II algorithm for the optimization of both training errors and output weights within the ELM model.The combined NSGA-II-ELM Pinball-Huber model was then employed for power load forecasting in the context of Taixing City.By employing the multi-objective optimization algorithm NSGA-II, we acquired the Pareto optimal solution set for the number of hidden layer nodes in the ELM model, enabling an in-depth analysis of the forecasting outcomes.Our analysis of the experimental outcomes revealed that the performance of the suggested Pinball-Huber loss function within the ELM framework surpassed that of other loss functions.Moreover, the NSGA-II algorithm effectively enhanced the performance of diverse ELM-loss function combinations.The innovative combined approach, NSGA-II-ELM Pinball-Huber model, can be seen as a promising and effective method for power load forecasting, offering a novel solution to this domain.
Multiobjective optimization greatly improves the predictive performance of the model, but it takes up a significant amount of computational resources.In the future, we aim to delve deeper into simplifying the computational resources and time required for training the proposed method, which is crucial for the widespread applicability of the model.Furthermore, this forecasting model can only provide predicted values for future power loads, and recent research has focused on uncertain predictions.Future studies will delve deeper into the probability predictions of the model, which holds significant value for practical applications in power systems [52,53].

A The ELM-loss function model
The combined ELM-loss function is a single objective model, and its mathematical model can be written as follows: where r i represents the training error of the sample and Lagrangian multipliers are introduced for each equality constraint condition in the model, and the Lagrangian function is constructed to transform it into an unconstrained optimization problem: where α = [α i , α 2 , ..., α N ] is the Lagrange multiplier vector.We solve (17), and we obtain the output weight vector β as follows: where W N is the sample weight matrix and W L is the weight matrix of hidden nodes.The details of w i of the loss function can be found in Table 1.Their specific forms are as follows: In general, the specific steps of the ELM-Pinball-Huber model are as follows: Step 1: Initialize the relevant parameters w, b, and L of the ELM-Pinball-Huber model.
Step 3: Update sample weight matrix W N and hidden nodes' weight matrix W L .
Step 5: Substitute the test set into the trained model to get the forecasting results.
Similar to the above ELM-Pinball-Huber model, we can combine the loss functions in Table 1 with ELM, respectively, to compare their performance.
Open Access This article is under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material.If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.To view a copy of this licence, visit http://creativecomm ons.org/licenses/by/4.0/.

Fig. 1
Fig. 1 The structural diagram of ELM.((x 1 , x 2 , ..., x N ) is the input of ELM.(w N , b N ) is the weight and threshold of hidden layer.β M is the output layer weight.)

Require:
Hidden layer nodes L, input weight ω, and hidden layer threshold b Ensure: Training error and output weight of ELM 1: Set the population and iteration times 2: Initialize the population: 3: Randomly generate multiple groups of ELMs with different numbers of hidden layer nodes 4: for each ELM 5: Select the input weight ω and hidden layer threshold b randomly 6: Take the proposed Pinball-Huber loss function as the objective function and solve the output weight vector β = (H T W H) −1 H T W Y by IRLS 7: Get the training error N i=1 P H(r i ) based on Pinball-Huber loss function and output weight L i=1 | β i | of ELM as the two optimization objectives 8: end for 9: Fast nondominated sort and crowding distance sort on the population and take it as the Parent 10: for each i = 1, 2, 3, • • • , gen do 11: Tournament selection, Simulated Binary Crossover, and Polynomial mutation to produce the Offspring 12: Merge Parent and Offspring 13: Fast nondominated sort and crowding distance sort 14: Generate new Parent by Elitist strategy 15: end for 16: Get optimized Pareto solution set of all the groups 17: Select the best sample from the set, that is, the ELM model with the best parameters L, ω, b 5 Case study This section employs the power load dataset from Taixing City in southern Jiangsu Province to validate the efficacy of the integrated NSGA-II-ELM-Pinball-Huber forecasting model within the power load system.

Fig. 4
Fig. 4 Training error distribution diagram of ELM-Pinball-Huber in the Taixing data set

Fig. 5
Fig. 5 Multistep forecasting results of ELM-Pinball-Huber and ELM with other loss functions in the Taixing data set

Fig. 6 Fig. 7
Fig. 6 Distribution of Pareto solution set and optimization iteration curves of NSGA-II-ELM-Pinball-Huber in the Taixing data set

N
i=1 P H(r i ) is the total error under Pinball-Huber loss function of N different training samples, here representing experience risk.L j=1 | β j | is a lasso penalty term, representing the complexity of the model.C > 0 is called the regularization parameter or the penalty parameter and is used to balance empirical risk and model complexity.

Table 1
Loss functions and their weight functions

Table 3
The hyperparameters of Pinball-Huber loss function

Table 4
Multistep forecasting results of ELM-loss function models in the Taixing data set