Abstract
The purpose of this work is to compare predictive performance of neural networks trained using the relatively novel technique of training single hidden layer feedforward neural networks (SFNN), called Extreme Learning Machine (ELM), with commonly used backpropagation-trained recurrent neural networks (RNN) as applied to the task of financial market prediction. Evaluated on a set of large capitalisation stocks on the Australian market, specifically the components of the ASX20, ELM-trained SFNNs showed superior performance over RNNs for individual stock price prediction. While this conclusion of efficacy holds generally, long short-term memory (LSTM) RNNs were found to outperform for a small subset of stocks. Subsequent analysis identified several areas of performance deviations which we highlight as potentially fruitful areas for further research and performance improvement.
Similar content being viewed by others
1 Introduction
The prediction of stock price movements is of particular interest to traders in financial markets, in addition to a wide range of finance applications not limited to the market timing strategies of funds managers and the selective hedging practices of corporations. Paradoxically the most successful theories of this field assume that stock prices undergo random walks when markets are in equilibrium and include work pioneered by the seminal publications of Black, Scholes [2, 3], and Merton [16] in continuous time and the Binomial Options Pricing model of Cox, Ross, and Rubinstein [7] in discrete time. Nonetheless, this remains a developing area, with frequent deviations from the random walk assumption observed (such as those identified in Jacobs [14]) and with recent advances in data provision and computational techniques many of these approaches have found application in this area of research.
One such application has been machine learning techniques (Strader (2020)). This field has seen the widespread application of Neural Network models, which are explicitly nonlinear in their mathematical representations (see, for example, Chollet [6] with such specifications as Long Short-Term Memory (LSTM) models being particularly suited to financial time-series analysis. Nevertheless, much of the published work is accomplished using deep neural networks, with relatively few studies using shallow neural networks notably across regionally separated markets. Strader [23] in particular, highlights these areas as requiring additional empirical research. The Extreme Learning Machine (ELM) of Huang et al. [13] is one such alternative training technique for shallow neural networks that has recently seen promising empirical results across various practical applications. These applications have been in diverse areas of research, ranging from image quality assessment Suresh, Venkatesh Babu, and Kim [25], improving wireless sensor network localization Phoemphon, So-In, and Niyato [21] to predicting landslides (KV, Pillai, and Peethambaran [17]). A natural extension of these applications then is in the area of financial research.Footnote 1
Tang et al. [27] present a case for using a Random Vector Functional Link (RVFL) network, a special case of single hidden layer feedforward neural networks, together with a Variational Mode Decomposition for the day-ahead crude oil price prediction and compare the results with other neural networks. They find the RVFL network outperforms simpler ELM-based models (as it distinctively includes a direct input–output link, unlike an ELM), albeit taking longer to train. Bisoi et al. [4, 5] support these findings on the same dataset utilising similar network choices. Bisoi et al. [4, 5] explore the same concept of Variational Mode Decomposition yet applying it in the study of next-day ahead stock price prediction using ELM-based models. They analyse a sample of daily prices from regional markets and find promising results for the S&P BSE 500, Hang Seng and FTSE 100 Indices. Mohanty, Parida, and Khuntia [19] show an improvement on a plain ELM-based model performance for stock price prediction by morphing it into a kernel ELM (KELM) combined with an auto encoder (AE). Mohanty et al. [19] test this augmented model (KELM-AE) on several bank stock indices using a normalised OHLC (open-high-low-close) “candlestick” dataset as input variables with the next-day ahead prediction. Göçken et al. [10] compare the effectiveness of using ELM-trained SFNN to predict stock prices on the Borsa Istanbul (BIST 100) stock exchange, albeit focusing on a relatively narrow subset of 3 stocks. Results are compared to other popular models used for the task such as DNN, Jordan RNN, GLM, RT, and Gaussian Process Regression. The conclusion drawn on the best performing model architectures and variables is used for prediction and comparison in this study. An empirical study conducted by Zhang [30] involved an attempt to apply an ELM-trained SFNN to predict price movements of a stock on the Hong Kong stock exchange. Zhang’s results (2021) tantalizingly provide some empirical support to the work of Gocken et al. [10].
This paper utilises emerging findings from Zhang [30] within the framework of Gocken [10] and designs, validates and evaluates a series of SFNN models, trained using the ELM methodology, on the 20 largest stocks on the Australian equity market, also known as constituents of the ASX20. The ELM-trained model results are shown to indeed fulfil the promise of fast learning times (as compared to training the same models using classic backpropagation algorithm) with a comparably high, and often superior, level of accuracy. In general, improving prediction accuracy in this task, even if by a slight margin, can bring about material benefits to the interested stakeholders. This is the driving motivation behind this study; it shows that less computationally intensive model training techniques can deliver potentially higher economic benefits to capital market participants. With varying degree of computational power availability among market participants, designing more efficient techniques, such as those based on ELM training methodology, has direct industry applications. Additionally, performance findings may be used to academically interpret the operating mechanism of neural networks, thus advancing this strand of research.
We draw particular distinction between existing studies in the underlying dataset used as individual stocks are likely to possess different movement characteristics driven by inherent risks and influential factors, as compared to, for example, combinatory stock indices or foreign exchange rates. To the best of our knowledge, there has been no study comparing stock price prediction performance of the most commonly known and successfully used LSTM to the relatively novel ELM training methodology on a broad set of large and frequently traded stocks.
Another differentiating element between these relevant studies is the extent of initial data reconstruction applied to the raw financial price series. Given the well-studied noisiness and dynamics in the financial price series, such structured methods as Empirical or Variation Mode Decomposition Das et al. [8], Bisoi et al. [4, 5], and Discrete Wavelet Transform Wu et al. [28] have previously been applied to the raw stock price series. The rest of the studies rely on the logic of the Takens’ theorem [26], either explicitly or implicitly, by constructing technical features and statistical metrics from the raw price series to be used as the model input Khuwaja et al. [15] provide the most comprehensive detail on the application of this methodology, Das et al. [9], Mohanty et al. [19], Panda et al. [20] appear to implicitly follow this path). We apply the latter approach of explicitly applying the spirit of the Takens’ theorem by constructing technical features from the raw price series, intentionally very similar to the ones used in Zhang [30], albeit with slight modification and an addition.
This paper is organised as follows. Section 2 details the training and testing methodology and data used in this study. Subsequent Results and Discussion section evaluates model performance for the two mentioned training methodologies on the holdout test dataset. Finally, Conclusion and Future Research section concludes the paper and identifies promising future research opportunities.
2 Methodology
This section first discusses the models construction and training methodology, followed by the data description and preparation.
2.1 Models and training methodology
2.1.1 Extreme learning machine (ELM)
Extreme Learning Machine, as introduced by Huang et al. [13], modifies the training methodology for single hidden layer feedforward neural networks by converting it to an analytical solution. It has been shown ELMs (i.e., ELM-trained SFNNs) produce similar, if not better results for the typical NN-allocated tasks, albeit at much faster training speed and avoiding local optima convergence challenges.
In an ELM, the vectors of input weights and hidden node biases are first randomly assigned and used to calculate the hidden layer output matrix, \(H\), using a specified transfer function. The Moore–Penrose generalised inverse of this matrix \(H\), denoted as \({\text{H}}^{\dag }\) as adapted from Huang et al. [13], is then used to analytically determine the vector of output weights \(\gamma\) that would best fit with the data output matrix \(T\).
Mathematical representation of a single hidden layer feedforward neural network (SFNN) trained with Extreme Learning Machine (ELM) training methodology is as follows. A model with \(N\) distinct samples and \(j\) randomly assigned hidden neurons \(({X}_{j},{t}_{j})\) where \({X}_{j}={\left[{x}_{j1},{x}_{j2},\dots ,{x}_{jn}\right]}^{T}\) as an input and \(T={\left[{t}_{j1},{t}_{j2},\dots ,{t}_{jn}\right]}^{T}\in {R}^{n}\) as targets vector can be represented by the following equation:
where \({\gamma }_{j}\) and \({W}_{j}={\left[{w}_{j1},{w}_{j2},\dots ,{w}_{jn}\right]}^{T}\) are the output and input weights, respectively, \({b}_{j}\) is the randomly assigned bias of \(jth\) hidden neuron and \(h(.)\) is the nonlinear activation function. The goal of the ELM training methodology is to reduce the error between the predicted and the target (actual) values, such that \({\sum }_{n=1}^{N}\Vert {o}_{n}-{t}_{n}\Vert =0\) and \({\sum }_{j=1}^{L}{\gamma }_{j}h\left({W}_{j}.{X}_{n}+{b}_{j}\right)={t}_{n}, for n=1,\dots ,N.\)
In short, the ELM training methodology comprises the following steps:
-
1)
Randomly assign hidden layer weights \({W}_{j}\) and bias \({b}_{j}\) values.
-
2)
Calculate the hidden layer output matrix H:
$$H = \left[ {\begin{array}{*{20}c} {h\left( {x_{1} } \right)} \\ \vdots \\ {h(x_{N} )} \\ \end{array} } \right] = \left[ {\begin{array}{*{20}c} {h_{1} \left( {W_{1} x_{1} + b_{1} } \right)} & \cdots & {h_{L} \left( {W_{L} x_{1} + b_{L} } \right)} \\ \vdots & \ddots & \vdots \\ {h_{1} \left( {W_{1} x_{N} + b_{1} } \right)} & \cdots & {h_{L} \left( {W_{L} x_{N} + b_{L} } \right)} \\ \end{array} } \right]$$(2)where \({h}_{j}(x)\) stands for the nonlinear transfer function of the j-th hidden neuron, and L stands for the number of hidden neurons chosen in the SFNN.
Each column of the H output matrix represents the j-th hidden neuron output vector with regards to the vector of inputs \({x}_{1}, {x}_{2}, \dots , {x}_{N}\).
-
3)
Calculate the vector of output weights \({\gamma }_{j}\):
$$\gamma = {\text{H}}^{\dag} T$$(3)
where \(\gamma ={\left[{\gamma }_{1},{\gamma }_{2},\dots , {\gamma }_{L}\right]}^{T}\) is the vector of output weights, \(T={\left[{t}_{1},{t}_{2},\dots , {t}_{N}\right]}^{T}\) represents the output matrix, and \({\text{H}}^{\dag}\) is the Moore–Penrose generalised inverse of the hidden layer output matrix H.
Effectively, the objective function of the SFNN trained using ELM methodology is the minimisation of the cost function as follows:
where \(j=1,\dots ,L\). The minimisation of the cost function is based on the sum of squared errors calculation represented by \(\epsilon\) and is defined in Eq. (5)
Figure 1 depicts a high level overview of the ELM-trained neural network structure, with the specific inputs and outputs chosen for this study. For full details and proof of theorems underpinning the ELM training methodology please see Huang et al. [13].
“The design of an ANN [Artificial Neural Network (ANN)] is more of an art than a science” Zhang, Patuwo and Hu [31], and, in the case of the ELM-trained SFNN, it is mainly the number of hidden nodes in the single hidden layer that needs to be chosen. Some works in this area focus on developing theoretical bounds on the minimum and maximum number of hidden nodes required (for example, LeCun et al. [18] while others develop complex algorithms for supporting this decision using the dataset at hand Xu and Chen [29]. Yet another category focuses on providing rule-of-thumb advice for determining optimal number of hidden nodes based on the number of inputs and outputs in the model, or merely on the number of training samples. Major risk here is discounting some of the other key attributes of the dataset used (e.g., signal-to-noise ratio or complexity of the function to be learnt) that may materially impact on this decision.
Thus, we use cross-validation to determine the optimal number of hidden nodes in the network. Given we are dealing with the time-series data, we split the training set (~ 7.5 years out of the total 10 years of data obtained) into training (the first 5 years) and validation (the remaining ~ 2.5 years) subsets. We train ELM neural networks starting with very small number of hidden nodes (2) progressing incrementally (initially, with increments of 1) to larger network sizes (> 300 hidden nodes in the largest network) for each individual stock. The best topology is then chosen based on the mean squared error results from the validation dataset for each stock across networks trained. Before the best chosen model is used for the final test on the holdout data (the last 2.5 years of data available), training and validation datasets are combined, and this final model is trained on the full training dataset.
2.1.2 Recurrent neural network (LSTM)
The Long Short-Term Memory (LSTM) algorithm was developed by Hochreiter and Schmidhuber [11] to address the vanishing gradient problem, a persistent effect observed in simpler (i.e., non-recurrent) feedforward neural networks as their depth increases (theoretical explanation behind this effect is discussed extensively by Bengio et al. [1].
At its core, the LSTM algorithm allows the network to “carry” information across many timesteps (hence the name) to be later “reinjected” back into the network when needed. This is particularly useful for tackling tasks where time-series are studied and locally learnt features at some previous point can then be “remembered” by the network and used later when a similar pattern arises. Hu et al. [12] conduct a survey of literature studying deep learning models used for stock price prediction and conclude hybrid LSTM-based models are most widely researched.
We train LSTM neural network models using raw stock daily close price data with identical lookback period of 5 days. Decision to use the raw daily stock price data is based on understanding of the feature extraction mechanism of an LSTM Recurrent Neural Network (RNN). An LSTM network is designed to be able to remember information and features learned several timesteps before the current unit processing which is what we are attempting to accomplish with feature extraction for the ELM-trained SFNNs. 5 day lookback period is chosen to provide direct comparison between ELM-trained SFNN and LSTM neural networks. A three-layer stacked LSTM structure with 50 hidden nodes in each layer is used as a commonly used architecture of this type of a neural network in finance research (see, e.g., Sirignano and Cont [22].
Figure 2 provides a schematic overview of the training process behind an LSTM-based neural network.
2.2 Data
Individual stocks in the S&P/ASX20 index are chosen for empirical evaluation of the models. This is consistent with existing literature in the field (as discussed in detail in Sect. 1) while also capturing broad coverage of industries on the Australian share market. Daily candlestick chart price data (Open-High-Low-Close, without Volume) is obtained using yahoo-finance module in Python (Table 1).
The period under analysis is chosen between the two most significant events of the past 20 years that had major impact on financial markets—Global Financial Crisis and the Covid-19. Thus, 1 January 2010 to 31 December 2019 forms the analysis period, and the price data is split into training, validation, and testing datasets to create an objective evaluation protocol.
Maintaining the sequential logic of the time-series, the first 5 years of the period (1 Jan 2010–31 Dec 2014) form the training dataset, the following 2.5 years are used as the validation dataset (1 Jan 2015–30 Jun 2017), with the remainder 2.5 years (1 Jul 2017–31 Dec 2019) being used as the holdout test dataset for the models.
History of the index constituents for the period under study is obtained and stocks that have been present in the index for more than a year are chosen. Table 2 contains descriptive list of the dataset obtained, and the final set of individual stocks chosen for model testing.
For illustration purposes, example price charts for several stocks included in this study are provided below (Fig. 3).
This study diverges from the previous research (for example, Zhang [30] or Bisoi et al. [4, 5] by testing models on the stock price data that may have moved out of a certain narrow range (e.g., within 3 standard deviations of the historical mean). Dataset under study is chosen broadly to capture stock price movements of any kind and range to eliminate any bias from the research.
To present this data in a meaningful manner, we calculate technical indicators utilising the following parameters of a “lookback period” \(L\) = 5 and an “forward period” \(O\) = 1. These parameters have been chosen arbitrarily yet driven by the logic of having at least one week of daily stock price data (hence the lookback period of 5 days) and predicting for the time period ahead where this data sensibly matters (thus, limiting this to just 1 day ahead). The choice of parameters is consistent with the existing literature in the field (e.g., Göçken et al. [10], Khuwaja et al. [15]).
Using these parameters, the daily stock price data are then converted into a set of technical indicators as below. Assuming daily stock price data contains substantial amount of noise, Takens’ delay embedding theorem is relied upon to construct these smoothed “attractors” (i.e., technical indicators in this case) that could be used for stock price prediction Takens [26]. The technical indicators developed for this study were built to resemble the ones presented in the study by Zhang [30], yet with some addition, and comprise the following:
1. Maximum: the maximum stock price for the i-th lookback period as \(Max({P}_{i-L}, {P}_{i-L+1},\dots , {P}_{i-1}, {P}_{i})\);
2. Minimum: the minimum stock price for the i-th lookback period as \(Min({P}_{i-L}, {P}_{i-L+1},\dots , {P}_{i-1}, {P}_{i})\);
3. Average: the average value of the stock price for the i-th lookback period as \(\frac{\sum \left({P}_{i-L}, {P}_{i-L+1},\dots , {P}_{i-1}, {P}_{i}\right)}{L}\);
4. Standard Deviation (SD): the standard deviation of the stock price from its mean for the i-th lookback period calculated as \(\sqrt{\frac{1}{L-1}\sum_{j=0}^{L}({P}_{i-j}-\overline{{P }_{i}})}\) where \(\overline{{P }_{i}}\) is the mean of the stock price for the i-th lookback period (calculated as the Average above);
5. Pseudo Log Return (PLR): the logarithmic difference between average prices of consecutive lookback periods;
6. Trend Indicator (TI): a simple trend indicator calculated as the difference between the last close price for the i-th lookback period, \({P}_{i}\), and the first close price of the (i)-th lookback period, \({P}_{i-4}\), and taken as an “upward trend” (i.e., assigned the value of “1”) when the difference is positive (> 0), “no trend” (“0”) when there is no difference (i.e., \({P}_{i}\) = \({P}_{i-4}\)), or “downward trend” when it is negative (< 0; and assigned the value of “-1”). This indicator is created to measure a trend within each lookback period;
7. Previous Day Close (PDC): the stock price on the last day of the i-th lookback period.
Table 3 shows a sample set of technical indicators calculated for a stock given the original candlestick chart price data. As shown in Fig. 1, these technical indicators for each i-th lookback period (i.e., for each day in a sequence of daily stock price data) are fed into the ELM-trained SFNN model directly.
3 Results and discussion
Figure 4 visualises prediction results for a sample of stocks comparing two commonly used neural network training methodologies—ELM and LSTM—to the actual stock price data. This result is shown on the final, holdout test dataset (period from mid-2017 until the end of 2019) that has never been seen by any of the models and represents a one-off final test run. The original stock price (grey) is compared to the ELM-trained network prediction (blue) and prediction by the LSTM model (orange).
Summary numerical results for all stocks under analysis are presented in Table 4 on a number of critical performance metrics, including speed of training which gains importance as the prediction horizon decreases. All training and testing procedures were run on the Dell Latitude 5310 laptop with IntelI CITM) Processor, i5-10310U CPU @ 1.70 GHz, 2208 MHz, 4 Core(s), 8 Logical Processor(s) with 16.0 Gb of RAM (Installed Physical Memory).
Overall, the performance of the ELM-trained models is exceeding that of the LSTM networks, with only a few cases of LSTM slightly outperforming the former. On average, Group 1 in Table 4—where ELM models provide superior prediction results to the LSTM ones—shows that ELM performs materially better than LSTM for these time-series. However, for Group 2—where LSTM yielded better performance than ELM—the difference in performance between the two models under study is minimal. This is also supported by the charts in Fig. 4 where the top two stocks (STO.AX and IAG.AX) belong to Group 1 and the other two stocks (WPL.AX and WBC.AX) are from Group 2. Group 1 stocks have visible difference in LSTM and ELM prediction, with LSTM being off point visibly. Group 2 sees ELM and LSTM predictions move almost too close to call.
The speed of training is also noteworthy to mention. ELM-trained models have been known to increase the speed of training by a factor of 10–100 (at least) compared to their more widely accepted counterparts (LSTM in this case) Wu et al. [28]. Our test showed ELM models have been trained and tested more than 100 times faster than LSTM ones for the full set of stocks under analysis. The value of this advantage is appreciated when dealing with constant flow of live data on shorter frequencies and the need to make prompt decisions (e.g., in the high-frequency trading environment).
In general, ELM appears to capture well trends in the data, as well as better reacting to more short-term changes in the price move process (i.e., peaks and troughs). This may be attributed to the set of technical indicators used for prediction where network is almost able to build an expected stock price distribution and choose the most likely next point given the latest point in the data (i.e., previous day close price).
Let us contrast the stocks where LSTM achieved better prediction results than ELM with the ones where it did not. They do belong to either the financial services (ANZ, NAB, WBC) or metals and mining (NCM, WPL) industries, however, stocks belonging to the same industries are found within Group 1 as well (for example, CBA, MQG or BHP, STO) where ELM performed better than LSTM. Therefore, a hypothesis that changes between these two groups may be driven by industry membership does not stand.
Table 5 presents initial further analysis on the issue. Descriptive statistics are calculated for stocks from the two groups, split by the dataset used to train, validate, and test the models. It should be noted that the final model is trained on the full combined training and validation dataset, before running the final test on the holdout dataset. This timewise split allows us to explore the differences in model performance between stocks.
First, it is worth noting that either model, but especially LSTM, performed comparatively worse on stocks that experienced strong growth through the period of analysis (2010–2019). For example, CSL.AX and MQG.AX had substantial increase in the mean between datasets. Second, it appears that stocks from Group 2 had substantial change in data distribution skewness (highlighted by bold italics in the table) between periods. For example, ANZ.AX showed positive skewness around 0.3–0.4 in the training and validation datasets which then retraced to a negative 0.36 in the test period. WPL.AX time-series followed the similar path to ANZ.AX. NAB.AX and WBC.AX showed matching trajectory where a positive skewness in the full training period changed to almost 0 in the test data. NCM.AX skewness has sharply increased from the negative 0.15 for training to the positive above 1 for the test datasets. This change in skewness represents material change in data interaction dynamics between time-series mean and standard deviation that were used as critical input variables into the ELM-trained models. This may explain the difference in predictive performance results between models for these stocks, however, bodes well for further research.
To test economic significance and practicality of the aforementioned findings, a simple trading strategy is developed using ELM model-based stock price predictions. Table 6 shows profitability results by threshold level δ that is used in the following manner:
where \({a}_{t}^{i}\) is the action taken at time \(t\), \(\Delta {s}_{t}^{i}\) is the expected change in stocks price \(s\) of stock \(i\) for time period \(t\) based on the ELM model prediction, and δ is the chosen threshold level in absolute $ terms to indicate whether the predicted stock price change is considered material. Performance of the prediction-based strategy is tested on several threshold levels as there does not appear to be agreement in the literature on a common value or even an approach. For example, Mohanty et al. [19] and Bisoi et al. [4, 5] only compare the next day predicted price to the current day value, inherently assigning 0 value to the action threshold. We use threshold in this strategy to avoid unnecessary trades based on minor changes in price and, since no agreement appears to exist in the literature as to its value, we test a range of threshold levels.
We test model performance using simple return metric by various thresholds to indicate persistence of the findings and conduct robustness checks. The main result of ELM outperformance over its LSTM counterpart holds across the full range of thresholds used. The difference is that fewer trades are made with higher thresholds. Profitable results are bolded in Table 6. In addition to being profitable for the majority of stocks, it is important to note that the overall average results of the model are positive across all threshold values. These results may be best understood as investing the same amount of funds for each individual stock trading from the beginning. These results are designed to reveal that real-world implementations are possible, and future research should investigate an optimal trading strategy based on such a model.
4 Conclusion and future research
Prediction of future stock prices returns is arguably one of the most challenging areas of finance. This is driven by common belief that stock prices have a relatively low signal-to-noise ratio and a substantial array of influential factors, with time- and cross-section-varying weightings assigned to each. This presents a difficult task for predictive models, albeit an almost impossible one for traditional, linear models. Nonlinear machine learning models have shown some success in this field, but they most often require vast amounts of data to extract usable features from the high-noise price data. However, the relatively new Extreme Learning Machine training methodology for single hidden layer neural networks has not yet been comprehensively evaluated for this purpose. This study compares this relatively novel ELM training methodology to the more well-known Long Short-Term Memory (LSTM) recurrent neural network. Utilising a set of simple technical indicators built on raw stock price data, ELM-trained neural networks are constructed, trained, and then tested on a set of blue-chip stocks listed on the Australian Securities Exchange.
The analysis confirms the proposed benefits of ELM training, specifically reduced training time with comparable predictive power without requiring as much data. It further confirms emerging findings of Zhang [29] as to the overall efficacy of the models and potential shortfalls, albeit on a broader real stock price dataset. ELM-trained models have shown substantial improvement in predictive accuracy on the majority of individual stock price datasets used in the study. In the relatively few cases where LSTM does better, it appears that changes in the stock price data distribution might be the reason for this deviation.
The performance discrepancy between LSTM and ELM-trained models bodes well for potential future research. There are relatively few number of cases where LSTM outperforms and, in these cases, we have observed substantial drift in the degree of skewness between training and test datasets as compared to less unexpected changes for the rest of the stock price series. LSTM’s ability to capture changes in the underlying asymmetry of the distribution may prove a significant advantage in practice. Identifying important variables in the ELM neural networks and trialling additional input metrics directly relating to skewness of the input series distribution also represent interesting research directions. Additionally, testing the results on a different set of financial instruments may shed further light on the findings about the LSTM’s ability to respond better to changes in the underlying data distribution which would be a useful feature in investment practice.
Given the encouraging results, future research could investigate ways to further enhance the ELM-based trading system. It would be valuable to investigate adding a stronger trend indicator (such as a medium or a long-term Exponential Moving Average) given that the current ELM did not perform well with the changing distribution of the price data. It is likely to help better identify the underlying trend in the data and adjust price prediction accordingly. Another parameter that could be further tuned is the threshold value in the action decision step. The strategy presented used an initial set of dollar-based thresholds, but the optimal threshold to use might vary according to underlying timeseries data and its distribution properties. Finally, the choice of input variables used in the model can be further investigated to improve its predictive power. For example, it might be beneficial to trial various volatility measures instead of the standard deviation currently used in the model. Measures that may better reflect more recent volatility or more accurately capture both short- and long-term inherent price series variance are most promising. Such future research identifying better methods to more accurately predict security prices has widespread applications, from improving on pure profitability-driven investment objectives to better informed portfolio construction and risk management.
Data availability
The datasets generated during and/or analysed during the current study are available in the Yahoo Finance repository, https://finance.yahoo.com/lookup?s=DATA. Data have been sourced using the API module connection direct from Python IDE, Spyder using the yahoo-finance module commands. Additionally, the datasets generated during and/or analysed during the current study are also available from the corresponding author on reasonable request.
Notes
Suganthan and Katuwal [24] provide an overview of original works in the area of randomisation-based feedforward neural networks, shedding light on potential origins of the ELM training methodology.
References
Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw 5(2):157–166. https://doi.org/10.1109/72.279181
Black F, Scholes M (1972) The valuation of option contracts and a test of market efficiency. J Finance 27(2):399–417. https://doi.org/10.2307/2978484
Black F, Scholes M (1973) The pricing of options and corporate liabilities. J Polit Econ 81(3):637–654
Bisoi R, Dash PK, Mishra SP (2019) Modes decomposition method in fusion with robust random vector functional link network for crude oil price forecasting. Appl Soft Comput J 80:475–493. https://doi.org/10.1016/j.asoc.2019.04.026
Bisoi R, Dash PK, Parida AK (2019) Hybrid variational mode decomposition and evolutionary robust kernel extreme learning machine for stock price and movement prediction on daily basis. Appl Soft Comput J 74:652–678. https://doi.org/10.1016/j.asoc.2018.11.008
Chollet, F. (2018). Deep learning with Python. Manning Publications Co.
Cox JC, Ross SA, Rubinstein M (1979) Option pricing: a simplified approach. J Finan Econ 7(3):229–263. https://doi.org/10.1016/0304-405X(79)90015-1
Das PP, Bisoi R, Dash PK (2018) Data decomposition based fast reduced kernel extreme learning machine for currency exchange rate forecasting and trend analysis. Expert Syst Appl 96:427–449. https://doi.org/10.1016/j.eswa.2017.10.053
Das SR, Mishra D, Rout M (2019) Stock market prediction using firefly algorithm with evolutionary framework optimized feature reduction for OSELM method. Expert Syst with Appl: X. https://doi.org/10.1016/j.eswax.2019.100016
Göçken M, Özçalıcı M, Boru A, Dosdoğru AT (2019) Stock price prediction using hybrid soft computing models incorporating parameter tuning and input variable selection. Neural Comput Appl 31(2):577–592. https://doi.org/10.1007/s00521-017-3089-2
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
Hu Z, Zhao Y, Khushi M (2021) A survey of forex and stock price prediction using deep learning. Appl Syst Innov 4(1):1–30. https://doi.org/10.3390/ASI4010009
Huang GB, Zhu Q, Siew C (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501. https://doi.org/10.1016/j.neucom.2005.12.126
Jacobs H (2015) What explains the dynamics of 100 anomalies? J Bank Finan 57:65–85
Khuwaja P, Khowaja SA, Khoso I, Lashari IA (2020) Prediction of stock movement using phase space reconstruction and extreme learning machines. J Exp Theor Artif Intell 32(1):59–79. https://doi.org/10.1080/0952813X.2019.1620870
Merton RC (1973) Theory of rational option pricing. Bell J Econ Manag Sci 4(1):141–183. https://doi.org/10.2307/3003143
Shihabudheen KV, Pillai GN, Peethambaran B (2017) Prediction of landslide displacement with controlling factors using extreme learning adaptive neuro-fuzzy inference system (ELANFIS). Appl Soft Comput 61:892–904. https://doi.org/10.1016/j.asoc.2017.09.001
LeCun, Y., Bottou, L., Orr, G.B., Müller, K.R. (1998). Efficient BackProp. In: Orr, G.B., Müller, KR. (eds) Neural Networks: Tricks of the Trade. Lecture Notes in Computer Science, vol 1524. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-49430-8_2
Mohanty DK, Parida AK, Khuntia SS (2021) Financial market prediction under deep learning framework using auto encoder and kernel extreme learning machine. Appl Soft Comput. https://doi.org/10.1016/j.asoc.2020.106898
Panda A, Rath A, Uday Kiran Reddy CH (2020) On efficient prediction of Indian stock indices using ELM-based technique. Adv Intel Syst Comput 1118:371–381. https://doi.org/10.1007/978-981-15-2475-2_35
Phoemphon S, So-In C, Niyato DT (2018) A hybrid model using fuzzy logic and an extreme learning machine with vector particle swarm optimization for wireless sensor network localization. Appl Soft Comput J 65:101–120. https://doi.org/10.1016/j.asoc.2018.01.004
Sirignano J, & Cont R (2018). Universal features of price formation in financial markets: perspectives from deep learning. https://doi.org/10.2139/ssrn.3141294.
Strader TJ, Rozycki JJ, Root TH, Huang Y-H (2020) Machine learning stock market prediction studies: review and research directions. J Int. Technol Inf Manag 28(4):3
Suganthan PN, Katuwal R (2021) On the origins of randomization-based feedforward neural networks. Appl Soft Comput. https://doi.org/10.1016/j.asoc.2021.107239
Suresh S, VenkateshBabu R, Kim HJ (2009) No-reference image quality assessment using modified extreme learning machine classifier. Appl Soft Comput J 9(2):541–552. https://doi.org/10.1016/j.asoc.2008.07.005
Takens, F (1981). Detecting strange attractors in turbulence. Dynamical systems and turbulence, lecture notes in mathematics, vol. 898. Springer-Verlag. pp. 366–381.
Tang Z, Zhang T, Wu J, Du X, Chen K (2020) Multistep-ahead stock price forecasting based on secondary decomposition technique and extreme learning machine optimized by the differential evolution algorithm. Math Probl Eng 2020:13. https://doi.org/10.1155/2020/2604915
Tang L, Wu Y, Yu L (2018) A non-iterative decomposition-ensemble learning paradigm using RVFL network for crude oil price forecasting. Appl Soft Comput J 70:1097–1108. https://doi.org/10.1016/j.asoc.2017.02.013
Wu D, Wang X, Wu S (2021) A hybrid method based on extreme learning machine and wavelet transform denoising for stock prediction. Entropy. https://doi.org/10.3390/e23040440
Yang F, Chen Z, Li J, Tang L (2019) A novel hybrid stock selection method with stock prediction. Appl Soft Comput J 80:820–831. https://doi.org/10.1016/j.asoc.2019.03.028
Zhang F (2021) Extreme learning machine for stock price prediction. Int J Elect Eng Edu. https://doi.org/10.1177/0020720920984675
Acknowledgements
The author(s) disclosed receipt of the following financial support for the research, authorship and/or publication of this article: This work was supported by the Australian Government Department of Education and Training through the Australian Government Research Training Program (RTP) Scholarship to Volodymyr Novykov.
Funding
Open Access funding enabled and organized by CAUL and its Member Institutions.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix 1: Individual stock price ELM model predictions.
Appendix 1: Individual stock price ELM model predictions.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Novykov, V., Bilson, C., Gepp, A. et al. Empirical validation of ELM trained neural networks for financial modelling. Neural Comput & Applic 35, 1581–1605 (2023). https://doi.org/10.1007/s00521-022-07792-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-07792-3