1 Introduction

The problem of information present in the markets and of information that can be released at a company level has always been fundamental for understanding how a price is determined, as well as at a regulatory level. As early as 1945, Hayek [1] described how prices, in a system where relevant information is dispersed, can coordinate the actions of different subjects. In modern financial markets, investors benefit from the information available by selecting only what they believe is important.

In recent times, the amount of information available to investors has increased exponentially [2] thanks to the news provided by companies, creating a problem of optimal selection of the most important information compared to those that could only “deceive” the investor. Thus, an open question remains how much current information dissemination capabilities are able to impact price efficiency in incorporating new information.

The first forms of analysis of the price effects of information concern disclosures of corporate reporting [3]. The disclosure of information highlights managerial talent and is able to explain crisis situations [4, 5], or it is able to reduce the information asymmetry by preventing investors from exchanging private information [6]. In contrast to this view, the increase in complexity in the information provided could lead to problems with the legibility of a news [7], or it could be used to obfuscate negative news [8]. Some studies [9,10,11] examine the effectiveness of numerical and textual information on price discovery, highlighting how the combination of high levels of numerical information combined with graphic information can have a great impact [12] on investors and on the choice of price; while in this case we face the inverse situation, that is to understand from the prices what is the level of information contained and the consequences that this can have on the forecast. This is where this paper focuses, in investigating the market hours problem associated with financial information. However, our starting point is not to consider textual information (deriving for example from company statements), but to use only the prices recorded on the markets to understand what level of information these prices embody. In particular, this consideration is carried out over small time intervals to highlight the difference between closing markets and always open markets.

Many markets are subject to opening and closing times. In such types of markets, when news spreads or events occur after closing hours, price reactions can only occur after the next opening of the market. In contrast, the cryptocurrency market (and currencies in general) is not subject to closing times. In this type of market, the “opening” price of the new day and “closing” price of the previous day are recorded at midnight each day, creating continuity in the historical price series. Thus, the recorded prices are assumed to be the sum of events occurring in a 24-h period. Therefore, these prices may contain more information useful for forecasting. The objective of this study is to verify how this difference of information in price can result in forecast imbalances when using an appropriate neural network. We focus on the discriminative and predictive power of prices. Previous studies, such as those of Tang et al. [13] and Gorr [14], have demonstrated that neural networks can model seasonality and other factors, e.g., trends and cyclicity, autonomously. Therefore, different “quantities of information” contained in various types of prices would seem to be the only cause resulting in forecast imbalances. Datasets with different temporal structures would contain different amounts of information, resulting in differences with respect to the degree of predictability related to associative learning tasks [15].

The remainder of this paper is organized as follows. In Sect. 2, we present some of the most important literature in the sector, in Sect. 3, we analyze the architecture of Generative Adversarial Network (GAN). Herein, we focus on TimeGAN, which is used to extrapolate the characteristics of different features in time-series data. In Sect. 4, we define our model and methodology. In Sect. 5, we apply this special GAN to the time-series data of stocks and cryptocurrencies and compare the obtained results, extending the previous datasets by adding a feature and verifying the results. Finally, conclusions are presented in Sect. 6. In “Appendix”, the most used neural network architectures to make financial market predictions, Long–Short Term Memory (LSTM) and Convolutional Neural Network (CNN), are presented.

2 Literature review

Moving away from works in which give the information we try to understand what the impact on the price is, finding ourselves in the diametrically opposite situation, we can consider only time series, whose analyzes has always attracted the attention of academia, especially for predicting the future values in a series. Financial time series are optimal candidates for such an analysis. They base their assumptions on the random walk hypothesis, a concept introduced by Bachelier [16] in 1900, which has remained the central pivot in the theory of time series. Based on the random walk hypothesis, Kendall [17] assumed that the stock price movement was random, whereas Cootner [18] indicated how the stock price movement could not be explained in detail but could be better approximated based on the Brownian motion. Traditionally, the best practice has been to focus on logarithmic returns, which provides the benefit of linking statistical analysis with financial theory. In his efficient market hypothesis (EMH) theory, Fama [19] introduced a concept that historical prices are factored into the current prices in a given market. However, deploying such historical data in any analysis would be less useful (if not completely useless) for predicting future prices. LeRoy [20] demonstrated that the concentration on yields was unjustified and concluded that the stock markets are inefficient. Taylor [21] proposed an alternative price trend model and provided empirical evidence that the price trend model was useful for analyzing the future prices in markets.

From an econometric perspective, Box and Jenkins [22] introduced power transformations to statistical models and applied them to a time series. Specifically, they suggested to use power transformation for obtaining an adequate Autoregressive Moving Average (ARMA) model. Based on this, Hamilton [23] provided a formal mathematical definition of this model. Several evolutions have followed this pattern, e.g., Autoregressive Integrated Moving Average (ARIMA) and Seasonal Autoregressive Integrated Moving Average (SARIMA).

2.1 Machine learning

Recently, because of the developments of Artificial Neural Networks (ANNs) and their suitability in nonlinear modeling [24], there has been considerable interest in applying such methods to time-series prediction in the Machine Learning (ML) framework. For example, Foster et al. [25] were among the first to compare the use of neural networks as function approximators with the use of the same networks to optimally combine classical regression methods. Further, they highlighted the manner in which using networks to combine forecasting techniques provided performance improvements. In addition, Refenes et al. [26] used a neural network system to forecast the exchange rates via a feedforward network, which faced difficulty in predicting the turning points despite an accuracy of 66%. Sharda and Patil [27] compared the predictions obtained using neural networks and the Box–Jenkins model. Based on this comparison, they verified that neural networks performed better than expected for time series with long memory. In contrast, the neural networks outperformed the prevision for time series with short memory. Andrawis et al. [28] combined the forecasts obtained via different time aggregation, and Adeodato et al. [29] proposed a methodology based on an ensemble of Multilayer Perceptron networks (MLPs) to achieve robust time-series forecasting. In addition, Wichard [30] proposed a hybrid model to forecast a time series with recurring seasonal periods using separately trained models. A considerable innovation was introduced in the support vector machine (SVM) model [31], thereby solving the pattern classification problem. In particular, its use was immediately extended to regression, with subsequent application to time-series forecasting [32]. Over time, several SVM-based models have been developed, e.g., the least-square SVM (LS-SVM) [33]. More recently, Xiao et al. [34] used a cumulative auto-regressive moving average which combines the least squares Support Vector Machine model (ARI-MA-LS-SVM), to make basic predictions for the stock market; highlighting how this model is more effective for stock price forecasting than a classical forecasting model. In addition, Kovalerchuk and Vityaev [35] used various models, such as the Evolutionary Computation (EC) and Genetic Programming (GP) models, Zaccagnino et al. [36, 37] developed automatic ML-based methods to provide privacy awareness to users and total control over their data during online activities, and Li and Ma [38] developed a model to forecast the stock price using an ANN. Mittelmayer and Knolmayer [39] compared different text mining techniques for extracting market response to improve prediction; and Mitra [40] focused on studying news to predict the anomalous returns associated with trading strategies.

2.2 Deep learning

In case of Deep Learning (DL) techniques, increasingly complex architectures have been used, especially in the previous decade [41]. For example, Liu et al. [42] used a CNN-LSTM for strategic analysis in financial markets; Zhang et al. [43] used structure for motion (SFM) to predict stock prices by extracting different types of patterns; Jin et al. [44] propose a decomposition for stock price by an Empirical Model Decomposition (EDM), subsequently used an LSTM with Attention Mechanism to improve prediction; or Lu et al. [45] that used a combination of CNN for extracting features, biLSTM to predict a stock price for the next day and the Attention Mechanism to capture the influence of features on the closing price. However, many other types of more complex networks can be readjusted to the time series to make predictions, e.g., GANs for speech synthesis [46], denoising images [47], or an imbalanced generative adversarial fusion network (IGAFN) to integrate credit data into an unified latent feature space, as demonstrated by Lei et al. [48]. In addition, other combinations of architectures are represented by the deep convolutional neural network (DCNN) and conditional random field (CRF) networks as proposed by Papandreou et al. [49] for high-resolution segmentation and a deep relative distance learning (DRDL) network as proposed by Liu et al. [50] to calculate the distance between vehicles via graphical analysis by translating them into the Euclidean space. Finally, this methodology that we intend to apply to the financial sphere about quantities of information, can be implemented in other promising scenarios, e.g., Blockchain [51], R&D [52], or supply chain [53].

3 Neural network architectures

The most generally used network architectures in financial time series forecasting are Long–Short Term Memory (LSTMs), particularly for the ability of adapting to the features considered. However, more recently, this sector has opened up to new architectures such as Convolutional Neural Networks (CNNs, whose use is justified by considering features no longer as mere time series but, in many cases, images) and Generative Adversarial Networks (GANs). A more accurate description of how LSTM and CNN work can be found in “Appendix”. Below, however, we will focus only on GANs (and on a particular version) to highlight the amount of information available on the markets.

3.1 Generative adversarial network

Generative models are formulated based on an approach developed in accordance with the Bayes’ theorem. Generative models consider sensorial hypotheses about the input to modify the parameters characterizing them. The learning mechanism involves maximization of the likelihood of data relative to the generative model; this corresponds to the discovery of efficient methods for encoding the input information. For financial time-series forecasting, the most common generative model is Generative Adversarial Network (GAN), which was introduced by Goodfellow et al. [54] in 2014. The GAN model comprises two networks, i.e., a generative network G that produces new data based on a certain distribution \(p_g\) and a discriminative network D that evaluates them, resulting the probability of \(x \sim p_\mathrm{data}\), where \(p_\mathrm{data}\) indicates the distribution of training data. The network’s objective [46] is to encourage D to find a binary classifier that provides optimal discrimination between the real and generated data and simultaneously encourage G to fit the true data distribution. D and G play the following two-player minimax game [54] with the value function V(GD).

$$\begin{aligned} \min _G \max _D V(D, G)&= {\mathbb {E}}_{x \sim p_\mathrm{data}(x)} [\log D(x)] \nonumber \\&\quad + {\mathbb {E}}_{z \sim p_z(z)} [\log (1-D(G(z)))] \end{aligned}$$
(1)

This type of neural network is one of the most complex and is generally used in complex sectors. Yoon et al. [55] adopted a version of it to accept time-series inputs and generate synthetic data.

3.1.1 TimeGAN

One of the main problems associated with time-series forecasting is the selection of optimal variables such that the neural network can capture their links and dynamics over time. In particular, Yoon et al. [55] proposed the Time-Series Generative Adversarial Network (TimeGAN) to generate realistic time-series data in various domains. TimeGAN considers unsupervised adversarial loss and stepwise supervised loss and uses the original data for supervision. TimeGAN comprises four networks [55]: the embedding function, the recovery function, the sequence generator, and the sequence discriminator. The autoencoding components are trained jointly with the adversarial components such that TimeGAN simultaneously learns to encode features, generate representations, and iterate over time. Typically, GANs are used (regarding a financial time series) for the generation and replacement of any missing values (NaN). However, in this case, the main objective is to recreate a time series based on the features considered as input.

In this model, the generator is exposed to two types of inputs during training, i.e., synthetic embeddings to generate the next synthetic vector and sequences of embeddings of actual data for generating the next latent vector. In the first case, the gradient is computed based on unsupervised loss, whereas the gradient is computed based on supervised loss in the second case.

4 Methods and materials

To highlight the generated results, Yoon et al. [55] proposed a graphical measure for visualization, i.e., t-SNE (van der Maaten and Hinton [56], to visualize the similarity of the generated distribution with respect to the original distribution) and two scores (obtained by optimizing a two-layer LSTM).

  • Discriminative score: This factor indicates the error of a standardized classifier (RNN) when distinguishing the real sequence and the sequence generated based on a test set using a post-hoc RNN network.

  • Predictive score: This factor evaluates the prediction performance of synthetic data. Here, a post-hoc RNN architecture is employed for prediction. Further, the performance is reported in terms of mean absolute error (MAE), which measures the ability of synthetic data to predict the next-step temporal vectors. For example, the MAE between paired observations is defined as

    $$\begin{aligned} \mathrm{MAE} ={\frac{\sum _{i=1}^{n}|y_{i}-x_{i} |}{n}}. \end{aligned}$$
    (2)

In addition, the t-SNE algorithm also returns Kullback–Leibler (KL) divergence

$$\begin{aligned} D_{\mathrm {KL} }(P\Vert Q)=\sum _{i}P(i)\log _{2}\left( {\frac{P(i)}{Q(i)}}\right) , \end{aligned}$$
(3)

which is a measure of the difference between two probability distributions P and Q [57] that indicates the information lost when using a distribution (in this case, the synthetic distribution) for approximating another distribution (the original distribution). We employ this network to demonstrate that the financial instruments listed on a market subject to time constraints have less predictive power than the instruments traded on a 24/7 market. The financial instruments that are subject to timetables during the continuous trading phase are representative of the information presented during those hours; however, the events that occur after closing is not reflected (immediately) by the price and will be recorded only on the following day during the opening auction. In contrast, for instruments not subject to schedules, this problem does not occur because any event that may affect the price will be recorded, affecting the price dynamics. Exchanges offer the possibility to conduct negotiations outside closing hours (trading in premarket and after hours), as in the case of Borsa Italiana, where the preauction phase is from 08:00 to 09:00 and after-hours trading is from 05:50 to 8:30 p.m., and NASDAQ, where premarket trading is from 04:00 to 09:30 (ET) and after-hours trading is from 4:00 to 8:00 p.m. (ET). However, certain time slots remain uncovered. Therefore, the “amount of information” associated with each price is an essential element for time-series forecasting. This hypothesis is confirmed by the fact that the corresponding financial instrument becomes more suitable for forecasting when considering a variable that links two successive days.

Here, GAN is used to generate synthetic data based on the input such that the network understands the links existing between data. This network can optimally identify the best variables (features) for forecasting by observing the improvements/reduction in the predictive score and by replacing different features in the datasets. Therefore, GAN can be used to validate financial instruments and variables based on which predictions can be subsequently conducted using any tool for obtaining the optimal result. In particular, many tools used to make predictions originate from econometrics. Thus, this type of network can be used to screen financial instruments such that the accuracy of the classical econometric tools is improved.

4.1 Dataset and experimental setup

Table 1 List of stocks and cryptos used
Table 2 Extract of the Apple price dataset without Yield
Table 3 Extract of the Bitcoin price dataset with Yield

Herein, empirical analyses of stocks and cryptocurrencies were conducted. However, first, we can consider different instruments because they are both time series and, therefore, have the same characteristics. The instruments used are four stocks and three cryptos listed in Table 1. All the stocksFootnote 1 are from NASDAQ, while the cryptocurrenciesFootnote 2 are related to USD. The related datasets were divided into two categories, i.e., classic variables (features) obtained from any site that tracks negotiations and a category in which the yield feature was added. Each dataset for stocks is about 500 days, while those for cryptocurrencies are from 700 days (due to the 24 h opening). The classical features used in the financial sector are Open (O), High (H), Low (L), Close (C), and Volume (V), which give the name to the corresponding datasets. For simplicity, we indicate with Close the Adjusted Close feature, which represents the price recorded at closing after some updates made by the Stock Exchange. Prices were considered based on a daily time frame from 12/20/2017 to 12/31/2019. Table 2, shows an extract of how the dataset is composed. Here, the Yield feature (indicate with (Y)) was determined as \(Y = \ln \frac{P_t}{P_{t-1}}\), where \(P_t\) represents the current stock price and \(P_{t-1}\) is the stock price of the previous day. Table 3 shows how the dataset changes with the addition of the new feature (the Close feature in cryptocurrencies is called Last). An immediately striking peculiarity is that in cryptocurrency datasets, the opening and closing prices are the same in most cases. This is linked to the 24 h opening described above since the registration of the closing and opening price of the following day take place with a few moments of difference. Previous datasets were again divided into several types to test the predictive ability of prices. In particular, the macro division concerns the use or not of the Yield feature and some combinations between the previous OHLCVs.

  • Dataset without yield:

    • OHLCV dataset, with five features: open, high, low, close (last in the case of cryptocurrencies), and volume;

    • OC dataset, with only two features: open and close (last).

  • Dataset with yield:

    • OHLCVY dataset, with six features (yield was added to the previous OHLCV).

    • OCY dataset, with three features (yield was added to the previous OC).

    • CY dataset, with only two features: close/last and yield.

In the “with yield” case, to test the predictive power of the Close feature (the price that is most considered in the financial field), we decided to combine it alone with Yield (CY dataset), which represents a summary of the trend between the previous day and the current one.

As previously defined, the TimeGAN we will use comprises 4 component networks, of which 2 are encoders and 2 decoders. These encoders and decoders, in practice, are based on Recurrent Neural Networks (RNNs) with LSTM cell. Figure 1 represents the structure of the TimeGAN instantiated with RNNs and their connections are visible. In this case, the \({\mathcal {L}}\) functions represent the losses; \(g_{{\mathcal {S}}, {\mathcal {X}}}\), \(r_{{\mathcal {S}}, {\mathcal {X}}}\), \(d_{{\mathcal {S}}, {\mathcal {X}}}\) and \(e_{{\mathcal {S}}, {\mathcal {X}}}\) represent the functions that operate for each type of network (e.g., g for generator); \(u_t\) and \(h_{{\mathcal {S}}, t}\) the hidden state sequences; while \({\tilde{s}}\), s, \({\tilde{x}}_t\), \(x_t\), \({\tilde{y}}_{{\mathcal {S}}}\), \({\tilde{y}}_t\), \(z_{{\mathcal {S}}}\), \(z_t\) are the random vectors extracted from a certain distribution. It is precisely the various g, r, d and e functions implemented with particular network architectures.

Fig. 1
figure 1

TimeGAN [55] structure instantiated with RNNs

Table 4 Parameter setting

Regarding the choice of the hyperparameters, the first choice was the module (cell), for which we opted for the LSTM, given its diffusion in the financial field and the excellent results obtained on data of this type. On the other hand, the hyperparameters contained in Table 4(a) were selected by choosing the configuration of Yoon et al. [55], as they also analyzed the financial case and proposed this choice as the best result. Only for seq_len (sequence length of time-series data) we have operated as choice 24, this to make it represent approximately one month of trading. Otherwise, the number of neurons in the RNN (the so-called parameter hidden_dim) was chosen variable based on the dataset type. To make a choice, we made, with each dataset, 10-fold cross-validation by using the GridSearchCV method (as proposed by Guarino et al. [58]). The number of neurons size was tested with values in the range between 2 and 40 (for even number), and the best results are those shown in Table 4(b) (as the values that minimize the KL divergence). Figure 2 summarizes how it works, highlighting a concatenation of LSTM cells whose input \(X_t\) is represented by the tensor constituted by the different datasets. Training and testing were performed using Google Colab.

Fig. 2
figure 2

LSTM cells concatenated in a RNN [59, 60] with an example of the input data

5 Results

At this point we can compare the behavior of the network in the different types of subdivisions created in the dataset.

5.1 Case 1: Dataset without yield

Here, we compare different datasets from a graphical perspective. Then, based on the scores, we compare the capabilities of different time series in terms of the difference with respect to the amount of information. In this comparison, we compared the OHLCV datasets with the OC datasets for all financial instruments beginning from the dataset without yield. The graphical analysis was performed by comparing the t-SNE plots, and score-based comparison was performed by comparing the KL divergence, discriminative score, and predictive score.

Fig. 3
figure 3

t-SNE analysis of synthetic data (blue) and original data (red) (OHLCV dataset) (color figure online)

The first t-SNE analysis considered the OHLCV dataset (Fig. 3). This analysis indicates the potential of TimeGAN relative to data generation. In the cases shown in Fig. 3b, f (both cryptocurrencies), there is very precise adherence of the synthetic data to the original real data. Obviously, with this type of dataset, greater features are combined, allowing the network to improve the forecast. Graphical analysis allows us to observe how synthetic data extend over a greater surface than the original data in the case of cryptocurrencies  Fig. 3b, d, f compared with the stocks (where they are much more concentrated).

The second t-SNE analysis was based on the OC dataset, as shown in Fig. 4.

Fig. 4
figure 4

t-SNE analysis of synthetic data (blue) and original data (red) (OC dataset) (color figure online)

In this case, at first glance, synthetic data are observed to be different from original data. However, careful analysis indicates that in Fig. 4a, c, f (all cryptocurrencies), synthetic data are similar to the original data or imitate (to a limited extent) their distribution. In case of stocks, the distribution of synthetic data is dispersive and not entirely consistent with the distribution of original data.

Here, the hypothesis is that the prices of cryptocurrencies contain a greater amount of information and have greater discriminative and predictive power. This hypothesis is supported based on a graphical analysis. However, to eliminate doubt, we introduce the results of analysis based on discriminative and predictive scores, where \(D_s\) denotes the discriminative score, \(P_s\) denotes the predictive score, and \(D_{KL}\) denotes the KL divergence, as shown in Table 5. The most important score is the predictive score (corresponding to MAE) because it specifies the ability to use the input data for making predictions, whereas the discriminative score specifies the ability of data to deceive the GAN discriminator.

Table 5 Scores of both datasets (the lower the value, the better the result)

By analyzing the values, it can be noticed how the cryptocurrency scores were the lowest and therefore the most significant, especially in the case of the OC dataset.

We also assume that in markets subject to opening/closing time, investors are struck by “euphoria” by being trapped in a bottleneck and going to generate a price that may not be representative of the real trend, especially close to the closing minutes. In contrast, such situations may not exist in 24/7 markets because such markets are always open.

5.1.1 Outliers

We observe a particular situation when analyzing the stocks of Tesla Inc. (TSLA) listed on NASDAQ. Despite being listed on a market subject to timetables, this analysis using TimeGAN (both graphically and based on scores) resulted in a price type containing plenty of information such that it achieved almost better results than cryptocurrencies (Fig. 5).

Fig. 5
figure 5

t-SNE analysis of the TSLA stock on different datasets

In the second dataset, the synthetic data reproduced the distribution of the original data very well. The results of the score analysis are presented in Table 6.

Table 6 Scores of TSLA in both datasets

In this case, especially in the OC dataset, the price range indicated very low discriminative and predictive scores, even lower than that associated with cryptocurrencies. Because of this “outlier”, we deduce that some financial instruments listed on stock exchanges are subject to timetables, which can completely absorb information despite the above limitation. This situation could be linked to, for example, the hypothesis that negative events never occurred outside the opening hours of the stock exchange or (in a less realistic but still possible hypothesis) that no external situations occurred in the considered time range that could influence the price. In these cases, the use of this type of GAN can be a “form of control” on prices, especially when they are to be used for forecasting. In addition, there may be hidden elements that affect the price. However, we assume that a financial instrument whose price has a “large amount of information” could result in improved prediction power when compared with a financial instrument with less information.

5.2 Case 2: Dataset with yield

Here, as in the previous case, we compare the financial instruments from a graphical perspective relative to the discriminative and predictive scores in reference to the datasets with the yield feature. We included the yield feature and verified its usefulness with respect to the amount of information. The hypothesis we seek to validate is that yield can link the information of the current day with the previous day, eliminating (or at least attenuating) the information gap created in markets subject to closing/opening times, e.g., stocks. In addition, under this new condition, the Tesla stock is no longer regarded as an outlier (the proof is presented later).

Fig. 6
figure 6

t-SNE analysis of synthetic data (blue) and original data (red) (OHLCVY dataset) (color figure online)

The first t-SNE analysis considered the OHLCVY dataset, as shown in Fig. 6. From the graphical perspective, it can be observed how the situation changed compared to the previous datasets (without yield). In case of stocks (Fig.  6a, c, e, g), synthetic data adhere nearly perfectly to the original real data and cover a more extensive surface on them unlike cryptocurrencies, which extend over more “uncovered” areas.

Fig. 7
figure 7

t-SNE analysis of synthetic data (blue) and original data (red) (OCY dataset) (color figure online)

The second t-SNE analysis was based on the OCY dataset, as shown in Fig. 7. In this case, despite the small differences between the two types of financial instruments, the data generated by stocks overlap better with the original data when compared with those generated by cryptocurrencies even though the covered areas are limited in both cases.

Fig. 8
figure 8

t-SNE analysis of synthetic data (blue) and original data (red) (CY dataset) (color figure online)

Finally, the third t-SNE analysis was conducted based on the CY dataset, as shown in Fig. 8. In the latter case, the best performances were obtained in the Bitcoin Cash Fig. 8f and Google Fig. 8e cases, where the ability of cryptocurrencies to be more predictive was eliminated. The Tesla stock Fig. 8g presented original data that can be classified as outliers, making t-SNE less readable. However, it was possible to observe how the synthetic data can cover the rest of the originals very well.

In support of the graphical analyses, the scores shown in Table 7 demonstrate how the yield feature improved the information potential of stocks.

Table 7 Scores of both datasets (lower scores are better)

On an average, the performance allowed GAN to achieved improved predictive score for stocks and discriminative score (especially in the OCY and CY datasets). However, this feature, which was also associated with the dataset of cryptocurrencies, resulted in a worse performance compared to the datasets in which it was not present, confirming that TimeGAN is an excellent prediction tool. The Tesla stock was no longer an outlier because it obtained very similar scores compared to the remaining stocks. The yield feature, which is commonly used in financial analysis, made it possible to eliminate the information gap between cryptocurrencies and stocks.

6 Conclusion

In this paper, we demonstrated that TimeGAN can identify which financial instruments have a time series of prices containing abundant information. First, the prices of cryptocurrencies were observed to have much higher discriminatory and predictive powers than stocks, especially in the dataset comprising only the opening and closing prices. In addition, in case of the complete OHLCV dataset, prices with high discriminative power (combined with the remaining features) made it possible to significantly improve the adherence of the synthetic data with original data. From this analysis, we observed that some stocks have the same discriminative and predictive power as cryptocurrencies. Thus, because the time-series forecasting is primarily performed on stocks, this neural network can be used to screen the optimal titles, which combined with different features, improve the forecasting procedure. Second, by adding the yield feature, the previous situation can be modified to obtain datasets with stock prices as predictive as cryptocurrencies (or more in some cases). In future, we plan to investigate other features based on which the information capacity of various financial instruments may be improved to mitigate errors in forecasts.