Generating probabilistic forecasts from arbitrary point forecasts using a conditional invertible neural network

In various applications, probabilistic forecasts are required to quantify the inherent uncertainty associated with the forecast. However, many existing forecasting methods still only generate point forecasts. Although methods exist to generate probabilistic forecasts from these point forecasts, these are often limited to prediction intervals or must be trained together with a speciﬁc point forecast. Therefore, the present article proposes a novel approach for generating probabilistic forecasts from arbitrary point forecasts. In order to implement this approach, we apply a conditional Invertible Neural Network (cINN) to learn the underlying distribution of the data and then combine the uncertainty from this distribution with an arbitrary point forecast to generate probabilistic forecasts. We evaluate our approach by generating probabilistic forecasts from multiple point forecasts and comparing these forecasts to six probabilistic benchmarks on four data sets. We show that our approach generally outperforms all benchmarks with regard to CRPS and Winkler scores and generates probabilistic forecasts with the narrowest prediction intervals whilst remaining reasonably calibrated. Furthermore, our approach enables simple point forecasting methods to rank highly in the Global Energy Forecasting Competition 2014.


Introduction
Probabilistic forecasts are required to quantify the inherent uncertainty associated with any prediction of the future [23,45].These probabilistic forecasts are crucial for many applications such as stabilising energy systems [11], managing congestion in traffic systems [39], or sizing servers of web applications to cope with a certain number of daily visits [36].Despite this necessity for probabilistic forecasts, many modern forecasting methods still generate point forecasts [44].Although many recent machine learning libraries offer support for probabilistic loss functions to simplify the generation of probabilistic forecasts, this may not be possible if an existing point forecast model that cannot easily be modified or retrained is already in use.
One solution to overcome this challenge is to generate probabilistic forecasts based on these existing point forecasts.For many years, such forecasts have been generated B Kaleb Phipps kaleb.phipps@kit.eduExtended author information available on the last page of the article by analysing the residual errors of the point forecast.Based on these errors' standard deviation or quantiles, prediction intervals can be calculated to generate probabilistic forecasts [29,56].Moreover, such probabilistic forecasts can be generated by using machine learning methods exploiting the residual errors [5,54], by applying the Bayesian theory of probability to a point method [37], or by considering Monte-Carlo sampling methods [3].Although these methods may be effective, they also have various limitations.For example, the prediction interval-based approaches can only generate prediction intervals as probabilistic forecasts, while machine learning methods depend on the point forecast and must be retrained if the point forecast is altered.Ideally, such probabilistic forecasts should be generated directly from arbitrary point forecasts and should not require retraining if the point forecast changes.
Therefore, in the present article, we present an approach that generates probabilistic forecasts from arbitrary point forecasts by using a Conditional Invertible Neural Network (cINN) to learn the underlying distribution of the time series data.Since time series have an inherent component of randomness [29], we propose using this uncertainty within the distribution of the time series data to generate probabilistic forecasts.However, the underlying system responsible for this uncertainty typically generates observations of an unknown probability distribution.Therefore, with our approach, we first map this unknown probability distribution of the underlying time series data to a known and tractable distribution by applying a cINN.Then, we use the output of a trained arbitrary point forecast method as an input to the trained cINN and consider the representation of this forecast in the known and tractable distribution.We then analyse the neighbourhood of this representation in the known and tractable distribution to quantify the uncertainty associated with the representation.Finally, we use the backward pass of the cINN to convert this uncertainty information into the forecast.In our approach, the cINN is trained independently of the point forecast and must not be retrained when the point forecast is altered.
Thus, the main contribution of the present article is twofold.First, we provide a novel approach for generating probabilistic forecasts from arbitrary point forecasts whose training is independent of the point forecast.Second, we empirically evaluate the approach using different data sets from various domains.In this empirical evaluation, we compare our approach to six probabilistic benchmarks, evaluate multiple metrics, and recreate the Global Energy Forecasting Competition 2014 (GEFCom2014) competition setting.
The remainder of our article is structured as follows.First, we present related work and highlight the research gap that the present article addresses in Section 2. In Section 3, we then explain our approach in detail and highlight how we use a cINN to generate probabilistic forecasts from an arbitrary point forecast.We detail the experimental setup in Section 4, before presenting our results in Section 5.In Section 6 we discuss our evaluation and key insights.Finally, we conclude and suggest possible directions for future work in Section 7.

Related work
Our article is closely related to two research fields: previous work that generates probabilistic forecasts based on point forecasts and previous work focusing on probabilistic forecasts using a cINN.Table 1 presents an overview of the identified related articles, and in this section, we present and discuss these articles in more detail and highlight the research gap the present article addresses.

Generating probabilistic forecasts from point forecasts
Determining the uncertainty associated with a point prediction is one of the key research areas of uncertainty quantification [52].Many methods focus on generating probabilistic prediction intervals from existing point forecasts by using the residual errors between the point forecast and the true value [29].These prediction intervals can be generated by assuming a Gaussian distribution of the errors [29], using the empirical distribution of the errors [56], or considering nonconformity errors [8,53,59].While effective, these methods are designed to generate prediction intervals rather than approximate the full probability distribution, which may be a limitation.Furthermore, if the point forecaster used is changed, new residual or nonconformity errors must be calculated to apply these methods.Although this calculation is not a retraining process, it does require additional effort.
Similar approaches also use residual errors in combination with further machine learning algorithms.[5], for example, train a neural network to forecast the standard deviation of the residual errors and generate probabilistic forecasts as realisations of a Gaussian distribution centred around the original point forecasts.Similarly, [54] use the residual errors from a point forecast to train a Generative Adversarial Network (GAN).This trained GAN is then used to generate multiple residual scenarios, which are combined with the point forecast to form probabilistic forecasts.The main limitation of both approaches is that the additional machine-learning models used to predict the uncertainty (i.e. standard deviation or residual scenarios) depend on the selected point forecast [5,54].Therefore, these machine-learning models must be retrained whenever the point forecast is altered.
Further approaches include a Bayesian method involving assumed priors [32], integrating uncertainty into the prediction via an ensemble of predictions [10], and considering uncertainty through Monte Carlo sampling approaches or similar [4].The main limitation of these approaches, apart from the assumption regarding the Bayesian prior, is the computational complexity resulting from sampling or generating a large ensemble pool.

Probabilistic forecasts using cINNs
To generate probabilistic forecasts, cINNs, also referred to as normalising flows [1], are combined with other machine learning methods.[2], for example, apply normalising flows to learn the parameters of Bernstein polynomials, which are, in turn, used to generate a probabilistic forecast.Moreover, [46] combine normalising flows with recurrent neural networks to generate probabilistic forecasts.Normalising flows are also combined with quantile regression networks and copulas [55], or used to generate a conditional approximation of a Gaussian mixture model [31] to improve the accuracy of the resulting probabilistic forecasts.Whilst these methods are all effective, normalising flows are used to enrich existing complex probabilistic forecasting methods, but not to provide probabilistic forecasts themselves.
An alternative method that directly uses normalising flows in the context of probabilistic forecasts is to learn multi-dimensional distributions of electricity price differences to predict the trajectory of intraday electricity prices [9].Similarly, normalising flows may be applied multiple times to generate scenario-based probabilistic forecasts [15,21,60], or to generate a proxy for weather ensemble prediction systems based on numerical weather prediction models [18].These methods use the generative nature of normalising flows to generate multiple predictions drawn from the same distribution.However, the forecasts are only probabilistic as an ensemble, with each individual forecast still being a point forecast.Furthermore, these forecasts assume that the underlying learned distribution remains constant and only partly considers external features.Finally, these methods all focus on directly generating probabilistic forecasts.Therefore, such methods cannot be applied to generate probabilistic forecasts from existing, well-designed point forecasts.

Research gap
As shown in Table 1, we identify a lack of existing work that directly generates probabilistic forecasts from arbitrary point forecasts without the training process being dependent on this point forecast or being limited to only generating prediction intervals.In the present article, we aim to fill this research gap by presenting an easy-to-use approach described in the following section.

Generating probabilistic forecasts with a cINN
To generate probabilistic forecasts from arbitrary point forecasts, we directly apply the uncertainty in the underlying time series.This uncertainty usually reflects the inherent randomness or unpredictability of the measured underlying system.However, this underlying system typically generates observations of an unknown distribution.Although this data is not random, the distribution is still unknown, and it is challenging to include the corresponding uncertainty directly in a forecast.
To solve this challenge, we aim to find a bijective mapping from the unknown distribution to a known and tractable distribution.Since many time series are affected by exogenous features such as weather, this bijective mapping should also be able to consider such exogenous features, as shown in Fig. 1.If such a mapping g exists, we will be able to map a point forecast from the unknown distribution to its representation in a known and tractable distribution.In the known and tractable distribution, we could then analyse the neighbourhood of this representation and gain information about its uncertainty.Finally, we could map this uncertainty information back to the unknown distribution using the inverse mapping g −1 to generate probabilistic forecasts.Fig. 1 Overview of the proposed approach.In the first step, an arbitrary point forecaster (a) and cINN (b) are trained independently, both considering exogenous features, past observations, or past historical data.To generate a probabilistic forecast, as shown in (c), the arbitrary point forecaster first generates a point forecast based on historical data and exogenous features.The resulting point forecast is combined with the exogenous features as inputs to a bijective mapping realised by the trained cINN.This mapping generates a representation of the forecast in a known and tractable distribution.We analyse the neighbourhood of this known and tractable representation to gain information about its uncertainty.Finally, we map this representation back to the unknown distribution to generate a probabilistic forecast In this section, we demonstrate that this mapping g does exist under certain conditions, and we show how this mapping can be used to generate probabilistic forecasts.We then explain how to apply this approach with a cINN, starting with the training of this cINN, before describing how we generate probabilistic forecasts using arbitrary point forecasts.

Including uncertainty from the underlying distribution of the data
This section demonstrates that a bijective mapping from an unknown distribution in a known and tractable distribution exists.Given the existence of this mapping, we highlight the equivalence of the uncertainty in the image and the inverse image of the considered mapping.Finally, we describe how this mapping is realised with a cINN

Bijective mapping
To introduce the bijective mapping, let us consider a times series y = {y t } t∈T consisting of T observations as realisations of a random variable Y ∼ f Y (y) with a probability density function (PDF) f Y (y) in the realisation space Y. Furthermore, we consider a bijective mapping g : Y → Z from the realisation space Y to the space of the tractable distribution Z where y → g(y, •) = z, and g is a continuously differentiable function. 1To calculate the PDF f Z (z) in terms of f Y (y), we can apply the change of variables formula [12,42], i.e.
where ∂ g −1 ∂z is the Jacobian matrix.Since g is bijective, this equation describes a bijective mapping from the unknown distribution f Y (y) to the known and tractable distribution f Z (z).Therefore, the change of variable formula provides us with the required mapping.

Equivalence of uncertainty
After introducing the bijective mapping, we need to show the equivalence of the uncertainty in the unknown distribution and known tractable distribution when applying (1).More specifically, we show the equivalence of quantiles in both the realisation space and the tractable distribution space since quantiles serve as a non-parametric representation of the uncertainty.
To show this equivalence, we first consider the cumulative distribution function (CDF) of the random variable Z = g(Y , •) ∼ f Z (z), defined as ( If we use the expression for f Z (z) from the change of variables formula (1) in the definition of the CDF (2), we obtain describing the CDF of F Z (z) in terms of the CDF F Y (y).
Since g is, per definition, a continuously differentiable function, we can apply integration by substitution for multiple variables to rewrite (3) as which is simply the CDF of Y evaluated at the inverse of g.Further, the quantiles z α of Z are defined by the inverse of the CDF, i.e.
where inf refers to the infimum, the smallest value of z that fulfils the condition, and α ∈ [0, 1] is the considered quantile.Consequently, if we know that the α quantile of F Z is z α , then we can also calculate the α quantile of F Y as g −1 (z α , •) = y α .From this follows an equivalence between the quantiles of Z and the quantiles of Y , which implies an equivalence in the uncertainty.
Given the mathematical equivalence of the uncertainty in the two considered distributions, we can include uncertainty in a tractable and known distribution f Z (z) and use the inverse mapping g −1 : Z → Y to map this uncertainty to the original distribution f Y (y).

Realising the bijective mapping
To realise this bijective mapping g, we use a cINN [1,27].A cINN is a neural network that consists of multiple specially designed conditional affine coupling blocks [1].As shown by [1], these coupling blocks ensure that the mapping g : Y → Z learnt by the cINN is bijective.Furthermore, with the conditional information, the cINN is able to consider additional information, such as exogenous features, extracted statistical features from the time series, or calendar information, when learning the mapping [1,27].As a result, the cINN is designed to learn an approximation of f Z (z) and a mapping g, which is per definition bijective, thus ensuring we can apply (1) as described previously.Since cINNs are specifically designed to efficiently calculate the inverse of a function [1], a well-trained cINN should be capable of learning the bijective mapping g, even if this mapping is nontrivial.

Applying our approach
In the following, we describe how we realise the inclusion of uncertainty with a cINN. 2 We first detail how we train a cINN that learns the distribution of the underlying data.Second, we describe how we use this trained cINN to generate probabilistic forecasts.

Training
To apply our approach, firstly an arbitrary base point forecaster F(•, ψ) with trainable parameters ψ must be trained, as shown in Fig. 1a.However, since the cINN is trained on historical time series realisations and not with the point forecasts, the cINN is trained completely independently of the point forecaster and must not be retrained if the point forecast is altered.Furthermore, the cINN in our approach can be applied to any arbitrary point forecast, including a point forecast that has been previously trained and implemented.Since these arbitrary point forecasters may have very different training procedures, we refrain from a more detailed description of the training of the base point forecaster in the present article.As a result of this training, however, the base point forecaster F(•, ψOPT ) has optimised parameters ψOPT .
We use a cINN to realise the continuous differentiable function g described previously.In addition to the original realisation y, we also consider conditional information c as an input to the function g.This conditional information always includes calendar features such as time of the day, and day of the week, but depending on the time series may also include additional exogenous features that are available for the forecast period.Thereby, the calendar information extracted from the time series is necessary conditional information to account for the temporal dependencies of the time series, whilst the exogenous features are optional.Furthermore, statistical features extracted from the time series can also be included as conditional information.We train the cINN using past exogenous features and past observations, Algorithm 1 Forecasting with our approach.
Trained base point forecaster, cINN, and selected σ ŷ ← F (y hist , c, ψOPT ) ẑ = g(ŷ, c, θOPT ) Ŷσ ← Initialise empty set to store samples for i ∈ {1, . . ., I } do r i ∼ N (0, σ ) ← Initialise random noise for sampling zi ← ẑ + r i ỹi ← g −1 (z i , c, θOPT ) Ŷσ ← Ŷσ ∪ ỹi end for ŷprob ← CalculateQuantiles( Ŷσ ) return ŷprob as shown in Fig. 1b.The aim of the training is to ensure that the cINN learns the function g, so that resulting realisations z = g(y, c) follow a known and tractable latent space distribution f Z (z).In our approach, we define this known and tractable latent space distribution as a multi-dimensional Gaussian distribution, where the number of dimensions is equal to the forecast horizon.Therefore, we apply the change of variables formula to derive the loss function where J = det(∂ g/∂y) is the determinant of the Jacobian, θ is the set of all trainable parameters, and λ θ 2 2 is an L2 regularisation [1,27]. 3Training a cINN with this loss function results in a network with the optimised parameters θOPT and ensures that the realised latent space distribution f Z (z) achieves the best possible approximation of the desired multi-dimensional Gaussian distribution [1].

Forecasting
The process of generating probabilistic forecasts with our approach is shown in Algorithm 1.The process begins with the output of the trained base point forecaster We then combine this output with the associated conditional information c and pass it through the trained cINN to obtain a latent space representation of the output, i.e.

ẑ = g(ŷ, c, θOPT ).
Given this latent space representation of the point forecast, we explore the uncertainty in the neighbourhood of the forecast with 3 Full details on the derivation of this loss function are presented in [1].
Using (4), we select a random noise r i from a standard normal distribution with mean 0 and variance σ and add this noise to the realisation ẑ.We define the variance used for the sampling process σ as the sampling hyperparameter, which must be selected in advance before generating a probabilistic forecast.Due to the equivalence of uncertainty in both spaces shown in Section 3.1, we can process this perturbed sample via a backward pass of the cINN, i.e.
to obtain a perturbed sample in the realisation space ỹi .Based on the selected σ , we repeat the sampling process I times to obtain multiple realisations of zi and, in turn, multiple realisations ỹi that are all similar but not identical to the original forecast.If we combine all these samples in a set Ŷσ , i.e.
then this set of realisations provides a representation of the uncertainty in the neighbourhood of the forecast, as schematically shown in Fig. 2 on the left.Given this set, there are multiple possibilities for generating a probabilistic forecast.
We can use all the samples as an ensemble forecast, perform a density estimation over the samples to generate a distribution forecast, or calculate the quantiles of these samples.In the present paper, we calculate the quantiles of these samples in the original realisation space as schematically shown in Fig. 2 on the right.The resulting quantiles represent a probabilistic forecast ŷprob , derived from the original arbitrary point forecast.

Experimental setup
This section describes the experimental setup we use to evaluate our approach.We first introduce the data used, before explaining the evaluation metrics.Furthermore, we describe the selected base forecasters used to generate the point forecasts, introduce the benchmarks we compare our approach to, and detail the implementation of the used cINN.

Data
We evaluate our proposed approach on four different openly available data sets.In this section, we briefly introduce each of these data sets before we describe their preprocessing.The first considered data set is Electricity, namely the UCI Electricity Load Dataset4 [13].From this data set, we select Fig. 2 A schematic representation of the probabilistic forecast generated with our approach.Initially, a set of samples represents the uncertainty.In the present paper, we then calculate the quantiles of these samples to generate the probabilistic forecast ŷprob .However, it would also be possible to consider all samples as an ensemble forecast or to perform density information to obtain a distribution forecast the time series MT_158 and resample it to an hourly resolution.
The second data set, Price, contains zonal electricity price data recorded at a single location at an hourly resolution and taken from the electricity price track of the GEFCom2014 [28].To evaluate our approach on a period longer than a single day, we combine data from all tasks in the GEFCom2014 price track.
Third, we consider a Solar data set which contains hourly real-world solar power generation from a solar plant in Australia.This data set is taken from the solar power forecasting track of the GEFCom2014 [28] and, again, we combine data from all tasks to enable evaluation on a period longer than a day.
The fourth considered data set, Bike, contains hourly records of rented bikes from the UCI Bikesharing Dataset [13,17]. 5 We normalise each of the above data sets before creating separate test, validation, and training subsets for the training and testing of our approach.An overview of these splits and the exogenous variables considered for each data set is presented in Table 15 in Appendix A.

Evaluation metrics
When evaluating probabilistic forecasts, it is important to consider both sharpness and calibration [23].According to [23], probabilistic forecasts should aim to maximise sharpness subject to calibration.This aim implies, for example, that a prediction interval should be as narrow as possible while still maintaining coverage close to the nominal coverage rate.Probabilistic forecasts that are too sharp provide misleading information about the uncertainty present, whilst probabilistic forecasts that only focus on calibration may not be sharp enough to deliver any useful information [23].With these considerations in mind, we aim to evaluate our approach comprehensively, considering both sharpness, calibration, and the trade-off between these two, and therefore 5 https://archive.ics.uci.edu/ml/datasets/bike+sharing+datasetconsider multiple evaluation metrics.In the following, we briefly present these metrics in the order they appear in the evaluation.

Continuous ranked probability score
To evaluate the quality of the probabilistic forecasts, we consider the Continuous Ranked Probability Score (CRPS) [41].The CRPS is a proper scoring rule that measures both the calibration and sharpness of a predictive cumulative distribution function F [22].The CRPS is defined as where 1{y ≤ z} is the indicator function which is one if y ≤ z and otherwise zero.Since our approach and the benchmarks provide samples drawn from a distribution, we use the sample-based variant of the CRPS implemented in the properscoring library.6

Quantile deviation
To analyse the calibration of our forecasts, we consider the deviation of the forecast quantiles from the theoretical quantiles.We define the quantile deviation for the α-quantile as where ŷi,α is the α-quantile forecast, y i the true value, and 1 the indicator function.Ideally, QD α should be zero for all values of α.However, a positive value indicates the quantile forecast overestimates the theoretical quantile, whilst a negative value indicates that the quantile forecast underestimates the theoretical quantile.To account for the total quantile devi-ation across all considered quantiles α ∈ Q, we calculate the Mean Absolute Quantile Deviation (MAQD) defined as

Normalised prediction interval width
To measure the sharpness of the probabilistic forecasts, we consider the normalised Mean β-PI Width (nMPI(β)).The nMPI(β) is defined as where ŷi, 1+β 2 is the predicted upper quantile, ŷi, 1−β 2 the predicted lower quantile for the forecast value ŷi , and ȳ the mean of the target time series.We consider the nMPI(β) to enable a comparison between different data sets.

Winkler score
To jointly assess calibration and sharpness, we assess the quality of the prediction intervals of our probabilistic forecasts with the Winkler score [57].As defined by [29], if the 100 then the Winkler score for the α-quantile is defined as where y i is the true value.In this manner, a Winkler score without any violations is simply the width of the prediction interval, whilst true values falling outside the prediction interval are penalised.Therefore, low Winkler scores suggest narrow but reasonably calibrated prediction intervals.In our evaluation, we consider the Mean Winkler (MW) score across all considered quantiles α ∈ Q, defined as

Pinball loss improvement
To evaluate our approach in the recreated GEFCom2014 competition, we consider the scoring mechanism used in this competition.This mechanism relies on the Pinball Loss (PL), a scoring rule that minimises the loss when issuing a point forecast for the α-quantile [24].For a set of considered quan-tiles Q = [0.01, . . ., 0.99], the PL is calculated with where y i is the true value and ŷi,α is the quantile forecast for the quantile α.For the GEFCom2014, the relative improvement of the PL compared to a given baseline forecast is considered, i.e.
where PL Forecast is the PL for the considered forecast and PL Baseline the PL for the baseline provided in the GEF-Com2014.

Selected base forecasters
Our proposed approach can be applied to forecasts from arbitrary point forecasters.Thus, we evaluate our approach on four simple and two state-of-the-art point forecasting methods.As simple base point forecasters we consider a Linear Regression (LR), a Random Forest (RF), a Feed-Forward Neural Network (NN), and the eXtreme Gradient Boosting (XGBoost) Regressor.We select these methods due to their robust performance in multiple studies, e.g.[16,19,47,49,50], and [51].The two state-of-the-art base point forecasters are Neural Hierarchical Interpolation for Time Series Forecasting (N-HiTS) [6] and Temporal Fusion Transformer (TFT) [38].We provide implementation details for each of the selected base point forecasters in Table 16 in Appendix A.
When applying the base forecasters with the cINN to generate probabilistic forecasts, we manually select the sampling parameter σ that minimises the CRPS on the validation data set.An overview of the selected sampling parameters is presented in Table 17 in Appendix A. Furthermore, all selected base point forecasters are implemented in a pipeline with pyWATTS7 [26].

Probabilistic benchmarks
To assess the quality of the probabilistic forecasts generated with our approach, we compare them to multiple probabilistic benchmarks.These benchmarks can be classified into the following two groups: probabilistic forecasts generated from existing point forecasts and directly generated probabilistic forecasts.We focus on a selection of benchmarks that have achieved state-of-the-art performance whilst being relatively computationally inexpensive and therefore exclude computationally expensive benchmarks that may generalise poorly, such as Bayesian Neural Networks [30].In the following, we introduce the benchmarks of both groups.

Probabilistic forecasts based on existing point forecasts
The first group of probabilistic benchmarks considers methods that generate probabilistic forecasts from existing point forecasts.All of these benchmarks operate on a similar principle.They consider the empirical errors between the point forecasts ŷi and true values y i on a validation data set.These empirical errors are then used to generate prediction intervals.The benchmarks differ in how these empirical errors are used to generate prediction intervals. 8 The first considered benchmark is the Gaussian Prediction Interval (Gaussian PI).In this case, the empirical errors are assumed to be distributed according to a Gaussian distribution and the prediction intervals are calculated based on the standard deviation of these errors [29].
Second, we consider the Empirical Prediction Interval (Empirical PI).This benchmark does not assume any parametric distribution but instead uses the empirical distribution of these empirical errors to calculate the prediction intervals [56].
Finally, we consider a Conformal Prediction Interval (Conformal PI).This benchmark, introduced for multihorizon time series forecasts by [53], calculates a critical nonconformity score for each of the empirical errors and applies Bonferroni and finite sample correction to ensure temporal dependence between these critical scores across the forecast horizon.These critical nonconformity scores are combined with the point forecast to generate the prediction intervals [53].

Direct Probabilistic Forecasts
The second group of probabilistic benchmarks considers methods that directly generate probabilistic forecasts.The first of these benchmarks is DeepAR [48], which is an autoregressive recurrent neural network-based approach for probabilistic forecasting.We implement DeepAR using the PyTorch Forecasting library 9 . 8These benchmarks are selected due to their simplicity and proven performance.Due to computational cost, we explicitly exclude machine learning methods that require retraining for each point forecast, such as [5] and [54]. 9https://pytorch-forecasting.readthedocs.io/en/stable/api/pytorch_forecasting.models.deepar.DeepAR.html The second benchmark method is a Quantile Regression Neural Network (QRNN).It trains a NN to directly forecast selected or multiple quantiles instead of the mean or median [35].To realise the QRNN, we use a separate simple feed-forward NN to forecast each of the selected quantiles, training each NN with the appropriate pinball loss function.The QRNN is implemented using TensorFlow10 with the Keras11 library and the pinball loss function.
The third benchmark method uses the Nearest Neighbour Quantile Filter (NNQF) proposed by [25].Similar to the QRNN, this method also forecasts quantiles.However, instead of using a custom quantile loss function to directly learn the quantiles, the NNQF finds similar values for each time step based on similarity in the target variable to determine quantiles in the data.A forecasting method is then trained to predict these calculated quantiles [25].To realise the NNQF, we use a multi-layer feed-forward NN with one output per quantile, which is implemented using sklearn [43] and pyWATTS [26].

Used cINN
In the evaluation, we use the same cINN architecture (see Table 18 in Appendix A) for each of the considered data sets.It is based on GLOW coupling layers that consider conditional input [34].Similar to [1,27], the conditional input is provided by a fully connected NN, which uses the same exogenous information available to the base forecaster as conditional information (see Table 15).We detail the implementation information for the used cINN in Tables 18,19 and 20 in Appendix A. When training the used cINN, we apply the Adam optimiser with a maximum of 100 Epochs.Furthermore, when sampling in the latent space to generate probabilistic forecasts, we consider a sample size of 100.We implement the cINN in a pipeline with pyWATTS [26].

Evaluation
We evaluate our proposed approach in three steps.First, we compare the probabilistic forecasts generated from our approach when using different base point forecasters.Second, we compare our approach with existing probabilistic benchmarks.For each of these two steps, we first consider an overview of the normalised evaluation metrics for each model across all metrics and all data sets to establish an overview of the results.We then consider the probabilistic forecasts' quality, calibration, sharpness, and prediction intervals separately.Finally, in the third step, we recreate the price track of GEFCom2014 and compare our approach to the competition winners.

Comparison of different base point forecasters
In this section, we compare the performance of the different base point forecasts combined with the cINN in our approach.An overview of the normalised evaluation metrics for each base point forecaster when combined with the cINN for all considered evaluation metrics and data sets is shown in Fig. 3.
For comparison purposes, the performance in each of the considered metrics is normalised so that the best-performing base point forecaster achieves a score of 0.1 and the worstperforming base point forecaster a score of 1.We observe that the best-performing base forecaster depends on the considered metric and the data set.However, the TFT performs consistently well according to all metrics across all data sets by ranking within the top three in all cases apart from the CRPS and nMPI(β) for both β on the Electricity data set.Furthermore, certain base point forecasters exhibit highly Fig. 3 An overview of the normalised evaluation metrics for each base point forecaster when combined with the cINN for each considered evaluation metric across all data sets.The value of each metric is normalised between 0.1 and 1 for illustrative purposes to facilitate the comparison, with lower values indicating better performance.The base point forecasters are ranked from best to worst for each metric variable performance.For example, the LR achieves the best performance with regards to MAQD on the Price data set but consistently performs as one of the worst models on all other data sets.In the following, we report the results in more detail by comparing the forecast quality, the calibration, the sharpness, and the prediction intervals.

Quality
For each of the base point forecasters combined with the cINN, we report the average CRPS across five runs in CRPS Table 2.
We observe that the best-performing point forecaster combined with the cINN again depends on the data set considered, although the cINN base point forecaster combined with the cINN results in the lowest CRPS on two of the four data sets.Furthermore, although all base point forecasters combined with the cINN perform similarly on the Price data set, there are noticeable differences in the results of the other data sets.For example, the N-HiTS and TFT point forecasters combined with the cINN result in a noticeably lower CRPS on the Bike data set than the other point forecasters when combined with the cINN.

Calibration
To analyse the calibration of the probabilistic forecasts generated by combining different point forecasters with the cINN, we report the cINN in Table 3.
We observe that the base forecaster that results in the lowest MAQD when combined with the cINN depends on the data set considered.However, XGBoost, when combined with the cINN, achieves the lowest CRPS on two of the four data sets.Furthermore, the MAQD varies noticeably between the data sets.On the Electricity data set, almost all base point forecasters achieve a similar MAQD when combined with the cINN.However, on the Bike data set, only the TFT combined with the cINN result in a MAQD under 0.1.

Sharpness
We evaluate the sharpness of the probabilistic forecasts generated when combining different base learners with the cINN by reporting the average nMPI(β) over five runs in Table 4.
Depending on the considered data set, we observe that different base point forecasters, when combined with the cINN, result in the best nMPI(β).Only the RF, when combined with the cINN, achieves the best nMPI(β) on more than one data set, whilst the TFT as base point forecaster performs best on the Price data set, and N-HiTS as a base forecaster on the Bike data set.We observe varying nMPI(β)s across the data sets, with the largest nMPI(β) of 1.2953 for β = 98% on the Electricity data set and the narrowest nMPI(β) of 0.1329 for β = 70% on the Price data set.

Prediction intervals
To evaluate calibration and sharpness at the same time, we report the average MW score across five runs as a measure of the quality of the prediction intervals generated by different point forecasters in Table 5.
Again, the performance varies depending on the considered data set.Probabilistic forecasts generated when combining the TFT with the cINN result in the lowest MW score for the Price and Bike data sets, and RF as the base point forecaster results in the lowest MW score for the Electricity and Solar data sets.Regarding the MW scores, the performance varies noticeably depending on which base point forecaster is used on all data sets, although this variance is smaller on the Price data set.

Comparison to benchmarks
In the second step of our evaluation, we compare probabilistic benchmarks with the probabilistic forecasts generated when combining a cINN with the XGBoost, N-HiTS, and TFT base point forecasters.First, we compare the probabilistic forecasts from our approach to benchmarks that also use these same point forecasters to generate probabilistic forecasts.Second, we compare our approach to methods that directly generate probabilistic forecasts.

Probabilistic forecasts based on existing point forecasts
We report an overview of the normalised evaluation metrics for the TFT as the base point forecaster when combined with the cINN and the other benchmarks based on existing point forecasts for all considered evaluation metrics and data sets in Fig. 4. For comparison purposes, the performance in each of the considered metrics is normalised so that the best-performing model achieves a score of 0.1 and the worst-performing model a score of 1.We first observe that for CRPS, MW, and nMPI(β) with β = 98%, our approach using the cINN results in the best performance on all data sets.Additionally, our approach also achieves the lowest nMPI(β) for β = 70% on the Price and Bike data sets and only performs slightly worse than Conformal PI and Empirical PI on the remaining data sets.The benchmarks only perform better than our approach in terms of MAQD, with our approach never achieving the lowest MAQD.The results for the two remaining base point forecasters, namely XGBoost and N-HiTS, are similar and can be found in Appendix C. In the remainder of this section, we analyse quality, calibration, sharpness, and the prediction intervals for each considered data set in more detail.

Quality
We evaluate the quality of the different probabilistic forecasts by reporting the average CRPS across five runs in Table 6.First, our approach using a cINN generally performs better or similarly to the benchmarks.The cINN results in probabilistic forecasts with the lowest CRPS for each considered point forecaster on the Price and Solar data sets and for two of the three point forecasters on the Electricity and Bike data sets.In the remaining two cases, the Conformal PI results in the lowest CRPS, however, the CRPS resulting from our approach is similar in all cases.Second, we observe that the Gaussian PI consistently results in the highest CRPS.Finally, we note that, across all data sets and for all considered point forecasters, the Empirical PI generates probabilistic forecasts resulting in almost identical CRPSs to those from the Conformal PI.

Calibration
To evaluate the calibration of the considered probabilistic forecasts, we report the average MAQD for each data set calculated across five runs in Table 7.
We first observe that the results depend strongly on the base point forecaster and the data set considered.Whilst the Conformal PI results in the lowest MAQD for all base point forecasters on the Electricity data set, the results for the other data sets are not as clear.On the Price and Solar data sets, the Conformal PI and Empirical PI achieve the lowest MAQD depending on the considered point forecaster, whilst each of the three considered benchmarks performs best on the Bike data set for one of the applied base point forecaster applied.Our approach using the cINN never achieves the lowest MAQD.

Sharpness
To assess the sharpness of probabilistic forecasts generated from point forecasts, we report the average nMPI(β) over five runs in Table 8.
Our approach using a cINN results in the lowest nMPI(β) for all base point forecasters, considered values of β and data Fig. 4 An overview of the normalised evaluation metrics for the TFT base forecaster when combined with the cINN or used with the benchmarks based on existing point forecasts for each considered evaluation metric across all data sets.The value of each metric is normalised between 0.1 and 1 for illustrative purposes to facilitate the comparison, with lower values indicating better performance.The considered models are ranked from best to worst for each metric sets almost all the time, with the exceptions being for β = 70% on the Electricity set with the TFT and on the Solar data set with XGBoost and the TFT.Moreover, the nMPI(β) from the Empirical PI and Conformal PI are generally the largest for β = 98%, and noticeably so.For example, the nMPI(β) for β = 98% for Conformal PI on all data sets are often more than double the nMPI(β) generated with the cINN.Further, the nMPI(β) for β = 70% are generally the largest with the Gaussian PI.Finally, the nMPI(β) vary noticeably depending on the data set and selected point forecaster.

Prediction intervals
To simultaneously consider calibration and sharpness, we analyse the prediction intervals of the considered probabilis-  9.
We first note that the probabilistic forecasts generated with the cINN result in the lowest MW scores on all data sets and for all considered point forecaster.Furthermore, the Winkler scores from our approach are noticeably smaller than the benchmarks.Although all the prediction interval-based benchmarks generate probabilistic forecasts with similar Winkler scores, the Gaussian PI results in slightly lower Winkler scores on all data sets, with the difference being most noticeable on the Price data set.Finally, we observe that similar to the CRPS results, the MW scores for the Empirical PI and Conformal PI are almost identical for every data set and each considered point forecaster, with the Conformal PI sometimes performing slightly better.

Direct probabilistic forecasts
An overview of the normalised evaluation metrics for our approach using a cINN combined with three base point forecasters (XGBoost, N-HiTS, TFT) and the three considered benchmarks that directly generate probabilistic forecasts for all considered evaluation metrics and data sets is shown in Fig. 5.For comparison purposes, the performance in each of the considered metrics is normalised so that the best-performing model achieves a score of 0.1 and the worstperforming model a score of 1.We observe that combining an appropriate base point forecaster with the cINN results in the best-performing model for all data sets and across all metrics in all but two cases.The exceptions are the MAQD on the Price data set, where the NNQF performs best, and Fig. 5 An overview of the normalised evaluation metrics for our approach combining the cINN with three base point forecasters (XGBoost, N-HiTS, TFT) and the direct probabilistic benchmarks for each considered evaluation metric across all data sets.The value of each metric is normalised between 0.1 and 1 for illustrative purposes to facilitate the comparison, with lower values indicating better performance.The considered models are ranked from best to worst for each metric the CRPS on the Solar data set, where the QRNN provides the best performance.Furthermore, regarding MW, all three of the base point forecasters, when combined with the cINN, perform better than all of the direct benchmarks on all data sets.Finally, our approach performs consistently across all metrics and data sets, only once achieving the worst ranking, whilst the performance of the direct benchmarks is far more variable, with each of them being ranked worst at least three times.To evaluate the performance in more detail, we consider the forecast quality, calibration, sharpness and prediction intervals for each of the direct benchmarks and our approach with the cINN on the four considered data sets in the following.

Quality
To analyse the quality of the probabilistic forecasts, we report the average CRPS across five runs for all data sets in Table 10.The first observation is that our approach results in the lowest CRPS on three of the four data sets.Thereby, the choice of the base point forecaster is important, with XGBoost combined with the cINN performing best on the Electricity data set, whilst the TFT combined with the cINN performs best on the Price and Bike data sets.On the Solar data set, the QRNN benchmark outperforms all others, although our approach using the XGBoost as a base point forecaster performs similarly.Furthermore, when combined with the cINN, we observe that all base point forecasters outperform all the direct benchmarks on the Price data set, and two of the three base point forecasters outperform all the direct benchmarks on the Electricity data set.In general, the performance of the direct benchmarks is also highly dependent on the considered data set.Of the direct benchmarks, the NNQF performs best for the Electricity data set, the QRNN for the Price and Solar data sets, and DeepAR for the Bike data set.

Calibration
To assess the calibration of the direct probabilistic benchmarks and our approach, we report the average MAQD across five runs in Table 11.
Similar to the CRPS results, our approach using a cINN results in the lowest deviation for three of the four data sets.However, unlike the CRPS results, our approach using the cINN results in the lowest MAQD on the Electricity, Solar, and Bike data sets whilst the NNQF achieves the lowest overall MAQD, on the Price data set.With regards to the direct benchmarks, the NNQF outperforms the other direct benchmarks on all data sets except for the Solar data set, where the lowest MAQD is achieved with the QRNN.

Sharpness
To compare the sharpness of probabilistic forecasts generated with our approach and those from the direct benchmarks, we report the average nMPI(β) over five runs in Table 12.
With regards to the nMPI(β), our approach results in the smallest nMPI(β) for all data sets.Using the XGBoost as a base point forecaster generates the narrowest prediction intervals for the Electricity data set and the lowest nMPI(β) for β = 70% for the Solar data set.Additionally, using the TFT as a base point forecaster results in the narrowest prediction intervals for the Price data set and the lowest nMPI(β) for β = 98% for the Solar data set.Finally, using N-HiTS as a base point forecaster results in the lowest nMPI(β) for the Bike data set.The width of the prediction intervals for the direct probabilistic benchmark depends on the data set.For the Electricity and Price data sets, DeepAR generates probabilistic forecasts with the lowest nMPI(β).However, for the Solar data set, the nMPI(β) from the QRNN is the smallest.The Bike data set is interesting for the benchmarks since the nMPI(β) with β = 98% is the smallest for DeepAR, but the nMPI(β) with β = 70% is the smallest for the QRNN.

Prediction intervals
To evaluate calibration and sharpness simultaneously, we consider the quality of the prediction intervals generated with our approach and the direct probabilistic benchmarks.For this purpose, we report the average MW score across five runs in Table 13.
We first observe that our approach results in the lowest MW scores for every data set.Furthermore, the MW scores for each point forecaster, when combined with the cINN, are lower than any of the direct benchmarks on all data sets.Regarding the direct probabilistic benchmarks, the bestperforming benchmark depends on the data set considered.DeepAR results in the lowest MW scores for the Electricity, Price, and Bike data sets, whilst QRNN results in the lowest MW scores for the Solar data set.

Qualitative analysis
As a final comparison to the benchmarks, we qualitatively compare prediction intervals and calibration for the Price data set to gain further insight into the characteristics of probabilistic forecasts generated by our approach and the We plot the 98%, 70%, and 40% prediction intervals for a single day in the test data set in Fig. 6.Compared to the Conformal PI, our approach generates probabilistic forecasts with the narrowest prediction intervals regardless of the base point forecaster used.In fact, for probabilistic forecasts generated with the N-HiTS or TFT base point forecaster, our approach using a cINN results in the narrowest prediction intervals overall.Furthermore, whilst the 40% and 70% Conformal PIs are only slightly wider than those generated by the cINN, the 98% prediction intervals are by far the widest of all considered benchmarks.The three direct probabilistic benchmarks generate prediction intervals that are generally wider than those generated by the cINN but narrower than the Conformal PIs.
To further analyse the calibration of our forecasts, we plot the forecast quantile coverage against the theoretical quantile coverage as a calibration plot in Fig. 7.We observe that, for all base point forecasters, the Conformal PI provides the most calibrated forecasts, with hardly any deviation from the diagonal.However, our approach using a cINN also results in forecasts that only slightly deviate from the diagonal by slightly overestimating the lower quantiles and slightly underestimating the upper quantiles.From the direct probabilistic benchmarks, the NNQF achieves similar results to our approach using a cINN, whilst the results of DeepAR and the QRNN are noticeably worse.

GEFCom2014 probabilistic price forecasting
For the final step of our evaluation, we test the proposed approach by retrospectively determining its placement in the GEFCom2014.The GEFCom2014 was a probabilistic energy forecasting competition with four different tracks for load, price, wind, and solar forecasting [28].For the evaluation, we recreate the setup of the GEFCom2014 price forecasting track in which 14 teams competed and compare the performance of our approach to the leading entrants from the competition.This comparison is based on the scoring mechanism from GEFCom2014 that considers the final twelve of fifteen tasks, each generating 24-hour quantile forecasts.Given the pinball loss improvement PL % for each task, the final ranking is determined as the average pinball loss improvement across all tasks weighted by the task number [28].For our approach, we apply all previous base point forecasters, select the best-performing sampling parameter over the first three non-evaluated tasks, and use this for all remaining tasks.The final weighted pinball loss improvement and the resulting rank of our approach are shown in Table 14 (see Table 22 for the results per task).Overall, our approach with a cINN and various base forecasters performs well.With simple point forecasting methods such as RF regression and LR, we achieve an average weighted pinball loss improvement that would have placed these methods within the top five of the competition.Furthermore, with a more advanced TFT base forecaster, we achieve an average weighted pinball loss improvement that would have resulted in a third-place finish.
Fig. 6 Exemplary 98%, 70%, and 40% prediction intervals for the Price data set.Probabilistic forecasts are generated by using XGBoost, N-HiTS, and the TFT as base point forecasters and either combining them with our cINN or applying Conformal PI.Further, we compare the three direct probabilistic benchmarks: DeepAR, QRNN, and NNQF

Discussion
In this section, we first discuss our results and the implications these have before we highlight some of the key insights gained from the evaluation.

Results
With regard to our results, we discuss two aspects.First, we focus on the forecasting performance of our approach before considering the GEFCom2014.Fig. 7 Exemplary calibration plots comparing the theoretical and forecast quantiles on the Price data set, with the red diagonal indicating zero deviation.We compare probabilistic forecasts generated by using XGBoost, N-HiTS, and the TFT as base point forecasters and either combining them with our cINN or applying Conformal PI.Further, we compare the three direct probabilistic benchmarks: DeepAR, QRNN, and NNQF

Forecasting performance
With regard to forecasting performance, we first discuss the performance of our approach with different point forecasters before comparing our approach to other probabilistic benchmarks.
When comparing different point forecasters in our approach, we note that the quality of the point forecasts affects the quality of the probabilistic forecasts when combined with the cINN.This observation is unsurprising since the cINN in our approach includes uncertainty around the initial point forecast, and, therefore, the more accurate the point forecast is, the easier it is to include uncertainty effectively.Additionally, the quality of the point forecasts is influenced by the exogenous features considered.Therefore, it may be useful to consider additional features or factors influencing the forecast, similar to [58].
When comparing our approach to the selected benchmarks, we make several observations.First, we observe that our approach generally outperforms all benchmarks regarding CRPS.In the three occasions where our approach does not result in the lowest CRPS, the difference between the best performing Conformal PI or QRNN and our approach is small.Second, our approach is not optimally calibrated.Although the forecasts generated with our cINN are well calibrated compared to the direct probabilistic benchmarks, the Empirical PI and Conformal PI achieve lower MAQDs on all data sets.However, it is worth noting that prediction intervalbased approaches are specifically designed to achieve certain coverage levels, and further, considering the calibration plots in Fig. 7 suggests that the difference in calibration may not be as noticeable as the raw MAQD numbers suggest.
Third, our approach consistently generates the sharpest probabilistic forecasts with the lowest nMPI(β).This obser-vation is further highlighted by Fig. 6 where the forecasts from the N-HiTS and TFT base point forecaster combined with the cINN are far narrower than those of any other benchmarks.
Fourth, our approach outperforms all considered benchmarks on all data sets with regards to MW scores.Since Winkler scores consider both calibration and sharpness, this result suggests that the poorer calibration from our approach is counteracted by the narrow prediction intervals.Considering again Fig. 6, although the prediction intervals of our approach are narrow, the ground truth is still almost always contained within the interval.In comparison, the other benchmark methods, specifically the prediction interval-based approaches, appear to overestimate the width of the prediction intervals, which adversely affects the Winkler score.
Fifth, we note that both our approach and each of the considered benchmarks have strengths and weaknesses.Our approach results in narrow prediction intervals and low CRPS scores, but this comes at the cost of calibration performance.In contrast, the prediction interval-based benchmarks are highly calibrated but generate far wider prediction intervals, resulting in a worse performance regarding Winkler scores.Therefore, the best probabilistic forecast may vary, depending on the requirements of the situation it is being used for.
Finally, when considering the performance of our approach across all metrics compared to the benchmarks (see Figs. 4  and 5), we conclude that the resulting probabilistic forecasts are high quality and generally outperform all benchmarks.Our approach's small loss in calibration performance results in a far sharper forecast, which is reflected in the generally better performance in CRPS and MW.Therefore, our approach can be considered to deliver state-of-the-art probabilistic forecasts.

GEFCom2014
Not only does our approach perform competitively in the considered track of the GEFCom2014, but several factors undercut its true performance.All top-placing contestants in the competition perform specialised operations to improve forecasting performance, i.e. peak pre-processing [20], filtering methods to weight certain days higher [40], or tailored training data periods to improve performance [14].In contrast, we consider all available data for training, refrain from complex pre-processing steps, and only use the default hyperparameters for the base forecasters.Furthermore, the true value of our approach is its ability to enable simple base point forecasters, such as a LR or RF, to rank within the top five compared to the original entrants.Finally, different base point forecasters perform better for different tasks.Therefore, we expect an ensemble method that automatically selects the best base forecaster for a given task to deliver even better performance.

Insights
In addition to the results, there are a few insights regarding the sampling in the latent space and the flexible nature of our approach, which we discuss here.

Sampling in latent space
Our approach includes uncertainty in point forecasts via latent space distribution sampling.Currently, this sampling is performed by adding normally distributed random noise r i ∼ N (0, σ ) to the point forecast.This approach has several limitations.First, the sampling parameter σ is manually selected to generate optimal forecasts according to CRPS.However, by varying this sampling parameter, it is possible to generate different probabilistic forecasts which follow the same general shape but vary in the amount of uncertainty considered.Therefore, it may be interesting to investigate methods to automatically select an optimal sampling parameter given the observed data, a selected base forecaster, and a specific evaluation metric.Such methods would enable the generation of probabilistic forecasts with properties that are specifically designed for a certain application.However, such probabilistic forecasts will only be possible if the requirements of this application can be formulated via a probabilistic loss metric that can be used to select σ .
Second, the current approach to optimise σ is rather rudimentary and based on a single evaluation metric.Therefore, it would be interesting to adapt this optimisation, perhaps by adapting concepts from conformal prediction to calculate the nonconformity scores of the samples.Furthermore, optimising the samples based on the resulting quantiles used as an output of the probabilistic forecast might be interesting.
With such a strategy, Bonferroni correction could possibly be applied to improve the calibration of our approach.

Flexible nature
In the present article, we evaluate the probabilistic forecasting performance of a single selected base point forecaster combined with the cINN.However, our approach is independent of the base point forecast considered, i.e. once the cINN has been trained for a given data set, we can generate probabilistic forecasts from any arbitrary point forecast without retraining.This is advantageous compared to other methods using cINNs or GANs, which require the generative model to be retrained whenever the point forecast is altered.Moreover, such an approach allows us to easily generate an ensemble of probabilistic forecasts based on different point forecasts.
Furthermore, for similar data sets, it may be possible to generate probabilistic forecasts with a generalised cINN that is only trained once on all data sets or a subset thereof.Such a generalised cINN could be beneficial for global forecasting and be applied to scenarios where no data is available, e.g. a new building similar to existing buildings is included in a suburb.However, such a generalised cINN will only be possible if the underlying distribution across the multiple data sets is similar and can be accurately mapped to a single tractable latent space distribution.
Another important aspect is that our approach is not limited to prediction intervals or specific quantiles.Whilst we choose to calculate quantiles based on samples as the output of our approach, the generative nature of the cINN enables us to generate an arbitrary number of samples and either use these directly to form an ensemble forecast or to output an empirical forecast distribution.This is advantageous compared to other probabilistic forecast methods that are, by nature, limited to generating prediction intervals.Specifically, applying an empirical density estimation algorithm to obtain a non-parametric density forecast is simple based on our approach.Such forecasts may be particularly useful if the entire distribution is required, for example, for a stochastic optimisation problem.

Conclusion
In the present article, we introduce an approach to generate probabilistic forecasts from arbitrary point forecasts by using a Conditional Invertible Neural Network (cINN) to learn the underlying distribution of the time series data.Our approach maps the underlying distribution of the data to a known and tractable distribution before combining the uncertainty from this known and tractable distribution with an arbitrary point forecast to generate probabilistic forecasts.Importantly, the cINN is independent of the considered point forecast and must not be retrained when the point forecast is altered.
We evaluate our approach by combining multiple point forecasts with a cINN and comparing the resulting probabilistic forecasts with six probabilistic benchmarks on four data sets.We show that our approach generally outperforms all benchmarks regarding CRPS and Winkler scores.Further, our approach generates probabilistic forecasts with the narrowest prediction intervals whilst maintaining reasonable performance in calibration.Finally, we recreate the GEF-Com2014 and show that our approach enables simple base point forecasters to rank within the top five.
Our approach offers a solution to generate flexible probabilistic forecasts based on arbitrary point forecasts.In future work, this flexibility should be further investigated by developing a more advanced strategy for selecting the sampling parameter to improve the calibration of our probabilistic forecasts.Furthermore, automating the selection of this sampling parameter and considering how different metrics for optimising this parameter affect the resulting forecasts should be investigated.It would also be interesting to extend our approach to multivariate probabilistic forecasts.Finally, it may be interesting to explore the performance of our approach using a generalised cINN to generate probabilistic forecasts on multiple data sets without retraining.

Appendix B point forecast evaluation
Since the quality of the base point forecaster influences the resulting probabilistic forecast when combined with a cINN, we briefly evaluate the point performance of the selected base with a true value y i , a forecast value ŷi , and n observations.We report the average RMSE (B1) over five runs in Table 21.
In general, we observe that the best-performing point forecast depends on the data set considered, with the performance of most point forecasters varying noticeably across the data sets.However, the TFT performs most consistently, achieving the lowest RMSE on two of the four data sets.Fig. 9 An overview of the normalised evaluation metrics for N-HiTS base forecaster when combined with the cINN or used with the benchmarks based on existing point forecasts for each considered evaluation metric across all data sets.The value of each metric is normalised between 0.1 and 1 for illustrative purposes to facilitate the comparison, with lower values indicating better performance.The considered models are ranked from best to worst for each metric

Table 1
An overview of previous research related to the present article.None of the identified articles proposes methods capable of generating probabilistic forecasts from existing point forecasts without being limited to only generating prediction intervals or involving a training process that is dependent on the training of the point forecast

Table 2
The average CRPS calculated on the test data set for each of the considered base point forecasters combined with the cINN over five runs.The best values for each data set are highlighted in bold

Table 3
The average MAQD between the theoretical and forecast quantiles calculated on the test data set for each of the considered base point forecasters combined with the cINN over five runs.The best values for each data set are highlighted in bold

Table 4
The average nMPI(β) calculated on the test data set for each of the considered base point forecasters combined with the cINN for β = 98% and β = 70% over five runs.The best values for each data set and β are highlighted in bold

Table 5
The average MW score calculated on the test data set for each of the considered base point forecasters combined with the cINN over five runs.The best values for each data set are highlighted in bold

Table 7
Comparison

Table 10
Comparison of the average CRPS between the probabilistic forecasts from the cINN and the direct probabilistic benchmarks.The average CRPS is calculated over five runs on the test data set, and the best values for each data set are highlighted in bold

Table 11
Comparison of the average MAQD between the theoretical and forecast quantiles from the cINN and the direct probabilistic benchmarks.The average MAQD is calculated across five runs on the test data set, and the best values for each data set are highlighted in bold

Table 13
Comparison of the average MW score between the probabilistic forecasts from the cINN and the direct probabilistic benchmarks.The average MW score is calculated over five runs on the test data set, and the best values for each data set are highlighted in bold

Table 14 The
This ranking is determined by how each individual method would have placed in the GEFCom2014 price forecasting challenge in 2014 and, therefore, assumes that only one of the presented methods is considered at a time when creating the respective ranking 1

Table 20 )
Table 19 Implementation details of the subnetwork in the used cINN

Table 21
Comparison of the average RMSE on the test data set for each of the considered base point forecasters.The best values for each data set are highlighted in bold