Keywords

6.1 Introduction

Machine learning is increasingly used to predict fuel spray characteristics, such as spray penetration [1, 2]. However, the cost of fuel spray experiments limits the available training data. When a neural network is trained on limited data, there is a trade-off between over-fitting the data and over-simplifying the neural network (providing too few nodes or layers within the network). A nonlinear phenomenon, such as spray breakup, requires a degree of neural network complexity that cannot be trained reliably with the quantity of experimental spray penetration data typically available. To get around this, researchers have augmented experimental spray results with complementary computational fluid dynamics (CFD) studies [3]. However, CFD occupies time and resources and risks distorting the original experimental data. There is the potential to remove CFD from the process and to train neural networks directly on experimental data [4]. This study proposes to encompass predictive uncertainty due to neural network under-training (due to limited training data) within confidence intervals around neural network predictions. The aim is to improve the practical usefulness of neural networks as predictive tools in engineering design, in cases where they are trained incompletely due to insufficient training data.

Confidence intervals for neural networks are not new. Chryssolouris et al. [5] derived confidence intervals for a neural network from fundamental statistics. To use this approach, the experimenter must assess the confidence intervals of each of the neural network’s training inputs and the sensitivity of each neural network output to variations in each input. Confidence intervals are typically available for spray data, because once a new experimental condition is created, experiments are easily repeated. A matrix of dm⁄dn sensitivities sized m outputs by n inputs must be manipulated (transposed and multiplied) to calculate confidence intervals for the neural network. Trichakis et al. [6] reviewed numerical methods of estimating the confidence intervals on neural networks and applied the bootstrap method of Efron [7] to estimate confidence intervals on a neural network predicting the hydraulic head of aquifers. Although beyond the scope of Efron’s original proposal, Trichakis et al. showed that the bootstrap method is a practical way to estimate confidence intervals for perceptron neural networks. Confidence intervals have been used with some success in other applications of neural networks, including applications related to solar power generation [8].

The present study applies the bootstrap method to calculate confidence intervals on spray penetration predictions. The novel contribution of the present study is to incorporate within the confidence intervals not only (1) uncertainty arising from sampling error on measured input values but also (2) residual random errors in the neural network due to initialisation and incomplete training. Because of the limited data available for training, the neural network training algorithm does not converge reliably. This means that neural network output contains random error due to residual randomness from the initialisation state. We allow this random error to combine with variations that arise from resampling inputs (due to errors in experimental measurements). The confidence intervals produced by this method thus incorporate both sources of error. Thus, we prefer repeated, abortive attempts at neural network training, with suitable confidence intervals, over a single neural network that over-fits the data. We recognise that in general simpler neural networks (containing fewer nodes) can be trained on less abundant data, but there is typically a lower limit on neural network complexity whereby an over-simple neural network introduces model bias (a situation in which the neural network is insufficiently complex to represent adequately the output despite successful training). The methods suggested in this paper offer a means of training suitably complex neural networks from limited experimental data without relying on CFD or any data augmentation techniques. By providing confidence intervals, the methods suggested in this paper support the judicious application of neural networks to practical problems in engineering.

6.2 Methodology

This investigation uses liquid and vapour spray penetration data from Honecker et al. [9]. The authors began with a common rail injector having specific flow rate of 310 mL/30 s at 100 bar from 8 holes of 125 µm diameter. They modified the injector to have only three holes, and they measured the injection plume from only the “downward facing” hole. Measurement was by means of three cameras with no filter, 308 nm filter and 520 nm filter, respectively. Images were processed in Matlab to identify contrast thresholds in each camera view representing the boundaries of the liquid and vapour phases.

The researchers ran diesel and polyoxymethylene dimethyl ethers (OME) fuel injection events across three different injection pressure conditions and captured images at 100 ns intervals. For this work we have used the OME data only. For each injection pressure condition, 20 injection events were measured, resulting in a total of 1260 images. For each group of 20 images having the same injection pressure condition and time stamp, the experimenters produced an average vapour penetration length with standard deviation and, where available, an average liquid penetration length with standard deviation. This resulted in 63 average penetration length data points for the vapour phase and 57 for the liquid phase. Wishing to use the raw, resampled data but having access only to the mean and standard deviation figures published by Honecker et al., the present study calculates a normally distributed population of 20 penetration lengths for every experimental condition using the normdist (normal distribution) function in Microsoft Excel [ver. 16.61.1 (2022)]. This resulted in 63 × 20 data points for the vapour phase and 57 × 20 for the liquid phase.

We build 100 separate training data sets of 57 data points each from the available liquid phase data, and again (100 training sets of 63 points each) from the vapour phase data. The choice of 100 is arbitrary, but it is intentionally larger than the number of repeat samples of each experimental data point. We use the resulting 2 × 100 data sets to train 100 neural networks on the liquid phase data and 100 neural networks on the vapour phase data. Each of the 2 × 100 data sets contains one each of the permutations of injection pressure and time from the original experimental study (all 63 vapour phase permutations from the vapour phase or all 57 from the liquid phase). However, each of the data points in each of the 2 × 100 sets is selected randomly from among the 20 available samples for that data point.

Each of the 2 × 100 neural networks trained in this way involves just 8 nodes and 15 training epochs. Training is stopped at 15 epochs, with more typical convergence criteria supressed. These hyperparameter values are informed by sensitivity studies (ranging from 2 to 16 nodes and 10 to 20 epochs). Because the neural networks produced within these ranges of hyperparameter values are undertrained, we observe variability in the results of the sensitivity study. The values were chosen subjectively after repeated runs, but more systematic approaches are suggested in the ‘Discussion’ section below.

We use Matlab’s Neural Network Toolbox [ver. R2022a (2022)]. We create feed-forward neural networks with one hidden layer of 8 nodes, represented in Fig. 6.1. We use the sigmoid activation function, ϕ. Inputs are normalised (φ) and outputs are denormalised (φ′). We use the Levenberg–Marquardt training algorithm with a fixed learning rate of 0.001 to set weights w, which are initialised randomly with a mean value of 0.5 and a normal distribution. Training stops at 15 epochs, ensuring that the networks do not over-fit the limited available training data but also ensuring that every network is under-trained. The 2 × 100 neural networks thus trained are expected to differ because they each retain some of the randomness introduced by the initialisation of weights. We do not consider alternative activation functions, because, in a single-layer network, saturated weights and vanishing gradients are not expected [10]. We do not consider alternative initialisation strategies.

Fig. 6.1
A connected network of the independent variables, input layers, hidden layers, output layers, and the dependent variable. It includes t, P subscript i n j, x cap subscript i = 1, h subscript j = 1, y cap, and l, among others.

Layout of single-layer feed-forward neural networks used in this study, relating injection pressure Pinj and time t to liquid or vapour penetration length l

To reiterate: the 20 repeat samples of each of the 63 + 57 data points are not aggregated to create a larger data set. Rather, each neural network is trained on a set of distinct data points (63 for the vapour phase or 57 for the liquid phase). The repeat sampling is reflected not in the training data for any one neural network but rather in the profusion of neural networks, each trained on a different data set wherein each point was selected randomly from among a population of 20 repeat samples of that data point.

After training a family of neural networks to predict spray penetration, the entire population of neural networks may be queried for any desired set of input values. The result of this query is a set of predictions having a mean and standard deviation. Thus, the family of neural networks collectively offers a prediction with confidence intervals corresponding to any input condition.

6.3 Results and Discussion

The mean outputs of the family of 100 × 2 neural networks are shown in Fig. 6.1a, b with the experimental data (mean values) from Honecker et al. [9]. The same mean predictions are shown together with their upper and lower confidence intervals (single standard deviation) in Figs. 6.2, 6.3, and 6.4 for injection pressures of 700 bar, 1000 bar, and 1500 bar, respectively.

Fig. 6.2
2 scatterplot and multi-line graphs of penetration versus time for the liquid and vapor phases. Both graphs plot E x p P i and N N P i equals 700, 1000, and 1500 bars with an increasing trend.

NN mean predictions (lines) with experimental data (points) a liquid phase, b vapour phase

Fig. 6.3
2 scatterplot and multi-line graphs of penetration versus time for the liquid and vapor phases with a P subscript i of 700 bar. Both graphs plot experiments, N N mean, and 2 N N standard deviations with an increasing trend.

NN mean predictions and confidence intervals at 700 bar a liquid phase, b vapour phase

Fig. 6.4
2 scatterplot and multi-line graphs of penetration versus time for the liquid and vapor phases with a P subscript i of 1000 bar. Both graphs plot experiments, N N mean, and 2 N N standard deviations with an increasing trend.

NN mean predictions and confidence intervals at 1000 bar a liquid phase, b vapour phase)

The figures show wider confidence intervals on neural network outputs in conditions where the experimental data of Honecker et al. [9] showed wider confidence intervals: in the vapour phase more than in the liquid phase, and in the spray breakup region of the liquid phase curves. Figure 6.6 reproduces graphs of experimental data from Honecker et al. [9].

Furthermore, the results show wider confidence intervals at the maximum penetration length of both liquid and vapour phases—conditions where the experimental data did not exhibit similarly wide confidence intervals. Thus, there is a source of uncertainty captured in this study that is not directly related to the uncertainty in the input data (Fig. 6.5).

Fig. 6.5
2 scatterplot and multi-line graphs of penetration versus time for the liquid and vapor phases with a P subscript i of 1500 bar. Both graphs plot experiments, N N mean, and 2 N N standard deviations with an increasing trend.

NN mean predictions and confidence intervals at 1500 bar a liquid phase, b vapour phase)

Fig. 6.6
3 multi-line graphs of L P L, G P L, and L O L versus time. It plots pressure variations of 700, 1000, and 1500 bars in decreasing, increasing, and increasing trends with fluctuations, respectively. It has P c h = 50 bar, T c h = 800 kelvin, t i n j = 800 microseconds, and O M E 3 to 5.

Experimental data from Honecker et al. [9]

This paper documents the application of bootstrap confidence intervals (following the method of Efron [7]) to neural networks, extending the work of Trichakis et al. [6] to deal with data sets that are limited in relation to the complexity of the neural network required to fit the data. The spray penetration data of Honecker et al. [9] are used to train a family of neural networks, each incorporating randomness from (1) random sampling of redundant input data per the bootstrap method and (2) residual randomness from initialisation of the neural networks (persisting in networks because of undertraining). With every set of predictions made by the family of neural networks, there is a mean and a confidence interval associated with that set.

This study does not select the node number rigorously. Training stopped according to an epoch number limit, and again this was not assessed rigorously. Had this study sought to produce a rigorously optimised neural network, then a better approach would have been to produce a different family of neural networks representing bootstrapped confidence intervals for each permutation of hyperparameters. Rather than reserving precious data for the purpose of validating the selection of hyperparameters, the full population of families of neural networks may be examined to identify the set of hyperparameter values offering the narrowest confidence intervals. Rather than comparing the average size (across the predictive domain) of confidence intervals for each hyperparameter value, researchers may go a step further to compare weighted average standard deviations, where the weights are specified to give greater importance to the areas of interest within the curves, such as the droplet breakup region in the case of liquid spray penetration curves.

6.4 Conclusion

We have developed and applied a novel machine learning technique to predict synthetic fuel spray penetration from limited experimental data without computational fluid dynamics. When experimental data are scarce, confidence intervals on a neural network predictions are much more useful than confirming the neural network’s performance with reference to a set of reserved test data. As the quantity of test data approaches infinity, the performance against test data approaches a confidence interval, but this is not an efficient use of data in the sense that it does not represent, for each data point, the greatest contribution of that data point toward improving the neural network’s predictive accuracy. The bootstrap method applies every data point to the related (by statistical theory) tasks of minimising the error and quantifying the error. Thus, the bootstrap approach is preferable when data are limited.

The bootstrap confidence intervals produced in this study reflect both (1) the randomness associated with the input data, as measured by the confidence intervals on that data or the repeat sampling results of that data and (2) the residual randomness due to the random initialisation of the neural networks, surviving in the ‘trained’ neural networks because of under-training. It is not necessary that the training data incorporates repeat sampling or confidence intervals of its own; the method could be applied to show confidence intervals that arise entirely from under-training rather than from both sources of error. Likewise, the bootstrap method could be applied where randomness exists in the training data but under-training is not a factor, as in the study by Trichakis et al. [6].

We propose that, when data are limited, the best application of all the available data is toward the better training of neural networks and the narrowing of the confidence intervals around their predictions. When applied ad infinitum (with unlimited data), a validation study for a neural network approaches an accurate representation of a confidence interval on neural network predictions. Far better, we propose, when data are limited, to apply all data to the narrowing of confidence intervals which are calculated from theory or from an a demonstrably sound numerical approximation of the relevant theory (such as the bootstrap method). These considerations are offered in the hope of improving the application of machine learning to limited data sets and the assessment of the uncertainty that arises from so doing.