Below we explore the usefulness of parallel MCMC in inverting the \( \epsilon_{r} \) field. We shall model the \( \epsilon_{r} \) field using multivariate Gaussians placed at the first 8 pilot points. The Gaussians are governed by the same variogram, whose range is also estimated from the \( \epsilon_{r} \) field data. Thus our RFM contains nine parameters including \( \epsilon_{r} \) at 8 pilot points and the variogram’s correlation range. They are treated as random variables in our statistical formulation of the inverse problem and their nine-dimensional joint PDF is inferred via MCMC. Although the same forward model is used for generation of the synthetic true field and MCMC inversion, there is still model-form error, as we use SGSIM, a stochastic generator of relative permittivity fields. It means that even if we use the same parameter values as the true case in the forward model, we will not reproduce the exact relative permittivity field or measurements as the true case. In other word, the MCMC inversion in this studied case is not an “inverse crime”. A more detailed explanation is in Sect. 4.1.
Inversions with observations with 2% noise (the base case)
As a first step, we solve the inverse problem with 2% noise and limited observations. 30 sources and 30 receivers are used to calculate the first-arrival-travel time to compare against the 900 observations, as shown in Fig. 1b. In this study, we assume the horizontal correlation range of the variogram is 10 times larger than the vertical one. Prior distributions for relative dielectric permittivity at the pilot points is U[4, 18] and U[1, 3] for the correlation range. 20 MCMC chains were used, and Fig. 2 shows the posterior density distribution after 50,000 iterations per chain i.e., a total of 1 million parameters samples were explored for constructing the posterior density distribution. The red vertical lines are the true value. The density distributions show convergences to the true values for all the parameters except for the parameter spatial correlation range, although its distribution does encapsulate the true value. A possible reason is that the locations of the 8 pilot points already impose a length-scale for the \( \epsilon_{r} \) field, which may conflict with the 9th parameter (spatial correlation range). Further, there is no consistent over- or underestimation of \( \epsilon_{r} \) at the 8 pilot points. The MAP (maximum a posteriori) estimate for \( \epsilon_{r} \) i.e., the peak of the marginalized PDF, at pilot points 4 and 7 are overestimates, whereas \( \epsilon_{r} \) at pilot point 8 and the correlation range are underestimated. There is no substantial difference in the MAP estimates and true values for the rest of the parameters. Thus our formulation and implementation seem to be correct and do not introduce bias in the results. In this study, as mentioned in Sect. 3, the permittivity field covers 4 m by 15 m area, and is discretized into a 20 X 75 grid (1500 points total). The permittivity value on each point is calculated by sequential Gaussian simulation (SGSIM) algorithm (Deutsch and Journel 1998), which internally depends on a random number generator. SGSIM takes as its inputs the permittivity values at the pilot points, as well as the variogram for a multiGaussian distribution, and outputs a realization that serves as the permittivity field. For commonly used random number generator, an integer number is used as random seed (or seed state, or seed) for initializing a “pseudorandom” number generation. With a fixed random seed, the random number generator can always give the same random numbers series, which will provide the same permittivity field with given value at pilot points and correlation range. Figure 3 shows the results for an inversion test case with a fixed random seed, which is the same as the one used for the generation of the true field. The posterior distribution is very sharp and almost collapses to the true model parameters’ values (2% noise is added to the observations, which leads to a slightly imperfect collapse). The posterior distribution of the correlation range is wide, since the pilot points’ permittivity values partially constrain the correlation range.
However, in this study, the random seed are deemed unknown, similar to a real inversion problem. The random seed cannot be calibrated as the relationship between random seed and generated random number series is chaotic. In summary, the generation of the true case/field is not repeatable if the random seed is unknown. Figure 4 shows an simple example to demonstrate how the random seed affects the posterior distribution. In this simple example, all the true values for the 8 pilot points and the parameter spatial correlation range are used to generate the stochastic field through SGSIM, but without knowing the random seed, there can be infinite number of the stochastic fields that look different from each other. All these stochastic fields can be used to calculate the travel time, and the travel times for the fields would be different from each other as well. The root-mean-square errors (RMSEs) between the computed travel times for the stochastic fields and the travel time calculated from the synthetic true field can be evaluated. The red line in Fig. 4 shows the distribution of the RMSEs for 1000 stochastic fields, which are all generated through SGSIM using true values of the pilot points and the parameter spatial correlation range. As a comparison, the blue line in Fig. 4 shows the distribution of the RMSEs of the 1000 fields where the pilot point 1 is 10% bigger than the true value (Keeping all other parameters the same as true case, only changing pilot point 1). Similar evaluations were done by increasing the pilot point 1 to be 25% and 50% bigger than the true value, shown as the green and black lines, respectively. There are obvious overlaps among these distributions, such as the pink shadow area indicating the overlap between the case with all true values (red line), and the case with the pilot point #1 to be 50% higher than the true value (black line). This represents the possibility that biased pilot points may yield a better-performing stochastic field than the stochastic field(s) generated with all the true values, although this possibility is only 5% (the pink shadow area in Fig. 4) in the example. Such possibilities are 42 and 23% respectively, when the pilot point 1 is 110 and 125% of the true value. This is the reason why the posterior does not perfectly collapse to the true value (the red line would stack at zero in that case). Please note that the values of the possibilities listed here are only for this simple example. Summarily, since we let the random seed to vary during MCMC iterations, it causes the posterior distribution to be wide, as shown in Fig. 2.
Figure 5 shows the convergence of the posteriors for the base case. Because there are one million data points for each parameter, it is difficult to check the convergence through the trajectories. Hence, the boxplot is used to show the convergence of the quantiles of the posteriors distributions. After about 20,000 iteration (totally 400,000 samples for 20 chains), the posteriors converged.
Inversions with different level of noise in observation
In practice, observations are noisy; they affect the quality of the inferences and the sophistication of the RFM that can be used with them. Here we investigate the impact of noisy observations on the inferred permittivity field. We do so by varying the noise added to observations. The noise is modeled as a normal distribution, with mean set to 0 and the standard deviation defined as a percentage of the average (true) observation. 4 cases were investigated with the noise standard deviation set to 2, 5, 10, and 15 percent of the mean of the synthetic true observation. The number of the sources and receivers is kept at 30. The results are based on 20 chains, each executing 50,000 iterations. The mean of the true, noiseless observations is 0.0765 (µs), and the standard deviation is 0.030025 (µs). Table 2 lists the standard deviation of the noise and the ratio of noise standard deviation over observation standard deviation.
Table 2 Standard deviation of the noise and the ratio of noise standard deviation over observation standard deviation
Figure 6 shows the boxplots for the inferred permittivities and correlation range, as a function of the standard deviation of the noise added to observations. The horizontal red lines are the true value for the 8 pilot points and the correlation range. The horizontal axis of each plot shows the magnitude of the noise. When the noise’s standard deviation is smaller than 10% of the observations’ mean, the proposed approach captures the true values within the interquartile range (IQR) of the samples produced by MCMC. At noise levels of about 15%, the inversion is destabilized i.e., the information content in the observations is sufficiently masked that they can no longer constrain the nine-dimensional RFM with no model form error.
Data worth and redundancy
In this section, we investigate the effects of varying the number of sources and receivers to evaluate the data worth and redundancy issues. Equal numbers of sources and receivers are used. The sources and the receivers are uniformly distributed in their respective wells from 0 to -15 m at the left side (x = 0 m) and right side (x = 4 m) of the field. 12 cases were investigated with 5, 10, 15, 20, 25, 30, 35, 40, 45, 50, 75, and 100 sources and receivers. 2% noise is added to the true observations. The results are based on 20 chains, each executing 50,000 iterations. Figure 7 shows the boxplot of the 8 pilot points and the correlation range’s posterior density distribution. With the increase of sources and receivers, the posterior density distribution is seen to capture the true value better when the number of sources/receivers reach 30. The bound of the posterior cannot be improved when the number of sources/receivers exceeds 35. With the increase of number of sources/receivers, the distance between nearby receivers become smaller and smaller, which means that the measured travel time at nearby receivers are closer and closer. The small difference of the travel time between nearby receivers may be covered by the noise. The only exception is the 9th parameter, range, which is not affected much as the number of sources is varied. This is because the 8 pilot points already include the information on the field’s spatial correlation range.
Pilot points
In this section, we investigate the effects of changing the number of pilot points. The number of pilot points is increased from 4 to 24 incrementally, for a total of 6 test cases. The position and true value of the 24 pilot points are shown in Fig. 1a and Table 1. 2% noise is added to the true observations. The number of the sources and receivers is kept at 30 (900 observation data). The aim of this section is to show that with a given observational dataset, there is an optimal number of pilot-points. Commonly, when the number of pilot points increases, the variance of permittivity field in the domain should decrease. However, this assumes that the permittivity of the pilot points is known. In our case, the permittivity of the pilot points are unknown, and needs to be inferred from observations. As the number of pilot-points increases, and the number of indirect observations do not, it becomes progressively more difficult to infer them accurately. Figure 8 shows the boxplot of posterior density distribution for the pilot points in the 6 cases in the study. The horizontal red lines are the true value of the 24 pilot points. There are 24 boxplots in Fig. 8, representing the posterior density distribution of the 24 pilot points obtained for the 6 cases. For example, the first bar in the first plot (named “pilot point 1”) stands for the posterior distribution of the pilot point 1 in the case with a total of 4 pilot points modeling the field. For the plot named “pilot point 23”, there is one bar in the plot, because pilot point 23 can only be calibrated when the total numbers of pilot points is at least 23. With the increase in the total number of pilot points, the uncertainty ranges slightly increase, especially for the pilot points 1–8. Figure 9 (top) shows the mean dielectric permittivity field for the cases with 4, 8, 12, 16 and 24 pilot points controlling the field. The mean dielectric permittivity field is the average of 200,000 realizations of \( \epsilon_{r} \) fields generated using samples randomly picked from the MCMC chains. Compared to the true field (Fig. 1a), we see that the mean field computed with 4 and 8 pilot points can capture the main spatial variation of the field. The cases with 12 and 16 pilot points controlling the field capture more spatial details, though they might be spurious. In Fig. 9 (bottom) we plot the pointwise variance computed from the 200,000 realizations. One can see that as the number of pilot points in the RFM (i.e., its complexity, flexibility and consequently, dimensionality) increases, we see higher variance in \( \epsilon_{r} \). This is especially true for the most complex RFMs with 20 and 24 pilot points. Figure 10 shows the best dielectric permittivity field of the 200,000 realizations, i.e., the one whose simulations best match the observations. Note that the individual fields do not necessarily resemble Fig. 9 (top row). Figure 11 shows the root mean squared error (RMSE) between the 5 inverted fields and the true field. The black circles are for the RMSE between the mean field and the true field. The red circles are for the RMSE between the best inverted filed and the true field. The RFM with 8 pilot points, with limited observations (900 in this study), provides the best matches in terms of the estimated mean and best fields compared to other RFMs. However, a good agreement between observations and mean or best field does not automatically imply that the eight-pilot-point RFM is the one to use for the given observational dataset; rather the determination must be made based on all the realizations that may be obtained from a calibrated RFM.
This is accomplished using the cumulative rank probability score (CRPS); see Ray et al. (2015) for the definition of CRPS and how it can be used, along with an ensemble of predictions from a (Bayesian) calibrated model, to gauge the quality of the calibration. CRPS, loosely speaking, computes the discrepancy between an ensemble of predictions (by computing the empirical cumulative distribution function) and observations. It has units of the observed quantity (time, in our case) and smaller values of CRPS are preferred. In Fig. 12, we plot the CRPS of the ensemble predictions obtained from calibrated models that used RFMs of increasing complexity. We see that the 4-pilot-point RFM has the lowest CRPS, showing the difficulty of estimating permittivity accurately as the pilot points are increased.
Discussion
The uncertainty in the inferred parameters—\( \epsilon_{r} \) at the pilot points and the correlation range of the variogram are caused by three factors: (1) the quality of the observational data, i.e., the magnitude of noise in it; (2) the quantity of observational data, and (3) the adequacy of the RFM in estimating a spatially complex \( \epsilon_{r} \) field. In Sect. 4.1–4.4 we performed a set of experiments, and we interpret the results to gauge the interplay of the three factors in deciding the quality of the inversion.
In Sect. 4.1, we check if the formulation of the likelihood and the MCMC implementation produces correct results i.e., if the inferences are bias-free when observations are noiseless. In Fig. 2 we find the inferences drawn with limited observations to be free of any systematic errors and we proceed to the problem of the information required to constrain a nine-dimensional RFM (Fig. 6). We find that for less than about 10% noise, the PDF for \( \epsilon_{r} \) get wider with the noise. For 15% noise, the median of the inferred PDFs shift away from the true value. Note that the observations may still be sufficiently informative to constrain a simpler RFM.
Having established an approximate lower bound on the amount of information required to constrain the RFM, we refine the analysis by removing the paucity of observational data/information. In Fig. 7 and Sect. 4.3 we perform inversions with the 8-pilot-point RFM while increasing the amount of observations. Figure 4 shows that the median \( \epsilon_{r} \), as inferred at the 8 pilot points, asymptote to position-specific constants by about 50 source-receiver pairs, while their uncertainty keeps shrinking as the number of source-receiver pairs increases. The parameters with the largest estimation errors are pilot point #4 and the correlation range. As seen in Fig. 1, pilot point #4 is near a sharp gradient in \( \epsilon_{r} \), and capturing it with a mixture of eight pilot points is difficult. The difficulty in estimating correlation range is explained by an ambiguity. There are two length scales in the inverse problem—the correlation range and the distance between the pilot points. The correlation range is therefore difficult to estimate and increasing the number of source-receiver pairs does not sharpen the PDF (see Fig. 7).
In reality, the appropriate RFM is not known a priori, and one typically has to investigate RFMs of increasing complexity to arrive at the best one. In our case, this implies performing the inversion using RFMs constructed with increasing numbers of pilot points. This is also investigated in Sect. 4.4, where we investigate RFMs constructed using 4–24 pilot points. As seen in Fig. 8, 9 and 11, a more sophisticated RFM does not necessarily lead to better reconstructions of \( \epsilon_{r} \) fields, if the quantity of observational data is held constant; instead it runs the danger of overfitting and providing poor predictions. In Fig. 8 (plots of \( \epsilon_{r} \) for pilot points #2, #4, #5 and #8), we see that the width of the uncertainty bounds seems to become constant after about 10 pilot points in the RFM. This is also reflected in Fig. 9. In the plots on top, where we plot the mean of 200,000 realizations of \( \epsilon_{r} \), increasing the complexity of the RFM seems to reconstruct more spatial details. However, the variance in the reconstructions [Fig. 9 (bottom row)] increases with the complexity of the RFM, and the details captured by the mean field are not necessarily more accurate, given the increasing uncertainty associated with them. Figures 9 and 10 reveal the danger of using a mean \( \epsilon_{r} \) field from the MCMC solution as a representative of the entire ensemble of \( \epsilon_{r} \) field realizations. As Fig. 9 (bottom) shows, the pointwise standard deviations are large, and consequently, the best field (Fig. 10) has little resemblance to the mean field (Fig. 9 (top row)).
Figure 11 also shows that the agreement between the true and estimated fields actually become worse as we add pilot points beyond 8 to the RFM. Figure 12, which plots the CRPS as the RFM complexity is increased, shows that the RMSE of the mean field is not a good guide for selecting RFMs, as it ignores the variability/uncertainty in the inferred field. The CRPS plot shows us that of the RFMs considered, the 4-pilot-point RFM is most appropriate for use with the dataset, even though the RMSE of the mean field it produces is not the most optimum. Thus while we may have 900 travel time observations, they may not be of much use in constraining a complex RFM. This may be due to the physics of the problem—EM waves can find alternative paths with much the same travel times as we place more pilot points—or it could be due to the variability of the multiGaussian permittivity fields generated by SGSIM.
In Fig. 13, we plot the estimate of the noise (\( \sigma \)) for the tests shown in Figs. 6, 7 and 8. In Fig. 13a, we see that when the observations are corrupted by 2, 5, 10 and 15% noise, we infer \( \sigma \) to be about 5, 8, 15 and 25%. This overestimate is due to the variability introduced by SGSIM and the limited nature of the observations. In Fig. 13b, we see that increasing the observations actually improves the estimate of \( \sigma \), drawing it closer to its true value of 2%; however, there is still some residual variability due to the stochastic nature of SGSIM. In Fig. 13c, we see that increasing the number of pilot points somewhat reduces \( \sigma \).