Abstract
This paper deals with the taking into account a given target set of realizations as constraints in the Kullback–Leibler divergence minimum principle (KLDMP). We present a novel probabilistic learning algorithm that makes it possible to use the KLDMP when the constraints are not defined by a target set of statistical moments for the quantity of interest (QoI) of an uncertain/stochastic computational model, but are defined by a target set of realizations for the QoI for which the statistical moments associated with these realizations are not or cannot be estimated. The method consists in defining a functional constraint, as the equality of the Fourier transforms of the posterior probability measure and the target probability measure, and in constructing a finite representation of the weak formulation of this functional constraint. The proposed approach allows for estimating the posterior probability measure of the QoI (unsupervised case) or of the posterior joint probability measure of the QoI with the control parameter (supervised case). The existence and the uniqueness of the posterior probability measure is analyzed for the two cases. The numerical aspects are detailed in order to facilitate the implementation of the proposed method. The presented application in high dimension demonstrates the efficiency and the robustness of the proposed algorithm.
Similar content being viewed by others
References
Agmon N, Alhassid Y, Levine RD (1979) An algorithm for finding the distribution of maximal entropy. J Comput Phys 30(2):250–258. https://doi.org/10.1016/0021-9991(79)90102-5
Arnst M, Abello-Álvarez B, Ponthot J-P, Boman R (2017) Itô-SDE MCMC method for Bayesian characterization of errors associated with data limitations in stochastic expansion methods for uncertainty quantification. J Comput Phys 349:59–79. https://doi.org/10.1016/j.jcp.2017.08.005
Arnst M, Soize C, Bulthies K (2021) Computation of Sobol indices in global sensitivity analysis from small data sets by probabilistic learning on manifolds. Int J Uncertain Quantif 11(2):1–23. https://doi.org/10.1615/Int.J.UncertaintyQuantification.2020032674
Batou A, Soize C (2013) Calculation of Lagrange multipliers in the construction of maximum entropy distributions in high stochastic dimension. SIAM/ASA J Uncertain Quantif 1(1):431–451. https://doi.org/10.1137/120901386
Bernardo JM, Smith AFM (2000) Bayesian theory. Wiley, Chichester
Bilionis I, Zabaras N (2017) Bayesian uncertainty propagation using Gaussian processes. In: Ghanem R, Higdon D, Houman O (eds) Handbook of uncertainty quantification, Ch. 15. Springer, Cham, pp 555–600
Bowman A, Azzalini A (1997) Applied smoothing techniques for data analysis: the kernel approach with S-plus illustrations, vol 18. Oxford University Press, Oxford. https://doi.org/10.1007/s001800000033
Burrage K, Lenane I, Lythe G (2007) Numerical methods for second-order stochastic differential equations. SIAM J Sci Comput 29(1):245–264. https://doi.org/10.1137/050646032
Capiez-Lernout E, Soize C (2022) Nonlinear stochastic dynamics of detuned bladed disks with uncertain mistuning and detuning optimization using a probabilistic machine learning tool. Int J Non-Linear Mech 143:104023. https://doi.org/10.1016/j.ijnonlinmec.2022.104023
Cappé O, Garivier A, Maillard O-A, Munos R, Stoltz G et al (2013) Kullback–Leibler upper confidence bounds for optimal sequential allocation. Ann Stat 41(3):1516–1541. https://doi.org/10.1214/13.AOS1119
Carlin BP, Louis TA (2008) Bayesian methods for data analysis. Chapman and Hall
Congdon P (2007) Bayesian statistical modelling, vol 704. Wiley
Cover TM, Thomas JA (2006) Elements of information theory, 2nd edn. John, Hoboken
Dashti M, Stuart AM (2017) The Bayesian approach to inverse problems. In: Ghanem R, Higdon D, Houman O (eds) Handbook of uncertainty quantification, Ch. 10. Springer, Cham, pp 311–428. https://doi.org/10.1007/978-3-319-12385-1_7
Depraetere N, Vandebroek M (2017) A comparison of variational approximations for fast inference in mixed logit models. Comput Stat 32(1):93–125. https://doi.org/10.1007/s00180-015-0638-y
Dieudonné J (1978) Treatise on analysis, vol 6. Academic Press, New York
Farhat C, Tezaur R, Chapman T, Avery P, Soize C (2019) Feasible probabilistic learning method for model-form uncertainty quantification in vibration analysis. AIAA J 57(11):4978–4991. https://doi.org/10.2514/1.J057797
Fearnhead P (2006) Exact and efficient Bayesian inference for multiple changepoint problems. Stat Comput 16(2):203–213. https://doi.org/10.1007/s11222-006-8450-8
Filippi S, Cappé O, Garivier A (2010) Optimism in reinforcement learning and Kullback–Leibler divergence. In: Proceedings of the 48th annual Allerton IEEE conference on communication, control, and computing, pp 115–122
Gelfand IM, Vilenkin NI (1964) Generalized functions. Volume 4. Applications of harmonic analysis, vol 380. AMS Chelsea Publishing
Gentle JE (2019) Computational statistics. Springer, New York. https://doi.org/10.1007/978-0-387-98144-4
Ghanem R, Soize C (2018) Probabilistic nonconvex constrained optimization with fixed number of function evaluations. Int J Numer Methods Eng 113(4):719–741. https://doi.org/10.1002/nme.5632
Ghanem R, Higdon D, Owhadi H (2017) Handbook of uncertainty quantification, vol 1 to 3. Springer, Cham. https://doi.org/10.1007/978-3-319-12385-1
Ghanem R, Soize C, Safta C, Huan X, Lacaze G, Oefelein JC, Najm HN (2019) Design optimization of a scramjet under uncertainty using probabilistic learning on manifolds. J Comput Phys 399:108930. https://doi.org/10.1016/j.jcp.2019.108930
Ghanem R, Soize C, Mehrez L, Aitharaju V (2022) Probabilistic learning and updating of a digital twin for composite material systems. Int J Numer Methods Eng 123(13):3004–3020. https://doi.org/10.1002/nme.6430
Girolami M, Calderhead B (2011) Riemann manifold Langevin and Hamiltonian Monte Carlo methods. J R Stat Soc 73(2):123–214. https://doi.org/10.1111/j.1467-9868.2010.00765.x
Givens G, Hoeting J (2013) Computational statistics, 2nd edn. Wiley, Hoboken
Golightly A, Wilkinson DJ (2006) Bayesian sequential inference for nonlinear multivariate diffusions. Stat Comput 16(4):323–338. https://doi.org/10.1007/s11222-006-9392-x
Golub GH, Van Loan CF (1993) Matrix computations, 2nd edn. Johns Hopkins University Press, Baltimore
Guilleminot J, Dolbow JE (2020) Data-driven enhancement of fracture paths in random composites. Mech Res Commun 103:103443. https://doi.org/10.1016/j.mechrescom.2019.103443
Guilleminot J, Soize C (2013) Stochastic model and generator for random fields with symmetry properties: application to the mesoscopic modeling of elastic random media. Multiscale Model Simul (A SIAM Interdiscipl J) 11(3):840–870. https://doi.org/10.1137/120898346
Hairer E, Lubich C, Wanner G (2003) Geometric numerical integration illustrated by the Störmer–Verlet method. Acta Numer 12:399–450. https://doi.org/10.1017/S0962492902000144
Kaipio J, Somersalo E (2005) Statistical and computational inverse problems, vol 160. Springer. https://doi.org/10.1007/b138659
Kapur JN, Kesavan HK (1992) Entropy optimization principles with applications. Academic Press, San Diego
Kelley CT (2003) Solving nonlinear equations with Newton’s method. SIAM. https://doi.org/10.1137/1.9780898718898
Kennedy MC, O’Hagan A (2001) Bayesian calibration of computer models. J R Stat Soc Ser B (Stat Methodol) 63(3):425–464. https://doi.org/10.1111/1467-9868.00294
Kloeden P, Platen E (1992) Numerical solution of stochastic differentials equations. Springer, Heidelberg
Krée P, Soize C (1986) Mathematics of random phenomena. Reidel Pub. Co (first published by Bordas in 1983 and also published by Springer in 2012)
Kullback S, Leibler RA (1951) On information and sufficiency. Ann Math Stat 22(1):79–86. https://doi.org/10.1214/aoms/1177729694
Luenberger DG (2009) Optimization by vector space methods. Wiley, New York
Marin J, Pudlo P, Robert C, Ryder R (2012) Approximate Bayesian computational methods. Stat Comput 22(6):1167–1180. https://doi.org/10.1007/s11222-011-9288-2
Marzouk YM, Najm HN, Rahn LA (2007) Stochastic spectral methods for efficient Bayesian solution of inverse problems. J Comput Phys 224(2):560–586. https://doi.org/10.1016/j.jcp.2006.10.010
Matthies HG, Zander E, Rosić BV, Litvinenko A, Pajonk O (2016) Inverse problems in a Bayesian setting. In: Computational methods for solids and fluids, vol. 41. Springer, pp 245–286. https://doi.org/10.1007/978-3-319-27996-1_10
Neal R (2011) MCMC using Hamiltonian dynamics. In: Brooks S, Gelman A, Jones G, Meng X-L (eds) Handbook of Markov chain Monte Carlo, Ch. 5. CRC Press, Boca Raton, pp 1–51. https://doi.org/10.1201/b10905-6
Neil M, Tailor M, Marquez D (2007) Inference in hybrid Bayesian networks using dynamic discretization. Stat Comput 17(3):219–233. https://doi.org/10.1007/s11222-007-9018-y
Owhadi H, Scovel C, Sullivan T (2015) On the brittleness of Bayesian inference. SIAM Rev 57(4):566–582. https://doi.org/10.1137/130938633
Perrin G, Soize C (2020) Adaptive method for indirect identification of the statistical properties of random fields in a Bayesian framework. Comput Stat 35(1):111–133. https://doi.org/10.1007/s00180-019-00936-5
Perrin G, Soize C, Ouhbi N (2018) Data-driven kernel representations for sampling with an unknown block dependence structure under correlation constraints. Comput Stat Data Anal 119:139–154. https://doi.org/10.1016/j.csda.2017.10.005
Picchini U, Samson A (2018) Coupling stochastic em and approximate Bayesian computation for parameter inference in state-space models. Comput Stat 33(1):179–212. https://doi.org/10.1007/s00180-017-0770-y
Robert C, Casella G (2005) Monte Carlo statistical methods. Springer. https://doi.org/10.1007/978-1-4757-4145-2
Saleem N, Ijaz G (2018) Low rank sparse decomposition model based speech enhancement using gammatone filterbank and Kullback–Leibler divergence. Int J Speech Technol 21(2):217–231. https://doi.org/10.1007/s10772-018-9500-2
Sambasivan R, Das S, Sahu SK (2020) A Bayesian perspective of statistical machine learning for big data. Comput Stat 35(3):893–930. https://doi.org/10.1007/s00180-020-00970-8
Scott SL, Blocker AW, Bonassi FV, Chipman HA, George EI, McCulloch RE (2016) Bayes and big data: the consensus Monte Carlo algorithm. Int J Manag Sci Eng Manag 11(2):78–88. https://doi.org/10.1080/17509653.2016.1142191
Shen Y, Cornford D, Opper M, Archambeau C (2012) Variational Markov chain Monte Carlo for Bayesian smoothing of non-linear diffusions. Comput Stat 27(1):149–176. https://doi.org/10.1007/s00180-011-0246-4
Shohat JA, Tamarkin JD (1943) The problem of moments. A mathematical surveys and monographs, vol 1. American Mathematical Society (RI)
Soize C (1993) Mathematical methods in signal analysis (in French, Méthodes Mathématiques en Analyse du Signal). Masson, Paris
Soize C (1994) The Fokker–Planck equation for stochastic dynamical systems and its explicit steady state solutions, vol. series on advances in mathematics for applied sciences, vol 17. World Scientific, Singapore. https://doi.org/10.1142/2347
Soize C (2006) Non Gaussian positive-definite matrix-valued random fields for elliptic stochastic partial differential operators. Comput Methods Appl Mech Eng 195(1–3):26–64. https://doi.org/10.1016/j.cma.2004.12.014
Soize C (2008a) Construction of probability distributions in high dimension using the maximum entropy principle. Applications to stochastic processes, random fields and random matrices. Int J Numer Methods Eng 76(10):1583–1611. https://doi.org/10.1002/nme.2385
Soize C (2008b) Tensor-valued random fields for meso-scale stochastic model of anisotropic elastic microstructure and probabilistic analysis of representative volume element size. Probab Eng Mech 23(2–3):307–323. https://doi.org/10.1016/j.probengmech.2007.12.019
Soize C (2011) A computational inverse method for identification of non-Gaussian random fields using the Bayesian approach in very high dimension. Comput Methods Appl Mech Eng 200(45–46):3083–3099. https://doi.org/10.1016/j.cma.2011.07.005
Soize C (2015) Polynomial chaos expansion of a multimodal random vector. SIAM-ASA J Uncertain Quantif 3(1):34–60. https://doi.org/10.1137/140968495
Soize C (2017) Uncertainty quantification. An accelerated course with advanced applications in computational engineering. Springer, New York. https://doi.org/10.1007/978-3-319-54339-0
Soize C (2021) Stochastic elliptic operators defined by non-Gaussian random fields with uncertain spectrum. The American Mathematical Society Journal. Theory Probab Math Stat 105:113–136. https://doi.org/10.1090/tpms/1159
Soize C (2022) Probabilistic learning inference of boundary value problem with uncertainties based on Kullback–Leibler divergence under implicit constraints. Comput Methods Appl Mech Eng 395:115078. https://doi.org/10.1016/j.cma.2022.115078
Soize C, Ghanem R (2016) Data-driven probability concentration and sampling on manifold. J Comput Phys 321:242–258. https://doi.org/10.1016/j.jcp.2016.05.044
Soize C, Ghanem R (2020a) Physics-constrained non-Gaussian probabilistic learning on manifolds. Int J Numer Methods Eng 121(1):110–145. https://doi.org/10.1002/nme.6202
Soize C, Ghanem R (2020b) Probabilistic learning on manifolds. Found Data Sci 2(3):279–307. https://doi.org/10.3934/fods.2020013
Soize C, Ghanem R (2021) Probabilistic learning on manifolds constrained by nonlinear partial differential equations for small datasets. Comput Methods Appl Mech Eng 380:113777. https://doi.org/10.1016/j.cma.2021.113777
Soize C, Ghanem R (2022) Probabilistic learning on manifolds (PLoM) with partition. Int J Numer Methods Eng 123(1):268–290. https://doi.org/10.1002/nme.6856
Soize C, Poloskov IE (2012) Time-domain formulation in computational dynamics for linear viscoelastic media with model uncertainties and stochastic excitation. Comput Math Appl 64(11):3594–3612. https://doi.org/10.1016/j.camwa.2012.09.010
Soize C, Ghanem R, Desceliers C (2020) Sampling of Bayesian posteriors with a non-Gaussian probabilistic learning on manifolds from a small dataset. Stat Comput 30(5):1433–1457. https://doi.org/10.1007/s11222-020-09954-6
Spall JC (2005) Introduction to stochastic search and optimization: estimation, simulation, and control, vol 65. Wiley
Spantini A, Cui T, Willcox K, Tenorio L, Marzouk Y (2017) Goal-oriented optimal approximations of Bayesian linear inverse problems. SIAM J Sci Comput 39(5):S167–S196. https://doi.org/10.1137/16M1082123
Stuart AM (2010) Inverse problems: a Bayesian perspective. Acta Numer 19:451–559. https://doi.org/10.1017/S0962492910000061
Talay D (2002) Stochastic Hamiltonian systems: exponential convergence to the invariant measure, and discretization by the implicit Euler scheme. Markov Process Relat Fields 8(2):163–198
Talay D, Tubaro L (1990) Expansion of the global error for numerical schemes solving stochastic differential equations. Stoch Anal Appl 8(4):483–509. https://doi.org/10.1080/07362999008809220
Vasconcelos N, Ho P, Moreno P (2004) The Kullback–Leibler kernel as a framework for discriminant and localized representations for visual recognition. In: Proceedings of the European Conference on Computer Vision, pp 430–441. https://doi.org/10.1007/978-3-540-24672-5_34
Zhang W, Shan S, Chen X, Gao W (2007) Local Gabor binary patterns based on Kullback–Leibler divergence for partially occluded face recognition. IEEE Signal Process Lett 14(11):875–878. https://doi.org/10.1109/LSP.2007.903260
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The author declares that he has no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix A. Generation of the training set, target set, and numerical values of the parameters
Appendix A. Generation of the training set, target set, and numerical values of the parameters
The training set \(D_d = \{ {\varvec{x}}^1,\ldots , {\varvec{x}}^{N_d}\}\) with is made up of \(N_d\) independent realizations of random variable \({\varvec{X}}=({\varvec{Q}},{\varvec{W}})\), which are generated by using a stochastic computational model corresponding to the finite element discretization of a stochastic elliptic boundary value problem for which \(n_x=430{,}098\), \(n_q=10{,}098\), and \(n_w= 420{,}000\). The target set \(D_{\textrm{targ}}= \{ {\varvec{q}}^1_{\textrm{targ}},\ldots , {\varvec{q}}^{N_r}_{\textrm{targ}}\}\) is generated using the stochastic computational model with another values of the parameters (see “Appendix A.3”).
1.1 Appendix A.1. Definition of the stochastic boundary value problem
Let \(\Omega = ]\,0\, , 1\,[\, \times \, ]\,0\, , 0.2\,[\,\times \,]\,0 \, , 0.1\,[ \, m^3\) be the bounded open set of \({\mathbb {R}}^3\), with generic point \({\varvec{\omega }}= (\omega _1,\omega _2,\omega _3)\), and with boundary \(\partial \Omega =\Gamma _0 \cup \Gamma _1\cup \Gamma _2\) in which \(\Gamma _0 =\{\omega _1=1\, , \, 0\le \omega _2 \le 0.2 \, , \, 0 \le \omega _3 \le 0.1 \}\), \(\Gamma _1 =\{\omega _1=0\, , \, 0\le \omega _2 \le 0.2 \, , \, 0 \le \omega _3 \le 0.1 \}\), and \(\Gamma _2 =\partial \Omega \backslash \{\Gamma _0\cup \Gamma _1 \}\). Let be \({\overline{\Omega }}=\Omega \cup \partial \Omega \). The outward unit normal to \(\partial \Omega \) is denoted by . We use the usual convention of summation on repeated Latin indices. Domain \(\Omega \) is occupied by a heterogeneous and anisotropic elastic random medium for which the elastic properties are defined by the fourth-order tensor-valued non-Gaussian random field . Let be the \({\mathbb {R}}^3\)-valued displacement random field defined in \(\Omega \). A Dirichlet condition is given on \(\Gamma _0\) while a Neumann condition is given on \(\Gamma _1\cup \Gamma _2\). The stochastic boundary value problem is written, for \(k=1,2,3\) and almost surely, as
in which the stress tensor is related to the strain tensor by by the constitutive equation, . For \(k=1,2,3\), the applied stresses \(p_k\) on \(\Gamma _1\) are defined as follows:
\(p_1 = 0\) on \(\Gamma _1\), except:
\(\quad p_1 = -1.8\times 10^8\, N/m^2\) for \({\varvec{\omega }}\in \{\omega _1=0\, , \, 0\le \omega _2\le 0.02 \, , \, 0\le \omega _3 \le 0.1\}\).
\(\quad p_1 = +9.0\times 10^7\, N/m^2\) for \({\varvec{\omega }}\in \{\omega _1=0\, , \, 0.18\le \omega _2\le 0.2 \, , \, 0\le \omega _3 \le 0.1\}\).
\(p_2 = 0\) on \(\Gamma _1\), except:
\(\quad p_2 = +1.0\times 10^7\, N/m^2\) for \({\varvec{\omega }}\in \{\omega _1=0\, , \, \{0\le \omega _2\le 0.02\}\cup \{0.18\le \omega _2\le 0.20\} \, , \, 0\le \omega _3 \le 0.02\}\).
\(\quad p_2 = -1.5\times 10^7\, N/m^2\) for \({\varvec{\omega }}\in \{\omega _1=0\, , \, \{0\le \omega _2\le 0.02\}\cup \{0.18\le \omega _2\le 0.20\} \, , \, 0.08\le \omega _3 \le 0.1\}\).
\(p_3 = 0\) on \(\Gamma _1\), except:
\(\quad p_3 = -2.40\times 10^7\, N/m^2\) for \({\varvec{\omega }}\in \{\omega _1=0\, , \, 0\le \omega _2\le 0.02 \, , \, 0\le \omega _3 \le 0.1\}\).
\(\quad p_3 = +2.64\times 10^7\, N/m^2\) for \({\varvec{\omega }}\in \{\omega _1=0\, , \, 0.18\le \omega _2\le 0.2 \, , \, 0\le \omega _3 \le 0.1\}\).
Using the matrix representation in Voigt notation, the random elasticity field is rewritten, for k, m, n, and q in \(\{1,2,3\}\), as with \({{\textbf{i}}}=(k,m)\) with \(1\le k\le m\le 3\) and \({{\textbf{j}}}=(n,q)\) with \(1\le n \le q \le 3\) in which indices \({{\textbf{i}}}\) and \({{\textbf{j}}}\) belong to \(\{1,\ldots , 6\}\). The -valued random field \(\{ [{\varvec{A}}({\varvec{\omega }})] ,{\varvec{\omega }}\in \Omega \}\) is a non-Gaussian, second order, and statistically homogeneous. Its mean function is the given \({\varvec{\omega }}\)-independent matrix corresponding to a homogeneous isotropic elastic material whose Young modulus is \(10^{10}\, N/m^2\) and Poisson coefficient 0.15 (note that the fluctuations around the mean are those of a heterogeneous anisotropic elastic material). The non-Gaussian -valued random field \(\{ [{\varvec{A}}({\varvec{\omega }})]\, ,{\varvec{\omega }}\in \Omega \}\) is constructed using the stochastic model (Soize 2006, 2008b, 2017) of random elasticity fields for heterogeneous anisotropic elastic media that are isotropic in statistical mean and exhibit anisotropic statistical fluctuations, for which the parameterization consists of spatial-correlation lengths and of a positive-definite lower bound. The random field \(\{[{\varvec{A}}({\varvec{\omega }})],{\varvec{\omega }}\in \Omega \}\) is written as,
in which is the upper triangular \((6\times 6)\) real matrix such that , where \(\epsilon \) is a given positive number (which can be chosen arbitrarily small), and where \(\{ [{\varvec{G}}({\varvec{\omega }})],{\varvec{\omega }}\in {\mathbb {R}}^3\}\) is a -valued random field (by construction), defined on , indexed by \({\mathbb {R}}^3\). Then \( [{\varvec{G}}]\) is homogeneous, mean-square continuous, and such that \(E\{[{\varvec{G}}({\varvec{\omega }}))]\} = [I_6]\) for all \({\varvec{\omega }}\in {\mathbb {R}}^3\). Note that the lower bound \(\epsilon \,[\,{\underline{{\varvec{A}}}}\, ]/(1+\epsilon )\) used in Eq. (A.5) could be replaced by a more general lower bound \([A_b]\) in as proposed in Guilleminot and Soize (2013) and Soize (2017). For all \({\varvec{\omega }}\) fixed in \({\mathbb {R}}^3\), the -valued random variable \([{\varvec{G}}({\varvec{\omega }})]\) has been constructed by using the Maximum Entropy Principle under the following available information, \(E\{[{\varvec{G}}({\varvec{\omega }})]\} = [I_6]\) and \(E\{ \log (\det [{\varvec{G}}({\varvec{\omega }})] ) \} = b_G\) with \(\vert b_G \vert \, < +\infty \), which has been introduced in order that the random matrix \([{\varvec{G}}({\varvec{\omega }})]^{-1}\) (that exists almost surely) be such that \(E\{\Vert [{\varvec{G}}({\varvec{\omega }})]^{-1}\Vert ^2\} \le \) \(E\{\Vert [{\varvec{G}}({\varvec{\omega }})]^{-1}\Vert _F^2\} < +\infty \). In this construction, for all \({\varvec{\omega }}\) fixed in \({\mathbb {R}}^3\), is a -valued nonlinear function [g(.)] of \(6\times (6+1)/2 = 21\) independent normalized Gaussian real-valued random variables denoted by and such that and . The spatial correlation structure of random field \(\{[{\varvec{G}}({\varvec{\omega }})],\) \({\varvec{\omega }}\in {\mathbb {R}}^3\}\) is introduced by considering 21 independent real-valued random fields for \(1\le m \le n \le 6\), corresponding to 21 independent copies of a unique normalized Gaussian homogeneous mean-square continuous real-valued random field whose normalized spectral measure is given and has a support that is controlled by three spatial correlation lengths \(L_{c1} = L_{c2} = L_{c3} = 0.4\). Note that this Gaussian field can be replaced by a non-Gaussian field for taking into account uncertainties in the spectral measure (Soize 2021). The constant \(b_G\) is eliminated in favor of a hyperparameter \(\delta _G > 0\), which allows for controlling the level of statistical fluctuations of \([{\varvec{G}}({\varvec{\omega }})]\), defined by \(\delta _G =(E\{\Vert [{\varvec{G}}({\varvec{\omega }})] - [I_6]\Vert _F^2\} / 6)^{1/2}\), which is independent of \({\varvec{\omega }}\) and such that \(\delta _G= 0.6\).
1.2 Appendix A.2. Stochastic computational model for generating the training set \(D_d\) and observed quantities of interest
The stochastic boundary value problem defined by Eqs. (A.1) to (A.4) is discretized by the finite element method. Domain \(\Omega \) is meshed with \(50\times 10\times 5 = 2500\) finite elements using 8-nodes finite elements. There are 3366 nodes and 10, 098 dofs (degrees of freedom). The displacements are locked at all the 66 nodes belonging to surface \(\Gamma _0\) and therefore, there are 198 zero Dirichlet conditions. There are 8 integration points in each finite element. Consequently, there are \(N_p= 20{,}000\) integration points \({\varvec{\omega }}^1,\ldots , {\varvec{\omega }}^{N_p}\). The -valued random variable \({\varvec{W}}\) is generated as follows. For all \(p=1,\ldots , N_p\), let in which \(\log _M\) is the logarithm of positive-definite matrices. The -valued random variable \({\varvec{W}}\) is then defined as the vector that is the reshaping of the upper triangular part of the \(N_p\) matrices \(\{\, [{\varvec{G}}_p^{\textrm{log}}], p=1,\ldots , N_p\}\).We then have \(n_w = 21\times N_p = 420{,}000\). The finite element discretization of random field is the -valued random variable \({\varvec{Q}}\) with \(n_q= 10{,}098\). Consequently \({\varvec{X}}=({\varvec{Q}},{\varvec{W}})\) is a random variable with values in with \(n_x=n_q+n_w = 430{,}098\). The stochastic computational model is then represented by a stochastic linear matrix equation that is solved by using the Monte Carlo numerical simulation method yielding the training set \(D_d = \{ {\varvec{x}}^1,\ldots , {\varvec{x}}^{N_d}\}\) in which is a realization of random variable \({\varvec{X}}=({\varvec{Q}},{\varvec{W}})\), the computed realizations being independent. For studying the convergence properties, the considered values of \(N_d\) are \(N_d\in \{100, 200, 300, 400 \}\).
The components of the quantity of interest \({\varvec{Q}}\), which will be observed for presenting the results, are the 3 components denoted by \(Q_{{\textrm{obs}},1}\), \(Q_{{\textrm{obs}},2}\), and \(Q_{{\textrm{obs}},3}\) that correspond to the 3 dofs along directions \(\omega _1\), \(\omega _2\), and \(\omega _3\) of the finite element node of coordinates (0, 0, 0.1) (located at top corner in which the displacements are significant and result from tension, torsion, and two bendings contributions).
1.3 Appendix A.3. Target set of realizations
The target set \(D_{\textrm{targ}}= \{{\varvec{q}}_{\textrm{targ}}^1,\ldots {\varvec{q}}_{\textrm{targ}}^{N_r}\}\) is generated using the stochastic boundary value problem defined in Section Appendix A.1 for which the elasticity matrix \([{\underline{{\varvec{A}}}}^{\textrm{targ}}]\) is the one of a homogeneous and isotropic elastic material with a Young modulus \(9\times 10^9\, \hbox {N/m}^2\) and a Poisson coefficient \(\nu =0.15\). The level of statistical fluctuations of the random field \(\{{\varvec{G}}^{\textrm{targ}}({\varvec{\omega }}),{\varvec{\omega }}\in {\mathbb {R}}^3\}\) is \(\delta _G^{\textrm{targ}}= 0.3\). In order to analyze the convergence with respect to \(N_r\), we have considered, in consistency with the values of \(N_d\), the intervals \(N_r \in [50\, , N_{\textrm{targ}}]\) with \(N_{\textrm{targ}}\in \{100,200,300,400\}\).
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Soize, C. Probabilistic learning constrained by realizations using a weak formulation of Fourier transform of probability measures. Comput Stat 38, 1879–1925 (2023). https://doi.org/10.1007/s00180-022-01300-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00180-022-01300-w