Abstract
A fuzzy arithmetic framework for the efficient possibilistic propagation of shape uncertainties based on a novel fuzzy edge detection method is introduced. The shape uncertainties stem from a blurred image that encodes the distribution of two phases in a composite material. The proposed framework employs computational homogenisation to upscale the shape uncertainty to a effective material with fuzzy material properties. For this, many samples of a linear elasticity problem have to be computed, which is significantly sped up by a highly accurate low-rank tensor surrogate. To ensure the continuity of the underlying mapping from shape parametrisation to the upscaled material behaviour, a diffeomorphism is constructed by generating an appropriate family of meshes via transformation of a reference mesh. The shape uncertainty is then propagated to measure the distance of the upscaled material to the isotropic and orthotropic material class. Finally, the fuzzy effective material is used to compute bounds for the average displacement of a non-homogenized material with uncertain star-shaped inclusion shapes.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Composite materials are ubiquitous in many applications. Whether they are formed by chance (like with unfavourable impurities in metal) or by design (like pebbles in concrete), the mechanical properties of the whole material are determined by the resulting composite structure [1]. The different types of composite materials—e.g. metal matrix composites [2], fiber-reinforced polymers [3], composite wood [4] and other advanced composite materials [5]—find their use in a wide range of applications, e.g. masonry, aerospace industry, wind power plants and sports equipment. In many applications, the added composite material improves a base matrix material in terms of wear resistance, damping properties and mechanical strength, while keeping the same weight. However, such improved composite materials are a result of empirical studies, experimentation and chance. It hence is obvious that a thorough understanding of composite materials via theoretical and accurate computational models is desirable to predict and systematically improve the properties and applicability of composites.
What makes this task challenging are the influence of material behaviour on the micro- and macro scales, as well as the multifaceted sources of uncertainties. For instance, the involved length scales may span up to ten orders of magnitudes, i.e. the size of a nano particle of \(10^{-6}\,\textrm{m}\) is embedded in a material with a length scale of \(10^{-1}{-}10^{1}\,\textrm{m}\) [2]. The same holds for the material constants if the constituents are fundamentally different as in metals and polymers. A standard finite element approach becomes very costly in such a setting since the details of the composite have to be resolved by the mesh for accurate simulations. Additionally, one has to handle the uncertainty of the material constants such as the position, form and size of the inclusions. This increases the costs even further since many realisations are required to correctly capture the uncertainty statistically. Each new realisations requires a costly (possible automatic) remeshing of the computational domain. The mapping from geometric parameters—namely position and shape of each inclusion—onto the generated meshes are usually discontinuous. This introduces unwanted effects that we tackle in Sect. 4.1.
The first challenge—the span of scales—is tackled by homogenisation methods, a term introduced by Babuška in [6]. In some form or another, these methods incorporate micro-scale behaviour into an adjusted macro-scale model, replacing the full composite material model by a corrected homogeneous one, obtained by an asymptotic limit of an assumed (periodic) domain. The central idea is to derive equations that describe the effective material properties analytically. Alternatively, computational approaches have been devised to solve particular micro-scale problems to deduce the adjusted macro-scale behaviour numerically [7, 8]. Some notable examples are stochastic homogenisation, see [9,10,11,12], projection based homogenisation [13] and (stochastic) representative volume element methods [14].
The second challenge—the uncertainty of the material—is sometimes neglected [15, 16] by only considering deterministic material properties. If the probability distribution of the material constants is known, the uncertainty can be modelled with precise probabilities [17]. Ignoring the uncertainty seems valid if it has little influence on the system’s response or the model data is known sufficiently accurate and is free of inherent fluctuations. Using precise probabilities is valid if the distributions of the material constants are known precisely. However, as was pointed out by Motamed and Babuška [18], stochastic models based on precise probabilities are not always able to model the uncertainty in composite materials. Instead, they propose a model based on an imprecise probability theory. Examples for this are evidence theory [19], random set theory [20], possibility theory [21] and—more recently—optimal Uncertainty Quantification (UQ) [22]. In comparison to precise probability, imprecise probability methods are able to provide estimates of uncertainties based only on a small set of data and few assumptions, which would be far too limited for a probabilistic method. For an illuminating work that dissects the difference between precise and imprecise probabilities, we refer the reader to [23]. As a motivation the approach developed in what follows, we especially point to the false confidence theorem, which qualitatively states that “probability dilution is a symptom of a fundamental deficiency in probabilistic representations of statistical inference, in which there are propositions that will consistently be assigned a high degree of belief, regardless of whether or not they are true” [23].
In this work, we consider a practical problem that is supposed to illustrate when imprecise probabilities are appropriate (in particular more so than probabilistic methods), the challenges performing a non-probabilistic uncertainty propagation and quantification, and ways to overcome the computational difficulties. We assume a material consisting of two phases, namely a soft matrix phase and a hard inclusion phase. Both phases are linear elastic with precisely known Young’s modulus and Poisson ratio. The inclusions repeat periodically in a checkerboard fashion. Consequently, a classical numerical homogenisation method would yield a homogenised (globally constant) macro-scale material. Considering real applications, the main problem stems from the unknown shape of the inclusions. Often, the shape is retrieved with a non-intrusive imaging process, e.g. computed tomography scans or magnetic resonance imaging. To illustrate this, Fig. 1 shows a computed tomography scan of an imperfect adhesive bond within a Henkel beam, cf. [24]. Apparently, the image is noisy, blurred, pixelated and exhibits artefacts. While it is possible to de-noise and de-blurr the image with image post-processing, which might make the identification of inclusions possible, different numerical methods yield different shapes [25]. This reveals the shape uncertainty inherent to the image. Here, we just focus on one blurred image per inclusion, assuming that the artefacts and noise were successfully removed already, see Fig. 3 for an example. Opposite to this, in the case of multiple CT scans of the same inclusion, the additional information can be employed to reduce the shape uncertainty and thus refine the epistemic uncertainty model until the uncertainty becomes irreducible and hereby aleatoric uncertainty models are preferable.
To tackle a problem in this setting, we introduce a novel fuzzy edge detection based on possibility theory, presented in Sect. 2.1 and a restriction of possible interface boundaries via bounded total variation in Sect. 3. This yields a fuzzy model of the boundary which in turn introduces a computational challenge in terms of a numerical homogenisation problem as discussed in Sect. 2.2. For this, the micro-model has to be solved very often in order to propagate the uncertainty of the boundary to the homogenised material model. To alleviate this expensive task, we introduce a highly accurate rank adaptive low-rank tensor surrogate in Sect. 4. In the numerical experiments Sect. 5, the surrogate model is validated numerically. Moreover, we measure the distance of the homogenised material to the class of isotropic and orthotropic elastic tensors and eventually use the homogenised material to perform a worst/best case analysis for a full matrix composite model with 64 inclusions, i.e., with help of the homogenised material we will find bounds for the average displacement of the non-homogenised material.
2 Basics
This section serves as a brief introduction of three fundamental topics that form the basis of this work. First, fuzzy set theory which is used to model the uncertainty ingrained in the blurred image. Second, the computational up-scaling method (numerical homogenisation). Third, classes of constitutive tensors to measure the distance of the up-scaled material to the isotropic and orthotropic material are discussed.
2.1 Fuzzy set theory
We introduce a possibilistic framework with fuzzy sets based on 4 central definitions, see e.g. [26] or more recently [27,28,29].
Fuzzy sets are common sets that are equipped with a membership function which assigns each element in the set a value between zero and one. This values only task is to communicate a degree of belongingness to the set. The meaning of this value depends on the problem and—more importantly—on the community using this value to formalise uncertainty. Arguably, this ambiguity is a feature and not a short-coming of this theory since it avoids assumptions that are usually made by other approaches. In probability theory for instance, one has to assume a prior distribution before updating the posterior distribution with new samples. The following definition introduces the terminology of fuzzy sets.
Definition 1
(Fuzzy set/variable, \(\alpha \)-cuts and interactivity) Let \(Z\ne \emptyset \) be a set and \(\mu :Z\rightarrow [0,1]\) be a map such that there exists \(z\in Z\) with \(\mu (z)=1\). The map \(\mu \) is called (normalised) membership function. We define a (normalised) fuzzy set \(\tilde{z}\) on Z by
If \(\mu (Z) =\{0,1\}\) with unique \(z^*\in Z\) and \(\mu (z^*)=1\) then \(\tilde{z}\) is called a crisp set. Furthermore, we denote by \(\mathcal {F}(Z)\) the set of all fuzzy sets on Z. We thus simply write \(\tilde{z}\in \mathcal {F}(Z)\). For each fuzzy set \(\tilde{z}\in \mathcal {F}(Z)\) we denote the associated membership function by \(\mu _{\tilde{z}}\). If \(Z\subset \mathbb {R}^N\) for \(N\in \mathbb {N}\), we call \(\tilde{z}\in \mathcal {F}(Z)\) a fuzzy variable \((N=1)\) or vectorial fuzzy variable \((N>1)\) described by a (joint) membership function \(\mu _{\tilde{z}}\). The support of \(\mu _{\tilde{z}}\) is defined as
For \(\alpha \in [0,1]\), the \(\alpha \)-cut \(C_\alpha \) of \(\mu _{\tilde{z}}\) is defined as
Let \(\tilde{z}_i\in \mathcal {F}(Z_i)\) for sets \(Z_i\) with \(i=1,\ldots ,M<\infty \), and \(\tilde{z}=(\tilde{z}_1,\ldots , \tilde{z}_M)\). If the joint membership function associated with \(\tilde{z}\) has the form \(\mu _{\tilde{z}}=\min _i \mu _{\tilde{z}_i}\) then \(\tilde{z}\) is called non-interactive and interactive otherwise.
Fuzzy sets are quite versatile since there are no restrictions on the (type and structure of the) used sets. However, for numerical methods to become efficient, certain assumptions are beneficial. The first restriction mimics numbers and vectors.
Definition 2
(Fuzzy number/vector) Let \(\tilde{z}\in \mathcal {F}(Z)\) with \(Z\subset \mathbb {R}^N\) for some \(N\in \mathbb {N}\) such that Z is bounded and convex and the (joint) membership function \(\mu _{\tilde{z}}\) is upper semi-continuous, i.e.
and quasi-concave, i.e.
If there exists a unique \(z^*\in Z\) such that \(\mu _{\tilde{z}}(z^*)=1\) then we call \(\tilde{z}\) a fuzzy number for \(n=1\) and a fuzzy vector for \(n>1\).
This notion is easily extended to intervals and domains.
Definition 3
(Fuzzy interval/domain) With the same assumptions as in Definition 2, there exists a subset \(S\subset Z\) such that \(\mu _{\tilde{z}}(z) = 1\) for all \(z\in S\). Then \(\tilde{z}\) is called a fuzzy interval for \(n=1\) and fuzzy domain for \(n>1\) .
Note that the quasi-concavity of the membership function implies convexity of any \(\alpha \)-cut \(C_\alpha \). In particular, given \(S\subset Z\) as in Definition 3, the convex hull \({\text {conv}}(S)\) is a proper subset of \(C_1\), too. The quasi-concavity also leads to the nestedness property of \(\alpha \)-cuts from Definitions 2 and 3, i.e.
This property is essential for the \(\alpha \)-cut propagation method, see Theorem 2. In the following example, we introduce the most common fuzzy structures. Namely, the fuzzy trapezoidal interval which we will use throughout this work.
Example 1
A particular class of fuzzy sets is the trapezoidal fuzzy set \(\tilde{z}\), forming a subset of \(\mathcal {F}(Z)\) with \(Z\subset \mathbb {R}\). The respective membership function \(\mu _{\tilde{z}}\) is described by two upper semi-continuous functions \(f_L:(-\infty ,0]\rightarrow [0,1]\) and \(f_R:[0,\infty )\rightarrow [0,1]\). Here, \(f_L(0) = f_R(0)=1\) with \(f_L\) \(( f_R )\) is monotonously increasing (decreasing) and \(\lim _{z\rightarrow -\infty }f_L(z)=0\) \(( \lim _{z\rightarrow \infty }f_R(z)=0 )\) such that there exist \(\ell ^*,r^*\in Z\) with \(z_\ell ^*\le z_r^*\) and
The triangle fuzzy number \(\tilde{z}=\langle \ell ,z^*,r \rangle \) specified by left and right limit \(\ell ,r\) and peak position \(z^*= z_\ell ^*= z_r^*\) is a special case of the trapezoidal fuzzy set. Figure 2 depicts the propagation of a trapezoidal fuzzy set.
For arbitrary fuzzy sets Zadeh’s extension principle is the way to propagate uncertainty through a mapping.
Theorem 1
(Zadeh’s extension principle [26])
Consider a function \(f:Z \rightarrow V\) with a non-empty set V. Let \(\tilde{z}\in \mathcal {F}(Z)\) with membership function \(\mu _{\tilde{z}}\). Define
with membership function \(\mu _{\tilde{f}}\) defined for all \( v\in V \) as
Then \(\tilde{f}\in \mathcal {F}(V)\) with membership function \(\mu _{\tilde{f}}\).
If more underlying structure is given, the extension principle can be formulated equivalently in terms of constrained optimization. Zadeh’s principle can be reformulated into the so-called \(\alpha \)-cut propagation for fuzzy vectors and fuzzy domains to reduce the computational costs.
Theorem 2
(\(\alpha \)-cut propagation [30])
Let \(f:Z\rightarrow V\) be continuous between metric spaces \((Z,d_1)\) and \((V,d_2)\) and let \(\tilde{z}\in \mathcal {F}(Z)\) with \(C_0[\tilde{z}]\subset K\subset Z\) for a compact set K with convex Z. Furthermore, let the membership function \(\mu _{\tilde{z}}\) be quasi-concave and upper semi-continuous. Then \(\mu _{\tilde{f}}\) can be characterised via \(\alpha \)-cuts as
Moreover, if \(V=\mathbb {R}\) then
Note that (9) follows from the assumption that \(C_\alpha [\tilde{z}]\) is closed and consequently compact as it is a closed subset of a compact set K in a metric space. Since f is assumed to be continuous, compact sets are mapped to compact sets.
Comparing Zadeh’s principle with the \(\alpha \)-cut propagation, we immediately see that the former requires the inverse image and an optimization step for each point v whereas the later only depends on two optimization steps of f for the number of \(\alpha \)-cuts. Since it is infeasible to perform Zadeh’s principle for all points \(v\in V\), it is combined or replaced with a sampling approach. We distinguish two variants to sample in V:
-
Semi sampling approach: directly solve the constrained optimization problem with a global optimiser. For a given sequence \((v_k)_k\subset V\), compute the supremum over \(Z_k:=\{z\in Z: f(z) = v_k\}\), see the red line in Fig. 2.
-
Full sampling approach: choose a sequence \((z_k)_k\subset Z\), compute \((f_k)_k = [f(z_k)]_k\) and \((\mu _k)_k = [\mu _{\tilde{z}}(z_k)]_k\). Use the data sample pairs \((v_k, \mu _k)\) and reconstruct \(\mu _{\tilde{f}}\), e.g. by convex hull or an envelope approach, see the orange/purple graphics in Fig. 2.
If f has high evaluation costs, the propagation inevitably becomes difficult and costly. This for example is the case if f represents a finite element solution with a large number of elements. Thus, to still render the propagation feasible the number of evaluations of f needs to be computationally feasible. In case of applicability of \(\alpha \)-cut propagation, the number of evaluations depends on the (global) optimization method. This number is possibly lower than the sampling counterparts. However, if the requirements of Theorem 2 are not met one may fall back to the sampling methods. By introducing a surrogate model \(f_h \approx f\), the computational costs can be further reduced for both strategies. In essence, this approach can be interpreted as a reduction of evaluations of f, determined by the number of evaluations needed to obtain \(f_h\), assuming that the evaluation costs of \(f_h\) are negligible.
2.2 Upscaling of heterogeneous linear elastic material
In the following we develop a numerical homogenisation method which yields a macroscopic material based on microscopic properties. The goal of the approach is to dispose of the computationally involved microstructure and construct an upscaled material surrogate with similar homogenised behaviour on a larger scale. Equivalent terms for “homogenised behaviour” are macroscopic, effective or upscaled behaviour [31], where the examined “behaviour” is subject to some quantity of interest (e.g. average displacement or stress) based on the system response.
In classical (asymptotic) homogenisation, an effective property is calculated based on the assumption of an infinite periodic domain [32]. The local microscopic structure is defined in terms of a representative volume element, which in our setting would consist of a single inclusion that has identical shape in the entire domain. For problems in non-periodic media, the methods of numerical homogenisation or numerical upscaling as e.g. described in [33] are used, where local boundary value problems are solved to calculate effective characteristics in each local domain [34], also see [9, 12, 35] for analytical stochastic homogenisation. Periodic boundary conditions (PBCs) are commonly used for numerical upscaling methods of matrix composite material [36, 37]. In this work, we focus on numerical homogenisation with PBCs for linear elasticity. Let \(D=[-1,1]^2\) be a unit cell domain and let the heterogeneous material law be encoded in a tensor \(\varvec{C}=\varvec{C}(x)\in \mathbb {R}^{2,2,2,2}, x\in D\). For given macroscopic strain \(\varvec{E}\) write the displacement \(\varvec{u}\) as \(\varvec{u}(x)=\varvec{E}\cdot x + \varvec{v}\) where the D-periodic fluctuation \(\varvec{v}\) solves
such that \(\varvec{\sigma }\cdot \textbf{n}\) is antiperiodic on D with \(\varvec{n}\) denoting the outer normal with respect to \(\partial D\). Let \(\langle \varvec{\epsilon }\rangle \) and \(\langle \varvec{\sigma }\rangle \) denote the average strain and stress of the computed strain \(\varvec{\epsilon }\) and stress \(\varvec{\sigma }\). Then, there exists a tensor \(\mathcal {H}\in \mathbb {R}^{2,2,2,2}\) satisfying
The tensor \(\mathcal {H}\) is called the effective or upscaled (macroscopic) tensor, representing the elastic moduli of the homogenised medium. By construction it holds that \(\varvec{E} = \langle \varvec{\epsilon }\rangle \). Consequently, \(\mathcal {H}\) can be obtained from (12) by choosing macro strains \(\varvec{E}\) corresponding to different elementary load cases and subsequently computing \(\langle \varvec{\sigma }\rangle \) by solving (11). In what follows, the homogenisation technique will be applied for every shape parametrisation encoded in \(\varvec{C}\).
2.3 Measuring the distance between constitutive tensors
Constitutive tensors \(\varvec{C}=(C_{ijkl})\in \mathbb {R}^{d,d,d,d}\) in the planar case of \(d=2\) determine the behaviour of the linear elastic material by relating stress and strain. Such a tensor is said to be in \(\mathbb {E}\textrm{la}(d)\) if the symmetry property
holds. A tensor \(\varvec{C}\in \mathbb {E}\textrm{la}(d)\) may exhibit further symmetry properties. If there exist Lamé constants \(\lambda , \mu \in \mathbb {R}\) such that
the tensor is called isotropic. For orthotropic tensors, we define a rotation matrix
and the full orthogonal group operation o(g) via
We then say that \(\varvec{C}_{\text {ortho}}\in \mathbb {E}\textrm{la}(d)\) is orthotropic if it can be represented as
for some isotropic \(\varvec{C}_{\text {iso}} \in \mathbb {E}\textrm{la}(d)\), \(\Lambda _1,\Lambda _2\in \mathbb {R}\), \(\theta \in [0,2\pi ]\) and forth order tensors \(\varvec{O}^1\) and \(\varvec{O}^2\) with a Kelvin-Mandel matrix representation [38] given as
Note that any isotropic \(\varvec{C}_{\text {iso}}\in \mathbb {E}\textrm{la}(d)\) is invariant under action of \(\rho (g)\), in particular \( \rho (g)\varvec{C}_{\text {iso}} = \varvec{C}_{\text {iso}}\). With that representation, any orthotropic tensor has 5 degrees of freedom \((\lambda , \mu , \Lambda _1, \Lambda _2, \theta )\). The orthotropic tensors represented in normal form have only 4 degrees of freedom since \(\theta \) is chosen as zero. Let \(\mathbb {E}\textrm{la}(d,\text {iso}), \mathbb {E}\textrm{la}(d,\text {ortho})\subset \mathbb {E}\textrm{la}(d)\) be the set of all isotropic and orthotropic tensors, respectively. Given any anisotropic tensor \(\varvec{C}\), we define the distance to the symmetry class of isotropic and orthotropic tensors by
with the Frobenius norm \(\Vert \cdot \Vert \). The distances to the symmetry classes can be characterised as follows.
Proposition 3
(Distance to isotropy class [39])
Let \(\varvec{C}\in \mathbb {E}\textrm{la}(d)\) for \(d=2\). Define \(\varvec{C}_{\text {iso}} = 2\mu \varvec{I}+\lambda \varvec{1}\otimes \varvec{1}\) with
as the orthogonal projection of \(\varvec{C}\) onto the isotropic material class. Then,
Proposition 4
(Distance to orthotropy class [39])
Let \(\varvec{C}\in \mathbb {E}\textrm{la}(d)\) for \(d=2\) and \((\theta _k)_{k=1}^K\), \(K\in \mathbb {N}\) be the finite roots of
with \(a = 4X_1Y_1\), \(b=2(Y_1^2-X_1^2)\), \(c=2X_2Y_2\) and \(d=Y_2^2-X_2^2\) and
For \(\theta \in [0,2\pi ]\), define
and
Then,
With the introduced material class concepts of constitutive tensors we are able to measure the distance of our effective material to the isotropic respectively orthotropic material class for each shape parametrisation.
3 Fuzzy edge detection
Given a blurred image as depicted in Fig. 3, the most common approach to reconstruct the original image is to use some method from the wide class of blind deconvolution methods [40]. These methods assume that an image y is the convolution of an original image x and some kernel k, distorted with additive noise n, i.e.,
Finding a pair (x, k) satisfying equation (17) is equivalent to de-blurring the image, leading to a original image x. This however is an ill-posed problem, since an infinite number of such pairs can be found [41], making some form of regularisation inevitable. The classical approaches assume zero noise and employ regularised least squared methods [42,43,44]. Each regularisation is based on assumptions of the kernel and the original image. More recently, natural image statistics and the Bayesian framework were used to formulate such assumptions more precisely and thus improved the performance of blind deconvolution methods, see [45,46,47] for a first overview. Simply put, prior knowledge and the ability to incorporate this knowledge improve the deconvolution result. However, we are in a different situation since we miss this prior knowledge and we are shall not be inclined to make unverifiable assumptions. Furthermore, we are not primarily interested in the most probable reconstructed image x. Instead, we want a set of possible boundaries given a blurred image and a small set of assumptions. Therefore, we introduce the fuzzy edge detection which is based on three simple assumptions
-
(a)
Each inclusion is a star-shaped, i.e. there exist a midpoint from which each boundary point is reached by a straight line.
-
(b)
The gradient of the blurred image yields a connected domain and the boundary of the inclusion lies in this connected domain.
-
(c)
Each boundary curve has a limited prescribed variation.
These three assumptions implicitly define a possible set of boundary curves provided a blurred image. Any of these boundary curves is possible but nothing can be said about the probability of each curve. Since we assume a star-shaped inclusion, a radial function \(r(\theta ):[0,2\pi )\mapsto \mathbb {R}_{+}\) is able to describe the boundary of this inclusion. The uncertainty of the boundary is then captured in a fuzzy function \(\tilde{r}\).
In order to construct a fuzzy function \(\tilde{r}\) from a blurred image with domain \(D=[-1,1]^2\), we use trigonometric interpolation on a vector of \(N\in \mathbb {N}\) fuzzy numbers. Each component of the fuzzy vector denoted by
is constructed from a radial cut (i.e. a line segment) that starts in the center and ends at the outer boundary for a prescribed set of angles \(\Theta = \left\{ \theta _i = \frac{2\pi i}{N+1}\mid i =1,\ldots ,N\right\} \). The continuous function \(\mathcal {I}_i: [0,R_i^\textrm{max}] \rightarrow \mathbb {R}_+\) with \(R_i^\textrm{max}= 1/\max \{\vert \cos \theta _i\vert ,\vert \sin \theta _i\vert \}\) determines the intensity of the gradient along this radial cut, see Fig. 4 for an illustration. The information encoded in \(\mathcal {I}_i\) is used to construct a trapezoidal fuzzy interval. For this, consider lower and upper percentage threshold values \(0<\underline{\delta }< \overline{\delta } <1\) and determine \(M :=\max _{r\in [0,R_i^\textrm{max}] }\mathcal {I}_i(r)\). If the intensity in one point is larger than \(\overline{\delta } M\), it is assigned a possibility of one. If it is smaller than \( \underline{\delta } M \), it gets assigned a possibility of zero. For simplicity, in between these thresholds we assign a possibility by linear interpolation. This algorithm yields a membership function \(\mu _{{\tilde{r}}_{\theta _i}}\) of trapezoidal form, which is illustrated in Fig. 4.
Gathering these fuzzy numbers into a vector yields the non-interactive fuzzy vector \(\tilde{R}_N\), encoded in the membership function
where non-interactivity means that the value of one sample does not influence the membership function of another value. Let \(R=(r_1\ \cdots \ r_N ) \in C_{0}[\tilde{R}_N]\) and denote the trigonometric interpolation by
It represents the mapping from interpolation points onto the trigonometric polynomials. This mapping is bijective if the degree of freedom coincides with the number of interpolation points. The coefficients are efficiently computed via a discrete Fourier transformation and the interpolation scheme yields convergence rates as follows.
Proposition 5
(Trigonometric interpolation) Let \(f:(0,2\pi )\rightarrow \mathbb {R}\) be a k-times differentiable periodic function that describes the boundary of a star shaped inclusion in \(\mathbb {R}^2\), then there exists \(c = c(k,f,f^{(k)})>0\) such that
Given all possible interpolation points \(R\in C_0[ \tilde{R}_N]\), we define the set of radial boundary functions describing the interface as
Note that \(\mathcal {T}(\cdot ;\tilde{R}_N)\) defines a fuzzy function [29]. The number of cuts and thus the number of interpolation points N is setted according to the accuracy required by the considered problem. With sufficiently many interpolation points, the set \(B_N\) can represent highly oscillatory boundaries. For the sake of efficiency, the interpolation should be carried out with as few points as possible, which demands an adequate knowledge about the properties (in particular smoothness/roughness) of the physical system at hand. We choose the total variation of the radial function as a measure for roughness, which is defined for \(R\in C_0[\tilde{R}_N]\) as
Given some bound \(0\le b \le \infty \), the TV restricted set of radial boundary function based on N grid points is defined by
With this construction, we may define the interactive fuzzy set
where \(\textbf{1}_X\) denotes the characteristic function on a set X. In Sect. 3.1 we make clear that the fuzzy set in (21) in fact defines a fuzzy vector. This in turn motivates the \(\alpha \)-cut propagation of fuzzy uncertainty, introduced in Sect. 2.1. In particular, consider a continuous real valued function \(\mathcal {Q}\) defined on \(C_0[\tilde{R}_N]\) which is the 0-cut of the non-interactive fuzzy vector \(\tilde{R}_N\) from (18). Note that \(C_0[\tilde{R}_N]\) is compact since the image domain \(D=[-1,1]^2\) is bounded. The propagated uncertainty of the interactive fuzzy set \(\tilde{R}_{N,b}\) through \(\mathcal {Q}\) is then captured on each \(\alpha \)-level
of the fuzzy set \(\tilde{\mathcal {Q}}_b := \mathcal {Q}(\tilde{R}_{N,b})\). We emphasise that \(C_0[\tilde{R}_N]\) defines a tensor domain on which \(\mathcal {Q}\) is well defined, i.e.
a fact that becomes useful when applying low-rank tensor formats in the upcoming Sect. 4.2 as surrogate models for \(\mathcal {Q}\). However, the computation of \(\tilde{Q}_b\) only requires evaluation of \(\mathcal {Q}\) on \(C_0[\tilde{R}_{N,b}]\subset C_0[\tilde{R}_N] \), which in general is not a tensor domain.
3.1 Properties of the TV bounded fuzzy set \(\tilde{R}_{N,b}\)
The total variation determines the shape of the inclusion. With a very small total variation, the shapes become more and more circular. Any circular shape, independent of the radius, has a total variation of zero. We want to point out, that we do not measure the total variation of the trajectory, where the total variation of a circle would be larger than zero. Instead we measure the total variation of the radial function.
In Fig. 5 shapes with different total variations are depicted. The blue shape is generated by taking alternating radii, the white shape is the result of optimization. The maximal total variation is 5.2 for \(N=12\), whereas a randomly generated shape has a total variation of around 2. With increasing N the maximal total variation would also increase.
Note that the resulting shapes may violate the boundaries in between the sampling points. Especially, the maximum total variation solution. This can be resolved with more sampling points, a sufficiently strict TV bound or by replacing the trigonometrical interpolation with a corresponding spline interpolation.
Bounding the total variation leads to interaction of the fuzzy set \(\tilde{R}_{N,b}\). Figure 6 illustrates the interaction for the case \(N=3\). It shows that fixing one point constraints the remaining points to the light gray area. Hence the possibility outside this area is zero. In a non-interactive setting the possibility could be strictly larger than zero. Consequently, the total variation bound restricts the set of possible curves. The Figure indicates that the set of valid radial points is convex, which is shown in the following proposition.
Proposition 6
(Convexity of TV constrained domain) Let \(N\in \mathbb {N}\). Then for \(b\ge 0\) the 0-cut \(C_0[\tilde{R}_{N,b}]\) defines a convex set.
Proof
Let \(R_1, R_2 \in {\text {supp}}\tilde{R}_N\), \(t\in [0,1]\) and define \(R_t = tR_1 + (1-t)R_2\). The radial function to describe the boundary then takes the form
with imaginary number i and coefficients \(a_i[R_t]\). Since the Fourier interpolation defines a linear operator, it holds \(a_i[R_t] = ta_i[R_1] + (1-t)a_i[R_2] \). Thus, by triangle inequality it follows that
\(\square \)
Consequently the \(\alpha \)-cuts of \(\tilde{R}_{N,b}\) are nested and convex.
Proposition 7
(Characterisation of \(\tilde{R}_{N,b}\)) Let \(b\ge 0\) and \(\tilde{R}_N\) be a fuzzy (domain) vector with \(C_1[\tilde{R}_N] \subset C_0[\tilde{R}_{N,b}]\). Then \(\tilde{R}_{N,b}\) defines a fuzzy (domain) vector.
Note that for given N, there always exists \(b=b(N)\) such that the conditions of Proposition 7 hold true. In particular, for b large enough it holds \(\tilde{R}_N = \tilde{R}_{N,b}\). Consequently, the \(\alpha \)-cut propagation (23) can be applied.
Remark 1
The convexity property from Proposition 6 is essential to apply Theorem 2. It is a consequence of the linear interpolation and the triangle inequality property of the TV based restriction. In the case, one of the latter properties does not hold, the fuzzy set may not define a fuzzy domain. Then, the propagation of the fuzzy set still can be realised by sampling approaches as discussed in the end of Sect. 2.1.
4 Accelerated emulation of fuzzy effective material
The fuzzy edge detection described above yields a fuzzy vector \(\tilde{R}_N\), where N is the number of radial cuts. Trigonometric interpolation for each realization \(R\in \tilde{R}_N\) then results in a boundary representation of the inclusion. In the following, we define a composite material based on this representation.
Consider the square \([-1,1]^2\) on which we define a two phase composite material \(\varvec{C}(x)= \varvec{C}(x,R)\) with \(\varvec{C}_\text {incl},\varvec{C}_\text {matrix} \in \mathbb {E}\textrm{la}(d,\text {iso})\). This description represents a piecewise isotropic material with star-shaped inclusion defined as
such that \(0< R_{\text {min}}< \mathcal {T}(\theta , R)< R_\text {max} < 1\) uniform in \(\theta \) and R.
Given a fixed R, the homogenisation from Sect. 2.2 yields the constitutive tensor \(\varvec{C}(\cdot ,R) \in \mathbb {E}\textrm{la}(d)\) for the upscaled macroscopic material. In Voigt notation, this mapping is denoted as
with symmetric positive definite matrix \(\mathcal {H}(R)\in \mathbb {R}^{3,3}\). It in general describes an anisotropic material since the involved geometry of the inclusion may lack any type of symmetry.
We would like to point out that the parametric dependency \(R\rightarrow \varvec{C}(\cdot , R)\) does not define a continuous function with images in \(L^\infty (D)^{d,d,d,d}\) since marginal changes of the shape of the inclusion immediately yield an \(L^\infty \) error equal to the contrast \(\Vert \varvec{C}_\text {incl}-\varvec{C}_\text {matrix}\Vert _F\) with Frobenius norm \(\Vert \cdot \Vert _F\). Despite this discontinuity and its effect on the regularity of the parametric solution \(R\rightarrow \varvec{u}(R)\), the parametric homogenised tensor (26) defines a continuous map.
Recall that the evaluation of \(\mathcal {H}\) involves multiple simulations of periodic linear elastic boundary value problems of the form (11). To accelerate the upscaling process, in the following we replace the simulator with an emulator. The emulator relies on a mesh discretisation of the domain D and a group of tensor train surrogates to approximate \(\mathcal {H}\), which is discussed in Sect. 4.2.
If D is discretised by an automatic mesh generator under the constrained of equal amount of vertices on the inclusion’s boundary and on the boundary of D for any R, the resulting mesh map \(R\mapsto \mathcal {M}(R)\) in general is discontinuous. As a consequence, a mesh dependent finite element computation may inherit the lack of continuity even through the whole geometry has smooth dependence on R, see Fig. 7 for an example. This in turn aggravates optimization based on gradient information. We solve this issue in Sect. 4.1 by constructing a family of transformed meshes with smooth dependence on R.
4.1 Constructive smooth transformation of meshes
For \(n\in \mathbb {N}\) interpolation points, consider the composite interface discretisation
We now construct a smooth diffeomorphism that creates meshes based on a single reference mesh \(\hat{\mathcal {M}}\) denoted as \(\Psi :\mathbb {R}^d\times \mathbb {R}^N\rightarrow \mathbb {R}^d\) such that \(\Psi (\hat{\mathcal {M}}, R)=\mathcal {M}(R)\). Let \(R_\text {ref} = [0.5,\ldots ,0.5]\in \mathbb {R}^N\) and \(\mathcal {T}(\cdot ,R_\text {ref})\) define a circular inclusion with radius 0.5. Consider a fixed reference mesh discretisation \(\mathcal {M}_\text {ref}\) of \([-1,1]^2\) such that
-
there is a fixed number of nodes on the boundary \(\partial [-1,1]^2\)
-
any node in \(\mathcal {F}_n(R_\text {ref})\) is part of the mesh.
Compute the sets \(\mathcal {F}_{2n}(R_\text {ref})\) and \(\mathcal {F}_{2n}(R)\) corresponding to higher discrete resolution of the reference or target interface. The transformation \(\Psi \) then consists of two parts as follows. First, let \(\{\phi _i\}\) form a smooth partition of unity of \([0,2\pi ]\) with \(\phi _i(i\pi / n) = 1\) and \({\text {supp}}\phi _i = [(i-1)\pi /n, (i+1)\pi /n]\). This function set deals with the angular part of the transformation. Second, let \(\chi _i\in \mathcal {C}^k[0,\sqrt{2}]\) be a set of splines with \(k\ge 2\) for the radial part of the transformation with the following properties:
-
1.
The reference radius is mapped onto the transformed radius, i.e.
$$\begin{aligned} \chi _i(\mathcal {T}(\theta _i, R_{\text {ref}})) = \mathcal {T}(\theta _i,R). \end{aligned}$$ -
2.
The spline is strong monotonically increasing on \((R_{\text {min}}, R_{\text {max}})\) and otherwise equals the identity map.
Altogether, the transformation reads
Note that the number of radial cuts N determines the range of shapes, whereas the number of interpolation points n has to be sufficient large to trace the shape and to account for the resolution of the finite element mesh. We refer to Fig. 8 for an illustration of the capacity of the constructed map \(\Psi \), which is characterised next.
Proposition 8
(\(\mathcal {C}^k\) diffeomorphism) The map \(\Psi : [-1,1]^2 \rightarrow [-1,1]^2\) is bijective and k-times continuously differentiable.
Proof
The bijectivity follows from strong monotonicity in radial direction and the smooth partition of unity in angular direction. The smoothness of \(\phi _i\) and of the polar coordinate mapping away from zero is given since \(\chi _i\) is k-times continuous differentiable. For \(i=1,\ldots ,2n\) the desired property follows immediately. \(\square \)
4.2 Tensor trains based emulation
Assume some function \(f:{C_0[{\tilde{R}}]}\subset \mathbb {R}^N\rightarrow \mathbb {R}\) with . For \(f = H_{ij}\), \(1\le i \le j\le 3\) we consider surrogates of the form
based on a polynomial feature class \(\Pi :=\{\Pi _{\beta }, {\beta }\in \mathbb {N}^N\}\) where each polynomial \(P_{\beta } \in \Pi \) is of tensorised form \(\Pi _{\beta } = \bigotimes _{i=1}^N q_{{\beta }_i}^i\) with one dimensional polynomials \(\{q_{{\beta }_i}^i:C_0[\tilde{R}_i] \rightarrow \mathbb {R}, {\beta }_i \in \Lambda _i\}\) for \(i=1,\ldots , N\) and unknown coefficient tensor \(\varvec{U}:\Lambda \rightarrow \mathbb {R}\). This model suffers from the curse of dimensionality since the cardinality \(\vert \Lambda \vert \) grows exponential with respect to the dimension N. A possibility to circumvent this challenge lies in a compressed representation of the coefficient tensor, here based on the tensor train format (TT format) [48]. Let \(\rho =(\rho _1,\ldots , \rho _{N-1})\in \mathbb {N}^{N-1}\) be the tensor train rank and let \(\rho _0=\rho _N=1\). We then choose \(\varvec{U}\) given in tensor train decomposition by
with component tensors of order 3, such that \(U_i[:]\in \mathbb {R}^{\rho _{i-1}, \vert \Lambda _i\vert , \rho _i}\) for \(i=1,\ldots , N\). The number of degrees of freedom for this design is given as \(\sum _{i=1}^N \rho _{i-1} \vert \Lambda _i\vert \rho _i\), which does not grow exponentially but only linearly in dimension N. Given data \((R^{(k)}, y^{(k)})_{k=1}^K\), \(K\in \mathbb {N}\) with \(y^{(k)}=f(R^{(k)})\), we obtain a surrogate as in (28) by carrying out a regularised empirical regression, namely by solving the optimisation problem
with regularisation parameter \(\lambda >0\) and the Frobenius norm \(\Vert \cdot \Vert _{{\textrm{Fro}}}\). The optimisation problem (30) can be solved by regression with an alternating linear scheme (ALS) [49]. As a modification of the basic ALS, we introduce a scheme for rank adaptivity. This concept offers several advantages to obtain an adjusted tensor train rank which is in general unknown a priori. From a practical point of view, it reduces the computational cost during optimisation while obtaining a prescribed accuracy in the approximation class. Furthermore, starting with a small rank, the iterative process empirically enables the ALS to converge to a solution based on successive computations of initial guesses for models with higher ranks based on the given restricted number of samples. The proposed approach is summarised in Algorithm 1 and the rank adaptivity is presented in Algorithm 2.
A principle of the rank adaptivity is to keep one (control) singular value to monitor the importance of the related rank coupling value during the optimisation process. This additional singular value per rank coupling remains until the end of the regression scheme to prevent oscillation between rank growth and reduction. It ensures an upper bound of the related rank value throughout the entire process. After successful termination of Algorithm 2, the existing (possibly small) control singular value is removed by a final rounding [48] of the resulting tensor train. We refer to [49] for more technical details on the basic ALS, e.g. with regard to setting and moving the non-orthogonal component (the core) via left and right matricisation and orthogonalisation. Eventually, the overall emulator is obtained by evaluation of the 6 scalar valued tensor train surrogate maps, replacing \(R\mapsto [\mathcal {H}(R)]_{ij}\) and \(1\le i\le j\le 3\) corresponding to the upper triangular part of the symmetric matrix \(\mathcal {H}(R)\).
Note that non-zero values of \(H_{13}\) and \(H_{23}\) solely appear when the effective material behaviour is not isotropic or rotational-free orthotropic, i.e., this is described by (13) with \(\theta =0\).
5 Numerical experiments
This section is concerned with the assessment of the numerical performance of the approach presented above, for which two experiments are examined. In the first experiment, we measure the distance of the upscaled material to the isotropy and orthotropy class. Additionally, we identify configurations that maximise and minimise the respective distances. In the second experiment, we test if the upscaled material is suitable for a worst case analysis, i.e., we test if it is possible to find bounds that envelop the most extreme behaviour of the quantity of interest under consideration. For this, a material with \(8\times 8\) arbitrary inclusions placed on a checkerboard is compared to an upscaled material with the same geometrical dimensions. Both experiments demonstrate the efficacy of the fuzzy approach to model uncertainty of the inclusion to identify extremal behaviour and its source.
Three types of computational tasks are performed in the experiments. The foundation is laid by Finite Element (FE) simulations to generate the realisations of constitutive tensors by solving (11). Furthermore, an alternating linear scheme is used to train the surrogate model and for optimization computations to carry out the uncertainty propagation based on \(\alpha \)-cuts.
All FE computations are done with the ‘FEniCS‘ package [50]. The mesh generation is realised using ‘Gmsh‘ [51] and the ‘Bubbles‘ package [52]. Moreover, the python package ‘TensorTrain‘ [53] is utilised for the rank adaptive tensor train regression and ‘ALEA‘ [54] for the underlying polynomial features. The optimization tasks are performed with restarted trust region optimization implemented in ‘Scipy‘ [55].
The computations are performed with \(N=6\) and \(N=12\) radial sampling points. Here, the \(N=6\) case imposes significantly less computational burden on the surrogate and the optimization than the \(N=12\) case. In particular, our optimization scheme took up to \(10^{6}\) evaluations of the \(\mathcal {H}(R)\) for \(N=12\) and up to \(1\times 10^{5}\) evaluations for \(N=6\) per propagated quantity of interest. We underline that all optimization tasks involved are non-convex with non-linear cost functions and constraints. Consequently, we take arbitrary points \(R \in [0.3, 0.7]^6\) for \(N=6\) and \(R \in [0.4, 0.6]^{12}\)—a smaller domain—for \(N=12\). This enables sufficiently non-trivial shapes without rendering the surrogate model and the optimization infeasible. Each realisation R determines a boundary stellar inclusion in terms of a trigonometric interpolation. Inside and outside of the inclusion we assume isotropic material behaviour. Concretely, we let the Young’s modulus \(E_\text {incl} = 3230\) and Poisson ratio of \(\nu _\text {incl}=0.3\) in the inside. Outside of the inclusion we choose \(E_\text {matrix}= E_\text {incl}/4\) and \(\nu _\text {matrix}=0.2\). These values are transformed into Lamé constants by
By adaptation to the plane, adapted Lamé constants are obtained, namely
For a given \(R\in C_0[\tilde{R}_N]=[0.3,0.7]^N\) the homogenisation problem (11) is solved with FE of uniform polynomial degree \(p=2\). The reference mesh is based on 80 nodes on the related reference composite interface \(\mathcal {T}(\cdot ;R_\textrm{ref})\) is shown in Fig. 8. The computational mesh with composite interface \(\mathcal {T}(\cdot , R)\) is then obtained by applying the transformation \(\Psi (\cdot , R)\).
5.1 Surrogate validation
As a preparative step, we build an accelerated emulator in the tensor train format for \(N=6\) and \(N=12\) dimensional fuzzy input vectors \(\tilde{R}_N\), yielding a compressed surrogate of the map \(R\rightarrow \mathcal {H}(R)\).
Given \(K = 15635\) \((N=6)\) and \(K = 12000 + 4096\) \((N=12)\) samples that are normalised with respect to the sample mean and variance, we iteratively apply Algorithm 1 and Algorithm 2 with tensorised Chebyshev polynomials up to degree 5 in each coordinate. More precisely, we train the initial tensor train surrogate with a degree of two. Next, a new tensor train for polynomial degree \(\textrm{deg}+1\) up to \(\textrm{deg}=4\) is set to the last tensor approximation as initial guess instead of starting with a random rank-1 tensor train. All tensor train component entries associated to the higher polynomial degree are initially set to zero. This training approach turn out to have two advantages. First, in the iterative procedure the training and validation sets are split randomly, resulting in a pseudo stochastic solver. A strategy that is successful in the context of machine learning. Second, there is a huge speed-up in the training procedure for finding a local minimum. In fact, with a naive tensor train training for \(N=12\) with initial tensor polynomial degree 5, we rarely observed convergence to meaningful surrogates given random initial values.
We use \(\textrm{tol}_{\textrm{MSE}}=1\times 10^{-8}\), \(\textrm{iter}_{\textrm{max}}=200\), \(L_{\textrm{HIST}}=10\), \(\textrm{tol}_{\textrm{DECAY}}=1\times 10^{-5}\), \(r_\textrm{max}\le 5,7,9,20\) for \(\textrm{deg}=2,\ldots ,5\) , \(r_\textrm{add}\in \{1,2\}\), and \(\delta = 1\times 10^{-8}\). The results of the surrogate learning approach are depicted in Tables 1 and 2, which complement Figs. 10 and 11, showing mean square error or pointwise relative and absolute errors. The highlighted entries in both tables show a pointwise lack of accuracy of the tensor train surrogate for the \(H_{13}\) and \(H_{23}\) components, which represent the pure rotational or anisotropic contribution in the overall upscaled tensor \(\mathcal {H}\) on the used test sets. A further inspection of the high relative errors showed that these strictly belong to pointwise approximations of values close to zero. Nevertheless, strikingly small mean squared errors can be observed as a result of the proposed optimization algorithm. Figures 10 and 11 display the value ranges of the subcomponents of \(\mathcal {H}\) and the overall absolute and relative errors of the subcomponents and the full tensor \(\mathcal {H}\). We observe that the reduced (pointwise) accuracy of the surrogates for \(H_{13}\) and \(H_{23}\) does not influence the overall error for the full parametric tensor \(\mathcal {H}\). The final obtained ranks obtained with the rank adaptive algorithm are depicted in Fig. 9 for \(N=6\) and \(N=12\).
5.2 Distances to the isotropy and orthotropy material class
In this section we apply the tensor train emulator to obtain an approximation of the parametric constitutive tensor \(\mathcal {H}(R)\) of the effective material defined in (26), given an parametrisation of the inclusion through the vector \(R\in C_0[\tilde{R}_{N,b}]\). For each constitutive tensor we compute the distance to the isotropy class, denoted by \(d_{\textrm{iso}}\), and orthotropy class, denoted by \(d_{\textrm{ortho}}\), according to (14). The fuzzy uncertainty of the boundary is propagated in terms of \(\alpha \)-cuts. On each \(\alpha \)-level, two optimization problems with constraints defined in (23) have to be solved, where \(\mathcal {Q}\) is either \(d_{\textrm{iso}}\) or \(d_{\textrm{ortho}}\). As optimiser we use a restarted trust region scheme to minimise the non-convex and non-linear target function with non-linear constraints.
Figure 12 shows the experimental results for \(\mathcal {Q}= d_{\textrm{iso}}\) with \(N=6\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.3,0.7]^N\) and the total variation bounds \(b=0.5, 1, \infty \). We point out that since \(D=[-1,1]^2\) is not of special hexagonal shape but of square shape, a circular inclusion does not yield isotropic effective material properties. Nevertheless, the circular shape—while exhibiting vanishing total variation—provides effective properties closest to the isotropy class. Since in our experiments we choose a stiffer material as inclusion, the circular inclusion minimises its size proportional to the possible maximal radius encoded in the \(\alpha \)-level. In this sense, the size of a circle is a measure of perturbation of the isotropic matrix material. Moreover, if the total variation and radii bounds allow it, the inclusion converges to a peanut shape. The green peanut shaped inclusion in Fig. 12 marks the inclusion with maximum distance to the isotropic material class for the given parameterisation family \(B_{N,\infty }\) from (20). Note that by rotational invariance also the peanut rotation of \(\pi /2\) yields the same maximising shape.
Figure 14 pictures the result for \(\mathcal {Q}= d_{\textrm{iso}}\) with \(N=12\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.4,0.6]^N\) and total variation bounds \(b=0.5, 1, \infty \). The shrinked domain has an immediate impact on the maximal possible total variation bound of all interface realisations. Consequently, maximum distances are caused by curves with similar shape up to rotation. Already for a total variation of 0.5, a symmetric shape emerges that becomes more prominent with increasing total variation.
Figure 13 depicts the experimental results for \(\mathcal {Q}= d_{\textrm{ortho}}\) with \(N=6\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.3,0.7]^N\) and total variation bounds \(b=0.5, 1, \infty \). A minimum of 0 is attained on each \(\alpha \) cut of the corresponding membership function, since the orthotropy class \(\mathbb {E}\textrm{la}(2,\textrm{ortho})\) contains the square symmetric class. Many boundaries in \(B_{N,b}\) lead to a shape that yields a square symmetric effective material.
As typical maximiser of the distance to the orthotropy class, a bean shape is found. We observe that the maximal possible total variation value of \(B_{N,b}\) is not exhausted when attaining the maximal distance. Note that due to rotational invariance, the bean shape can be rotated by multiples of \(\pi /2\) while still being a (in fact the same) maximiser.
Figure 15 shows the experimental results for \(\mathcal {Q}= d_{\textrm{ortho}}\) with \(N=12\), zero \(\alpha \)-cut \(C_0[\tilde{R}_N] = [0.4,0.6]^N\) and total variation bounds \(b=0.5, 1, \infty \). Again the shrinked domain has an immediate impact on the maximal possible total variation bound of all interface realisations. Consequently, maximum distances are attained by curves with similar shapes up to rotation. We observe that symmetric dented shape maximises the distance to the orthotropy class for the \(\alpha =1\) cut or a total variation smaller or equal to 0.5. However, as the total variation bound gets larger and we allow for larger fluctuations due to a wider parameter range the inclusion converges to a non-axis aligned structure with non-symmetric dented shape.
5.3 Best/worst case estimates for non-homogenised matrix composites
Let the domain \(D=[-N_s,N_s]\times [-N_s,N_s]\) with \(N_s=8\) and consider a checkerboard partitioning {\( D_{ij},i,j=1,\ldots ,N_s\}\) of D with square subdomains \(D_{ij}\) defined by
On each subdomain \(D_{ij}\) we consider a single two phase composite as defined by (25) using a local polar coordinate system w.r.t. the midpoint of \(D_{ij}\), encoded in a material tensor \(\varvec{C}_\text {inhomo}\) with piecewise values \(\varvec{C}_\text {incl}\) and \(\varvec{C}_\text {matrix}\) specified in Sect. 5.1.
For this experiment we choose the non-interactive fuzzy set \(\tilde{R}_{N,\infty }\) with \(N=6\) trapezoidal fuzzy components characterised by
In between these prescribed points, the \(\alpha \)-levels are obtained by linear interpolation. The TV bound is ignored in this experiment.
Furthermore, we consider a homogenised material tensor \(\varvec{C}_\text {homo} = \mathcal {H}(R)\) for \(R\in C_0[\tilde{R}_{N,b}]\) obtained from the homogenisation process of Sects. 2.2 and 5.1. In each domain, we model individual star shaped material inclusions as in (25) with boundaries of uniformly bounded total variation smaller or equal to b.
Given a material law \(\varvec{C}\in \{\varvec{C}_\text {inhomo},\varvec{C}_\text {homo}\}\), we then solve the linear elasticity problem
representing a linear elastic tensile test. Here, \(\Gamma _{\varvec{0}}\) = \(\{-N_s\}\times [-N_s,N_s]\) and
We are interested in the fuzzy propagation of the homogenised material through the average displacement \(\mathcal {Q}=\overline{u}\) defined by
with the solution \(\varvec{u}\) of (31). Figure 16 shows the resulting membership function based on the fuzzy homogenised material law. The black dots mark resulting average displacements obtained for various non-periodic inclusions. Moreover, some configurations of possible realisations of the \(8\times 8\) inclusions are displayed. The corresponding average displacement is marked with colours, accordingly. Given any non-periodic configuration of prescribed matrix composites, we observe that the membership function at each \(\alpha \)-level bounds the various average displacements. In particular, for \(\alpha = 0., 0.5, 1\) we illustrated a corresponding single composite shape that (assumed in a periodic structure) results in the minimum and maximum average displacements. As the interior composite is slightly stiffer than the surrounding matrix material, it is to be expected that the shapes of the minimiser and maximiser attempt to exhaust or avoid the maximum capacity of the parameter range encoded in \(C_\alpha [\tilde{R}_{N,\infty }]\). It is noteworthy that the homogenised material is able to serve as worst/best case estimator for this particular quantity of interest (i.e. the average displacement). As long as the repetition of the same cell leads to extremal values of the quantity of interest, the fuzzy homogenised material functions as a worst/best case estimator.
6 Conclusion
We considered the possibilistic shape uncertainty of one inclusion induced by a blurred image. In this setting, computational stochastic homogenisation was carried out to propagate the uncertainty through the linear elasticity model, resulting in a fuzzy effective material. Eventually, the effective material was examined with respect to its distance to the orthotropic/isotropic material class. Moreover, it was used as a worst/best case estimator with respect to the global average displacement for a non-homogenised material.
To achieve this result, two major obstacles had to be overcome. The first was caused by the discontinuous mapping from boundary to effective material, which was a consequence of the automatic re-meshing for each new boundary instance. This was successfully resolved with an arbitrarily smooth transformation of a single reference mesh onto the respective mesh for a domain with a star-shaped inclusion. The second challenge was given by the computational effort needed to perform the (possibilistic) uncertainty propagation. Depending on the optimiser and the quantity of interest at hand, each \(\alpha \)-cut required up to \(5\times 10^4\) computational homogenisation simulations. This problem was successfully resolved with a highly accurate tensor train emulator.
We provide an (numerical) analysis for a problem where only little knowledge is given and only simple assumptions can be made. This may be distinguished from other works, where more prior knowledge is given and more elaborated assumptions are suitable. In the latter scenario, one usually can utilise the Bayesian framework to model the uncertainty, see [56, 57], where in our scenario stochastic modelling may lead to a false confidence [23]. While usually the uncertainty of material properties is modelled with precise probabilities, leading to statements about expectation values and probabilities of failure, cf. [58, 59], we model geometrical uncertainties with imprecise probability, leading to statements about possible intervals and worst/best case configurations. Even if an imprecise probability approach for geometrical uncertainty is chosen, the geometrical shape is often restricted to simple parametrisations like a circle of varying diameter [60]. This in fact is a very useful simplification, for instance if the observed quantities only depend on the volume ratio of the two materials. Nonetheless, our experiments demonstrate that the shape determines if the effective material is isotropic, orthotropic or anisotropic. A circle always results in an orthotropic effective material in the considered case with unit square cells. Furthermore, our tensor train surrogate shows a remarkable accuracy of a relative error of order \(\mathcal {O}(10^{-4})\). This is partly the result of a suitable choice of features, the novel rank-adaptive training strategy and the continuity induced by using a mesh transformation instead of automatic re-meshing. Compared to other works such as [61, 62], where standard generic methods like artificial neural networks and polynomial chaos are used in a similar context, we gain several orders of accuracy.
To focus on the influence of the shape uncertainty, other parts in our setting and analysis were kept simple. These parts are ideal starting points for future research. The homogenisation, for example, was performed with two phases of linear elastic materials. This can easily be substituted with more sophisticated materials—one may think of anisotropy, damage and higher contrast—or different homogenisation settings. The assumption of a star-shaped inclusion can be discarded with an adjusted edge detection and a different parametrisation of the boundary. Furthermore, the roughness measurement via total variation can be exchanged or extended with other restrictions of the boundary curves. In addition, it is possible to model the material constants appropriately with precise probabilities. By doing so, we add a stochastic dimension which together with the possibilistic uncertainty results in a fuzzy-stochastic model.
Although we only investigated 2D materials, our approach can be extended to 3D. The numerical upscaling and the surrogate model are applicable in higher dimensions. For the fuzzy edge detection the radial cuts should be replaced by cuts through the sphere. Under the assumption of star-shaped inclusions in 3D, the inclusion surface can be described in spherical coordinates with suitable interpolation for the radial component, which in turn can be constrained by a corresponding 2D TV measure. Moreover, in the 3D case there are more material classes. Depending on the symmetry class, the associated distances are—to our knowledge—not yet analytically given and thus have to be computed numerically [39, 63]. Thus, the theoretical adaption is feasible, but we expect the computational challenges of the 3D case to be significantly larger.
Another interesting research direction for our framework are more involved material models like hyperelastic materials or materials with damage modeling. In this case the numerical upscaling leads to an non-linear effective material law. In order to represent this non-linearity in our framework the surrogate model design needs to be adapted, e.g. mapping to interpolants of effective potentials based on [64].
Finally, we hope that our presented contributions, namely the treatment of discontinuity from re-meshing, the fuzzy edge detection, and the highly accurate tensor train surrogate, can be used and extended beneficially in other research to simulate composite materials in the presence of shape uncertainty.
References
Jones RM (2018) Mechanics of composite materials, 2nd edn. CRC Press, Boca Raton
Casati R, Vedani M (2014) Metal matrix composites reinforced by nano-particles—a review. Metals 4(1):65–83
Seydibeyoglu MO, Mohanty AK, Misra M (2017) Fiber technology for fiber-reinforced composites. Woodhead Publishing, Swaston, Cambridge
Oksman Niska K, Sain M (2008) Wood-polymer composites. Woodhead Publishing Materials, Swaston, Cambridge
Vasiliev VV, Morozov EV (2018) Advanced mechanics of composite materials and structures. Elsevier, Netherlands
Babuška I (1976) Homogenization approach in engineering. In: Computing methods in applied sciences and engineering. Springer, Germany, pp 137–153
Braides A et al (2002) Gamma-convergence for beginners, vol 22. Clarendon Press, Oxford University Press
Engquist B, Souganidis PE (2008) Asymptotic and numerical homogenization. Acta Numer 17:147–190
Anantharaman A, Le Bris C (2012) Elements of mathematical foundations for numerical approaches for weakly random homogenization problems. Commun Comput Phys 11(4):1103–1143
Blanc X, Le Bris C, Lions P-L (2007) Stochastic homogenization and random lattices. J Math Pures Appl 88(1):34–63
Bourgeat A, Piatnitski A (2004) Approximations of effective coefficients in stochastic homogenization. In: Annales de l’IHP Probabilités et Statistiques, vol 40, pp 153–165
Gloria A, Neukamm S, Otto F (2014) An optimal quantitative two-scale expansion in stochastic homogenization of discrete elliptic equations. ESAIM Math Model Numer Anal 48(2):325–346
Engquist B, Runborg O (2002) Wavelet-based numerical homogenization with applications. In: Multiscale and multiresolution methods. Springer, Germany, pp 97–148
Ostoja-Starzewski M (2006) Material spatial randomness: from statistical to representative volume element. Probab Eng Mech 21(2):112–132
Ladeveze P, Nouy A (2003) On a multiscale computational strategy with time and space homogenization for structural mechanics. Comput Methods Appl Mech Eng 192(28–30):3061–3087
Tranquart B, Ladevèze P, Baranger E, Mouret A (2012) A computational approach for handling complex composite microstructures. Compos Struct 94(6):2097–2109
Babuška I, Andersson B, Smith PJ, Levin K (1999) Damage analysis of fiber composites part i: statistical analysis on fiber scale. Comput Methods Appl Mech Eng 172(1–4):27–77
Babuška I, Motamed M (2016) A fuzzy-stochastic multiscale model for fiber composites: a one-dimensional study. Comput Methods Appl Mech Eng 302:109–130
Shafer G (1976) A mathematical theory of evidence, vol 42. Princeton University Press, Princeton
Matheron G (1974) Random sets and integral geometry. Wiley, New York, p 261
Zadeh LA (1978) Fuzzy sets as a basis for a theory of possibility. Fuzzy Sets Syst 1(1):3–28
Owhadi H, Scovel C, Sullivan TJ, McKerns M, Ortiz M (2013) Optimal uncertainty quantification. SIAM Rev 55(2):271–345
Balch MS, Martin R, Ferson S (2019) Satellite conjunction analysis and the false confidence theorem. Proc R Soc A 475(2227):20180565
Drieschner M, Matthies HG, Hoang T-V, Rosić BV, Ricken T, Henning C, Ostermeyer G-P, Müller M, Brumme S, Srisupattarawanit T et al (2019) Analysis of polymorphic data uncertainties in engineering applications. GAMM-Mitteilungen 42(2):201900010
Drieschner M, Petryna Y (2019) Acquisition of polymorphic uncertain data based on computer tomographic scans and integration in numerical models of adhesive bonds. Preprint-Reihe des Fachgebiets Statik und Dynamik, Technische Universität Berlin
Zadeh LA (1975) The concept of a linguistic variable and its application to approximate reasoning I. Inf Sci 8(3):199–249
Moens D, Hanss M (2011) Non-probabilistic finite element analysis for parametric uncertainty treatment in applied mechanics: recent advances. Finite Elem Anal Des 47(1):4–16
Möller B, Beer M (2008) Engineering computation under uncertainty-capabilities of non-traditional models. Comput Struct 86(10):1024–1041
Möller B, Graf W, Beer M, Sickert J-U (2002) Fuzzy randomness-towards a new modeling of uncertainty. In: Fifth world congress on computational mechanics, Vienna, Austria
Nguyen HT (1978) A note on the extension principle for fuzzy sets. J Math Anal Appl 64(2):369–380
Boutin C (2019) Homogenization methods and generalized continua in linear elasticity. In: Altenbach H, Oechsner A (eds) Encyclopedia of continuum mechanics. Springer, Berlin
Sánchez-Palencia E (1980) Non-homogeneous media and vibration theory, vol 127. Lecture notes in physics. Springer, Berlin
Moravec F, Roman S (2009) Numerical computing of elastic homogenized coefficients for periodic fibrous tissue. Appl Comput Mech 3:141–152
Ostoja-Starzewski M, Schulte J (1996) Bounding of effective thermal conductivities of multiscale materials by essential and natural boundary conditions. Phys Rev B 54(1):278
Le Bris C, Legoll F, Minvielle W (2016) Special quasirandom structures: a selection approach for stochastic homogenization. Monte Carlo Methods Appl 22(1):25–54
Christensen RM (2012) Mechanics of composite materials. Courier Corporation, North Chelmsford
Pecullan S, Gibiansky L, Torquato S (1999) Scale effects on the elastic behavior of periodic and hierarchical two-dimensional composites. J Mech Phys Solids 47(7):1509–1542
Mandel J (1965) Généralisation de la théorie de plasticité de wt koiter. Int J Solids Struct 1(3):273–295
Antonelli A, Desmorat B, Kolev B, Desmorat R (2022) Distance to plane elasticity orthotropy by Euler–Lagrange method. C R Mécanique 350(G2):413–430
Kundur D, Hatzinakos D (1996) Blind image deconvolution. IEEE Signal Process Mag 13(3):43–64
Sondhi MM (1972) Image restoration: the removal of spatially invariant degradations. Proc IEEE 60(7):842–853
Cannon M (1976) Blind deconvolution of spatially invariant image blurs with phase. IEEE Trans Acoust Speech Signal Process 24(1):58–63
Hunt BR (1973) The application of constrained least squares estimation to image restoration by digital computer. IEEE Trans Comput 100(9):805–812
Miller K (1970) Least squares methods for ill-posed problems with a prescribed bound. SIAM J Math Anal 1(1):52–74
Lam EY, Goodman JW (2000) Iterative statistical approach to blind image deconvolution. JOSA A 17(7):1177–1184
Levin A, Weiss Y, Durand F, Freeman WT (2011) Understanding blind deconvolution algorithms. IEEE Trans Pattern Anal Mach Intell 33(12):2354–2367
Ruiz P, Zhou X, Mateos J, Molina R, Katsaggelos AK (2015) Variational bayesian blind image deconvolution: A review. Digit Signal Process 47:116–127
Oseledets IV (2011) Tensor-train decomposition. SIAM J Sci Comput 33(5):2295–2317
Holtz S, Rohwedder T, Schneider R (2012) The alternating linear scheme for tensor optimization in the tensor train format. SIAM J Sci Comput 34(2):683–713
FEniCS Project—Automated solution of differential equations by the finite element method. http://fenicsproject.org
Geuzaine C, Remacle J-F (2009) Gmsh: a 3-d finite element mesh generator with built-in pre-and post-processing facilities. Int J Numer Methods Eng 79(11):1309–1331
Gruhlke R, Schäfer T. Bubbles—a Python framework for composite modelling. https://github.com/Priusds/Bubbles
Gruhlke R, Sommer D. TensorTrain—a Python framework for tensor train approximations with PyTorch and NumPy backend. https://github.com/Drachiro/TensorTrain
Eigel M, Gruhlke R, Marschall M, Trunschke P, Zander E. ALEA–a python framework for spectral methods and low-rank approximations in uncertainty quantification. https://bitbucket.org/aleadev/alea
Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, Burovski E, Peterson P, Weckesser W, Bright J, van der Walt SJ, Brett M, Wilson J, Millman KJ, Mayorov N, Nelson ARJ, Jones E, Kern R, Larson E, Carey CJ, Polat İ, Feng Y, Moore EW, VanderPlas J, Laxalde D, Perktold J, Cimrman R, Henriksen I, Quintero EA, Harris CR, Archibald AM, Ribeiro AH, Pedregosa F, van Mulbregt P (2020) SciPy 1.0 contributors: SciPy 1.0: fundamental algorithms for scientific computing in python. Nat Methods 17:261–272. https://doi.org/10.1038/s41592-019-0686-2
Peng T, Saxena A, Goebel K, Xiang Y, Sankararaman S, Liu Y (2013) A novel Bayesian imaging method for probabilistic delamination detection of composite materials. Smart Mater Struct 22(12):125019
Yan G, Sun H, Büyüköztürk O (2017) Impact load identification for composite structures using Bayesian regularization and unscented Kalman filter. Struct Control Health Monit 24(5):1910
Dimitrov N, Friis-Hansen P, Berggreen C (2013) Reliability analysis of a composite wind turbine blade section using the model correction factor method: numerical study and validation. Appl Compos Mater 20(1):17–39
Zhou X, Gosling P, Ullah Z, Pearce C et al (2016) Exploiting the benefits of multi-scale analysis in reliability analysis for composite structures. Compos Struct 155:197–212
Pivovarov D, Steinmann P (2016) Modified SFEM for computational homogenization of heterogeneous materials with microstructural geometric uncertainties. Comput Mech 57(1):123–147
Pilania G, Wang C, Jiang X, Rajasekaran S, Ramprasad R (2013) Accelerating materials property predictions using machine learning. Sci Rep 3(1):1–6
Vasilyeva M, Tyrylgin A (2021) Machine learning for accelerating macroscopic parameters prediction for poroelasticity problem in stochastic media. Comput Math Appl 84:185–202
Kochetov M, Slawinski MA (2009) On obtaining effective orthotropic elasticity tensors. Q J Mech Appl Math 62(2):149–166
Yvonnet J, Monteiro E, He Q-C (2013) Computational homogenization method and reduced database model for hyperelastic heterogeneous structures. Int J Multiscale Comput Eng 11(3):201–225
Acknowledgements
The authors gratefully acknowledge the financial support of the German Research Foundation (DFG) within the Subproject 4 and 10 of the Priority Programme “Polymorphic uncertainty modelling for the numerical design of structures—SPP 1886”.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Eigel, M., Gruhlke, R., Moser, D. et al. Numerical upscaling of parametric microstructures in a possibilistic uncertainty framework with tensor trains. Comput Mech 71, 615–636 (2023). https://doi.org/10.1007/s00466-022-02261-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00466-022-02261-z
Keywords
- Fuzzy partial differential equations
- Possibility
- Polymorphic uncertainty modeling
- Uncertainty quantification
- Low-rank tensor formats
- Parametric partial differential equations
- Linear elasticity
- Homogenisation
- Tensor trains