1 Introduction

The initiation and development of cardiovascular diseases has been associated with changes in the mechanical properties of the underlying arterial tissue [1, 2]. Measures reflecting arterial stiffness, such as the pulse wave velocity (PWV) [3], the stiffness index (\(\upbeta \)) [4], and the pressure-strain elastic modulus (Ep) [5], are often used to assess patients in the clinic [2, 6, 7]. The amount of information obtainable from these popular clinical measures is limited, however, since they typically average stiffness over the artery or assume a constant stiffness despite the distinctive non-linear stiffening behavior of the arterial wall [8]. Several research groups have tried to address these shortcomings and proposed methods to identify more realistic patient specific mechanical properties of arteries in vivo (in the living human) [9,10,11,12,13]. These studies use non-linear continuum-mechanical models and identify the material parameters by fitting the models to clinical data. The fitting is performed by solving an optimization problem; typically minimization of the non-linear least-squares differences of measured and predicted pressure-radius response.

Although these new methods offer more insight into the mechanical properties of arteries in vivo, they suffer a fundamental problem: the non-linear optimization problem to be solved is non-convex and generally possesses local solutions which are not the global solution. This becomes a problem when the new information is used to develop diagnostic tools and intervention criteria. Intrasubject comparison requires a set of parameters for each patient which is uniquely relatable to a ‘normal’ set for healthy individuals, and the global solution is a natural candidate.

The issue of non-convexity is commonly addressed by using a gradient-based optimization algorithm which finds a local solution in combination with multiple starting points [11,12,13]. The best local solution is then taken to be the global solution. Unfortunately, there is no guarantee that this heuristic method identifies the global solution. Furthermore, no estimate can be provided for the difference between the best local and the true global solution.

In biomedicine, simulated annealing, particle swarm and genetic algorithms are frequently used to solve optimization problems [14,15,16,17], but these algorithms are also of heuristic nature and do not necessarily identify the global solution. Deterministic global optimization methods such as branch-and-bound (B&B) on the other hand, guarantee that in finite time, a solution is found whose optimal value is at worst a user-specified value higher than that of the global solution [18].

B&B algorithms have been used previously to solve least-squares minimization problems to global optimality [19, 20]. These studies show that the algorithm’s efficiency strongly depends on the reformulation of the original optimization problem and benefits from a low number of unknown parameters and simple functional expressions. The parameter identification method in Gade et al. [13] exhibits these characteristics, thus making B&B a suitable candidate when searching for the global solution. General purpose global optimization solvers exist that can be used as black-box solvers [21,22,23,24], but efficient global optimization requires exploitation of problem-specific characteristics, something which is difficult without access to source code.

In this paper, we propose a B&B-type algorithm for the parameter identification method presented in Gade et al. [13] and explore the feasibility of computing global solutions for the mechanical properties of arteries in vivo. In particular, we compute global solutions for the clinical data of three subjects. For each data set we analyze four successively larger parameter regions around a candidate global solution to study the algorithms efficiency.

2 Parameter identification method

The parameter identification method is taken from Gade et al. [13]. For a thorough description and discussion, the reader is referred to the original paper.

Fig. 1
figure 1

Unloaded (a) and loaded (b) state of an artery. The coordinates r, \(\theta \), and z of the cylindrical coordinate system are associated with the radial, circumferential, and axial direction, respectively. The thin dark-grey lines represent the collagen fibers embedded in the isotropic matrix which is colored in light grey

In the parameter identification method, an artery is treated as a homogeneous, incompressible, thin-walled cylinder which consists of an isotropic matrix with embedded collagen fibers. In the unloaded state, i.e. outside the human body, the artery has an inner radius \(R_\mathrm {i}\), a wall thickness H, and a length L, see Fig. 1a. In vivo, the artery is subjected to the blood pressure P from the inside and stretched to a length l which is invariant with respect to the changing blood pressure [25], see Fig. 1b. The stretch in the axial direction, i.e. \(\lambda _z\!=\!l/L\), results in an axial reaction force F. Furthermore, the inner radius and wall thickness in the loaded state are denoted with lower case letters, i.e. \(r_\mathrm {i}\) and h, and the cross-sectional area of the arterial wall \(A\!=\!2\pi r_\text {i}h+\pi h^2\) is constant because the arterial tissue is assumed to be incompressible and the axial stretch is constant. In this loaded state, two sets of stresses are calculated for an artery: equilibrium stresses and constitutively determined stresses.

By stating equilibrium in the circumferential and axial direction, the corresponding equilibrium stresses are calculated as, i.e. Laplace laws,

$$\begin{aligned}&\sigma _{\theta \theta }^\text {Lp}=\left( \frac{r_\text {i}}{h}+\frac{1}{2}\right) P, \end{aligned}$$
(1)
$$\begin{aligned}&\sigma _{zz}^\text {Lp}=\frac{\pi r_\text {i}^2P+F}{\pi h\left( 2r_\text {i}+h\right) }, \end{aligned}$$
(2)

respectively. The axial reaction force cannot be measured in vivo and is estimated by assuming that the ratio between the axial and circumferential stresses is \(\gamma \!=\!0.59\) at the mean arterial blood pressure \({\bar{P}}\!=\!13.3~\text {kPa}\) [9]. Using Eqs. (1) and (2), the axial force is estimated as

$$\begin{aligned} {\bar{F}}={\bar{P}}\pi \left[ \frac{\gamma }{2}\left( 2{\bar{r}}_\text {i}+{\bar{h}}\right) ^2-{\bar{r}}_\text {i}^2\right] , \end{aligned}$$
(3)

where the inner radius \({\bar{r}}_\text {i}\) and the wall thickness \({\bar{h}}\) are associated with \({\bar{P}}\).

The second set of stresses is determined using the Holzapfel–Gasser–Ogden (HGO) constitutive model [26]. Following Gade et al. [13], the constitutively determined stresses are: in the circumferential direction

$$\begin{aligned} \sigma _{\theta \theta }^{\text {mod}}= 2c\Bigl [\lambda _{\theta }^2-\left( \lambda _{\theta }\lambda _z\right) ^{-2}\Bigr ]+ 4k_1\left( I_4-1\right) \exp {\!\left[ k_2\left( I_4-1\right) ^2\right] }\lambda _{\theta }^2\cos ^2\beta , \end{aligned}$$
(4)

and in the axial direction

$$\begin{aligned} \sigma _{zz}^{\text {mod}}= 2c\Bigl [\lambda _z^2-\left( \lambda _{\theta }\lambda _z\right) ^{-2}\Bigr ] + 4k_1\left( I_4-1\right) \exp {\!\left[ k_2\left( I_4-1\right) ^2\right] }\lambda _z^2\sin ^2\beta , \end{aligned}$$
(5)

where \(c\!>\!0\) is a material constant describing the isotropic matrix, \(k_1,k_2\!>\!0\) are material constants associated with the embedded collagen fibers, and \(\beta \in [0,\pi /2]\) is the pitch angle of the symmetrically arranged collagen fibers relative to the circumferential direction, see Fig. 1a. Furthermore, \(I_4\) is the squared stretch along the collagen fibers calculated as

$$\begin{aligned} I_4 = \lambda _{\theta }^2\,\cos ^2\!\beta +\lambda _z^2\,\sin ^2\!\beta , \end{aligned}$$
(6)

and \(\lambda _\theta \) is the stretch in the circumferential direction defined as the ratio of the loaded to the unloaded mid-wall circumference, i.e.

$$\begin{aligned} \lambda _{\theta }= \frac{\pi \left( 2r_\mathrm {i}+h\right) }{\pi \left( 2R_\mathrm {i}+H\right) } =\frac{2r_\text {i}+h}{R_\text {i}+\displaystyle \sqrt{R_\text {i}^2+\lambda _z h \left( 2r_\text {i}+h \right) }}\text {,} \end{aligned}$$
(7)

where H has been replaced by \(R_\mathrm {i}\), \(\lambda _z\), \(r_\mathrm {i}\), and h using that the wall volume is constant due to incompressibility.

The equilibrium stresses in Eqs. (1) and (2) are fully determined given a data set comprising time-resolved blood pressure (P) and inner radius \((r_\mathrm {i})\) measurements sampled n times and information about the cross-sectional area (A) of the arterial wall. The constitutively determined stresses in Eqs. (4) and (5), however, contain six unknown model parameters: the unloaded inner radius \(R_\text {i}\), the axial stretch \(\lambda _z\), the material constants c, \(k_1\), \(k_2\), and the angle \(\beta \). These parameters are determined by solving the following non-convex weighted least-squares minimization problem:

figure a

where \({{\mathbb {U}}}{{\mathbb {P}}}\) denotes upper problem in anticipation of the B&B algorithm, \(\varvec{\kappa }\!=\!\bigl (R_\text {i},\lambda _z,c,k_1,k_2,\beta \bigl )\) is the parameter vector, \(j\!=\!1,\dots ,n\) indicates a data point sample, and the superscripts L and U denote lower and upper bound, respectively. The weighting factor \(\psi \!=\!0.99\) is used to let the objective function be dominated by the circumferential stresses [13]. The non-linear inequality constraints in (\({{\mathbb {U}}}{{\mathbb {P}}}\)) are not present in the original optimization problem and are introduced to reduce the parameter space to obtain physiologically reasonable values, see the Discussion. The fitting ranges and the limits on the non-linear constraints are based on experimentally observed values [27,28,29,30] and summarized in Table 1.

The parameter identification method in Gade et al. [13] has been numerically validated and the \(95\%\) limits of agreement have been determined for each of the six parameters, see Table 2. These limits represent the interval around the identified value in which the true parameter is lying and are used to create the vicinity regions around a candidate global solution later on in the Sect. 5. For parameter \(k_2\) the difference of the identified and the correct value increases as the correct value increases. To compensate for this systematic error, the \(95\%\) limits of agreement are based on the ratio instead [13].

Table 1 Fitting ranges of the model parameters and constraint limits
Table 2 Limits of agreement for each parameter taken from Gade et al. [13]

3 Global optimization approach

A B&B algorithm is used to solve the problem (\({{\mathbb {U}}}{{\mathbb {P}}}\)) to global optimality, meaning that the relative difference between the optimal values of the estimated and the true global solution is less than the tolerance \(\varepsilon \). The basic idea is to generate a sequence of non-increasing upper bounds \(U\!B\) and a sequence of non-decreasing lower bounds LB on the global solution. By successive subdivision of the parameter space along the B&B tree, \(\varepsilon \)-convergence is guaranteed in a finite number of iterations [18]. In each region of the parameter space, the original non-convex problem, i.e. the upper problem (\({{\mathbb {U}}}{{\mathbb {P}}}\)), is solved to local optimality and the upper bound is updated in case this local solution provides a lower value on the upper bound. Then the convex relaxation of the original problem, which will be referred to as the lower problem (\(\mathbb {LBP}\)), is solved to local, hence global optimality to establish a new lower bound on the global solution in the current region. If the lower bound is above the upper bound, the current region cannot contain the global solution and can therefore be excluded. Otherwise the current region is added to a list of active problems \({\mathcal {A}}\) which possibly contain the global optimum and the process is repeated until the relative difference between the upper and lower bound is less than the specified tolerance \(\varepsilon \).

3.1 Construction of convex relaxation

To facilitate the construction of the convex relaxation, the model equations (4) and (5) are not introduced in the objective function but enforced as constraints instead [19]. This makes the objective function convex in the auxiliary variables \(\sigma _{\theta \theta ,j}^\text {mod},\sigma _{zz,j}^\text {mod}\) which represent the model stresses and are introduced as additional unknowns to the optimization problem. The added constraints enforcing the model equations are non-linear due to the presence of bilinear, fractional, and componentwise convex terms. To facilitate the creation of convex relaxations of the model equations, the terms \(r_{\text {i},j}, R_\text {i},\lambda _z,\beta \) are replaced by scaled counterparts according to

$$\begin{aligned} {\tilde{r}}_j=\frac{\left( 2r_{\text {i},j}+h_j\right) ^2}{s}, \; {\tilde{R}} = s\left( R_\text {i}+\sqrt{R_\text {i}^2+\lambda _z\displaystyle \frac{\displaystyle A}{\displaystyle \pi }}\right) ^{-2}, \; {\tilde{\lambda }}_z=\lambda _z^2, \; {\tilde{\beta }} = \sin ^2\beta , \end{aligned}$$
(8)

where \(s\!=\!1000\) is a scaling factor chosen such that the magnitude of \({\tilde{R}}\) is similar to the other unknown parameters. After substitution of the scaled counterparts in the model equations, the following auxiliary variables are introduced:

$$\begin{aligned} \begin{aligned} w_1\rightarrow c{\tilde{R}},\ w_2\rightarrow {\tilde{R}}{\tilde{\lambda }}_z,\ w_3\rightarrow c{\tilde{\lambda }}_z,\ w_4\rightarrow \displaystyle \frac{c}{w_2},\ w_5\rightarrow {\tilde{R}}{\tilde{\beta }},\\ w_6\rightarrow {\tilde{\lambda }}_z{\tilde{\beta }},\ w_7\rightarrow k_1{\tilde{R}},\ w_8\rightarrow k_1w_5,\ w_9\rightarrow k_1w_6,\\ w_{10} \rightarrow w_7-w_8,\ w_{11,j}\rightarrow {\tilde{r}}_j\left( {\tilde{R}}-w_5\right) +w_6-1,\\ w_{12,j}\rightarrow w_{11,j}w_{10},\ w_{13,j}\rightarrow w_{11,j}w_9,\ w_{14,j}\rightarrow k_2w_{11,j}^2, \end{aligned} \end{aligned}$$
(9)

where \(\rightarrow \) means an auxiliary variable replacing the term following the arrow. The auxiliary variables \(w_1\), \(w_2\), \(w_3\), \(w_5\), \(w_6\), \(w_7\), \(w_8\), \(w_9\), \(w_{12,j}\), \(w_{13,j}\) replace bilinear terms, \(w_4\) a fractional term, and \(w_{14,j}\) componentwise convex terms. For each of the bilinear, fractional, and componentwise convex auxiliary variables in Eq. (9), four inequality constraints are added to the optimization problem, see “Appendix A”. The auxiliary variables \(w_{10}\) and \(w_{11,j}\) are enforced as linear equality constraints and are introduced to decrease the total number of auxiliary variables. The resulting intermediate optimization problem is

figure b

where \({\varvec{{u}}}\!=\!\left[ \tilde{\varvec{\kappa }},\varvec{\sigma }^\text {mod}_{\theta \theta },\varvec{\sigma }^\text {mod}_{zz},{\varvec{{w}}}\right] \), \(\tilde{\varvec{\kappa }}\!=\!\left( {\tilde{R}},{\tilde{\lambda }}_z,c,k_1,k_2,{\tilde{\beta }}\right) \), and \({\varvec{{w}}}\) contains the auxiliary variables \(w_a\). Due to the introduction of auxiliary variables, the degree of non-linearity in the model equations has been greatly reduced. However, the model equations in (\(\mathbb {INP}\)) still have componentwise convex terms of the form \(x\exp \!\left( y\right) \) which are addressed according to “Appendix A.4”. For these terms, instead of introducing new auxiliary variables, the relaxation is incorporated directly into the model equations. The complete convex relaxation of (\({{\mathbb {U}}}{{\mathbb {P}}}\)) now reads

figure c

where

$$\begin{aligned} e={\left\{ \begin{array}{ll} \exp {w_{14,j}^\text {L}}, &{} \text { if } w_{14,j}^\text {U}=w_{14,j}^\text {L}, \\ \exp {w_{14,j}^\text {L}}\frac{\displaystyle \exp {\!\left( w_{14,j}^\text {U}-w_{14,j}^\text {L}\right) }-1}{\displaystyle w_{14,j}^\text {U}-w_{14,j}^\text {L}}, &{} \text { otherwise.} \end{array}\right. } \end{aligned}$$
(10)

The non-linear inequality constraints present in (\({{\mathbb {U}}}{{\mathbb {P}}}\)) are enforced as simple box constraints on the corresponding (auxiliary) variables \({\tilde{R}}\), \(w_{11,j}\), and \(w_{14,j}\). The bounds on auxiliary variables \(w_{11,j}\), \(w_{12,j}\), and \(w_{13,j}\) are non-trivial due to the non-linear dependence on \({\tilde{R}}\), \({\tilde{\lambda }}_z\), and \({\tilde{\beta }}\), see “Appendix B”. Furthermore, tight bounds on the auxiliary variables representing the model stresses are added, see Sect. 4.1.

The lower bounding problem (\(\mathbb {LBP}\)) has in total \(16+6n\) variables and \(37+21n\) constraints. A comparison of the size of (\({{\mathbb {U}}}{{\mathbb {P}}}\)) and (\(\mathbb {LBP}\)) is provided in Table 3.

Table 3 Formulation statistics of (\({{\mathbb {U}}}{{\mathbb {P}}}\)) and (\(\mathbb {LBP}\))

3.2 Branching method

A crucial choice for the efficiency of a B&B algorithm is the branching method, i.e. the selection of the parameter to branch on. The proposed strategy follows the ideas in Esposito and Floudas [19] and considers the deviations between the values of the auxiliary variables \(w_a\in {\varvec{{w}}}\) and the non-convex terms they replace expressed in terms of the six model parameters \(p\in \tilde{\varvec{\kappa }}\). The deviations are thus calculated as

$$\begin{aligned} \delta _{a}=\left| w_a^*-w_a\!\left( \tilde{\varvec{\kappa }}^*\right) \right| , \end{aligned}$$
(11)

where \(w_a^*\) denotes the value of the ath auxiliary variable at the solution of (\(\mathbb {LBP}\)) and \(w_a\!\left( \tilde{\varvec{\kappa }}^*\right) \) is the ath auxiliary variable expressed as an explicit function, see Eq. (9), of the model parameters and evaluated at the solution of (\(\mathbb {LBP}\)). The deviation of auxiliary variable \(w_1\) is for example defined as

$$\begin{aligned} \delta _1=\left| w_1^*-c^*{\tilde{R}}^*\right| . \end{aligned}$$
(12)

In order to account for \(w_{14,j}\) being an exponent, the associated deviation is calculated using

$$\begin{aligned} \delta _{14,j}=\left| \exp \!\left[ w_{14,j}^*\right] -\exp \!\left[ w_{14,j}\!\left( \tilde{\varvec{\kappa }}^*\right) \right] \right| , \end{aligned}$$
(13)

instead.

Given the deviations in Eqs. (11) and (13), a region-reduction measure is introduced for every model parameter according to

$$\begin{aligned} r_p=\frac{p^\mathrm {U}-p^\mathrm {L}}{p^\mathrm {U,orig}-p^\mathrm {L,orig}}, \end{aligned}$$
(14)

where \(p^\mathrm {U},p^\mathrm {L}\) are the upper and lower bounds of parameter p in the current region and \(p^\mathrm {U,orig},p^\mathrm {L,orig}\) are the original upper and lower bounds. Finally, a total deviation \(\delta _{p}\) is calculated for every model parameter p by summing up all \(\delta _a\)’s in which this parameter is present and multiplying by the corresponding region reduction measure \(r_p\). The parameter which has the highest \(\delta _p\) is chosen to branch on and the corresponding region is bisected.

3.3 Proposed branch-and-bound algorithm

The rudimentary B&B algorithm described in the beginning of Sect. 3 is adapted to the properties of the problem at hand. In Gade et al. [13], the non-convex problem is solved from several starting points and the solution with the lowest error function value is taken to be the global solution. The same methodology is applied here when solving the upper problem for the first time. The supposedly good estimate of the global solution allows for not solving the upper problem until the lower bound reaches a user-specified threshold t.

The solution of both (\({{\mathbb {U}}}{{\mathbb {P}}}\)) and (\(\mathbb {LBP}\)) can be accelerated by using a good starting point. Since the upper problem is solved prior to the lower problem, little information about the current region is available. Therefore, when solving (\({{\mathbb {U}}}{{\mathbb {P}}}\)) the only criteria for choosing a starting point is whether it is feasible or not. For that purpose ten feasible or nearly feasible starting points are generated using Latin Hypercube Sampling [31] in the current parameter region. The starting point of the lower problem is determined by using the solution of the upper problem if available, otherwise it is generated using the same method as for the upper problem.

To speed up the B&B algorithm even further it is parallelized. Since it is not essential for the B&B algorithm that the sequence of lower bounds is always non-decreasing, it is possible to split the list of active problems \({\mathcal {A}}\) into \({\mathcal {A}}_1,\dots ,{\mathcal {A}}_v\), where v is the number of available CPUs, and assess them individually.Footnote 1 The individual assessment is done for a certain time T and then \({\mathcal {A}}_1,\dots ,{\mathcal {A}}_v\) are merged again into one list of active problems and the process is repeated until in less than 1000 active regions the lower bound is below a user-specified threshold t. From then on, the B&B algorithm is executed serially to make sure that the sequence of lower bounds is non-decreasing considering all active regions. The whole B&B algorithm is summarized in Algorithm 1.

figure d
figure e

3.4 Implementation

The proposed B&B algorithm is implemented in Matlab R2019b [32]. The upper problem (\({{\mathbb {U}}}{{\mathbb {P}}}\)) is solved using the function fmincon with the interior-point optimization algorithm. For numerical efficiency the model parameters c, \(k_1\), \(k_2\), and \(\beta \) are replaced in (\({{\mathbb {U}}}{{\mathbb {P}}}\)) by scaled counterparts [13],

$$\begin{aligned} c=\mathrm {e}^{{{\tilde{c}}}}\text {,}\quad k_1=\mathrm {e}^{\tilde{k}_1}\text {,}\quad k_2=\mathrm {e}^{\tilde{k}_2}\text {,}\quad \text {and}\quad \beta =\arcsin \sqrt{{\tilde{\beta }}}\text {,} \end{aligned}$$
(15)

where the superscribed tilde indicates a scaled parameter. The analytical gradient of the objective function, the analytical gradient of the non-linear inequality constraints, and the analytical Hessian of the Lagrangian are determined with Maple 2019.1 [33] and supplied to fmincon. To solve the upper problem from 10 starting points, the MultiStart-class feature is used.

The lower problem (\(\mathbb {LBP}\)) is solved using the Interior Point Optimizer (IPOPT) software package [34] which is interfaced to Matlab as mex code. Again, the analytical gradient of the objective function, the analytical gradient of the constraints, and the analytical Hessian of the Lagrangian are provided.

Although both optimization algorithms are based on the interior-point method, fmincon solves (\({{\mathbb {U}}}{{\mathbb {P}}}\)) and IPOPT (\(\mathbb {LBP}\)) more efficiently. The same stopping criteria are used for both algorithms: absolute change of objective function value must be less than \(10^{-10}\), and constraint violation must be less than \(10^{-8}\).

4 Material

The material for this study comes from the abdominal aorta of three healthy, non-smoking Caucasian males, representative for each age group. Subjects I, II, and III are 25, 41, and 69 years of age, respectively. The pressure-radius data is taken from Sonesson et al. [35] and for details about the data acquisition the reader is referred to that paper. From the raw data we create a pressure-radius curve consisting of \(n\!=\!18\) equidistant data points for each subject according to Stålhand [36], see Fig. 2. The total number of data points is chosen as small as possible, see the Discussion. Neither the deformed wall thickness nor the deformed wall cross-sectional area were recorded in Sonesson et al. [35]. The deformed wall cross-sectional area A is therefore estimated according to \(A\!=\!19.6\!+\!0.8\cdot \text {age}\), where A is in \(\text {mm}^2\) and age is in years [36].

Fig. 2
figure 2

Pressure-radius data for subjects I, II, and III. The markers denote the 18 path-equidistant samples and the lines denote the complete pressure-radius loops of the corresponding subject

4.1 Bounds on the auxiliary variables representing model stresses

The pressure-radius measurements show distinctive hysteresis, see Fig. 2. When the input data is converted into the Laplace stresses according to Eqs. (1) and (2), the hysteresis is still present, see Fig. 3. The constitutive model in Sect. 2 assumes a purely elastic material, however, and neglects hysteresis accordingly. The curves describing the model stresses in the circumferential and axial direction must, therefore, be (mostly) contained within the corresponding hysteresis loops if they are to be the best solution to the least-squares optimization problem. Individual lower and upper bounds on each auxiliary variable representing the model stresses are therefore introduced based on the hysteresis loops of the Laplace stresses. Afterwards the average difference between the upper and lower bounds, excluding the endpoints, is calculated and the stress ranges smaller than this average difference are symmetrically enlarged. The lower and upper limits for the endpoints are created by using twice the average stress difference. The resulting bounds on the auxiliary variables representing the circumferential and axial model stresses are shown in Fig. 3.

Fig. 3
figure 3

Bounds on the auxiliary variables representing the circumferential and axial model stresses for subject I. The black circles denote the pressure-radius measurements converted into Laplace stresses, the black lines are the hysteresis loops, and the red bars are the bounds for each auxiliary variable representing the model stresses. (Color figure online)

5 Results

The proposed B&B algorithm is applied to the clinical data and four successively increasing regions in the parameter space around a candidate global solution for each of the three subjects is investigated. The candidate global solutions are determined by solving (\({{\mathbb {U}}}{{\mathbb {P}}}\)) with 100 starting points generated with Latin Hypercube Sampling over the entire parameter space defined by the fitting ranges in Table 1. For subjects I, II, and III, respectively, \(81\%\), \(79\%\), and \(64\%\) of all starting points end up at the same solution with the lowest objective function value. The fact that the same solution is not always found demonstrates the non-convexity of the optimization problem and the associated existence of local solutions which are not the global solution. The candidate global solutions are summarized in Table 4. Around these candidate solutions four vicinity regions are established: \(1\%\), \(10\%\), \(100\%\), and \(500\%\) of the limits of agreement for each parameter, see Table 2. The vicinity regions respect the parameter boundaries in Table 1 and the explicit parameter spaces are summarized in Table 5. For each subject and each vicinity region, a stopping-tolerance in the range \(\varepsilon \!=\!0.01{-}0.095\) is required for the B&B algorithm, with larger tolerances for larger vicinity regions, see Table 6. The threshold t until the B&B algorithm is executed in parallel is put to \(90\%\) of the objective function value of the respective candidate global solution.

During the initialization step of the B&B algorithm, the respective candidate global solution is identified in every case and a better solution for (\({{\mathbb {U}}}{{\mathbb {P}}}\)) is never found. The qualitative progress of the difference between the upper and lower bound is similar for each subject in the same vicinity region, see Fig. 4. For many iterations, the relative difference is unity, i.e. \(LB\!=\!0\), followed by a comparably rapid drop until the lower bound has reached approximately \(90\%\) of the upper bound and the progress is substantially slowed down. The total amount of B&B iterations to achieve the required tolerance is large, especially for the two largest vicinity regions. Although each B&B iteration requires only a fraction of a second, the total CPU solution times are huge and are only manageable through parallelization, see Table 6.

Table 4 Candidate global solutions of investigated subjects
Table 5 Investigated parameter spaces for each subject
Table 6 Solution statistics for each subject and vicinity region

6 Discussion

In all cases tested, the candidate global solution is identified during the initialization step of the B&B algorithm and no better solution is found, even after solving (\({{\mathbb {U}}}{{\mathbb {P}}}\)) several million times in the larger parameter regions. This is not an unusual behavior of deterministic global optimization methods which often require many iterations to verify that a promising solution found early in the process is indeed the global solution [18]. That no better solution of (\({{\mathbb {U}}}{{\mathbb {P}}}\)) is found highlights the successful combination of using multiple starting points, selection of starting points and the rescaling when solving (\({{\mathbb {U}}}{{\mathbb {P}}}\)) in the initialization step.

Initially a tolerance of \(\varepsilon \!=\!0.01\) was required for each subject and vicinity region. It was, however, soon realized that \(\varepsilon \!=\!0.01\) is not feasible for every optimization run, since the required computational power exceeded our available resources. An in-house cluster with 108 2.6 GHz cores and 576 GB RAM, as well as two personal computers, were running for approximately four months to solve all optimization runs with a combined CPU time of 5.4 years.Footnote 2 Especially the optimization runs for subject III lasted for a long time because the parameter space within each vicinity region is larger compared to the other two subjects. The increased parameter spaces for subject III are due to the higher \(k_2\)-value of the candidate global solution and the relative confidence interval for that parameter, see Sect. 2. The required tolerance is therefore lowered on an individual basis such that (\({{\mathbb {U}}}{{\mathbb {P}}}\)) is at least solved a few million times during which no better solution must be found.

Fig. 4
figure 4

Relative difference between lower and upper bound, \(\left( U\!B-LB\right) /U\!B\), during the solution process. The black, blue, and red lines correspond to subjects I, II, and III, respectively. The dash-dotted, dotted, dashed, and solid lines represent the vicinity regions \(1\%\), \(10\%\), \(100\%\), and \(500\%\), respectively. (Color figure online)

In its current form, the proposed B&B algorithm does not utilize parallelization once the lower bound has been lifted to \(90\%\) of the upper bound. This threshold is chosen for two reasons. First, the upper problem is solved again when the threshold has been passed. It is assumed that (\({{\mathbb {U}}}{{\mathbb {P}}}\)) becomes practically convex in the small parameter space of the respective active region and that a better solution is found in this state. The best solution of (\({{\mathbb {U}}}{{\mathbb {P}}}\)), i.e. the lowest upper bound, should be available throughout the B&B algorithm to exclude regions as early as possible and communication between the active regions split into parts is not provided. Second, when the list of active regions is split into several parts, the sequence of lower bounds will be different in each individually processed part. The more parts the list of active regions is split into and the longer these parts are processed individually, the more the lower bounds will diverge and the iteration number to achieve a certain tolerance will be overestimated. This overestimation has a step-like appearance where the tolerance progress is shifted towards higher B&B iterations. This is especially pronounced for subject III whose optimization runs have been parallelized the most, see Fig. 4. In order to get the correct tolerance progress towards the end of each optimization run, the B&B iterations are performed serially once the threshold is passed. Both aspects are of minor concern and the algorithm should be parallelized throughout if the goal is solely to identify mechanical properties of arteries with certification of global optimality.

Non-linear inequality constraints are added into the original optimization problem (\({{\mathbb {U}}}{{\mathbb {P}}}\)). Their main purpose is to reduce the parameter space to the physiological range. Circumferential stretches of \(0.8\!<\!\lambda _\theta \!<\!1.6\) are reported for the abdominal aorta [27, 37] and we take \(\lambda _\theta ^\text {L}\!=\!0.5\) and \(\lambda _\theta ^\text {U}\!=\!2\). The invariant \(I_4\) is a macroscopic quantity representing the squared stretch of collagen fibers. Hence, the upper bound on \(I_4\) can be related to the squared failure-stretch of collagen fibers and is estimated to be \(I_4^\text {U}\!=\!2\) [38, 39]. The lower limit on \(I_4\) is introduced to facilitate the construction of the lower bounding problem. Collagen fibers are, generally, considered to support tensile loads but buckle in compression [26]. Hence, the contribution of collagen in the constitutive model, cf. Sect. 2, should be omitted if \(I_4\!<\!1\). Due the difficulty in constructing a valid underestimator which is smooth at the transition from non-recruited \(\left( I_4\!<\!1\right) \) to recruited collagen fibers \(\left( I_4\!\ge \!1\right) \), collagen fibers are assumed to be always in extension. In order to test the validity of this assumption, the model parameters of the three subjects are identified with the original parameter identification method in Gade et al. [13] which does not restrict \(I_4\). Collagen fibers are recruited throughout the cardiac cycle in all cases, thus \(I_4\!\ge \!1\) appears to be a valid assumption. Lastly, the upper bound of the exponent in the exponential term appears explicitly in some constraints of (\(\mathbb {LBP}\)), i.e. \(\exp w_{14}^\text {U}\!=\!\exp \!\left[ k_2^\text {U}\left( I_4^\text {U}-1\right) ^2\right] \). Due to the high upper bound on \(k_2\) in Table 1, the exponential expression might become very large. To prevent numerical problems, the exponent is limited to 20.

The major determinant for the efficiency of a B&B-type algorithm is the creation of the convex underestimator. The convex underestimator must be as close as possible to the original function while the number of additional auxiliary variables and constraints should be kept to a minimum. Several alternative formulations of (\(\mathbb {LBP}\)) were examined, but the presented one turned out to be most efficient. In particular, the newly proposed underestimator for componentwise convex terms [40] gave a substantial speed-up compared to the more general \(\alpha \)-estimator [41]. Furthermore, the introduction of physiologically inspired auxiliary variables and the tight bounds on the auxiliary variables representing the model stresses based on the hysteresis loop, cf. Sect. 4.1, decrease the required computational time. With respect to the branching method, other variants were tested, in particular strategies 1–4 proposed in Adjiman et al. [42]: (S1) bisect the parameter which has the largest region reduction measure defined in Eq. (14); (S2) determine the auxiliary variable whose underestimator possesses the highest separation distance to the actual term it replaces. Bisect the involved parameter according to S1; (S3) Determine the auxiliary variable which differs the most from the actual term it replaces at the solution of (\(\mathbb {LBP}\)), cf. Eq. (11). Bisect the involved parameter according to S1; and (S4) calculate the deviation of each auxiliary variable according to S3. For every parameter, sum up all the deviations that the parameter is involved in and bisect the parameter with the highest sum. The branching method described in Sect. 3.2 was found to be superior, however. Branching on auxiliary variables is not used since updating all lower and upper bounds is non-trivial for branching on a member of \({\varvec{{u}}}\) compared to a member of \(\tilde{\varvec{\kappa }}\).

From an implementation point of view, one point is worth discussing. By looking at the total amount of variables and constraints of (\(\mathbb {LBP}\)), see Table 3, it is obvious that the amount of samples, i.e. the number of pressure-radius pairs, must be as low as possible. For the three investigated subjects in this study, \(n\!=\!18\) is found to be the lower limit to get a solution which is independent from the number of samples. For other subjects this number might be different.

Considering the clinical application, the proposed B&B algorithm appears to be impractical. The solution time is orders of magnitude too large, even if only the vicinity around a candidate global solution is explored and the process is massively parallelized. Furthermore, the smaller the relative tolerance is required to be, the longer the identification process will take since the B&B algorithm slows down substantially towards smaller relative differences between lower and upper bound. However, no better solution than the one identified during the initialization is found, even after solving (\({{\mathbb {U}}}{{\mathbb {P}}}\)) several million times. This indicates that the heuristic approach of using multiple starting points works well for this kind of optimization problem. The original parameter identification method in Gade et al. [13] with its comparably small computational requirements, therefore, provides a good estimate of the global solution if it is not even obtained.