The framework of the empirical fragility assessment procedure proposed here consists of three steps. Firstly, the appropriate statistical models are selected. Then they are fitted to the available data, and finally their goodness-of-fit is assessed. The procedure is based on two, commonly found in the literature, assumptions that field response data are of high quality and the measurement error of the explanatory variable is negligible.
The procedure consists of two independent stages, developed in order to address the first two aims of this study. In the first stage, the fragility curves corresponding to the five levels of damage highlighted in Table 1 (DS), which can be used for risk assessment, are constructed in terms of the probability of a damage state being reached or exceeded by a building of a given class for different levels of tsunami intensity. The five curves are constructed by fitting a generalised linear model (GLM), originally developed as an extension to linear statistical models (McCullagh and Nelder 1989). This model is considered a more realistic representation of the relationship between a discrete ordinal (i.e. the counts of buildings being in the six damage states) and continuous (i.e. flow depth) or discrete (i.e. building class) explanatory variables. Indeed, such models recognise that the fragility curves are ordered, and therefore, they should not cross. They also recognise that the fragility curves are bounded between 0 and 1, and they take into account that some points have a larger overall number of buildings than others.
In the second stage, the probability that a collapsed building will be washed away, conditioned on tsunami intensity, is assessed. This conditional probability is also estimated by fitting the GLM to the collapsed data and their corresponding intensity measure levels. For this stage, the GLM is considered the most appropriate model to relate a discrete response variable (i.e. counts of collapsed buildings being washed away) with a continuous (i.e. flow depth) or discrete (i.e. building class) explanatory variable.
The importance of accounting for the structural characteristic of buildings in the fragility assessment is explored by the construction of generalised linear models of increased complexity in line with the recommendations of Rossetto et al. (2014). Initially, a simple parametric statistical model is introduced in order to relate the damage sustained by buildings irrespective of their class to a measure of tsunami intensity. The complexity of this model is then increased by adding a second explanatory variable expressing the structural characteristics of each building class, and the potential superiority of the most complex model is assessed using appropriate statistical measures.
Finally, the ability of the flow depth to predict the damage data, taking into account the building class, is investigated by assessing whether the constructed models adequately fit the available damage data through graphical model diagnostics.
Selection of generalised linear models
The GLMs methodology implies firstly, the selection of an appropriate statistical distribution for the response variable (i.e. damage state), which will be referred to as the random component of the model; secondly, the expression of the damage probability exceedance (or fragility function μ) through a chosen linear predictor η and link function g (this will be referred to as the systematic component of the model). When using GLM, the linear predictor is an additive function of all explanatory variables included in the model.
Overview of model components
In the first stage of the analysis, the damage response consists in six discrete damage levels: each building in the sample may be described by 1 out of 6 possible outcomes, which makes the multinomial distribution (Forbes et al. 2011) an adequate random component for this stage. Through the expression of the systematic component, complete ordering (ordered model) or partial ordering of these outcomes may be considered. Although the damage scale presented in Table 1 presents fully ordered outcomes (i.e. damage states of strictly increasing severity), the partially ordered model is more complex; thus, it may provide a significantly better fit to the data. Therefore, both types of models should be considered and quantitatively compared. In the second stage of the analysis, the damage response consists in two discrete damage levels: each building which has collapsed may have been washed away or not. This makes the binomial distribution the most appropriate random component for the second stage.
The relationship between the estimated fragility curve, the linear predictor and the link function is illustrated in Fig. 3 for the three link functions applicable to multinomial and binomial distributions, namely the probit, logit and complementary log–log functions. Note that the complementary log–log link shall not be considered in this study, unless the model appears to systematically underestimate observations for high damage probabilities. Indeed, this link provides estimations close to the logit link except for a heavier right tail and cannot be differentiated when the linear predictor takes the value of zero.
For each stage of the analysis, the chosen random components and possible expressions for the systematic components will be introduced. Candidate models will be evaluated through measures of relative goodness-of-fit in order to select the best fitting model.
First stage
The construction of fragility curves in the first stage of analysis is based on the selection of an ordered and partially ordered probit model of increasing complexity. For this stage, the response variable is expressed in terms of the counts of buildings suffering a given damage state (DS = ds
i
for i = 0, …, n). The explanatory variables are expressed in terms of a tsunami intensity measure and the building class.
The random component expresses the probability of a particular combination of counts y
ij
of buildings, irrespective of their class, suffering damage ds
i=0,…, n
for a specified tsunami intensity level, im
j
, which is considered to follow a multinomial distribution:
$$y_{ij} \sim \prod\limits_{i = 0}^{n} {\frac{{m_{j} !}}{{y_{ij} !}}P\left( {{\text{DS}} = {\text{ds}}_{i} |im_{j} } \right)^{{y_{ij} }} }$$
(1)
Equation (1) shows that the multinomial distribution is fully defined by the determination of the conditional probability, P(DS = ds
i
|im
j
), with \(m_{j}\) being the total number of buildings for each value of \(im_{j}\). This probability can be transformed into the probability of reaching or exceeding ds
i
given im
j
, P(DS ≥ ds
i
|im
j
), essentially expressing the required fragility curve, as:
$$P\left( {{\text{DS}} = {\text{ds}}_{i} |im_{j} } \right) = \left\{ {\begin{array}{*{20}l} {1 - P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} } \right) } \hfill & {i = 0} \hfill \\ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} } \right) - P\left( {{\text{DS}} \ge {\text{ds}}_{i + 1} |im_{j} } \right)} \hfill & {0 < i < n} \hfill \\ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} } \right)} \hfill & {i = n} \hfill \\ \end{array} } \right.$$
(2)
The fragility curves are expressed in terms of an ordered probit link function using the logarithm of the tsunami intensity measure (which is essentially the lognormal distribution, adopted by the majority of the published fragility assessment studies), forming a model which will be referred to as M.1.1.1, in the form:
$$\varPhi^{ - 1} \left[ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} } \right)} \right] = \theta_{0,i} + \theta_{1} \log \left( {im_{j} } \right) = \eta$$
(3)
where θ = [θ
0i
, θ
1] is the vector of the “true”, but unknown parameters of the model; θ
1 is the slope, and θ
0i
is the intercept for fragility curve corresponding to ds
i
; η is the linear predictor; Φ
−1[·] is the inverse cumulative standard normal distribution, expressing the probit link function. In order to build an ordered model, the systematic component, expressed by Eq. (3), assumes that the fragility curves corresponding to different damage states have the same slope, but different intercepts. This ensures that the ordinal nature of the damage is taken into account leading to meaningful fragility curves, i.e. curves that do not cross.
The assumption that the fragility curves have the same slope for all damage states is validated by comparing the fit of the ordered probit model with the fit of a partially ordered probit model, which relaxes this assumption. It should be noted that the latter model considers that the damage is a nominal variable and disregards its ordinal nature by allowing the slope of the fragility curves for each damage state to vary independently. This model (M.1.1.2) is constructed by transforming the linear predictor, η, in the form:
$$\varPhi^{ - 1} \left[ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} } \right)} \right] = \theta_{0,i} + \theta_{1,i} \log \left( {im_{j} } \right) = \eta$$
(4)
The complexity of the aforementioned systematic components [i.e. eqs. (3) and (4)] is then increased by adding a new explanatory variable which accounts for building class. Thus, a nested model (M.1.2) is obtained, which systematic component has the form:
$$\Phi^{ - 1} \left[ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} ,{\text{class}}} \right)} \right] = \theta_{0,i} + \theta_{1} \log \left( {im_{j} } \right) + \left( {\theta_{2} ,\theta_{3} } \right){\text{class}}$$
(5.1)
$$\Phi^{ - 1} \left[ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} ,{\text{class}}} \right)} \right] = \theta_{0,i} + \theta_{1,i} \log \left( {im_{j} } \right) + \left( {\theta_{2,i} ,\theta_{3,i} } \right){\text{class}}$$
(5.2)
In Eqs. (5.1) and (5.2), \((\theta_{2} ,\theta_{3} )\)and \((\theta_{2,i} ,\theta_{3,i} )\), respectively, are parameters of the model corresponding to the K − 1 categories of class, which is a discrete explanatory variable representing the various building classes. Let us note that in this case (inclusion of a categorical explanatory variable), class is dummy coded (i.e. the K − 1 values of the variable class become a binary variable, to each combination of 0–1 corresponds a different building class). The model whose systematic component is expressed by Eq. (5.1) assumes that the slope of the fragility curves is not influenced by the building class. By contrast, the intercept changes according to the building class.
The model further expands in order to account for the interaction between the two explanatory variables (the resulting model is termed M.1.3). Accounting for the interaction means that the slope changes according to the building class. This means that the rate of increase in the probability of exceedance of a given damage state, as flow depth increases, depends on the construction type of the building (i.e. the two variables do not act independently). Therefore, the systematic component is rewritten in the form:
$$\Phi^{ - 1} \left[ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} ,{\text{class}}} \right)} \right] = \theta_{0,i} + \theta_{1} \log \left( {im_{j} } \right) + \left( {\theta_{2} ,\theta_{3} } \right){\text{class}} + \theta_{12} \log \left( {im_{j} } \right){\text{class}}_{1} + \theta_{13} \log \left( {im_{j} } \right){\text{class}}_{2}$$
(6.1)
$$\Phi^{ - 1} \left[ {P\left( {{\text{DS}} \ge {\text{ds}}_{i} |im_{j} ,{\text{class}}} \right)} \right] = \theta_{0,i} + \theta_{1,i} \log \left( {im_{j} } \right) + \left( {\theta_{2,i} ,\theta_{3,i} } \right){\text{class}} + \theta_{12,i} \log \left( {im_{j} } \right){\text{class}}_{1} + \theta_{13,i} \log \left( {im_{j} } \right){\text{class}}_{2}$$
(6.2)
Second stage
The second stage focuses on buildings which collapsed under tsunami actions (i.e. damage state ds5 as presented in Table 1). Based on the information found in the database, the failure mode of the collapsed buildings is a binary variable (i.e. they can either be washed away or not). This is reflected by the selection of the random component of the generalised linear models developed below. For this stage, the response variable is expressed in terms of the counts of collapsed buildings which are washed away and the explanatory variables are the same as in the first stage.
A simple statistical model is introduced first which relates the response variable with a measure of tsunami intensity, ignoring the contribution of the building class. Thus, for a given level of intensity im
j
, the counts x
j
of washed away collapsed buildings are assumed to follow a discrete binomial distribution, as:
$$x_{j} \sim \left( {\begin{array}{*{20}c} {m_{j} } \\ {x_{j} } \\ \end{array} } \right)\mu_{j}^{{x_{j} }} \left[ {1 - \mu_{j} } \right]^{{m_{j} - x_{j} }}$$
(7)
In Eq. (7), \(m_{j}\) is the total number of buildings, and μ
j
is the mean of the binomial distribution expressing the probability of a collapsed building being washed away given im
j
. The mean, μ
j
, is essentially the systematic component of the selected statistical model which is formed here in line with the linear predictors developed for the first stage as:
$$g\left( {\mu_{j} } \right) = \eta = \left\{ {\begin{array}{*{20}l} {\theta_{4} + \theta_{5} \log \left( {im_{j} } \right). } \hfill & 1 \hfill \\ {\theta_{4} + \theta_{5} \log \left( {im_{j} } \right) + \left( {\theta_{6} ,\theta_{7} } \right){\text{class}}.} \hfill & 2 \hfill \\ {\theta_{4} + \theta_{5} \log \left( {im_{j} } \right) + \left( {\theta_{6} ,\theta_{7} } \right){\text{class}} + \theta_{56} \log \left( {im_{j} } \right){\text{class}}_{1} + \theta_{57} \log \left( {im_{j} } \right){\text{class}}_{2} .} \hfill & 3 \hfill \\ \end{array} } \right.$$
(8)
where θ = [\(\theta_{4}\),\(\theta_{5}\), \(\theta_{6} ,\theta_{7} ,\theta_{56} ,\theta_{57}\)] is the vector of the “true”, but unknown parameters of the models; g(.) is the link function, which can have the form of a probit as well as logit function:
$$\begin{aligned} g\left( {\mu_{j} } \right) & = g\left( {P\left( {{\text{DS}} = {\text{washed}}\;{\text{away}}|{\text{collapse}},im_{j} ,{\text{class}}} \right)} \right) \\ & = \left\{ {\begin{array}{*{20}l} {\varPhi^{ - 1} \left( {P\left( {{\text{DS}} = {\text{washed}}\;{\text{away}}|{\text{collapse}},im_{j} ,{\text{class}}} \right)} \right)} \hfill & { . 1 {\text{ (probit)}}} \hfill \\ {\log \left( {\frac{{P\left( {{\text{DS}} = {\text{washed}}\;{\text{away}}|{\text{collapse}},im_{j} ,{\text{class}}} \right)}}{{1 - P\left( {{\text{DS}} = {\text{washed}}\;{\text{away}}|{\text{collapse}},im_{j} ,{\text{class}}} \right)}}} \right)} \hfill & { . 2 {\text{ (logit)}}} \hfill \\ \end{array} } \right. \\ \end{aligned}$$
(9)
Statistical model fitting technique
Having formed the aforementioned generalised linear models, they are then fitted to the field data. This involves the estimation of their unknown parameters by maximising the log-likelihood functions, as:
$$\begin{aligned} \theta^{\text{opt}} & = \arg \hbox{max} \left[ {\log \left( {L\left( \theta \right)} \right)} \right] \\ & = \left( {\begin{array}{*{20}l} {\arg \hbox{max} \left[ {\log \left( {\prod\limits_{j = 1}^{M} {\prod\limits_{i = 0}^{n} {\frac{{m_{j} !}}{{y_{ij} !}}P\left( {{\text{DS}} = {\text{ds}}_{i} |im_{j} ,{\text{class}}} \right)^{{y_{ij} }} } } } \right)} \right]} \hfill & {1{\text{st}}\;{\text{stage}}} \hfill \\ {\arg \hbox{max} \left[ {\log \left( {\prod\limits_{j = 1}^{M} {\left( {\begin{array}{*{20}c} {m_{j} } \\ {x_{j} } \\ \end{array} } \right)\mu_{j}^{{n_{j} }} \left[ {1 - \mu_{j} } \right]^{{m_{j} - x_{j} }} } } \right)} \right]} \hfill & {2{\text{nd}}\;{\text{stage}}} \hfill \\ \end{array} } \right. \\ \end{aligned}$$
(10)
Apart from the mean estimates of the unknown parameters, θ, the uncertainty around these values is also estimated in this study. Given that the data are grouped and that the water depth is the only explanatory variable, the uncertainty in the data is likely to be larger than the uncertainty estimated by the adopted generalised linear model. This over-dispersion is taken into account in the construction of the confidence intervals by the use of bootstrap analysis (Rossetto et al. 2014). According to this numerical analysis, 1,000 random samples with replacement are generated from the available data points. For each realisation a generalised linear model is fitted. For specified levels of tsunami intensity, the 1,000 values of the fragility curves corresponding to a given damage state are ordered and the values with 90 and 5 % probability of exceedance are identified.
Goodness-of-fit assessment
The proposed procedure is based on developing a number of realistic statistical models, which are then fitted to the available data. Which one provides the best fit? To answer this question, the relative as well as the absolute goodness-of-fit of the proposed models is assessed. The relative goodness-of-fit assessment aims to identify the model which provides the best fit compared with the available alternatives. The absolute goodness-of-fit aims to explore whether the assumptions on which the model is based upon are satisfied, thus providing a reliable model for the given data.
Relative goodness-of-fit
The relative goodness-of-fit procedure is based on whether the two candidate models are nested (i.e. the more complex model includes at least all the parameters of its simpler counterpart) or not.
The goodness-of-fit of two nested models is assessed using the likelihood ratio test. Generally the more complex model fits the data better given that it has more parameters. This raises the question as to whether the difference between the two models is statistically significant. The likelihood ratio test is used in order to assess whether there is sufficient evidence to support the hypothesis that the most complex model fits the data better than its simpler alternative. According to this test, the difference, D, in the deviances [−2log (L)] is assumed to follow a Chi square distribution with degrees of freedom df = df
simple model − df
complex model:
$$p = P\left( {D \ge d = - 2\frac{{\log \left( {L_{{{\text{simple}}\;{\text{model}}}} } \right) - \log \left( {L_{{{\text{complex}}\;{\text{model}}}} } \right)}}{{df_{{{\text{simple}}\;{\text{model}}}} - \, df_{{c{\text{omplex}}\;{\text{model}}}} }}} \right)$$
(11)
In Eq. (11), L denotes the likelihood function, and p the p value of the statistical test. The difference can be noted by random chance and therefore is not significant if we obtain a p value >0.05. This means that there is not sufficient evidence to reject the hypothesis; hence, the simpler model can be considered a better fit than its nested alternative. By contrast, the difference is significant, and therefore, the hypothesis can be rejected if p value <0.05.
The comparison of the fit of non-nested models (here, two models with identical linear predictors but different link functions) is assessed by the use of the Akaike information criterion (AIC) following the recommendations of Rossetto et al. (2014). This criterion is estimated as:
$${\text{AIC}} = 2k - 2\ln \left( L \right)$$
(12)
where k is the number of parameters in the statistical model. The model with the smallest AIC value is considered to provide a relatively better fit to the available data.
Absolute goodness-of-fit
The absolute goodness-of-fit of the best model is assessed by informal graphical tools. In particular, the goodness-of-fit of the best candidate models from each stage is assessed by plotting the observed counts of buildings for which DS ≥ ds
i
with their expected counterparts. The closer the points are to the 45 degree line, the better the model.