Introduction

All evaluations of the risks of extreme high river flows require methods to statistically estimate the upper quantiles on the basis of the observed data. Such methods are widely used in the development of regulations concerning the design of hydraulic structures, embankments, bridges and culverts, flood mitigation strategies and civil protection policies. Furthermore, it is widely recognized that estimation of changes in the frequency of floods related to climatic changes and analyzing trends in the high flow extremes are crucial for the adaptation and mitigation measures undertaken to cope with the negative impacts of climate change. For this purpose new methods taking into account non-stationarity of flood events are developed with a view to replace long-standing and well-established characteristics and principles of engineering design and to move from an equilibrium- or stationary-paradigm to one of a constant evolution that recognizes the dynamic nature of physical and socio-economic processes. Non-stationary statistical analysis can be performed on the observed annual/seasonal maxima series and on projections obtained from coupled hydrological and climate models. However, for methodical, computational and practical reasons this analysis is difficult and uncertain, which casts a shadow over both practical and theoretical efforts. This paper is a sequel to the Debele et al. (2017) paper where an application of the GAMLSS tools to the FFA was compared with two more traditional approaches. The results of the research presented in the latter indicate that the GAMLSS performs better than the other two methods applied from the point of view of its flexibility and a superior treatment of non-stationarity. The aim of this paper is to put the FFA analysis into the context of real-life applications and to discuss the pros and cons of different FFA approaches from the end-user perspective.

We describe briefly the history of flood frequency analysis (FFA) in “Some historical aspects of flood frequency analysis”. Some methodological problems and practical aspects of non-stationary flood frequency analysis (NFFA) are discussed in “Non-stationary process of flood formation and NFFA”. The GAMLSS package is described in “Introduction to the GAMLSS package” and its application to NFFA is presented in “Hydrological Applications and Assessment of the GAMLSS Software”. The relevant conclusions are presented in “Conclusions”.

Some historical aspects of flood frequency analysis

The flood frequency analysis arose from practical problem of how to ensure the safety of structures prone to high waters, and for how long. For ages constructors have coped with this problem; however, the methods they used have changed from trial-and-error methods to more and more sophisticated ones accompanying the growing potential of hydrological observations, mathematics and computing capabilities. Many artifacts from ancient times demonstrate the skills of their creators, both in terms of design and construction. The earliest indications of Dutch dike building date from the late Iron Age. In Roman times, dikes and dams were created on the territories of present France, Spain, Portugal, Syria and North African countries. One of the oldest functioning bridges in the world is the Pons Fabricius in Rome, Italy, which was built in 62 BC; the year after Cicero was a consul, to replace an earlier wooden bridge destroyed by fire. The oldest, the Caravan Bridge over the river Meles in Izmir, Turkey is dated c. 850 BC. Intact from antiquity, the bridges have been in continuous use ever since.

In many countries the beginning of the twentieth century was a period of introducing the statistical analysis of the maximum annual flows probability distribution as a basis for determining design characteristic in the form of the upper quantile corresponding to the specified return period—the expected life time of a structure. The return period is adopted according to the structure class but most commonly the 100-year return period is chosen and the corresponding quantile called a 100-year water (or flood) is thought to be a fair balance between protecting society and overly stringent regulation. Former procedures based on water level measurement defined usually the design characteristic as the highest observed water level plus 1 m. The design characteristics assessment from the probability distribution of the annual maximum flows entailed a number of well-known problems. As the true distribution remains unknown, the selection of an appropriate probability distribution model and the estimation method of its parameters were, and still are, the most important. The limited length of hydrological series makes a discussion of the proper choice of statistical model unjustified; although some goodness-of-fit criteria can point to the best model but their discriminate power remains low. Moreover, when the series length increases (even only by 1 year) the best distribution model can be different from that previously chosen. As the candidate distributions can have different tail heaviness, the consequences to design quantiles are obvious. To avoid this situation and to enable control over the results, in many countries the type of annual maxima distribution, and the estimation method as well, were fixed and the parameters and the quantiles were updated, e.g., every 5 years. Essentially, the chosen distribution form stemmed from tradition and multiple tests of the regional conformity. The estimation methods evolved from simple graphical fitting based on plotting position and a comparison of the empirical and theoretical quantiles to the method of moments, linear moments and mathematically advanced optimization methods (e.g., maximum likelihood method). The inventory and analysis of the FFA models used in different countries at the end of the twentieth century were presented by Cunnane (1989). Many have reached the status of more or less obligatory guidelines and rules [e.g., in the USA—Bulletin No 15 (1969) with further extensions and updates, in the UK—Flood Studies Report (1975) succeeded by Flood Estimation Handbook and Flood Estimation Guidelines, in Poland—Regulations (1969, 2007), and many others].

The research on methods for flood frequency estimation under the various climatic and geographic conditions found in Europe, carried out by the COST FloodFreq project over 20 years later (Castellarin et al. 2012) revealed that nowadays the distribution types in use did not change significantly from Cunnane’s report. They are: Gamma, Pearson 3, Log Pearson 3, Gumbel, Lognormal, Generalized Logistic and Weibull 3 with the newly introduced Generalized Extreme Value distribution (GEV) on the top of the list, Generalized Pareto and Two-Component Extreme Value (TCEV) distributions. Currently, the seasonal approach is more often used in the case of different seasonal flood generating processes within a year (e.g., Guidelines for flood frequency analysis…, 2005; Strupczewski et al. 2009, 2011; Vormoor et al. 2015).

With the explosive development of computing technology the guidelines for the FFA started to be accompanied by software packages often prepared by hydrologists themselves. Later, statisticians and specialists in numerical methods and programming languages professionally developed the software and offered it on the market. In the early 1980s the HOMS (Hydrological Operational Multipurpose System) project was implemented by the World Meteorological Organisation (WMO) to enable free exchange and transfer of technology in hydrology and water resources. This technology is usually in the form of descriptions of hydrological instruments, technical manuals or computer programs provided by the Hydrological Services of member countries of the WMO. The principle of the project is that the technology transferred is not only ready for use but also works reliably.

The “FFA statistics” used and developed by hydrologists as a separate branch of the statistics based on their own paradigms attracted the interest of good, professional statisticians who enriched its methodological range. Particularly noteworthy is the paper of Langbein (1949).

The R-project, an open source programming language and software environment for statistical computing and graphics that is supported by the R Foundation for Statistical Computing, plays an important role of a platform of statistical software repository and exchange (R Core Team 2017, https://www.r-project.org/).

The GAMLSS (Rigby and Stasinopoulos 2005) software is implemented in a series of packages in R language available from CRAN—the R library (https://www.r-project.org/). The platform is of general statistical use and it is not dedicated to hydrological problems with their specific requirements and constraints and, in particular, it is not directly applicable to FFA. The methodology of FFA laboriously developed for years is currently facing the challenge of adopting GAMLSS solutions.

Non-stationary process of flood formation and NFFA

To carry out reliable statistical inference we need to make significant assumptions about relations between the observed and unobserved data. Among these, the major assumption is simplicity, consistency and uniformity of Nature which can be understood as a premise of stationarity of the processes generating hydrological extremes, at least in the time scale of human life. Nowadays, the symptoms of climate change are an increasingly prominent topic in scientific and public discourse. Many scientific bodies claim that instrumental data analysis provides evidence that climate is changing and that the process is accelerating. “Rivers are a product of the climate”, Vojejkov wrote in his work “Climates of the globe and Russia in particular” (1884). Climate change must therefore result in changes of water resources and flood generation processes as a consequence. If so, the natural question to be answered is the persistence of these changes and how meaningful they can be.

Carbon dioxide, the main suspect in climate change, displays an exceptional persistence that renders its warming nearly irreversible for more than 1000 years. The warming due to non-CO2 greenhouse gases, although not irreversible, persists notably longer than the anthropogenic changes in the greenhouse gas concentrations themselves (Solomon et al. 2010). Are these changes already visible in the magnitudes and frequencies of flooding? And further, can the observed trends be extrapolated and to what extent?

Koutsoyiannis (2006) and Koutsoyiannis and Montanari (2007) give a clear answer:

  • The fluctuations of the mean are a normal feature of stationary processes, which, together with the limited length of the series of observation may lead to the impression that the underlying process is non-stationary.

  • A deterministic function representing trend is a function that can be produced only by deduction, independently of the data (a priori); e.g., by a model that could predict them.

  • On the contrary, according to common practice, the “trends” and “shifts” in the means are inferred by induction based on the data (a posteriori).

  • Therefore such fitted lines are not deterministic and do not represent non-stationarity.

We do not have reliable stochastic or deterministic models of the process which could simulate the climate drivers and their hydrological impacts with the desired accuracy.

Coupled hydrological and climate models running for various emission scenarios provide us with the projections of seasonal/annual maxima for the near (to 2050) and far (up to 2100) future.

But the problem arises: how can we trust the projections made for the time horizon of 50 years and more, given that the weather forecasts for a period longer than 3 days are not credible? Furthermore, the most uncertain factor in prediction by global climate models is the spatial–temporal distribution of rainfall which is usually the driving cause of floods and drought phenomena. So, a stationary model is sometimes preferable to a non-stationary one when the evolution in time of hydrological processes cannot be predicted reliably (Bayazit 2015; Serinaldi and Kilsby 2015; Lins and Cohn 2011).

Numerous sources of uncertainty influencing the projections of the temporal evolution of future changes (e.g., Serinaldi and Kilsby 2015; Koutsoyiannis 2013), however, do not intimidate us from analyses, hoping that they do give the best assessment of the future extreme events we are able to achieve nowadays. Enthusiasts of the idea of climate change and the possibility of its prediction announced to the world that stationarity is dead (e.g., Milly et al. 2008), and, although difficult, the non-stationary approach to water management has become mandatory. However, it seems that the question is not whether to use these methods, but how to use them in design and planning practices considering their immense uncertainty and who will benefit (or maybe loose) in a result. Many projects on “managing flood risk to keep pace with climate change” declare tackling flood risk now for future generations.

It should be noted that a number of projects on the adaptation to climate change carried out under the banner of the benefit for future generations improved the much neglected so far domains, for which funding would be difficult to get under the less publicized slogan. Whether the stationarity is dead or not “playing this game with Nature”, we can only monitor the dynamics of our time series and improve climate and hydrological models. Designers can use past events and trends only as an indication of the severity of effects likely to occur in the future, acknowledging the fact that it is very easy to confuse statistical significance with practical or substantive importance. They should focus on the size of the expected effects instead of their statistical significance. The same statistically significant trends for large and small river can be practically important for the small but not important for the large river, where changes due to the trend are only a small fraction of the observed flow and lie completely within the range of accuracy of their assessments.

Taking into account the accuracy of flood flow data is another important problem ignored by the scientists. Development of the FFA and NFFA methods is focused on their mathematical formulation where there is no place for a typically hydrological reasoning. It is shown in the large number of papers published in hydrological journals which are strictly mathematical without reference to actual hydrological situation even in the form of case studies (e.g., Rasmussen 2001; Askhar and Mahdi 2003; Jawitz 2004, and many others).

Introduction to the GAMLSS package

Generalized additive models for location, scale and shape is a modern distribution-based approach to parametric and semi-parametric regression models, where the parameters of the assumed distribution for the response variable can be modeled as linear and/or non-linear and/or non-parametric smoothing functions of the explanatory variables (Rigby and Stasinopoulos 2005). GAMLSS was proposed by Rigby and Stasinopoulos (2005), Stasinopoulos and Rigby (2007) and Stasinopoulos et al. (2008) as a way of overcoming some of the limitations associated with generalized linear models (GLM) and generalized additive models (GAM) (Nelder and Wedderburn 1972; Hastie and Tibshirani 1992, respectively). In GAMLSS the exponential family distribution assumption for the response variable is relaxed and replaced by a general distribution family called the GAMLSS family.

A GAMLSS model assumes that independent observations y i of a random variable Y for i = 1, 2, 3…, n have probability distribution function \( f_{Y} \left( {y_{i} |\theta^{i} } \right) \) with \( \theta^{i} = ( {\theta_{1}^{i} , \ldots ,\theta_{p}^{i} } ) \) vector of p parameters accounting for location, scale and shape of the random variable distribution. The number of parameters p is limited to four, and guarantees enough flexibility of distributions in most applications. In GAMLSS explanatory variables are introduced into the model through the ‘predictors’ η k related to parameters by monotonic link functions \( g_{k} \left( \cdot \right) \)—identity, log, inverse function and others. The GAMLSS package involves several important sub-models, relating the distribution parameters to explanatory variables through an additive model (Eq. 1) or semi-parametric additive model in linear (Eq. 2) and non-linear form (Eq. 3).

$$ g_{k} \left( {\theta_{k} } \right) = \eta_{k} = X_{k} \beta_{k} + \mathop \sum \limits_{j = 1}^{m} Z_{jk} \gamma_{jk} $$
(1)
$$ g_{k} \left( {\theta_{k} } \right) = \eta_{k} = X_{k} \beta_{k} + \mathop \sum \limits_{j = 1}^{m} h_{jk} \left( {x_{jk} } \right) $$
(2)
$$ g_{k} \left( {\theta_{k} } \right) = \eta_{k} = h_{k} \left( {X_{k} \beta_{k} } \right) + \mathop \sum \limits_{j = 1}^{m} h_{jk} \left( {x_{jk} } \right), $$
(3)

where \( \theta_{k} \) and η k are vectors of length n, X k is a known design matrix (fixed effects design matrix) of order n × m (a matrix of explanatory variables, i.e., covariates), β k is a parameter vector of length m and \( \sum\nolimits_{j = 1}^{m} {Z_{jk} \gamma_{jk} } \) represent random effects term, \( h_{jk} \left( \cdot \right) \) is an unknown function of explanatory variables x jk (smoothing term) and \( h_{k} \left( \cdot \right) \) is a known non-linear function. The dependence shown in Eq. 2 can be linear or smooth through smoothing terms, but this case will not be considered here. If m = 0 (additive terms removed), then models described by Eqs. 13 are reduced to the fully parametric form. For Eq. 2, for example, it reads:

$$ g_{k} \left( {\theta_{k} } \right) = \eta_{k} = X_{k} \beta_{k} . $$
(4)

The meaning of θ k depends on parameterization of the distribution. In the FFA mean/standard deviation parameterization \( \theta_{k} \) is equal to the mean value and the standard deviation of the distribution assumed. Then the model in Eq. 4 simply reduces to VGLM (vector generalized linear models) case where θ 1 = E(Y) and \( \theta_{2} = \sqrt {{\text{Var}}(Y)} \) and both g k (θ k ) = η k are linear functions of the explanatory variables.

The list of distributions included in GAMLSS is impressive and covers over 90 types, and the open design of the system allows attaching others.

The estimation methods in GAMLSS are based on the maximum likelihood principle. Fully parametric models are estimated by the maximum likelihood method. For models with random effects and smoothing (Eqs. 24) the penalized likelihood is used. Two basic algorithms are used for maximizing the likelihood. The CG algorithm is a generalization of the Cole and Green (1992) algorithm and it uses the first and (expected or approximated) second and cross-derivatives of the log-likelihood function with respect to the distribution parameters. The limitation of CG is that it requires the existence of expected values of the cross-derivatives of the log-likelihood function with respect to location, scale and shape parameters. However, for many probability density functions \( \left\{ {f_{Y} \left( {y|\theta } \right)} \right\} \) the parameters are informationally orthogonal; since the expected values of the cross-derivatives of the log-likelihood function are zero. The CG algorithm performs better for distributions with potentially highly correlated parameters. In the case of zero cross-derivatives of likelihood function, the simpler RS algorithm, which is a generalization of the algorithm by Rigby and Stasinopoulos (1996a, b), is more suitable. The RS algorithm does not use the cross-derivatives and does not require accurate starting values for the parameters to ensure convergence; and it is faster for larger data sets.

The GAMLSS package gives the opportunity for fast-fitting of different models to the data set and the model selection can be carried out by checking the significance of the fitting improvement, e.g., between stationary and non-stationary model by means of deviance statistics; however, other methods are possible.

The GAMLSS framework of statistical modeling is implemented in a series of packages in R. The packages can be downloaded from the R library, CRAN (https://www.r-project.org/). There are more packages for stationary and non-stationary analysis in R; e.g., extRemes (Gilleland and Katz 2016), FAdist (Aucoin 2015), PearsonDS (Becker and Klößner 2017).

Hydrological applications and assessment of the GAMLSS software

As stated above, the list of distributions included in GAMLSS contains over 90 types. Truncated, censored, log and logit transformed and finite mixture versions of these distributions can be also used. However, a hydrologist can be confused since he/she cannot find there the names of the distributions commonly used in FFA, or they are two-parameter distributions while the three-parameter, lower-bounded distributions are mostly in operational use. Furthermore, the names and acronyms of the GAMLSS family distributions and their parameterization, as well, are difficult for beginners and less experienced users of the package. Sometimes three-parameter FFA distributions are hidden in the generalized parameterization of GAMLSS. For instance, the Reverse Generalized Extreme Family distribution RGE is a re-parameterization of the three-parameter Weibull distribution. It is due to the fact that some parameterizations are computationally preferable to others in the sense that maximization of the likelihood function is easier.

Most GAMLSS-based research in hydrology addresses the application of two-parameter distributions from the GAMLSS family. Some GAMLSS applications in hydrological/climatological time-series are presented by Villarini et al. (2009a, b, 2010a, b, 2012), Machado et al. (2015), López and Francés (2013), Osorio and Galiano (2012), Hudson et al. (2008). GAMLSS have been used to model seasonal rainfall and temperature in Rome by Villarini et al. (2010a); they showed that the GAMLSS models could represent the magnitude and spread in the seasonal time series with parameters being a smooth function of time or teleconnection indices. GAMLSS models have been used for flood frequency analysis in Villarini et al. (2009a, b) and López and Francés (2013). The study by Machado et al. (2015) also applied GAMLSS to model flood data using historical information. Some studies applied GAMLSS to gridded datasets; e.g., Osorio and Galiano (2012) and Galiano et al. (2015) applied GAMLSS modeling framework to develop a methodology to account for non-stationarity existing in climate and hydrological processes and assessing the non-stationary spatial patterns of extreme droughts using gridded data of observed rainfall from regional climate models (RCMs). The summary of important applications of GAMLLS in NFFA is given in Table 1.

Table 1 Summary of important applications of GAMLSS in non-stationary FFA

A very few hydrological studies have applied the three-parameter GAMLSS distributions, e.g., López and Francés (2013) and Zhang et al. (2015). In both studies, the Generalized Gamma (GG) distribution was applied to develop a framework for flood frequency analysis for annual maximum daily flows and maximum daily precipitation. It should be noted that although GG is a three-parameter distribution, its parameters are scale and two shape parameters. The distribution is defined for values y > 0, and does not have a lower bound being a location parameter. So, from the point of view of FFA applications it can be treated as a two-parameter distribution, more flexible due to the two shape parameters. As shown by Strupczewski et al. (2008) two-shape-parameter distributions in the stationary case are likely to have greater flexibility for fitting the right tail of empirical distribution than their lower-bounded parameter counterparts. The accuracy of the skewness coefficient estimation from short series (e.g., Wallis et al. 1974) is low. Therefore, it is commonly approved in the NFFA that the shape parameter remains constant. Allowing the skewness to vary in time seems unrealistic in terms of reliable estimation.

A review of the extensive available literature (Table 1) shows that the package and distributions selected from GAMLSS family can be used in the NFFA. It should be stressed that the applications described above are research-based. The reliability of the results obtained using two-parameter models should be evaluated in simulation experiments. None of the above noted papers addressed the practicality and efficiency of GAMLSS algorithm using Monte Carlo simulations.

In this topical issue of the Acta Geophysica we present a comparison of three non-stationary approaches to FFA (Debele et al., this issue) with time as a covariate one of them is the GAMLSS distributions family estimated by GAMLSS software. Using the possibility of introducing new distributions to the GAMLSS family we have added Pearson three distribution. The fitting works only with a CG algorithm, with and without covariates. However, our attempts to introduce other three-parameter distributions, namely Lognormal type 3 and GEV, were unsuccessful and the estimation procedures failed. The lower-bounded distributions are generally difficult to fit by the maximum likelihood method and the linear moments method (LMM) is recommended (e.g., Hosking and Wallis 1997; Strupczewski et al. 2001a, b; Markiewicz et al. 2010). In a comparison study we examined the performance of GAMLSS algorithm in the case of false distribution assumption (T = Lognormal 3 with imposed linear trends in mean and standard deviation, H = RGE). The relative bias (RB) and relative root mean square error (RRMSE) were computed for time-dependent moments and 99% quantiles. The RB and the RRMSE were calculated for 10,000 simulations with a sample size of n = 50, 100 and 200 generated by considered parent distribution. Compared to the weighted least squares (WLS), described in Strupczewski and Kaczmarek (2001), Kochanek et al. (2013) and Strupczewski et al. (2015), the algorithm GAMLSS showed better efficiency in the estimation of the trend in the standard deviation and worse in the mean value. This finding also applies to the real flood data when sometimes the algorithm GAMLSS identified the trend in the mean as the opposite to that observed. The accuracy measures RB and RRMSE of the shape parameter were estimated best by GAMLSS. This result, together with the high efficiency in the evaluation of the trend in the standard deviation, has been translated into good estimates of the 99% quantile for the short and long series.

While applying GAMLSS to the NFFA a hydrologist can encounter some other problems. Perhaps the most important is the problem of the confidence intervals (CI) of time-dependent quantiles. The GAMLSS package allows us to evaluate CI by the delta method, likelihood or deviance profiles (Rigby et al. 2014), but in more complicated cases, which are typical in FFA, e.g., seasonal or multi-model approaches, their application is not obvious.

Conclusions

The non-stationarity of flood peak flows due to climate change creates numerous problems of both the theoretical and applied kind. In addition to describing the observed time series the hydrologist’s goal is also to predict (extrapolate) future flows for the length of a planning horizon. The uncertainty of predictions derived from theoretical methods governs the results as well as the assumed distribution type. The inductive (data-based) methods of trend detection cannot provide proof of the non-stationarity of flood generation processes, nor for its persistence in the future. However, from a practical point of view, we are interested only in the long-term trends covering the expected life time of design structures (100 years or more for major structures), despite the fact that both accuracy of prediction and demands are decreasing with time.

Climate change and its hydrological impacts became a front page topic in the mass media and scientific journals and political issues creating a sense of danger at the individual, local and global scales. Only a positive trend detected in the flood data can sustain this feeling, so only positive trends are “politically correct”. Hunters of trends in the time series of high water should be disappointed when a downward trend is detected, or if a positive trend identified as statistically significant turns out to be physically insignificant and therefore irrelevant to hydrological design.

Application of non-stationary methods in the practice of the FFA is conditioned by the availability of software that allows the assessment of the impact of changes in the flood generating processes on design characteristics, i.e. the upper quantiles. In this paper the possibilities of the most popular R software—the GAMLSS package are described with a view to its being applied in flood frequency analysis.

The GAMLSS package is the universal, flexible and complex statistical tool for different fields of application. However, its potential for stationary and non-stationary flood frequency analysis is limited by the distribution types included and the estimation method used. The GAMLSS family of distributions does not include three-parameter distributions with location parameter as the lower bound, the most frequently used in FFA. Only one distribution RGE, re-parametrized Weibull 3 parameter distribution, belongs to the GAMLSS family. Although the software is open to the inclusion of new distribution types, the estimation algorithms can fail. We were able to add only Pearson type 3 distribution to the GAMLSS family. The attempts with other distributions were unsuccessful. We presume that it is due to the maximum likelihood estimation method which is not the most suitable for this kind of distributions.

It is difficult to apply the GAMLSS package. The user should understand distributions and their properties and should make decisions regarding the distribution of the response variable, the choice of explanatory variables and the link functions, the amount of smoothing and random effects. The application of GAMLSS results in estimates of time-varying quantiles which are strongly distribution dependent, so the selection of the suitable distribution is of the first importance.

Looking at the many hydraulic structures built at a time when risk assessment methods were not as complex as they are today and hydrological data were sparse and irregular, we admire the knowledge of water and the environment, engineering skills and common sense of their creators. Perhaps the loss of this sense is the price we pay for technological development.