Introduction

Geotechnical data obtained by field and laboratory investigations is often both complex and incoherent. This can be seen, for example, when trying to find correlations and transformation models between various parameters. Different researchers suggest different correlations where the key regressor varies. Many times, none of them can give very accurate description, while the original data seems rather scattered. Part of the problem is the variation and uncertainty related to geotechnical properties. This is often divided into an inherent (natural) variability and an epistemic uncertainty. The inherent variability represents the natural variation of the property. The epistemic uncertainty is more complex. It includes statistical uncertainty, measurement error, and transformation model uncertainty. It is then difficult to unveil the true complexity of geotechnical data.

The use of machine learning (ML) has engaged popularity in many fields of application, also in geotechnical engineering (Zhang and Phoon 2022, Zhang et al. 2022). It is tempting to develop ML-based models that describe complicated geotechnical phenomena. The learning ability of ML models makes it an attractive tool to reveal the hidden connections for examples between various soil parameters. However, if the data is incoherent and there exists a lot of variation due to errors in measurements or some other reasons, this becomes a great challenge to the learning process. This might well cause overfitting of the ML model, and even though the statistical parameters might indicate improved correlation, for some combinations the model might give predictions that are completely out of plausible range. Such models might thus give an improved solution for a vast majority of cases, but for the other ones, the result might be quite misleading.

Transformation models for estimating undrained shear strength have been of particular interest in geotechnical engineering for long. Traditional transformation models can be roughly divided into two types, as shown by Eqs. (1) and (2).

$${S}_{u}=\alpha {{\sigma }^{^{\prime}}}_{p}$$
(1)
$$\frac{{S}_{u}}{{{\sigma }^{^{\prime}}}_{v}}=S{\mathrm{OCR}}^{m}$$
(2)

where \(\alpha\), \(S\), and \(m\) are either constants or functions of some soil properties. Their values depend on the material and the test type. For example, Skempton  (1954) suggested a category (1) equation for the field vane shear strength suFV of normally consolidated clays, where the factor \(\alpha\) depends on the plasticity index (PI). For Scandinavian clays, Hansbo (Hansbo 1957) suggested the use of liquid limit (LL) instead of PI. Later, Chandler (Chandler 1988) extended Skempton’s and Larsson (Larsson 1980) Hansbo’s ideas to overconsolidated (OC) clays. Equation (2) was first suggested by Ladd and Foot (Ladd and Foot 1974) and is often referred as the SHANSEP approach.

It has been recognized for long, that the suFV should be corrected with respect to plasticity to account for anisotropy and rate effects (Bjerrum 1972; Taylor 2001). This corrected value is often referred to as su(mob). It is often assumed that the su(mob) value is approximately equal to the undrained shear strength determined by the direct simple shear (DSS) tests, i.e., the suDSS value. However, it should be acknowledged that rate effects are present in the DSS as well.

Mesri (Mesri 1975) suggested the use of Eq. (1) for su(mob) using a constant value of \(\alpha = 0.22\). Jamiolkowski et al. (Jamiolkowski et al. 1985) suggested category (2) equation with \(S = 0.23\pm 0.04\) and \(m = 0.8\). However, for suDSS, Larsson et al. (Larsson et al. 2007) suggested that \(S\) depends linearly on the LL, while Karlsrud and Hernandez-Martinez (Karlsrud and Hernandez-Martinez 2013) proposed the use of the water content (\(w\)). Ching and Phoon (Ching and Phoon 2012) also suggested the use of category (2) equation but proposed that \(S\) would be dependent on LI, sensitivity (St), or both. D’Ignazio et al. (D’Ignazio et al. 2016) concluded that while the suFV correlates with the index properties (PI, LL, \(w\), and LI), the su(mob) values seem to be unaffected by them.

Based on the above, it seems that while majority of moe that equations of category (re recent research seems to prefer the use of category (2) equations, it is unclear whether index properties should be included when predicting the design value su(mob). It should also be acknowledged that there is now evidence that equations of category (1) or (2) would be the best one to correlate the preconsolidation pressure and in case of (2) also the effective vertical stress, to the undrained shear strength. Machine learning (ML) offers a valuable tool to test the dependencies between undrained shear strength and other soil properties. Recently, Zhang et al. (Zhang and Phoon 2022) studied the use of extreme gradient Boosting (XGBoost) and random forest (RF) learning methods to study these relationships for suFV for a database consisting of Scandinavian clays. In short, they obtained the best performance with the XGBoost and RF models in comparison to three other ML models and two transformation models of category (2) where the value of \(S\) was made dependent on either \(w\) or LL. They also concluded that the preconsolidation pressure is the most dominant factor, with a relative importance of 54.3% for their XGBoost model, while the value was 20.3% for effective vertical stress, 12.5% for PL, 6.8% for \(w\), and 6.1% for LL. Based on a fivefold cross-validation of the ML models, they concluded that the models are influenced by the data quality.

While the study of Zhang et al. (Zhang and Phoon 2022) gives some important insights on the possibilities of ML models and the importance of various soil properties, it remains unclear which properties to include in predicting the su(mob) used in design. In addition, the differences of various proposed models could be partly explained by the quality of data used. This paper attempts to bring more clarification to these questions. The scope of this paper is thus not on finding the best transformation model for engineering purposes but rather to (1) study which combination of soil properties gives the best estimate of su(mob), (2) study what is the influence of dataset quality on the models, and (3) study and exemplify the risks of overfitting. It is believed that ML models provide a good tool for such study providing much more freedom in the intervariable correlations than normally achieved by traditional transformation models. Artificial neural networks (ANN) will be used as the main ML method while multivariate adaptive regression spline (MARS) and genetic programming (GP) will be used as comparative ML methods and category (1) and (2) equations will be used for comparisons.

Datasets used in study

Two different datasets are used in the study. Ching and Phoon (Ching and Phoon 2014) proposed labeling a multivariate database as “soil type”/ “number of parameters of interest”/ “number of data points.” The first dataset taken from D’Ignazio et al. (D’Ignazio et al. 2016) and labelled as F-CLAY/7/216 contains thus 216 data points each containing 7 parameters for Finnish clays. The dataset can be found on the webpage of ISSMGE technical committee 304 at http://140.112.12.21/issmge/tc304.htm. However, the preconsolidation pressures used in this study are all normalized with respect to strain rate to correspond to a typical strain rate applied in CRS oedometer tests (1 × 10−6 s−1–4 × 10−6 s−1), using the same first-order correction as D’Ignazio et al. (D’Ignazio et al. 2016). The clay parameters in the dataset consist of undrained shear strength determined by the field vane (suFV), in situ effective vertical stress (σ′v), preconsolidation pressure (σ′p), liquid limit (LL), plastic limit (PL), water content (\(w\)), and sensitivity (St). In addition, the location and the depth of the data points are known. In this study, the focus is on the design value of undrained shear strength, i.e., on su(mob). The correction of suFV is done according to Finnish guidelines (Ratahallintokeskus 2005) following the ideas of Bjerrum (Bjerrum 1972) and Helenelund (Helenelund 1977) applying a correction factor dependent of the liquid limit. The basic statistics, including mean, median, minimum, and maximum values, and coefficient of variation (COV) of the datasets are summarized in Table 1. In addition to the parameters listed above, also the statistics for calculated over consolidation ratio (OCR = σ′p/σ′v) and the ratio su(mob)/σ′p is provided.

Table 1 Basic statistic for the two datasets

The second dataset labelled as F-CLAY/7/135 is a subset of the previous dataset. Therein an attempt is made, to remove outliers. It is well known that the undrained shear strength is closely related to the preconsolidation pressure. The changes of these parameters with respect to depth should thus be consistent. As already discussed, Mesri (Mesri 1975) suggested the ratio su(mob)/σ′p to be 0.22 for clays in general. There should thus not be too high variation of the ratio within an otherwise homogenous clay layer, even if the expected value of this ratio is not restricted. Any clear deviations found in F-CLAY/7/216 with respect to the above will be considered as outliers and removed. Some sites also include data from the dry crust layers. As this data is very limited, it is not considered in the second dataset. The criteria described above includes the use of engineering judgement and involves thus subjectivity. With respect to statistical criteria removing data outside the limits su(mob)/σ′p  = mean ± 2COV, corresponding to a 95% confidence interval for a normally distributed data, the criteria applied removed all except one data point on the high side. The aim herein is not to produce a “perfect” dataset, but rather to study the influence of data coherence on the performance of various models. The basic statistics for the second datasets are summarized in Table 1. The performance of the criteria is further discussed in the “Discussion” section.

As can be seen from Table 1, the removal of outliers has not considerably narrowed the range of values for the variables, comparing datasets F-CLAY/7/216 and F-CLAY/7/135. For the OCR, the maximum value has reduced to almost half, but the mean and median values are close for the two datasets. With respect to data coherence for the present study, the most interesting variable is the ratio between the undrained shear strength and the preconsolidation pressure. As discussed above, this is sometimes considered a constant. In Fig. 1, the statistics for the ratio su(mob)/σ′p is illustrated for the two datasets.

Fig. 1
figure 1

Ratio between undrained shear strength and preconsolidation pressure for the two datasets

As can be seen, the dataset F-CLAY/7/135 is much more coherent with respect to the ratio su(mob)/σ′p than F-CLAY/7/216. It is believed that the development of ML-based transformation models in the next sections will reveal if true outliers where removed, or if the data F-CLAY/7/216 will better show hidden connections between the variables.

Transformation models and ML methods used in study

Statistical parameters used in the development and evaluation of the models

In the development and evaluation of the different models, the following statistical parameters are utilized: bias factor (\(b\)), coefficient of variation (COV), standard deviation \(\sigma\) (STD), the Pearson correlation coefficient (\(R\)) or its squared value (R2), and the mean square error (MSE).

The bias \(b\) is the mean of the ratio between the actual target value (\(x\)) and the predicted value (\(y\)), i.e.,

$$b=\frac{1}{n}\sum_{i=1}^{n}\frac{{x}_{i}}{{y}_{i}}$$
(3)

So, if the bias \(b = 1\), the model provides on average an unbiased prediction.

The amount of uncertainty a parameter/property has can be described with the coefficient of variation (COV), which is a dimensionless ratio between the standard deviation \(\sigma\) and the mean value \(\mu\) of the property, i.e.,

$$\mathrm{COV}=\frac{\sigma }{\mu }$$
(4)

It shows thus the extent of variability in relation to the mean of the population, the closer the value of COV is to zero, the smaller is the scatter of the data, with COV = 0 indicating that the prediction is deterministic.

The standard deviation \(\sigma\) of the entire population of variable (\(x\)) is given by

$${\sigma }_{x}=\sqrt{\frac{\sum_{i=1}^{n}{\left({x}_{i}-{\mu }_{x}\right)}^{2}}{n}}$$
(5)

The Pearson correlation coefficient (\(R\)) is a measure of the linear correlation between two variables. It is defined as the covariance of the two variables (\(x, y\)) divided by the product of their standard deviations, i.e.,

$$R=\frac{\mathrm{COV}\left(x,y\right)}{{\sigma }_{x}{\sigma }_{y}}$$
(6)

\(R\) has values in between + 1 and \(-1\), where + 1 indicates total positive linear correlation, 0 no linear correlation, and − 1 total negative linear correlation.

MSE is a risk function that measures the average squared difference between the predicted (\(y\)) and the values that are estimated (\(x\)). It can be calculated as

$$\mathrm{MSE}=\frac{1}{n}\sum_{i=1}^{n}{\left({x}_{i}-{y}_{i}\right)}^{2}$$
(7)

MSE obtains positive values, the closer to zero the better is the prediction.

Traditional transformation models

As discussed in “Datasets used in study,” the traditional transformation models will be based on categories (1) and (2) types of equations. Equation (1) will be utilized in two different ways. Firstly, \(\alpha\) will be fitted to give the best average fit with bias equal to one, and the performance of this model is then evaluated by its COV value. Secondly, \(\alpha\) will be evaluated to give perfect match for each data point to study if the resulting \(\alpha\) correlates to the classification data. Equation (2) will be firstly evaluated by using the values \(S = 0.244\) and \(m = 0.763\) proposed by D’Ignazio et al. (D’Ignazio et al. 2016). Secondly, specific correlations for \(S\) and \(m\) will be studied.

ANN-based transformation models

General

ANN, as a form of artificial intelligence, imitates the behavior of the brain and nervous system (Shahin et al. 2001). The brain is a highly complex, nonlinear, and parallel information-processing system. To perform the necessary computations, like pattern recognition, perception, and body control, it organizes some constituents named as neurons.

A neuron is a fundamental information-processing unit in the operation of a neural network (Haykin 2005). Figure 2 shows a neuron model and the architecture of the ANN model used in this study.

Fig. 2
figure 2

Architecture of a typical 3-layer feedforward network of ANN (Haykin 2005) for simulation of corrected field vane strength

In the mathematical modeling, the three considered basic elements of the neural model are as below:

  1. 1.

    A set of synapses or connecting links which are modeled by a parameter considered as weight. An input signal xj at the input of synapse j connected to neuron k is multiplied by the synaptic weight wkj.

  2. 2.

    A function for summing the weighted input signals for each neuron.

  3. 3.

    An activation function for limiting the amplitude of the output of a neuron. Different activation functions might be utilized, such as sigmoid which is used in this study, and the normalized amplitude of the output of a neuron often ranges in the closed interval \([\mathrm{0,1}]\) or \([-\mathrm{1,1}]\).

An externally applied bias, denoted by bk, is also included in the neural model. The bias increases or decreases the input of the activation function, depending on whether it is positive or negative, respectively.

In other words, a neuron \({\varvec{k}}\) is modeled mathematically by the following equations:

$${u}_{k}=\sum_{j=1}^{m}{w}_{kj}{x}_{j}$$
(8)

and

$${y}_{k}=\varphi \left({u}_{k}+{b}_{k}\right)$$
(9)

where x1, x2, , xm are the input signals; wk1, wk2, , wkm are the synaptic weights of neuron k; uk is the linear summation function; bk is the bias; φ is the activation function; and yk is the output signal of the neuron. The use of bias bk has the effect of applying an affine transformation to the output uk of the linear combiner in the model of Fig. 2 (Haykin 2005).

Input–output mapping

A popular paradigm of learning called learning with a teacher or supervised learning involves modification of the synaptic weights (and biases) of a neural network by applying a set of training samples. Each example set of samples consists of a unique input signal and a corresponding estimated response (output signal). The network training is performed using some examples picked at random (or selected by user) from the samples sets, and the synaptic weights of the network are modified to minimize the difference between the estimated and the actual values of the desired output parameter(s). The difference between the estimated and actual outputs is computed by means of an appropriate statistical error criterion such as MSE. The training of the network is repeated for many examples in the input–output datasets until the network reaches a steady state, where there are no further significant changes in the synaptic weights.

Multilayer feedforward networks

The arrangement of the neurons of a neural network is intimately linked with the learning algorithm used to train the network and should be structured. As a commonly utilized architecture which is employed in this study also, the multilayer feedforward network, known also as Multilayer Perceptron (MLP), consists of one or more hidden layers, whose computation nodes are called hidden neurons or hidden units. The hidden layer(s)/neurons intervene the network input and output layers.

The nodes in the input layer (the input vector) constitute the input signals applied to the neurons (computation nodes) in the second layer (i.e., the first hidden layer). The output signals of the second layer are used as inputs to the third layer and so on for the rest of the network. The set of output signals of the neurons in the output (final) layer of the network constitutes the response of the network to the source nodes in the input (first) layer. The three-layer feedforward ANN architecture used in this study is shown in Fig. 2 (Haykin 2005).

As the learning algorithm, back propagation approach was used so that the variables of the model, i.e., weights and biases, would be optimized. In this algorithm, MSE is propagated back from the output to the connection weights and updates the weights to minimize the prediction error (Foroozesh et al. 2014). The Levenberg–Marquardt (LM) method was used to minimize the error criterion because it is fast, accurate, and reliable (Adamowski and Chan 2011; Adamowski and Karapataki 2010). The LM algorithm is a modification of the classic Newton algorithm for finding an optimum solution to a minimization problem (Adamowski and Karapataki 2010; Daliakopoulos et al. 2005).

The mathematical computations of the ANN modeling are implemented by means of the MATLAB ANN toolbox. The primary objective of the ANN modeling has been to perform the input–output mapping, i.e., to capture the relationships/correlations between the input and output variables with the least error criterion.

GP-based transformation models

Genetic programming (GP) is developed based on the concept of genetic algorithm (Koza 1992). In GP, transformation models are encoded as a set of genes that are then modified using an evolutionary algorithm. Firstly, many equations are created based on the input variables. The fitness of each equation is then determined in the second step. Thereafter, the best equation is selected. New equations are created through the procedure of reproduction, crossover, and mutation. In Fig. 3, the flowchart for the GP procedure is presented.

Fig. 3
figure 3

Flowchart for simple genetic programming (GP) model

MARS-based transformation models

Multivariate adaptive regression splines (MARS) is developed based on the concept of nonparametric regression (Friedman 1991) and can be seen as an extension of linear models that automatically model nonlinearities and interactions between variables. It adopts the following relation between input (\(x\)) and output (\(y\)).

$$y={c}_{0}+\sum_{m=1}^{M}{c}_{m}{B}_{m}\left(x\right)$$
(10)

where c0 is constant, Bm(x) is \(m\)-th basis function, cm is the coefficient of \(m\)-th basis function, and \(M\) is the number of basis functions.

The expression of basis function is given below (Sekulic and Kowalski 1992):

$$\begin{array}{cc}{b}_{q}^{-}\left(x-t\right)={\left[-\left(x-t\right)\right]}_{+}^{q}=\left\{\begin{array}{c}{\left(t-x\right)}^{q}\\ 0\end{array}\right.& if x>t\end{array}$$
(11)

Otherwise,

$${b}_{q}^{+}\left(x-t\right)={\left[+\left(x-t\right)\right]}_{+}^{q}=\left\{\begin{array}{c}{\left(x-t\right)}^{q}\\ 0\end{array}\right.$$
(12)

where \(q\) is power, t is the knot location, and \({b}_{q}^{-}\left(x-t\right)\) and \({b}_{q}^{+}\left(x-t\right)\) are the spline functions. The development of MARS model involves two steps, namely, a forward and a backward step as follows.

Forward step: this step is used to introduce basis function in Eq. (10). MARS can suffer overtraining due to large number of basis functions.

Backward step: for preventing overfitting, extra basis functions have been deleted from Eq. (10). Generalized cross-validation (GCV) value is used to delete the extra basis functions. The expression of GCV is given below:

$$\mathrm{GCV}=\frac{\frac{1}{N}\sum_{i=1}^{N}{\left[{y}_{i}-f\left({x}_{i}\right)\right]}^{2}}{{\left[1-\frac{C\left(B\right)}{N}\right]}^{2}}$$
(13)

where \(N\) is the number of data and \(C(B)\) is a complexity penalty that increases with the number of BF in the model and which is defined as

$$C\left(B\right)=\left(B+1\right)+dB$$
(14)

where d is a penalty for each basis function (BF) included into the model. The details about d are given by Friedman (Friedman 1991).

The flowchart for MARS is given in Fig. 4.

Fig. 4
figure 4

Flowchart for the MARS model

Performance of the models

Performance of traditional models

Firstly, the performance of Eq. (1) was evaluated by fitting the value of \(\alpha\) to give an unbiased prediction of the undrained shear strength. The obtained values for \(\alpha\) and corresponding b between the measured and the predicted COV for the ratio between measured and predicted and R2 and MSE between the measured and the predicted mobilized undrained shear strength values are presented in Table 2.

Table 2 Performance of various traditional transformation models, \(b\) between the measured and the predicted, COV for the ratio between measured and predicted, and R2 and MSE between the measured and the predicted mobilized undrained shear strength values

In the second round, the value of \(\alpha\) was fitted to each data point to give an exact fit. The obtained \(\alpha\) values did not show any notable correlation with most of the soil data. For the dataset F-CLAY/7/216, no correlation was found. For the dataset F-CLAY/7/135, some correlation was found for the OCR. The best fit was found using Eq. (15).

$$\alpha ={n}_{1}{OCR}^{{n}_{2}}$$
(15)

The obtained values for n1 and n2 are 0.2444 and − 0.247, respectively, giving a COV = 0.446 and R2 = 0.23.

Introducing Eq. (15) into Eq. (1), gives the following modified version of Eq. (1):

$${S}_{u}={n}_{1}{{\sigma }^{i}}_{p}{\mathrm{OCR}}^{{n}_{2}}$$
(16)

It is worth noticing that this equation is the same equation as (2) and can be rewritten into

$${S}_{u}={n}_{1}{{\sigma }^{^{\prime}}}_{v}{\mathrm{OCR}}^{{1+n}_{2}}$$
(2b)

The performance of Eq. (2b) with the obtained n1 and n2 values and corresponding b, COV, R2, and MSE values is presented in Table 2 for dataset F-CLAY/7/135.

The performance of Eq. (2) was first evaluated by using the values \(S = 0.244\) and \(m = 0.763\) obtained by D’Ignazio et al. (D’Ignazio et al. 2016). As can be expected, these values give a very good performance for dataset F-CLAY/7/135, as the data is almost the same and Eq. (2)_135 is almost identical to the previous Eq. (2b)_135; see Table 2. On the other hand, for dataset F-CLAY/7/216, the performance was much poorer. It is interesting to note that for this dataset the inclusion of OCR did not improve the prediction at all. It was also observed that if the preconsolidation pressures of dataset F-CLAY/7/216 had not been normalized, the prediction of Eq. (2) would have been poorer than the one from Eq. (1).

The performance of Eq. (2) was further evaluated by fitting both the value of \(S\) and \(m\) for each data point to yield the perfect match. However, the obtained \(S\) and \(m\) values did not show any correlation to any of the data. A special effort was tried out to correlate \(S\) to the plasticity index. The reason behind is that for the active undrained shear strength the value of \(S\) is likely to increase with the friction angle, as the yield surface gets more inclined. There is no data of the friction angles, but generally, the friction angle is found to correlate with the plasticity index. However, even if the value of \(S\) for active undrained strength might increase with the friction angle due to higher inclination of the yield surface, for the same reason, the \(S\) value for the passive undrained shear strength should decrease. The field vane value, representing more the average of active, direct shear and passive, is thus likely to be uninfluenced by the friction angle.

As any attempt to further improve the performance of Eq. (2) did not yield any notable improvement, an unbiased parameter set was finally introduced as set (2)* in Table 2. The COV values are shown with 4 digits just to show the infinite small differences between models (2b)_135, (2)_135, and (2)*_135. In practice, their performance is equally good. It is worth noting that for dataset F-CLAY/7/216, the best fit of Eqs. (1) and (2)* gives practically the same result. Their similarities are also shown in the value of parameter \(m\) for (2)*_216 being close to one. The predictions of Eqs. (1) and (2)* are shown in Fig. 5 together with a range of ± 2 COV.

Fig. 5
figure 5

Performance of Eq. (1) for dataset F-CLAY/7/216 (a) and F-CLAY/7/135 (c) and Eq. (2) for F-CLAY/7/216 (b) and F-CLAY/7/135 (d). The dotted lines represent the range of predictions corresponding to ± 2 COV

Performance of ANN models

Based on the available databases, different scenarios of input–output mappings were performed, and their parameter estimation precisions were calculated. In all scenarios, the mobilized vane shear strength was considered as the output, but the input parameters, in situ vertical effective stress (σ'v), preconsolidation pressure (σ'p), overconsolidation ratio (OCR), liquid limit (LL), plastic limit (PL), water content (\(w\)), depth, and sensitivity (St) were arranged in different scenarios such that the estimation accuracy of the ANN models and GP, MARS, and some traditional empirical equations could be compared. As another target of organizing different scenarios, a sensitivity analysis was performed among the input variables to identify the least number of soil properties that might lead to the best estimation of the modified vane shear strength of soil. Altogether 16 different scenarios were verified. The considered input parameter(s) in each scenario and the statistics of their performance are presented in Table 3.

Table 3 Results of ANN analysis with various trials for databases of F-CLAY/7/216 and F-CLAY/7/135

Based on the observations from the previous studies, regarding the complexity of the models, usually 3 to 10 neurons lead to the best results (Foroozesh et al. 2014; Karaagac et al. 2009; Khosrojerdi et al. 2016). So, in this study, 7 neurons were considered in all ANN models so that the results of different models might be compared. For the training and verification, the data was divided into training (70%), validation (15%), and testing (15%) datasets. The same 70% training dataset was later used for training of the GP and MARS models, while testing of them was done for the rest of the data (30%). The validation dataset is used for the generalization of the ANN model in the training process. However, the validation and test dataset error criteria of the ANN model are compared with the test dataset for the GP and MARS models, mainly because of limited amount of data, especially for F-CLAY/7/135. Traditionally, a single training is performed for the ANN models, accepting the results as such. However, due to random initial guess of the optimization model in training of the ANN, repeating training with the same data will yield different results. To study more this aspect, training was done 1000 times for each trial. The differences between the trials will be discussed later. For Table 3, the training sets giving the lowest weighted sum for MSEtrain, MSEvalidation, and MSEtest, with equal weights for each, were chosen. In Fig. 6, the COV, R2, and MSE values are presented for the different trials for the training and test data.

Fig. 6
figure 6

Performance of ANN trials considering training data (a, c, e) (left) and test data (b, d, f) (right) for databases of F-CLAY/7/216 and F-CLAY/7/135

Studying the data in Table 3 and Fig. 6, several interesting conclusions can be made. Trial 2 seems to give a relatively good prediction using only the preconsolidation pressure as input. All statistical indicators for trial 2 are better than for the traditional transformation model using Eq. (1), indicating that the relationship is not linear. Adding information about the vertical effective stress, i.e., comparing trials 3, 4, and 5 to trial 2, improves the overall performance for both data. However, for the more scattered data F-CLAY/7/216, no improvement is found for the testing data. The different setups using preconsolidation stress and effective vertical stress, i.e., trials 3, 4, and 5, give in practice equal performance. It should though be noted that the results were received from different ANN training sets and relying on a single training would have indicated differences in between the trials. For trials 3, 4, and 5, the statistical measures are again much better than for the traditional transformation model, indicating that the relationship is not as given by Eq. (2).

Adding information about plasticity, i.e., trials 6, 7, and 8, clearly improves the performance of the training dataset for F-CLAY/7/216. However, for the testing data, the performance drops, except for trial 6, which shows a slight improvement. For the more coherent dataset F-CLAY/7/135, no improvement can be found. However, also for this dataset, some of the training sets gave a clear improvement for the training sets of trials 6, 7, and 8, but then, the testing data gave very poor results. Adding information about water content and depth to previous (trials 9 and 10) does not improve the performance.

Comparing trials 4, 11, 12, 14, and 16 reveals that adding sensitivity data, i.e., trial 11, does not give any improvement. However, the inclusion of water content in trial 12 improves the performance for dataset F-CLAY/7/216, while no improvement is found for F-CLAY/7/135. The inclusion of information about depth (trial 14) improves the performance of training and validation, while the testing performance gets worse. The inclusion of both water content and depth information (trial 16) improves the overall performance of both datasets and seems to give the best performance of all trials. However, in practice, the improvement to trial 4 is rather insignificant. It is also interesting to note that the inclusion of information about plasticity, sensitivity, water content, and depth (trial 15) gives the best performance of the training set for F-CLAY/7/216, while the performance of testing set is not so good.

The performance of models can be visualized by Taylor diagrams (Taylor 2001). They are often used to compare climate models but can well be used to compare how any model represents reality. In the Taylor diagram, three statistical parameters are used and plotted in the same diagram, namely, the Pearson correlation coefficient (\(R\)), the centered root-mean-square error (RMSE, i.e., the root of MSE), and the standard deviation (STD). Because of their mathematical relation, these parameters can be plotted together in a quadrant in the following way. The observation, i.e., the parameter that is modelled, in our case sumob, is plotted on the horizontal axis, at a distance of STD from the origin. The horizontal axis represents \(R = 1\), while other \(R\) values are related to the angle to the horizontal axis, i.e., the azimuthal angle. The model points are plotted at a distance of their STD from the origin on a line which angle corresponds to the \(R\) value. Then, the centered RMSE value will be related to the distance of this point to the observation. As a centered RMSE value is used, i.e., the means are subtracted out before computing their values, the diagram does not provide information about overall biases, but solely characterizes the centered pattern error (Taylor 2001).

As the actual value of STD is not important when comparing the models, a normalized plot is often used. Therein, both RMSE and STD values are normalized with the STD of the observation. Then, the observation point is at unit distance from the origin. Herein, such a normalized plot is used to compare different datasets.

In Fig. 7, the Taylor diagram representation is given for trials 4, 15, and 16 for both datasets F-CLAY/7/216 and F-CLAY/7/135. In addition, trial 15 is given for an ANN training set, giving the best performance for the training data out of the 1000 training sets. In all cases, the performance for training, validation, and testing is given separately, using white fill for training, grey for validation, and black for testing. Trials 4 and 16 give a relatively consistent performance, while the points for training, validation, and testing are relatively closely located, trial 16 showing the best performance. Trial 15 is giving a good performance for the training data, but it declines more for validation and training data, indicating overtraining. The implications of overtraining are clearly shown for the case trial 15 max train, showing the best performance for training data, but worst performance for testing data.

Fig. 7
figure 7

Taylor diagrams for trials 4, 15, and 16 for databases F-CLAY/7/216 (a) and F-CLAY/7/135 (b). White, grey, and black fills show training, validation, and testing, respectively

The differences between trials 4 and 16 are further studied by Fig. 8 comparing the predicted data to the observed. Again, white fill is used for training, grey for validation, and black for testing. The dotted lines in both trials 4 and 16 represent ± 2COVAll_trial 4.

Fig. 8
figure 8

Performance of ANN trials 4 (a) and 16 (b) for F-CLAY/7/135 and trials 4 (c) and 16 (d) for F-CLAY/7/216. In addition to predicted vs. measured data, a 1:1 line and the lines for ± 2 COV are given

As can be seen also from this presentation, the difference between these two trials is minimal. However, it can though be seen that including data on water content and depth slightly improves the performance especially for smaller values. Generally, the value of depth is not related to the strength, as we may have highly overconsolidated clays at shallow depths with high undrained shear strength and soft normally consolidated clays at deeper depths. However, as the dataset consisted of clays with generally low OCR, it is logical that the depth correlates with the strength. As information of water content and depth is available if oedometer data is used, adding them will not require any extra effort. It would though not be advisable to use such model for highly overconsolidated clays, predicting, e.g., the strength of dry crust. For practical engineering, the difference between trials 4 and 16 is insignificant.

Performance of GP models

Various combinations of input parameters were also tried for the GP models, but not as extensively as for the ANN. Interestingly, for F-CLAY/7/216, the best performances were achieved by including data of σ'v, σ'p, and OCR, while for F-CLAY/7/135 including LL improved slightly the performance of the testing set, while the performance of training was slightly declined. It is good to note that also for GP the outcome will vary for repetitive trainings and a similar 1000 training trials as for ANN were not tried out. As no clear improvement was found for additional parameters, models including data of σ'v, σ'p, and OCR are used for comparisons.

The performance of the GP models for F-CLAY/7/216 and F-CLAY/7/135 using σ'v, σ'p, and OCR as input is given in Fig. 9, as the ratio of measured vs. predicted undrained shear strength including lines indicating ± 2 COV. The statistical performance of the models evaluated by their bias, COV, R2, and MSE is presented in Table 4.

Fig. 9
figure 9

Performance of GP model for datasets F-CLAY/7/216 (a) and F-CLAY/7/135 (b). In addition to predicted vs. measured data, a 1:1 line and the lines for ± 2 COV are given

Table 4 Performance of various AI transformation models

Performance of MARS models

Similarly, as for GP models, different combinations of input parameters were also tried for the MARS models, but not as extensively as for the ANN. Similarly, for MARS models, the best performance was not achieved by including all parameters. In general, it can be reported that the most consistent performance was achieved with a combination including data of σ'v, σ'p, and OCR. For F-CLAY/7/216, the inclusion of LL improved the performance of training set, while for the training set, COV and \(R\) values were slightly worse and MSE was slightly better. The differences though were minor. For F-CLAY/7/135, the inclusion of LL and PL data improved the performance of the training set, while the performance of testing set and overall performance dropped. Again, the differences are rather small. Unlike for ANN and GP models, MARS yields the same results on repeated trainings.

The performance of the final MARS models for F-CLAY/7/216 and F-CLAY/7/135 using σ'v, σ'p, and OCR as input is given in Fig. 10 as the ratio of measured to predicted undrained shear strength including lines indicating ± 2 COV. The statistical performance of the models evaluated by their bias, COV, R2, and MSE is presented in Table 4.

Fig. 10
figure 10

Performance of MARS model for datasets F-CLAY/7/216 (a) and F-CLAY/7/135 (b). In addition to predicted vs. measured data, a 1:1 line and the lines for ± 2 COV for training set are given

Discussion

In the development of the ML models, the data was divided into a training (70%) and test (30%) data for MARS and GP methods and to training (70%), validation (15%), and test (15%) data for the ANN. As the division to the different sets was done randomly, it might result in that all relevant data is not covered in the training set. In Table 5, the mean, min, and max values of the training datasets are given with respect to their values for the complete datasets. A min value of 1.24 means thus that for training the smallest value was 1.24 times larger than the minimum value of the whole dataset. As can be seen from Table 5, the training dataset for F-CLAY/7/216 covered almost the entire range of the various parameters with a mean very close to the whole dataset. For dataset F-CLAY/7/135, the training data did not cover as well the range of values for all data. Still, the models performed much better than for F-CLAY/7/216 dataset.

Table 5 Range of values used for training of the models given as training dataset value/all data value, where value corresponds to the mean, min, and max values

As an overview of the performance of different types of models, the Taylor diagram plots for some selected transformation models are presented in Fig. 11 for both datasets F-CLAY/7/216 and F-CLAY/7/135, indicating the overall performance of the models. As can clearly be seen also from this presentation, all models perform better for database F-CLAY/7/135. Based on this, two obvious conclusions can be made. Firstly, the removal of data from F-CLAY/7/216 to obtain F-CLAY/7/135 consisted at least of the major part of true outliers. Secondly, incoherent data deteriorated the performance of all models evaluated. A further conclusion from Fig. 11 is that the ML models perform better than the traditional equations. Generally, the ANN models gave the best performance, although the difference to GP and MARS models was small. It should also be remembered that for the ANN, 1000 training sets were performed while only one was done for the GP model, MARS lacking the influence of different training sets.

Fig. 11
figure 11

Taylor diagrams for modelling of sumob for the various transformation models for datasets F-CLAY/7/216 (a) and F-CLAY/7/135 (b)

By studying the performance of different ANN trials in testing, it can be noted that adding information generally tends to improve the performance in the training dataset. Remembering that due to the initial guess of the optimization model there is some randomness in a single ANN training and that in the chosen training sets the overall performance was optimized, it is understood that the trend could be more apparent. The important conclusion from this is that if one uses all available data to train a model, like often done for traditional transformation models, the predictions may improve by adding variables. However, these may still be false predictors causing overfitting of the data. Then, the model may fail in predicting any additional data, resulting in even dangerous outliers.

It is obvious that σ'p is by far the most dominant factor in predicting su. The inclusion of σ'v also improves the estimates. Both conclusions are logical with respect to the yielding of clays, i.e., the relationship between the yield surface of clays, in situ stress, and failure. However, it should be noted that for the more scattered dataset F-CLAY/7/216 the inclusion of the effective vertical stress did not improve the performance of the testing data. The situation is more unclear with respect to plasticity properties. While there was clear indication of overtraining when additional parameters were included in the model, it can be argued that including LL could slightly improve the performance of some models. It is good to note that the correction made for suFV to derive su(mob) is based on LL. There is no reason to believe that this correction would be perfect, and some influence of LL to su(mob) is thus likely.

Comparing the results for the two datasets F-CLAY/7/216 and F-CLAY/7/135, it was observed that for none of the methods and different sets of parameters, the dataset F-CLAY/7/216 led to a better transformation model and revealed some connections. A logical conclusion then is that the partially subjective cleaning of data for dataset F-CLAY/7/135 did not remove any potentially important data, but true outliers were removed from dataset F-CLAY/7/216. Acknowledging that the dataset F-CLAY/7/216 is more scattered, it can be further concluded that the risk for overtraining increases for more scattered data. Thus, if the data is scattered, simpler models should be preferred, as also indicated by the relatively good performance of using only σ'p as input for F-CLAY/7/216 dataset, in comparison to the more extended models. This supports the discussions of results by D’Ignazio et al. (D’Ignazio et al. 2016), Mesri and Wang (Mesri and Wang 2017), and D’Ignazio et al. (D’Ignazio et al. 2017).

The comparison of the performance of traditional and ML-based transformation models is in that sense problematic, that while the whole datasets were used for the traditional models, only 70% was used to train the ML models with additional 15% for the validation of the ANN models. Still, it is obvious that the ML models give a clearly better performance when comparing the results presented in Tables 2 and 4. For engineering purposes, it is though often more practical to apply traditional transformation models, of which type (2) gives a very fair prediction for soft sensitive clays. For ML models, ANN can be criticized of being a black box while no equation is provided. GP and MARS models provide an equation, but the equations are often rather complex and sensitive, so one needs to be careful in their implementation. At the end, it is much up to the problem in hand if ML models give any additional benefits in predicting su. For big database applications, ML offers intriguing possibilities. The authors also strongly believe that ML is a very useful tool to provide new insights over various dependencies and, as in the present study, give explanation to some of the present discrepancies researchers may have on correlations between geotechnical properties.

Conclusions

In this study, transformation models for sumob have been developed based on traditional correlation models as well as ML models. Two different datasets have been used, F-CLAY/7/216 being the larger but more incoherent one while some data has been cleaned from its subdataset F-CLAY/7/135. Based on the study, the following conclusions can be made:

  • Based on all models, the cleaning of the F-CLAY/7/216 data into dataset F-CLAY/7/135 removed true outliers.

  • As expected, preconsolidation pressure is the most important variable in predicting the mobilized undrained shear strength.

  • For both traditional and ML models, the inclusion of effective vertical stress/OCR in addition to the preconsolidation pressure did not improve the prediction for the more scattered dataset F-CLAY/7/216.

  • For the more coherent dataset F-CLAY/7/135, the inclusion of effective vertical stress and/or OCR improved the prediction notably for both traditional and ML models.

  • The inclusion of depth and \(w\) improved slightly the performance of the ANN models, especially for lower su values. This can though well be partly due to the used datasets, with a very limited OCR.

  • The ANN-based sensitivity study with 1000 different training sets revealed that the outcome of repetitive trainings may vary significantly. The difference was more significant for the more scattered dataset F-CLAY/7/216.

  • The ANN-based sensitivity study of different training scenarios revealed that the performance of the training dataset can generally be improved by including more parameters. However, the performance of the testing data rapidly decreases when false predictors are added. This trend is also more significant for the more scattered dataset F-CLAY/7/216.

  • The above indicates that there is a greater risk for overtraining for more incoherent data. That is why, it is recommended to use simple models with few input data, when the data is scattered.

  • Generally, all ML models performed better than the traditional models with the same input parameters, indicating that the dependency of su on these parameters is not as straightforward as in the traditional models.

Data availability

1. The datasets analyzed during the current study are available in the webpage of ISSMGE technical committee 304 at http://140.112.12.21/issmge/tc304.htm

2. The datasets generated during the current study are available from the corresponding author on reasonable request.