Introduction

Infiltration is the vital property of the water. It is the process in which surface water such as precipitation, flood, and snowfall percolate into the soil. Infiltration is the most affecting process during irrigation actions which is to be considered for scheduling of irrigation, irrigation system design and optimization, and management of irrigation system (Al-Azawi 1985; Bhave and Sreeja 2013). It separates the water into two parts: groundwater flow and surface flow (Singh et al. 2018a). There are many parameters such as density, texture, and type of soil and moisture content that affect the infiltration process (Angelaki et al. 2013). The estimation of the infiltration characteristics is also useful to evaluate the performance of the hydrogeological investigations (Pedretti et al. 2012).

Various researchers analyzed the infiltration data and gave some infiltration model for solving the problem related to the infiltration. These models are Green and Ampt, Harton, Philip, Kostiakov, Holton, modified Kostiakov and novel model. Mishra et al. (2003) divided these models into three categories: physical, semiempirical, and empirical models. Sihag et al. (2017a) used four infiltration rate models and found that novel model was the best-suited model for the soil of NIT Kurukshetra campus. Vand et al. (2018) also found the novel model was the best-fit model in two provinces (Lorestan and Ilam) in Iran. Sihag and Singh (2018) also investigated about these models using double-ring infiltrometer and found that Mezencev and modified Kostiakov model can be used to evaluate the infiltration rate of the soil for the given study area. Chowdary et al. (2006) have studied the infiltration process under different experimental conditions. Singh et al. (2018a) compares the four infiltration model and found that modified Philip’s model is the best-fit model among other selected models to calculate the infiltration rate of the soil.

Soft computing techniques such as random forest, adaptive neuro-fuzzy inference system, gene expression programming, support vector machine, generalized neural network, Gaussian process regression, artificial neural network fuzzy logic, and M5P model tree have been widely used in civil and water resources engineering problems. Many researchers used these soft computing techniques in civil and water resources engineering-related problem successfully (Haghiabi et al. 2018; Parsaie et al. 2016a, Parsaie et al. 2017a, b, 2018a, b; Tiwari et al. 2018; Parsaie and Haghiabi 2014b, 2015a, 2017a, b; Haghiabi et al. 2017a; Tiwari et al. 2017; Nain et al. 2018; Azamathulla et al. 2016) and found that these techniques are less time consuming and gave good result. Also, these techniques have the very less optimum user-defined parameters. Sihag et al. (2018b, 2017b) used the different proportion of sandy soil and analyzed the infiltration characteristics in the laboratory. Thus, the aim of this study is to analyze the impact of fly ash in the cumulative infiltration of the soil. The major objective is to compare the performance of various soft computing techniques to predict the cumulative infiltration of the fly ash mixed in the soil.

Soft computing techniques

Artificial neural network

Artificial neural network (ANN) is the most common computing technique which is based on the nerve cells of the human brains. ANN is successfully used in hydrological and water resources problems (Sihag 2018; Sihag et al. 2018d; Haghiabi et al. 2017b; Sihag et al. 2017c; Parsaie et al. 2016b; Parsaie and Haghiabi 2014a, 2015b). Neurons are arranged in the form of layers. Every layer carries out the different kinds of transformations on their inputs. Multiple transformations occurred during signals pass through the first input layer to the last output layer. ANNs have three interconnected layers. The first layer consists of input neurons which receive the input data. Received input data from the first layer are forwarded to the second layer which consists of different hidden layers. After processing of data in hidden layers, they are transferred to the third layer which consists of output neurons. Training an artificial neural network involves choosing from allowed models for which there are several associated algorithms. For further explanation about ANN, readers are referred to Haykin (2004) and Tiwari and Sihag (2018).

Gaussian process

Gaussian process (GP) is an artificial machine learning technique to build computer systems that can adapt and learn from their experience. Rasmussen and Williams (2006) assumed for the processing of GP regression model that the adjoining observations give knowledge to each other. This technique has emerged in recent years and currently successfully applied in various research fields of medicine, chemistry, construction, etc. Gaussian process is based on probability theorem which could make predictions on unknown input data as well as provide prediction exactness which highly increases the statistical significance in prediction. Also Gaussian processes are based on multivariate Gaussian distributions which extend it to infinite dimensionality. Formally, Gaussian process setups the data by using the domain which has any finite subset range following a multivariate Gaussian distribution. In this paper, radial basis kernel \(\left( {K\left( {x,x^{\prime}} \right)} \right) = e^{{ - \gamma \left| {x - x^{\prime}} \right|^{2} }} )\) and Pearson VII function kernel \(\left( {1/\left[ {1 + \left( {2\sqrt {x_{i} - x_{j}^{2} } \sqrt {2^{{\left( {1/\omega } \right)}} - 1} /\sigma } \right)} \right]^{\omega } } \right)\) is used, where γ, σ, and ω are kernel-specific parameters. For further explanation about GP, readers are referred to Kuss (2006) and Singh et al. (2018b).

Support vector machines

The support vector machines (SVMs) are based on statistical learning concept and structural risk minimization hypothesis. The basic concept of SVMs is to arrange the data sets from the input zone to infinite-dimensional feature zone by constructing set of hyper planes so that classification, regression, or other problems become simpler in the feature zone. The hyperplanes are defined as the set of points whose dot product with a vector is constant in that space. Support vector regression has been proposed by Vapnik et al. (1995) and it is a learning system using a high dimensional feature space. The model shaped by SVR depends only on a training dataset because any training data close to the model prediction are ignored by the function for generating the model. Various kernel functions are used with SVM-based regression approaches. In this study radial basis kernel \(\left( {K\left( {x,x^{\prime}} \right)} \right) = e^{{ - \gamma \left| {x - x^{\prime}} \right|^{2} }} )\) and Pearson VII function kernel \(\left( {1/\left[ {1 + \left( {2\sqrt {x_{i} - x_{j}^{2} } \sqrt {2^{{\left( {1/\omega } \right)}} - 1} /\sigma } \right)} \right]^{\omega } } \right)\) are used, where γ, σ and ω are kernel-specific parameter. For further explanation about SVM, readers are referred to Smola and Schölkopf (2004) and Sihag et al. (2018a).

Random forest

Random forests (Breiman 2001) are developed by a collection of tree-based models (Breiman et al. 1984) which can be used for categorization tasks in which the base models are classification trees or regression tasks which depend on base models of regression trees. The forest consists of various trees, which have any value between one to several thousand. To organize a new data set, data set of each condition is passed down every tree. All trees give a classification for that condition. Modeling of a single tree is highly sensitive and complicated. Small changes in the training data turn out a high variation in single classification trees and often lead to rather low classification accuracies (Breiman 1996). Random forests have been justified to be magnificent predictive models in regression tasks and several classifications. They are reasonably fast to obtain results and can be easily assimilated if more speed is required. For further explanation about the random forest, readers are referred to Singh et al. (2017) and Sihag et al. (2018c).

M5P model

M5P is the combined form of the conventional decision tree and the linear regression functions. The model tree algorithm applied in this paper is based on M5P algorithm. The aim of M5P algorithm is to establish a model that evaluates the relation between a target value of the training cases and the values of their input attributes. The performance of the model is checked by the accuracy parameters through which it predicts the values of the curtained cases.

M5P model combines the multiple linear regression and decision tree for data analysis. Decision tree makes relation between the observed inputs and the outputs by logic learning which is appropriate for categorized numerical input and outputs. Decision trees categorize the input dataset and output dataset by the maximized entropy to understand the regression and logical-type rules between inputs and outputs which unambiguously portray the patterns and relationships between data by the regression equations, while other models like SVR and ANN hide them. So, model trees are not only simple but also efficient and accurate technique for modeling and prediction of large data sets (Quinlan 1992).

Conventional models

Two conventional model SCS model and Kostiakov model were also used in this investigation. The empirical constants mentioned in the equation were found by implementing the least-square technique.

SCS model

$$I = at^{b} + 0.6985$$
(1)
$$I = 0.1211t^{0.3449} + 0.6985$$
(2)

Kostiakov model

$$I = ct^{d}$$
(3)
$$I = 0.3824t^{0.2412}$$
(4)

where ‘I’ is cumulative infiltration, ‘t’ is time, ‘a’, ‘b’, ‘c’ and ‘d’ are dimensionless constant.

Measurement of cumulative infiltration

The cumulative infiltration was measured using a mini disk infiltrometer (MDI, Decagon Devices Inc., Pullman, Washington, USA) in a Hydraulic lab of National Institute of Technology, Kurukshetra, India. Figure 1 shows that it consists of two chambers (water reservoir and bubble chamber), connected via a Mariette tube to provide a constant water pressure head of − 0.5 to − 7 cm (equivalent to − 0.05 to − 0.7 kPa). A porous sintered steel disk of diameter 4.5 cm and thickness 3 mm is fitted at the bottom end of the infiltrometer. A suction rate of 1 cm was chosen for this study. The soil used for determination of cumulative infiltration is the mixture of fly ash, sand, and clay at different proportion. The moisture content of the soil samples was measured by the oven-dry method. A proctor (volume 1000 cm3) was used for measuring the density of the soil samples. All the initial conditions like suction head, dry density, moisture content, etc., were predetermined. Properties of the soil samples and details of the soil samples with their moisture content are listed in Tables 1 and 2, respectively.

Fig. 1
figure 1

(Infiltrometer User’s Manual 2014)

Mini disk infiltrometer

Table 1 Details of the material used for experiments
Table 2 Detail of the soil samples with their moisture content

The water-filled infiltrometer (MDI) is placed upon the surface of soil resulting in infiltration of water into the soil and the volume of water in the reservoir chamber was recorded at regular interval. No rainfall occurred during the test period. Infiltration is computed from the cumulative infiltration records versus time following Sihag 2018 and Sihag et al. 2017c, according to Decagon Devices Inc. (2014), and fitted by the function of Eq. 5.

$$I = C_{1} t + C_{2} \sqrt t$$
(5)

where I is the cumulative infiltration (cm), t is the time (s), and C1 (cm/s) and C2 (cm/s−0.5) are parameters. C1 is related to hydraulic conductivity and C2 is related to soil sorptivity.

Dataset

The collected dataset contains a total of 138 field measurement instances having six attributes. Out of 138 datasets, 96 datasets randomly selected for the training of the models whereas the remaining 42 datasets were selected for the testing. The input variables are time, sand, fly ash, clay, density, and moisture content (Mc), and the output variable is cumulative infiltration in this study. Statistical characteristics for input and output variables are shown in Table 3 for the entire dataset.

Table 3 Statistical characteristics of experimental data

Flowchart of the research is given in Fig. 2. Initial step was the selection of the soil (clay and sand), impurity (fly ash), and instrument for experimentation. Total data were divided into two randomly selected parts of training and testing. Training data set was used to calibrate the conventional and soft computing-based infiltration models and testing data set was selected to validate the models. The step (optimum performance evaluation parameters) of flowchart was repeated till optimum users define parameters were achieved. The performance evaluation parameters such as R, RMSE, and NS were selected to compare the conventional and soft computing-based infiltration models.

Fig. 2
figure 2

Flowchart of the research

Results and discussion

The performance of soft computing models (SVM, GP, ANN, M5P model tree, and RF) were compared and evaluated by performance evaluation parameters. Three most frequent performance evaluation parameters were used such as root-mean-squared error (RMSE), coefficient of correlation (R), Nash–Sutcliffe efficiency coefficient (NSE) in this study. The optimum values for the RMSE, R, and NSE are 0, 1, and 0, respectively. The RMSE, R, and NSE can be calculated as:

$$R = \frac{{a\sum mn{-}(\sum m)(\sum n)}}{{\sqrt {a(\sum m^{2} ) - (\sum m)^{2} } \sqrt {a(\sum n^{2} ) - (\sum n)^{2} } }}$$
(6)
$${\text{RMSE}} = \sqrt {\frac{1}{a}\sum\nolimits_{i = 1}^{a} {\left( {m - n} \right)} }$$
(7)
$${\text{NSE}} = 1 - \frac{{\mathop \sum \nolimits_{i = 1}^{z} \left( {m - n} \right)^{2} }}{{\mathop \sum \nolimits_{i = 1}^{z} \left( {m - \bar{m}} \right)^{2} }}$$
(8)

m is the actual values, N is the predicted values, a is the number of observations, and \(\bar{m}\) is the average actual values.

Figures 3, 4, 5, 6, and 7 show the curves of cumulative infiltration for the different amount of fly ash at particular moisture content. Output from Figs. 3, 4, 5, 6, and 7 suggests that cumulative infiltration increases with increase in the concentration of the fly ash. Similarly, in case of the moisture content, cumulative infiltration increases with increase in the moisture content from 2 to 15%, but it starts to decrease when the moisture content increase from 15 to 20%. The cumulative infiltration was observed maximum when the concentration of the fly ash was 50% by weight and moisture content was 15%.

Fig. 3
figure 3

Cumulative infiltration versus time for varying fly ash at 2% moisture content

Fig. 4
figure 4

Cumulative infiltration versus time for varying fly ash at 5% moisture content

Fig. 5
figure 5

Cumulative infiltration versus time for varying fly ash at 10% moisture content

Fig. 6
figure 6

Cumulative infiltration versus time for varying fly ash at 15% moisture content

Fig. 7
figure 7

Cumulative infiltration versus time for varying fly ash at 20% moisture content

Soft computing modeling techniques such as GP, SVM, Random forest, and M5P had applied on observed data to evaluate the prediction efficiency. The user-defined parameters for the various soft computing techniques are summarized in Table 4. The values of sigma, noise, omega, and c were 1, 0.01, 2, and 50, respectively, for the SVM and GP techniques and values of m, k, I, hidden layer, learning rate, and momentum were 10, 1, 100, 7, 0.01, and 0.6 for M5P model tree, random forest and ANN, respectively, whose values the lesser error in prediction of the infiltration rate.

Table 4 User-defined parameters of training and testing dataset

The details of the performance evaluation parameters for the different soft computing techniques are given in Table 5. The values of the R, RMSE, and NSE lie in the range of 0.9996–0.8936, 0.052–0.5919, and 0.9992–0.7868, respectively, for training dataset and 0.9817–0.7868, 0.2387–0.7611, and 0.9613–0.6070, respectively, for the testing dataset.

Table 5 Performance evaluation parameters of training and testing dataset

Table 5 suggests that all the soft computing techniques work well to predict the cumulative infiltration except M5P model tree. M5P model tree gave the worst result with lowest values of R and RMSE (0.7869 and 0.6070) and highest RMSE value (0.7611). SVM with RBF kernel was the best-fit model with the highest values of R and NSE (0.9817 and 0.9613) and lowest value of RMSE (0.2387). Figure 8 shows the agreement plot between actual cumulative infiltration and predicted cumulative infiltration. The output from Fig. 8 also suggests the similar trends that all the soft computing techniques work well except M5p model tree. Almost, all the scattered from M5P model tree lies outside from the ± 20% error line. Similarly, SVM with RBF kernel gave the best agreement with the ± 20% error line with minimum scattered lies outside the agreement line. The soft computing techniques others than SVM with RBF kernel and M5P model tree gave the almost similar result with the range of R, RMSE, and NSE in between 0.9800–0.9509, 0.2535–0.4007, and 0.9564–0.8910 respectively.

Fig. 8
figure 8figure 8

Actual values versus predicted values using different soft computing techniques using testing dataset

Table 6 gives the result of the SCS model and Kostiakov model. The values of R and NSE for these two conventional models were very low, and the value of the RMSE was very high for the training and testing period. Figure 9 shows the agreement diagram of two models (SCS and Kostiakov model). The scatters of these two equations did not lie in ± 50% error line. The output from Fig. 9 and Table 5 suggests that the two conventional model such as the SCS model and Kostiakov model fail to predict the cumulative infiltration rate.

Table 6 Performance evaluation parameters of training and testing dataset of conventional models
Fig. 9
figure 9

Actual values versus predicted values of cumulative infiltration for the SCS model and Kostiakov model of the testing dataset

Figure 10 compares the predicted values and actual values of the cumulative infiltration for the various soft computing techniques and two conventional infiltration model which clearly suggest that all the scatters from conventional model and M5P model tree lies outside of ± 30% error line and almost of the scatters from other conventional model lies within the ± 30% error line. But SVM with RBF kernel gives the best-fit scatters. Hence, SVM with RBF kernel is the superior soft computing techniques and can be successfully used to predict the cumulative infiltration for the given combination of the soil.

Fig. 10
figure 10

Actual values versus predicted values using all model for the testing dataset

Conclusions

Infiltration process has been experimentally investigated on sand, clay, and fly ash mixed samples using mini disk infiltrometer. Various soft computing techniques such as SVM, GP, M5P model tree, random forest, and ANN for modeling the experimental result and prediction of the cumulative infiltration. The various conclusions drawn from this investigation are given as follows:

  • It is established from the test results that increment in the cumulative infiltration was observed with increment in the percentage of the fly ash. But there is a contradiction that it increases up to 40% of the fly ash and decreases when the percentage of the fly ash increases from 40 to 50%.

  • The cumulative infiltration also increases with the decrement in the percentage of the clay.

  • The prediction of the cumulative infiltration is tested by various soft computing techniques (SVM, GP, M5P model tree, random forest, and ANN), and SVM with RBF kernel was found best to predicting the cumulative infiltration followed by GP, random forest, ANN, and M5p Model Tree.

  • Obtained results suggest that the performance of conventional models, SCS model and Kostiakov model, is not satisfactory as compared to the soft computing techniques.