1 Introduction

Ultra-high-performance concrete (UHPC) has gained significant attention and practical applications in the construction industry due to its exceptional compressive strength and durability. Its use extends to various structures, including super high-rise buildings, long-span bridges, maritime structures, bridge girders, bridge decks, and pre-stressed girders maintenance (Xue et al. 2020; Nematollahi et al. 2012). The mechanical performance, toughness, ductility, and long-term durability of UHPC make it a preferred material for critical infrastructure. The key factors in constructing advanced UHPC include developing a highly dense particle-carrying architecture and incorporating steel fibers for concrete reinforcement. Achieving an optimal steel fiber content in the UHPC composition is crucial to establish the desired characterizes of the material. These steel fibers enhance the UHPC's tensile durability and ductility, resulting in enhanced crack resistance and overall performance. (Yu et al. 2014). Indeed, the material and structural characteristics of UHPC differ significantly from those of conventional concrete structures. UHPC is famous for its exceptional physical characteristics, including higher compressive strength, higher tensile strength, and excellent durability. These unique properties make UHPC well-suited for various applications, especially in situations where enhanced shear and flexural behavior are essential for structural stability and design (Yavas and Goker 2020a). In such stochastic settings, understanding and predicting shear behavior become even more critical for ensuring the structural integrity of UHPC members. Proper consideration of shear behavior in the design process helps prevent shear failures, shear cracks, and other potential structural issues that may arise due to the complex randomness in geometrical materials. The shear behavior of UHPC has been the subject of numerous studies to understand its mechanical response and ensure the safety and reliability of structures made from this advanced material. Some of the key studies and findings related to the shear behavior of UHPC are summarized. Simwanda et al. (2022) focused on the structural reliability of UHPC fiber-reinforced concrete beams with stirrups. They conducted a numerical analysis using non-linear finite element methods and compared the results with standard design codes to assess the accuracy of existing design approaches. The inclusion of fibers and stirrups in UHPC beams significantly influenced their shear behavior, and the study aimed to validate and enhance the design guidelines for UHPC structures. In a research by Qi et el. (2016), UHPC fiber-reinforced concrete beams underwent a four-spot bending trial to investigate their flexural response, especially changes in bending resistance. Graybeal et al. (2006) used pre-stressed I-UHPC beam and conducted three shear tests without any reinforcement involvement. The findings revealed that conventional code requirements often underestimate shear resistance of UHPC beam. This highlights importance of studying and understanding specific shear behavior of UHPC to develop accurate design guidelines. Some researchers (Yang et al. 2022; Liu et al. 2020) explored interaction characteristics with steel bars and UHPC, particularly in context of geopolymer-based UHPC (G-UHPC). Understanding the bond characteristics is crucial for assessing the comprehensive shear behavior of UHPC structures, as the interaction of concrete with reinforcement significantly influences the load transfer capacity. Understanding the complexities of shear behavior, especially in challenging scenarios such as intense pressures and early concrete cracking (Elsayed et al. 2022; Yoo and Banthia 2016), is essential for optimizing UHPC's performance and promoting its use in practical engineering applications (Raheem et al. 2019; Schmidt and Fehling 2005). Properly designed UHPC structures with enhanced shear behavior can lead to more durable and resilient infrastructure, offering significant benefits in terms of safety and performance. The outstanding mechanical characteristics of glass FRP, such as its stiffness and strength, make it a valuable material for reinforcement (Silva and Rodrigues 2006). Understanding the bond behavior and stress-slip relationship between UHPC and GFRP is essential for using this combination in practical engineering applications. Such research could lead to the commercialization of UHPC and its successful implementation in real-world structures (Tian et al. 2019; Liao et al. 2022). Overall, these studies highlight the importance of investigating the shear behavior of UHPC and its interaction with various reinforcement materials.

The complexity of the shear mechanism in UHPC presents a significant challenge in developing a proper theoretical calculation method for predicting shear behavior accurately (Chen et al. 2019). Existing UHPC guidelines from different countries, such as the French UHPC guideline AFGC-1013 (Resplendino 2011), Japan Society of Civil Engineers guideline (Uchida, et al. 2005), and Switzerland UHPC standard (S.I.A.) (Graybeal et al. 2020), often rely on classical Strut and Tie theory and employed semi-theoretical shear strength and hybrid empirical calculation methods to estimate shear capacity. Despite efforts by researchers worldwide to propose various shear calculation methods, such as plasticity theory, Strut and Tie method, limit equilibrium theory, and pressure field theory (Nielsen and Hoang 2016), these formulas often lack consistency and may even yield conflicting results. Consequently, the reliability of these shear strength calculation methods remains questionable. Given the limitations of traditional theoretical methods, there is a need for a more reliable and efficient approach to predict shear behavior in UHPC structures. ML models offer a sophisticated solution to this challenge. Numerous fields of Civil Engineering domain, for example, Bridge engineering (Melhem and Cheng 2003), Tunnel engineering (Xu et al. 2019), and Transportation engineering (Arciszewski et al. 1994) problems, have been successfully solved using the ML model. ML models, such as artificial neural networks (ANN) (Ghafari et al. 2012), support vector machines (SVM) (Umar et al. 2022), and tree-based ensembles (Umar et al. 2022) like XGBoost and random forest, deep learning (DL) (Wu et al. 2022), etc. have shown great promise in solving complex engineering problems. ML models can capture non-linear relationships between input parameters and shear behavior in UHPC structures. By training on experimental data from various UHPC beams, ML models can learn patterns and correlations that conventional methods may overlook. This data-driven approach enables ML models to provide accurate predictions of shear capacity while minimizing the computational effort compared to traditional numerical methods. In summary, due to the complexity and inconsistencies in existing theoretical calculation methods for predicting the shear behavior of UHPC, a more sophisticated approach like ML models is needed. By leveraging the power of data and pattern recognition, ML models can offer reliable predictions of shear behavior with reduced computational requirements, making them a valuable tool for designing and analyzing UHPC structures.

Various research studies and methodologies leverage artificial intelligence and advanced numerical techniques to address complex engineering problems. For example, Tran et al. (2023) focused on using damage indicators in combination with the ALOANN (Augmented Lagrangian-based Optimal Artificial Neural Network) method to detect damage in structures. By incorporating damage indicators, it becomes easier to identify the location and extent of damage in the structure. ALOANN is likely an optimization-based approach that utilizes ANN to efficiently and accurately detect structural damage. Nguyen and Wahab (2023) proposed a method to enhance the calibration process of 2D VARANS-VOF (Variational Asymptotic Navier–Stokes-Volume of Fluid) models. These models are used to simulate how waves interact in fluid–structure interaction problems. The proposed methodology likely improves the accuracy and reliability of the simulation results, enabling better predictions of wave interactions in complex scenarios. In research by Dang et al. (2023), ANN is integrated with balancing composite motion optimization (BCMO) to address optimization tasks related to the vibration and buckling behavior of mathematically varied microplates in unknown physical attributes. The combination of ANN and BCMO likely offers an efficient and effective approach to solve optimization problems in the context of composite materials with varying properties. Convolutional neural networks (CNN) and region-based CNN (R-CNN) were hired to forecast various types of damage and accurately predict bounding boxes encompassing the affected regions (An efficient stochastic-based coupled model for damage identification in plate structures). By using CNN and R-CNN, a robust and accurate damage forecasting in images or visual data can be achieved.

The existing ML methods for predicting the shear performance are concisely presented in Table 1. The ML model evaluation involved eight ML models and a few hybrid ML models. Among these models, XGBoost showed the highest test accuracy with a high R2 value (R2 > 0.99), and there was relatively low variance between the training and test datasets. Additionally, ANN, R.F., and SVM also demonstrated good accuracy in both training and test datasets, with minimal variance in R2 values. Other hybrid ML models achieved high accuracy in both training and test results.

Table 1 Previous literature for predicting C.S. and F.S. of concrete materials

As far as the author knowledge, no previous research has been carried out specifically on UHPC with GFRP bars using standard country code and compared with ML models. To address this research gap, the study used four different ML models to estimate shear response of I-UHPC GFRP and compared the ML results with conventional country code results.

Employing the experimental data collected from the literature review, proposed ML models outperformed the traditional country code results in predicting the shear response of IUHPC-GFRP. This suggests that the ML models provide greater accuracy in predicting shear response of UHPC beams, especially in combination with GFRP reinforcement, compared to the conventional methods based on standard country codes.

In summary, the study demonstrates the potential of ML models to enhance the accuracy of predicting shear behavior in UHPC structures, particularly when considering the use of GFRP reinforcement. The results provide valuable insights into the application of ML techniques for optimizing the design and performance of UHPC structures in practical engineering applications.

2 Shear strength models of UHPC beams for different codes and proposed equations

Figure 1 depicts the interconnection between the maximum force and the central deflection for UHPC beams without strips (a) and with stirrups (b). The figure displays an ideal model of GFRP UHPC beam specimens with and without stripes.

Fig. 1
figure 1

A model of GFRP-reinforced UHPC beam specimens

The available design code, such as ACI318M-14 and Eurocode2, established functional equations to estimate the response of the UHPC I-shaped beam in terms of material characteristics and geometric range.

2.1 ACI318M-14

$$V_{u} = \left[ {\sqrt {f^{\prime}_{c} } + 120\rho_{s} \left( \frac{d}{a} \right)} \right]\frac{{b_{w} d}}{7} + \frac{{A_{sv} f_{yv} d}}{s}$$
(1)

The \(f^{\prime}_{c}\) is denoted as compressive strength, and the longitudinal reinforcement ratio is termed as \(\rho_{s}\). A, \(S_{w}\) and w is regarded as distance of shear span, operative width, and operative beam depth. Additional parameters include \(A_{sv}\) the stirrup's cross-sectional area and its yield strength, and s denoted as stirrup spacing.

2.2 Eurocode 2

$$V_{RD} = \left[ {0.18k(100\rho_{s} f^{\prime}_{c} )^{\frac{1}{3}} } \right]b_{w} d + \frac{{A_{sv} f_{yv} Z}}{s}$$
(2)

Here,

$$k = 1 + \sqrt{\frac{200}{d}}$$
(3)

Other values are the same as the ACI318M-14 specifications. Deviation), StandardScaler was adopted to handle the missing data processing in 3DP-FRC.

3 Database development

The formation of ML prediction models highly relies on constructing an authentic dataset. For this reason, a comprehensive literature study was conducted to compile information from previously published studies. Therefore, 54 UHPC beams were collected from 8 different published literature (Akbar et al. 2023; Yavas and Goker 2020b; Chenggong et al. 2022; Mészöly and Randl 2018; Ashour et al. 1992; Lee et al. 2012; Bahij et al. 2018; Jin et al. 2020) to understand the shear performance of glass FRP bars reinforced UHPC I-shaped beam. The datasets stand for varying situations involving shear performance conditions of FRP bars reinforced UHPC I-shaped beam, and all the beams ranged from having 0% additive to having 2% admixture by weight. The input variables were used to forecast the shear performance, and eleven input variables are comprised in our database, which is also used in predicting shear capacity through the county code conventional formula. The input parameters are cubic compressive strength \((f_{c} )\), cylinder compressive strength \((f_{cc} )\), the operative width of beam \((b_{w} )\), the operative depth of the beam \((d)\), ratio of shear span to depth(\(\lambda\)), longitudinal reinforcement proportion \((\rho_{s} )\), the angle between compressive stress and beam axis \((\theta )\), the cross-sectional area of stirp \((A_{sv} )\), stirrup spacing \((s)\), yield strength of stirrup \((f_{yv} )\) and proportion of steel fiber volume \((p_{f} )\). The final output included in this study's database was the experimental value \((V_{e} )\) of shear behavior of glass FRP bars. Table 2 presents the analytical properties of the dataset's input variables, output variable, and model performance metrics, in average, standard variation, lowest, and hishest values.. The \((f_{c} )\) ranges between 108.2 and 193 MPa vary between 96.8 MPA and 173.7 MPa while remaining between 225 and 640 MPA, respectively. Another parameter, such as \((\lambda )\) is 3.5 and \((P_{f} )\) can be added to 4.5%.

Table 2 A detailed examination of analytical features

Figure 2 demonstrates the correlation matrix (Li et al. 2021) between inputs and outputs variables, and a small square represents the correlation between the parameters on each axis. The same two parameters are linked together in those squares; as a result, the plot shape becomes symmetrically diagonal. The correlation is calculated in the range of negative one to positive one, and no evident linear connection involving the two variables if the values are closer to zero. The correlation drawing nearer to the value of 1 demonstrated that inputs are more closely linked with one another. If a specific inputs value rises, another does accordingly. Corresponding results can be obtained with a correlation nearer to a negative one, but instead of both variables increasing simultaneously, one will decrease when the other rises.

Fig. 2
figure 2

Correlation between input and output parameters

The highest correlation coefficient's absolute value is about 0.65, thereby revealing that there is a notable linear relationship between \(f_{c}\) and \(f_{cc}\). The \(f_{yv}\) and the \(A_{sv}\) have a strong positive correlation. Moreover, with a correlation value range of -0.87, \(b_{w}\) has a negative correlation with \(\lambda\).

3.1 Data preparation and handling the missing value

The entire database undergo z-score normalization to standardize the values as experimental data collected from multiple origin exhibits diverse unit and range.

$$Z_{\ln } = \frac{{x_{\ln } - \overline{{x_{n} }} }}{{\delta_{j} }}$$
(4)

The \(Z_{\ln }\) represents the standardized value of \(x_{\ln }\), the nth variable of lth data's inputs, \(\overline{{x_{n} }}\) is the average value of nth input parameters, and n is the nth dispersion measure of nth input parameters.

$$\delta_{j} = \sqrt {\frac{1}{D - 1}\sum\limits_{l = 1}^{D} {x_{\ln } - \overline{{x_{n} }} } }$$
(5)

where D denotes the number of data entries. Z normalization method control the mean of every inputs about zero and standard variation about 1.

Numerous difficulties were encountered during the data-gathering process, including insufficient details and missing data. Several publications address the issue of missing values as a critical problem in dealing with machine learning models. Few studies have been conducted on dealing with missing data through statistical approaches(Audet et al. 2022; Fei et al. 2023; Enders 2022; Memon et al. 2022); other scholars have taken different approaches to fill up the missing data by replacing missing values with zero, mean values, and medians (Enders 2022; Somer et al. 2022; Little and Rubin 2019). Several numerical models were conducted by a few authors, such as Chakraborty and Gu (Chakraborty and Gu 2019) proposed a mixed numerical model to deal with a high percentage of missing values, Wei et al. (2018) impended eight imputation models. They compared the models for handling different types of missing values, and Nazbzul et al. (2020) suggested a general framework for adapting insufficient heterogeneous missing data.

For this study, we used XGBoost to fill up the missing data, and XGBoost is a popular machine-learning model for dealing with missing values before normalization. Some other imputation approaches, such as K.N.N. and decision trees, are also employed to deal with missing data (Emmanuel et al. 2021), but XGBoost is more accurate when handling missing data (Aydin and Ozturk 2021). Built on the internal sparsity-aware algorithm, the XGBoost method can handle missing values in a more sophisticated way (Chen and Guestrin 2016).

The XGBoost machine learning model will estimate the split object for the non-missing data throughout the model training process, choose the largest split object, and appoint it as a split node by selecting a specified threshold for a specific variable. A sparsity-aware split-finding algorithm can efficiently manage missing values. Therefore, only non-missing samples are evaluated to determine the best direction, reducing the algorithm's complexity. The flow chart in Fig. 3 will graphically illustrate the process.

Fig. 3
figure 3

Flowchart of Handling Missing Data Through XGBoost

4 Applying machine learning (ML) model

4.1 Support vector machine (SVM)

Structured risk minimization approach provides a basis for the support vector regression method, where the empirical risk needs to be reduced for every section of the structure(Awad and Khanna 2015). S.V.R. seeks to minimize the generalization error's rather than revealing empirical faults. Let's presume a data set

$$\{ G = (x_{1} ,y_{1} )........(x_{n} ,y_{n} )\} ,where|n = 1.......Z$$
(6)

Input data base \(x \in R^{G \times Z}\) is our devised matrix and \(y \in R^{n}\) is our designed vector output which assigned for training sets. Here, x symbolizes the higher dimensional input pattern space. SVM regression aims to construct a function p(x) with the highest error derivation from the actual targets for all training data.

The standard equation of the non-linear support vector regression function is

$$p\left( x \right) = w^{T} \sigma \left( x \right) + k$$
(7)

Here in Eq. (7), \(\sigma\) is a non-linear transformation operator, while w indicates weight, k suggests bias, and T indicates transpose vector. W and k are computed using data by reducing the normalized risk function of the variables.

$$M(c) = \frac{1}{2}||W||^{2} + C\left[ {\frac{1}{D}\sum\limits_{l = 1}^{D} {L\sigma \left( {y_{i} ,\left( {Px_{i} } \right)} \right)} } \right]$$
(8)

The 'regularized term' is the previous term on the right-hand side of Eq. (8) to evaluates the function's flattening. It minimizes flattens level to a maximum extended level. The equation has an empirical error, and there has a loss function to calculate this error.

$$L_{\sigma } \left( {y_{i} ,P(x_{i} )} \right) = \left\{ \begin{gathered} y_{i} - p(x_{i} ) - \sigma ;where,\sigma \le |y_{i} - p(x_{i} )| \hfill \\ y_{i} - p(x_{i} );where,\sigma > |y_{i} - p(x_{i} )| \hfill \\ \end{gathered} \right.$$
(9)

The \(y_{i}\) represents the actual value and \(p(x_{i} )\) shows the estimated value. From Eq. (8), the variable C, also known as the regularized constant, is merely a variation between empirical error and the normalized function. Together C and \(\sigma\) hyperparameter can be specified by us where the value \(\varepsilon\) determine the training accuracy. Data sets inside the function are used as a support vector. Therefore, the training set doesn't play any role in making decisions.

The final form of the non-linear Support vector regression will be shaped as below:

$$P(x) = \sum\limits_{l = 1}^{D} {\left( {a_{i} - a_{i}^{*} } \right)S\left( {x_{i} ,x_{j} } \right) + b}$$
(10)

Here \(a_{i}\) and \(a_{i}^{*}\) is Lagrange multipliers, which require to satisfy the positive constraints. In certain ways, the function depends on the input size and input dimension. The \(k\left( {x_{i} ,x_{j} } \right)\) denoted as kernel operator whose value is determined by the inner product \(\phi \left( {x_{i} ,x_{j} } \right)\). There are numerous varieties of kernel operator. Gaussian radial basis function (RBF) stands out as one of the most frequently utilized kernel functions and appropriate for dealing with non-linear regression problems as it can translate the training data into multi-dimensional space vector non-linearly.

4.2 Random forest (R.F.)

The Random Forest (R.F.) model combines many decision tree nodes with bootstrapping and aggregation concepts, has been successfully used for decision tree optimization methods. In this model, often known as a random forest, a randomized collection of predictor factors determines how each tree develops. Each tree is constructed using random forests employing sample replacement. Let's assume generating \(z\) tree with tree \(k_{z} \left( x \right)\) the formula is shown below yields the R.F. regression model.

$$U(x) = \frac{{\sum\limits_{l = 1}^{l} {K_{d} (x)} }}{d}$$
(11)

The nonparametric regression method known as RF regression utilizes an ensemble to generate K outputs for every tree within a set of Z trees.\(\left\{ {J_{1} \left( i \right),J_{2} \left( i \right),.......J_{z} \left( i \right)} \right\}\) where \(I = \left\{ {i_{1,} i_{2} ,.......,i_{b} } \right\}\) is a vector of Y dimensions that generates an output vector of size \(i\) for each tree in the forest,\(\beta_{i} = \left( {i = 1,2,......J} \right)\). Each decision tree is built using a bootstrap sample created by randomly choosing examples from the previous training database in a repetition process.

The remaining one-third of the training set forms the out-of-bag (OOB) sample. A combined amount of two-thirds of the training set are used to obtain the regression function. A regression tree is built every time through a arbitrary chosen training set taken from original database.

$$JI\left( {k_{d(xl)} } \right) = 1 - \sum\limits_{l = 1}^{n} {U(} k_{d(xl)} ,n)^{2}$$
(12)

When using unknown test data, the built-in validation features enhance the random forests' capacity to predict more accurately. The random forest prediction model's precision is evaluated based on its accurate predictions of unknown test sets.

4.3 Artificial neural networks (ANN)

To detect patterns, learn from the estimated error, and forecast outcomes in higher-dimensional space, artificial neural networks (ANNs) are complicated data processing methods that are successfully performed with great accuracy in the research area of classification, non-linear regression, and time series forecasting analysis. Employing a network of interlinked neurons generated from complicated non-linear relationships between input variables \(x_{l}\), weights function \(w_{l}\), and accepted bias \(b_{l}\), in the \(l_{th}\) hidden layer with a target value of \(y\) in \(n_{th}\) neuron through \(n_{th}\) hidden layer. Each neuron's estimated weights and biases are added to generate a network architecture below.

$$N = \sum\limits_{l = 1}^{D} {W_{\ln } X_{l} + b_{l} }$$
(13)

A hyperbolic tangent function was employed to improve the output. The formula's values fall within the range of 0 and 1, and its definition is as follows:

$$y = V(N) = \frac{2}{{1 + e^{ - 2N} }} - 1$$
(14)

The \(V\) is an active function of the output layer, which enables the network to extract relevant data and control non-linearity to the output layer. Figure 4 shows the ANN with eight input parameters hidden layer and an output layer for predicting Ve (K.N.).

Fig. 4
figure 4

ANN architecture

The multilayer perceptron's weights and biases are updated to quantify and minimize the error value of ANN. The process is done on each multilayer perceptron's layers until the final layer produces the output signal. The network is optimized continuously in the training process, which is carried out using various optimization algorithms through repetitive training samples. The quantity of hidden layers and a total number of neuron counted in every layer are the principal hyperparameters of ANN. In our model, we implemented three hidden layers and the RELU activation function to control the input weighted sum and convert it into an output result.

4.4 Extreme gradient boosting (XGBoost)

XGBoost provide exceptional performance for regression model. The technique use "Parallel boosting" concept, that constructs potent learner by amalgamating additive training techniques. XGBoost proves invaluable for prediction, as it effectively mitigates over-fitting and enhances computational efficiency through the integration of regularization terms. By minimizing objective functions, it ensures optimal processing performance. In practical applications, it can be regarded as a soft comprehensive software library that synergistically merges distinctive algorithms with the decision tree approach, leading to enhanced estimation accuracy. Presume the same data base as shown in Eq. 6, the input data set and designed vector output which assigned for training sets.

The suggested ensemble of trees employs n adaptive functions to approximate the system's behaviour.

$$\mathop {\overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\frown}$}}{y}_{l} }\limits_{{}}^{{}} = \phi (x_{l} ) = \sum\limits_{n = 1}^{n} {f_{n} \left( {x_{l} } \right),f_{n} \in T}$$
(15)

The set of regression trees is denoted by T and characterized as follows

$$C = \left\{ {f(x) = A_{c(x)} } \right\}\left( {t:R^{n} \to J,A \in R^{m} } \right)$$
(16)

here, U denotes the tree configurations, J and A indicate leaf node quantities and corresponding values. Furthermore, the term \(f_{n}\) illustrates the association between A and t for the individual tree.

The OA of XGBoost can be reduced as follows,

$$y^{z} = \sum\limits_{l = 1}^{n} {E\left( {y_{n} ,\overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\frown}$}}{y}_{n}^{(t + 1)} } \right)} + f_{n} (x_{n} ) + Q(f_{t} )$$
(17)

Here, \(y_{n}\) is considered as a measured value, \(\overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\frown}$}}{y}_{n}\) stands as a estimated value and \(E\) is utilized as a convex function (loss function) to assess and compare the precise and predicted values. The iteration count t is employed for error reduction, while Q represents the retribution for the regression tree model's intricacy.

$$F = y^{j} + \frac{1}{2}d||r||^{2}$$
(18)

where, r represents for the number of Leaf vector ratings, d manages regularization process, while \(y\) computes the minimum loss required to split the leaf node calculated in aforementioned equation.

5 Results and discussions

5.1 Hyper-parameters tuning

Generally, To build an effective machine learning model, it is essential to choose the optimal algorithm and carefully adjust the hyperparameters to attain the most favourable model architecture. (Chen and Guestrin 2016). Hyperparameters control the behavior and structure of the training period in ML models. A few common hyperparameters, for example, the support vector machine's penalty parameter (C), neural network's learning rate, and kernel type of Support Vector Machine are applied to optimize the algorithm designed and minimize the loss function (Awad and Khanna 2015). In this study, Random Search (R.S.) methods (Schratz et al. 2019) are employed to fine-tune the chosen hyperparameters of models. As possible hyper-parameter settings, R.S. randomly chooses a predetermined samples number from a range between the minimum and maximum values. These possibilities are subsequently trained up until the budget allocation is reached. As per R.S.'s conceptual framework, concerning the configuration space is sufficiently large, it is viable to identify global optimums or, at least, close approximations (Schratz et al. 2019). The primary benefit of R.S. is that it can quickly assign resources and processes parallelly compared with the grid search algorithm (Awad and Khanna 2015). When adjusting hyper-tuning, cross-validation is a potential rescaling approach used to regulate prediction accuracy. Out of many training and test splits set, this database was divided into training and testing group using an 70/30 split. In this study, five subsets of the data are divided into the five-fold cross-validation method, with one subset being kept evaluating the trained model after every iteration. The training data were divided into five segments, with the first segment of every iteration is being employed for testing purpose while the remaining segments employed to train the model. The ANN, SVM, R.F., and XGBoost algorithms given in Table 3 had their best parameter settings for our database determined through fine tuning. The average of the predictions is employed to assess the model's performance. This research employs four distinct statistical performance metrics, including mean absolute percentage error (MAPE), coefficient of determination (R2), root mean squared error (RMSE), and mean absolute error (M.A.E.).

$$R^{2} = 1 - \frac{{\sum\limits_{l = 1}^{n} {\left( {y_{l} - \widehat{{y_{l} }}} \right)^{2} } }}{{\sum\limits_{l = 1}^{n} {\left( {y_{l} - \overline{y} } \right)^{2} } }}$$
(19)
$$MSE = \frac{1}{n}\sum\limits_{l = 1}^{n} {\left( {y_{l} - \widehat{{y_{l} }}} \right)^{2} }$$
(20)
$$MAE = \frac{1}{n}\sum\limits_{l = 1}^{n} {\left| {y_{l} - \widehat{{y_{l} }}} \right|}$$
(21)
$$RMSE = \sqrt {\frac{1}{n}\sum\limits_{l = 1}^{n} {\left( {y_{l} - \widehat{{y_{l} }}} \right)^{2} } }$$
(22)

where \(y_{l}\) is the true value (experimental value) and \(\widehat{y}\) is the ML predicted value, and \(\overline{y}\) is the arithmetic mean of y values.

Table 3 Fine-tuned hyperparameters of GFRP reinforced UHPC beam ML model

5.2 Machine learning (ML) model results

The utilized computational environment used in this study, involved popular programming languages like Python and reliable ML libraries commonly applied in the machine learning domain. The ML models' findings show that among various sets of training testing split, one model that employed 70% of the data for training while 30% of data employed for testing phase produced the best regression accuracy. An in-depth examination of the model is conducted as the performance results improve through hyperparameter tuning, enhancing the overall analysis.

Table 4 displays the accuracy of ML models based on the R2, M.A.E., RMSE, and MAPE. When the error value is reasonably low, it is considered to have a high-performance model. The R2 value demonstrates the ML model's ability to evaluate the experimental data. The cost function, usually known as the RMSE, is critical to how an ML system learns. Parameters of both RMSE and M.A.E. measure the precision and quality of fit.

Table 4 ML model's performance

Figure 5 displays the performance of the training dataset's ML models with estimated and experimental value. The distance of data points from the adjusted curve proved high consistency between experimental findings and calculated values across all datasets. Among all the predicted models, XGBoost predicted most accurately (R2 = 0.99). Other models also have high prediction accuracy, such as Random Forest has R2 = 0.95, ANN has R2 = 0.87, and SVM has R2 = 0.85. Similar results are shown for the testing dataset. Figure 6 displays significant XGBoost prediction accuracy for the test dataset with R2 = 0.85 while R.F., ANN, and SVM have an R2 value 0.832, 0.823, and 0.762.

Fig. 5
figure 5

Performance of the ML model from training evaluation

Fig. 6
figure 6

Performance of the ML model from testing evaluation

Additionally, in Fig. 7, the solid lines evaluates the relationship between predicted value with the experimental value for train and test data based on suggested ML methods. The grey curve depicted the shear capacities of 95% overestimation or underestimation. These figures show a strong correlation between the predictions made by all models and the relevant experimental values (average R2 0.81 for all models). The machine learning outcome indicated that the XGBoost model outperformed all other models in estimating shear behavior. The XGBoost model's prediction accuracy and shear capabilities of experimental data are strongly correlated, as shown by the training and test dataset's respective lower error values of MAE 13.89 KN and 57.80 KN.

Fig. 7
figure 7

Evaluation of training and testing performance based on linear regression analysis

Furthermore, accuracy assessment of the ML methods revealed a minimally low variation between actual strength capacity and the estimated RMSE values. The XGBoost method prediction accuracy in training and test stage are between 17.66 KN and 76.57 KN, respectively. For the SVM, ANN, and R.F. models, these values were 98.97 KN, 78.89 KN, and 44.64 KN for the training dataset and 84.07 KN, 97.75 KN, and 81.94 KN for the test dataset (Table 4). According to Fig. 7 and Table 4, the R.F. model acquired second place in predictive ability, having the second highest R2 and the comparative lowest M.A.E., MAPE, and RMSE value, followed by ANN.

Figure 8 illustrates the residual assumption of the shear behavior of a glass F.R.P. I-shaped UHPC beam. The ML model's residual dispersion is roughly the same across different models along the x-axis of λ. All machine learning models display almost the same trend and pattern for XGBoost, ANN, SVM and R.F.. The model indicates that the mean is normally allocated, without significant curvature or non-

Fig. 8
figure 8

Residual analysis of ML model in term of \({\uplambda }\)

normality, showing that the regression model is not inadequate.

5.3 Standard country code resultsV

Figure 9aand b depicted the correlation between observed shear capacity and predicted shear force using the conventional formula of Eurocode2 and AFGC-2013 country codes. These outcomes are calculated in R squared value, while best-fitting data values are found in AFGC-2013 between them. The regression model is used to assess intensively the predicted shear behavior from various models and design codes listed in Table 5. Table 5 also shows that between the two design codes, ACI318M-14 has the highest prediction accuracy (R2 = 0.39), with the lowest average of 0.486 where the lowest standard variation of 0. 223. ML model results show that the average and standard variation for SVM and ANN are same, with a value of 0.996 and 0.224, respectively. Thus, it demonstrates a strong relationship between the predicted value and the suggested ML model in this study.

Fig. 9
figure 9

Standard country code prediction results

Table 5 Comparison between standard country code and ML prediction results

In this dataset, XGBoost demonstrate a mean of 0.941 and a variance of 0.915, while R.F. models have a mean and standard variation of 0.191 and 0.229. XGBoost for determining shear behavior performs better than all other models. The M.S.E. and RMSE are also calculated in Table 5 to fulfill the robust machine learning model. Therefore, the strongest interconnection between the experimental and predicted value for training and testing phase can be seen in R2 value. Furthermore, the future utilization of larger datasets is recommended to enhance prediction accuracy, allowing the integration of hybrid ML algorithms for improved forecasting.

Figure 9c and d illustrate the standard residual error. The data along the x-axis essentially evenly distribute the residual assumption with a consistent variation and variability. This figure verifies the validity of the assumption model by showing that the mean is uniformly distributed and free of any systematic curvature or non-normality.

5.4 Taylor diagram

Figure 10 represents a Taylor diagram that statistically compares the four-machine learning models. The diagram presented a concise statistical evaluation of the similarity between trend parameters based on correlation, root-mean-square variance, and variation proportion. This diagram displays three popular statistical goodness of fit measurement parameter, such as R2, RMSE, and S.T.D to evaluate the accuracy among ML model prediction result with experimental values of shear behavior. As observed in the figure, all machine learning models have a higher rate of prediction accuracy (higher R2 value) along with lower RMSE, and XGBoost shows the best prediction accuracy among all models with the highest correlation value (0.928) and least RMSE (74.627) value. The ANN demonstrated the lowest prediction accuracy in a correlation coefficient around 0.877, identical to SVM with the same high RMSE (96.461) value. The R.F. model showed satisfactory prediction accuracy while maintaining a position between the best and least accurate model. However, with considerable accuracy, all four ML models can generally predict the shear behavior of glass F.R.P. bars-reinforced ultra-high-performance concrete I-shaped beams.

Fig. 10
figure 10

Taylor Diagram presents a statistical assesment among various ML models

5.5 SHAP analysis

The Shapley Additive explanations (SHAP) algorithm is used to determine the relative importance of the factors that affect the shear strength behavior of UHPC computed by XGBoost. As a result, it is possible to statistically analyze the input–output correlations that are inherently hidden in the conventional machine learning model. According to the analysis of this study, the XGBoost model has shown higher accuracy than other ML models. Therefore, the SHAP value may be used to evaluate the XGBoost predictions in various ways. It shows the significance of each parameter and quantifies how each parameter affects each-others for determining model prediction accuracy. Figure 11 provides the input parameters' feature importance, which illustrates each feature's overall influence on the model predictions.

Fig. 11
figure 11

SHAP summary plot and parameter significance distribution results

It can be comprehended from Fig. 11a that Stirrup Spacing (S) significantly influences the shear behavior of glass FRP bars reinforced UHPC, while they \(f_{yv}\) are of the minor importance. \(f_{cc}\) also affects the shear behavior but the significance is much less than \(\rho_{f}\) and slightly higher than \(d\). Among all other parameters, \(\lambda\), \(f_{c}\) and θ has an inadequate amount of influence on output results, and other remaining variables have very poor effects as well. A SHAP summary plot of the features is shown in Fig. 11b through a beea swarm plot. The distribution plot demonstrated the SHAP values for every feature with underlying influence patterns to demonstrate density and dots piled up along every feature sequence. The figure shows the individual SHAP value on the x-axis and the input parameters ranked in order of importance on the y-axis. The database's outcome results are represented in a violin plot, and the color warm indicates the rising or decreasing of a particular value in input features, and values range from small to large in colors blue to red. The plot demonstrates that, on average, stirrup spacing (s) is the most significant feature. Consequently, the summary plot provides information about which feature is highly essential and how one feature affects another, determining shear behavior. The general observation from the figure shows that the shear strength of glass fiber I-shaped UHPC will rise with an increase in the features of \(\rho_{f}\), \(f_{cc}\) and \(d\). On the other hand, shear strength leans to decline as parameters for features like \(\lambda\) and θ increase in values.

6 Conclusions

This research's main driving force and motivation is utilizing the soft computing method to predict the shear behavior of UHPC I-shaped glass bars. Based on experimental data from eight pieces of literature, this study made a machine-learning model to predict the shear strength and performance behavior of UHPC I-shaped beams and compared the ML results to traditional country code results. This study looked at how well four machine learning models, such as ANN, XGBoost, and SVM, could predict the shear strength of a glass FRP UHPC I-shaped beam. The results show that the integrated machine learning approach is very good at predicting UHPC shear behavior, and we conclude that ANN, R.F., SVM, and XGBoost can all be used to figure out how an R.C. beam is sheared. The results of the study led to the following conclusions.

  1. 1.

    During the training phase, the ideal hyperparameters of ML-based models were chosen using a hybrid search with fivefold cross-validation. Comparing with experimental results revealed that the four ML-based approaches with input features and appropriate hyper-parameters can predict the shear behavior of a glass FRP-reinforced UHPC I-shaped beam.

  2. 2.

    XGBoost was the most reliable and efficient model with an R2 = 0.99, RMSE = 17.66 KN, and M.A.E. = 13.89 KN among all other machine learning models used in this study.

  3. 3.

    In addition, this study incorporates current design criteria and empirical methods, which were developed to evaluate the shear behavior of glass FRP UHPC I-shaped beams. Consequently, the machine learning model can predict the actual strength with greater accuracy and reliability than conventional design code models such as Eurocode2 and ACI318M-14.

  4. 4.

    The SHAP algorithm was implemented to show the feature dependency of ML results and indicate feature importance in shapely values. It shows that stirrup spacing(s) is the most significant feature among all input parameters.

  5. 5.

    The findings of this study indicated that all of the essential design factors that affect the shear behavior of glass F.R.P. bars-reinforced UHPC I-shaped beams should be considered to improve the empirical model and standard equations' ability to estimate the shear behavior.

  6. 6.

    UHPC is a relatively new material with limited tests done so far; therefore, future limitations and challenges will need further study.