Introduction

Ultra-high-performance concrete (UHPC) is a new development in concrete technology. UHPC is a durable cement-based composite with high tensile and compressive strength [1]. Improved mechanical properties increase the shear strength, flexural strength, and concrete structures’ durability. UHPC is currently utilized in some concrete structures, typically containing precast waffle panels for bridge decks, precast/prestressed bridge girders, and connecting materials between precast concrete deck slabs and beams [2, 3]. In 2001, the USA began using UHPC for highway infrastructure. In addition, replacing normal concrete with UHPC saves materials and reduces labor costs and installation [4]. Regardless, the benefits have not been commonly identified due to the specific needs regarding the material variables employed for producing UHPC blends and the UHPC’s high cost [5].

But getting the right mixture for UHPC sampling is tedious and time-consuming. For this reason, artificial intelligence (AI) has replaced laboratory work to predict the mechanical properties of UHPC [6,7,8]. Machine learning (ML) algorithms have been widely utilized to assess estimative results that nearly match the experiment, like artificial neural networks (ANNs). Nevertheless, an investigation may contain a complete test matrix with many parameters, of which the majority show little assistance to the test results. These computer scientists for developing new selection algorithms according to data-driven models [9]. Demand for software calculating tools in estimating engineering components, systems, and materials continues to grow.

Therefore, ANN has appeared as one of the most popular software computational models successfully used in many engineering problems [10]. In general, ANN has been executed in pattern and character recognition prediction and approximation, classification, image processing, prediction, optimization, and control of corresponding issues. This has prompted investigators have been offered ANN models for solving many civil engineering problems [11,12,13]. Moreover, wide applications in modeling the ANN behavior of specific texture elements have been reported in several studies. Investigator interests have turned to using various ANN models to solve predictive building materials challenges in recent years consisting of concrete, steel, and composite [14,15,16].

Most of the issues corresponding to concrete properties, such as new properties and hardening, have been solved employed ANN models according to the collected observed dataset. In addition, the compressive strength (CS) estimation of concrete by the ANN model is a topic of continuous investigation. This has prompted investigators to use the ANN calculation to evaluate the CS of light-weight, normal-weight, and recycled concrete [17,18,19]. Other researchers have investigated different predictive models to explain high-performance concrete’s compressive strength, employing various ML methods. Subsequently, the emergence of UHPC has caused the fundamental development of the ANN model toward prediction. Investigators have generated ANN to simulate the UHPC performance accurately [20, 21].

Awodiji et al. [22] trained a series of ANN models to examine the relationship between CS and the ratio of material mass to set an age for various hydrated lime-cement concrete. Kasperkiewicz et al. [23] used ANN to optimize silica, cement, fine and coarse aggregates in superplasticizer, and water high-performance concrete (HPC) regardless of data complexity, incompleteness, and consistency predicted an excellent mixing ratio. They showed a significant correlation between the observed actual and estimative values, and ANN models can be used to approximate optimal mixtures. Ghafari et al. [24] studied a backpropagation neural network (BPNN) implementation and statistical mixture designed for estimating the UHPC needed performance. They aim to use BPNN and statistical blend design for assessing the CS and consistency of UHPC in two various curing modes, including primarily wet and steam curing. The 53 concrete samples were designed according to a statistical mixture design sizing matrix, and the components that create the mixture were accepted as separated parameters in the BPNN model. The results showed that BPNN can predict CS and slump more accurately than the statistical mixed method.

Regardless, these black box models provide very little information about what happens during the process of the ANN calculator. Thus, when evaluating the performance of UHPC blends, resolving this equivocation will be the next step in driving the motion to deploy intelligent algorithms while proving it mathematically. Deep ML applications have indicated promising work when optimization strategies during the ANN training phase are exploited to iteratively choose parameters that affect the model's accuracy [22, 25]. Then the selected parameters can be used in ANN or any other intelligent regression algorithm to improve the accuracy of the prediction model while understanding the physical phenomenon behind these selections [26].

UHPC is a material with complex and nonlinear behavior that poses a challenge for modeling using conventional analytical techniques. However, using an Adaptive neuro-fuzzy inference system (ANFIS) can provide a solution for developing a predictive model for the compressive strength (CS) of UHPC. ANFIS can capture the intricate nonlinear relationships between the input variables, such as mixed design parameters, and the output variable, CS. The study aims to present ML models containing ANFIS to identify critical parameters affecting the accuracy of UHPC CS estimation. ANFIS uses a set of fuzzy rules to represent the mapping between the input and output variables. Comprehensive data of multi-parameter experimental results have been compiled from publicly available CS of UHPC analysis. In addition, when combined with the model, three innovative algorithms increase the accuracy of the prediction and reduce the error of the results. Algorithms that form a hybrid model by combining with the corresponding model include generalized normal distribution optimization (GNDO), COOT optimization algorithm (COA), and Honey Badger Algorithm (HBA). The hybrid models’ framework consists of ANGN, ANCO, and ANHB. To evaluate the models, some metrics have been used to select the most appropriate model are discussed in the following sections.

Methods

Dataset

Table 1 shows the constitutive variables of the UHPC samples based on empirical tests from the published paper [27]. In Table 1, the minimum (Min), maximum (Max), average (Mean), and standard deviation (St. Dev.) of the variables are specified that the inputs are cement, silica fume, fly ash, sand, steel fiber, quartz powder, water, and admixture, and the output is compressive strength. In addition, the dataset contains 132 samples, of which 92 belonged to the training and 40 to the testing phase. Also, the distribution of the dataset is indicated in Fig. 1 [28].

Table 1 The properties of data set components engaged in the modeling process
Fig. 1
figure 1

The histogram for the input and output variables

Furthermore, Table 2 shows the correlation between the input and output variables. The values in the matrix indicate a negative correlation between the compressive strength (CS) of UHPC and variables such as C, SF/C, QP/C, and Ad/C. In contrast, a positive correlation exists between CS and variables such as FA/C, S/C, STF/C, and W/C. Moreover, the correlation matrix reveals interesting interdependencies between some independent variables, including a robust negative correlation between C and S/C and a strong positive correlation between S/C and W/C.

Table 2 The correlation between the input and output variables

Adaptive neuro-fuzzy inference system

A fuzzy set consists of elements with different membership levels. The degree of membership offers flexibility in modeling fuzzy collections [29]. Several inference approaches like Mamdani and Sugeno are created for fuzzy rule-based systems [30]. Distinguish the output of fuzzy rule from sharp function. In Sugeno’s system, a typical representation of fuzzy rules is represented by when \({x}_{1}, {x}_{2}, \mathrm{and} {x}_{N}\) are A1, A2, and AN, alternatively, then \(y=f(x)\) here \({A}_{1}, {A}_{2}, ..., \mathrm{and }{A}_{N}\) represents fuzzy sets, and y represents the hash function. The result of each rule is a weighted average used to calculate the results of all the rules and a separate value in this technique. The explanation of a nonlinear map of the system is like a Kanno-type system \(({f}_{FS})\) can be defined as follows:

$${f}_{FS}=\sum_{i}^{N}{w}_{i}{f}_{i}=\frac{\sum_{i}^{N}{h}_{i}{f}_{i}}{\sum_{i}^{N}{h}_{i}}$$
(1)

Here, \(N\) denotes the number of rules and \({h}_{i}\) denotes the membership function of fuzzy collection. From ANFIS, membership functions have been repeatedly determined to produce the correct output. Many membership functions exist, such as Bell, trigonometric, trapezoidal, and Gaussian. The functions of Gaussian membership were employed in this analysis. The function of Gaussian used as

$$f(x,m,s)={e}^{\frac{{-(x-s)}^{2}}{{2m}^{2}}}$$
(2)

In Eq. (2), \(s\) and \(m\) indicate the standard deviation and the dataset’s mean. Training techniques are generally performed via two strategies containing hybrid learning algorithms and backpropagation in the ANFIS methodology (Appendix 1).

Generalized normal distribution optimization

GNDO inspired the theory of normal distribution [31]. A normal distribution can be determined by expecting a random variable x to follow a probability distribution with location parameter (μ) and scale parameter (δ). Its probability density function can be determined as follows:

$$f\left(x\right)=\frac{1}{\sqrt{2\pi \delta }}.\mathrm{exp}\left(-\frac{{(x-\mu )}^{2}}{2\delta }\right)$$
(3)

In Eq. (3), \(x\) indicates a normal random variable, and the normal distribution, \(\mu\) and \(\delta\) show the position of parameters and the scale parameter utilized to define the mean and standard variance of the random variable alternatively. Based on the relationship between the normal distribution and the distribution of individuals within the population in Eq. (4), a generalized normal distribution model can be constructed for optimization:

$${V}_{i}^{t}={\mu }_{i}+{\delta }_{i}\times p, i=1, 2, 3, \dots , N$$
(4)

Here \({V}_{i}^{t}\) represents the tracking vector of the \(i-th\) individual at time t, \({\mu }_{i}\) represents the overall mean position of the i-th individual, \({\delta }_{i}\) represents the generalized standard variance, and \(p\) shows the indicative penalty coefficient. In addition, \({\mu }_{i}\), \({\delta }_{i}\), and \(p\) can be defined as follows:

$${\mu }_{i}=\frac{1}{3}({x}_{i}^{t}+{x}_{\mathrm{best}}^{t}+a)$$
(5)
$${\delta }_{i}=\sqrt{\frac{1}{3}[{({x}_{i}^{t}-\mu )}^{2}+{({x}_{\mathrm{best}}^{t}-\mu )}^{2}+{(a-\mu )}^{2}]}$$
(6)
$$p=\left\{\begin{array}{c}\sqrt{-\mathrm{log}\left({\lambda }_{1}\right)}\times \mathrm{cos}\left(2\pi {\lambda }_{2}\right), if r\le 1\\ \sqrt{-\mathrm{log}\left({\lambda }_{1}\right)}\times \mathrm{cos}\left(2\pi {\lambda }_{2}+\pi \right), if r>1\end{array}\right.$$
(7)

In the above equations, \(r\), \({\lambda }_{1}\), and \({\lambda }_{2}\) represent random numbers between 0 and 1, \({x}_{best}^{t}\) represents the current best position, and \(a\) represents the current average position of the population. Furthermore, the \(a\) is determined as

$$a=\frac{\sum_{i=1}^{N}{x}_{i}^{t}}{N}$$
(8)

Global Exploration finds promising regions in language regions around the world. GNDO's global scan assumes three randomly chosen people, which can be given in Eq. (9):

$${V}_{i}^{t}={x}_{i}^{t}+b\times \left(\left|{\lambda }_{3}\right|\times {V}_{1}\right)+(1-b)\times \left(\left|{\lambda }_{4}\right|\times {V}_{2}\right)$$
(9)

In Eq. (9), \({\lambda }_{3}\) and \({\lambda }_{4}\) represent two random numbers that follow a normal distribution, \(b\) shows the adjustment parameter representing a random number between 0 and 1, and \({V}_{1}\) and \({V}_{2}\) represent two tracking vectors. Alternatively, \({V}_{1}\) and \({V}_{2}\) can be calculated as follows:

$$V_1=\left\{\begin{array}{lc}x_i^t-x_{p1}^t,&if\;f\left(x_i^t\right)<f\left(x_{p1}^t\right)\\x_{p1}^t-x_i^t,&otherwise\end{array}\right.$$
(10)
$$V_2=\left\{\begin{array}{lc}x_{p2}^t-x_{p3}^t,&\;if\;f\left(x_{p2}^t\right)<f\left(x_{p3}^t\right)\\x_{p3}^t-x_{p2}^t,&otherwise\end{array}\right.$$
(11)

Here \(p1\), \(p2\), and \(p3\) show the three random integers selected from 1 to N.

COOT optimization algorithm

Coots are small aquatic birds of the family Rallidae. They form the genus Fulica, which in Latin means "coot". The algorithm begins with a first-order random population of x = {x1, x2, …, xn} [32]. A random population is often evaluated by the target function to determine the target values \(V=\left\{{V}_{1},{V}_{2}, \dots ,{V}_{n}\right\}\). The population is calculated in visible space as

$$P\left(i\right)=r\left(1,d\right)\times \left(ub-lb\right)+lb$$
(12)

where \(P\left(i\right)\) is the coot position, \(d\) shows the dimension of the problem and \(ub, lb\) shows the upper and lower bounds of the search space.

Furthermore, the fitness of each solution must be computed using the objective function Oi = f(x) after the initial population is generated and given the position of each agent. Choosing some coot to be the team leaders. To find a random position based on Eq. (13), move the coot to an arbitrary position in the search room.

$$G=r\left(1,d\right)\times \left(ub-lb\right)+lb$$
(13)

Coot motions explore different parts of the search distance. This movement takes the algorithm out of the local optimal point when the algorithm gets stuck in the local optimal point. The new position of the “coot” is calculated according to Eq. (14)

$$P\left(i\right)=P\left(i\right)+J\times r\times (G-P\left(i\right))$$
(14)

In Eq. (14), \(r\) is a random number between 0 and 1, and \(J\) can be calculated as:

$$J=1-T\times \left(\frac{1}{{Max}_{Iter}}\right)$$
(15)

where \(T\) shows the current iteration and \({\mathrm{Max}}_{\mathrm{Iter}}\) shows the maximum iteration.

Chain development can be performed utilizing the average position of two coots. Another way to realize chain movement is that first calculate the distance vector between the two coots, then bring the larger coot closer together by about half the vector distance. Utilizing the primary strategy, the new position of the coot is calculated as follows:

$$P\left(i\right)=0.5\times (P\left(i-1\right)+P\left(i\right))$$
(16)

where \(P\left(i-1\right)\) shows the second coot.

The remaining coots may have to control their position and approach according to the group leader, and several coots manage the group in front of the group. The idea is to maintain its position depending on the Leader. The Leader’s average position can be considered, and the coot can upgrade it according to this average position. Expecting mean position leads to premature convergence. Using a mechanism for the motion implementation as

$$I=1+(cN)$$
(17)

where \(I\) is the index number of the Leader, \(c\) is the current coot number, and \(N\) is the number of leaders.

Depending on the Leader’s position, the coot should upgrade its position. The coot's next position, according to the chosen Leader, can be determined as follows:

$$P\left(i\right)=p+2\times r\times \mathrm{cos}(2\pi {r}_{1})\times (l-P\left(i\right))$$
(18)

where \(P\left(i\right)\) indicates the coot’s current position, \(p\) determines the chosen Leader’s position, and \({r}_{1}\) shows a random number in the interval [− 1,1].

Groups need to be aligned themselves with their goals, so leaders need to update their positions on purpose. Equation (19) suggests upgrading the leader position as the formula finds a suitable location around the current sweet point. Managers must step away from their best fit to find the correct position. This equation provides a great way to get closer to or farther from the optimal position.

$$p=\begin{Bmatrix}K\times r\times\cos\left(2\pi r_1\right)\times\left(L-p\right)\;if&r<0.5(a)\\K\times r\times\cos\left(2\pi r_1\right)\times\left(L-p\right)-L\;if&r\geq0.5(b)\end{Bmatrix}$$
(19)

where \(L\) indicates the best location found so far, and \(D\) can be determined as

$$K=1-T\times \left(\frac{1}{{Max}_{Iter}}\right)$$
(20)

In addition, the COA pseudo-code has been shown in Algorithm 1.

Honey Badger Algorithm

The HBA imitates the Honey Badger search method [33].

Initialize each position based on the number of badgers (N) as

$${x}_{i}={lb}_{i}+{r}_{1}\times ({ub}_{i}-{lb}_{i})$$
(21)

In Eq. (21), \({x}_{i}\) shows the honey badger’s ith position associated with the candidate solution for the N population, \({r}_{1}\) indicates a random number between 0 and 1, \({lb}_{i}\) and \({ub}_{i}\) determine the explore region’s lower and upper bounds, alternatively.

Intensities included space between prey concentrations and prey and honey badgers. \({I}_{i}\) indicates the intensity of the prey’s odor. The motion is slow, and vice versa when the odor is strong, described by the inverse square law [34] expressed as

$$\begin{array}{l}{I}_{i}={r}_{2}\times \frac{s}{4\pi {S}_{i}^{2}}\\ s={({x}_{i}-{x}_{i+1})}^{2}\\ {S}_{i}={x}_{prey}-{x}_{i}\end{array}$$
(22)

where \({r}_{2}\) shows the random number between 0 and 1, \(s\) indicates the concentration's strength, \({S}_{i}\) is the space between the ith badger and the prey.

The density factor manages time-varying randomness from Exploration to exploitation to allow a smooth transition. In addition, updating the density factor, which decreases with iteration, to account for randomness over time can be determined as follows:

$$a=C\times \mathrm{exp}\left(\frac{-t}{{t}_{\mathrm{max}}}\right)$$
(23)

Here, \(C\) indicates a constant \(\ge 1\) (default = 2) and \({t}_{\mathrm{max}}\) shows a maximum iteration number.

The output of the local optimal step and the position of the agent's actions are used to exit the locally optimal region. In the explore space, the developed algorithm exploits an indicator to change the direction of discovery to benefit from significant opportunities for tight roaming agents.

HBA position \(({x}_{\mathrm{new}})\) update techniques are divided into two sections containing the “mining phase” and the “crypt phase.” During the digging phase, a badger indicates an action similar to cardioid conformation [35]. Cardioid movement can be calculated as follows:

$${x}_{\mathrm{new}}={x}_{\mathrm{prey}}+e\times c\times S\times {x}_{\mathrm{prey}}+e\times {r}_{3}\times a\times {d}_{i}\times \left|\mathrm{cos}(2\pi {r}_{4})\times \left[1-\mathrm{cos}(2\pi {r}_{5})\right]\right|$$
(24)

In Eq. (24), \({x}_{\mathrm{prey}}\) indicates the position of prey, this is the best position found so far, in other words, the best overall position. c ≥ 1 (default = 6) is the ability of the badger to achieve food, \({r}_{3}\), \({r}_{4}\), and \({r}_{5}\) shows various random numbers between 0 and 1, and \(e\) acts as a explore direction change flag, which can be calculated as

$$\mathrm I=\left\{\begin{array}{lc}1&if\;r_6\leq0.5\\-1&else\end{array}\right.$$
(25)

During the digging phase, honey badgers strongly depend on prey odor intensity, space between prey and badger, and time-varying food-influence factors. Additionally, badgers can detect every F sound, making it easier to locate prey when foraging.

If the honey badger follows the honeyguide to achieve the hive, this is presented as:

$${x}_{\mathrm{new}}={x}_{\mathrm{prey}}+e\times {r}_{7}\times a\times {S}_{i}$$
(26)

Algorithm 2 has indicated the HBA pseudo-code.

Performance evaluation methods

Evaluating the performance of the hybrid models during the training and testing sections is an essential step in ensuring that the model performs well against future unpublished datasets in terms of robustness, accuracy, and generalizability. Specifically, statistical metrics can be used to assess the ML model’s error in estimating the target. This paper used the coefficient of determination (R2), mean squared error (RMSE), the median of absolute percentage error (MDAPE), mean absolute error (MAE), and uncertainty 95% (U95) to assess the predictive accuracy of each model in the following:

$${R}^{2}={\left(\frac{{\sum }_{i=1}^{n}\left({p}_{i}-\overline{p }\right)\left({r}_{i}-\overline{r }\right)}{\sqrt{\left[{\sum }_{i=1}^{n}{\left({p}_{i}-\overline{p }\right)}^{2}\right]\left[{\sum }_{i=1}^{n}{\left({r}_{i}-\overline{r }\right)}^{2}\right]}}\right)}^{2}$$
(27)
$$\mathrm{RMSE}=\sqrt{\frac{1}{n}\sum\limits_{i=1}^{n}{\left({r}_{i}-{p}_{i}\right)}^{2}}$$
(28)
$$\mathrm{MAE}=\frac{1}{n}\sum_{i=1}^{n}\left|{p}_{i}-{r}_{i}\right|$$
(29)
$$\mathrm{MDAPE}=\mathrm{medain}\left(\Vert \frac{{r}_{i}-\overline{r}}{{r }_{i}}\Vert \times 100\right)$$
(30)
$$U95=\frac{1.96}{n}\sqrt{\sum_{i=1}^{n}{({r}_{i}-{p}_{i})}^{2}+\sum_{j=1}^{n}{({r}_{j}-{p}_{j})}^{2}}$$
(31)

In the above equations, \(n\) determines the sample number, \({r}_{i}\) and \({p}_{i}\) are actual and predicted values, \(\overline{p }\) and \(\overline{r }\) are the mean values of predicted and actual, alternatively.

Results and discussion

This section will be discussed the results obtained from the model in two parts containing training and testing, in which 70% of the sample evaluation involved the training phase and 30% were assigned to the test. In addition, the models are evaluated and compared with each other to choose the model with the highest accuracy and most minor error. The models were assessed by the evaluators introduced in “Performance evaluation methods” section. Table 3 has been shown the results obtained from the proposed models. The ideal values of the results in the evaluator are that except for R2, the remaining metrics should get the lowest value and close to zero due to the indicating error of models. If the values obtained during the test phase are better than the training, it indicates that the learning of the samples has been done suitable in the training section, which shows the model’s power.

Table 3 The results obtained from the proposed models

In R2, where values are specified as percentages, models should get values close to 100%. As shown in Table 3, the models obtained better values during the testing phase. Comparing between models, it can be seen that ANHB reached the highest value, equivalent to 99.58%, during the test phase, but not much different from the rest of the models. In RMSE, \({\mathrm{ANHB}}_{\mathrm{train}}=2.112\) (MPa) has the lowest value and weakest performance of both parts of ANGN, and the differences between ANHB with ANCO and ANGN were 29% and 43%, respectively. In MDAPE and MAE, the lowest values obtained during the ANHB test phase were equal to 1.153 and 1.845, respectively. Finally, for U95, the lowest value equivalent to 5.901 (MPa) was obtained for the \({\mathrm{ANHB}}_{\mathrm{test}}\), which reached 29 and 43% differences with ANCO and ANGN, alternatively. In general, the strongest to the weakest performance of the evaluation of models in two phases are related to ANHB, ANCO, and ANGN, respectively.

Table 4 compares our present study and previously published articles that explored similar fields. It serves as a reference to assess the performance and workability of our developed hybrid model concerning recent studies. The results from the ANHB model demonstrate its superior ability to predict the compressive strength of UHPC compared to the other models studied.

Table 4 Comparison of present study results with recently published articles with similar datasets

Figure 2 shows the scatter plot in the developed models' training and testing phase. The corresponding figure is based on the R2 and RMSE evaluators, which specify the dispersion and density of points. In addition, the center line is determined in \(X=Y\) coordinates, and the angle between the linear fit and the center line indicates the performance of the models. The points related to ANHB are close to or on the center line, which is not observed in the overestimated or underestimated points. On the other hand, ANGN had more dispersion due to the high RMSE and low R2, and the angle difference was high between the linear fit of ANGN with center compared to other models. In addition, the high density and accuracy of ANHB can be seen in Fig. 3, which shows the comparison between predicted and measured samples. ANHB had a low difference between the predicted and measured. The scatter of points in the training section in ANCO is more due to the low R2 and high RMSE than the test.

Fig. 2
figure 2

The scatter plot in the training and testing phase of developed models

Fig. 3
figure 3

The comparison between predicted and measured samples

Furthermore, it is possible to find a significant difference in some points of the training phase, but improving the performance in the test has minimized it. On the other hand, for ANGN, the dispersion of points in training is such that the points are over and underestimated, and as can be seen in Fig. 3, the difference between predicted and measured is higher than in other models. In general, it can be concluded that the ANHB model has been able to have high accuracy with the density of points and the slight difference between predicted and measured.

Figure 4 presents the scatter error plot for the developed models during the training and testing phases. In an ideal scenario, the error values should be close to zero, indicating accurate predictions. During the ANHB model’s training phase, most predictions exhibited less than 5% errors, signifying its robust performance. However, a few samples, like sample 42, exhibited increased dispersion and were identified as outlier data, as demonstrated in Fig. 5. During the testing phase, the ANHB model showed no particular distribution of errors, and most data points fell within the range of 0%. As a result, the mean error was nearly zero, demonstrating the model's capability to generalize well to unseen data. In the case of the ANCO model, the dispersion of errors increased, leading to the identification of four outliers in both negative and positive ranges. Despite this, the ANCO model significantly improved, reducing its error from 13% during the training phase to 5% during testing. This reduction in error showcased the model’s ability to enhance its performance and better handle diverse datasets. Contrastingly, the ANGN model demonstrated higher error values than the other two models during the training phase, achieving an error rate of 18%. This higher error rate can be attributed to the presence of outlier data. Figure 5 highlighted outlier data points, further underscoring a performance weakness in the ANGN model. However, the ANGN model showed remarkable improvement during the testing phase, outperforming the other two. No outlier data was observed during this phase, and the error rate was reduced to 10%, demonstrating the model's adaptability and ability to overcome its initial limitations.

Fig. 4
figure 4

The scatter error plot of presented models based on the training and testing phase

Fig. 5
figure 5

The error box plot for developed models in the training and testing phase

The scatter error plot provided valuable insights into the models' performance during training and testing. While the ANHB model performed well with some outlier data during training, it demonstrated robustness in testing. The ANCO model improved performance during testing, despite encountering increased dispersion during training. On the other hand, the ANGN model initially suffered from higher errors and outlier data during training but exhibited remarkable improvement and outperformed the other two models during testing. These observations underscore the strengths and weaknesses of each model, guiding future refinements and optimizations for enhanced performance.

Applying the predictive model, comprising ANFIS with GNDO, COA, and HBA, offers practical engineering benefits for UHPC construction.

Summary of benefits:

  1. 1.

    Enhanced quality control: the model estimates UHPC’s CS before testing, optimizing mix designs, and reducing material waste and costs.

  2. 2.

    Improved structural design: accurate CS predictions enable precise structural design, ensuring compliance with safety standards and regulations.

  3. 3.

    Cost and time savings: the model’s predictions reduce the need for physical tests, saving time and resources during construction.

  4. 4.

    Early issue detection: early assessment of CS helps identify and address potential issues in mixed design and curing processes.

  5. 5.

    Optimal construction scheduling: accurate CS predictions facilitate efficient scheduling of construction activities for enhanced project management.

  6. 6.

    Risk mitigation: engineers can assess risks related to specific concrete batches or conditions, making informed decisions to avoid problems.

  7. 7.

    Research advancements: the model supports UHPC research by exploring the effects of different mixtures, additives, or curing methods on CS.

On the other vise, the developed model has several limitations:

  1. 1.

    It is computationally complex and requires significant processing power, making it challenging for real-time applications or low-resource environments.

  2. 2.

    The model’s interpretability is reduced due to incorporating complex algorithms, which can be a concern in domains where transparency is essential.

  3. 3.

    While it performs well on the training data, its generalization to unseen data is limited, presenting a challenge for broader applications.

  4. 4.

    The model's success depends heavily on access to large, high-quality datasets, which can be resource-intensive and difficult to obtain in specific fields.

  5. 5.

    Integrating multiple algorithms introduces numerous hyperparameters that require careful tuning, making the optimization process time-consuming and hindering deployment.

Conclusions

Ultra-High-Performance Concrete (UHPC) is a new development in concrete technology. UHPC is a durable cement-based composite with high tensile and compressive strength. However, getting the right mixture for UHPC sampling is tedious and time-consuming. For this reason, artificial intelligence (AI) has replaced laboratory work to predict the mechanical properties of UHPC. This study aimed to forecast the CS of UHPC employing ANFIS with the most significant influential concrete mixing factors. The results of the research are the following:

  • In R2, the most appropriate value in the training and testing phase belonged to ANHB, which did not differ significantly from the other two models.

  • The lowest RMSE value is related to ANHB, which had a difference of 29% and 43% with ANCO and ANGN, respectively.

  • In MDAPE, the most appropriate value was obtained by ANHB equal to 1.153 MPa, which had a difference of 27 and 32% with ANCO and ANGN, respectively.

  • In MAE and U95, the most appropriate values and other metrics belonged to ANHB and were equaled to 1.845 and 5.901 Mpa, alternatively.

  • Machine learning models are reliable for predicting the mechanical properties of UHPC and can replace laboratory methods to save time and energy.