1. Introduction
Currently, reinforced concrete (RC) buildings are the prevalent choice for constructing civil infrastructure all over the world. They will suffer from various load scenarios throughout their service life [
1]. To maintain their safety, the accurate prediction of the various RC components’ capacities is significant [
2]. Among various RC components, the prediction of the shear strength of an RC deep beam is one of the most-challenging tasks due to its sophisticated failure mechanism [
3,
4]. To predict an RC deep beam’s shear capacity, in the past few decades, plenty of experimental studies have been carried out to figure out its failure process [
5,
6,
7]. It has been found that the classical plane section assumption is no longer satisfied in deep beams, which makes their failure mechanism very complicated.
To resolve this issue, several theoretical models have been proposed, for instance the well-developed strut-and-tie model. Foster proposed the use of brace-and-tie models to describe the mechanisms of the observed non-flexural behavior and presented the rationale behind the empirically determined minimum reinforcement requirements, as well as the minimum mesh reinforcement requirements [
8]. Tuchscherer et al. compiled a database of 868 deep beam shear tests from the literature and fabricated and tested 37 additional deep beam specimens. Through comprehensive analysis of the database, deficiencies in the existing specifications were discovered. Therefore, the authors proposed suggestions for these methods, resulting in an improved brace–tie-modeling procedure [
9] An experimental verification of reinforced concrete deep beams designed by the strut-and-tie method was performed by Starvcev et al. [
10].
However, due to the unavoidable theoretical simplifications and assumptions, the shear capacity of the deep beam is controlled by various complicated behaviors, such as the flexure–shear coupling effect, reinforcement bonding behavior, and the size effect [
11,
12,
13]. Consequently, a single theoretical model cannot consider all the effects of these behaviors, and the corresponding prediction accuracy would not be ideal.
Currently, machine learning algorithms are becoming more and more prevalent in civil engineering [
14,
15,
16,
17]. They have been used to achieve various tasks, such as component and structure behavior prediction, model updating, and traffic load identification [
18,
19,
20]. Chen et al. developed data-driven models using ensemble learning algorithms (GBDT and RF) to accurately predict the bond strength between carbon-fiber-reinforced polymer (CFRP) and steel in CFRP-strengthened steel structures, with the GBDT achieving the highest accuracy (R
2 = 0.98) and potential for structural design and evaluation [
21]. Benbouras et al. designed a new machine learning model, in particular a DNN model, to predict the pile-bearing capacity more effectively and efficiently than traditional methods, providing a user-friendly interface called “BeaCa2021” for researchers and civil engineers [
22]. Czarnecki et al. proposed a hybrid model combining non-destructive methods and neural networks to accurately estimate the subsurface tensile strength of cementitious composites containing waste granite powder, demonstrating its practical suitability for civil engineering applications with error values ranging from 10 to 12% [
23]. In this scenario, machine learning algorithms have been utilized to tackle the prediction of the deep beam’s shear capacity owing to their excellent regression ability and adequate experimental RC deep beam data. Karim et al. utilized a machine learning (ML) algorithm to accurately estimate the shear strength of fiber-reinforced polymer (FRP) reinforced concrete (RC) beams and collected 302 shear test results from the literature to develop the most-effective prediction model [
24]. Sanad and Saka [
25] initially tried using a neural network for predicting the ultimate shear strength of RC deep beams. However, a neural network is usually treated as a black box model, which is uninterpretable by the users. Then, Ashour et al. [
26] and Gandomi et al. [
27] adopted a genetic programming technique to achieve an explicit predicting formula for the deep beam’s shear strength. However, the accuracy of a formula obtained from genetic programming is usually not comparable to the one obtained from the black box machine learning model.
Alshboul et al. proposed a closed-form model based on GEP to predict the shear strength of SFRC deep beams with an R
2 value of 78.9%. The analytical results showed that the GEP model accurately predicted the effects of the concrete strength, the flexural steel percentage, and the ratio of the shear span to the beam depth [
28]. Esteghamati et al. proposed a framework based on training simulation-based seismic and environmental assessments of 720 mid-rise concrete office buildings. The framework aimed to develop generalizable surrogate models to predict the seismic vulnerability and environmental impacts of a class of buildings at specific sites [
29]. Meanwhile, Esteghamati et al. compared knowledge-driven, data-driven, and physics-driven alternative models to assess the relative ability to estimate earthquake losses under complete and incomplete design information scenarios [
30]. Feng et al. [
31] implemented explainable ensemble learning algorithms such as the gradient boosting regression tree (GBRT) to achieve precise shear strength prediction of an RC deep beam. In addition, in the case that the relevant experimental data are not sufficient, Chen and Feng [
32] proposed a method for utilizing multi-fidelity data to train a machine learning model for deep beam shear strength prediction.
Among all these studies, it can be found that, without a doubt, the performance of all these data-driven models would dramatically surpass the formulas used in each design code [
33,
34]. In most cases, boosting ensemble learning algorithms such as the GBRT and extreme gradientboosting (XGBoost) are the most-promising choices. This trend also coincides with the results of famous machine learning competitions such as Kaggle [
35]. However, these models can only provide a deterministic prediction about the shear strength. The aleatory or epistemic uncertainty within training data such as measurement and recording errors is not considered in the models, which could generate large deviations [
36]. Consequently, a scalar prediction without any confidence interval is useless and even harmfulto the structural assessment and design.
To address this shortcoming, Li et al. [
4] proposed a probabilistic prediction model for a deep beam’s shear strength based on Markov chain Monte Carlo (MCMC) calibration. The obtained model could provide probabilistic prediction results in a probabilistic density function (PDF) form. Therefore, the aforementioned uncertainty could be effectively considered. Wu et al. proposed a probabilistic fatigue life prediction model to predict the fatigue life of reinforced concrete beams in a chlorinated environment and took into account statistical uncertainty using Bayesian inference to determine and update the model parameters [
37]. Yu et al. proposed a probabilistic prediction model based on Gaussian process regression (GPR) with an anisotropic composite kernel. This method overcomes the limitations of traditional mechanics and data-driven approaches in rationally quantifying the complex uncertainties regarding the shear strength of reinforced concrete (RC) beam–column joints [
38]. However, the convergence issues within MCMC severely demand the users to have expertise [
39]. To attain convenient probabilistic predictions, a novel algorithm dubbed natural gradient boosting (NGBoost) was proposed and introduced into this field recently [
40,
41].
In this context, in this paper, NGBoost was used for the probabilistic prediction of the shear capacity of deep beams. For the convenience of the application, the Bayesian optimization of NGBoost’s hyper-parameters was adopted. The proposed model was thoroughly examined and compared with the existing prediction formulas and conventional machine learning models. In the following sections, the collected experimental database of the deep beams is presented first. Then, the methodology of the proposed probabilistic prediction approach is introduced. The implementation of the proposed method on the shear capacity prediction is presented next, along with a comparison of the results.
2. Collected Experimental Database
To establish an accurate predicting model of a deep beam’s ultimate capacity, a database with 267 deep beam experiment samples was collected and utilized [
31].
Figure 1 illustrates a typical RC deep beam with the relevant experiment arrangement for measuring its shear capacity. Based on previous studies, the factors influencing a deep beam’s shear capacity include: (1) the geometric factors, such as the span of the beam
L, the shear span
a, the height
h, the effective height
, and the width
b of the beam section; (2) the material factors, such as the concrete strength
, the yield strength of the longitudinal reinforcement
, the yield strength of the horizontal and vertical web reinforcement
; (3) the reinforcement details, such as the longitudinal reinforcement ratio
, the horizontal and vertical web reinforcement ratio
, and the spacing for the horizontal and vertical web reinforcement
. Hence, these 14 factors were selected as the input features for the machine learning prediction models, while the output of the model was the beam’s shear capacity
.
The marginal distribution of the 14 input factors is displayed in
Figure 2. As can be seen, the distribution of all factors covers a wide value range. In addition, the size of the database is 267, which is 10-times larger than the input feature dimension of 14. Therefore, this database is sufficient for establishing a data-driven prediction model via a machine learning algorithm.
3. Methodology
To achieve both accurate mean prediction and a robust probabilistic prediction of the deep beam’s shear capacity, a novel algorithm, NGBoost, was adopted here. The fundamentals of NGBoost are briefly introduced.
Fundamentals of NGBoost
Natural gradient boosting (NGBoost) is a gradient boosting method, which is a machine learning technique for supervised learning. NGBoost uses natural gradients to improve the model performance and convergence speed under the gradient-boosting framework.
The basic idea is to use the classification and regression tree (CART) as the base learner. Each time a new CART is trained, it reduces the error of the existing model:
Specifically, assuming that the existing model is
, the new CART model
tries to approximate the negative gradient of the current model, that is:
where loss represents the loss function.
Then, add
to the existing model to form a new model:
By continuously adding
to approximate the negative gradient direction, the loss can be quickly reduced and the model improved. The innovation of NGBoost is that it uses “natural gradients” instead of numerical gradients. The reason is that numerical gradients are sensitive to parameter scale changes, while natural gradients improve this problem. The calculation formula for natural gradients is:
where
F represents the Fisher information matrix.
Thus, on this background, NGBoost is proposed to solve this issue.
The major improvement of NGBoost is the substitution of the numerical gradient with the natural gradient as the object of training, which is why it is named NGBoost. Compared with numerical gradients, natural gradients have nothing to do with the measurement of the parameter space and are more suitable as optimization goals for model improvement. This allows NGBoost to make probabilistic predictions, not just point estimates. Specifically, NGBoost uses the log-likelihood loss function, and the prediction target is the probability distribution . Each new CART approaches the natural gradient of the current model distribution relative to the true distribution, making the model as a whole approach the true posterior distribution, thereby making probability predictions.
To achieve probabilistic prediction, the prediction result of a model should be a probabilistic distribution
. Therefore, instead of a single scalar prediction, the prediction should be at least a two-dimensional type in order to describe a distribution. As a result, the classic boosting framework cannot be applied because the traditional loss function and numerical gradient are not suitable. To solve this problem, the negative log-likelihood
as the common index for quantifying the difference between two distributions was used to replace the traditional loss function. Furthermore, the natural gradient was used for training
to achieve a prediction of the multi-dimensional type [
42].
With this improvement, NGBoost could achieve probabilistic prediction while maintaining the highly accurate performance of a boosting algorithm.
4. Applying NGBoost in Deep Beam Shear Capacity Prediction
For the probabilistic prediction of a deep beam’s shear capacity, the collected database in
Section 2 was separated into training and testing datasets via a 7:3 ratio. Then, in training NGBoost, just like conventional machine learning models, its hyper-parameters need to be carefully tuned in order to achieve a satisfying performance. In this study, instead of the commonly used grid searching or random searching strategy, the Bayesian optimization algorithm was used, which is more efficient for high-dimensional parameter optimization [
43]. The main idea behind Bayesian optimization is using the Gaussian process to construct a surrogate model of the objective function for optimizing, and this model was used to find the optimal point in an iterative manner. Bayesian optimization usually contains four steps, as shown in
Figure 3.
1. Surrogate model fitting: In this step, the Gaussian process (GP) algorithm is used to fit a surrogate model. The Gaussian process is a non-parametric Bayesian method that captures the uncertainty of the objective function by providing a distribution for each input point. The observed data points are used to train this model to predict function values and uncertainties for unobserved points.
2. Acquisition evaluation: The acquisition function defines whether the current information should be explored or exploited during the next search. Common acquisition functions include the expected improvement (EI), probability of improvement (PI), and upper confidence bound (UCB). These functions help determine the next sampling point by balancing exploration and exploitation.
3. Searching: In this step, the optimal point of the next candidate is found according to the acquisition function. This usually involves an optimization process aimed at finding input values that maximize the value of the acquisition function. This point may be a minimum value of the objective function, or it may be in an area of greater uncertainty, thereby achieving a balance between exploration and utilization.
4. Update: The objective function is evaluated on the newly found candidate points, and then, the surrogate model is updated with this newly observed data point. This updated process typically involves refitting the Gaussian process model to more accurately capture the shape and uncertainty of the objective function.
More-specific details about Bayesian optimization can be consulted in [
43].
Cross-validation (CV) was also adopted in this paper. CV is a statistical method for evaluating the performance of machine learning models. Its main purpose is to ensure that the model has a good generalization ability, that is the model not only performs well on the training data, but also performs well on the test data.
The basic idea of CV is to divide the original dataset into multiple subsets and, then, use one subset as the test set and the remaining subsets as the training set to train and evaluate the model. This process is repeated multiple times, and each subset is used as a test set for validation, ultimately resulting in multiple model performance evaluation results. The average of these results is usually calculated to evaluate the overall performance of the model. In order to take into account the effectiveness of thecomputing performance, this article used a five-fold CV.
After these training processes, the obtained NGBoost model would be further examined in the testing to illustrate its prediction ability on the deep beam’s shear capacity. The training process for NGBoost is summarized in
Figure 4.
5. Shear Capacity Prediction Result Discussion
The optimization process of the hyper-parameters of NGBoost for shear capacity prediction is given in
Figure 5. The optimized hyper-parameters of NGBoost include the base learner’s (CART) maximum depth, the minimum sample number for splitting, the minimum sample number for each leaf, the learning rate, and the CART number for the entire model. It can be seen in
Figure 5a and the parameter searching range given in
Table 1 via Bayesian optimization that the optimal hyper-parameter sets can be efficiently obtained in a high-dimensional parameter space.
Figure 5b also provides the typical optimization contour of two key hyper-parameters: the CART number and the learning rate. The optimal hyper-parameters are listed in
Table 1.
Then, the average prediction performance of NGBoost on the deep beam’s shear capacity was compared with the performance of representative machine learning and ensemble learning algorithms, including linear regression, support vector regression, a neural network, random forest, and XGBoost. These algorithms are mature algorithms that have been realized in the Python software package “Scikit-learn” v1.3.1 [
44].
5.1. Comparison with Representative Machine Learning Algorithm
The performance of each algorithm on the training and testing dataset of the deep beam’s shear capacity experiment is provided in
Figure 6, where the red line represents a perfect prediction. It should be noted that, besides NGBoost, all five compared machine learning models’ hyper-parameters were also carefully optimized by Bayesian optimization, just like the process provided for NGBoost.
From
Figure 6, it can be seen that, except for linear regression, all other models achieved good performance in predicting the shear capacity. Their performance varied a little. Moreover, NGBoost could achieve comparable performance on the scalar prediction of the shear capacity of a deep beam in comparison with the excellent ensemble learning algorithm, XGBoost. Furthermore, to better quantify the performance of different models, the indexes the determination coefficient R
2 and root-mean-squared error (RMSE) were adopted to illustrate the overall fitting accuracy and absolute error of each model.
From
Table 2, we can observe the error performance of different machine learning models in predicting the shear load capacity. Among all the listed models—linear regression (LR), support vector regression (SVR), neural network (NN), random forest (RF), XGBoost, and NGBoost—NGBoost showed the best prediction performance. First, it can be noted that NGBoost reached 0.9045 in the coefficient of determination (R
2), which was the highest among all models, indicating that the NGBoost model has high relevance and accuracy in predicting actual values. Secondly, the NGBoost model also performed well in terms of the root-mean-squared error (RMSE), reaching the lowest value of 38.7976 kN, which further confirmed the accuracy and reliability of the model in predicting shear bearing capacity.
In comparison, the LR and SVR models performed more generally in terms of the R2 and RMSE, especially the linear regression model, whose R2 was only 0.8302. The RMSE was 51.7434 kN, which is relatively high, indicating that its prediction accuracy and reliability were low.
NN, RF, and XGBoost also showed good performance, but still failed to surpass NGBoost. For example, although the neural network reached 0.9003 on the R2, its RMSE was still 39.6392 kN, which was higher than NGBoost. Similarly, XGBoost’s R2 was 0.8976, and the RMSE was 40.1715 kN, which also failed to surpass NGBoost.
Taken together, the NGBoost model performed most prominently in terms of the error in shear bearing capacity prediction. Its higher coefficient of determination and lower root-mean-squared error proved that it was superior to the other models. The NGBoost model demonstrated higher prediction accuracy and reliability.
5.2. Comparison with Empirical Formulas
To further illustrate the superiority of the NGBoost in shear capacity prediction, several empirical formulas globally adopted in current design codes were used for comparison. In this section, the prediction formulas from the codes in China, the U.S., Canada, and Europe were selected [
45,
46,
47,
48], as listed here:
Chinese Guobiao (GB) code for the design of concrete structures:
American Concrete Institute (ACI) building code requirements for structural concrete:
This formula is derived from the strut-and-tie model, in which is the bottom width of the strut; are the loading and supporting area width; is the angle of the strut; and is the strut coefficient.
Canadian Standards Association (CSA) design of concrete structures:
Similarly, this formula is also derived from the strut-and-tie model, where and is the ultimate tensile strain of the tie.
Eurocode (EU) design of concrete structures:
where
is the strut width and equals to
.
Subsequently, the prediction results of these four formulas on the identical testing set of deep RC beams can be calculated and is presented in
Figure 7. For better comparison, the mean prediction result of NGBoost is also provided here. Firstly, it can be seen that, owing to the purpose of maintaining the reliability level of the designed structure, all formulas were more likely to underestimate the shear capacity of a deep beam as the prediction point falls below the red line in
Figure 7. Among the four formulas, though the CSA code’s formula had the best overall regression accuracy, its performance was still evidently worse than that of NGBoost based on the distribution of the scatters given in
Figure 7.
Similarly, to better quantify the accuracy of each formula, their R
2 and RMSE scores are listed in
Table 3. It clear that the NGBoost model’s prediction accuracy on the deep beam’s shear capacity dramatically surpassed all the representative prediction formulas derived from relevant mechanics. The improvement ratio of NGBoost over the best formula from CSA reached a 66% R
2 and 53% for the RMSE. These results strongly prove the effectiveness and superiority of using NGBoost for the scalar prediction of a deep beam’s shear capacity.
5.3. Probabilistic Prediction Results
Besides the excellent scalar prediction ability of NGBoost on the deep beam’s shear capacity, its probabilistic prediction ability is the other advantage of NGBoost. As mentioned in
Section 3, the NGBoost algorithm can predict the probabilistic density distribution of the deep beam’s shear capacity. In this study, a normal distribution type was selected, and thus, NGBoost can output a mean value
and a standard deviation
about the deep beam’s shear capacity. On this basis, a confidence interval can be provided for the users to make decisions. As given in
Figure 8, apart from the mean prediction on the shear capacity, a 98% confidence interval was provided for the entire database based on
. Although the mean prediction accuracy of NGBoost was high enough, with the help of a confidence interval, a robust prediction result would be more helpful for the subsequent reliability-based structural design and assessment.
To better illustrate the probabilistic ability of NGBoost for the deep beam’s shear capacity, four deep beam samples were selected, and the corresponding prediction results are given in
Figure 9. According to
Figure 9, the mean prediction accompanied by the predicted probabilistic density function is shown, while the actual shear capacity is pointed out. From these four samples, it can be demonstrated that NGBoost can adaptively consider the uncertainty within the given database and provide the corresponding probabilistic results on the deep beam’s shear capacity. In comparison with the classic machine learning algorithms, which can only attain scalar prediction, the probabilistic prediction result could provide the uncertainty about the shear capacity. Consequently, the users could be more confident about the corresponding decisions based on pre-defined target reliability levels. For a higher reliability level, a wider confidence interval would be obtained. As a result, the actual shear capacity value would be contained in the interval, and the following design and assessment result would be more reliable. With the help of NGBoost, the data-driven prediction result could actually facilitate this field.
6. Conclusions
The prediction of an RC deep beam’s shear capacity is still an important, but difficult task in civil engineering due to the complicated failure mechanism behind it. In order to solve this task and maintain the safety of the relevant structures, a data-driven prediction approach based on machine learning algorithms has been tried recently, but these models can only provide a scalar prediction result about the shear capacity. Because these machine learning models are usually uninterpretable, a scalar prediction is untrustworthy and cannot be used in practice. Under this circumstance, a probabilistic prediction approach of the shear strength of reinforced concrete deep beams was established based on the newly developed NGBoost algorithm. The performance of the proposed model was thoroughly compared with classic machine learning models and existing empirical formulas. In addition, its probabilistic prediction ability was illustrated. Based on the results, the following conclusions can be reached:
- (1)
Though the advantage of NGBoost lies in its probabilistic ability, the mean prediction accuracy of NGBoost on the deep beam’s shear capacity was still better than the accuracy of the commonly used machine learning algorithms including the advancing ensemble learning algorithms such as random forest and XGBoost.
- (2)
In comparison with the prediction formulas for the deep beam shear capacity used in the relevant codes, NGBoost’s prediction performance dramatically surpassed all the prediction formulas. These results revealed the effectiveness of using NGBoost in predicting the shear capacity of a deep beam.
- (3)
NGBoost can successfully provide a probabilistic prediction about the shear capacity of a deep beam via a probabilistic density function. Based on this probabilistic density function, the confidence level of the deep beam’s strength can be easily obtained by the users. Consequently, a more-robust design and assessment of the deep beam can be conducted. The issue of the prediction of a machine learning algorithm being untrustworthy can be avoided. Overall, the proposed NGBoost-based probabilistic prediction approach could effectively provide an accurate and robust result about the deep beam’s shear capacity.
Although the model in this article achieved results that exceeded other research results, there are still some issues worthy of further study. To attain a probabilistic prediction via NGBoost, the type of probabilistic distribution should be pre-set, which requires some prior knowledge about the targeted issue, such as the shear capacity prediction in this study. If the type pf distribution is not appropriate, the probabilistic prediction results of NGBoost may not be satisfactory. Therefore, in the future, NGBoost should be further modified to be able to adaptively construct the probabilistic distribution from the data, attaining more applicability.