Next Article in Journal
Effect of Natural Weathering on the Mechanical Strength of Bamboo Bio-Concrete
Previous Article in Journal
Assessment of Health-Oriented Layout and Perceived Density in High-Density Public Residential Areas: A Case Study of Shenzhen
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Concrete Creep Prediction Based on Improved Machine Learning and Game Theory: Modeling and Analysis Methods

School of Civil Engineering, Architecture and the Environment, Hubei University of Technology, Wuhan 430068, China
*
Author to whom correspondence should be addressed.
Buildings 2024, 14(11), 3627; https://doi.org/10.3390/buildings14113627
Submission received: 17 October 2024 / Revised: 10 November 2024 / Accepted: 11 November 2024 / Published: 14 November 2024
(This article belongs to the Section Building Materials, and Repair & Renovation)

Abstract

:
Understanding the impact of creep on the long-term mechanical features of concrete is crucial, and constructing an accurate prediction model is the key to exploring the development of concrete creep under long-term loads. Therefore, in this study, three machine learning (ML) models, a Support Vector Machine (SVM), Random Forest (RF), and Extreme Gradient Boosting Machine (XGBoost), are constructed, and the Hybrid Snake Optimization Algorithm (HSOA) is proposed, which can reduce the risk of the ML model falling into the local optimum while improving its prediction performance. Simultaneously, the contributions of the input features are ranked, and the optimal model’s prediction outcomes are explained through SHapley Additive exPlanations (SHAP). The research results show that the optimized SVM, RF, and XGBoost models increase their accuracies on the test set by 9.927%, 9.58%, and 14.1%, respectively, and the XGBoost has the highest precision in forecasting the concrete creep. The verification results of four scenarios confirm that the optimized model can precisely capture the compliance changes in long-term creep, meeting the requirements for forecasting the nature of concrete creep.

1. Introduction

Creep significantly affects concrete’s mechanical properties. It provides a reference to evaluate the mechanical features of concrete structures under long-term loading, which is the basis of the mechanical response of concrete under this condition [1,2]. The generation of creep increases the deflection of the concrete and the loss of prestressing force, causing a break in the structure and reducing the performance of the construction [3,4,5,6]. Concrete creep arises from the expansion of cracks in the transition zone at the interface between mortar and aggregate within the concrete; cracks are first generated in the transition region between coarse aggregate and mortar. The cracks are not merely related to the acceptable deformation of the cementitious material itself, but also to the uneven shrinkage caused by the internal drying of the concrete; simultaneously, the internal restraint generated by creep can also slow down the development of microcracks [7,8,9,10,11].
Factors exerting an influence on creep have been extensively investigated, such as the Young’s modulus intrinsic to the cementitious material itself, the water-to-binder ratio, the curing temperature, the relative humidity, and the age of loading of the concrete [12,13,14,15,16,17]. Recently, scholars have attempted to decouple the effects of various variables from different perspectives to make more accurate predictions of creep under complex conditions. The interaction of these variables and the high sensitivity of creep with respect to these parameters pose an even greater challenge for large-scale measurements. In an effort to further surmount the impact that hydration reactions and the interaction of input variables exert on the results of a creep experiment during a loading procedure, Wyrzykowski et al. [18] partially substituted unhydrated cement with inert quartz powder, and then conducted uniaxial compression creep experiments to calibrate the mechanical properties of the non-aging system. Ladaoui et al. [19] carried out tests on the essential creep of four kinds of high-efficiency concrete under various temperatures and found that the basic creep coefficients at the temperature increases of 20 °C and 50 °C were 2 and 3. Gan et al. [20] investigated the impact of factors such as the water–cement ratio and stress level on creep measurements by means of the micro-cantilever bending method, indicating that microscopic variations affect the overall creep. However, they are confined to decoupling a finite number of variables, thereby restricting the precise prediction of creep in actual loading.
With the extensive development of creep research, a series of experiment-based analysis models, semi-empirical models considering the coupling effects of different variables, and classical models have been applied to creep studies. Bažant and Wu [21] proposed the utilization of the Maxwell chain model to represent the rate-dependent creep law of concrete. Zhang and Hamed [22] presented the rate-dependent creep law. Nikola Tošic’ et al. [23] analyzed the difference in creep flexibility coefficients between RAC and NAC under the same conditions using the MC2010 model based on the establishment of a database on the creep of recycled aggregate concrete (RAC) and natural aggregate concrete (NAC). Jian Cao et al. [24,25] investigated the consequences of different fly ash contents and loading ages on the creep recovery of fly ash concrete using the modified B4 model. Cherif Belebchouche et al. [26] investigated the creep behavior of self-compacting concrete beams under different parametric mix proportions by means of the B3, ACI209, EC02, and CEB MC 90 models. The above models provide an effective alternative to experiments for the in-depth study of the creep mechanism under the influence of multiple parameters. However, they assume ideal conditions, and there are inevitable accuracy errors in actual engineering, while MC2010, MC09, EC02, and other models only consider a limited number of input characteristics, and the B3 and B4 models need to be modified under actual engineering conditions according to a large number of experimental data on the model coefficients when considering the creep of admixed concrete, self-consolidating concrete, etc.
ML models have been extensively utilized in fields such as rubberized concrete frost resistance [27], chloride diffusion in concrete [28], the compressive strength of silica fume concrete [29], lateral confinement factors of reinforced concrete columns [30], and the mechanical features of recycled aggregate concrete [31]. ML models provide new directions for creep prediction. The Support Vector Machine (SVM) [32,33,34] can utilize kernel functions instead of the inner product of mapping to a high-dimensional space, avoiding dimensionality catastrophes and providing good robustness. Random Forest (RF) [35,36] has characteristics including controllable generalization error, a strong ability to handle high-dimensional eigenvector space, resistance to overfitting, and fewer hyperparameters for optimization, but there are still some limitations. Extreme Gradient Boost (XGBoost) [37,38] introduces the second-order partial derivatives of the loss function for weighting based on the CART regression tree, characterized by high accuracy and speed. However, XGBoost has the disadvantages of being prone to falling into local optimal solutions and high requirements for learning samples. The prediction ability of the above model is influenced by hyperparameters. Intelligent optimization algorithms can search for optimal hyperparameters to upgrade the prediction precision and robustness of machine learning (ML) models. The Snake Optimization (SO) [39] algorithm aims to search for the optimal solution by simulating the foraging behavior of snake groups. However, the standard SO algorithm has problems with the slow convergence speed in the early stage and becoming trapped in local optimal solutions. Therefore, three improved strategies are proposed based on the snake optimizer to promote the algorithm’s generalization ability in optimizing various model parameters.
In previous studies, feature importance, LIME, PDP, etc., have often been used for the interpretability analysis of models. However, these methods cannot stably explain the prediction analysis of complex models. The Shapley theory was first proposed in 1953 [40,41]. It has the characteristics of strong additivity and high stability and has been widely used in concrete strength prediction [42,43], concrete penetration resistance [44], and concrete early performance analyses [45].
In this study, the author proposes four improvement strategies for the SO algorithm to enhance its capability to leap out of local extremums. On this basis, three ML models (SVM, RF, and XGBoost) for creep prediction are developed according to the NU creep–shrinkage database. Finally, based on the SHAP method, the influence of single features and cross-features on prediction results is explored.

2. Data Processing Analysis

From 2011 to 2013,the NU database was expanded from the RILEM database by Professor Bažant’s team at Northwestern University with the support of the U.S. Department of Transportation. This expansion was based on collating more than 1400 creep test curves and 30,000 data points from many journal articles, conference records, and other documents [46,47]. Based on consolidation theory, this paper studies 4207 sets of experimental data in the NU database to construct the ML model. Thirteen parameters are chosen as model inputs: water–cement ratio (w/c), aggregate–cement ratio (a/c), cement content ((C) Kg/m3), cement type, 28-day compressive strength (fc28 MPa), surface-to-volume ratio (V/S), sample height (mm), loading temperature (T °C), experimental humidity (RH_Test °C), loading stress ((Sigma) MPa), strength–stress ratio (sigma/fct0), time of loading or time exposed to the environment ((t′) (days)), and loading time ((t) days). Compared with other numerical parameters, in this study, the cement type is treated as a classification parameter of the input variable. In the classification of cement types, rapid-hardening cement (R), slow-hardening cement (SL), and rapid-hardening high-strength cement (RS) are, respectively, characterized as 1, 2, and 3 for classification. The measured creep compliance (Jcreep (μm/m/MPa)) in the experiment is taken as the output variable. Table 1 describes the database employed for creep prediction, encompassing the minimum, maximum, mean, and derivative values (as depicted in Figure 1).

3. Methodology

The traditional grid search algorithm and random search algorithm for complex models with high dimensions and large amounts of data need to enumerate the as-summed values of each hyperparameter combination to optimize the model. Meanwhile, due to the shortcomings of the low convergence rate and local extremum of the snake optimizer, there is often a large error when enumerating hyperparameter combinations. Not only is the optimization efficiency low, but the optimization results are often unsatisfactory. In consideration of the limitations of the aforementioned optimization methods, this study puts forward the use of the HSOA optimization method to find the hyperparameters of models such as SVM, RF, and XGBoost during the cross-validation process to improve the model’s prediction efficiency. Finally, based on the optimized model, the Shapley value is calculated to characterize the final impact of different input variables on creep compliance.

3.1. Machine Learning Models

3.1.1. Support Vector Machine Models

The SVM was presented by Corinna Cortes and Vladimir Vapnik in 1995 as a generalized linear classifier [48,49]. A SVM uses interval maximization as its learning approach to transform linear problems into minimized regularized loss functions. Hence, the SVM’s learning method provides the optimum algorithm to solve convex quadratic programming. The parameters that need to be optimized under cross-validation are (1) the penalty coefficient (C), (2) the regularization parameter (penalty), (3) the kernel function parameter (sigmoid), and (4) the classification decision parameter (decision_function_shape).

3.1.2. Random Forest Model

A Random Forest (RF) is a tree-based ensemble model proposed by Breiman in 2001 [50]. The Random Forest (RF) adopts a random feature selection strategy [51]. In the sample selection stage, the Random Forest algorithm selectively and repeatedly samples k samples from the original dataset to construct weak learners, and the output result is the voting average of multiple weak learners [52,53]. The random selection of features implies that a limited number of features are utilized to split nodes during each round of data training, ensuring that all decision trees are uncorrelated with each other; the final prediction of the RF model is the mean of all predictions made by the decision trees.
The hyperparameters that need to be optimized for the Random Forest model are (1) the maximum depth of the decision tree (max_depth); (2) the minimum number of leaf nodes (min_leaf_nodes); (3) the number of decision trees (n_estimators) (the final forecasting of the RF model is the mean of all predictions made by the decision trees); and (4) the max feature subset parameter (max_features).

3.1.3. Extreme Gradient Boosted Tree Model (XGBoost)

Extreme Gradient Boosting creates a strong learner through training multiple weak learners, incorporating second-order derivatives to increase the loss function’s accuracy, regular terms to prevent overfitting, and Blick storage for parallel computing. The optimal model is built by minimizing the data’s loss function [54]. A shrinkage strategy is introduced to avoid overfitting the loss function, where the residual values fitted by each weak learner are multiplied by eta in the range [0–1]. The expression of the loss function is shown in Equation (1) [38]:
L = min f F 1 N i = 1 N L ( y i , f ( x i ) )
where X i and y i describe the input and output values, respectively; f ( x i ) describes the ith weak learner; and N represents the total number of weak learners. To reduce the model’s complexity, Equation (1) can be rewritten after introducing J f [38]:
L = min f F 1 N i = 1 N L ( y i , f ( x i ) ) + λ J ( f )
Among them, λ is the structural risk coefficient parameter. After this is introduced, the objective function is as follows [38]:
o b j ( θ ) = i n L ( y i , y i ) + k = 1 K Ω ( f k )
The objective function O b j   is divided into two terms, where the right and left terms indicate the model’s training error and regularization. The XGBoost model contains multiple CARTs, defining the complexity of each tree as follows [38]:
Ω ( f ) = γ T + 1 2 λ ω 2
where T denotes the total number of leaf nodes, ω denotes the leaf node vector modulus, and γ is a model hyperparameter. A second-order Taylor expansion of the objective function can be obtained [38,55]:
f ( x + x ) = l ( y i , y i k 1 ) + g i f k ( x i ) + 1 2 h i f k ( x i ) 2
Assume that the output of the k-integrated tree models is the following [38,55]:
y ^ i = ϕ ( x i ) = k = 1 K f k ( x i )
where y ^ i   represents the prediction results of the first i decision trees for the sample. The objective function is described as follows [56]:
o b j = i = 1 n l ( y i , y ^ i ( t 1 ) + f t ( x i ) ) + Ω ( f t )
After performing a second-order Taylor expansion and omitting the k − 1 rounds of pseudo-residuals, the objective function becomes the following [56]:
o b j = i = 1 n ( L ( y ^ 1 + f t ( x i ) i , y i ) + g i f i ( x i ) + 1 2 h i f t 2 ( x i ) ) + Ω ( f t )
g i = 𝜕 L ( y ^ i t 1 ) 𝜕 L ( y ^ t 1 )
h i = 𝜕 2 L ( y ^ i t 1 , y i ) 𝜕 L ( y ^ t 1 ) 2
where g i and h i denote the first- and second-order derivatives.
To minimize the objective function, four hyperparameters of the XGBoost model need to be iteratively optimized: (1) max_depth; (2) the learning rate of the model (learning_rate); (3) the minimum sum of weights for node splitting (min_child_weight); and (4) the structural risk coefficient parameter. At the same time, the node splitting coefficient (gamma) of the objective function is optimized.

3.2. Snake Optimization Algorithm

The snake optimizer [39] is a recently proposed nonlinear, multi-objective population optimization algorithm inspired by snakes’ foraging, fighting, and mating behavior. The SO algorithm is divided into a food search phase, a fighting phase, and a mating phase. However, the snake optimizer has problems with slow convergence in the early stage and quickly falling into the local optimal solution, so four optimization methods are used: chaotic mapping replaces the population initialization stochastic distribution function, optimizing the food search phase of the algorithm with an average subtraction optimization strategy; a reverse learning strategy is utilized to expand the algorithm’s search range to avoid falling into the local optima; and the adaptive τ-distribution perturbation approach is used to promote the population’s iterative diversity and ensure the local search efficiency of the algorithm.

3.2.1. Food Search

Food (Q) and temperature (T) influence the behavioral patterns of snakes. Female and male snakes are uniformly distributed in a particular spatial area, and when the temperature is low and the total amount of food (Q) < 0.25, females and males randomly search for food and update their position information according to each other’s location. When the temperature (T) is less than the 0.6% threshold, females and males enter fighting or mating modes; otherwise, they focus on feeding.
At this stage, the population-position update equation becomes the following [39]:
X i , j ( t + 1 ) = X f o o d ± C 3 × T e m p × r a n d × ( X f o o d X i j ( t ) )
where X i j ( . ) and X f o o d indicate the position of an individual (male or female) and the best individuals, respectively, and C 3   is a constant.

3.2.2. Combat Phase

When Temp < 0.6% threshold, males fight each other for the location of females; the fighting ability of the snake depends on the amount of food searched for in the food search phase. At this stage, due to the insufficient search range of the algorithm in the local space, there is significant differentiation in the combat capability of the male snakes, causing the algorithm to fall into the local optimum. The following is the updated formula for the optimal position of males [39]:
X i , m ( t + 1 ) = X i , m ( t ) + C 3 × F M × r a n d × ( Q × X b e s t , f X i , m ( t ) )
where X i , m ( . )   and X b e s t . f   represent the position of the ith male and the best female individuals in the population, respectively, and F M indicates the male snake’s fighting capability.

3.2.3. Mating Patterns

During the mating phase, the female snakes assume a dominant position. They preferentially search for locations with a cold climate and an abundance of food as the optimal breeding spots. As the snakes’ search ability is affected by both temperature and food, the snake optimizer can easily fall into the local optimal solution due to the overly small search range. The equation for the update of the female position is the following [39]:
X i , f ( t + 1 ) = X i , f ( t ) + C 3 × M m × r a n d × ( Q × X i , f ( t ) X i , m ( t ) )
where X i , f ( . ) indicates the position of the ith female in the population, and M m indicates the male and female mating capabilities, respectively.

3.3. Improved Algorithms

3.3.1. Chaotic Mapping

Chaotic mapping is a complex nonlinear state with irregularity, randomness, and ergodicity, often used to generate chaotic sequences. Bernoulli chaotic mapping has the characteristic of a more uniform distribution; in this study, this approach is used instead of the population random distribution formula of the algorithm so that the population search is more efficient. Expressions such as [57]:
Z n + 1 = Z n 1 β 0 Z n 1 β Z n ( 1 β ) β 1 β Z n 1

3.3.2. Average Subtraction Optimization Strategy

The SABO algorithm presents a novel computational notion called subtraction of search agents B and A, defined as follows [58]:
A ν B = sign ( F ( A ) F ( B ) ) ( A ν B )
ν   is an m-dimensional vector, including random numbers between [1, 2], and F(A) and F(B) describe the objective functions corresponding to search agents A and B, respectively. This strategy calculates the displacement of an individual search agent   X i in the search space by taking the arithmetic average of ν   subtraction of the position information of female snakes ( X i , f ) and male snakes ( X i , m ). Reducing the total amount of food (Q) and the influence of temperature (T) on the population’s random search promotes the algorithm’s global search ability. So, Equation (11) becomes:
X b e s t ( i , j ) = X i ( t ) + ( ( r a n d ( 1 : dim ) . * ( X i ( t ) ν X i ( X i f , X i , m ) ) ) . / ( N m )
where X b e s t ( . ) and X i , ( t ) represent the positions of the best and ith individuals in the population at the ith iteration, respectively.

3.3.3. Reverse Learning Strategy

The reverse learning strategy is a strategy for particles to find an optimal solution at the current location for other regions [59], which can enlarge the SO algorithm’s search scope to prevent it from falling into the local optimal solution. In a given space, an individual (p) of height h is projected onto the x-axis to obtain the globally optimum position X b e s t ,   and a convex lens with focal length f is located at the origin; imaging through a convex lens gives a P * with a height of h * in the X-axis by imaging the resulting reverse optimal position X * b e s t . Thus, the equation can be defined as follows [39,59]:
X * b e s t ( t ) = ( a j + b j ) 2 + ( a j + b j ) 2 n X b e s t ( t ) n
a j and   b j denote the current solution’s upper and lower bounds in the j-dimensional dimension. Thus, Equations (12) and (13) become the following [39,59]:
X b e s t ( i , j ) = X m i , j + C 3 * F M * r a n d * Q * X * b e s t _ f t X m i , j 1 2
X b e s t i , j = X m i , j + C 3 * M f * r a n d * Q * X * b e s t _ f t X f i , j 1 2

3.3.4. Adaptive τ - Distribution Perturbation Approach

The population diversity reduces in the latter iterations of the SO. This leads to the algorithm having a tendency to become trapped in the local optima. Thus, an adaptive t-distribution dimension-by-dimension mutation approach is proposed for perturbing the individuals with the optimum fitness in the food search, combat, and mating stages, respectively, improving the SO algorithm’s capability to jump out of local optima. When t n N 0 , 1 ,   t = 1 = C ( O , 1 ) , where N (0, 1) and C (0, 1) describe the Gaussian and Cauchy distributions, respectively. The two boundaries of the t-distribution are Gaussian distribution and Cauchy distribution; these possesses strong local exploitation ability based on maintaining population diversity, which guarantees the population’s convergence rate in the later step. The formula transforms as follows [60]:
X i b e s t ( i , j ) = X best ( i , j ) + τ ( C _ i t e r ) × X b e s t ( i , j )
where iter describes the current number of iterations; t(iter) describes the t-distribution with the degree of freedom t. Figure 2 presents the flowchart of the HSOA.
In this paper, four assessment indices, including the coefficient of determination (R2), mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE), are used to evaluate the efficiency of different models. MAE expresses the mean value of the residuals of creep flexibility and creep flexibility at the sample points. RMSE denotes the average degree of residuals at the sample points. Although R2 does not maximize the residuals with the database variance, it more accurately expresses the differences in the efficacy of each model. The expressions for the four metrics are as follows [61,62]:
R 2 = k = 1 N ( q o , k q ¯ 0 ) ( q t , k q ¯ t ) K = 1 N ( q 0 , k q ¯ 0 ) 2 K = 1 N ( q , t , k q ¯ t ) 2
M A E = 1 N ( K = 1 N | q 0 , k q t , k q 0 , k | )
R M S E = 1 N K = 1 N ( q 0 , k q t , k ) 2
M A P E = 100 % N K = 1 N | q t q 0 q 0 |
where N is the total sample size, q 0 , k indicates the real value corresponding to k samples, q 0 ¯ denotes the mean value, q t , k denotes the sample output value, q t ¯ denotes the sample output mean value, and k = 1/N.

4. Results and Discussion

The training platform was hosted using a 12th Gen Intel(R) Core(TM) i7-12700H 2.30 GHz(TM), 16.0 GBRAM, and an NVIDIA GeForce RTX 3060 with 16G GPU. The software configuration is as follows: MATLAB 2024b and Python 3.9.
The purpose of the computational analysis lies in validating the predictive capabilities of the three models, namely HSOA-SVM, HSOA-RF, and HSOA-XGBoost. HSOA will optimize the hyperparameters of each ML model using five-fold cross-validation. The calculation process also uses cross-validation operation on 70% of the training data and 30% of the test data, and the outcome is the mean of the five-fold cross-validation.

4.1. Model Prediction Results

4.1.1. Optimal Hyperparameter Settings for ML Models

Table 2 lists the hyperparameters optimized for each ML model, and the value of each parameter is the optimal value based on multiple iterations. The population quantity and the number of iterations is determined using the experimental approach.

4.1.2. Six Model Prediction Results

After determining the three models’ optimum hyperparameters using the HSOA method, 4207 data separated in advance from the database are utilized to evaluate the ML model’s predictive efficiency and generalization capability. The results of five-fold cross-validation are shown in Table 3. From the validation results, it can be seen that all six ML models attained relatively high precision, and the accuracies of the three models, namely SVM, RF, and XGBoost, on the test set witnessed increases of 9.927%, 9.58%, and 14.1%, respectively. Simultaneously, the accuracies of the optimized model on the training set and the test set exhibit only a subtle difference, indicating that the HSOA effectively reduces the overfitting of the model. Figure 3 presents the metric radar plots for the six models’ training and testing sets. The radar chart intuitively reflects the differences in four evaluation indicators between the original model and the optimized model in training set a and testing set b.
Figure 4 describes the prediction outcomes of the original models SVM (a), RF (b), and XGBoost (c), as well as the optimized models HSOA-SVM (d), HSOA-RF (e), and HSOA-XGBoost (f) on the test set. Analyzing and comparing the six models’ regression prediction outcomes in Figure 4, it is seen that the predicted values of the improved models have a smaller overall difference from the actual values, and the prediction accuracy is higher, demonstrating that the optimized model has considerable anti-overfitting ability and most of the data points have sufficient accuracy, even though there are some large discrete values in all three improved models.

4.1.3. Improved Model Performance Analysis

Residual analysis is an essential index for appraising the efficiency of diverse models. The residuals of different nonlinear regression models are closely related to the optimization algorithms and the mean and dispersion of the models themselves. The computational residuals of the three optimization models are shown in Figure 5. The particles in the residual plot of HSOA-XGBoost are more evenly distributed on both sides of the central axis, and the number of abnormal particles is significantly lower than that of HSOA-RF and HSOA-SVM. This is because, compared with the SVM and RF models, XGBoost effectively enhances the base learner due to the optimization of the λ coefficient. Meanwhile, Figure 6 shows the three ML models’ residual interval distributions. The HSOA-XGBoost model’s residual interval distribution is more reasonable, indicating higher prediction stability.
Figure 7 Depicts the curves of the absolute errors of the three models versus the experimental values, and this curve presents an overview of the three models’ overall errors. The minimum and maximum values of the absolute errors of the three models are all between 1 and 4, respectively, indicating that all three improved models have reached a high-precision prediction level. This is also largely consistent with the residual ranges of the three models. At the same time, HSOA-XGBoost performs optimally in particle regression, residual analysis, and absolute error.
Although the HSOA-XGBoost model shows high prediction performance, the attention given to the black box mechanism behind the ML model prediction has been insufficient. Therefore, it is crucial to further explain the logic of HSOA-XGBoost.

4.2. Feature Importance Analysis Using SHAP

4.2.1. Characteristic Importance Analysis

To interpret the ML model, its output is manifested through a linear summation of its input features multiplied by the corresponding SHAP values. The exclusive solution for the SHAP value is derived by weigthing the average over all possibilities and summing all possible combinations of the eigenvalues on this basis [63].
ϕ i = S N \ { i } | S | ! ( | N | | S | 1 ) ! | N | ! ( ν ( S { i } ) ν ( S ) ) ) .
where S is the number of non-zero indices, | S | ! N S 1 ! N !   represents the weight of an index, ν S { i } ν S represents the marginal contribution of I to S , and S N { i } represents summing over all possible indices.
The SHAP value for feature importance is regarded as the average predicted value of the sample for that feature minus the average predicted value of the other specimens. The SHAP value for each feature is shown in Figure 8, and the features are ranked according to their average SHAP value. In the compliance flexibility experiments, the loading time is the most crucial parameter for forecasting the creep compliance of concrete, the water–cement ratio is the second most significant variable, the cement content ranks third, the compressive strength follows closely behind, and input characteristics such as loading stress and strength–stress ratio also play a vital role. Although input variables such as stress level, temperature, relative humidity, and the weight ratio of aggregate to cement are placed behind other characteristic elements, this does not mean that these characteristic elements cannot exert significant functions in the creep process.
Figure 9 shows the boundary of the feature values separating the negative and positive contributions for the five influential features. In this paper, the impact of the top five features explained by the global SHAP value on the final forecasting of creep compliance is compatible with the ranking given by feature importance, and is also consistent with the characteristic factors of creep in the curing theory. Meanwhile, the linear incorporation of SHAP values can contribute to all features. For a fixed base value, the creep flexibility’s final prediction result is the offset outcome of the SHAP values with positive and negative contributions, and for the following text force plot is used to visualize this offset process.

4.2.2. Single Factor Analysis

To gain a further understanding of how HSOA-XGBoost can deliver accurate predictions, it is essential to further analyze the SHAP values of the five influential input variables. It is evident from Figure 10 that the distribution of SHAP values exhibits a specific development tendency along with the variation in parameter values. The influence rules of the six parameters on the prediction of HSOA-XGBoost are as follows: The prolongation of the creep age and the increase in cement content and water–cement ratio can significantly increase the creep prediction value. Meanwhile, at the low stress level, the internal of the concrete specimen will cause the surface aggregation of aggregate particles due to compressive stress, which will result in the reduction in the creep effect. At a high stress level, the increase in stress will significantly increase the creep prediction result. Concurrently, high-level stress causes the local materials to present nonlinear viscoelastic behaviors, exposing concrete to the interaction between creep and fatigue, eventually leading to creep damage. A growth in compressive strength will reduce the creep prediction result. The above phenomenon is in line with the creep hydration law in consolidation theory.

4.2.3. Explanation of Local Features

The overall influence of the five most influential factors on creep prediction was addressed in the preceding text. However, as the temperature increases, the cementitious material’s Young’s modulus, ultimate yield strength, and linear expansion coefficient change. This will cause a change in the capillary water content to affect the generation and development of microcracks. The variation in the water–cement ratio will affect the viscosity and exudation amount of water in concrete. Meanwhile, the sliding of cement gel sheets between the concrete gel water-absorbing layers and draining the water layer within the cement gel sheet can have a significant effect on the essential creep deformation. Thus, for a fixed base value of 67.23, the final forecasting of creep flexibility is the outcome of the interaction of the characteristic variables. Consequently, there exists a compelling need to delve deeper into the impact exerted by diverse input variables on creep prediction.
As shown in Scenario 1 in Figure 11a, the predicted result is greater than the base value, which may be due to the higher temperature, the increase in the activation energy of water between the water molecules of the gel material, the increase in the infusibility of water, the reduction in the water content inside the gel material, and the reduction in the instantaneous elastic stiffness and delayed decreasing in the stiffness of the concrete. Under load, the ability to resist stress relaxation is reduced at 65% relative humidity, resulting in a significant escalation of the creep deformation magnitude.
In contrast to Scenario 1, the predicted outcomes presented for Scenario 2 in Figure 11b fall short of the benchmark value. The meager compressive strength along with the elevated relative humidity constitute significant negative correlative factors in the prediction of creep compliance. Under the humidity condition of RH_Test = 99%, the increase in sliding water molecules in the gel makes it more likely for interlayer sliding to occur, and simultaneously, the gel pore spacing is more prone to being transformed into the space between adjacent layers. This results in a relatively increased specific surface area of the concrete, lowering the difficulty of interlayer sliding of the matrix and diminishing the creep modulus of the concrete.
In contrast to Scenario 1 and Scenario 2, the high compressive strength in Scenario 3 (as shown in Figure 11c) offers a positive correlation for creep prediction. In Scenario 3, the water–cement ratio exhibits the highest negative correlation, and the water–cement ratio and stress are key factors in the creep of early-age concrete. Under a low initial water–cement ratio, the creep modulus increases rapidly and nonlinearly as the degree of hydration grows. This implies that concrete, which contains a relatively high water–cement ratio during the hydration process, possesses a higher creep compliance.

4.3. Exemplification of General Scenarios

To illustrate the efficiency of the improved model in predicting the concrete’s creep nature, the author selected four typical cases from the test set and predicted the creep compliance within the range of 1 to 2190 days. Table 4 presents the detailed circumstances of the four distinct schemes (Scenario S1–Scenario S4). Scenario S1 and Scenario S2 are creep experiments commencing from the concrete ages of 28 (t0) and 7 days (t0), with the creep durations being 200 and 553 days, respectively. Scenario S3 and Scenario S4 are creep experiments starting at the concrete ages of 14 and 7 days, respectively, and the creep periods are 1315 and 2190 days, respectively. In four typical scenarios, the author uses the optimized XGBoost model and the B4 model for creep prediction. Table 5 presents the formula of the B4 model. This model considers the influences of relevant factors such as loading age, cement type, water–cement ratio, aggregate–cement ratio, 28-day compressive intensity, and environmental temperature when forecasting the creep nature of the concrete. Meanwhile, it has received theoretical mathematical support from nanostructure micro-stress relaxation theory, water molecule diffusion theory, and microcrack damage model. Therefore, the B4 model can effectively predict the creep compliance changes in concrete under different time spans.
Where t ^ and t ^ are the current age adjusted for temperature and age at loading adjusted for temperature; J ( t ^ , t ^ ) represents total creep;   q 1 represents instantaneous compliance; C 0 ( t ^ , t ^ ) indicates the essential compliance function; and C d ( t ^ , t ^ , t ˜ 0 ) indicates the additional creep compliance caused by drying. Q ( t ^ , t ^ ) indicates a binomial integral. Q f ( t ^ ) , Z ( t , t ) , and r ( t ^ ) describe the time-dependent parameters obtained through calculation. q 2 ~ q 5   are parameters related to the water–cement ratio (w/c), aggregate–cement ratio (a/c), and cement type (C). R T   is a variable related to creep activation energy and temperature. τ 0 is the shrinkage factor of concrete, ε s h ( t ˜ 0 ) is ultimate shrinkage strain, and τ s h is shrinkage strain. P 1 - P 5 , P 2 w - P 5 w , P 3 a - P 5 a , P τ a , P τ w , P τ c , p 5 ε , τ c e m , and P 5 H are the coefficients related to the type of cement.
Figure 12 shows the forecast outcomes of the HSOA-XGBoost and B4 models for the experimental values of creep compliance in four scenarios. The fitting curves of HSOA-XGBoost in the four scenarios are consistent with the experimental results. In four typical scenarios, although the B4 model shows high accuracy in the early hydration process, it still fails to capture the creep compliance under long-term creep. At the same time, the creep prediction value of the B4 model in Scenario S1–Scenario S4 is generally greater than the actual value observed in the experiments. This changing tendency of the B4 model is even more pronounced in Scenario S3 and Scenario S4 within a time span where the experimental data have relatively extensive coverage. The reason for this phenomenon is that the B4 model is inadequate in depicting the rapid changes in the physical state of the material under the influence of the hydration effect at the stage of creep, thus resulting in a larger calculated result for the parameters.

4.4. Limitations and Discussion

As presented in Figure 12, the predicted creep curve based on HSOA-XGBoost is non-smooth. This is a result of the nature of the input database. Even though each calculation for each set of experimental data in the figure is accomplished under the identical mixture parameters and experimental characteristic conditions, creep experiments persist for several months or years, so errors in the experimental outcomes are unavoidably present. Concurrently, ML models are highly dependent (R2 = 0.708) on the integrity of the input data. Incomplete and scattered data points will influence the fitting effect of ML models for compliance creep test values. Although the HSOA optimization strategy has been proposed to improve the robustness and generalization ability of models such as SVM, RF, and XGBoost to reduce the influence of outliers, the data in the NU database are scattered points combined under various experimental conditions. Therefore, before training the data, interpolation processing of missing values needs to be performed using logarithmic functions to improve the integrity of the database. However, unstructured databases may pose certain challenges to prediction. Therefore, two possible measures need to be adopted to improve the applicability of ML models in creep prediction:
  • Improving the quality of the database can be carried out by collecting more creep data under different experimental conditions, by considering more factors such as the shape of the test block (prism or cylinder) and the mechanical state during the experiment (three-point bending test or axial compression test) that will have an influence on creep.
  • Prior to the training of an ML model, the data within the database should be preprocessed to eliminate the impacts of outliers and noise on the model training process. One prevalent approach is to substitute the missing values with the predicted values derived from a specific functional model. However, this method entails certain risks.

5. Conclusions

In this study, three machine learning models were established based on the introduction of the HSOA, and the creep behavior of concrete was predicted. Firstly, the NU database was preprocessed and divided into the training and test datasets. The five-fold cross-validation method was adopted to enhance the model’s prediction precision, and the ML prediction model’s generalization ability was verified based on 1262 sets of creep data. Subsequently, the predictions of the HSOA-XGBoost model were explicated considering the SHAP theory. The effects of diverse input variables on creep prediction were analyzed and accounted for by means of common creep theories. In the end, the efficiency of the B4 and HSOA-XGBoost models was evaluated through four scenarios. The subsequent conclusions can be stated:
  • Through the optimization of the HSOA and cross-validation, all three optimized models achieved a relatively high level of accuracy. In the test dataset, HSOA-XGBoost demonstrated higher precision, with R2 reaching 0.908, 0.926, and 0.968, respectively. HSOA-XGBoost exhibited a more robust performance. The fitting ability for the creep experimental values is far superior to that of the widely used B4 model.
  • The SHAP theory offered a rational explanation of the ML model and provided five input features exerting a considerable influence on the prediction of concrete creep: (1) creep age; (2) loading stress; (3) cement type; (4) water–cement ratio; and (5) compressive strength. The five most impactful input features revealed by the SHAP theory were fundamentally in line with the influencing factors in the creep theory.
  • As the ML model considers the balance between bias and variance, the prediction results of the HSOA-XGBoost model in long-term creep are close to the experimental observed values. Meanwhile, the creep law captured by the HSOA-XGBoost model is consistent with the general creep law in the consolidation theory, which verifies the rationality of the ML model. Given the scattered and imperfect data in the NU database, the creep compliance curve is non-smooth. Further supplementation of the database may contribute to more complete predictions. Although the HSOA optimizes the hyperparameters of models such as SVM, RF, and XGBoost in multiple iterations, due to the limitations of the prediction principle of ML models themselves and the limitations of the search principle of the SO algorithm, 100% prediction accuracy does not exist.

Author Contributions

Conceptualization, W.L. and K.M.; methodology, W.L.; software, W.L. and H.L.; validation, W.L., C.L., and K.M.; formal analysis, W.L.; investigation, W.L. and H.L.; resources, K.M. and W.L.; data curation, C.L.; writing—original draft preparation, W.L.; writing—review and editing, C.L. and H.L.; visualization, W.L. and C.L.; supervision, H.L. and K.M.; project administration, W.L. and H.L; funding acquisition, H.L. and K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used in the article can be obtained from the corresponding author.

Acknowledgments

The authors would like to acknowledge Hubei University of Technology.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

SVM: Support Vector Machine (supervised binary classification machine learning model). RF: Random Forest (integrated machine learning models based on decision trees). XGBoost: Extreme Gradient Boosting (integrated model based on the gradient boosting framework). SHAP: SHapley Additive exPlanations (interpretable method for machine learning models). SO: Snake Algorithm (a meta-heuristic algorithm based on snake population). HSOA: Hybrid Snake Optimization Algorithm (the Snake Algorithm improved by the author). MC2010: Mode Code 2010 (European code for concrete structure design). MC90: Mode Code 90 (European code for concrete structure design). B3: B3 stress–strain model for concrete (description of the stress–strain constitutive model of concrete in the process of stress). B4: Bažant–Baweja Non-linear Creep and Shrinkage Model (describes the constitutive model of shrinkage and creep under sustained loading).

References

  1. Zhou, H.; Fan, X.; Liu, Y. Dynamic Serviceability and Safety Reliability Analysis of Aging PC Girder Bridges with Non-Prestressed Reinforcement Considering Concrete Shrinkage, Creep and Stochastic Vehicle Load Flows. Structures 2024, 64, 106515. [Google Scholar] [CrossRef]
  2. Asamoto, S.; Kato, K.; Maki, T. Effect of Creep Induction at an Early Age on Subsequent Prestress Loss and Structural Response of Prestressed Concrete Beam. Constr. Build. Mater. 2014, 70, 158–164. [Google Scholar] [CrossRef]
  3. Zhang, S.; Hamed, E. Application of Various Creep Analysis Methods for Estimating the Time-Dependent Behavior of Cracked Concrete Beams. Structures 2020, 25, 127–137. [Google Scholar] [CrossRef]
  4. Zhou, H.; Fan, X.; Liu, Y.; Lu, D. Time-Variant Reliability Analysis of Simply Supported PC Girder Bridges Considering Shrinkage, Creep, Resistance Degradation and Vehicle Load Flows. Structures 2023, 56, 104885. [Google Scholar] [CrossRef]
  5. Xie, F.; Chang, T.; Zhang, C.; Zhang, F.; Kuang, G.; Zhang, Y. Medium and Short-Term Effective Prestress Losses Considering Multiple Factors: Laboratory and on-Site Beam Experiments. Structures 2024, 68, 107166. [Google Scholar] [CrossRef]
  6. Yang, Q.; Yao, D.; Yan, J.; Liu, M.; Fan, F.; Fu, Z. Analysis of Time-Varying Mechanical Properties of Prestressed Concrete Containment during the Tensioning Process and Service Considering the Influence of Creep. Int. J. Press. Vessel. Pip. 2023, 204, 104962. [Google Scholar] [CrossRef]
  7. Denarié, E.; Cécot, C.; Huet, C. Characterization of Creep and Crack Growth Interactions in the Fracture Behavior of Concrete. Cem. Concr. Res. 2006, 36, 571–575. [Google Scholar] [CrossRef]
  8. Di Luzio, G.; Cusatis, G. Solidification-Microprestress-Microplane (SMM) Theory for Concrete at Early Age: Theory, Validation and Application. Int. J. Solids Struct. 2013, 50, 957–975. [Google Scholar] [CrossRef]
  9. Mahesh, S.; Anoop, M.B.; Sasmal, S. Creep and Fracture of UHSC—A Microindentation Study. Constr. Build. Mater. 2024, 427, 136173. [Google Scholar] [CrossRef]
  10. Dummer, A.; Neuner, M.; Hofstetter, G. An Extended Gradient-Enhanced Damage-Plasticity Model for Concrete Considering Nonlinear Creep and Failure Due to Creep. Int. J. Solids Struct. 2022, 243, 111541. [Google Scholar] [CrossRef]
  11. Tong, T.; Hua, G.; Liu, Z.; Liu, X.; Xu, T. Localizing Gradient Damage Model Coupled to Extended Microprestress-Solidification Theory for Long-Term Nonlinear Time-Dependent Behaviors of Concrete Structures. Mech. Mater. 2021, 154, 103713. [Google Scholar] [CrossRef]
  12. Guo, W.; Wei, Y. Investigation of Compressive Creep of Calcium-Silicate-Hydrates (C-S-H) in Hardened Cement Paste through Micropillar Testing. Cem. Concr. Res. 2024, 177, 107427. [Google Scholar] [CrossRef]
  13. Cao, J.; Zeng, P.; Liu, T.; Tu, B. Influence of Mineral Powder Content and Loading Age on Creep Behavior of Concrete Members under Axial Compression. Results Eng. 2023, 19, 101304. [Google Scholar] [CrossRef]
  14. Frech-Baronet, J.; Sorelli, L.; Charron, J.P. New Evidences on the Effect of the Internal Relative Humidity on the Creep and Relaxation Behaviour of a Cement Paste by Micro-Indentation Techniques. Cem. Concr. Res. 2017, 91, 39–51. [Google Scholar] [CrossRef]
  15. Hu, Z.; Wyrzykowski, M.; Griffa, M.; Scrivener, K.; Lura, P. Young’s Modulus and Creep of Calcium-Silicate-Hydrate Compacts Measured by Microindentation. Cem. Concr. Res. 2020, 134, 106104. [Google Scholar] [CrossRef]
  16. Frech-Baronet, J.; Sorelli, L.; Chen, Z. A Closer Look at the Temperature Effect on Basic Creep of Cement Pastes by Microindentation. Constr. Build. Mater. 2020, 258, 119455. [Google Scholar] [CrossRef]
  17. Tamtsia, B.T.; Beaudoin, J.J.; Marchand, J. The Early Age Short-Term Creep of Hardening Cement Paste: Load-Induced Hydration Effects. Cem. Concr. Compos. 2004, 26, 481–489. [Google Scholar] [CrossRef]
  18. Wyrzykowski, M.; Scrivener, K.; Lura, P. Basic Creep of Cement Paste at Early Age—the Role of Cement Hydration. Cem. Concr. Res. 2019, 116, 191–201. [Google Scholar] [CrossRef]
  19. Ladaoui, W.; Vidal, T.; Sellier, A.; Bourbon, X. Effect of a Temperature Change from 20 to 50 °C on the Basic Creep of HPC and HPFRC. Mater. Struct. /Mater. Constr. 2011, 44, 1629–1639. [Google Scholar] [CrossRef]
  20. Gan, Y.; Rodriguez, C.R.; Schlangen, E.; van Breugel, K.; Šavija, B. Assessing Strain Rate Sensitivity of Cement Paste at the Micro-Scale through Micro-Cantilever Testing. Cem. Concr. Compos. 2021, 121, 104084. [Google Scholar] [CrossRef]
  21. Bazant, Z.P.; Wu, S.T. Rate-type creep law of aging concrete based on maxwell chain. Mater. Struct. 1974, 7, 45–60. [Google Scholar] [CrossRef]
  22. Do, Q.H.; Bishnoi, S.; Scrivener, K.L. Microstructural Modeling of Early-Age Creep in Hydrating Cement Paste. J. Eng. Mech. 2016, 142, 04016086. [Google Scholar] [CrossRef]
  23. Tošić, N.; de la Fuente, A.; Marinković, S. Creep of Recycled Aggregate Concrete: Experimental Database and Creep Prediction Model According to the Fib Model Code 2010. Constr. Build. Mater. 2019, 195, 590–599. [Google Scholar] [CrossRef]
  24. Cao, J.; Tu, N.; Liu, T.; Han, Z.; Tu, B.; Zhou, Y. Prediction Models for Creep and Creep Recovery of Fly Ash Concrete. Constr. Build. Mater. 2024, 428, 136398. [Google Scholar] [CrossRef]
  25. Bažant, Z.P.; Jirasek, M.; Hubler, M.H.; Carol, I. RILEM Draft Recommendation: TC-242-MDC Multi-Decade Creep and Shrinkage of Concrete: Material Model and Structural Analysis*: Model B4 for Creep, Drying Shrinkage and Autogenous Shrinkage of Normal and High-Strength Concretes with Multi-Decade Applicability. Mater. Struct. /Mater. Constr. 2015, 48, 753–770. [Google Scholar] [CrossRef]
  26. Chabane, A.; Belebchouche, C.; Bensebti, S.E.; Czarnecki, S.; Boukezzoula, A. Comparison of the Accuracy of Regulation Models for Self-Compacting Concrete Creep Evaluation. J. Build. Eng. 2022, 59, 105069. [Google Scholar] [CrossRef]
  27. Huang, X.; Wang, S.; Lu, T.; Wu, K.; Li, H.; Deng, W.; Shi, J. Frost Durability Prediction of Rubber Concrete Based on Improved Machine Learning Models. Constr. Build. Mater. 2024, 429, 136201. [Google Scholar] [CrossRef]
  28. Liu, W.; Liu, G.; Zhu, X. Applicability of Machine Learning Algorithms in Predicting Chloride Diffusion in Concrete: Modeling, Evaluation, and Feature Analysis. Case Stud. Constr. Mater. 2024, 21, e03573. [Google Scholar] [CrossRef]
  29. Xiao, S.; Cheng, H.; Que, Z.; Liu, T.; Zou, D. Enhancing Marine Anti-Washout Concrete: Optimal Silica Fume Usage for Improved Compressive Strength and Abrasion Resistance. Constr. Build. Mater. 2024, 428, 136262. [Google Scholar] [CrossRef]
  30. Naeej, M.; Bali, M.; Naeej, M.R.; Amiri, J.V. Prediction of Lateral Confinement Coefficient in Reinforced Concrete Columns Using M5′ Machine Learning Method. KSCE J. Civ. Civil. Eng. 2013, 17, 1714–1719. [Google Scholar] [CrossRef]
  31. Miao, X.; Zhu, J.X.; Zhu, W.B.; Wang, Y.; Peng, L.; Dong, H.-L.; Xu, L.Y. Intelligent Prediction of Comprehensive Mechanical Properties of Recycled Aggregate Concrete with Supplementary Cementitious Materials Using Hybrid Machine Learning Algorithms. Case Stud. Constr. Mater. 2024, 21, e03708. [Google Scholar] [CrossRef]
  32. Sánchez, V.D.A. Advanced Support Vector Machines and Kernel Methods. Neurocomputing 2003, 55, 5–20. [Google Scholar] [CrossRef]
  33. Gold, C.; Sollich, P. Model Selection for Support Vector Machine Classification. Neurocomputing 2003, 55, 221–249. [Google Scholar] [CrossRef]
  34. Baudat, G.; Anouar, F. Feature Vector Selection and Projection Using Kernels. Neurocomputing 2003, 55, 21–38. [Google Scholar] [CrossRef]
  35. Biau, G.; Fr, G.B. Analysis of a Random Forests Model. J. Mach. Learn. Res. 2012, 13, 1063–1095. [Google Scholar]
  36. Breiman, L. Bagging Predictors; Kluwer Academic Publishers: New York, NY, USA, 1996; Volume 24. [Google Scholar]
  37. Ali, Z.A.; Abduljabbar, Z.H.; Taher, H.A.; Sallow, A.B.; Almufti, S.M. EXtreme Gradient Boosting Algorithm with Machine Learning: A Review. Acad. J. Nawroz Univ. 2023, 12, 320–334. [Google Scholar] [CrossRef]
  38. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar]
  39. Hashim, F.A.; Hussien, A.G. Snake Optimizer: A Novel Meta-Heuristic Optimization Algorithm. Knowl. Based Syst. 2022, 242, 108320. [Google Scholar] [CrossRef]
  40. Chen, H.; Covert, I.C.; Lundberg, S.M.; Lee, S.-I. Algorithms to Estimate Shapley Value Feature Attributions. Nat. Mach. Intell. 2023, 5, 590–601. [Google Scholar] [CrossRef]
  41. Rozemberczki, B.; Watson, L.; Bayer, P.; Yang, H.-T.; Kiss, O.; Nilsson, S.; Sarkar, R. The Shapley Value in Machine Learning. arXiv 2022, arXiv:2202.05594. [Google Scholar]
  42. Wang, Z.; Liu, H.; Amin, M.N.; Khan, K.; Qadir, M.T.; Khan, S.A. Optimizing Machine Learning Techniques and SHapley Additive ExPlanations (SHAP) Analysis for the Compressive Property of Self-Compacting Concrete. Mater. Today Commun. 2024, 39, 108804. [Google Scholar] [CrossRef]
  43. Karim, R.; Islam, M.H.; Datta, S.D.; Kashem, A. Synergistic Effects of Supplementary Cementitious Materials and Compressive Strength Prediction of Concrete Using Machine Learning Algorithms with SHAP and PDP Analyses. Case Stud. Constr. Mater. 2024, 20, e02828. [Google Scholar] [CrossRef]
  44. Abdulalim Alabdullah, A.; Iqbal, M.; Zahid, M.; Khan, K.; Nasir Amin, M.; Jalal, F.E. Prediction of Rapid Chloride Penetration Resistance of Metakaolin Based High Strength Concrete Using Light GBM and XGBoost Models by Incorporating SHAP Analysis. Constr. Build. Mater. 2022, 345, 128296. [Google Scholar] [CrossRef]
  45. Nazar, S.; Yang, J.; Wang, X.E.; Khan, K.; Amin, M.N.; Javed, M.F.; Althoey, F.; Ali, M. Estimation of Strength, Rheological Parameters, and Impact of Raw Constituents of Alkali-Activated Mortar Using Machine Learning and SHapely Additive ExPlanations (SHAP). Constr. Build. Mater. 2023, 377, 131014. [Google Scholar] [CrossRef]
  46. Bazant, Z.P.; Li, G.-H. Comprehensive database on concrete creep and shrinkage. ACI Mater. J. 2008, 105, 635–637. [Google Scholar]
  47. Hubler, M.H.; Wendner, R.; Bažant, Z.P. Comprehensive Database for Concrete Creep and Shrinkage: Analysis and Recommendations for Testing and Recording. ACI Mater. J. 2015, 112, 547. [Google Scholar] [CrossRef]
  48. Padierna, L.C.; Carpio, M.; Rojas-Domínguez, A.; Puga, H.; Fraire, H. A Novel Formulation of Orthogonal Polynomial Kernel Functions for SVM Classifiers: The Gegenbauer Family. Pattern Recognit. 2018, 84, 211–225. [Google Scholar] [CrossRef]
  49. Li, Y.S.; Peiyi, P.H. SVM classification: Its contents and challenges. Appl. Math.-A J. Chin. Univ. 2003, 18, 332–342. [Google Scholar]
  50. Gupta, D.L.; Malviya, A.K.; Singh, S. Performance Analysis of Classification Tree Learning Algorithms. Int. J. Comput. Appl. 2012, 55, 1–6. [Google Scholar]
  51. Zhou, Q.; Zhou, H.; Li, T. Cost-Sensitive Feature Selection Using Random Forest: Selecting Low-Cost Subsets of Informative Features. Knowl. Based Syst. 2016, 95, 1–11. [Google Scholar] [CrossRef]
  52. Zhang, P.F.; Iqbal, M.; Zhang, D.; Zhao, X.L.; Zhao, Q. Bond Strength Prediction of FRP Bars to Seawater Sea Sand Concrete Based on Ensemble Learning Models. Eng. Struct. 2024, 302, 117382. [Google Scholar] [CrossRef]
  53. Cai, R.; Han, T.; Liao, W.; Huang, J.; Li, D.; Kumar, A.; Ma, H. Prediction of Surface Chloride Concentration of Marine Concrete Using Ensemble Machine Learning. Cem. Concr. Res. 2020, 136, 106164. [Google Scholar] [CrossRef]
  54. Charbuty, B.; Abdulazeez, A. Classification Based on Decision Tree Algorithm for Machine Learning. J. Appl. Sci. Technol. Trends 2021, 2, 20–28. [Google Scholar] [CrossRef]
  55. Zou, Z.; Wang, B.; Hu, X.; Deng, Y.; Wan, H.; Jin, H. Enhancing Requirements-to-Code Traceability with GA-XWCoDe: Integrating XGBoost, Node2Vec, and Genetic Algorithms for Improving Model Performance and Stability. J. King Saud. Univ. Comput. Inf. Sci. 2024, 36, 102197. [Google Scholar] [CrossRef]
  56. Lin, L.; Li, S.; Wang, K.; Guo, B.; Yang, H.; Zhong, W.; Liao, P.; Wang, P. A New FCM-XGBoost System for Predicting Pavement Condition Index. Expert. Syst. Appl. 2024, 249, 123696. [Google Scholar] [CrossRef]
  57. Brown, R.; Chua, L.O. Clarifying chaos ii: Bernoulli chaos, zero lyapunov exponents and strange attractors. Int. J. Bifurc. Chaos 1998, 8, 1–32. [Google Scholar] [CrossRef]
  58. Trojovský, P.; Dehghani, M. Subtraction-Average-Based Optimizer: A New Swarm-Inspired Metaheuristic Algorithm for Solving Optimization Problems. Biomimetics 2023, 8, 149. [Google Scholar] [CrossRef]
  59. Ouyang, C.; Zhu, D.; Qiu, Y. Lens Learning Sparrow Search Algorithm. Math. Probl. Eng. 2021, 2021, 9935090. [Google Scholar] [CrossRef]
  60. Zhang, W.; Sheng, L.; Chunhui, R. Mixed Strategy Improved Sparrow Search Algorithm. J. Comput. Eng. Appl. 2021, 57, 74. [Google Scholar] [CrossRef]
  61. Ji, X.C.; Chen, R.S.; Lu, C.X.; Zhou, J.; Zhang, M.Q.; Zhang, T.; Yu, H.L.; Yin, Y.L.; Shi, P.J.; Zhang, W. Recent Advances in Machine Learning for Defects Detection and Prediction in Laser Cladding Process. Next Mater. 2025, 7, 100404. [Google Scholar] [CrossRef]
  62. Veza, I.; Deniz Karaoglan, A.; Akpinar, S.; Spraggon, M.; Idris, M. Machine Learning of Weighted Superposition Attraction Algorithm for Optimization Diesel Engine Performance and Emission Fueled with Butanol-Diesel Biofuel. Ain Shams Eng. J. 2024, in press. [CrossRef]
  63. Sun, D.; Ding, Y.; Wen, H.; Zhang, F.; Zhang, J.; Gu, Q.; Zhang, J. SHAP-PDP Hybrid Interpretation of Decision-Making Mechanism of Machine Learning-Based Landslide Susceptibility Mapping: A Case Study at Wushan District, China. Egypt. J. Remote Sens. Space Sci. 2024, 27, 508–523. [Google Scholar] [CrossRef]
Figure 1. Histogram of the distribution of input variables.
Figure 1. Histogram of the distribution of input variables.
Buildings 14 03627 g001
Figure 2. Indicators for Model Evaluation.
Figure 2. Indicators for Model Evaluation.
Buildings 14 03627 g002
Figure 3. Radar charts of the training (a) and testing sets (b) for the HSOA-SVM, HSOA-RF, and HSOA-XGBoost models’ performance.
Figure 3. Radar charts of the training (a) and testing sets (b) for the HSOA-SVM, HSOA-RF, and HSOA-XGBoost models’ performance.
Buildings 14 03627 g003
Figure 4. The regression results of the test sets for (a) SVM; (b) RF; (c) XGBoost; (d) HSOA-SVM; (e) HSOA-RF; and (f) HSOA-XGBoost.
Figure 4. The regression results of the test sets for (a) SVM; (b) RF; (c) XGBoost; (d) HSOA-SVM; (e) HSOA-RF; and (f) HSOA-XGBoost.
Buildings 14 03627 g004aBuildings 14 03627 g004b
Figure 5. Residual analysis plots for (a) HSOA-SVM; (b) HSOA-RF; and (c) HSOA-XGBoost models.
Figure 5. Residual analysis plots for (a) HSOA-SVM; (b) HSOA-RF; and (c) HSOA-XGBoost models.
Buildings 14 03627 g005
Figure 6. Residual distribution plot for HSOA-SVM, HSOA-RF, and HSOA-XGBoost.
Figure 6. Residual distribution plot for HSOA-SVM, HSOA-RF, and HSOA-XGBoost.
Buildings 14 03627 g006
Figure 7. Comparing the errors between the testing sets of (a) HSOA-SVM; (b) HSOA-RF; and (c) HSOA-XGBoost.
Figure 7. Comparing the errors between the testing sets of (a) HSOA-SVM; (b) HSOA-RF; and (c) HSOA-XGBoost.
Buildings 14 03627 g007
Figure 8. Aggregated HSOA-XGBoost-based concrete creep SHAP.
Figure 8. Aggregated HSOA-XGBoost-based concrete creep SHAP.
Buildings 14 03627 g008
Figure 9. Global SHAP values using the HSOA-XGBoost model.
Figure 9. Global SHAP values using the HSOA-XGBoost model.
Buildings 14 03627 g009
Figure 10. Results of importance analysis using SHAP. (a) Time since loading (days). (b) Cement content (Kg/m3). (c) Water–cement ratio. (d) Loading stress (MPa). (e) Compressive strength (MPa).
Figure 10. Results of importance analysis using SHAP. (a) Time since loading (days). (b) Cement content (Kg/m3). (c) Water–cement ratio. (d) Loading stress (MPa). (e) Compressive strength (MPa).
Buildings 14 03627 g010
Figure 11. SHAP characteristic force diagram. (a) Scenario 1; (b) Scenario 2; (c) Scenario 3.
Figure 11. SHAP characteristic force diagram. (a) Scenario 1; (b) Scenario 2; (c) Scenario 3.
Buildings 14 03627 g011
Figure 12. Four typical scenarios for predicting creep flexibility: (a) Scenario S1; (b) Scenario S2; (c) Scenario S3; and (d) Scenario S4.
Figure 12. Four typical scenarios for predicting creep flexibility: (a) Scenario S1; (b) Scenario S2; (c) Scenario S3; and (d) Scenario S4.
Buildings 14 03627 g012
Table 1. The minimum, maximum, mean, and derivative values of input parameters.
Table 1. The minimum, maximum, mean, and derivative values of input parameters.
VariableMinMaxMeanSDType
w/c0.2360.80.43510.1145Input
a/c1.228.324.4181.031Input
C (kg/m3)247725410.15886.4Input
fc28 (MPa)118.910.853.0723.13Input
h mm)761800510.819275.19Input
V/S1312931.5810.97Input
T (°C)1913035.47823.346Input
RH_Test (%)2010176.7520.5Input
Sigma (MPa)0.6946.316.569.93Input
sigma/fct00.0310.840.3350.1167Input
t (days)1.421 × 10−146979200.27454.38Input
Cem13//Input
t′ (days)0.669030.6218.78Input
JCreep (μm/m/MPa)−16.7590.4270.7945.62Output
Table 2. HSOA-SVM, HSOA-RF, and HSOA-XGBoost hyperparameter settings.
Table 2. HSOA-SVM, HSOA-RF, and HSOA-XGBoost hyperparameter settings.
HSOA-SVMHSOA-RFHSOA-XGBoost
Number of iterations60Number of iterations60Number of iterations60
Population size40Population size40Population size40
c1.0N_estimators18Max_depth7
Kernel function coefficient1/KMin_leaf_nodes8Learning rate0.47
Decision_function_shape0.9Max_depth3.2Min_child_weight0.8
Penalty0.4Max_features7 γ 0.85
//// λ 0.9
Table 3. Comparison of ML model performance.
Table 3. Comparison of ML model performance.
MLR2MAEMAPERMSE
TrainTestTrainTestTrainTestTrainTest
SVM0.8120.8265.765.087.436.789.878.98
RF0.8250.8455.675.137.316.429.658.55
XGBoost0.8770.8495.355.367.236.259.458.54
HSOA-SVM0.9010.9081.781.962.233.575.095.16
HSOA-RF0.9240.9261.431.662.052.794.174.08
HSOA-XGBoost0.9450.9681.261.452.242.334.013.88
Table 4. Four classical scenarios used as examples.
Table 4. Four classical scenarios used as examples.
w/ca/cC (kg/m3)fc28 (MPa)H (mm)V/ST (°C)RH_Test (%)Sigma (MPa)sigma/fct0cemt′ (Days)
S10.314.4440588600242110117.680.20SL28
S20.555.39336338004420657.360.402RS7
S30.415.593324114004720659.40.32R14
S40.485.86325546003354405.520.117SL7
Table 5. Creep flexibility coefficient calculation formula B4.
Table 5. Creep flexibility coefficient calculation formula B4.
B4 Models [25]:
J ( t ^ , t ^ ) = q 1 + R T C 0 ( t ^ , t ^ ) + C d ( t ^ , t ^ , t ˜ 0 )
R T = exp [ U R ( 1 239 1 T + 273 ) ]
C 0 ( t ^ , t ^ ) = q 2 Q ( t ^ , t ^ ) + q 3 ln [ 1 + ( t ^ t ^ 1 days ) 0.1 ] + q 4 ln ( t ^ t ^ )
C d ( t ^ , t ^ , t ˜ 0 ) = q 5 exp [ p 5 H H ( t ^ , t ˜ 0 ) ] exp [ P 5 H H c ( t ^ 0 , t ˜ 0 ) ] 0.5
H ( t ^ , t ^ 0 ) = 1 ( 1 h ) tanh t ^ t ^ 0 τ s h
H c ( t , t ) = 1 ( 1 h ) tanh t ^ 0 t ˜ 0 τ s h
q 2 = p 2 1 Gpa ( w / c 0.38 ) p 2 w
q 3 = p 3 q 2 ( a / c 6 ) p 3 a ( w / c 0.38 ) p 3 w
q 5 = p 5 1 Gpa ( a / c 6 ) p 5 a ( w / c 0.38 ) p 5 w k h ε s h ( t ˜ 0 ) p 5 ε
τ 0 = τ c e m ( a / c 6 ) p τ a ( w / c 0.38 ) p τ w ( 6.5 c p ) p τ c
Q ( t ^ , t ^ ) = Q f ( t ^ ) [ 1 + ( Q f ( t ^ ) Z ( t , t ) ) r ( t ^ ) ] 1 r ( t ^ )
Q f ( t ^ ) = [ 0.086 ( t ^ 1 days ) 2 9 + 1.21 ( t ^ 1 days ) 4 9 ] 1
Z ( t , t ) = ( t ^ 1 days ) 0.5 ln [ 1 + ( t ^ t ^ 1 days ) 0.1 ]
r ( t ^ ) = 1.7 ( t ^ 1 days ) 0.12 + 8
q 1 = p 1 E 28
E 28 = 4734 Mpa f ¯ c Mpa
q 4 = p 4 1 Gpa ( a / c 6 ) p 4 a ( w / c 0.38 ) p 4 w
τ s h = τ 0 K τ a ( k s D 1 mm ) 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, W.; Li, H.; Liu, C.; Min, K. Concrete Creep Prediction Based on Improved Machine Learning and Game Theory: Modeling and Analysis Methods. Buildings 2024, 14, 3627. https://doi.org/10.3390/buildings14113627

AMA Style

Li W, Li H, Liu C, Min K. Concrete Creep Prediction Based on Improved Machine Learning and Game Theory: Modeling and Analysis Methods. Buildings. 2024; 14(11):3627. https://doi.org/10.3390/buildings14113627

Chicago/Turabian Style

Li, Wenchao, Houmin Li, Cai Liu, and Kai Min. 2024. "Concrete Creep Prediction Based on Improved Machine Learning and Game Theory: Modeling and Analysis Methods" Buildings 14, no. 11: 3627. https://doi.org/10.3390/buildings14113627

APA Style

Li, W., Li, H., Liu, C., & Min, K. (2024). Concrete Creep Prediction Based on Improved Machine Learning and Game Theory: Modeling and Analysis Methods. Buildings, 14(11), 3627. https://doi.org/10.3390/buildings14113627

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop