Next Article in Journal
Notes on the Abrasive Water Jet (AWJ) Machining
Next Article in Special Issue
Effect of Carbon Black and Hybrid Steel-Polypropylene Fiber on the Mechanical and Self-Sensing Characteristics of Concrete Considering Different Coarse Aggregates’ Sizes
Previous Article in Journal
Diagonal Tensile Test on Masonry Panels Strengthened with Textile-Reinforced Mortar
Previous Article in Special Issue
Coatings Based on Phosphate Cements for Fire Protection of Steel Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computation of High-Performance Concrete Compressive Strength Using Standalone and Ensembled Machine Learning Techniques

1
School of Civil Engineering, Southwest Jiaotong University, Chengdu 610031, China
2
Department of Civil Engineering, COMSATS University Islamabad, Abbottabad 22060, Pakistan
3
Faculty of Civil Engineering, Cracow University of Technology, 24 Warszawska Street, 31-155 Cracow, Poland
4
Department of Civil Engineering, College of Engineering in Al-Kharj, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
5
Department of Civil and Environmental Engineering, Faculty of Engineering, Srinakharinwirot University, Nakhonnayok 26120, Thailand
*
Authors to whom correspondence should be addressed.
Materials 2021, 14(22), 7034; https://doi.org/10.3390/ma14227034
Submission received: 23 October 2021 / Revised: 16 November 2021 / Accepted: 17 November 2021 / Published: 19 November 2021
(This article belongs to the Special Issue Emerging Construction Materials for Sustainable Infrastructure)

Abstract

:
The current trend in modern research revolves around novel techniques that can predict the characteristics of materials without consuming time, effort, and experimental costs. The adaptation of machine learning techniques to compute the various properties of materials is gaining more attention. This study aims to use both standalone and ensemble machine learning techniques to forecast the 28-day compressive strength of high-performance concrete. One standalone technique (support vector regression (SVR)) and two ensemble techniques (AdaBoost and random forest) were applied for this purpose. To validate the performance of each technique, coefficient of determination (R2), statistical, and k-fold cross-validation checks were used. Additionally, the contribution of input parameters towards the prediction of results was determined by applying sensitivity analysis. It was proven that all the techniques employed showed improved performance in predicting the outcomes. The random forest model was the most accurate, with an R2 value of 0.93, compared to the support vector regression and AdaBoost models, with R2 values of 0.83 and 0.90, respectively. In addition, statistical and k-fold cross-validation checks validated the random forest model as the best performer based on lower error values. However, the prediction performance of the support vector regression and AdaBoost models was also within an acceptable range. This shows that novel machine learning techniques can be used to predict the mechanical properties of high-performance concrete.

1. Introduction

Concrete is the most commonly used material in construction [1,2,3,4,5]. One of the necessary components of concrete is its binder, i.e., cement. There are threats to the environment caused by the process of cement production, including its high energy demand and the emission of numerous gases [6,7,8]. In order to overcome these threats, one option is to utilize additional materials that have binding properties, such as supplementary cementitious materials (SCMs), including silica fume, fly ash, blast furnace slag (BFS), etc., in place of cement, either during cement production or while manufacturing concrete [9,10,11]. Utilizing SCMs in concrete provides numerous advantages, including greater ultimate strength, increased durability, economic benefits, the prevention of surface cracking, and increased sustainability [12]. SCMs can be used to make a variety of concretes, including low-carbon concrete (LCC), self-compacting concrete (SCC), high strength concrete (HSC), and high-performance concrete (HPC) [13,14,15]. The American Concrete Institute defines HPC as ‘concrete that meets unique combinations of performance and uniformity requirements that cannot always be met using standard constituents and traditional mixing, pouring, and curing techniques’ [16]. Although several definitions for HPC have been provided, it is generally understood that HPC is concrete that is more durable than conventional concretes (and this includes its increased compressive strength capacity) [17]. For many years, HPC has been employed in the construction of several concrete structures because of its superior features, which include high strength, durability, and efficiency [18,19]. These qualities are often obtained by incorporating SCMs (especially silica fume) into HPC. In other words, SCMs can be regarded as a necessary component of HPC. The utilization of SCMs in HPC results in reduced costs, reduced heat production, decreased porousness, and enhanced chemical resistance (due to high tightness caused by the usage of a new generation of admixtures in concrete mixtures), all of which contribute to the lower maintenance costs associated with structures created using HPC [20,21].
In recent years, machine learning (ML) algorithms have demonstrated significant potential for forecasting cementitious material properties [22,23,24,25,26,27,28]. Among the numerous machine learning methods, support vector regression (SVR) and artificial neural network (ANN) methods have been widely utilized to predict concrete parameters such as compressive strength (C-S) [29], split-tensile strength, elastic modulus, and so on [30,31,32]. ANN and SVR, however, are standalone models. Other fields of study have demonstrated that prediction accuracy can be greatly improved by integrating the results of standalone models into an ensemble machine learning (EML) model [30]. So far, there has been limited research in this sector that employs ensemble learning to predict concrete parameters. Adaptive Boosting (AdaBoost) and random Forest (RF) are ensemble learning techniques that can improve prediction accuracy by combining numerous regression tree predictions and voting on the final outcome [6,33]. Ahmad et al. [6] performed standalone and EML techniques to predict the C-S of concrete and compared their accuracy. It was determined that EML techniques predicted the outcomes with a higher level of accuracy than the standalone technique. However, the results of the standalone technique were also within an acceptable range. Song et al. [34] conducted an experimental study alongside the application of standalone techniques to forecast the C-S of concrete containing ceramic waste. It was concluded that the prediction model’s outcomes were in good agreement with the experimental results. Abuodeh et al. [35] used the ANN technique to forecast the C-S of ultra-HPC and reported that the ANN model performed effectively when predicting outcomes. Hence, the present research focuses on the use of advanced techniques to forecast the properties of concrete.
Substantial research is currently being carried out to determine the mechanical characteristics of HPC. However, the casting of specimens in the laboratory and the curing of them to the desired age, in addition to testing, are time-consuming activities that require great effort. The use of novel techniques such as ML to forecast the mechanical characteristics of HPC may overcome these issues and eliminate the cost of experimentation. This study adopted both standalone (SVM) and EML (AdaBoost and RF) techniques to foretell the 28-day C-S of HPC. The performance of each model was validated using the coefficient of determination (R2) value. Additionally, statistical error checks (including checks on the mean absolute error (MSE) and the root mean square error (RMSE)) and k-fold cross-validation checks were used to compare the performance of each technique employed. Sensitivity analysis was also performed to determine the contribution of input parameters towards the prediction of outcomes.

2. Data Description

ML techniques require a variety of input variables to produce the expected output [36,37]. The data used to forecast the C-S of HPC were retrieved from the literature [38,39,40,41]. The total database available in the literature was 1030, but the data retrieved were filtered to retain only the 28-day C-S results for further studies. The models comprised fine aggregate, coarse aggregate, cement, water, superplasticizer, fly ash, and BFS as inputs, with only one variable, C-S, as the output. The quantity of data points and the input variables have a substantial impact on the model’s output [22,23,42]. In the case of the 28-day C-S prediction of HPC, this study employed a total of 425 data points (mix proportions). A descriptive statistical analysis for each input parameter was performed, and the results are shown in Table 1. This table gives the information of mean, median, mode, standard deviation, range, minimum, and maximum values for each input variable used in this study. In addition, the relative frequency distribution of each of the input variables used is shown in Figure 1.

3. Research Strategy

Anaconda software [43] was used to run the ML models using Python code. The Anaconda navigator is a desktop graphical user interface included in the Anaconda software, which enables launching applications that guide Conda packages, environments, and channels without the need to utilize command-line methods. It is also a distribution point for Python and R programming languages used for data science and ML applications that focus on clarifying package building and maintenance. To estimate the C-S of HPC, this study applied three approaches: SVR, AdaBoost, and RF. Spyder (version: 4.3.5) was picked from the Anaconda navigator for executing the models. The R2 value of the expected outcome from all models indicated the degree of accuracy. R2 values normally vary between 0 and 1, with a bigger number indicating better precision between the actual and expected results. Furthermore, to evaluate the performance of all models used in this research, statistical checks, error evaluation (including MAE, RMSE), and k-fold cross-validation checks were performed. A sensitivity analysis was also conducted in order to determine the contribution of each input variable. This approach is depicted in Figure 2 as a flowchart.

3.1. Random Forest

The fandom forest technique has been extremely successful as a general-purpose classification and regression tool. The strategy, which mixes many randomized decision trees and averages their predictions, has demonstrated a superior performance in scenarios where the number of variables is significantly greater than the number of observations. Additionally, it is adaptable to both large-scale problems and to a variety of ad hoc learning challenges, returning measurements of varying relevance.

3.2. AdaBoost

Boosting is a machine learning technique based on the concept of constructing a highly accurate prediction rule by combining many very ineffective and erroneous rules. Freund and Schapire’s AdaBoost algorithm was the first practical boosting algorithm and continues to be one of the most widely used and studied algorithms, with applications in a wide variety of industries. The AdaBoost regressor is a supervised machine learning technique that works in an ensemble. It is also referred to as Adaptive Boosting since the weights are re-allocated to each instance, with greater weights being assigned to instances that were mistakenly classified. Most of the time, boosting methods are used for supervised learning to reduce bias and variation. These ensemble algorithms are used to make the weak learner stronger, and they are quite effective. They employ an n-fold increase in the number of decision trees during the training phase for the given data. As the initial decision tree/model is prepared, recorded data that have been improperly categorized are given a high priority. Only these data are transmitted as input to the next model. The procedure is repeated until a specified number of base learners has been generated. When it comes to binary classification tasks, the AdaBoost regressor is the most effective way to improve the performance of decision trees. It can also be used to improve the performance of any other machine learning algorithms that are currently in use.

3.3. Support Vector Machine

Support vector machines are a traditional machine learning technology that can still be used to address classification issues involving large amounts of data and are especially beneficial for multidomain applications running in a big data environment. However, support vector machines are theoretically complicated and computationally expensive. A support vector machine (SVM) is a machine learning technique that uses examples to learn how to label objects. For example, an SVM may be trained to spot fraudulent credit card activity by reviewing hundreds or thousands of data on both fraudulent and legitimate credit card activity. Alternatively, an SVM can be trained to recognize handwritten numerals by inspecting a vast collection of scanned images of handwritten zeros, ones, and so on. Additionally, SVMs have been effectively applied to an expanding number of biological applications. Automatic classification of microarray gene expression profiles is a common biomedical application of support vector machines.

4. Results

4.1. Statistical Analysis

Figure 3 demonstrates the statistical analysis interpretation of the real and anticipated results for the 28-day C-S of HPC using the SVR model. The SVR produced outcomes within an acceptable range and with a low divergence amongst the real and anticipated values. The R2 value of 0.83 reflects the model’s satisfactory performance in terms of predicting results. Figure 4 depicts the scattering of experimental values (targets), projected values, and errors for the SVR model. The distribution’s greatest, lowest, and average error values were 20.34, 0.01, and 3.33 MPa, respectively. It was observed that 34.1% of the error data were less than 1 MPa, 29.4% of the error data were between 1 and 3 MPa, 27.1% were between 3 and 10 MPa, and only 9.4% of the error data were larger than 10 MPa. These values indicate the good agreement between the predicted and actual results.
Figure 5 and Figure 6 illustrate the difference between the AdaBoost model’s actual and projected outcomes. Figure 5 illustrates the correlation between actual and projected results, with an R2 value of 0.90, which is higher than the SVR model, indicating the superior performance of the AdaBoost technique compared to the SVR. Figure 6 illustrates the distribution of actual values (targets), predicted values, and errors for the AdaBoost model. The maximum, minimum, and average error values of the distribution were 9.63, 0.01, and 2.95 MPa, respectively. It was found that 30.6% of error values were less than 1 MPa, 24.7% were between 1 and 3 MPa, 23.3% were between 3 and 5 MPa, and 21.2% were greater than 5 MPa. The R2 and error distribution of the SVM and AdaBoost models suggests that the AdaBoost model can predict the C-S of HPC more accurately.
Figure 7 shows the correlation between actual and projected results for the RF model. The R2 value for the RF model is 0.93, which demonstrates a higher level of accuracy than the SVM and AdaBoost models. Moreover, Figure 8 depicts the RF model’s distribution of actual values (targets), forecast values, and errors. The distribution’s maximum, minimum, and average error values were 11.09, 0.02, and 2.22 MPa, respectively. It was noted that 37.6% of error data were less than 1 MPa, 36.5%% were between 1 and 3 MPa, 14.1% were between 3 and 5 MPa, and only 11.8% were greater than 5 MPa. This analysis reveals that the RF model has a higher level of accuracy than the SVM and AdaBoost models due to greater R2 and lower error values. Furthermore, both EML algorithms (AdaBoost and RF) employed a total of twenty sub-models to discover the optimal value that yields an uncompromised output result. Hence, these results confirm that EML techniques can predict the outcomes with higher level of accuracy than the standalone techniques.

4.2. K-Fold Cross-Validation Checks

During execution, the model’s legitimacy was determined using the k-fold cross-validation approach. Generally, the k-fold cross-validation process is performed to determine the model’s validity [36], in which pertinent data are randomly dispersed and split into 10 groups. During this study, nine groups were used for training, while one was used to validate the model. In total, 80% of the data was used to train the models, while the remaining 20% was used to evaluate the employed models. The fewer errors made (MAE and RMSE), the larger the R2 value and the more accurate the model. Additionally, the process must be repeated ten times to reach a satisfactory result. This comprehensive approach results in an exceptional level of precision. Additionally, as illustrated in Table 2, statistical analysis of errors (MSE and RMSE) was undertaken for all models. These checks also supported the higher level of accuracy of the RF model due to lower error values when compared to the SVM and AdaBoost models. The model’s response to prediction was assessed using statistical analysis in accordance with Equations (1) and (2), which were acquired from the literature [44,45].
MAE = 1 n i = 1 n x i x
RMSE = y p r e d y r e f 2 n
where n = total number of data samples, x , y r e f = reference values in the data sample, and x i , y p r e d = predicted values from models.
The k-fold cross-validation was evaluated using R2, MAE, and RMSE, and their distributions for the SVR, AdaBoost, and RF models are presented in Figure 9, Figure 10 and Figure 11, respectively. The maximum, minimum, and average R2 values for the SVM model were 0.80, 0.55, and 0.69, respectively, as depicted in Figure 9. In comparison, the AdaBoost model’s greatest, lowest, and average R2 values were 0.90, 0.60, and 0.76, respectively (Figure 10). The RF model’s highest, lowest, and average R2 values were 0.92, 0.61, and 0.79, respectively (Figure 11). When the error values (MAE and RMSE) were compared, the average MAE and RMSE values for the SVM model were 9.04 and 13.62, respectively, whereas the average MAE and RMSE values for the AdaBoost model were 8.18 and 11.63, respectively, and the average MAE and RMSE values for the RF model were 6.51 and 8.39, respectively. The RF model with the lowest error and a high R2 value performed best when predicting results. The k-fold analysis results for all the employed models containing the values of MAE, RMSE, and R2 are listed in Table 3.

4.3. Sensitivity Analysis

The purpose of this evaluation is to determine the impact of input variables on forecasting the C-S of HPC. The input parameters have a considerable influence on the projected outcome [24]. Figure 12 illustrates the effect of each input parameter on the C-S prediction of HPC. It was revealed from this analysis that cement was the most significant factor, with a 23.8% contribution, followed by superplasticizer, with 20.0%, and BFS, with a 17.1% contribution. However, the remaining input variables contributed to the prediction of C-S of HPC to a lesser degree, with fly ash accounting for 15.6%, water accounting for 12.6%, coarse aggregate accounting for 6.5%, and fine aggregate accounting for 4.4%. Sensitivity analysis yielded findings proportionate to the amount of input variables and number of data points used in the model’s construction. The results of the contribution level of all the input parameters can be obtained directly from the software used to run the models. However, Equations (3) and (4) were applied to ascertain the influence of each input variable on the model’s output:
N i = f m a x x i f m i n x i
S i = N i j i n N j
where f m a x x i and f m i n x i are the highest and lowest of the anticipated outputs over the i t h output.

5. Discussion

The goal of this research was to illustrate how supervised ML approaches could be applied to forecast the compressive strength of HPC. The study employed three machine learning techniques: one standalone, i.e., SVR, and two ensembled, including AdaBoost and RF. To ascertain which algorithm was the most accurate predictor, the prediction performance of each technique was compared. The result of the RF model was more precise, with an R2 value of 0.93, compared to the SVM and AdaBoost models, which produced R2 values of 0.83 and 0.90, respectively. Additionally, the performance of each model was confirmed using statistical analysis and the k-fold cross-validation technique. The lower the error levels, the better the model performed. However, assessing and proposing the optimal ML regressor for predicting results across a range of topics is difficult, as the success of each model was highly dependent on the input parameters and the data points used to run it. However, EML techniques often exploit the weak learner by building sub-models that can be trained on data and optimized to maximize the R2 value. Figure 13 and Figure 14 illustrate the distribution of R2 values for sub-models in the cases of the AdaBoost and RF approaches, respectively. The maximum, minimum, and average R2 values for AdaBoost sub-models were 0.904, 0.881, and 0.895, respectively (Figure 13), whereas the maximum, minimum, and average R2 values for RF sub-models were 0.934, 0.891, and 0.918, respectively (Figure 14). These values suggest the higher accuracy of RF sub-models when compared to AdaBoost sub-models. The literature indicates that RF models yield more accurate results than other ML approaches [46]. Additionally, a sensitivity analysis was conducted to ascertain the effect of each input parameter on the predicted C-S of HPC. The performance of the model can be affected by the input variables and the size of the data set. The sensitivity analysis examined the extent to which each of the seven input parameters contributed to the anticipated result. This study compared the performance of three ML approaches in order to determine the best technique for forecasting the C-S of HPC. The RF approach was noted to be the more accurate technique for the prediction of the mechanical properties of concrete. The contribution of this study is the selection of a superior/more accurate approach for predicting concrete strength.

6. Conclusions

This study aimed to employ standalone and ensemble machine learning (EML) techniques to predict the 28-day compressive strength (C-S) of high-performance concrete (HPC). One standalone technique, i.e., support vector regression (SVR), and two EML techniques, namely AdaBoost and random forest (RF), were employed to predict outcomes. The following conclusions have been drawn from this research:
EML techniques were more accurate in predicting the C-S of HPC than the standalone technique, with the RF model exhibiting the highest accuracy. The coefficient correlation (R2) values for the SVR, AdaBoost, and RF models were 0.83, 0.90, and 0.93, respectively. The results of all of the models employed were within an acceptable range, with little variance from the actual results.
Statistical analysis and k-fold cross-validation checks also demonstrated the models’ good performance. Additionally, these checks confirmed the RF model’s superior performance compared to the other models studied.
The contribution of input parameters was determined by sensitivity analysis and observed that cement, superplasticizer, blast furnace slag, fly ash, waster, coarse aggregate, and fine aggregate contributed towards the outcome’s prediction by 23.8%, 20.0%, 17.1%, 15.6%, 12.6%, 6.5%, and 4.4%, respectively.
Novel machine learning methods can accurately forecast the strength properties of concrete without the need for excessive sample casting and testing time.
This paper proposes the use of both standalone (SVM) and ensembled (AdaBoost and RF) machine learning approaches to forecast the 28-day C-S of HPC. Other ML techniques should also be employed to compare their accuracy in predicting outcomes. It is recommended that in future investigations, the quantity of data points and outcomes be increased through experimental work, field tests, and numerical analysis employing a variety of methodologies (e.g., the Monte Carlo simulation, among others). In order to improve the models’ responses, environmental variables (such as high temperatures and humidity) could also be included in the input parameters, together with a full explanation of the raw materials. Moreover, additional in-depth investigations, checks, and effects should be incorporated in order to improve the evaluation and comprehension of the outcomes obtained through the use of ML techniques.

Author Contributions

Y.X.—validation, writing-review and editing; W.A.—conceptualization, formal analysis, software and writing-original draft; A.A.—project administration, visualization, validation and supervision; K.A.O.—writing—review and editing; M.D.—funding acquisition and resources; F.A.—writing—review and editing; P.J.—writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by Cracow University of Technology and COMSATS University Islamabad.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Not Applicable.

Acknowledgments

The authors would like to thank the supportive roles of their departments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mangi, S.A.; Wan Ibrahim, M.H.; Jamaluddin, N.; Arshad, M.F.; Shahidan, S. Performances of concrete containing coal bottom ash with different fineness as a supplementary cementitious material exposed to seawater. Eng. Sci. Technol. Int. J. 2019, 22, 929–938. [Google Scholar] [CrossRef]
  2. Molay, T.G.G.; Leroy, M.N.L.; Fidele, T.; Franck, H.G.; Bienvenu, N.J.-M. Mechanical and physical performances of concretes made from crushed sands of different geological nature subjected to high temperatures. Eng. Sci. Technol. Int. J. 2019, 22, 1116–1124. [Google Scholar] [CrossRef]
  3. Ahmad, W.; Farooq, S.H.; Usman, M.; Khan, M.; Ahmad, A.; Aslam, F.; Yousef, R.A.; Abduljabbar, H.A.; Sufian, M. Effect of coconut fiber length and content on properties of high strength concrete. Materials 2020, 13, 1075. [Google Scholar] [CrossRef] [Green Version]
  4. Khan, M.; Ali, M. Use of glass and nylon fibers in concrete for controlling early age micro cracking in bridge decks. Constr. Build. Mater. 2016, 125, 800–808. [Google Scholar] [CrossRef]
  5. Khan, M.; Ali, M. Improvement in concrete behavior with fly ash, silica-fume and coconut fibres. Constr. Build. Mater. 2019, 203, 174–187. [Google Scholar] [CrossRef]
  6. Ahmad, W.; Ahmad, A.; Ostrowski, K.A.; Aslam, F.; Joyklad, P.; Zajdel, P. Application of Advanced Machine Learning Approaches to Predict the Compressive Strength of Concrete Containing Supplementary Cementitious Materials. Materials 2021, 14, 5762. [Google Scholar] [CrossRef]
  7. Liu, T.; Nafees, A.; Javed, M.F.; Aslam, F.; Alabduljabbar, H.; Xiong, J.-J.; Khan, M.I.; Malik, M.Y. Comparative study of mechanical properties between irradiated and regular plastic waste as a replacement of cement and fine aggregate for manufacturing of green concrete. Ain Shams Eng. J. 2021. [Google Scholar] [CrossRef]
  8. Shaker, F.; Rashad, A.; Allam, M. Properties of concrete incorporating locally produced Portland limestone cement. Ain Shams Eng. J. 2018, 9, 2301–2309. [Google Scholar] [CrossRef]
  9. Ahmad, W.; Ahmad, A.; Ostrowski, K.A.; Aslam, F.; Joyklad, P. A scientometric review of waste material utilization in concrete for sustainable construction. Case Stud. Constr. Mater. 2021, 15, e00683. [Google Scholar] [CrossRef]
  10. Mohamed, H.A. Effect of fly ash and silica fume on compressive strength of self-compacting concrete under different curing conditions. Ain Shams Eng. J. 2011, 2, 79–86. [Google Scholar] [CrossRef] [Green Version]
  11. Dalvand, A.; Ahmadi, M. Impact failure mechanism and mechanical characteristics of steel fiber reinforced self-compacting cementitious composites containing silica fume. Eng. Sci. Technol. Int. J. 2021, 24, 736–748. [Google Scholar] [CrossRef]
  12. Ahmad, W.; Ahmad, A.; Ostrowski, K.A.; Aslam, F.; Joyklad, P.; Zajdel, P. Sustainable approach of using sugarcane bagasse ash in cement-based composites: A systematic review. Case Stud. Constr. Mater. 2021, 15, e00698. [Google Scholar] [CrossRef]
  13. Salimi, J.; Ramezanianpour, A.M.; Moradi, M.J. Studying the effect of low reactivity metakaolin on free and restrained shrinkage of high performance concrete. J. Build. Eng. 2020, 28, 101053. [Google Scholar] [CrossRef]
  14. Uva, G.; Porco, F.; Fiore, A.; Mezzina, M. The assessment of structural concretes during construction phases. Struct. Surv. 2014, 32, 189–208. [Google Scholar] [CrossRef]
  15. Sangiorgio, V.; Uva, G.; Adam, J.M.; Scarcelli, L. Failure analysis of reinforced concrete elevated storage tanks. Eng. Fail. Anal. 2020, 115, 104637. [Google Scholar] [CrossRef]
  16. American concrete institute manual of concrete practice. In ACI Concrete Terminology; ACI CT-13; American Concrete Institute: Farmington Hills, MI, USA, 2013.
  17. Pedro, D.; De Brito, J.; Evangelista, L. Durability performance of high-performance concrete made with recycled aggregates, fly ash and densified silica fume. Cem. Concr. Compos. 2018, 93, 63–74. [Google Scholar] [CrossRef]
  18. Li, J.; Wu, Z.; Shi, C.; Yuan, Q.; Zhang, Z. Durability of ultra-high performance concrete—A review. Constr. Build. Mater. 2020, 255, 119296. [Google Scholar] [CrossRef]
  19. Semendary, A.A.; Hamid, W.K.; Steinberg, E.P.; Khoury, I. Shear friction performance between high strength concrete (HSC) and ultra high performance concrete (UHPC) for bridge connection applications. Eng. Struct. 2020, 205, 110122. [Google Scholar] [CrossRef]
  20. Park, S.; Wu, S.; Liu, Z.; Pyo, S. The role of supplementary cementitious materials (SCMs) in ultra high performance concrete (UHPC): A review. Materials 2021, 14, 1472. [Google Scholar] [CrossRef]
  21. Khatri, R.P.; Sirivivatnanon, V.; Gross, W. Effect of different supplementary cementitious materials on mechanical properties of high performance concrete. Cem. Concr. Res. 1995, 25, 209–220. [Google Scholar] [CrossRef]
  22. Ahmad, A.; Farooq, F.; Niewiadomski, P.; Ostrowski, K.; Akbar, A.; Aslam, F.; Alyousef, R. Prediction of compressive strength of fly ash based concrete using individual and ensemble algorithm. Materials 2021, 14, 794. [Google Scholar] [CrossRef]
  23. Ahmad, A.; Farooq, F.; Ostrowski, K.A.; Śliwa-Wieczorek, K.; Czarnecki, S. Application of Novel Machine Learning Techniques for Predicting the Surface Chloride Concentration in Concrete Containing Waste Material. Materials 2021, 14, 2297. [Google Scholar] [CrossRef] [PubMed]
  24. Ahmad, A.; Ostrowski, K.A.; Maślak, M.; Farooq, F.; Mehmood, I.; Nafees, A. Comparative Study of Supervised Machine Learning Algorithms for Predicting the Compressive Strength of Concrete at High Temperature. Materials 2021, 14, 4222. [Google Scholar] [CrossRef]
  25. Amin, M.N.; Iqtidar, A.; Khan, K.; Javed, M.F.; Shalabi, F.I.; Qadir, M.G. Comparison of Machine Learning Approaches with Traditional Methods for Predicting the Compressive Strength of Rice Husk Ash Concrete. Crystals 2021, 11, 779. [Google Scholar] [CrossRef]
  26. Shah, H.A.; Rehman, S.K.U.; Javed, M.F.; Iftikhar, Y. Prediction of compressive and splitting tensile strength of concrete with fly ash by using gene expression programming. Struct. Concr. 2021, 1–15. [Google Scholar] [CrossRef]
  27. Algaifi, H.A.; Alqarni, A.S.; Alyousef, R.; Bakar, S.A.; Ibrahim, M.H.W.; Shahidan, S.; Ibrahim, M.; Salami, B.A. Mathematical prediction of the compressive strength of bacterial concrete using gene expression programming. Ain Shams Eng. J. 2021. [Google Scholar] [CrossRef]
  28. Ruggieri, S.; Cardellicchio, A.; Leggieri, V.; Uva, G. Machine-learning based vulnerability analysis of existing buildings. Autom. Constr. 2021, 132, 103936. [Google Scholar] [CrossRef]
  29. Alexandridis, A.; Stavrakas, I.; Stergiopoulos, C.; Hloupis, G.; Ninos, K.; Triantis, D. Non-destructive assessment of the three-point-bending strength of mortar beams using radial basis function neural networks. Comput. Concr. 2015, 16, 919–932. [Google Scholar] [CrossRef]
  30. Chaabene, W.B.; Flah, M.; Nehdi, M.L. Machine learning prediction of mechanical properties of concrete: Critical review. Constr. Build. Mater. 2020, 260, 119889. [Google Scholar] [CrossRef]
  31. Song, H.; Ahmad, A.; Farooq, F.; Ostrowski, K.A.; Maślak, M.; Czarnecki, S.; Aslam, F. Predicting the compressive strength of concrete with fly ash admixture using machine learning algorithms. Constr. Build. Mater. 2021, 308, 125021. [Google Scholar] [CrossRef]
  32. DeRousseau, M.A.; Kasprzyk, J.R.; Srubar Iii, W.V. Computational design optimization of concrete mixtures: A review. Cem. Concr. Res. 2018, 109, 42–53. [Google Scholar] [CrossRef]
  33. Sun, J.; Ma, Y.; Li, J.; Zhang, J.; Ren, Z.; Wang, X. Machine learning-aided design and prediction of cementitious composites containing graphite and slag powder. J. Build. Eng. 2021, 43, 102544. [Google Scholar] [CrossRef]
  34. Song, H.; Ahmad, A.; Ostrowski, K.A.; Dudek, M. Analyzing the Compressive Strength of Ceramic Waste-Based Concrete Using Experiment and Artificial Neural Network (ANN) Approach. Materials 2021, 14, 4518. [Google Scholar] [CrossRef]
  35. Abuodeh, O.R.; Abdalla, J.A.; Hawileh, R.A. Assessment of compressive strength of Ultra-high Performance Concrete using deep machine learning techniques. Appl. Soft Comput. 2020, 95, 106552. [Google Scholar] [CrossRef]
  36. Ahmad, A.; Chaiyasarn, K.; Farooq, F.; Ahmad, W.; Suparp, S.; Aslam, F. Compressive Strength Prediction via Gene Expression Programming (GEP) and Artificial Neural Network (ANN) for Concrete Containing RCA. Buildings 2021, 11, 324. [Google Scholar] [CrossRef]
  37. Sufian, M.; Ullah, S.; Ostrowski, K.A.; Ahmad, A.; Zia, A.; Śliwa-Wieczorek, K.; Siddiq, M.; Awan, A.A. An Experimental and Empirical Study on the Use of Waste Marble Powder in Construction Material. Materials 2021, 14, 3829. [Google Scholar] [CrossRef] [PubMed]
  38. Machine Learning Repository, Center for Machine Learning and Intelligent Systems. Available online: https://archive.ics.uci.edu/ml/datasets/concrete+compressive+strength (accessed on 3 August 2007).
  39. Yeh, I.C. Modeling of strength of high-performance concrete using artificial neural networks. Cem. Concr. Res. 1998, 28, 1797–1808. [Google Scholar] [CrossRef]
  40. Yeh, I.C. Prediction of strength of fly ash and slag concrete by the use of artificial neural networks. J. Chin. Inst. Civil Hydraul. Eng. 2003, 15, 659–663. [Google Scholar]
  41. Yeh, I.C. Analysis of strength of concrete using design of experiments and neural networks. J. Mater. Civ. Eng. 2006, 18, 597–604. [Google Scholar] [CrossRef]
  42. Gandomi, A.H.; Roke, D.A. Assessment of artificial neural network and genetic programming as predictive tools. Adv. Eng. Softw. 2015, 88, 63–72. [Google Scholar] [CrossRef]
  43. Available online: https://anaconda.org/anaconda/anaconda-navigator (accessed on 3 August 2007).
  44. Farooq, F.; Ahmed, W.; Akbar, A.; Aslam, F.; Alyousef, R. Predictive modeling for sustainable high-performance concrete from industrial wastes: A comparison and optimization of models using ensemble learners. J. Clean. Prod. 2021, 292, 126032. [Google Scholar] [CrossRef]
  45. Aslam, F.; Farooq, F.; Amin, M.N.; Khan, K.; Waheed, A.; Akbar, A.; Javed, M.F.; Alyousef, R.; Alabdulijabbar, H. Applications of gene expression programming for estimating compressive strength of high-strength concrete. Adv. Civ. Eng. 2020, 2020, 8850535. [Google Scholar] [CrossRef]
  46. Farooq, F.; Nasir Amin, M.; Khan, K.; Rehan Sadiq, M.; Faisal Javed, M.; Aslam, F.; Alyousef, R. A comparative study of random forest and genetic engineering programming for the prediction of compressive strength of high strength concrete (HSC). Appl. Sci. 2020, 10, 7330. [Google Scholar] [CrossRef]
Figure 1. Relative frequency distribution of input variables: (a) fine aggregate; (b) coarse aggregate; (c) cement; (d) water; (e) superplasticizer; (f) fly ash; (g) blast furnace slag.
Figure 1. Relative frequency distribution of input variables: (a) fine aggregate; (b) coarse aggregate; (c) cement; (d) water; (e) superplasticizer; (f) fly ash; (g) blast furnace slag.
Materials 14 07034 g001aMaterials 14 07034 g001b
Figure 2. Sequence of research methodology.
Figure 2. Sequence of research methodology.
Materials 14 07034 g002
Figure 3. Actual and predicted outcomes relationship for support vector regression model.
Figure 3. Actual and predicted outcomes relationship for support vector regression model.
Materials 14 07034 g003
Figure 4. Actual, predicted, and error values’ distribution for support vector regression model.
Figure 4. Actual, predicted, and error values’ distribution for support vector regression model.
Materials 14 07034 g004
Figure 5. Actual and predicted outcomes relationship for AdaBoost model.
Figure 5. Actual and predicted outcomes relationship for AdaBoost model.
Materials 14 07034 g005
Figure 6. Actual, predicted, and error values’ distribution for AdaBoost model.
Figure 6. Actual, predicted, and error values’ distribution for AdaBoost model.
Materials 14 07034 g006
Figure 7. Actual and predicted outcomes relationship for random forest model.
Figure 7. Actual and predicted outcomes relationship for random forest model.
Materials 14 07034 g007
Figure 8. Actual, predicted, and error values’ distribution for random forest model.
Figure 8. Actual, predicted, and error values’ distribution for random forest model.
Materials 14 07034 g008
Figure 9. Statistical representation for K-fold cross-validation for support vector regression model.
Figure 9. Statistical representation for K-fold cross-validation for support vector regression model.
Materials 14 07034 g009
Figure 10. Statistical representation for K-fold cross-validation for the AdaBoost model.
Figure 10. Statistical representation for K-fold cross-validation for the AdaBoost model.
Materials 14 07034 g010
Figure 11. Statistical representation for K-fold cross-validation for random forest model.
Figure 11. Statistical representation for K-fold cross-validation for random forest model.
Materials 14 07034 g011
Figure 12. Contribution of input variable towards the prediction where SP: superplasticizer; BFS: blast furnace slag; FA: fly ash; CA: coarse aggregate; FAg: fine aggregate.
Figure 12. Contribution of input variable towards the prediction where SP: superplasticizer; BFS: blast furnace slag; FA: fly ash; CA: coarse aggregate; FAg: fine aggregate.
Materials 14 07034 g012
Figure 13. AdaBoost sub-model’s coefficient correlation (R2) values.
Figure 13. AdaBoost sub-model’s coefficient correlation (R2) values.
Materials 14 07034 g013
Figure 14. Random forest sub-model’s coefficient correlation (R2) values.
Figure 14. Random forest sub-model’s coefficient correlation (R2) values.
Materials 14 07034 g014
Table 1. Descriptive analysis of input variables.
Table 1. Descriptive analysis of input variables.
ParameterInput Variable (kg/m3)
Fine AggregateCoarse AggregateCementWaterSuperplasticizerFly AshBlast Furnace Slag
Mean764.4956.1265.4183.17.062.886.3
Standard Error3.54.15.10.90.33.24.3
Median769.3953.2261.0185.07.860.094.7
Mode755.8932.0313.0192.00.00.00.0
Standard Deviation73.183.8104.719.35.466.287.8
Range398.6344.0438.0125.232.2200.1359.4
Minimum594.0801.0102.0121.80.00.00.0
Maximum992.61145.0540.0247.032.2200.1359.4
Table 2. Statistical checks of techniques employed.
Table 2. Statistical checks of techniques employed.
ModelMAERMSE
Support vector regression3.3295.325
AdaBoost2.9473.908
Random forest2.2233.183
Table 3. K-fold cross-validation outcomes.
Table 3. K-fold cross-validation outcomes.
K-FoldSVRAdaBoostRandom Forest
MAERMSER2MAERMSER2MAERMSER2
14.375.680.776.797.540.833.945.330.80
27.3810.250.736.7910.100.755.788.110.88
313.3819.870.8010.8814.830.898.6812.080.66
420.1337.400.6114.4626.740.606.658.790.61
59.2812.690.677.6210.930.906.319.620.91
69.6712.130.759.8511.600.776.057.140.84
77.9111.280.587.6011.240.886.798.700.62
84.665.950.744.735.210.657.098.590.92
96.089.750.557.078.950.615.575.390.79
107.5111.150.696.069.220.728.2510.110.86
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, Y.; Ahmad, W.; Ahmad, A.; Ostrowski, K.A.; Dudek, M.; Aslam, F.; Joyklad, P. Computation of High-Performance Concrete Compressive Strength Using Standalone and Ensembled Machine Learning Techniques. Materials 2021, 14, 7034. https://doi.org/10.3390/ma14227034

AMA Style

Xu Y, Ahmad W, Ahmad A, Ostrowski KA, Dudek M, Aslam F, Joyklad P. Computation of High-Performance Concrete Compressive Strength Using Standalone and Ensembled Machine Learning Techniques. Materials. 2021; 14(22):7034. https://doi.org/10.3390/ma14227034

Chicago/Turabian Style

Xu, Yue, Waqas Ahmad, Ayaz Ahmad, Krzysztof Adam Ostrowski, Marta Dudek, Fahid Aslam, and Panuwat Joyklad. 2021. "Computation of High-Performance Concrete Compressive Strength Using Standalone and Ensembled Machine Learning Techniques" Materials 14, no. 22: 7034. https://doi.org/10.3390/ma14227034

APA Style

Xu, Y., Ahmad, W., Ahmad, A., Ostrowski, K. A., Dudek, M., Aslam, F., & Joyklad, P. (2021). Computation of High-Performance Concrete Compressive Strength Using Standalone and Ensembled Machine Learning Techniques. Materials, 14(22), 7034. https://doi.org/10.3390/ma14227034

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop