Next Article in Journal
Percussion-Based Pipeline Ponding Detection Using a Convolutional Neural Network
Next Article in Special Issue
Study on the Dynamic Response Characteristics and p–y Curve of Straight and Inclined Pile Groups in Saturated Sands
Previous Article in Journal
Effects of High-Intensity Electromagnetic Stimulation on Reducing Upper Limb Spasticity in Post-Stroke Patients
Previous Article in Special Issue
Dynamic Lateral Response of the Partially-Embedded Single Piles in Layered Soil
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Pile Bearing Capacity Using XGBoost Algorithm: Modeling and Performance Evaluation

1
Department of Civil Engineering, University of Engineering and Technology, Peshawar 25120, Pakistan
2
Department of Civil Engineering, Faculty of Engineering, International Islamic University Malaysia, Jalan Gombak, Selangor 50728, Malaysia
3
Department of Civil Engineering, University of Engineering and Technology Peshawar (Bannu Campus), Bannu 28100, Pakistan
4
Faculty of Engineering, University of Technology and Economics H. Chodkowska in Warsaw, Jutrzenki 135, 02-231 Warsaw, Poland
5
Faculty of Mechatronics, Armament and Aerospace of the Military University of Technology, Sylwestra Kaliskiego 2, 00-908 Warsaw, Poland
6
Faculty of Civil Engineering and Resource Management, AGH University of Science and Technology, 30-059 Kraków, Poland
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(4), 2126; https://doi.org/10.3390/app12042126
Submission received: 7 December 2021 / Revised: 1 February 2022 / Accepted: 9 February 2022 / Published: 18 February 2022
(This article belongs to the Special Issue Recent Progress on Advanced Foundation Engineering)

Abstract

:
The major criteria that control pile foundation design is pile bearing capacity (Pu). The load bearing capacity of piles is affected by the various characteristics of soils and the involvement of multiple parameters related to both soil and foundation. In this study, a new model for predicting bearing capacity is developed using an extreme gradient boosting (XGBoost) algorithm. A total of 200 driven piles static load test-based case histories were used to construct and verify the model. The developed XGBoost model results were compared to a number of commonly used algorithms—Adaptive Boosting (AdaBoost), Random Forest (RF), Decision Tree (DT) and Support Vector Machine (SVM) using various performance measure metrics such as coefficient of determination, mean absolute error, root mean square error, mean absolute relative error, Nash–Sutcliffe model efficiency coefficient and relative strength ratio. Furthermore, sensitivity analysis was performed to determine the effect of input parameters on Pu. The results show that all of the developed models were capable of making accurate predictions however the XGBoost algorithm surpasses others, followed by AdaBoost, RF, DT, and SVM. The sensitivity analysis result shows that the SPT blow count along the pile shaft has the greatest effect on the Pu.

1. Introduction

A pile is a long, structural element used to allow structural loads to be transferred to the soils at a depth below the structure’s base. Axial, lateral, and moment loads are examples of structural loads. The load transmission mechanism is based on pile toe and pile shaft resistances [1]. Deep foundations are another word for pile foundations that are often used in practice. Pile foundations are used to support structures that cannot be supported economically on shallow foundations. The most significant factor when designing a pile foundation is pile carrying capacity (Pu) [2]. Various ways to determine pile carrying capacity have been used during the years of research and development [3,4,5,6,7,8,9,10,11,12,13], including dynamic analysis, high strain dynamic test, pile load test, cone penetration test (CPT) and in situ tests. Some research, claims that the aforementioned connections exaggerate the bearing capability [14]. However, the pile load test is considered as one of the best methods to determine the pile bearing capacity, although this strategy is costly for small-scale projects and time-consuming [10], it is critical to find a more practical approach. As a result, many studies using in situ test data to assess pile carrying capacity have been performed [9].
Lopes and Laprovitera [15], and Decort [16] proposed different formulas for determining pile carrying capacity for several soils, including clay and sand. Conventional approaches have used numerous main parameters to determine the mechanical properties of piles, including the diameter of pile, length of pile, type of soil, and SPT blow counts of each layer. Nevertheless, the selection of relevant parameters, along with the failure in covering other parameters, have led to the disagreement of results given by various approaches [17]. As a result, the development of an optimal model for selecting an appropriate set of parameters is critical.
A recently developed approach based on data mining techniques has been increasingly employed to resolve real-world problems for the past half-decade, particularly in the field of civil engineering [18,19,20,21,22,23,24,25,26,27,28]. Several practical problems have already been effectively performed using machine learning algorithms, paving the way for new prospects in the construction industry. Furthermore, a variety of machine learning algorithms, for example, random forest, artificial neural network (ANN), decision tree, adaptive neuro-fuzzy inference system (ANFIS), AdaBoost, SVM, XGBoost have been developed for addressing technical issues, such as pile mechanical behavior prediction.
Goh et al. [29,30] produced an ANN-based algorithm of piles driven in clays to predict the capacity of friction, using on-field data records to train the algorithm. Furthermore, Shahin et al. [31,32,33,34] employed the ANN-based model for forecasting pile load capacity using data that included in situ load testing and cone penetration test (CPT) results. Similarly, Nawari et al. [35] published an ANN approach that uses SPT data and shaft geometry to measure the settling of drilled shafts. Pham et al. [17] produced an ANN and RF to predict driven pile’s capacity. Momeni et al. [36] created an ANN model modified with Genetic Algorithm (GA) which select appropriate biases and weights for predicting pile bearing capacity. Based on CPT data, Kordjazi et al. [37] employed an SVM model to forecast the pile ultimate load-bearing capability. Liu et al. [21] developed XGBoost, Backpropagation Neural Network (BPNN) and RF algorithm to estimate driven piles bearing Capacity. Liang et al. estimated stability of hard rock pillars applying XGBoost, gradient boosting decision tree (GBDT), and light gradient boosting machine (LightGBM) Algorithms [23]. Pham et al. [38] has also developed Deep Learning Neural Network to estimate the carrying capacity of piles.
In addition to machine learning (ML) techniques mentioned above, the GBDT method demonstrates excellent results in a variety of disciplines [39,40,41]. It uses the boosting strategy to incorporate many DTs into a strong classifier as one of the ensemble learning algorithms [42]. DTs belong to the ML approach which employs a tree-like framework to handle a wide range of input types while tracing each path to the prediction outcomes [43]. DTs, on the other hand, are easy to overfit and sensitive to dataset noise because errors of the DTs were offset by one another, the total prediction performance of GBDT improves with the integration of DTs. XGBoost [44] and LightGBM [45] have recently been proposed in the context of GBDT. They have also attracted a lot of attention as a result of their outstanding performances. These three techniques, in particular, operate well with tiny datasets. To some extent, overfitting, which occurs when results match existing data very closely but fail to correctly estimate future trends, can also be prevented [43].
The aim of the present study is to develop a robust model to estimate axial pile bearing capacity using the XGBoost algorithm based on reliable pile load test results. The scope of the present research includes the following:
  • To develop a model that is able to learn the complex relationship among axial pile bearing capacity and its influencing factors with reasonable precision.
  • To validate the proposed model by comparing the efficacy with prominent modeling techniques, such as AdaBoost, RF, DT, and SVM in terms of performance measure metrics.
  • To conduct sensitivity analyses for the determination of the effect of each input parameter on Pu.
The framework of the paper is as follows: In Section 2, data collection and preparation are presented. Section 3 describes the machine learning approaches. The construction of the prediction model is presented in Section 4. Results and discussion are given in Section 5. Lastly, there are some closing remarks.

2. Data Collection and Preparation

2.1. Dataset

In this study, the dataset of 200 reinforced concrete piles at the test site in Ha Nam province–Vietnam (the complete database is available in Table A1) was used to train and test the model. As a first step, all known parameters affecting Pu were taken into account. Furthermore, it was discovered that the majority of traditional methods utilized three categories of parameters: geometry of pile, pile material quality, and soil attributes [3]. To achieve the measurements, hydraulic pile presses were used to drive pre-cast square-section piles with closed tips to the ground at a constant rate of penetration. The testing began at least seven days after the piles were driven, and the experimental setup is shown in Figure 1. The load increased gradually in each pile test, as can be observed. The load might be increased up to 200 percent of the pile load design depending on the design requirements. The time it takes to achieve 100 percent, 150 percent, and 200 percent of the load could take from around 6 to 12 h or 24 h, depending on the load [38]. These two principles were used to determine pile bearing capacity:
(i)
the pile bearing capacity was taken as the failure load when the settlement of pile top at the current load level was five times or higher than the settlement of pile top at the previous load level;
(ii)
when the load–settlement curve became linear at the last test load, condition (i) would not be used. In such a case, the test load at which progressive movement occurs or the total settlement exceeds 10 % of the pile diameter or width would be taken as the pile bearing capacity.
As a result, previous studies (e.g., [38]) show that pile bearing capacity (Pu) is a function of (1) diameter of the pile (D); (2) depth of the first layer of soil embedded (X1); (3) depth of the second layer of soil embedded (X2); (4) depth of the third layer of soil embedded (X3); (5) pile top elevation (Xp); (6) ground elevation (Xg); (7) extra pile top elevation (Xt); (8) pile tip elevation (Xm); (9) SPT blow count at pile shaft (NS) and (10) SPT blow count at pile tip (Nt) as shown in Figure 2. Therefore, in the current study, these input variables were used to develop the proposed models.
Collected data were divided into training and testing sets, researchers have used a different percentage of the available data as the training set for different problems. For instance, Pham et al. [38] used 60%; Liang et al. [23] used 70%; while Ahmad et al. [28] used 80% of the data for training. The statistical consistency of training and testing datasets has a substantial impact on the results when using soft computing techniques which improves the performance of the model and helps in evaluating them better [22,46]. To choose the most consistent representation, statistical studies of input and output variables of the training and testing data were performed. It was accomplished through the use of a trial-and-error strategy. For training and testing datasets, the best statistically consistent combination was selected. The data division was performed in such a way that 140 (70%) samples were used for training, and 60 (30%) samples were used for testing the models considered in this study. The results of the statistical analysis of the finally selected combinations are shown in Table 1, which includes minimum, mean, maximum and standard deviation of the input and output variables.

2.2. Correlation Analysis

Correlation (ρ) was used to verify the intensity of correlation between different parameters (see Table 2). Given pair of random variables (m, n), the following equation for ρ is used:
ρ ( m ,   n ) = c o v ( m ,   n ) σ m σ n  
where cov denotes covariance, σ m denotes the standard deviation of m, and   σ n denotes the standard deviation of n. | ρ | > 0.8 represents a strong relationship among m and n, values between 0.3 and 0.8 represents medium relationship, while | ρ | < 0.30 represents weak relationship [47]. According to Song et al. [48], correlation is considered as “strong” if | ρ | > 0.8. Table 2 displays the correlations between input and output characteristics. The correlation coefficient has a maximum absolute value of 0.989, as shown in Table 2. There is a “strong to weak” relationship among various variable combinations so none of the input variables was removed.

3. Machine Learning Methods

3.1. Extreme Gradient Boosting Algorithm

Chen and Guestrin [44] suggested the XGBoost algorithm, which is based on the GBDT structure. It has attracted a lot of attention as a result of its outstanding results in Kaggle’s ML competitions [49]. Unlike GBDT, the XGBoost goal function includes a regularization term to avoid overfitting. The main objective function is described as follows:
O = i = 1 n L ( y i , F ( x i ) ) + k = 1 t R ( f k ) + C
where R ( f k )   represents the regularization term at iteration k, and C being a constant that can be removed selectively.
Regularization term R ( f k ) written as,
R ( f k ) = α H + 1 2 η j = 1 H w j 2
where α is the complexity of leaves, H denotes the number of leaves, η signifies penalty variable, and ω j represents output results in each leaf node. Leaves denote the expected categories based on classification criteria, whereas the leaf node denotes the tree node which cannot be divided.
Furthermore, unlike GBDT, XGBoost employs a second-order Taylor series of main functions rather than the first-order derivative. If the loss function is the mean square error (MSE), then the main function may be written as:
O = i = 1 n [ p i ω q ( x i ) + 1 2 ( q i ω q ( x i ) 2 ) ] + α H + 1 2 η j = 1 H ω j 2
where q ( x i ) is a function that maps data points to leaves, g i and h i represents loss function’s first and second derivatives, respectively.
The final loss value is calculated by adding all of the loss values together. Because samples in the DT corresponds to nodes of leaf, the ultimate loss value can be calculated by adding loss values of the leaf nodes. As a result, the main function can be written as:
O = j = 1 T [ P j ω j + 1 2 ( Q j + η ) ω j 2 ) ] + α H  
where P j = i ϵ I j p i , Q j = i ϵ I j q i , and I j are the total number of samples in leaf node j.
To summarize, the challenge of optimizing the main function is reduced to identifying the minimum of a quadratic function. Due to the addition of regularization phenomena, XGBoost has a stronger capability to avoid overfitting. The structure of XGBoost can be seen in Figure 3.

3.2. Random Forest (RF) Algorithm

Because of its simplicity and diversity, RF is the most applied ML method. Breiman in 2001, developed this supervised learning approach for classification and regression analysis [50]. RF is an integrated learning strategy that collects data from a single DT and improves prediction accuracy by using majority voting or mean findings, depending on the task.
Assume you have an input data set with the following values Q = q1, q2, q3, …, qn where n is the number of datasets. An RF model would be a set of T trees T1(Q), T2(Q), T3(Q) …, Tn(Q). R 1 ^ , R 2 ^ R n ^ is the predicted outcome of these decision-making trees. The eventual output of the RF model for the regression problem will be the average of all the above trees’ prediction outcomes. The concept of splitting initial training sets into smaller sets, with only a few predictive elements picked at random in each split, has been used to construct tree-growing algorithms. Because the programmer fails to prune decision trees according to predetermined stopping criteria, they continue to grow indefinitely. Tree growth stops such as the Gini Diversity Index, RMSE and MSE are frequently utilized. Trees with appropriate predictions are picked in the final RF model, and trees with low predictive outcomes are excluded. The overfitting problem of the single DT model is eliminated by randomly selecting predictor parameters and the final set of DTs [50,51]. Figure 4 illustrates the random forest’s structure.

3.3. AdaBoost Algorithm

AdaBoost or adaptive boosting is a sequential ensemble technique which is based on the principle of developing several weak learners using different training sub-sets drawn randomly from the original training dataset [52,53]. During each training, weights are assigned which are used when learning each hypothesis. The weights are used for computation of the error of the hypothesis on the dataset and are an indicator of the comparative importance of each instance. The weights are recalculated after every iteration, such that incorrectly classified instances by the last hypothesis receive higher weights. This enables the algorithm to focus on more difficult-to-learn instances. Assigning revised weights to the incorrectly classified instances is the most vital task of the algorithm. Unlike in classification, in regression, the instances are not correct or incorrect, rather they constitute a real-value error. By comparing the computed error to a predefined threshold prediction error, it can be labeled as an error or not an error and thus, the AdaBoost classifier can be used. Instances with larger errors on previous learners are more likely (i.e., higher probability) to be selected for training the subsequent base learner. Finally, weighted average or median is used to provide an ensemble prediction of the individual base learner predictions [54].

3.4. Support Vector Machine (SVM) Algorithm

Vapnik invented the SVM in 1995 [55], and it is a popular and successful learning algorithm for the classification of linear and nonlinear regression problems. The SVM algorithm delivers reliable prediction outcomes and is practicable for high-dimensional feature spaces, is robust and has good noise resistance [56,57]. In many disciplines, many effective SVM implementations with classification and regression issues have been documented [58,59,60]. The following is a summary of SVM’s basic theory.
As illustrated in Figure 5, a training set {(uk, vk), k = 1,2, … …, n} is chosen for an SVM model, where uk = [u1k, u2k, … …, unk] ∈ R n h is the input data, vk R n m is the output data corresponding to uk, and n is the number of training samples. The goal of the SVM is to identify an optimal hyperplane function f(x) (defined by the weight vector w and the offset b), that passes through all data items with the insensitive loss coefficient ε (based on two supporting hyperplanes, w.u – b = ε and w.u – b = −ε).
The function f(u) in nonlinear regression is determined as follows:
f ( u ) = i = 1 n ( α i α i ) K ( u i , u j ) + b )
with
i = 1 n ( α i α i ) = 0 ,     C α i ,   α i 0 , i
The penalty constant C is used to manage the penalty error, α i ,   α i are the Lagrange multipliers, and K (ui, uj) is the kernel function defined as follows:
K ( u i , u j ) = < Φ ( u i ) , Φ ( u j ) =  
The mapping function F is a nonlinear mapping function. The most often used kernel functions are linear, polynomial, sigmoid, and Gaussian functions:
Linear kernel function:
K ( u i , u j ) = u i u j
Polynomial kernel function:
K ( u i , u j ) = ( γ u i u j + c ) d    
Sigmoid kernel function:
K ( u i , u j ) = t a n h   ( γ u i u j + c ) d  
Gaussian kernel function:
K ( u i , u j ) = e x p ( γ ( u i u j ) 2 ) d

3.5. Decision Tree (DT) Algorithm

A decision tree is a tool with a tree-like structure that predicts likely outcomes, resource costs, utility costs, and potential consequences. One of the benefits of the machine learning approach over traditional statistical approaches such as regression is that they can handle more than two-dimensional data. For data-driven prediction analysis of diverse geotechnical problems, many researchers have adopted the tree-based approach [20,61,62]. As a result, tree-based ML techniques, such as DT, were used to build models and identify the key predictors of pile–soil friction in this work. DT can be seen graphically, showing specific decision requirements as well as the complicated branching that occurs in a constructed decision. This is one of the most popular and commonly used supervised learning techniques for forecasting model accuracy.
DT is capable of performing all tasks including recognition, classification, and prediction. DT is a “tree”-shaped structure made up of a succession of questions, each of which is described by a set of parameters. Roots, branches, and leaves comprise a real tree. Similarly, the graph for DT is comprised of nodes, which are leaves, and branches, which represent connections between nodes [63]. A variable is chosen as a root, also known as the initial node, during the DT process. By reference to the appointed features, the initial node is divided into many internal nodes. DT is a top-down tree, meaning the roots are at the very top. Roots, branches, and nodes are the end products of the branches [64]. Each node can be divided into two branches and each node has a relationship to a specific characteristic and branches that have been specified by a specific range of input. Figure 6 depicts a flowchart linked to the DT approach.

4. Construction of Prediction Models

Orange software was used to create the proposed models for predicting pile bearing capacity. Orange is an open-source software package. Machine learning, preprocessing, and visualization methods are included in the default installation, which is divided into six widget sets i.e., Data, Visualize, Classify, Regression, Evaluate and Unsupervised. Orange is visual programming software for machine learning, visualization, data mining, data analysis.
The predictor variables were provided via an input set (x) defined by x = {D, X1, X2, X3, Xp, Xg, Xt, Xm, NS and Nt}, while the target variable (y) is Pu. The most important task in every modeling step is to pick the right number of training and testing datasets. As a result, 70% of the whole data was chosen to generate the models in this study, with the developed models being tested on the remaining data. On the other way, 140 and 60 sets were utilized for creating and testing the models, respectively. All models (XGBoost, AdaBoost, RF, DT, and SVM) were tweaked to optimize the Pu prediction using the trial-and-error process. Figure 7 shows how the prediction models were built.

4.1. Hyperparameter Optimization

ML algorithms have parameters that must be tuned. The optimization procedure seeks to find ideal settings for XGBoost, AdaBoost, RF, DT, and SVM to achieve accurate prediction. This study tunes various critical parameters in the XGBoost, AdaBoost, RF, DT and SVM, as well as clarifies the definitions of these hyperparameters. The tuning parameters for the models were chosen and then changed in the trials until the best metrics shown in Table 3 were achieved.

4.2. Model Evaluation Indexes

The results of the proposed models are evaluated using R2, MAE, RMSE, MARE, NSE and RSR, as more commonly used criteria in the literature. The following equations are used to calculate these metrics:
R 2 = 1 i = 1 n ( x i x ^ i ) 2 i = 1 n ( x i x ^ ) 2
M A E = 1 n i = 1 n ( x i x ^ i )
R M S E = 1 n i = 1 n ( x i x ^ i ) 2
M A R E = 1 n i = 1 n | x i x ^ i x i × 100 |
N S E = 1 i = 1 n ( x i x ^ i ) 2 i = 1 n ( x i x ¯ ) 2
R S R = i = 1 n ( x i x ^ i ) 2 i = 1 n ( x i x ¯ ) 2
where n denotes the number of points, x i and x ^ i denotes the actual and expected outputs of i-th sample, respectively; x ¯ is data averaged actual output. R2 is a number that ranges from 0 to 1, a higher R2 value indicates a more efficient model. The model is considered effective when R2 is more than 0.8 and close to 1 [22]. The mean squared difference between projected outputs and targets is the criterion RMSE, and the mean magnitude of errors is the criterion MAE, RMSE and MAE are similar in that the closer these criterion values of these errors are to 0, the better the model’s performance. In circumstances where the MAE and RMSE are minimal, the model’s accuracy is greater. Models yielded the lowest MARE value, indicating that the model has superior predictive power. The RSR ranges from 0 to a considerable positive number. Lower RSR indicates lower RMSE, indicating that the model is more productive. RSR and NSE categorization ranges as very good, good, satisfactory, and unsatisfactory with ranges 0   R S R   0.5 , 0.5 < R S R     0.6 , 0.6 < R S R     0.7 , R S R > 0.7 and 0.75 < N S E     1 , 0.65 < N S E     0.75 , 0.5 < N S E     0.65 , and N S E     0.5 , respectively [65].
In addition, the Taylor diagram was used to compare the model’s performance visually [66]. Taylor diagram shows how similar patterns are and how closely a model pattern relates to reference. The standard deviation (σ), R2, and the RMSE are three equivalent model performance statistics that can be shown on a two-dimensional plot using the law of cosines. Taylor diagram is the best method for comparing the performance of various models in particular.

5. Result and Discussion

5.1. Comparison of Models

This section evaluates the model’s efficacy. Figure 8 and Figure 9 depict the training and testing dataset’s prediction performance in regression form, respectively, while Table 4 and Table 5 provide a summary of the relevant data.
In terms of training, the XGBoost model produced the best prediction results (i.e., R2 = 0.971, MAE = 47.518 and RMSE = 66.844) compared to AdaBoost (i.e., R2 = 0.957, MAE = 56.671 and RMSE = 82.495), RF (i.e., R2 = 0.952, MAE = 58.366 and RMSE = 79.240), DT (i.e., R2 = 0.932, MAE = 68.912 and RMSE = 94.304) and SVM (i.e., R2 = 0.887, MAE = 88.801 and RMSE = 123.375). This is also confirmed by the results of MARE, NSE and RSR in Table 4. In training part, XGBoost produced lesser MARE, NSE and RSR values compared to AdaBoost, RF, DT and SVM.
In the testing part, the XGBoost model had the best prediction results with respect to R2, MAE, RMSE, MARE, NSE and RSR (i.e., R2 = 0.955, MAE = 59.929, RMSE = 80.653, MARE = 6.6, NSE = 0.950, and RSR = 0.225) compared to AdaBoost (i.e., R2 = 0.950, MAE = 70.383, RMSE = 90.665, MARE = 8.252, NSE = 0.936, and RSR = 0.253), RF (i.e., R2 = 0.945, MAE = 69.030, RMSE = 86.348, MARE = 8.014, NSE = 0.942, and RSR = 0.241), DT (i.e., R2 = 0.0.925, MAE = 74.450, RMSE = 99.822, MARE = 8.775, NSE = 0.923, and RSR = 0.278) and SVM (i.e., R2 = 0.878, MAE = 98.320, RMSE = 128.027, MARE = 10.991, NSE = 0.873, and RSR = 0.357) as shown in Table 5.
Comparing the above performance measures the proposed XGBoost model performed better than the AdaBoost, RF, DT and SVM. From these statistical analysis and prediction capabilities, we can state that the XGBoost model has good accuracy prediction for pile bearing capacity.
The sensitivity results of the XGBoost model were assessed using Yang and Zang’s [67] method for assessing the impact of input variables on Pu. This approach, which has been used in several investigations [22,28,68,69,70], is as follows:
r i j = k = 1 n ( x i m × x o m ) k = 1 n x i m 2 k = 1 n x o m 2  
as n represents the number of values (i.e., 140); x i m and x o m denotes input and output variables, respectively. For each input parameter, the r i j value ranges from zero to one, with the greatest r i j values indicating the efficient output variable (i.e., Pu).
Figure 10 shows the r i j scores for all input variables. Figure 10 demonstrates that SPT blow count at pile shaft (NS) ( r i j = 0.985) has the greatest effect on the Pu.
With the use of the Taylor diagram (see Figure 11), we investigated the model’s efficiency further. The better the performance, the closer each produced model’s point is to the observed point location. The models demonstrated the best predictive capability, while the XGBoost method had a greater correlation and a lesser RMSE.

5.2. Comparison with Other Researchers

Table 6 shows some findings from a study on machine learning applications on pile bearing capacity. On the test data set, the expected efficiency of ML algorithms in foundation engineering having predictive outcomes of foundation load is mostly ranging R2 from 0.71 to 0.918, according to the results of previous studies while in the present study it is 0.955. However, due to the use of different datasets, a comparison between these results is unwarranted. A project that uses different datasets is needed to give a generalized model to foundation engineering.

6. Conclusions

Pile bearing capacity values were estimated in this paper using five models. The prediction model was built with ten input parameters and one output parameter. The modeling results show that the XGBoost model has the best capability for accurate prediction of Pu when compared to other models such as AdaBoost, RF, DT and SVM. The following are some of the major findings of this study:
  • In testing phase, the XGBoost model (R2 = 0.955, MAE = 59.929, RMSE = 80.653, MARE = 6.6, NSE = 0.950, and RSR = 0.225) has the highest performance capability as compared to other soft computing techniques considered in this study i.e., AdaBoost, RF, DT and SVM as well as the models used in the literature.
  • Sensitivity analysis results show that SPT blow count at pile shaft (NS) was the most important parameter in predicting pile bearing capacity.
  • Taylor diagram also verified that all the models are good but the predictive power of the XGBoost algorithm had a higher correlation and lower RMSE.
  • Based on the results and analysis the XGBoost model can also be applied to solve a variety of geotechnical engineering problems.
Furthermore, the XGBoost technique has the advantage of being easily updated, it is obvious that the proposed model is open to further development, and that the collection of more data will result in significantly stronger prediction capability, avoiding the requirement for expertise and time to update an existing design aid or equation.

Author Contributions

Conceptualization, M.A. (Mahmood Ahmad) and M.A. (Maaz Amjad); methodology, M.A. (Maaz Amjad), M.A. (Mahmood Ahmad) and I.A.; software, M.A. (Maaz Amjad) and M.A. (Mahmood Ahmad); validation, P.W., P.K. (Paweł Kami´nski), M.A. (Maaz Amjad) and U.A.; formal analysis, M.A. (Maaz Amjad) and I.A.; investigation, P.W., P.K., U.A. and M.A. (Mahmood Ahmad); resources, P.K.; data curation, M.A. (Maaz Amjad) and I.A.; writing—original draft preparation, M.A. (Maaz Amjad); writing—review and editing, M.A. (Maaz Amjad), I.A., M.A. (Mahmood Ahmad) and U.A.; supervision, M.A. (Mahmood Ahmad), P.W. and P.K.; project administration, M.A. (Mahmood Ahmad) and P.W.; funding acquisition, P.W. and P.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are included within the article.

Acknowledgments

The writers gratefully acknowledge Tuan Anh Pham from the University of Transport Technology in Vietnam, who provided pile load test results conducted on 200 reinforced concrete piles at the test site in Ha Nam province–Vietnam for this study.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviation

SymbolExplanation
PuPile bearing capacity
MLMachine learning
XGBoostExtreme gradient boosting
AdaBoostAdaptive boosting
RFRandom forest
DTDecision tree
SVMSupport vector machine
ANNArtificial neural network
ANFISadaptive neuro-fuzzy inference system
GAGenetic algorithm
BPNNBackpropagation neural network
GBDTGradient boosting decision tree
LightGBMLight gradient boosting machine
DLNNDeep Learning Neural Network
PSO-ANNParticle swarm optimization—ANN
GPRGaussian process regression
R2Coefficient of determination
MAEMean absolute error
MSEMean square error
RMSERoot mean square error
MAREMean absolute relative error
NSENash–Sutcliffe model efficiency
RSRRelative strength ratio
SPTStandard penetration test
CPTCone penetration test
DDiameter
X1Depth of first layer of soil embedded
X2Depth of second layer of soil embedded
X3Depth of third layer of soil embedded
XpPile top elevation
XgGround elevation
XtExtra pile top elevation
XmPile tip elevation
NSSPT blow count at pile shaft
NtSPT blow count at pile tip

Appendix A

Table A1. Data Catalog.
Table A1. Data Catalog.
S. No.DX1X2X3XpXgXtXmNsNtPu
Unitmmmmmmmmm--kN
14003.4580.32.953.652.9514.711.757.591017.9
24004.25812.153.562.1615.413.257.671152
34004.2581.022.153.582.1615.4213.277.681344
44004.2580.12.153.583.0814.512.357.141551
54004.3581.062.053.552.0915.4613.417.661321
63003.45.2503.43.493.4412.058.656.75559.8
74004.2581.022.153.582.1615.4213.277.681248
83003.45.1803.43.363.3811.988.586.73559.8
94004.757.2502.053.623.5714.05126.731425
103003.45.2503.43.473.4212.058.656.75559.8
113003.45.203.43.423.42128.66.73660.6
124003.455.2403.353.443.412.048.696.721240
134004.3581.072.053.522.0515.4713.427.671425
144004.12.1702.73.72.738.976.274.92661.6
154003.555.3903.253.443.2512.198.946.721083
164004.25812.153.562.1615.413.257.671152
174003.47.303.43.613.5114.110.77.281115.2
183003.45.203.43.433.43128.66.73610.7
193003.45.203.43.423.42128.66.73661.6
204004.11.802.73.392.798.65.94.64620
214003.4580.32.953.662.9614.711.757.59960
223003.45.2703.43.493.4212.078.676.75559.8
234004.25812.153.562.1615.413.257.671248
244004.657.402.153.593.3914.212.056.801551
254004.1202.73.562.768.86.14.80620
264004.3580.32.053.452.7514.712.657.221473
274004.3581.032.053.482.0515.4313.387.651318
284004.3581.012.053.462.0515.4113.367.641473
294004.11.7202.73.272.758.525.824.57423.9
304003.47.2803.43.483.414.0810.687.271318
314004.3581.052.053.552.115.4513.47.661221.5
323003.45.203.43.433.43128.66.73559.8
334004.2580.962.153.532.1715.3613.217.651344
344004.657.3502.153.553.414.15126.791392
354003.857.502.953.683.3814.311.357.131425
363003.45.3503.43.573.4212.158.756.78661.6
374004.757.502.053.63.314.312.256.791425
384004.3580.952.053.412.0615.3513.37.601323.2
394004.2580.92.153.572.2715.313.157.611473
404004.3580.962.053.422.0615.3613.317.611244
414004.3581.052.053.54.3515.4513.47.661297.8
424004.657.402.153.593.3914.212.056.801551
434004.657.202.153.583.581411.856.751551
444004.1202.73.52.78.86.14.80610.7
454004.3580.952.053.442.0915.3513.37.601152
464004.0580.662.353.462.415.0612.717.561318
474003.580.22.93.512.9114.611.77.50960
484004.3580.982.053.482.115.3813.337.621224.8
494004.657.502.153.593.2914.312.156.821551
504004.657.4602.153.563.314.2612.116.811551
514004.2580.22.153.552.9514.612.457.201392
524004.2581.022.153.582.1615.4213.277.681344
534003.47.2403.43.443.414.0410.647.26967
544004.2580.992.153.542.1515.3913.247.661248
554004.657.202.153.583.581411.856.751392
564004.1202.73.542.748.86.14.80712.5
574004.656.302.153.554.4513.110.956.531440
583003.45.203.43.453.45128.66.73559.8
593003.45.303.43.53.412.18.76.76661.6
604004.2580.962.153.542.1815.3613.217.651395
614004.2581.022.153.582.1615.4213.277.681344
624004.657.402.153.593.3914.212.056.801551
634003.47.3503.43.563.4114.1510.757.291052.4
644004.3581.072.053.522.0515.4713.427.671082.3
654004.757.602.053.443.0414.412.356.811473
664004.2580.92.153.562.2615.313.157.611395
673003.45.3503.43.573.4212.158.756.78661.6
684003.580.182.93.52.9214.5811.687.491032.4
693003.45.203.43.423.42128.66.73559.8
704003.47.3303.43.553.4214.1310.737.281094.25
714004.25812.153.552.1515.413.257.671248
724003.4580.22.953.522.9214.611.657.52967
734003.580.172.93.472.914.5711.677.48960
744003.4580.142.953.522.9814.5411.597.48885
754003.4580.072.953.422.9514.4711.527.441240
764005.46.302.153.521.0613.114.75.501056
773003.45.203.43.433.43128.66.73600.7
783003.45.303.43.523.4212.18.76.76508.9
794003.555.3603.253.413.2512.168.916.71930
804004.3581.182.053.662.0815.5813.537.731056
814004.1202.73.522.728.86.14.80610.7
823003.45.2503.43.493.4412.058.656.75610.7
834004.25812.153.552.1515.413.257.671344
843003.45.203.43.383.38128.66.73610.7
854004.2580.92.153.592.2915.313.157.611473
864004.11.8502.73.352.78.655.954.68508.9
873003.45.203.43.433.43128.66.73661.6
884004.2580.942.153.542.215.3413.197.641395
894004.2580.92.153.592.2915.313.157.611551
904004.757.2502.053.653.614.05126.731425
914004.2581.022.153.582.1615.4213.277.681152
924004.3581.052.053.532.0815.4513.47.661473
934003.4580.142.953.522.9814.5411.597.48885
944004.11.902.73.432.738.764.72620
954004.3580.972.053.422.0515.3713.327.611317
964004.656.4902.153.594.313.2911.146.581551
974003.47.3103.43.563.4514.1110.717.281032.4
983003.45.2503.43.483.4312.058.656.75610.7
994003.4580.192.953.562.9714.5911.647.521318
1004003.456.2903.353.443.3513.099.747.021240
1013003.45.2403.43.493.4512.048.646.75610.7
1024004.2580.72.153.582.4815.112.957.501392
1033003.45.2503.43.473.4212.058.656.75585.4
1044004.25812.153.562.1615.413.257.671152
1054004.11.802.73.322.728.65.94.64559.8
1064003.47.303.43.493.3914.110.77.281068.8
1074004.35812.053.452.0515.413.357.631119.7
1084003.47.3103.43.543.4314.1110.717.281032.8
1094003.4580.12.953.543.0414.511.557.461017.9
1103003.45.203.43.483.48128.66.73611.6
1114004.757.602.053.493.0914.412.356.811473
1124004.3581.042.053.522.0815.4413.397.651321
1134003.580.212.93.482.8714.6111.717.511032.4
1144004.657.202.153.553.551411.856.751392
1154004.3581.082.053.532.0515.4813.437.671248
1163003.45.2503.43.463.4112.058.656.75661.6
1173003.45.203.43.413.41128.66.73610.7
1184004.3581.12.053.552.0515.513.457.691425
1194004.3580.052.053.583.1314.4512.47.071344
1204004.12.0802.73.632.758.886.184.86432
1213003.45.2503.43.483.4312.058.656.75559.8
1224003.857.3502.953.643.4914.1511.27.091425
1233003.45.2503.43.483.4312.058.656.75508.9
1244004.657.502.153.593.2914.312.156.821551
1253003.45.303.43.53.412.18.76.76559.8
1263003.45.3203.43.553.4312.128.726.77661.6
1273003.45.2503.43.483.4312.058.656.75559.8
1284003.580.162.93.482.9214.5611.667.47960
1294004.657.502.153.553.2514.312.156.821551
1304004.757.502.053.453.1514.312.256.791297.8
1313003.45.203.43.423.42128.66.73610.7
1324004.3581.012.053.462.0515.4113.367.641550
1333003.45.203.43.413.41128.66.73610.7
1344003.47.303.43.543.4414.110.77.28967
1354004.2581.032.153.582.1515.4313.287.691248
1363003.45.2503.43.463.4112.058.656.75559.8
1373003.45.303.43.513.4112.18.76.76661.6
1384004.2580.42.153.552.7514.812.657.321392
1394004.3580.952.053.412.0615.3513.37.601110.6
1403003.45.203.43.43.4128.66.73559.8
1414003.857.302.953.683.5814.111.157.081440
1424004.12.0802.73.582.78.886.184.86480
1434004.4581.181.953.58215.5813.637.691032.4
1443003.45.203.43.43.4128.66.73559.8
1453003.45.203.43.433.43128.66.73661.6
1463003.45.2503.43.463.4112.058.656.75407.2
1474003.4580.222.953.572.9514.6211.677.531318
1484004.2581.012.153.572.1615.4113.267.681248
1494003.47.303.43.53.414.110.77.28958
1504004.12.202.73.722.7296.34.94610.7
1514004.3581.022.053.474.0515.4213.377.641318
1524004.2580.92.153.532.2315.313.157.611395
1534004.2580.42.153.592.7914.812.657.321551
1543003.45.2403.43.483.4412.048.646.75559.8
1554004.2580.42.153.552.7514.812.657.321392
1563003.45.2503.43.463.4112.058.656.75661.6
1574004.0580.72.353.472.3715.112.757.581318
1583003.45.2303.43.443.4112.038.636.74585.35
1594004.3580.72.053.492.3915.113.057.461392
1604004.25812.153.572.1715.413.257.671248
1614004.25812.153.582.1815.413.257.671395
1624004.25812.153.562.1615.413.257.671395
1634004.2581.012.153.572.1615.4113.267.681248
1644004.2580.12.153.533.0314.512.357.141551
1654003.580.172.93.482.9114.5711.677.481056
1664004.2581.022.153.582.1615.4213.277.681248
1673003.45.2503.43.463.4112.058.656.75532.4
1684004.3580.82.053.452.2515.213.157.521392
1693003.45.203.43.453.45128.66.73610.7
1704004.2580.982.153.542.1615.3813.237.661344
1714004.25812.153.562.1615.413.257.671344
1724003.4580.252.953.62.9514.6511.77.55960
1734004.657.2402.153.543.514.0411.896.761551
1744004.2580.92.153.582.2815.313.157.611395
1754003.47.303.43.53.414.110.77.28900
1764003.47.403.43.613.4114.210.87.301088.8
1774004.2580.12.153.543.0414.512.357.141551
1784003.47.2303.43.433.414.0310.637.26960
1793003.45.303.43.523.4212.18.76.76610.7
1804004.1202.73.552.758.86.14.80610.7
1814004.2581.032.153.582.1515.4313.287.691248
1824003.4580.122.953.472.9514.5211.577.471318
1834004.25812.153.582.1815.413.257.671395
1844004.3581.112.053.562.0515.5113.467.691128.6
1854004.457.2102.353.412.414.0111.666.831318
1864004.657.3802.153.583.414.1812.036.791551
1874004.25812.153.562.1615.413.257.671248
1884004.2580.22.153.582.9814.612.457.201551
1894004.657.602.153.583.1814.412.256.841446
1903003.45.2203.43.443.4212.028.626.74617
1914004.757.402.053.523.3214.212.156.761425
1924004.657.402.153.593.3914.212.056.801392
1934003.47.303.43.613.5114.110.77.281115.2
1943003.45.2503.43.493.4412.058.656.75559.8
1953003.45.2503.43.463.4112.058.656.75559.8
1964004.25812.153.582.1815.413.257.671395
1973003.45.1803.43.383.411.988.586.73559.8
1984004.2580.912.153.562.2515.3113.167.621473
1994004.0580.72.353.482.3815.112.757.581238
2004004.12.0102.73.532.728.816.114.80528

References

  1. Momeni, E.; Nazir, R.; Armaghani, D.J.; Maizir, H. Application of artificial neural network for predicting shaft and tip resistances of concrete piles. Earth Sci. Res. J. 2015, 19, 85–93. [Google Scholar] [CrossRef]
  2. Drusa, M.; Gago, F.; Vlček, J. Contribution to Estimating Bearing Capacity of Pile in Clayey Soils. Civ. Environ. Eng. 2016, 12, 128–136. [Google Scholar] [CrossRef] [Green Version]
  3. Meyerhof, G.G. Bearing Capacity and Settlement of Pile Foundations. J. Geotech. Eng. Div. 1976, 102, 197–228. [Google Scholar] [CrossRef]
  4. Shooshpasha, I.; Hasanzadeh, A.; Taghavi, A. Prediction of the axial bearing capacity of piles by SPT-based and numerical design methods. Int. J. GEOMATE 2013, 4, 560–564. [Google Scholar] [CrossRef]
  5. Chai, X.J.; Deng, K.; He, C.F.; Xiong, Y.F. Laboratory model tests on consolidation performance of soil column with drained-timber rod. Adv. Civ. Eng. 2021, 2021, 6698894. [Google Scholar] [CrossRef]
  6. ASTM. American Society for Testing and Materials—ASTM D4945-08 Standard Test Method for High-Strain Dynamic Testing of Deep Foundations; ASTM: West Conshohocken, PA, USA, 2008; Volume 1, p. 10. [Google Scholar]
  7. Schmertmann, J. Guidelines for Cone Penetration Test: Performance and Design; (No. FHWA-TS-78-209); Federal Highway Administration: Washington, DC, USA, 1978.
  8. Budi, G.S.; Kosasi, M.; Wijaya, D.H. Bearing capacity of pile foundations embedded in clays and sands layer predicted using PDA test and static load test. Procedia Eng. 2015, 125, 406–410. [Google Scholar] [CrossRef]
  9. Kozłowski, W.; Niemczynski, D. Methods for Estimating the Load Bearing Capacity of Pile Foundation Using the Results of Penetration Tests—Case Study of Road Viaduct Foundation. Procedia Eng. 2016, 161, 1001–1006. [Google Scholar] [CrossRef] [Green Version]
  10. Birid, K.C. Evaluation of Ultimate Pile Compression Capacity from Static Pile Load Test Results. In International Congress and Exhibition “Sustainable Civil Infrastructures: Innovative Infrastructure Geotechnology”; Springer: Cham, Switzerland, 2018; pp. 1–14. [Google Scholar] [CrossRef]
  11. Ma, B.; Li, Z.; Cai, K.; Liu, M.; Zhao, M.; Chen, B.; Chen, Q.; Hu, Z. Pile-Soil Stress Ratio and Settlement of Composite Foundation Bidirectionally Reinforced by Piles and Geosynthetics under Embankment Load. Adv. Civ. Eng. 2021, 2021, 5575878. [Google Scholar] [CrossRef]
  12. Tang, Y.; Huang, S.; Tao, J. Geo-Congress 2020; GSP 320 121; ASCE: Reston, VA, USA, 2020; pp. 121–131. [Google Scholar] [CrossRef]
  13. Nurdin, S.; Sawada, K.; Moriguchi, S. Design Criterion of Reinforcement on Thick Soft Clay Foundations of Traditional Construction Method in Indonesia. MATEC Web Conf. 2019, 258, 03010. [Google Scholar] [CrossRef]
  14. Momeni, E.; Maizir, H.; Gofar, N.; Nazir, R. Comparative study on prediction of axial bearing capacity of driven piles in granular materials. J. Teknol. 2013, 61, 15–20. [Google Scholar] [CrossRef] [Green Version]
  15. Lopes, F.R.; Laprovitera, H. Prediction of the Bearing Capacity of Bored Piles from Dynamic Penetration Tests. In Proceedings of the 1st International Geoteclmical Seminar on Deep Foundations on Bored and Auger Piles, Ghent, Belgium, 7–10 June 1988; pp. 537–540. [Google Scholar]
  16. Decourt, L. Prediction of load-settlement relationships for foundations on the basis of the SPT. In Proceedings of the Ciclo de Conferencias Internationale, Leonardo Zeevaert, UNAM, Mexico City, Mexico, 1995; pp. 85–104. [Google Scholar]
  17. Pham, T.A.; Ly, H.-B.; Tran, V.Q.; Van Giap, L.; Vu, H.-L.T.; Duong, H.-A.T. Prediction of Pile Axial Bearing Capacity Using Artificial Neural Network and Random Forest. Appl. Sci. 2020, 10, 1871. [Google Scholar] [CrossRef] [Green Version]
  18. Ahmad, M.; Ahmad, F.; Huang, J.; Iqbal, M.J.; Safdar, M.; Pirhadi, N. Probabilistic evaluation of CPT-based seismic soil liquefaction potential: Towards the integration of interpretive structural modeling and bayesian belief network. Math. Biosci. Eng. 2021, 18, 9233–9252. [Google Scholar] [CrossRef] [PubMed]
  19. Ahmad, M.; Tang, X.-W.; Ahmad, F.; Jamal, A. Assessment of Soil Liquefaction Potential in Kamra, Pakistan. Sustainability 2018, 10, 4223. [Google Scholar] [CrossRef] [Green Version]
  20. Ahmad, M.; Al-Shayea, N.A.; Tang, X.-W.; Jamal, A.; Al-Ahmadi, H.M.; Ahmad, F. Predicting the Pillar Stability of Underground Mines with Random Trees and C4.5 Decision Trees. Appl. Sci. 2020, 10, 6486. [Google Scholar] [CrossRef]
  21. Liu, Q.; Cao, Y.; Wang, C. Prediction of Ultimate Axial Load-Carrying Capacity for Driven Piles Using Machine Learning Methods. In Proceedings of the 3rd Information Technology, Networking, Electronic and Automation Control Conference, Chengdu, China, 15–17 March 2019; Institute of Electrical and Electronics Engineers: Piscataway, NJ, USA, 2019; pp. 334–340. [Google Scholar] [CrossRef]
  22. Ahmad, M.; Ahmad, F.; Wróblewski, P.; Al-Mansob, R.A.; Olczak, P.; Kamiński, P.; Safdar, M.; Rai, P. Prediction of Ultimate Bearing Capacity of Shallow Foundations on Cohesionless Soils: A Gaussian Process Regression Approach. Appl. Sci. 2021, 11, 10317. [Google Scholar] [CrossRef]
  23. Liang, W.; Luo, S.; Zhao, G.; Wu, H. Predicting hard rock pillar stability using GBDT, XGBoost, and LightGBM algorithms. Mathematics 2020, 8, 765. [Google Scholar] [CrossRef]
  24. Pham, B.T.; Nguyen, M.D.; Van Dao, D.; Prakash, I.; Ly, H.-B.; Le, T.-T.; Ho, L.S.; Nguyen, K.T.; Ngo, T.Q.; Hoang, V.; et al. Development of artificial intelligence models for the prediction of Compression Coefficient of soil: An application of Monte Carlo sensitivity analysis. Sci. Total Environ. 2019, 679, 172–184. [Google Scholar] [CrossRef]
  25. Ahmad, M.; Tang, X.-W.; Qiu, J.-N.; Ahmad, F. Evaluating Seismic Soil Liquefaction Potential Using Bayesian Belief Network and C4.5 Decision Tree Approaches. Appl. Sci. 2019, 9, 4226. [Google Scholar] [CrossRef] [Green Version]
  26. Ahmad, M.; Kamiński, P.; Olczak, P.; Alam, M.; Iqbal, M.; Ahmad, F.; Sasui, S.; Khan, B. Development of Prediction Models for Shear Strength of Rockfill Material Using Machine Learning Techniques. Appl. Sci. 2021, 11, 6167. [Google Scholar] [CrossRef]
  27. Ahmad, M.; Hu, J.-L.; Hadzima-Nyarko, M.; Ahmad, F.; Tang, X.-W.; Rahman, Z.; Nawaz, A.; Abrar, M. Rockburst Hazard Prediction in Underground Projects Using Two Intelligent Classification Techniques: A Comparative Study. Symmetry 2021, 13, 632. [Google Scholar] [CrossRef]
  28. Ahmad, M.; Hu, J.-L.; Ahmad, F.; Tang, X.-W.; Amjad, M.; Iqbal, M.; Asim, M.; Farooq, A. Supervised Learning Methods for Modeling Concrete Compressive Strength Prediction at High Temperature. Materials 2021, 14, 1983. [Google Scholar] [CrossRef] [PubMed]
  29. Goh, A.T.C.; Kulhawy, F.H.; Chua, C.G. Bayesian Neural Network Analysis of Undrained Side Resistance of Drilled Shafts. J. Geotech. Geoenviron. Eng. 2005, 131, 84–93. [Google Scholar] [CrossRef]
  30. Goh, A.T.C. Back-propagation neural networks for modeling complex systems. Artif. Intell. Eng. 1995, 9, 143–151. [Google Scholar] [CrossRef]
  31. Shahin, M.A.; Jaksa, M.B. Neural network prediction of pullout capacity of marquee ground anchors. Comput. Geotech. 2005, 32, 153–163. [Google Scholar] [CrossRef]
  32. Shahin, M.A. Intelligent computing for modeling axial capacity of pile foundations. Can. Geotech. J. 2010, 47, 230–243. [Google Scholar] [CrossRef] [Green Version]
  33. Shahin, M.A. Load–settlement modeling of axially loaded steel driven piles using CPT-based recurrent neural networks. Soils Found. 2014, 54, 515–522. [Google Scholar] [CrossRef] [Green Version]
  34. Shahin, M.A. State-of-the-art review of some artificial intelligence applications in pile foundations. Geosci. Front. 2016, 7, 33–44. [Google Scholar] [CrossRef] [Green Version]
  35. Nawari, N.O.; Liang, R.; Nusairat, J. Artificial intelligence techniques for the design and analysis of deep foundations. Electron. J. Geotech. Eng. 1999, 4, 1–21. [Google Scholar]
  36. Momeni, E.; Nazir, R.; Armaghani, D.J.; Maizir, H. Prediction of pile bearing capacity using a hybrid genetic algorithm-based ANN. Measurement 2014, 57, 122–131. [Google Scholar] [CrossRef]
  37. Kordjazi, A.; Nejad, F.P.; Jaksa, M. Prediction of ultimate axial load-carrying capacity of piles using a support vector machine based on CPT data. Comput. Geotech. 2014, 55, 91–102. [Google Scholar] [CrossRef]
  38. Pham, T.A.; Tran, V.Q.; Vu, H.L.T.; Ly, H.B. Design deep neural network architecture using a genetic algorithm for estimation of pile bearing capacity. PLoS ONE 2020, 15, e0243030. [Google Scholar] [CrossRef] [PubMed]
  39. Tama, B.A.; Rhee, K.H. An in-depth experimental study of anomaly detection using gradient boosted machine. Neural Comput. Appl. 2019, 31, 955–965. [Google Scholar] [CrossRef]
  40. Sun, R.; Wang, G.; Zhang, W.; Hsu, L.T.; Ochieng, W.Y. A gradient boosting decision tree based GPS signal reception classification algorithm. Appl. Soft Comput. 2020, 86, 105942. [Google Scholar] [CrossRef]
  41. Lombardo, L.; Cama, M.; Conoscenti, C.; Märker, M.; Rotigliano, E. Binary logistic regression versus stochastic gradient boosted decision trees in assessing landslide susceptibility for multiple-occurring landslide events: Application to the 2009 storm event in Messina (Sicily, southern Italy). Nat. Hazards 2015, 79, 1621–1648. [Google Scholar] [CrossRef]
  42. Sachdeva, S.; Bhatia, T.; Verma, A.K. GIS-based evolutionary optimized Gradient Boosted Decision Trees for forest fire susceptibility mapping. Nat. Hazards 2018, 92, 1399–1418. [Google Scholar] [CrossRef]
  43. Kotsiantis, S.B. Decision trees: A recent overview. Artif. Intell. Rev. 2013, 39, 261–283. [Google Scholar] [CrossRef]
  44. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  45. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  46. Javadi, A.A.; Rezania, M.; Nezhad, M.M. Evaluation of liquefaction induced lateral displacements using genetic programming. Comput. Geotech. 2006, 33, 222–233. [Google Scholar] [CrossRef]
  47. van Vuren, T. Modeling of transport demand—Analyzing, calculating, and forecasting transport demand. Transp. Rev. 2020, 40, 115–117. [Google Scholar] [CrossRef]
  48. Song, Y.; Gong, J.; Gao, S.; Wang, D.; Cui, T.; Li, Y.; Wei, B. Susceptibility assessment of earthquake-induced landslides using Bayesian network: A case study in Beichuan, China. Comput. Geosci. 2012, 42, 189–199. [Google Scholar] [CrossRef]
  49. Kaggle. Your Machine Learning and Data Science Community. Available online: https://www.kaggle.com/ (accessed on 2 December 2021).
  50. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  51. Ahmad, M.W.; Mourshed, M.; Rezgui, Y. Trees vs Neurons: Comparison between random forest and ANN for high-resolution prediction of building energy consumption. Energy Build. 2017, 147, 77–89. [Google Scholar] [CrossRef]
  52. Freund, Y.; Schapire, R.E. Experiments with a New Boosting Algorithm. In Proceedings of the 13th International Conference on International Conference on Machine Learning, Bari, Italy, 3–6 July 1996; pp. 148–156. [Google Scholar]
  53. Schapire, R.E. Explaining AdaBoost. In Empirical Inference; Schölkopf, B., Luo, Z., Vovk, V., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 37–52. [Google Scholar] [CrossRef]
  54. Seo, D.K.; Kim, Y.H.; Eo, Y.D.; Park, W.Y.; Park, H.C. Generation of Radiometric, Phenological Normalized Image Based on Random Forest Regression for Change Detection. Remote Sens. 2017, 9, 1163. [Google Scholar] [CrossRef] [Green Version]
  55. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  56. Hsu, C.W.; Chang, C.C.; Lin, C.J. A Practical Guide to Support Vector Classification; National Taiwan University: Taipei, Taiwan, 2003; Available online: http//www.csie.ntu.edu.tw/~cjlin (accessed on 1 July 2021).
  57. Fowler, B. A sociological analysis of the satanic verses affair. Theory Cult. Soc. 2000, 17, 39–61. [Google Scholar] [CrossRef]
  58. Barakat, N.; Bradley, A.P. Rule extraction from support vector machines: A review. Neurocomputing 2010, 74, 178–190. [Google Scholar] [CrossRef]
  59. Martens, D.; Huysmans, J.; Setiono, R.; Vanthienen, J.; Baesens, B. Rule extraction from support vector machines: An overview of issues and application in credit scoring. Rule Extr. Support Vector Mach. 2008, 80, 33–63. [Google Scholar] [CrossRef]
  60. Uslan, V.; Seker, H. Support Vector-Based Takagi-Sugeno Fuzzy System for the Prediction of Binding Affinity of Peptides. In Proceedings of the 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Osaka, Japan, 3–7 July 2013; Institute of Electrical and Electronics Engineers: Piscataway, NJ, USA, 2013; pp. 4062–4065. [Google Scholar] [CrossRef]
  61. Gandomi, A.H.; Fridline, M.M.; Roke, D.A. Decision Tree Approach for Soil Liquefaction Assessment. Sci. World J. 2013, 2013, 346285. [Google Scholar] [CrossRef] [Green Version]
  62. Amirkiyaei, V.; Ghasemi, E. Stability assessment of slopes subjected to circular-type failure using tree-based models. Int. J. Geotech. Eng. 2020, 1862538. [Google Scholar] [CrossRef]
  63. Tiryaki, B. Predicting intact rock strength for mechanical excavation using multivariate statistics, artificial neural networks, and regression trees. Eng. Geol. 2008, 99, 51–60. [Google Scholar] [CrossRef]
  64. Hasanipanah, M.; Faradonbeh, R.S.; Amnieh, H.B.; Armaghani, D.J.; Monjezi, M. Forecasting blast-induced ground vibration developing a CART model. Eng. Comput. 2017, 33, 307–316. [Google Scholar] [CrossRef]
  65. Khosravi, K.; Mao, L.; Kisi, O.; Yaseen, Z.M.; Shahid, S. Quantifying hourly suspended sediment load using data mining models: Case study of a glacierized Andean catchment in Chile. J. Hydrol. 2018, 567, 165–179. [Google Scholar] [CrossRef]
  66. Taylor, K.E. Summarizing multiple aspects of model performance in a single diagram. J. Geophys. Res. Atmos. 2001, 106, 7183–7192. [Google Scholar] [CrossRef]
  67. Yang, Y.; Zhang, Q. A hierarchical analysis for rock engineering using artificial neural networks. Rock Mech. Rock Eng. 1997, 30, 207–222. [Google Scholar] [CrossRef]
  68. Faradonbeh, R.S.; Armaghani, D.J.; Majid, M.Z.; Tahir, M.M.; Murlidhar, B.R.; Monjezi, M.; Wong, H.M. Prediction of ground vibration due to quarry blasting based on gene expression programming: A new model for peak particle velocity prediction. Int. J. Environ. Sci. Technol. 2016, 13, 1453–1464. [Google Scholar] [CrossRef] [Green Version]
  69. Chen, W.; Hasanipanah, M.; Rad, H.N.; Armaghani, D.J.; Tahir, M.M. A new design of evolutionary hybrid optimization of SVR model in predicting the blast-induced ground vibration. Eng. Comput. 2021, 37, 1455–1471. [Google Scholar] [CrossRef]
  70. Rad, H.N.; Bakhshayeshi, I.; Jusoh, W.A.W.; Tahir, M.M.; Foong, L.K. Prediction of Flyrock in Mine Blasting: A New Computational Intelligence Approach. Nat. Resour. Res. 2020, 29, 609–623. [Google Scholar] [CrossRef]
  71. Momeni, E.; Armaghani, D.J.; Fatemi, S.A.; Nazir, R. Prediction of bearing capacity of thin-walled foundation: A simulation approach. Eng. Comput. 2018, 34, 319–327. [Google Scholar] [CrossRef]
  72. Momeni, E.; Dowlatshahi, M.B.; Omidinasab, F.; Maizir, H.; Armaghani, D.J. Gaussian Process Regression Technique to Estimate the Pile Bearing Capacity. Arab. J. Sci. Eng. 2020, 45, 8255–8267. [Google Scholar] [CrossRef]
  73. Kulkarni, R.U.; Dewaikar, D.M. Prediction of Interpreted Failure Loads of Rock-Socketed Piles in Mumbai Region using Hybrid Artificial Neural Networks with Genetic Algorithm. Int. J. Eng. Res. 2017, 6, 365–372. [Google Scholar] [CrossRef]
  74. Armaghani, D.J.; Shoib, R.S.N.S.B.R.; Faizi, K.; Rashid, A.S.A. Developing a hybrid PSO–ANN model for estimating the ultimate bearing capacity of rock-socketed piles. Neural Comput. Appl. 2017, 28, 391–405. [Google Scholar] [CrossRef]
Figure 1. Schematic layout of pile load test.
Figure 1. Schematic layout of pile load test.
Applsci 12 02126 g001
Figure 2. Diagram for stratigraphy and pile parameters [38].
Figure 2. Diagram for stratigraphy and pile parameters [38].
Applsci 12 02126 g002
Figure 3. Structure of XGBoost Algorithm.
Figure 3. Structure of XGBoost Algorithm.
Applsci 12 02126 g003
Figure 4. Random Forest structure.
Figure 4. Random Forest structure.
Applsci 12 02126 g004
Figure 5. SVM for a regression problem.
Figure 5. SVM for a regression problem.
Applsci 12 02126 g005
Figure 6. Decision tree structure.
Figure 6. Decision tree structure.
Applsci 12 02126 g006
Figure 7. The flowchart for applying a data-driven technique to anticipate pile bearing capacity.
Figure 7. The flowchart for applying a data-driven technique to anticipate pile bearing capacity.
Applsci 12 02126 g007
Figure 8. Measured Pu versus Estimated Pu for training models using (a) XGBoost, (b) AdaBoost, (c) RF, (d) DT, and (e) SVM.
Figure 8. Measured Pu versus Estimated Pu for training models using (a) XGBoost, (b) AdaBoost, (c) RF, (d) DT, and (e) SVM.
Applsci 12 02126 g008
Figure 9. Measured Pu versus predicted Pu for testing models using (a) XGBoost, (b) AdaBoost, (c) RF, (d) DT, and (e) SVM.
Figure 9. Measured Pu versus predicted Pu for testing models using (a) XGBoost, (b) AdaBoost, (c) RF, (d) DT, and (e) SVM.
Applsci 12 02126 g009
Figure 10. Sensitivity analysis of input variables.
Figure 10. Sensitivity analysis of input variables.
Applsci 12 02126 g010
Figure 11. Taylor diagram of the models.
Figure 11. Taylor diagram of the models.
Applsci 12 02126 g011
Table 1. Statistical study of inputs and output data.
Table 1. Statistical study of inputs and output data.
DatasetStatistical
Parameters
Input and Output Parameters
D (mm)X1 (m)X2 (m)X3 (m)Xp (m)Xg (m)Xt (m)Xm (m)NsNtPu (kN)
TrainingMinimum3003.41.801.953.3228.65.94.64432
Average378.574.0026.430.3772.6153.5172.83413.42510.8116.9081064.739
Maximum4004.7581.183.43.74.4515.5813.637.691551
Standard Deviation41.1790.4552.0390.4670.5520.0690.6092.2072.5500.914363.681
TestingMinimum3003.42.0802.053.382.058.886.184.86407.2
Average371.6673.856.7740.3072.773.5222.98713.70210.9327.0871023.266
Maximum4004.7581.183.43.724.0515.5813.537.731551
Standard Deviation45.4420.4721.5940.4380.5850.0780.5491.7192.1230.636362.003
Table 2. Correlation between parameters.
Table 2. Correlation between parameters.
ParametersDX1X2X3XpXgXtXmNsNtPu
D1.000
X10.6411.000
X20.4620.3291.000
X30.4210.4480.5641.000
Xp−0.714−0.935−0.515−0.6721.000
Xg0.4360.3570.3330.203−0.3771.000
Xt−0.481−0.469−0.331−0.8100.628−0.1351.000
Xm0.4740.3780.9890.672−0.5710.334−0.4221.000
Ns0.5770.5720.9470.719−0.7320.371−0.5330.9691.000
Nt0.1970.0500.9230.619−0.2890.198−0.3030.9310.8271.000
Pu0.7350.7060.7850.474−0.7800.460−0.3360.7850.8460.5581.000
Table 3. Hyperparameters optimal results.
Table 3. Hyperparameters optimal results.
AlgorithmHyperparametersMeaningsOptimal Values
XGBoostn estimatorsNumber of trees133
Learning rateShrinkage coefficient of tree0.03
Maximum depthMaximum depth of a tree4
RFn estimatorsNumber of trees in forest500
Minimum split Minimum samples of split for nodes5
Maximum depthMaximum depth of a tree5
Minimum leaf Minimum samples of nodes for leaf8
AdaBoostn estimatorsNumber of trees500
Learning rateShrinkage coefficient of tree1
SVMC2Regularization parameter2.5
DTMinimum split Minimum samples of split for nodes4
Maximum depthMaximum depth of a tree100
Minimum leafMinimum samples of nodes for leaf7
Table 4. Summary of Training model.
Table 4. Summary of Training model.
Training Set
ModelR2MAE (kN)RMSE (kN)MARE (%)NSERSR
XGBoost0.97147.51866.8444.3550.9660.184
AdaBoost0.95756.67182.4955.2520.9480.228
RF0.95258.36679.2405.7390.9520.219
DT0.93268.91294.3046.9110.9320.260
SVM0.88788.801123.3758.5070.8840.340
Table 5. Summary of Testing model.
Table 5. Summary of Testing model.
Testing Set
ModelR2MAE (kN)RMSE (kN)MARE (%)NSERSR
XGBoost0.95559.92980.6536.6000.9500.225
AdaBoost0.95070.38390.6658.2520.9360.253
RF0.94569.03086.3488.0140.9420.241
DT0.92574.45099.8228.7750.9230.278
SVM0.87898.320128.02710.9910.8730.357
Table 6. Comparison with other studies.
Table 6. Comparison with other studies.
AuthorModelFoundation TypeNumber of SamplesR2RMSE
Momeni et al. [71]ANFISThin-walls1500.8750.048
ANN0.710.529
Momeni et al. [72]GPRPiles2960.84-
Kulkarni et al. [73]GA-ANNRock-socketed piles1320.860.0093
Armaghani et al. [74]ANN0.8080.135
PSO-ANN0.9180.063
Pham et al. [38]GA-DLNNPiles4720.882109.965
Present studyXGBoostPiles2000.95580.653
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Amjad, M.; Ahmad, I.; Ahmad, M.; Wróblewski, P.; Kamiński, P.; Amjad, U. Prediction of Pile Bearing Capacity Using XGBoost Algorithm: Modeling and Performance Evaluation. Appl. Sci. 2022, 12, 2126. https://doi.org/10.3390/app12042126

AMA Style

Amjad M, Ahmad I, Ahmad M, Wróblewski P, Kamiński P, Amjad U. Prediction of Pile Bearing Capacity Using XGBoost Algorithm: Modeling and Performance Evaluation. Applied Sciences. 2022; 12(4):2126. https://doi.org/10.3390/app12042126

Chicago/Turabian Style

Amjad, Maaz, Irshad Ahmad, Mahmood Ahmad, Piotr Wróblewski, Paweł Kamiński, and Uzair Amjad. 2022. "Prediction of Pile Bearing Capacity Using XGBoost Algorithm: Modeling and Performance Evaluation" Applied Sciences 12, no. 4: 2126. https://doi.org/10.3390/app12042126

APA Style

Amjad, M., Ahmad, I., Ahmad, M., Wróblewski, P., Kamiński, P., & Amjad, U. (2022). Prediction of Pile Bearing Capacity Using XGBoost Algorithm: Modeling and Performance Evaluation. Applied Sciences, 12(4), 2126. https://doi.org/10.3390/app12042126

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop