Next Article in Journal
Impact of Computer-Based Assessments on the Science’s Ranks of Secondary Students
Next Article in Special Issue
Mechanical Properties of Cement-Treated Soil Mixed with Cellulose Nanofibre
Previous Article in Journal
Use of Granulated Rubber Tyre Waste as Lightweight Backfill Material for Retaining Walls
Previous Article in Special Issue
Numerical Analysis of an Upstream Tailings Dam Subjected to Pond Filling Rates
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of Prediction Models for Shear Strength of Rockfill Material Using Machine Learning Techniques

1
Department of Civil Engineering, University of Engineering and Technology Peshawar (Bannu Campus), Bannu 28100, Pakistan
2
Faculty of Mining and Geoengineering, AGH University of Science and Technology, Mickiewicza 30 Av., 30-059 Kraków, Poland
3
Mineral and Energy Economy Research Institute, Polish Academy of Sciences, 7A Wybickiego St., 31-261 Cracow, Poland
4
Department of Civil Engineering, University of Engineering and Technology, Mardan 23200, Pakistan
5
State Key Laboratory of Coastal and Offshore Engineering, Dalian University of Technology, Dalian 116024, China
6
Department of Architectural Engineering, Chungnam National University, Daejeon 34134, Korea
7
Department of Civil Engineering, CECOS University of IT and Emerging Sciences, Peshawar 25000, Pakistan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(13), 6167; https://doi.org/10.3390/app11136167
Submission received: 24 May 2021 / Revised: 30 June 2021 / Accepted: 1 July 2021 / Published: 2 July 2021
(This article belongs to the Special Issue Trends and Prospects in Geotechnics)

Abstract

:
Supervised machine learning and its algorithms are a developing trend in the prediction of rockfill material (RFM) mechanical properties. This study investigates supervised learning algorithms—support vector machine (SVM), random forest (RF), AdaBoost, and k-nearest neighbor (KNN) for the prediction of the RFM shear strength. A total of 165 RFM case studies with 13 key material properties for rockfill characterization have been applied to construct and validate the models. The performance of the SVM, RF, AdaBoost, and KNN models are assessed using statistical parameters, including the coefficient of determination (R2), Nash–Sutcliffe efficiency (NSE) coefficient, root mean square error (RMSE), and ratio of the RMSE to the standard deviation of measured data (RSR). The applications for the abovementioned models for predicting the shear strength of RFM are compared and discussed. The analysis of the R2 together with NSE, RMSE, and RSR for the RFM shear strength data set demonstrates that the SVM achieved a better prediction performance with (R2 = 0.9655, NSE = 0.9639, RMSE = 0.1135, and RSR = 0.1899) succeeded by the RF model with (R2 = 0.9545, NSE = 0.9542, RMSE = 0.1279, and RSR = 0.2140), the AdaBoost model with (R2 = 0.9390, NSE = 0.9388, RMSE = 0.1478, and RSR = 0.2474), and the KNN with (R2 = 0.6233, NSE = 0.6180, RMSE = 0.3693, and RSR = 0.6181). Furthermore, the sensitivity analysis result shows that normal stress was the key parameter affecting the shear strength of RFM.

1. Introduction

Rockfill materials (RFMs) are commonly used in civil engineering projects such as rockfill dams, slopes, and embankments as construction materials for filling. This material is either obtained from a river’s alluvial deposits or by blasting available rock [1,2]. RFMs are widely being used in the construction of rockfill dams to trap the river water because of their inherent flexibility, capacity to absorb large seismic energy, and adaptability to various foundation conditions. The behavior of RFMs used in rockfill dams is important for the safe and cost-effective construction of these structures. Generally, rockfill behaves like a Mohr/Coulomb material, albeit without cohesion and with relatively high internal friction angles. Crushed rockfill, loosely layered, can behave like coarse sand. The shear strength of both types of RFM is affected by many factors such as mineral composition, surface structure, particle size, shape, relative density, individual particle strength, etc. [3,4,5]. Because of the variable jointing, angularity/roundness, and rock particle size distribution, the RFM can be considered the most complex material [6]. In order to know the mechanical properties of RFMs, extensive field and laboratory research is essential for understanding RFM behavior and determining shear strength parameters in order to design safe and cost-effective structures. In situ direct shear system was used to monitor the shear strength of RFM, as well as the variation in the shear strength of rockfill along with the fill lift [7]. Linero [8] carried out some large-scale shear resistance experiments to simulate the material’s original grain size distribution and the expected load level. RFM with a large particle size (maximum particle size of 1200 mm) is incompatible in laboratory testing [9]. Owing to restricting the effects of large particle sizes on test apparatus, such behavior makes it much more difficult to design representative/realistic large-scale strength tests. Furthermore, determining the shear strength of RFM directly is considered a costly and difficult process. Large-scale shear tests are often time-consuming and complex, and estimating the nonlinear shear strength function without using an analytical method is difficult. As a result, several researchers have attempted to determine the mechanical properties of RFM using indirect methods based on machine learning (ML) techniques.
In recent years, several researchers used ML algorithms and achieved efficient successes in different civil engineering and other sectors such as environmental [10], geotechnical [11,12,13,14,15,16,17,18], and other fields of science [19,20,21,22,23,24,25,26,27,28]. Numerous researchers have documented the behavior of the RFM. Marsal [3], Mirachi et al. [4], Venkatachalam [5], Gupta [29], Abbas [30], and Honkanadavar and Sharma [31] carried out laboratory experiments on different rockfill materials and concluded that the behavior of stress-strain is nonlinear, inelastic and based on the level of stress. They also noted that with an increase in maximum particle size for riverbed rockfill material, the angle of internal friction increases, and a reverse pattern for quarried rockfill material is observed. Frossard et al. [32] proposed a rational approach for assessing rockfill shear strength on the basis of size effects; Honkanadavar and Gupta [9] developed power law to relate the shear strength parameter to some index properties of riverbed RFM. Describing the mechanical behavior of rockfill materials and challenges in large-scale strength tests have incited several approaches in modeling the respective behavior of such soils. In this context, the artificial neural network (ANN) approach used by Kaunda [33] needs fewer rockfill parameters and was found to be more efficient in predicting RFM shear strength. Zhou et al. [34] have recently used cubist and random forest regression algorithms and have found that both can deliver better predictive RFM shear strength results than ANN and conventional regression models. This field, however, continues to be further explored. Considering that large-scale strength tests to characterize the shear strength are challenging, ML algorithms based on support vector machine (SVM), random forest (RF), AdaBoost, and k-nearest neighbor (KNN) models are proposed. Furthermore, the ML algorithms SVM, RF, AdaBoost, and KNN have demonstrated excellent prediction efficiency in a variety of fields [35,36,37,38,39] because of their generalization capability. The application in civil engineering field more significantly in prediction of RFM shear strength is limited based on literature surveys.
The main intention of the present study is to explore the capability of using SVM, RF, AdaBoost, and KNN algorithms to establish a more precise and parsimonious behavioral model for predicting the RFM shear strength. A critical review of existing literature suggests that despite the successful implementation of these techniques in various domains, their implementation in the prediction of RFM shear strength is scarcely explored. One of the primary significances of this study is that the data division in the training and testing data sets has been made with due regard to statistical aspects such as maximum, minimum, mean, and standard deviation. The splitting of the data sets is made to determine the predictive capability and generalization performance of established models and later helps to better evaluate them. Additionally, sensitivity analysis is carried out to find the main parameter influencing RFM shear strength. Concisely, the present study investigated and expanded the scope of machine learning algorithms for the development of the RFM shear strength model, which will provide theoretical support for researchers to establish a basis in selecting optimal machine learning algorithms in improving the predictive performance of RFM shear strength.
The rest of this article is structured as follows: The next section introduces the description of the used database and preliminaries of the algorithms used in the proposed approach and discusses the model evaluation metrics. Development of SVM, RF, AdaBoost, and KNN models are described in Section 3. Section 4 is dedicated to the performances and comparison of proposed models. Finally, Section 5 draws conclusions and outlines promising directions for future work.

2. Materials and Methods

2.1. Data Set

In this study, 165 samples of rockfill material (RFM) shear strength case history acquired by Kaunda [33] presented in Table A1 in the Appendix A were used to develop and evaluate the effectiveness of the proposed models. The RFM shear strength case history data are summarized in Table 1, where D10, D30, D60, and D90 correspond to the 10%, 30%, 60%, and 90% passing sieve sizes, Cc and Cu refer to coefficients of uniformity and curvature (Cc), respectively, FM and GM describe fineness modulus and gradation modulus, respectively, R represents ISRM hardness rating, UCSmin and UCSmax (MPa) indicate the minimum and maximum uniaxial compression strengths (MPa), γ is the dry unit weight (kN/m3), σn is normal stress (MPa), and τ is the shear strength of RFM (MPa) as the output variable. In this study, the output parameter selected to determine shear strength was the shear stress value at the failure of test samples and was the single output variable. The database was divided into two different sets, consisting of 80 percent (132 cases) and 20 percent (33 cases) of data, respectively, represented as training and testing sets. The testing set was accustomed to determine when training should be stopped in order to avoid overfitting. In order to achieve a consistent data splitting, different combinations of training and testing sets were experienced. The abovementioned selection was in such a way that the maximum (Max), minimum (Min), mean, and standard deviation of the parameters were consistent in the training and testing data sets (Table 2).

2.2. Support Vector Machine

Boser, Guyon, and Vapnik were the first to formulate and introduce the support vector machine (SVM) [40]. In the case of non-separable data, to accommodate errors for certain objects i, the “ideal boundary” must be introduced:
{ minimize ( 1 2 | δ | 2 + C i = 1 n ξ i ) under the constraints y i ( b + δ · x i ) + ξ i 1 and ξ i 0
where C is the penetrating parameter; δ and b are, respectively, the normal vector and the bias of the hyperplane; and each ξ i refers to the distance within object i and the respective margin hyperplane [41,42].
Data are implicitly mapped to a higher-dimensional space through mercer kernels, which can be broken down into a dot product to learn nonlinearly separable functions K ( x i , x j ) = φ ( x i ) · φ ( x j ) [42]. The kernel of the radial basis function (RBF) that is used widely is listed below:
K ( x i , x j ) = exp ( σ x i x j 2 )
where σ is the kernel parameter.

2.3. Random Forest

The use of a large series of low-dimensional regression trees is the basis of the random forest (RF). The theoretical development of RF is described by Breiman [43]. RF is an example of ensemble learning, which requires the development of a large number of decision trees to be implemented. In general, there are two types of decision trees: regression trees and classification trees. Regression trees were designed in the RF model since the main goal of this analysis was to predict the shear strength of RFM. Figure 1 depicts a general architecture for RF analysis. The protocol for analysis can be divided into two stages:
Stage 1: To create a sequence of sub-data sets, the bootstrap statistical technique is used to randomly sample from the initial data set (training data). The forest is then built using regression trees based on these sub-data sets. Each tree is trained by choosing a set of variables at random (a fixed number of descriptive variables selected from the random subset). Two important parameters that can be adjusted during the training stage are the number of trees (ntree) and the number of variables (mtry).
Stage 2: Once the model has been trained, a prediction can be made. In an ensemble approach, input variables are evaluated for all regression trees first, and then the final output is calculated by measuring the average value of each individual tree’s prediction.

2.4. AdaBoost Algorithm

The sequential ensemble technique AdaBoost, or adaptive boosting, is based on the concept of developing many poor learners using different training sub-sets drawn at random from the original training data set. Weights are allocated during each training session, and these are used to learn each hypothesis. The weights are used to calculate the hypothesis error on the data set and are a measure of the relative importance of each instance. After each iteration, the weights are recalculated so that instances classified wrongly by the previous hypothesis obtain higher weights. This allows the algorithm to concentrate on instances that are more difficult to understand. The algorithm’s most important task is to assign updated weights to instances that were wrongly labeled. In regression, the instances represent a real-value error. The AdaBoost technique can be used to mark the calculated error as an error or not an error by comparing it to a predefined threshold prediction error. Instances that have made a greater mistake on previous learners are more likely (i.e., have a higher probability) to be chosen for training the next base learner. Finally, an ensemble estimate of the individual base learner predictions is made using a weighted average or median [44].

2.5. k-Nearest Neighbor

The supervised ML algorithm k-nearest neighbor (KNN) can be used to solve both classification and regression problems. It is, however, most commonly used in classification problems [45]. In regression problems, the input data set consists of k that is nearest to the training data sets deployed in the featured set. The output is dependent if KNN is deployed to function as a regression algorithm. For KNN regression, the ensuing result is the characteristic value for the object, which is the mean figure of k’s nearest neighbors. To locate the k of a data point, a parameter such as Euclidean, Mahalanobis can be used as the distance metric [46,47].

2.6. Performance Metric

The coefficient of determination (R2), Nash–Sutcliffe efficiency (NSE) coefficient, root mean square error (RMSE), and the ratio of the RMSE to the standard deviation of measured data (RSR) were taken into account to examine the predictive capacity of the models, as shown in Equations (3)–(6) [48,49,50]:
R 2 = [ i = 1 n ( O i O ¯ ) ( P i P ¯ ) i = 1 n ( O i O ¯ ) 2 i = 1 n ( P i P ¯ ) 2 ]
N S E = i = 1 n ( O i O ¯ ) 2 i = 1 n ( P i O i ) 2 i = 1 n ( O i O ¯ ) 2
R M S E = 1 n i = 1 n ( P i O i ) 2
R S R = i = 1 n ( O i P i ) 2 i = 1 n ( O i O ¯ ) 2
where n is the number of observations under consideration, Oi is the ith observed value, O ¯ is the mean observed value, Pi is the ith model-predicted value, and P is the mean model-predicted value.
R-squared, also called the determination coefficient, describes the change in data as the degree of fit. The normal “determination coefficient” range is (0–1). The model is considered to be efficient if the R2 value is greater than 0.8 and is close to 1 [51]. The NSE is a normalized statistic that controls the relative extent of the residual variance relative to the variance of the data measured [52]. The NSE varies between −∞ and 1. When NSE = 1, it presents a flawless match among observed and predicted values. Model predictive output with a range of 0.75 < NSE ≤ 1.00, 0.65 < NSE ≤ 0.75, 0.50 < NSE ≤ 0.65, 0.40 < NSE ≤ 0.50, or NSE ≤ 0.4 is graded as very good, good, acceptable, or unacceptable, respectively [53,54]. The RMSE is the square root of the ratio of the square of the deviation between the observed value and the true value of the number of observations n. The RMSE has a value greater than or equal to 0, where 0 is a statistically perfect fit for the data observed [55,56,57]. The RSR is interpreted as the ratio of the measured data’s RMSE and standard deviation. The RSR varies between an optimal value of 0 and a large positive value. A lower RSR presents a lower RMSE, which indicates the model’s greater predictive efficiency. RSR classification ranges are described as very good, good, acceptable, and unacceptable with ranges of 0.00 ≤ RSR ≤ 0.50, 0.50 ≤ RSR ≤ 0.60, 0.60 ≤ RSR ≤ 0.70, and RSR > 0.70, respectively [53].

3. Model Development to Predict RFM Shear Strength

The models for RFM shear strength prediction were developed using Orange software, which is a popular open-source environment for statistical computing and data visualization. All data processing is carried out using Orange software (version 3.13). The most prevalent supervised learning classification algorithms are given by Orange. In the package documentation manuals, one can find more information about input parameters, implementation, and references.
The structure of the model was based on an input matrix identified by predictor variables, x = { D 10 , D 30 , D 60 , D 90 , C c , C u , G M , F M , R , U C S min , U C S max , γ , and   σ n } and output, also called target variable (y), was the RFM shear strength. In every modeling process, achieving a consistent data division and the appropriate size of the training and testing data sets is the most important task. The statistical features, such as the minimum, maximum, mean, and standard deviation of the data sets, have therefore been taken into account in the splitting process. The statistical accuracy of the training and testing data sets optimizes the performance of the models and ultimately helps to evaluate them better. On the remaining data set, the proposed models were tested. In other words, to build and test the models, 132 and 33 data sets were used, respectively. To fairly assess the predictive performance of the models, the data set used for the testing of all models was kept the same.
In order to optimize the RFM shear strength prediction, all the models (AdaBoost, RF, SVM, and KNN) were tuned based on the trial and error process. Initially, the values were chosen for model tuning parameters and gradually varied in experiments until the best fitness measurements were achieved. Figure 2 shows the schematic diagram of the proposed methodology. The optimization method aims to find the best parameters for AdaBoost, RF, SVM, and KNN in order to achieve the best prediction accuracy. Some critical hyperparameters in the AdaBoost, RF, SVM, and KNN algorithms are tuned in this study, as shown in Table 3. The definitions of these hyperparameters are also clarified in Table 3. The values for the tuning parameters of the models were first chosen and then varied in the trials until the best fitness measures mentioned in Table 3 were achieved.

4. Results and Discussion

In this study, R2, NSE coefficient, RMSE, and RMSE to standard deviation of measured data are chosen as the criterion for defining the model’s output. The database is split into a training data set and a testing data set to evaluate the performance of the presented models. To make a fair comparison, all the models are developed by applying them to the same RFM shear strength training and testing data sets. Figure 3, displays the scatter plot of the actual and the predicted RFM shear strength for the training phase. The analysis of the R2 together with NSE, RMSE, and RSR for the RFM shear strength data set demonstrates that the SVM achieved a better prediction performance with (R2 = 0.9655, NSE = 0.9639, RMSE = 0.1135, and RSR = 0.1899) succeeded by the RF model with (R2 = 0.9545, NSE = 0.9542, RMSE = 0.1279, and RSR = 0.2140), the AdaBoost model with (R2 = 0.9390, NSE = 0.9388, RMSE = 0.1478, and RSR = 0.2474), and the KNN with (R2 = 0.6233, NSE = 0.6180, RMSE = 0.3693, and RSR = 0.6181).
Figure 4, presenting the predicted RFM shear strength, is plotted with the actual RFM shear strength data. According to the test data set, all models demonstrated very good predictive potential (R2 > 0.8) with the exception of KNN, which displayed slightly worse results (i.e., R2 = 0.6304) for the test data set. The result of R2 demonstrated that all SVM, RF, and AdaBoost models except KNN are appropriate, but the SVM model performed better because it had the highest R2 value (0.9656), and after that, the RF (0.9181) and AdaBoost (0.8951) models. In comparison to the other models, the KNN model presented the worst estimates with maximum dispersion (Figure 4).
In addition, the NSE measure was ranked from highest to lowest predictive strength, following the way: SVM (0.9654) > RF (0.9164) > AdaBoost (0.8835) > KNN (0.6076), which is similar to R2. With regard to RMSE score, the SVM model also had the maximum predictive ability by having the lowest RMSE (0.0153), succeeded by the models RF (0.0797), AdaBoost (0.0941), and KNN (0.1727).
Finally, the reliability of all applied models was divided into four groups based on RSR values: unsatisfactory, satisfactory, good, and very good with ranges of RSR > 0.70, 0.60 ≤ RSR ≤ 0.70, 0.50 ≤ RSR ≤ 0.60, and 0.00 ≤ RSR ≤ 0.50, respectively. The RSR value therefore demonstrates very good results throughout all our established models except the KNN model, whose performance is considered to be satisfactory. Figure 5 depicts the bar graphs comparing the R2, NSE, RMSE, and RSR for the training and testing data sets of all the models. The R2 defines the degree of co-linearity between our predicted and actual data. The value of RMSE is more focused on large errors than on small errors. A lower RSR indicates a lower RMSE, indicating the model’s better predictive efficiency. The SVM model has high R2 and NSE while lower RMSE and RSR values, revealing that the SVM model is preferable for predicting the RFM shear strength for the testing data. The SVM achieved a better prediction performance with (R2 = 0.9655, RMSE = 0.0513 and mean absolute error (MAE) = 0.0184) in comparison to the cubist method (R2 = 0.9645, RMSE = 0.0975, and MAE = 0.0644) and ANN method (R2 = 0.9386, RMSE = 0.1320 and MAE = 0.0841) reported by Zhou et al. [34] and Kaunda [33], respectively, for the test data. Additionally, the accuracy of modeling determined by the linear regression method reported by Andjelkovic et al. [58] between measured and calculated values of shear strength (R2 = 0.836) was slightly lower than the proposed SVM model. In general, the generalization and reliability of the SVM algorithm perform well, and larger data sets can yield better prediction results.
In the present research, a sensitivity analysis was also conducted using Yang and Zang’s [59] method to evaluate the influence of input parameters on RFM shear strength. This approach has been used in several studies [60,61,62,63] and is formulated as:
r i j = m = 1 n ( y i m × y o m ) m = 1 n y i m 2 m = 1 n y o m 2
where n is the number of data values (this study used 132 data values) and yim and yom are the input and output parameters. The rij value ranged from zero to one for each input parameter, and the highest rij values suggested the most efficient output parameter (which was RFM shear strength in this study). The rij values for all input parameters are presented in Figure 6. It can be seen from Figure 6 that the σn with rij is 0.990. Similar research of sensitivity analyses on RFM shear strength was also implemented by Kaunda [33] and Zhou et al. [34]. The findings demonstrated that normal stress is the most sensitive factor, which shows agreement with the present mentioned results.
Despite the fact that the proposed model produces desirable prediction results, certain limitations should be addressed in the future.
(1)
Similar to other machine learning methods, the major disadvantages of SVM, RF, AdaBoost, and KNN models are sensitive to the fitness of the data set. Generally, if the data set is small, the generalization and reliability of the model would be influenced. However, the SVM, RF, and AdaBoost algorithms work with a limited data set, i.e., 165 cases, except for KNN. The prediction performances could be better on a larger data set. Furthermore, the developed models can always be updated to yield better results as new data becomes available.
(2)
Other qualitative indicators such as the Los Angeles abrasion value and lithology may also have influences on the prediction results of the shear strength of RFM. Accordingly, it is significant to analyze the influence of these indicators on the prediction results for improving performance.

5. Conclusions

This study employed and examined the SVM, RF, AdaBoost, and KNN algorithms in the RFM shear strength prediction problem. To construct and validate a new model on the basis of the aforementioned algorithms, a comprehensive database containing 165 RFM case studies was collected from the available literature. Thirteen different predictive variables for rockfill characterization were selected as the input variables: D10 (mm), D30 (mm), D60 (mm), D90 (mm), Cc, Cu, GM, FM, R, UCSmin (MPa), γ (kN/m3), UCSmax (MPa), and σn (MPa). The predictive performance of the proposed models is verified and compared. The conclusions can be outlined as follows:
  • In this study, the SVM model (R2 = 0.9656, NSE = 0.9654, RMSE = 0.0153, and RSR = 0.1861) successfully achieved a high level of modeling prediction efficiency to RF (R2 = 0.9181, NSE = 0.9164, RMSE = 0.0797, and RSR = 0.2891), AdaBoost (R2 = 0.8951, NSE = 0.8835, RMSE = 0.0941, and RSR = 0.3414), and KNN (R2 = 0.6304, NSE = 0.6076, RMSE = 0.1727, and RSR = 0.6264) in the test data set. As the same methodology (having the same training and test data sets) for structuring all models is taken into consideration, the SVM model resulted the best and highest performance in this aspect. This implies that this algorithm is robust in comparison with others in RFM shear strength prediction.
  • The performance (in terms of R2) of the test data set for the SVM, RF, and AdaBoost algorithms studied falls in the range of 0.9656–0.8951 across the three models with 13 input valuables. Results conclude that it is rational and feasible to estimate the shear strength of RFM from the gradation, particle size, dry unit weight (γ), material hardness, FM, and normal stress (σn).
  • Sensitivity analysis results revealed that normal stress (σn) was the key parameter affecting the shear strength of RFM.
The findings show that the SVM model is a useful and accurate artificial intelligence technique for predicting RFM shear strength and can be used in various fields. Further, the generalization of the proposed approach for achieving improved performance results, more experimental data should be collected in future research. Finally, RFM shear strength prediction using advanced machine learning algorithms (i.e., deep learning) is left as a future research topic.

Author Contributions

Conceptualization, M.A. (Mahmood Ahmad); P.K. and P.O.; methodology, M.A. (Mahmood Ahmad), P.K.; software, M.A. (Mahmood Ahmad) and F.A.; validation, P.O., F.A., M.A. (Mahmood Ahmad), M.A. (Muhammad Alam) and F.A.; formal analysis, M.A. (Mahmood Ahmad); investigation, M.A. (Mahmood Ahmad), F.A., P.K., M.J.I. and S.S.; resources, P.K. and P.O.; data curation, M.A. (Mahmood Ahmad), M.A. (Muhammad Alam) and M.J.I.; writing—original draft preparation, M.A. (Mahmood Ahmad); writing—review and editing, M.A. (Mahmood Ahmad), B.J.K., F.A. and S.S.; supervision, P.K. and P.O.; project administration, P.K.; funding acquisition, P.K. and P.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Notation

ANN Artificial neural network
AdaBoostAdaptive boosting
KNN k-nearest neighbor
NSENash–Sutcliffe efficiency coefficient
R2 Coefficient of determination
RF Random forest
RFMRockfill material
RMSERoot mean square error
RSR Ratio of RMSE to the standard deviation of the measured data
ISRMInternational Society of Rock Mechanics
SVM Support vector machine
D10Sieve size at 10 percent passing
D30Sieve size at 30 percent passing
D60Sieve size at 60 percent passing
D90Sieve size at 90 percent passing
CcCoefficient of curvature
CuCoefficient of uniformity
GMGradation modulus
FMFineness modulus
RISRM hardness rating
UCSminMinimum uniaxial compression strength
γDry unit weight
UCSmaxMaximum uniaxial compression strength
σnNormal stress
τShear strength
φAngle of internal friction

Appendix A

Table A1. Rockfill shear strength database.
Table A1. Rockfill shear strength database.
Case No.LocationD10 (mm)D30 (mm)D60 (mm)D90 (mm)CcCuGMFMRUCSmin
(MPa)
UCSmax
(MPa)
γ
(KN/m3)
σn
(MPa)
τ
(MPa)
1Canada0.020.9441811.052004.784.1911515.40.0220.013
2Canada0.020.9441811.052004.784.1911515.40.0440.025
3Canada0.020.9441811.052004.784.1911515.40.0880.049
4Canada0.032.16.61822.272204.224.7311538.90.0220.013
5Canada0.032.16.61822.272204.224.7311538.90.0440.024
6Canada0.032.16.61822.272204.224.7311538.90.0880.048
7Canada0.090.923.2102.9435.5653.94115370.0220.014
8Canada0.090.923.2102.9435.5653.94115370.0440.027
9Canada0.090.923.2102.9435.5653.94115370.0880.053
10U.K.1619291.89192.616.36510025019.620.0590.163
11U.K.1619291.89192.616.36510025019.620.0980.218
12U.K.1619291.89192.616.36510025019.620.1980.367
13U.K.1619291.89192.616.36510025019.620.2990.513
14U.K.0.33.216302.1353.333.375.74510025018.05040.0580.15
15U.K.0.33.216302.1353.333.375.74510025018.05040.0970.204
16U.K.0.33.216302.1353.333.375.74510025018.05040.1950.33
17U.K.0.33.216302.1353.333.375.74510025018.05040.2970.456
18U.K.1619291.89192.656.36510025019.620.1790.262
19U.K.1619291.89192.656.36510025019.620.5380.697
20U.K.1619291.89192.656.36510025019.620.8870.112
21U.K.0.33.216302.1353.333.375.74510025018.05040.1770.245
22U.K.0.33.216302.1353.333.375.74510025018.05040.5290.666
23U.K.0.33.216302.1353.333.375.74510025018.05040.8760.102
24Iran0.11.27.517.31.92754.327.424501009.31950.1010.16
25Iran0.11.27.517.31.92754.327.424501009.31950.3010.34
26Iran0.11.27.517.31.92754.327.424501009.31950.5030.5
27Iran0.42.811301.7827.53.385.64501009.31950.1720.207
28Iran0.42.811301.7827.53.385.64501009.31950.4970.476
29Iran0.42.811301.7827.53.385.64501009.31950.830.751
30Japan1.34.615321.0911.542.686.28510025017.814960.0940.136
31Japan1.34.615321.0911.542.686.28510025017.814960.1770.242
32Japan1.34.615321.0911.542.686.28510025017.814960.3510.415
33Japan1.34.615321.0911.542.686.28510025017.814960.5120.552
34Japan0.6216300.4226.673.245.8645010021.87630.0930.165
35Japan0.6216300.4226.673.245.8645010021.87630.1820.308
36Japan0.6216300.4226.673.245.8645010021.87630.3590.523
37Japan0.6216300.4226.673.245.8645010021.87630.5350.744
38Iran0.42.99.7312.1724.253.415.57450100210.1770.214
39Iran0.42.99.7312.1724.253.415.57450100210.5140.525
40Iran0.42.99.7312.1724.253.415.57450100210.8390.773
41Iran0.42.99.7312.1724.253.415.57450100211.1721.07
42Iran0.42.99.7312.1724.253.415.57450100211.4941.312
43Iran0.42.99.7312.1724.253.415.57450100211.971.648
44Iran0.42.99.7312.1724.253.415.5745010020.80.180.24
45Iran0.42.99.7312.1724.253.415.5745010020.80.50.447
46Iran0.42.99.7312.1724.253.415.5745010020.80.8210.689
47Iran0.42.99.7312.1724.253.415.5745010020.81.1420.93
48Iran0.52.89.7301.6219.43.435.61510025021.10.4870.39
49Iran0.52.89.7301.6219.43.435.61510025021.10.9720.766
50Iran0.52.89.7301.6219.43.435.61510025021.11.4481.11
51Iran0.22.519.442.21.61973.195.775100250210.1680.157
52Iran0.22.519.442.21.61973.195.775100250210.3730.634
53Iran0.22.519.442.21.61973.195.775100250210.7311.088
54Iran0.22.519.442.21.61973.195.775100250210.9061.258
55Iran0.22.519.442.21.61973.195.775100250211.2621.699
56Iran0.22.519.442.21.61973.195.775100250211.4371.883
57Iran0.43.310.333.32.6425.753.325.6445010021.80.0920.14
58Iran0.43.310.333.32.6425.753.325.6445010021.80.1790.23
59Iran0.43.310.333.32.6425.753.325.6445010021.80.3440.357
60Iran0.43.310.333.32.6425.753.325.6445010021.80.5140.52
61Iran0.43.310.333.32.6425.753.325.6445010021.80.8590.887
62Iran0.43.310.333.32.6425.753.325.6445010021.81.1861.149
63Iran1.22.14.225.30.883.53.635.3445010021.80.0920.147
64Iran1.22.14.225.30.883.53.635.3445010021.80.1780.22
65Iran1.22.14.225.30.883.53.635.3445010021.80.5030.461
66Iran1.22.14.225.30.883.53.635.3445010021.81.1480.959
67Iran0.42.99.7312.1724.253.45.53450100210.340.332
68Iran0.42.99.7312.1724.253.45.53450100210.990.843
69Iran0.42.99.7312.1724.253.45.53450100211.6181.271
70Iran0.42.99.7312.1724.253.45.53450100212.3991.799
71Iran0.42.99.7312.1724.253.45.5345010021.50.3420.342
72Iran0.42.99.7312.1724.253.45.5345010021.50.9940.865
73Iran0.42.99.7312.1724.253.45.5345010021.51.631.321
74Iran0.42.99.7312.1724.253.45.5345010021.52.4221.891
75Australia33.942.45060.21.061.470.28.8510025021.70.1630.23
76Australia33.942.45060.21.061.470.28.8510025021.70.2150.275
77Australia33.942.45060.21.061.470.28.8510025021.70.4120.424
78Australia303440.8500.941.360.528.5510025021.70.1650.252
79Australia303440.8500.941.360.528.5510025021.70.2150.28
80Australia303440.8500.941.360.528.5510025021.70.4120.424
81Germany411.736.298.20.959.051.477.5545010024.21.0391.114
82Germany411.736.298.20.959.051.477.5545010024.22.0341.964
83Germany411.736.298.20.959.051.477.5545010024.23.0042.705
84Germany39.130.498.20.9110.131.677.2845010024.20.5330.658
85Germany39.130.498.20.9110.131.677.2845010024.21.0391.114
86Germany39.130.498.20.9110.131.677.2845010024.22.0181.882
87Germany4.212.841.2990.959.811.377.6245010024.20.5120.512
88Germany4.212.841.2990.959.811.377.6245010024.21.0010.902
89Germany4.212.841.2990.959.811.377.6245010024.21.9871.728
90USA0.9318.8990.5320.892.646.35510025021.70.8610.898
91USA0.9318.8990.5320.892.646.35510025021.71.671.509
92USA0.9318.8990.5320.892.646.35510025021.74.0493.198
93U.K.0.441.56.9927.50.7315.893.825.1645010018.70.1590.189
94U.K.0.441.56.9927.50.7315.893.825.1645010018.70.4710.424
95U.K.0.441.56.9927.50.7315.893.825.1645010018.71.130.905
96Iran0.42.312.244.41.0830.53.35.6945010026.20.8150.66
97Iran0.42.312.244.41.0830.53.35.6945010018.70.7940.577
98Iran0.42.312.244.41.0830.53.35.69510025024.50.9940.864
99India0.51.54.615.60.989.24.224.845010024.51.3840.881
100India0.952.812.534.90.6613.163.025.9745010024.51.3690.836
101India1.34.618.955.90.8614.542.46.5445010024.51.3580.803
102Iran (multiple)0.5310.431.21.7320.83.365.6345010026.21.0560.625
103Iran (multiple)0.42.89.230.12.13233.425.5345010024.50.5010.451
104Iran (multiple)0.42.89.230.12.13233.425.5345010024.50.9860.827
105Iran (multiple)0.42.89.230.12.13233.425.5345010024.51.4791.241
106Iran (multiple)0.53.310.2312.1420.43.285.7510025024.50.4850.379
107Iran (multiple)0.53.310.2312.1420.43.285.7510025024.50.8080.631
108Iran (multiple)0.53.310.2312.1420.43.285.7510025024.51.1310.884
109Iran (multiple)0.42.810.431.21.88263.45.5845010018.70.8080.631
110Iran (multiple)0.42.810.431.21.88263.45.5845010018.71.1310.884
111USA2.419.380.11001.9433.381.327.72625040025.60.850.836
112USA2.419.380.11001.9433.381.327.72625040025.61.6951.637
113USA2.419.380.11001.9433.381.327.72625040025.64.2053.921
114Iran (multiple)0.01110.443.99.62104044.9345010024.20.2410.183
115Iran (multiple)0.01110.443.99.62104044.9345010024.20.4680.316
116Iran (multiple)0.01110.443.99.62104044.9345010024.20.9210.582
117Iran (multiple)0.01110.443.99.62104044.9345010024.20.2650.313
118Iran (multiple)0.01110.443.99.62104044.9345010024.20.5110.506
119Iran (multiple)0.01110.443.99.62104044.9345010024.21.0010.902
120USA0.20.561.22.60.166345010016.10.0210.029
121USA0.20.561.22.60.166345010016.10.0420.051
122USA0.20.561.22.60.166345010016.10.0680.071
123India0.11.36.5152.6654.424.55510025019.90.0540.005
124India0.11.36.5152.6654.424.55510025019.90.0890.028
125India0.11.36.5152.6654.424.55510025019.90.110.049
126India0.11.36.5152.6654.424.55510025019.90.1520.067
127India0.11.36.5152.6654.424.55510025019.90.1910.081
128India0.11.36.5152.6654.424.55510025019.90.240.092
129India0.116.2171.61624.54.44510025022.30.7060.94
130India0.116.2171.61624.54.44510025022.31.311.296
131India0.116.2171.61624.54.44510025022.31.8681.536
132India0.22.912.3323.4261.53.465.53510025022.30.7020.903
133India0.22.912.3323.4261.53.465.53510025022.31.3051.25
134India0.22.912.3323.4261.53.465.53510025022.31.8621.486
135India0.44.421.259.82.28532.746.27510025022.30.6970.862
136India0.44.421.259.82.28532.746.27510025022.31.2831.167
137India0.44.421.259.82.28532.746.27510025022.31.8191.358
138Australia27.132.641.3530.951.520.578.5510025015.30.0020.007
139Australia27.132.641.3530.951.520.578.5510025015.30.020.052
140Australia27.132.641.3530.951.520.578.5510025015.30.0320.072
141Australia27.132.641.3530.951.520.578.5510025015.30.0540.095
142Australia27.132.641.3530.951.520.578.5510025015.30.1110.168
143Australia27.132.641.3530.951.520.578.5510025015.30.1620.217
144Australia27.132.641.3530.951.520.578.5510025015.30.2090.259
145Australia27.132.641.3530.951.520.578.5510025015.30.4010.409
146Australia20.726.732.8531.051.580.898.2510025015.30.0030.008
147Australia20.726.732.8531.051.580.898.2510025015.30.0210.062
148Australia20.726.732.8531.051.580.898.2510025015.30.0350.089
149Australia20.726.732.8531.051.580.898.2510025015.30.0580.116
150Australia20.726.732.8531.051.580.898.2510025015.30.1150.191
151Australia20.726.732.8531.051.580.898.2510025015.30.1550.206
152Australia20.726.732.8531.051.580.898.2510025015.30.2090.259
153Australia20.726.732.8531.051.580.898.2510025015.30.3940.401
154Thailand3.17.82246.40.897.11.987.01450100210.8330.745
155Thailand3.17.82246.40.897.11.987.01450100211.6491.407
156Thailand3.17.82246.40.897.11.987.01450100212.4512.01
157Thailand3.17.82246.40.897.11.987.01450100213.2232.492
158Thailand3.57.119.845.70.735.662.036.98450100210.8080.631
159Thailand3.57.119.845.70.735.662.036.98450100211.5921.169
160Thailand3.57.119.845.70.735.662.036.98450100212.3381.576
161Thailand3.57.119.845.70.735.662.036.98450100213.142.178
162Netherlands111523320.892.091.487.48510025016.80.0140.028
163Netherlands111523320.892.091.487.48510025016.80.0280.048
164Netherlands111523320.892.091.487.48510025016.80.0550.082
165Netherlands111523320.892.091.487.48510025016.80.1080.143

References

  1. Aghaei, A.A.; Soroush, A.; Reyhani, M. Large-scale triaxial testing and numerical modeling of rounded and angular rockfill materials. Sci. Iran. Trans. A Civ. Eng. 2010, 17, 169–183. [Google Scholar]
  2. Varadarajan, A.; Sharma, K.; Venkatachalam, K.; Gupta, A. Testing and modeling two rockfill materials. J. Geotech. Geoenviron. Eng. 2003, 129, 206–218. [Google Scholar] [CrossRef]
  3. Marsal, R.J. Large scale testing of rockfill materials. J. Soil Mech. Found. Div. 1967, 93, 27–43. [Google Scholar] [CrossRef]
  4. Marachi, N.D. Strength and Deformation. Characteristics of Rockfill Materials. Ph.D. Thesis, University of California, Berkeley, CA, USA, 1969. [Google Scholar]
  5. Venkatachalam, K. Prediction of Mechanical Behaviour of Rockfill Materials. Ph.D. Thesis, Indian Institute of Technology, Delhi, India, 1993. [Google Scholar]
  6. Leps, T.M. Review of shearing strength of rockfill. J. Soil Mech. Found. Div. 1970, 96, 1159–1170. [Google Scholar] [CrossRef]
  7. Liu, S.-H. Application of in situ direct shear device to shear strength measurement of rockfill materials. Water Sci. Eng. 2009, 2, 48–57. [Google Scholar]
  8. Linero, S.; Palma, C.; Apablaza, R. Geotechnical characterisation of waste material in very high dumps with large scale triaxial testing. In Proceedings of the 2007 International Symposium on Rock Slope Stability in Open Pit Mining and Civil Engineering; Australian Centre for Geomechanics: Perth, Australia, 2007; pp. 59–75. [Google Scholar] [CrossRef]
  9. Honkanadavar, N.; Gupta, S. Prediction of shear strength parameters for prototype riverbed rockfill material using index properties. In Proceedings of the Indian Geotechnical Conference, Mumbai, India, 16–18 December 2010; pp. 335–338. [Google Scholar]
  10. Froemelt, A.; Duürrenmatt, D.J.; Hellweg, S. Using data mining to assess environmental impacts of household consumption behaviors. Environ. Sci. Technol. 2018, 52, 8467–8478. [Google Scholar] [CrossRef] [PubMed]
  11. Ahmad, M.; Tang, X.-W.; Qiu, J.-N.; Gu, W.-J.; Ahmad, F. A hybrid approach for evaluating CPT-based seismic soil liquefaction potential using Bayesian belief networks. J. Cent. South Univ. 2020, 27, 500–516. [Google Scholar]
  12. Ahmad, M.; Tang, X.-W.; Qiu, J.-N.; Ahmad, F. Evaluating Seismic Soil Liquefaction Potential Using Bayesian Belief Network and C4. 5 Decision Tree Approaches. Appl. Sci. 2019, 9, 4226. [Google Scholar] [CrossRef] [Green Version]
  13. Ahmad, M.; Tang, X.; Qiu, J.; Ahmad, F.; Gu, W. LLDV-a Comprehensive framework for assessing the effects of liquefaction land damage potential. In Proceedings of the 2019 IEEE 14th International Conference on Intelligent Systems and Knowledge Engineering (ISKE), Dalian, China, 14–16 November 2019; pp. 527–533. [Google Scholar]
  14. Ahmad, M.; Tang, X.-W.; Qiu, J.-N.; Ahmad, F.; Gu, W.-J. A step forward towards a comprehensive framework for assessing liquefaction land damage vulnerability: Exploration from historical data. Front. Struct. Civil. Eng. 2020, 14, 1476–1491. [Google Scholar] [CrossRef]
  15. Ahmad, M.; Tang, X.; Ahmad, F. Evaluation of liquefaction-induced settlement using random forest and REP tree models: Taking Pohang earthquake as a case of illustration. In Natural Hazards-Impacts, Adjustments & Resilience; IntechOpen: London, UK, 2020. [Google Scholar]
  16. Ahmad, M.; Al-Shayea, N.A.; Tang, X.-W.; Jamal, A.; Al-Ahmadi, H.M.; Ahmad, F. Predicting the pillar stability of underground mines with random trees and C4. 5 decision trees. Appl. Sci. 2020, 10, 6486. [Google Scholar] [CrossRef]
  17. Pirhadi, N.; Tang, X.; Yang, Q.; Kang, F. A new equation to evaluate liquefaction triggering using the response surface method and parametric sensitivity analysis. Sustainability 2019, 11, 112. [Google Scholar] [CrossRef] [Green Version]
  18. Pirhadi, N.; Tang, X.; Yang, Q. Energy evaluation of triggering soil liquefaction based on the response surface method. Appl. Sci. 2019, 9, 694. [Google Scholar] [CrossRef] [Green Version]
  19. Mosavi, A.; Shirzadi, A.; Choubin, B.; Taromideh, F.; Hosseini, F.S.; Borji, M.; Shahabi, H.; Salvati, A.; Dineva, A.A. Towards an ensemble machine learning model of random subspace based functional tree classifier for snow avalanche susceptibility mapping. IEEE Access 2020, 8, 145968–145983. [Google Scholar] [CrossRef]
  20. Mosavi, A.; Golshan, M.; Janizadeh, S.; Choubin, B.; Melesse, A.M.; Dineva, A.A. Ensemble models of GLM, FDA, MARS, and RF for flood and erosion susceptibility mapping: A priority assessment of sub-basins. Geocarto Int. 2020, 1–20. [Google Scholar] [CrossRef]
  21. Mosavi, A.; Hosseini, F.S.; Choubin, B.; Goodarzi, M.; Dineva, A.A. Groundwater salinity susceptibility mapping using classifier ensemble and Bayesian machine learning models. IEEE Access 2020, 8, 145564–145576. [Google Scholar] [CrossRef]
  22. Mosavi, A.; Sajedi-Hosseini, F.; Choubin, B.; Taromideh, F.; Rahi, G.; Dineva, A.A. Susceptibility mapping of soil water erosion using machine learning models. Water 2020, 12, 1995. [Google Scholar] [CrossRef]
  23. Mosavi, A.; Hosseini, F.S.; Choubin, B.; Abdolshahnejad, M.; Gharechaee, H.; Lahijanzadeh, A.; Dineva, A.A. Susceptibility Prediction of Groundwater Hardness Using Ensemble Machine Learning Models. Water 2020, 12, 2770. [Google Scholar] [CrossRef]
  24. Mosavi, A.; Hosseini, F.S.; Choubin, B.; Taromideh, F.; Ghodsi, M.; Nazari, B.; Dineva, A.A. Susceptibility mapping of groundwater salinity using machine learning models. Environ. Sci. Pollut. Res. 2021, 28, 10804–10817. [Google Scholar] [CrossRef] [PubMed]
  25. Mosavi, A.; Hosseini, F.S.; Choubin, B.; Goodarzi, M.; Dineva, A.A.; Sardooi, E.R. Ensemble boosting and bagging based machine learning models for groundwater potential prediction. Water Resour. Manag. 2021, 35, 23–37. [Google Scholar] [CrossRef]
  26. Choubin, B.; Borji, M.; Hosseini, F.S.; Mosavi, A.; Dineva, A.A. Mass wasting susceptibility assessment of snow avalanches using machine learning models. Sci. Rep. 2020, 10, 18363. [Google Scholar] [CrossRef]
  27. Lovrić, M.; Pavlović, K.; Žuvela, P.; Spataru, A.; Lučić, B.; Kern, R.; Wong, M.W. Machine learning in prediction of intrinsic aqueous solubility of drug-like compounds: Generalization, complexity, or predictive ability? J. Chemom. 2021, e3349. [Google Scholar] [CrossRef]
  28. Lovrić, M.; Meister, R.; Steck, T.; Fadljević, L.; Gerdenitsch, J.; Schuster, S.; Schiefermüller, L.; Lindstaedt, S.; Kern, R. Parasitic resistance as a predictor of faulty anodes in electro galvanizing: A comparison of machine learning, physical and hybrid models. Adv. Model. Simul. Eng. Sci. 2020, 7, 1–16. [Google Scholar] [CrossRef]
  29. Gupta, A.K. Constitative Modelling of Rockfill Materials. Ph.D. Thesis, Indian Institute of Technology, Delhi, India, 2000. [Google Scholar]
  30. Abbas, S.; Varadarajan, A.; Sharma, K. Prediction of shear strength parameter of prototype rockfill material. IGC-2003 Roorkee 2003, 1, 5–8. [Google Scholar]
  31. Honkanadavar, N.; Sharma, K. Testing and modeling the behavior of riverbed and blasted quarried rockfill materials. Int. J. Geomech. 2014, 14, 04014028. [Google Scholar] [CrossRef]
  32. Frossard, E.; Hu, W.; Dano, C.; Hicher, P.-Y. Rockfill shear strength evaluation: A rational method based on size effects. Géotechnique 2012, 62, 415–427. [Google Scholar] [CrossRef] [Green Version]
  33. Kaunda, R. Predicting shear strengths of mine waste rock dumps and rock fill dams using artificial neural networks. Int. J. Min. Mineral. Eng. 2015, 6, 139–171. [Google Scholar] [CrossRef]
  34. Zhou, J.; Li, E.; Wei, H.; Li, C.; Qiao, Q.; Armaghani, D.J. Random forests and cubist algorithms for predicting shear strengths of rockfill materials. Appl. Sci. 2019, 9, 1621. [Google Scholar] [CrossRef] [Green Version]
  35. Alkhatib, K.; Najadat, H.; Hmeidi, I.; Shatnawi, M.K.A. Stock price prediction using k-nearest neighbor (kNN) algorithm. Int. J. Bus. Humanit. Technol. 2013, 3, 32–44. [Google Scholar]
  36. Vijayan, V.; Ravikumar, A. Study of data mining algorithms for prediction and diagnosis of diabetes mellitus. Int. J. Comput. Appl. 2014, 95, 12–16. [Google Scholar]
  37. Thongkam, J.; Xu, G.; Zhang, Y. AdaBoost algorithm with random forests for predicting breast cancer survivability. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 3062–3069. [Google Scholar]
  38. Samui, P.; Sitharam, T.; Contadakis, M. Machine learning modelling for predicting soil liquefaction susceptibility. Nat. Hazards Earth Syst. Sci. 2011, 11, 1–9. [Google Scholar] [CrossRef] [Green Version]
  39. Pal, M. Support vector machines-based modelling of seismic liquefaction potential. Int. J. Numer. Anal. Methods Geomech. 2006, 30, 983–996. [Google Scholar] [CrossRef]
  40. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; pp. 144–152. [Google Scholar]
  41. Zhou, J.; Li, X.; Shi, X. Long-term prediction model of rockburst in underground openings using heuristic algorithms and support vector machines. Saf. Sci. 2012, 50, 629–644. [Google Scholar] [CrossRef]
  42. Zhou, J.; Li, X.; Wang, S.; Wei, W. Identification of large-scale goaf instability in underground mine using particle swarm optimization and support vector machine. Int. J. Min. Sci. Technol. 2013, 23, 701–707. [Google Scholar] [CrossRef]
  43. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  44. Seo, D.K.; Kim, Y.H.; Eo, Y.D.; Park, W.Y.; Park, H.C. Generation of radiometric, phenological normalized image based on random forest regression for change detection. Remote. Sens. 2017, 9, 1163. [Google Scholar] [CrossRef] [Green Version]
  45. Maillo, J.; Ramírez, S.; Triguero, I.; Herrera, F. kNN-IS: An iterative spark-based design of the k-nearest neighbors classifier for big data. Knowl. Based Syst. 2017, 117, 3–15. [Google Scholar] [CrossRef] [Green Version]
  46. Chomboon, K.; Chujai, P.; Teerarassamee, P.; Kerdprasop, K.; Kerdprasop, N. An empirical study of distance metrics for k-nearest neighbor algorithm. In Proceedings of the 3rd International Conference on Industrial Application Engineering, Kitakyushu, Japan, 28–30 March 2015; pp. 280–285. [Google Scholar]
  47. Prasath, V.; Alfeilat, H.A.A.; Hassanat, A.; Lasassmeh, O.; Tarawneh, A.S.; Alhasanat, M.B.; Salman, H.S.E. Distance and similarity measures effect on the performance of K-nearest neighbor classifier—A review. arXiv 2017, arXiv:1708.04321. [Google Scholar]
  48. Farooq, F.; Nasir Amin, M.; Khan, K.; Rehan Sadiq, M.; Faisal Javed, M.; Aslam, F.; Alyousef, R. A comparative study of random forest and genetic engineering programming for the prediction of compressive strength of High Strength Concrete (HSC). Appl. Sci. 2020, 10, 7330. [Google Scholar] [CrossRef]
  49. Golmohammadi, G.; Prasher, S.; Madani, A.; Rudra, R. Evaluating three hydrological distributed watershed models: MIKE-SHE, APEX, SWAT. Hydrology 2014, 1, 20–39. [Google Scholar] [CrossRef] [Green Version]
  50. Nhu, V.-H.; Shahabi, H.; Nohani, E.; Shirzadi, A.; Al-Ansari, N.; Bahrami, S.; Miraki, S.; Geertsema, M.; Nguyen, H. Daily water level prediction of Zrebar Lake (Iran): A comparison between M5P, random forest, random tree and reduced error pruning trees algorithms. ISPRS Int. J. Geo. Inf. 2020, 9, 479. [Google Scholar] [CrossRef]
  51. Gandomi, A.H.; Babanajad, S.K.; Alavi, A.H.; Farnam, Y. Novel approach to strength modeling of concrete under triaxial compression. J. Mater. Civ. Eng. 2012, 24, 1132–1143. [Google Scholar] [CrossRef]
  52. Nush, J.; Sutcliffe, J.V. River flow forecasting through conceptual models part I—A discussion of principles. J. Hydrol. 1970, 10, 282–290. [Google Scholar] [CrossRef]
  53. Khosravi, K.; Mao, L.; Kisi, O.; Yaseen, Z.M.; Shahid, S. Quantifying hourly suspended sediment load using data mining models: Case study of a glacierized Andean catchment in Chile. J. Hydrol. 2018, 567, 165–179. [Google Scholar] [CrossRef]
  54. Moriasi, D.N.; Arnold, J.G.; Van Liew, M.W.; Bingner, R.L.; Harmel, R.D.; Veith, T.L. Model evaluation guidelines for systematic quantification of accuracy in watershed simulations. Trans. ASABE 2007, 50, 885–900. [Google Scholar] [CrossRef]
  55. Koopialipoor, M.; Fallah, A.; Armaghani, D.J.; Azizi, A.; Mohamad, E.T. Three hybrid intelligent models in estimating flyrock distance resulting from blasting. Eng. Comput. 2019, 35, 243–256. [Google Scholar] [CrossRef] [Green Version]
  56. Asteris, P.G.; Tsaris, A.K.; Cavaleri, L.; Repapis, C.C.; Papalou, A.; Di Trapani, F.; Karypidis, D.F. Prediction of the fundamental period of infilled RC frame structures using artificial neural networks. Comput. Intell. Neurosci. 2016. [Google Scholar] [CrossRef] [Green Version]
  57. Koopialipoor, M.; Armaghani, D.J.; Hedayat, A.; Marto, A.; Gordan, B. Applying various hybrid intelligent systems to evaluate and predict slope stability under static and dynamic conditions. Soft Comput. 2019, 23, 5913–5929. [Google Scholar] [CrossRef]
  58. Andjelkovic, V.; Pavlovic, N.; Lazarevic, Z.; Radovanovic, S. Modelling of shear strength of rockfills used for the construction of rockfill dams. Soils Found. 2018, 58, 881–893. [Google Scholar] [CrossRef]
  59. Yang, Y.; Zhang, Q. A hierarchical analysis for rock engineering using artificial neural networks. Rock Mech. Rock Eng. 1997, 30, 207–222. [Google Scholar] [CrossRef]
  60. Faradonbeh, R.S.; Armaghani, D.J.; Abd Majid, M.; Tahir, M.M.; Murlidhar, B.R.; Monjezi, M.; Wong, H. Prediction of ground vibration due to quarry blasting based on gene expression programming: A new model for peak particle velocity prediction. Int. J. Environ. Sci. Technol. 2016, 13, 1453–1464. [Google Scholar] [CrossRef] [Green Version]
  61. Chen, W.; Hasanipanah, M.; Rad, H.N.; Armaghani, D.J.; Tahir, M. A new design of evolutionary hybrid optimization of SVR model in predicting the blast-induced ground vibration. Eng. Comput. 2021, 37, 1455–1471. [Google Scholar] [CrossRef]
  62. Rad, H.N.; Bakhshayeshi, I.; Jusoh, W.A.W.; Tahir, M.; Foong, L.K. Prediction of flyrock in mine blasting: A new computational intelligence approach. Nat. Resour. Res. 2020, 29, 609–623. [Google Scholar]
  63. Ahmad, M.; Hu, J.-L.; Ahmad, F.; Tang, X.-W.; Amjad, M.; Iqbal, M.J.; Asim, M.; Farooq, A. Supervised learning methods for modeling concrete compressive strength prediction at high temperature. Materials 2021, 14, 1983. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic representation of RF analysis.
Figure 1. Schematic representation of RF analysis.
Applsci 11 06167 g001
Figure 2. The flowchart of the methodology.
Figure 2. The flowchart of the methodology.
Applsci 11 06167 g002
Figure 3. Scatter plots of actual vs. predicted RFM shear strength in training stage: (a) SVM, (b) RF, (c) AdaBoost, and (d) KNN.
Figure 3. Scatter plots of actual vs. predicted RFM shear strength in training stage: (a) SVM, (b) RF, (c) AdaBoost, and (d) KNN.
Applsci 11 06167 g003
Figure 4. Scatter plots of actual vs. predicted RFM shear strength in testing stage: (a) SVM, (b) RF, (c) AdaBoost, and (d) KNN.
Figure 4. Scatter plots of actual vs. predicted RFM shear strength in testing stage: (a) SVM, (b) RF, (c) AdaBoost, and (d) KNN.
Applsci 11 06167 g004
Figure 5. Comparison of R2, NSE, RMSE, and RSR values from the SVM, RF, AdaBoost, and KNN models in (a) training phase and (b) testing phase.
Figure 5. Comparison of R2, NSE, RMSE, and RSR values from the SVM, RF, AdaBoost, and KNN models in (a) training phase and (b) testing phase.
Applsci 11 06167 g005
Figure 6. Sensitivity analysis results.
Figure 6. Sensitivity analysis results.
Applsci 11 06167 g006
Table 1. Rockfill materials shear strength case history data.
Table 1. Rockfill materials shear strength case history data.
Case No.LocationD10/mmD30/mmD60/mmD90/mmCcCuGMFMRUCSmin/MPaUCSmax/MPaγ/KNm−3σn/MPaτ/MPa
1Canada0.020.9441811.052004.784.1911515.40.0220.013
2Canada0.020.9441811.052004.784.1911515.40.0440.025
3Canada0.020.9441811.052004.784.1911515.40.0880.049
163Netherlands111523320.892.091.487.48510025016.80.0280.048
164Netherlands111523320.892.091.487.48510025016.80.0550.082
165Netherlands111523320.892.091.487.48510025016.80.1080.143
Table 2. Statistical parameters of the training and testing data sets.
Table 2. Statistical parameters of the training and testing data sets.
ParameterData SetMin ValueMax ValueMeanStandard Deviation
D10 (mm)Training0.01033.9004.8579.179
Testing0.01033.9002.8877.453
D30 (mm)Training0.56042.4008.46510.577
Testing0.56042.4005.4429.050
D60 (mm)Training1.20080.10019.28715.135
Testing1.20050.00014.25210.349
D90 (mm)Training2.600100.00040.38622.018
Testing2.60099.00038.09124.289
CCTraining0.10022.2702.1993.075
Testing0.10022.2703.2264.492
CUTraining1.3601040.00053.324156.064
Testing1.4701040.000134.510294.958
GMTraining0.2006.0002.7881.243
Testing0.2006.0003.3651.331
FMTraining3.0008.8006.2501.261
Testing3.0008.8005.7091.374
RTraining1.0006.0004.3640.910
Testing1.0005.0004.1821.131
UCSmin (MPa)Training1.000250.00075.04539.230
Testing1.000100.00068.27332.444
UCSmax (MPa)Training5.000400.000170.68288.010
Testing5.000250.000159.54587.957
γ (KN/m3)Training9.32038.90020.7664.605
Testing9.32038.90020.9325.854
σn (MPa)Training0.0024.2050.7290.780
Testing0.0213.2230.7560.816
τ (MPa)Training0.0053.9210.6600.662
Testing0.0242.4920.6680.619
Table 3. Hyperparameter optimization results.
Table 3. Hyperparameter optimization results.
AlgorithmHyperparameterExplanationOptimal Value
AdaBoostNumber of estimatorsNumber of trees2
Learning rateIt establishes the degree to which newly acquired information can override previously acquired information0.1
Boosting algorithmUpdates the weight of the base estimator with probability estimates or classification results (SAMME.R/SAMME)SAMME
Regression loss functionLinear/square/exponentialLinear
RFNumber of treesNumber of trees in the forest15
Limit depth of individual treesThe depth to which the trees will be grown03
SVMCost (C)Penalty term for loss and applies for classification and regression tasks8
Regression loss epsilon (ε)The distance between true and predicted values within which no penalty is applied0.1
Kernal typeKernel is a function that transforms attribute space to a new feature space to fit the maximum-margin hyperplane, thus allowing the algorithm to construct the model with linear, polynomial, RBF, and Sigmoid kernelsRBF
KNNNumber of neighborsNumber of nearest neighbors5
MetricDistance parameter—Euclidean/Manhattan/Chebyshev/MahalanobisEuclidean
WeightUniform—all points in each neighborhood are weighted equally/distance—closer neighbors of a query point have a greater influence than the neighbors further awayUniform
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ahmad, M.; Kamiński, P.; Olczak, P.; Alam, M.; Iqbal, M.J.; Ahmad, F.; Sasui, S.; Khan, B.J. Development of Prediction Models for Shear Strength of Rockfill Material Using Machine Learning Techniques. Appl. Sci. 2021, 11, 6167. https://doi.org/10.3390/app11136167

AMA Style

Ahmad M, Kamiński P, Olczak P, Alam M, Iqbal MJ, Ahmad F, Sasui S, Khan BJ. Development of Prediction Models for Shear Strength of Rockfill Material Using Machine Learning Techniques. Applied Sciences. 2021; 11(13):6167. https://doi.org/10.3390/app11136167

Chicago/Turabian Style

Ahmad, Mahmood, Paweł Kamiński, Piotr Olczak, Muhammad Alam, Muhammad Junaid Iqbal, Feezan Ahmad, Sasui Sasui, and Beenish Jehan Khan. 2021. "Development of Prediction Models for Shear Strength of Rockfill Material Using Machine Learning Techniques" Applied Sciences 11, no. 13: 6167. https://doi.org/10.3390/app11136167

APA Style

Ahmad, M., Kamiński, P., Olczak, P., Alam, M., Iqbal, M. J., Ahmad, F., Sasui, S., & Khan, B. J. (2021). Development of Prediction Models for Shear Strength of Rockfill Material Using Machine Learning Techniques. Applied Sciences, 11(13), 6167. https://doi.org/10.3390/app11136167

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop