Next Article in Journal
Physics-Informed Generative Adversarial Network Solution to Buckley–Leverett Equation
Next Article in Special Issue
Three-Dimensional Stratigraphic Structure and Property Collaborative Modeling in Urban Engineering Construction
Previous Article in Journal
Optimization of Fresh Food Logistics Routes for Heterogeneous Fleets in Segmented Transshipment Mode
Previous Article in Special Issue
A Novel Axial Load Inversion Method for Rock Bolts Based on the Surface Strain of a Bearing Plate
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Surrounding Rock Squeezing Classification in Underground Engineering Using a Hybrid Paradigm of Generative Artificial Intelligence and Deep Ensemble Learning

1
Research Institute of Mine Construction, Tiandi Science and Technology Company Limited, Beijing 100013, China
2
State Key Laboratory of Intelligent Coal Mining and Strata Control, Beijing 100013, China
3
School of Civil Engineering, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(23), 3832; https://doi.org/10.3390/math12233832
Submission received: 7 November 2024 / Revised: 27 November 2024 / Accepted: 3 December 2024 / Published: 4 December 2024
(This article belongs to the Special Issue Mathematical Modeling and Analysis in Mining Engineering)

Abstract

:
Surrounding rock squeezing is a common geological disaster in underground excavation projects (e.g., TBM tunneling and deep mining), which has adverse effects on construction safety, schedule, and property. To predict the squeezing of the surrounding rock accurately and quickly, this study proposes a hybrid machine learning paradigm that integrates generative artificial intelligence and deep ensemble learning. Specifically, conditional tabular generative adversarial network is devised to solve the problems of data shortage and class imbalance for data augmentation at the data level, and the deep random forest is built based on the augmented data for subsequent squeezing classification. A total of 139 historical squeezing cases are collected worldwide to validate the efficacy of the proposed modeling paradigm. The results reveal that this paradigm achieves a prediction accuracy of 92.86% and a macro F1-score of 0.9292. In particular, the individual F1-scores on strong squeezing and extremely strong squeezing are more than 0.9, with excellent prediction reliability for high-intensity squeezing. Finally, a comparative analysis with traditional machine learning techniques is conducted and the superiority of this paradigm is further verified. This study provides a valuable reference for surrounding rock squeezing classification under a limited data environment.

1. Introduction

With the rapid development of social economy and the accelerating advancement of urbanization, the demand for all kinds of infrastructure has presented a significant growth trend, which undoubtedly brings unprecedented challenges to infrastructure construction [1,2]. In particular, tunnel engineering is an indispensable part of water conservancy, transportation, mining, and other fields, and its construction scale and technical difficulty are rising [3]. Today’s tunnels are not only designed to be deeper and longer but also increasingly complex in shape and compact in layout, which means that their construction needs to face harsher geological conditions and construction environments [4]. Weak, fractured, or high-stress areas are all important safety factors that cannot be ignored in tunnel construction, which greatly increase the construction risk [5,6].
Surrounding rock squeezing is a geological disaster frequently occurring in tunnel construction [7,8]. The squeezing of the surrounding rock causes the deformation of the tunnel structure, such as arch subsidence and waist convergence around the tunnel [9]. At the same time, the supporting structure may also be subjected to excessive pressure and fail as a result of the pressure, which will make the tunnel structure lose stability, further aggravate the deformation of the tunnel structure and even lead to its collapse [10].
To predict the risk of surrounding rock squeezing, many researchers have proposed various empirical criteria derived from engineering experience and case studies. For instance, Aydan et al. [11] suggested using the ratio of uniaxial compressive strength of intact rock to vertical ground stress, with a threshold value of two to differentiate between squeezing and non-squeezing conditions. Similarly, Barla [12] introduced a criterion based on the ratio of uniaxial compressive strength of rock mass to vertical ground stress. These criteria mentioned above essentially fall under the category of strength-stress ratios. Additionally, Farhadian et al. [13], Singh et al. [14], and Jimenez et al. [15] developed squeezing criteria that consider extra factors such as burial depth and surrounding rock quality. These criteria have been widely applied to numerous tunnel projects, and their effectiveness has been validated.
As research has progressed, it has become evident that surrounding rock squeezing is a complicated nonlinear problem controlled by multiple trigging agents [7,16,17,18]. Consequently, some researchers have turned to machine learning techniques to predict the squeezing potential of the surrounding rock, given their strong capabilities to handle complex nonlinear problems [19,20,21]. For example, Zhang et al. [22] developed a back-propagation neural network model for squeezing prediction, using burial depth, excavation diameter, strength-stress ratio, rock mass quality index, and support stiffness as input parameters. Similarly, Bo et al. [23] built a logistic regression model to achieve squeezing classification based on the above five parameters, while Feng et al. [24] and Azizi et al. [25] proposed Bayesian network models for the same purpose. Additionally, support vector machine [26,27,28,29], decision tree [30,31], and probabilistic neural network [8] have also been popular methods for predicting the squeezing potential of the surrounding rock. These practises mentioned above fully demonstrate the effectiveness of machine learning techniques in the prediction of surrounding rock squeezing.
However, the performance of machine learning models is highly sensitive to data quality [32,33]. Class imbalance and data shortage are two common data issues. In the context of surrounding rock squeezing, the collection of squeezing cases is difficult, so the available data are usually limited, especially for strong and extremely strong squeezing cases [34]. On the one hand, it exacerbates the data shortage problem. On the other hand, it also aggravates the data imbalance of different squeezing grades. When data are scarce, it is difficult for machine learning models to effectively capture the mapping relationships contained in the data. Additionally, when the classes are imbalanced, machine learning models have a low generalization level for the minority.
On the basis of considering data shortage and class imbalance, this study proposes a hybrid intelligent modeling paradigm for surrounding rock squeezing classification. Firstly, conditional tabular generative adversarial network is designed to cope with data shortage and class imbalance at the data level. After that, the deep random forest is built based on the augmented data to capture the mapping relationship between the trigging agents and the squeezing grades. Finally, the generalization performance of the devised model is evaluated quantitatively using an independent test set and two professional metrics. The rest of this paper is organized as follows: Section 2 introduces the proposed hybrid intelligent modeling paradigm in detail; Section 3 describes the collected database and conducts data augmentation; Section 4 analyzes the generalization performance of the model and makes some comparisons with other common models; and Section 5 summarizes the main conclusions.

2. Proposed Hybrid Modeling Paradigm

This study aims to propose a hybrid modeling paradigm to solve the problems of data shortage and class imbalance and build a model for surrounding rock squeezing classification. Particularly, the task of data augmentation is undertaken by conditional tabular generative adversarial network, and the deep random forest is designed to learn the mapping relationship from the augmented data to achieve squeezing classification. In order to clearly demonstrate the proposed hybrid modeling paradigm, Section 2.1 describes conditional tabular generative adversarial network, Section 2.2 introduces deep random forest, Section 2.3 depicts the comparison scheme used to verify the advantages of the proposed hybrid modeling paradigm, and Section 2.4 illustrates how to evaluate the generalization performance of the model quantitatively. Figure 1 gives the research framework of this study.

2.1. Conditional Tabular Generative Adversarial Network

Conditional tabular generative adversarial network (CTGAN), as the representative algorithm of generative artificial intelligence, is a variant of generative adversarial network (GAN) [35,36]. The CTGAN has two characteristic properties. One is data preprocessing; that is, all continuous features are processed through model (i.e., variational Gaussian mixture model)-specific normalization. The other is conditional training, in which category-level unbalance of categorical features is addressed.
Resembling the GAN, the CTGAN includes the following two modules (Figure 2): generator and discriminator. The aim of the generator is to learn the conditional distribution of real data and then generate synthetic data. The purpose of the discriminator is to distinguish the difference between real data and synthetic data. When synthetic data are very close to real data, the discriminator will output a large score. Instead, when synthesized data differs significantly from real data, it will output a small score. In general, the generator is optimized for deceiving the discriminator as much as possible. If the discriminator cannot tell the difference, then the CTGAN is considered as a finely trained model that can be used to generate synthetic data.
To reserve the essential information of real data and generate effective synthetic data, the loss functions of the generator and discriminator are designed separately. For the generator, the loss function L G is expressed by the following equation:
L G = 1 n / N i = 1 n / N D ν ^ i , κ i + 1 n j = 1 n C E τ j , ε j
For the discriminator, the loss function L D is expressed by the following equation:
L D = 1 n / N i = 1 n / N D ν ^ i , κ i 1 n / N i = 1 n / N D ν i , κ i
where n is the batch size; N is the packing size used to handle model collapse; ν is the real data; ν ^ is the synthetic data; κ is the conditional vector; τ is the one-hot conditional vector; ε is the mask vector associated with τ ; L is the loss function; G is the generator; D is the discriminator; and CE is the cross entropy.
In the CTGAN’s training process, the generator and discriminator are optimized and updated by an Adam optimizer. Based on real data and synthetic data, the L D can be calculated, based on which the discriminator is updated first. Then, synthetic data are delivered to the updated discriminator and the L G can be calculated, based on which the generator is updated. Thus, when building the CTGAN, the discriminator is updated followed by the generator.

2.2. Deep Random Forest

The random forest (RF) is an ensemble learning method for classification and regression tasks [37]. The RF combines multiple single predictors into one integrated predictor, so as to attain excellent generalization performance. These single predictors can be generated in parallel and allow for diversity among each other. Generally, the RF uses binary decision trees, such as classification and regression tree (CART), as single predictors, due to the low parameters required, strong robustness, and low training cost [38]. Figure 3 illustrates the topology of the RF.
Inspired by the layered structure of deep learning, the deep random forest (DRF) is proposed [39]. As illustrated in Figure 4, the DRF creates multiple layers. For each layer, multiple RFs are assembled. Through cascading connections between different layers, the potential features in the input data are represented and learned. For the first layer of the DRF, it receives the original input data as the input, and correspondingly, each RF in this layer will generate a class probability vector, denoted as p = p 1 , p 2 , , p n . In this class probability vector, p n represents the probability that the sample belongs to the n-th class, and the dimension is equal to the number of classes set by the classification tasks. For subsequent layers, the input is composed by two parts: the original input data and the class probability vectors output by random forests in the previous layer. Thus, the dimension of the input is characterized by the following formula:
d i m i n p u t i = d i m d a t a + n × m i 1
where d i m i n p u t i is the input dimension of the current layer; d i m d a t a is the dimension of the original input data; n is the number of classes set by the classification tasks; m i 1 is the number of RFs contained in the previous layer; and i and i − 1 are the serial numbers of the layers.
In the last layer of the DRF, the class probability vectors of all RFs here will be averaged to obtain the final class probability vector, as shown in Equation (4). The index corresponding to the maximum in the final class probability vector is the class of the sample predicted by the DRF.
v e c 0 = 1 U k = 1 U v e c k
where v e c 0 is the final class probability vector, output generated by the last layer of the DRF; v e c k is the k-th class probability vector, output generated by the k-th RF of the last layer; and U is the number of RFs contained in the last layer.
It is well known that the DRF has a strong ability of representation learning [40]. By cascading RFs between multiple layers, the DRF enables fine-grained modelling of data. On the one hand, it maximizes data value. On the other hand, the separability of samples in the mapping space is improved by mining deep-level features. Therefore, the DRF is favourably popular in solving practical engineering problems.

2.3. Benchmark Comparison

To demonstrate the superiority of the proposed model, referred to as the CTGAN-DRF, five frequently used machine learning algorithms, including the back-propagation neural network (BPNN) [41], support vector machine (SVM) [42], Gaussian process classifier (GPC) [43], Naïve Bayesian classifier (NBC) [44], and k-nearest neighbor (KNN) [45], are chosen for benchmark comparison.
The BPNN is an algorithm for forward propagation and back propagation by graph computation. In forward propagation, the input data are passed layer by layer through the network, and the output result is finally obtained. In back propagation, according to the error between the output result and the desired target, the gradient of the parameters of each layer is calculated using the chain rule, and the network parameters are updated to reduce the error.
The SVM is a supervised learning algorithm mainly used for classification and regression tasks. It achieves efficient data classification by finding the best separated hyperplane in the feature space. Moreover, it performs well when dealing with high-dimensional data, nonlinear problems, and few training samples.
The GPC is a supervised classification algorithm based on Bayesian framework. It assumes that there is an implicit function for a Gaussian process, which corresponds monotonically to the probability that the sample belongs to a certain class. By selecting an appropriate likelihood function to describe this mapping relationship, the GPC can calculate the posterior probability of the sample and achieve classification prediction. The GPC has the characteristics of automatic parameter optimization and can give the classification result in probabilistic form.
The NBC is a classification algorithm based on Bayes theorem. It assumes that the sample features are independent of each other, calculates the posterior probability of the sample class using Bayesian formula according to the known prior probability and the probability distribution of the sample class, and then takes the class of the sample with the largest posterior probability as the output of the classifier. The NBC has the advantages of being a stable mathematical model, with good classification effect, and needing few training parameters.
The KNN is a supervised learning algorithm for classification and regression tasks. The classification principle states that if most of the K nearest neighbors near the sample belong to a certain class, then this sample also belongs to that class. The KNN is simple and flexible, without requiring training process, and can be applied to various data types.

2.4. Evaluation Metric

In order to quantitatively evaluate the generalization performance of the model, we select two evaluation metrics, namely accuracy and macro F1-score. The accuracy describes the ability of the model to correctly classify the samples without considering class differences, calculated as follows:
accuracy = k = 1 K u k , k i = 1 K j = 1 K u i , j
where K is the number of sample classes and u i , j is the number of samples belonging to the i-th class but predicted as the j-th class.
Different from the accuracy, the macro F1-score takes into account the model’s local generalization performance on each class. The macro F1-score is defined as the average of all individual F1-scores on various classes, calculated as follows:
macro F 1 - score = 1 K k = 1 K F 1 - score k
where F 1 - score k is the F1-score on the k-th class. It is calculated as follows:
F 1 - score k = 2 P r e k R e c k P r e k + R e c k
where P r e k and R e c k are the precision and the recall on the k-th class, respectively. They are calculated as follows:
P r e k = u k , k i = 1 K u i , k
R e c k = u k , k j = 1 K u k , j

3. Data Collection and Augmentation

3.1. Data Collection

To verify the effectiveness of the proposed approach, we collected 139 cases from 42 tunnel projects worldwide to build the database (see Supplementary Materials) [46]. Each case contains five input parameters, including burial depth, excavation diameter, strength-stress ratio, rock mass quality index, and support stiffness. Regarding their acquisition, the detailed description can be found in [46]. In essence, the occurrence of surrounding rock squeezing is closely related to excavation scale, ground stress, rock strength, rock mass structure, and support conditions [29]. Among the input parameters, burial depth is an indicator associated with ground stress, excavation diameter reflects the excavation scale, strength-stress ratio is a composite index integrating rock strength and ground stress, rock mass quality index emphasizes rock mass structure, and support stiffness characterizes applied support conditions. Figure 5 presents the visual statistical distribution of various parameters in this database.
A personalized label, namely squeezing strength of surrounding rock, is used as the output parameter of each case. Based on the relative strain of the surrounding rock, the squeezing strength is divided into five grades (Table 1) [8,47]. When the relative strain of the surrounding rock is <1.0%, 1.0%~2.5%, 2.5%~5.0%, 5.0%~10.0%, and >10.0%, it indicates no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively. In the established database, there are 28 cases with no squeezing (accounting for 20.14%), 37 cases with mild squeezing (accounting for 26.62%), 33 cases with moderate squeezing (accounting for 23.74%), 26 cases with strong squeezing (accounting for 18.71%), and 15 cases with extremely strong squeezing (accounting for 10.79%), as shown in Figure 6.
In order to ensure the representativeness of the test set, the ratio of training set to test set is taken as 7:3; that is, the training set has 97 cases, and the test set has 42 cases. In this way, there are enough test samples to reliably test the generalization performance of the model. In addition, the samples of different classes in the test set should be as balanced as possible. Based on these guidelines, we randomly select 8 cases with no squeezing, 8 cases with mild squeezing, 8 cases with moderate squeezing, 9 cases with strong squeezing, and 9 cases with extremely strong squeezing from the database to construct the test set. Correspondingly, the rest cases in the database are used as the training set, including 20, 29, 25, 17, 6 cases with no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.

3.2. Data Augmentation

In order to avoid disturbance to the test set and ensure the independence of the test set, data augmentation is only implemented on the training set. Firstly, the CTGAN is trained based on the true data (i.e., the training set). After that, the trained CTGAN is used to generate the synthetic data. On the one hand, the class imbalance is eliminated. On the other hand, the data shortage issue is remedied. After data augmentation, the number of cases with no squeezing is increased from 20 to 200. Similarly, the number of cases with mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing is also increased to 200, respectively. Table 2 summarizes the changes in the training set after data augmentation. Correspondingly, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 display the kernel density distribution of each parameter before and after data augmentation under varying squeezing grades.
Correlation coefficient matrix is used to evaluate the quality of the synthetic data. The higher the similarity between the correlation coefficient matrices of the synthetic data and the true data, the better the quality of the synthetic data, which represents that the synthetic data have more consistent underlying topology with the true data. For the “no squeezing” class, Figure 12a,b illustrate the correlation coefficient matrices of the true data and the synthetic data, respectively, and Figure 12c calculates the absolute difference between the correlation coefficient matrices of the true data and the synthetic data. In particular, the value in each box of Figure 12a,b is characterized by the Pearson correlation coefficient, which is a commonly used indictor to measure correlation. Figure 13, Figure 14, Figure 15 and Figure 16 present the relevant analysis results for the “mild squeezing” class, “moderate squeezing” class, “strong squeezing” class, and “extremely strong squeezing” class, respectively. As shown in Figure 12c, all absolute differences are less than 0.2, which mean that the synthetic data have a similar internal topology to the real data, and the CTGAN attains a high-level capability to learn from the real data and is capable of producing high-quality synthetic data. In addition to the no-squeezing samples, the consistent pattern can be found for both the moderate-squeezing samples and strong-squeezing samples when utilizing the CTGAN to generate the synthetic data, as presented in Figure 14c and Figure 15c, respectively. However, there are some anomalies for the mild-squeezing samples and extremely strong-squeezing samples. For example, as demonstrated in Figure 13c, the absolute difference in correlation coefficients between burial depth and excavation diameter for the real data and the synthetic data slightly exceeds 2, reaching 0.2071; and as indicated in Figure 16c, the absolute difference in correlation coefficients between burial depth and strength-stress ratio for the real data and the synthetic data is 0.3099. Nevertheless, most of the differences are within 0.2. Overall, the CTGAN has a superior data augmentation ability for various squeezing intensities.

4. Results and Discussion

4.1. Prediction Results of CTGAN-DRF Model

After data augmentation using the CTGAN, the DRF is trained using the augmented data. Correspondingly, the built model is referred to as the CTGAN-DRF. Based on the independent test set, the generalization performance of the CTGAN-DRF is tested quantitatively. In particular, the training and testing of the model are executed on a personal computer with Windows 10, i7-9750H CPU, 16GB RAM, and RTX 2060. Figure 17 depicts the confusion matrix of the CTGAN-DRF on the test set. In this confusion matrix, the sum of the values in each row is equal to the number of test samples of a specific class. From the bottom up, the sum of each row is eight, eight, eight, nine, and nine, which indicates that the number of test samples belonging to no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing is eight, eight, eight, nine, and nine, respectively.
Particularly, the values on the diagonal of the confusion matrix denote the number of correctly predicted test samples, and those on the non-diagonal represent the number of wrongly predicted test samples. In terms of no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, the number of correctly predicted test samples by the CTGAN-DRF is eight, seven, eight, eight, and eight, respectively, accounting for 100.00%, 87.50%, 100.00%, 88.89%, and 88.89% of each corresponding class of test samples. For mild squeezing, there is one wrongly predicted test sample out of eight test samples, and the predicted squeezing grade is no squeezing. For both strong squeezing and extremely strong squeezing, there is also one incorrect prediction. Regrading to both the former and the latter, the misclassified label is moderate squeezing.
Relying on the confusion matrix, the accuracy and macro F1-score of the CTGAN-DRF are calculated using Equation (5) and Equation (6), respectively. The results reveal that the accuracy of the CTGAN-DRF is 92.86%, and the macro F1-score reaches a score of 0.9292, as shown in Figure 18a. Furthermore, the F1-score of the CTGAN-DRF on each individual class is analyzed, as illustrated in Figure 18b. It can be observed that the F1-scores on strong squeezing and extremely strong squeezing exceed 0.9, which means that the CTGAN-DRF attains superior prediction reliability for high-grade squeezing disasters and makes significant contribution to squeezing management.

4.2. Comparison Analysis with Benchmark Models

In this section, we compared the proposed CTGAN-DRF with the BPNN, SVM, GPC, NBC, and KNN in terms of the prediction performance of squeezing disaster. The confusion matrices of the BPNN, SVM, GPC, NBC, and KNN are shown in Figure 19, based on which the accuracy and macro F1-score of the BPNN, SVM, GPC, NBC, and KNN are calculated according to Equation (5) and Equation (6), respectively. Referring to Figure 20, in comparison with the BPNN, SVM, GPC, NBC, and KNN, the accuracy of the CTGAN-DRF is increased by 45.24%, 40.48%, 47.62%, 50.00%, and 57.14%, respectively, and the macro F1-score is raised by 0.4579, 0.4075, 0.4812, 0.5003, and 0.5834, respectively. Based on these strong evidences, the superiority of the CTGAN-DRF is demonstrated.

5. Conclusions

Surrounding rock squeezing is a kind of geological disaster that occurs frequently in underground construction, posing a serious threat to safety, schedule, and property. The physical mechanism of surrounding rock squeezing is complex, whose occurrence is influenced by multiple factors and exhibits high nonlinearity with them. In order to establish the mapping relationship between the squeezing and the influencing factors, machine learning is introduced. Aiming at data shortage, class imbalance, and nonlinear topology, this study proposes a hybrid machine learning modeling paradigm. In this paradigm, conditional tabular generative adversarial network (CTGAN) is designed to achieve data augmentation, and the deep random forest (DRF) is integrated to construct the nonlinear topology from the augmented data. Consequently, the CTGAN-DRF model is developed for surrounding rock squeezing classification. To validate the efficacy of the developed model, a total of 139 historical squeezing cases are collected. The results indicate that the CTGAN-DRF model achieves a prediction accuracy of 92.86% and macro F1-score of 0.9292. In particular, the individual F1-scores on strong squeezing and extremely strong squeezing are more than 0.9, which means that the CTGAN-DRF model has excellent prediction reliability for high-intensity squeezing. Additionally, in comparison with other common machine learning models, such as back-propagation neural network, support vector machine, Gaussian process classifier, Naïve Bayesian classifier, and k-nearest neighbor, the CTGAN-DRF model exhibits an advantage in its generalization ability.
Overall, this study leverages the geological, support, and construction data under a specific environment to evaluate the potential surrounding rock squeezing tendency. The developed model is effective and meets the construction requirement, and it can provide the prediction results of squeezing tendency to assist engineers in taking advanced measures so as to avoid squeezing disaster. In particular, it also considers data shortage and class imbalance in the modelling, providing a meaningful reference for surrounding rock squeezing classification and other similar engineering problems with limited data quality. However, this study still has some limitations. First, due to the difficulty of real-time acquisition of input parameters, the model developed here is biased towards long-term predictions. If we plan real-time predictions, the real-time monitoring surrounding rock deformation data can be considered as the input to establish the model. Second, both the training and validation of the current model are executed based on the collected historical squeezing cases. In future research, we will deploy this model in real projects to further examine its performance under new engineering environments. If necessary, incremental learning can be used to update the current model. Third, exploring more data mining and feature characterization techniques is still an important task in future research, which can improve data utilization and decrease data loss in the modeling, especially when there is limited data available. Finally, it should be noted that surrounding rock squeezing is a very complex, nonlinear process, and its occurrence is controlled by many factors. In the future, we need to pay more attention to establish the input system and take into account as many squeezing-inducing factors as possible. At the same time, the investigation on novel squeezing prediction models, such as fuzzy logic predictor, is also necessary for squeezing management.

Supplementary Materials

The following supporting information can be downloaded at the following website: https://www.mdpi.com/article/10.3390/math12233832/s1, Table S1: Supplementary Material (139 squeezing cases).

Author Contributions

Conceptualization: S.C. and X.Y.; Methodology: S.C., X.Y., F.G. and Y.P.; Investigation: S.C., X.Y., F.G. and Y.P.; Writing—original draft: S.C. and X.Y.; Writing—review and editing: S.C., X.Y., F.G. and Y.P.; Validation: S.C. and X.Y.; Supervision: X.Y.; Data curation: X.Y.; Funding acquisition: X.Y., F.G. and Y.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant Nos. 42472351, 42177140, and 52404127), the Natural Science Foundation of Hubei Province (Grant No. 2024AFD359), and the Young Elite Scientist Sponsorship Program by CAST (Grant No. YESS20230742).

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Materials; further inquiries can be directed to the corresponding author/s.

Conflicts of Interest

Authors Shouye Cheng and Feng Gao were employed by Tiandi Science and Technology Company Limited. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhou, S.T.; Zhang, Z.X.; Luo, X.D.; Niu, S.S.; Jiang, N.; Yao, Y.K. Developing a hybrid CEEMDAN-PE-HE-SWT method to remove the noise of measured carbon dioxide blast wave. Measurement 2023, 223, 113797. [Google Scholar] [CrossRef]
  2. Yin, X.; Cheng, S.; Yu, H.; Pan, Y.; Liu, Q.; Huang, X.; Gao, F.; Jing, G. Probabilistic assessment of rockburst risk in TBM-excavated tunnels with multi-source data fusion. Tunn. Undergr. Space Technol. 2024, 152, 105915. [Google Scholar] [CrossRef]
  3. Zhu, H.H.; Yan, J.X.; Liang, W.H. Challenges and Development Prospects of Ultra-Long and Ultra-Deep Mountain Tunnels. Engineering 2019, 5, 384–392. [Google Scholar] [CrossRef]
  4. Gong, Q.M.; Yin, L.J.; Ma, H.S.; Zhao, J. TBM tunnelling under adverse geological conditions: An overview. Tunn. Undergr. Space Technol. 2016, 57, 4–17. [Google Scholar] [CrossRef]
  5. Yin, X.; Liu, Q.; Huang, X.; Pan, Y. Perception model of surrounding rock geological conditions based on TBM operational big data and combined unsupervised-supervised learning. Tunn. Undergr. Space Technol. 2022, 120, 104285. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Zhang, Q.; Zhang, X.; Li, M.; Qi, G. How Do We Analyze the Accident Causation of Shield Construction of Water Conveyance Tunnels? A Method Based on the N-K Model and Complex Network. Mathematics 2024, 12, 3222. [Google Scholar] [CrossRef]
  7. Fathipour-Azar, H. Multi-level Machine Learning-Driven Tunnel Squeezing Prediction: Review and New Insights. Arch. Comput. Methods Eng. 2022, 29, 5493–5509. [Google Scholar] [CrossRef]
  8. Huang, X.; Yin, X.; Liu, B.; Ding, Z.W.; Zhang, C.F.; Jing, B.Y.; Guo, X.S. A Gray Wolf Optimization-Based Improved Probabilistic Neural Network Algorithm for Surrounding Rock Squeezing Classification in Tunnel Engineering. Front. Earth Sci. 2022, 10, 857463. [Google Scholar] [CrossRef]
  9. Ramoni, M.; Anagnostou, G. Tunnel boring machines under squeezing conditions. Tunn. Undergr. Space Technol. 2010, 25, 139–157. [Google Scholar] [CrossRef]
  10. Arora, K.; Gutierrez, M.; Hedayat, A.; Xia, C.C. Tunnels in squeezing clay-rich rocks. Undergr. Space 2021, 6, 432–445. [Google Scholar] [CrossRef]
  11. Aydan, O.; Akagi, T.; Kawamoto, T. The squeezing potential of rock around tunnels: Theory and prediction with examples taken from Japan. Rock Mech. Rock Eng. 1996, 29, 125–143. [Google Scholar] [CrossRef]
  12. Barla, G. Tunnelling under squeezing rock conditions. Eurosummer-Sch. Tunn. Mech. Innsbr. 2001, 169–268. [Google Scholar]
  13. Farhadian, H.; Nikvar-Hassani, A. Development of a new empirical method for Tunnel Squeezing Classification (TSC). Q. J. Eng. Geol. Hydrogeol. 2020, 53, 655–660. [Google Scholar] [CrossRef]
  14. Singh, B.; Jethwa, J.L.; Dube, A.K.; Singh, B. Correlation Between Observed Support Pressure and Rock Mass Quality. Tunn. Undergr. Space Technol. 1992, 7, 59–74. [Google Scholar] [CrossRef]
  15. Jimenez, R.; Recio, D. A linear classifier for probabilistic prediction of squeezing conditions in Himalayan tunnels. Eng. Geol. 2011, 121, 101–109. [Google Scholar] [CrossRef]
  16. Tian, Y.; Chen, W.Z.; Tian, H.M.; Tan, X.J.; Li, Z.; Lei, J. Parameter design of yielding layers for squeezing tunnels. Tunn. Undergr. Space Technol. 2021, 108, 103694. [Google Scholar] [CrossRef]
  17. Cao, C.Y.; Shi, C.H.; Lei, M.F.; Yang, W.C.; Liu, J.W. Squeezing failure of tunnels: A case study. Tunn. Undergr. Space Technol. 2018, 77, 188–203. [Google Scholar] [CrossRef]
  18. Arora, K.; Gutierrez, M.; Hedayat, A. New Physical Model to Study Tunnels in Squeezing Clay-Rich Rocks. Geotech. Test. J. 2021, 44, 1055–1078. [Google Scholar] [CrossRef]
  19. Yin, X.; Liu, Q.; Lei, J.; Pan, Y.; Huang, X.; Lei, Y. Hybrid deep learning-based identification of microseismic events in TBM tunnelling. Measurement 2024, 238, 115381. [Google Scholar] [CrossRef]
  20. Zhou, S.T.; Lei, Y.; Zhang, Z.X.; Luo, X.D.; Aladejare, A.; Ozoji, T. Estimating dynamic compressive strength of rock subjected to freeze-thaw weathering by data-driven models and non-destructive rock properties. Nondestruct. Test. Eval. 2024. [Google Scholar] [CrossRef]
  21. Liu, Q.S.; Lei, Y.M.; Yin, X.; Lei, J.S.; Pan, Y.C.; Sun, L. Development and application of a novel probabilistic back-analysis framework for geotechnical parameters in shield tunneling based on the surrogate model and Bayesian theory. ACTA Geotech. 2023, 18, 4899–4921. [Google Scholar] [CrossRef]
  22. Zhang, J.F.; Li, D.; Wang, Y.H. Predicting tunnel squeezing using a hybrid classifier ensemble with incomplete data. Bull. Eng. Geol. Environ. 2020, 79, 3245–3256. [Google Scholar] [CrossRef]
  23. Bo, Y.; Huang, X.; Pan, Y.C.; Feng, Y.F.; Deng, P.H.; Gao, F.; Liu, P.; Liu, Q.S. Robust model for tunnel squeezing using Bayesian optimized classifiers with partially missing database. Undergr. Space 2023, 10, 91–117. [Google Scholar] [CrossRef]
  24. Feng, X.D.; Jimenez, R. Predicting tunnel squeezing with incomplete data using Bayesian networks. Eng. Geol. 2015, 195, 214–224. [Google Scholar] [CrossRef]
  25. Azizi, F.; Koopialipoor, M.; Khoshrou, H. Estimation of Rock Mass Squeezing Potential in Tunnel Route (Case Study: Kerman Water Conveyance Tunnel). Geotech. Geol. Eng. 2019, 37, 1671–1685. [Google Scholar] [CrossRef]
  26. Shafiei, A.; Parsaei, H.; Dusseault, M.B. Rock Squeezing Prediction By a Support Vector Machine Classifier. In Proceedings of the 46th U.S. Rock Mechanics/Geomechanics Symposium, Chicago, IL, USA, 24–27 June 2012. [Google Scholar]
  27. Sun, Y.; Feng, X.D.; Yang, L.Q. Predicting Tunnel Squeezing Using Multiclass Support Vector Machines. Adv. Civ. Eng. 2018, 2018, 4543984. [Google Scholar] [CrossRef]
  28. Huang, Z.; Liao, M.X.; Zhang, H.L.; Zhang, J.B.; Ma, S.K.; Zhu, Q.X. Predicting Tunnel Squeezing Using the SVM-BP Combination Model. Geotech. Geol. Eng. 2022, 40, 1387–1405. [Google Scholar] [CrossRef]
  29. Zhou, J.; Zhu, S.L.; Qiu, Y.G.; Armaghani, D.; Zhou, A.N.; Yong, W.X. Predicting tunnel squeezing using support vector machine optimized by whale optimization algorithm. Acta Geotech. 2022, 17, 1343–1366. [Google Scholar] [CrossRef]
  30. Ghasemi, E.; Gholizadeh, H. Prediction of Squeezing Potential in Tunneling Projects Using Data Mining-Based Techniques. Geotech. Geol. Eng. 2019, 37, 1523–1532. [Google Scholar] [CrossRef]
  31. Chen, Y.; Li, T.B.; Zeng, P.; Ma, J.J.; Patelli, E.; Edwards, B. Dynamic and Probabilistic Multi-class Prediction of Tunnel Squeezing Intensity. Rock Mech. Rock Eng. 2020, 53, 3521–3542. [Google Scholar] [CrossRef]
  32. He, B.; Armaghani, D.J.; Lai, S.H.; Samui, P.; Mohamad, E.T. Applying data augmentation technique on blast-induced overbreak prediction: Resolving the problem of data shortage and data imbalance. Expert Syst. Appl. 2024, 237, 121616. [Google Scholar] [CrossRef]
  33. Du, W.; Zhu, P.; Pu, Z.; Gong, X.; Li, C. Data augmentation on fault diagnosis of wind turbine gearboxes with an enhanced flow-based generative model. Measurement 2024, 225, 113985. [Google Scholar] [CrossRef]
  34. Liang, M.; Peng, H.; Xie, W.W.; Yu, B.; Han, Y.; Zhu, M.L.; Song, G.X.; Huang, N.H. Dynamic multiclass prediction of tunnel squeezing intensity with stacking model and Markov process. Tunn. Undergr. Space Technol. 2024, 146, 105632. [Google Scholar] [CrossRef]
  35. Ma, H.A.; Geng, M.Y.; Wang, F.; Zheng, W.Y.; Ai, Y.B.; Zhang, W.D. Data Augmentation of a Corrosion Dataset for Defect Growth Prediction of Pipelines Using Conditional Tabular Generative Adversarial Networks. Materials 2024, 17, 1142. [Google Scholar] [CrossRef] [PubMed]
  36. Hong, E.B.; Yi, J.S.; Lee, D.H. CTGAN-Based Model to Mitigate Data Scarcity for Cost Estimation in Green Building Projects. J. Manag. Eng. 2024, 40, 04024024. [Google Scholar] [CrossRef]
  37. Talekar, B.; Agrawal, S. A Detailed Review on Decision Tree and Random Forest. Biosci. Biotechnol. Res. Commun. 2020, 13, 245–248. [Google Scholar] [CrossRef]
  38. Loh, W.Y. Fifty Years of Classification and Regression Trees. Int. Stat. Rev. 2014, 82, 329–348. [Google Scholar] [CrossRef]
  39. Guehairia, O.; Ouamane, A.; Dornaika, F.; Taleb-Ahmed, A. Feature fusion via Deep Random Forest for facial age estimation. Neur. Net. 2020, 130, 238–252. [Google Scholar] [CrossRef]
  40. Tong, F.; Zhang, Y. Exploiting Spectral-Spatial Information Using Deep Random Forest for Hyperspectral Imagery Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5509505. [Google Scholar] [CrossRef]
  41. Wythoff, B.J. Backpropagation neural networks: A tutorial. Chemom. Intell. Lab. Syst. 1993, 18, 115–155. [Google Scholar] [CrossRef]
  42. Pisner, D.A.; Schnyer, D.M. Chapter 6–Support vector machine. In Machine Learning; Mechelli, A., Vieira, S., Eds.; Academic Press: Cambridge, MA, USA, 2020; pp. 101–121. [Google Scholar]
  43. Kapoor, A.; Grauman, K.; Urtasun, R.; Darrell, T. Gaussian Processes for Object Categorization. Int. J. Comput. Vis. 2010, 88, 169–188. [Google Scholar] [CrossRef]
  44. Xu, S. Bayesian Naïve Bayes classifiers to text classification. J. Inf. Sci. 2016, 44, 48–59. [Google Scholar] [CrossRef]
  45. Kramer, O. (Ed.) K-Nearest Neighbors. In Dimensionality Reduction with Unsupervised Nearest Neighbors; Springer: Berlin/Heidelberg, Germany, 2013; pp. 13–23. [Google Scholar]
  46. Jiao, Y.; Ou, G.; Wang, H.; Zhang, G.; Zou, J.; Tan, F. Prediction of Tunnel Squeezing Based on Evidence Theory. J. Appl. Basic Eng. Sci. 2021, 29, 1156–1170. [Google Scholar]
  47. Hoek, E.; Marinos, P. Predicting tunnel squeezing problems in weak heterogeneous rock masses. Tunn. Tunn. Int. 2000, 32, 45–51. [Google Scholar]
Figure 1. Research framework of this study.
Figure 1. Research framework of this study.
Mathematics 12 03832 g001
Figure 2. The components of the CTGAN.
Figure 2. The components of the CTGAN.
Mathematics 12 03832 g002
Figure 3. The topology of the RF.
Figure 3. The topology of the RF.
Mathematics 12 03832 g003
Figure 4. The topology of the DRF.
Figure 4. The topology of the DRF.
Mathematics 12 03832 g004
Figure 5. Visual statistical distribution: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness. The R1, R2, R3, R4, and R5 denote no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Figure 5. Visual statistical distribution: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness. The R1, R2, R3, R4, and R5 denote no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Mathematics 12 03832 g005
Figure 6. The proportion of various squeezing intensities.
Figure 6. The proportion of various squeezing intensities.
Mathematics 12 03832 g006
Figure 7. Kernel density distribution for the no-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Figure 7. Kernel density distribution for the no-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Mathematics 12 03832 g007
Figure 8. Kernel density distribution for the mild-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Figure 8. Kernel density distribution for the mild-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Mathematics 12 03832 g008
Figure 9. Kernel density distribution for the moderate-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Figure 9. Kernel density distribution for the moderate-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Mathematics 12 03832 g009
Figure 10. Kernel density distribution for the strong-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Figure 10. Kernel density distribution for the strong-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Mathematics 12 03832 g010
Figure 11. Kernel density distribution for the extremely strong-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Figure 11. Kernel density distribution for the extremely strong-squeezing samples: (a) burial depth; (b) excavation diameter; (c) strength-stress ratio; (d) rock mass quality index; and (e) support stiffness.
Mathematics 12 03832 g011
Figure 12. Similarity evaluation for the no-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data. The X1, X2, X3, X4, and X5 denote burial depth, excavation diameter, strength-stress ratio, rock mass quality index, and support stiffness, respectively. In Figure 13, Figure 14, Figure 15 and Figure 16, the meaning of the abbreviations (i.e., X1, X2, X3, X4, and X5) is the same as that in Figure 12.
Figure 12. Similarity evaluation for the no-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data. The X1, X2, X3, X4, and X5 denote burial depth, excavation diameter, strength-stress ratio, rock mass quality index, and support stiffness, respectively. In Figure 13, Figure 14, Figure 15 and Figure 16, the meaning of the abbreviations (i.e., X1, X2, X3, X4, and X5) is the same as that in Figure 12.
Mathematics 12 03832 g012
Figure 13. Similarity evaluation for the mild-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Figure 13. Similarity evaluation for the mild-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Mathematics 12 03832 g013
Figure 14. Similarity evaluation for the moderate-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Figure 14. Similarity evaluation for the moderate-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Mathematics 12 03832 g014
Figure 15. Similarity evaluation for the strong-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Figure 15. Similarity evaluation for the strong-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Mathematics 12 03832 g015
Figure 16. Similarity evaluation for the extremely strong-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Figure 16. Similarity evaluation for the extremely strong-squeezing samples: (a) correlation coefficient matrix of real data; (b) correlation coefficient matrix of synthetic data; and (c) absolute difference between correlation coefficient matrices of true data and synthetic data.
Mathematics 12 03832 g016
Figure 17. Confusion matrix of the CTGAN-DRF. The R1, R2, R3, R4, and R5 denote no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Figure 17. Confusion matrix of the CTGAN-DRF. The R1, R2, R3, R4, and R5 denote no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Mathematics 12 03832 g017
Figure 18. Performance evaluation of the CTGAN-DRF: (a) global performance and (b) local performance. The F1-score1, F1-score2, F1-score3, F1-score4, and F1-score5 denote the F1-score of the model on no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Figure 18. Performance evaluation of the CTGAN-DRF: (a) global performance and (b) local performance. The F1-score1, F1-score2, F1-score3, F1-score4, and F1-score5 denote the F1-score of the model on no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Mathematics 12 03832 g018
Figure 19. Confusion matrix: (a) BPNN; (b) SVM; (c) GPC; (d) NBC; and (e) KNN. The R1, R2, R3, R4, and R5 denote no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Figure 19. Confusion matrix: (a) BPNN; (b) SVM; (c) GPC; (d) NBC; and (e) KNN. The R1, R2, R3, R4, and R5 denote no squeezing, mild squeezing, moderate squeezing, strong squeezing, and extremely strong squeezing, respectively.
Mathematics 12 03832 g019
Figure 20. Performance comparison of different models.
Figure 20. Performance comparison of different models.
Mathematics 12 03832 g020
Table 1. Classification criteria of surrounding rock squeezing.
Table 1. Classification criteria of surrounding rock squeezing.
Squeezing Grade of the Surrounding RockRelative Strain of the Surrounding Rock
No<1.0%
Mild1.0%~2.5%
Moderate2.5%~5.0%
Strong5.0%~10.0%
Extremely strong>10.0%
Table 2. Data augmentation overview.
Table 2. Data augmentation overview.
Squeezing IntensitySample Size
Before Data AugmentationAfter Data Augmentation
No20200
Mild29200
Moderate25200
Strong17200
Extremely strong6200
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cheng, S.; Yin, X.; Gao, F.; Pan, Y. Surrounding Rock Squeezing Classification in Underground Engineering Using a Hybrid Paradigm of Generative Artificial Intelligence and Deep Ensemble Learning. Mathematics 2024, 12, 3832. https://doi.org/10.3390/math12233832

AMA Style

Cheng S, Yin X, Gao F, Pan Y. Surrounding Rock Squeezing Classification in Underground Engineering Using a Hybrid Paradigm of Generative Artificial Intelligence and Deep Ensemble Learning. Mathematics. 2024; 12(23):3832. https://doi.org/10.3390/math12233832

Chicago/Turabian Style

Cheng, Shouye, Xin Yin, Feng Gao, and Yucong Pan. 2024. "Surrounding Rock Squeezing Classification in Underground Engineering Using a Hybrid Paradigm of Generative Artificial Intelligence and Deep Ensemble Learning" Mathematics 12, no. 23: 3832. https://doi.org/10.3390/math12233832

APA Style

Cheng, S., Yin, X., Gao, F., & Pan, Y. (2024). Surrounding Rock Squeezing Classification in Underground Engineering Using a Hybrid Paradigm of Generative Artificial Intelligence and Deep Ensemble Learning. Mathematics, 12(23), 3832. https://doi.org/10.3390/math12233832

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop