3.1. Geological Situation and Time Series Data of Baijiabao Landslide
The Baijiabao Landslide is situated within the second group of Xiangjiadian Village, Guizhou Town, Zigui County, located in the Three Gorges Reservoir area, approximately 2.5 km from the confluence of the Xiangxi River and the Yangtze River. It lies 41.2 km from the Three Gorges Dam, positioned on the right bank of the Xiangxi River, a northern tributary of the Yangtze River. The geographical coordinates of the landslide are 30°58′59.9″ N, 110°45′33.4″ E. The terrain at the landslide site is complex, featuring a mix of gentle slopes and gullies, leading to varied landforms. The sliding surface of the landslide typically presents an arc shape. The profile and the distribution of monitoring equipment locations are illustrated in
Figure 2. The data were derived from field survey data of the landslide site.
The Baijiabao landslide is situated on the right bank of the Xiangxi River, a tributary of the northern bank of the Yangtze River. Seasonally, the area experiences precipitation predominantly from April to September, characterized by cooler temperatures in winter and rainy, humid, and warmer summers. Structurally, two distinct groups of geological joints are identified at the landslide site; one group facilitates the collapse of the landslide’s trailing edge, while the other bisects the landslide into left and right segments [
11].
To safeguard the lives and property of the local villagers, GNSS-based surface displacement monitoring of the Baijiabao landslide commenced in October 2006. This monitoring strategy utilizes multiple GNSS sensors strategically placed across the landslide to record precise displacement data, culminating in comprehensive datasets on the cumulative displacement. Over time, this allows researchers to trace and analyze the progression of deformation within the landslide. For this initiative, four monitoring sites designated as ZG323, ZG324, ZG325, and ZG326 were established. Sites ZG324 and ZG325 are located in the principal area of landslide activity, while ZG323 and ZG326 are positioned in the middle and lower sections, near the Zixing. The instrumentation employed delivers exceptionally high measurement accuracy [
12].
From November 2006 to December 2011, data from four displacement sensors—ZG323, ZG324, ZG325, and ZG326—were collected, providing 75 data periods. Sensor ZG323, located near the highway, provided the primary dataset for this study, supplemented by corresponding monthly rainfall and local reservoir water levels [
13]. These comprehensive data are illustrated in
Figure 3, sourced from the archived records of the China Geological Environmental Monitoring Institute, Three Gorges Center. As illustrated in the figure, landslide stability is significantly influenced by rainfall and reservoir water levels. During periods of heavy rainfall, water infiltrates the soil, increasing its weight and reducing its shear strength, thereby promoting landslide activity. Similarly, fluctuations in the reservoir water levels can alter the hydrostatic pressure within the slope materials. An increase in the water level can lead to slope saturation, reducing its stability, while a decrease may lessen the support against the slope materials, potentially triggering landslides. Therefore, reservoir bank landslides induced by rainfall exhibit distinct periodic and trend characteristics.
Time series data form the data foundation for this study, originating from four GNSS monitoring stations distributed across the landslide body. Due to signal interruptions and equipment failures, the data from each monitoring point are not always continuous, with varying degrees of missing information. The ZG323 monitoring point has the least data missing and is conveniently located in the middle of the landslide, making it the primary dataset for this study. Missing data from ZG323 are supplemented with data from the other three stations, resulting in a continuous time series dataset spanning from January 2007 to December 2011. GNSS monitoring data are vectorial and three-dimensional, meaning their magnitude does not directly represent accurate landslide displacement. To characterize the changes more precisely in landslide displacement, we project the vectorial displacement data along the direction of the landslide. Typically, monitoring occurs once a month, but the frequency increases during the rainy season from June to September each year, with two monitoring sessions being conducted each month. Ultimately, the number of data periods applied to the model totals 75, comprising 60 + 15 periods, where 60 represents the sum of monthly displacement data periods from January 2007 to December 2011, and 15 accounts for the sum of the second monitoring periods during the period of June to September in certain years. This ensures consistency in the direction of the data used, establishing an accurate dataset for subsequent model applications.
Given the step-like nature and evident periodicity of displacement at Baijiabao, this study proposes a landslide prediction method based on displacement decomposition using ICEEMDAN-SSA. This technique enables the precise extraction of trend and periodic terms from the landslide displacement data. A univariate LSTM model predicts the trend term, while a multivariate LSTM model addresses the periodic term. In this predictive framework, data from 2007 to 2010 serve as the training set, and data from 2011 are designated as the prediction set.
3.2. ICEEMDAN
Empirical Modal Decomposition (EMD) is a signal decomposition technique that iteratively extracts a series of intrinsic modal functions (IMFs), each representing different vibration modes at corresponding time scales. Traditional EMD algorithms often suffer from issues such as excessive mode extraction and mode aliasing. To address these issues, enhancements such as Completely Adaptive Noise Ensemble Empirical Modal Decomposition (CEEMDAN) and Improved Adaptive Noise Ensemble Empirical Modal Decomposition (ICEEMDAN) have been developed. CEEMDAN introduces adaptive noise into the original signal, followed by independent decomposition and averaging, which enhances denoising, improves accuracy, and mitigates modal aliasing to a significant extent.
ICEEMDAN, an advancement of the CEEMDAN method, is capable of decomposing complex and non-smooth signals into a trend term and a series of IMFs. In ICEEMDAN, IMFs must satisfy two conditions: (1) the number of orthogonal extreme points in any IMF must be equal or differ by no more than one throughout the signal time domain, and (2) the local mean of the function must be zero at any given moment. Through step-by-step decomposition into IMFs, ICEEMDAN achieves more accurate extraction of the signal structure, enhancing the precision and stability of signal processing analysis. ICEEMDAN is extensively applied in medical signal measurement, signal processing, image processing, and time series data analysis, offering robust methods for both analysis and processing [
14].
3.3. SSA
SSA is a technique for processing nonlinear time series data. It can effectively decompose the total displacement time series into multiple sub-sequences and aggregate them into trend terms and periodic terms while removing noise. Its specific process mainly consists of four parts: embedding, decomposition, grouping, and reconstruction [
15]. It is carried out through the following steps:
(1) The Creation of a Trajectory Matrix (embed step). SSA focuses on the analysis of one-dimensional and finite sequences [
], where
N represents the length of the sequence. First, the matching window length,
L, is selected and the raw time series is hysterically processed to construct the trajectory matrix,
D. In most cases, the L value chosen is less than
N/2, where
N is the length of the sequence.
K is defined as
N −
L + 1, and the resulting trajectory matrix,
X, is a matrix of L rows and K columns.
(2) Singular Value Decomposition (SVD). The decomposition, , is of the form , where is called the right matrix, is called the left matrix, and non-zero values occurring only on the main diagonal are the singular values, and the rest of the elements have a value of 0. In addition, both and are unit orthogonal matrices, satisfying . It is impractical to decompose the trajectory matrix directly, so it is first necessary to calculate the covariance matrix of the trajectory matrix, which can be obtained by the formula . Then, the eigenvalue decomposition of the covariance matrix is carried out to obtain a series of eigenvalues and corresponding eigenvectors . In this step, , and for the singular spectrum of the original sequence. In addition, the trajectory matrix may be expressed by the following formula: . Here, the eigenvector corresponding to each is called the time-empirical orthogonal function, which reveals patterns of change in time series data.
(3) Grouping. The grouping step involves dividing the set of fundamental matrices into cut into m disjoint subsets, . The singular value decomposition of can be expressed as a combination of .
(4) Singular Entropy Calculation. The singular entropy is then calculated using the Shannon entropy formula applied to the normalized singular values:
where
is the normalized singular value, and
is the rank of the trajectory matrix.
(5) Reconstruction (diagonal averaging). First, the hysteresis sequence
is calculated in
projection:
. Considering
, the
th column of the trajectory matrix
is associated with the time evolution type, and
represents the weight of
in the original sequence:
. These weights are also called time principal components (TPCs). In short, the matrix consists of
, which is actually the right matrix without normalization, which is
. Then, reconstruction is performed by using the orthogonal function of time experience and the principal component of time [
16], as shown in Formula (2).
In this way, the sum of all reconstructed sequences is equal to the original sequence, i.e., .
3.4. LSTM
In 1997, Hochreiter and Schmidhuber first proposed the LSTM model. LSTM is a special kind of RNN, which is carefully designed. The original RNN training process, due to the extended training time and the increase in the number of network layers, is likely to have problems such as gradient bursting and gradient vanishing, which make it difficult to efficiently process very long data processing. To address this problem, LSTM adds the RNN-based memory feature, which can maintain the long-term memory of the neural network [
17]. The LSTM model structure is shown in
Figure 4.
LSTM neurons possess three control gates, the forgetting gate (), the input gate (), and the output gate (), with subscripts denoting moments.
Forgetting gate (
): Taking
and
as inputs, a value between 0 and 1 is output, which is used to determine how much to retain the cellular state
of the previous step, where 1 means completely retained and 0 means completely discarded. The details are shown in Formula (3) [
18].
Input gate (
): First, the input gate of the sigmoid function is utilized to filter out the information that needs to be updated immediately. Then, a vector is generated at the tanh layer that determines how much information from the network input
at the current time step can be saved to the current cell state
. Finally, these two parts are combined to update the information in the current cell state
[
19]. The details are shown in Formulas (4)–(6).
Output gate (
): First, the sigmoid layer acts as an output gate determining how much information about the cell state
at the current moment can be retained into the current hidden state
. Then, the cell state is processed by the tanh layer, and the final output is the result of multiplying these two parts [
20]. The details are shown in Formulas (7) and (8).
3.5. Error Analysis Index
In this paper, the root mean square error (RMSE), mean absolute error (MAE), and R2 are used to assess the error in the prediction results.
The RMSE is the ratio of the square root of the sum of the squares of the deviations of the predicted values from the true values to the sample size, as shown in Equation (9).
The MAE is the average of the absolute errors between the predicted and observed values, as shown in Equation (10).
R2 is calculated as
, where
SSE is the residual sum of squares and SST is the total sum of squares. If the model is well fitted, then the
SSE will be small and the
R2 value will be close to 1. If the model is poorly fitted, then the
SSE will be large and the
R2 value will be close to zero.
where
and
are the true and predicted values, respectively, and
is the number of samples.