Next Article in Journal
Temperature Dependence of the Thermo-Optic Coefficient of GeO2-Doped Silica Glass Fiber
Next Article in Special Issue
Coupling Fault Diagnosis of Bearings Based on Hypergraph Neural Network
Previous Article in Journal
YOLOv7-Branch: A Jujube Leaf Branch Detection Model for Agricultural Robot
Previous Article in Special Issue
Coupling Fault Diagnosis Based on Dynamic Vertex Interpretable Graph Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GCN-Based LSTM Autoencoder with Self-Attention for Bearing Fault Diagnosis

1
Department of Electrical and Computer Engineering, Sungkyunkwan University, 2066 Seobu-ro, Jangan-gu, Suwon-si 16419, Gyeonggi-do, Republic of Korea
2
Department of Smart Factory Convergence, Sungkyunkwan University, 2066 Seobu-ro, Jangan-gu, Suwon-si 16419, Gyeonggi-do, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(15), 4855; https://doi.org/10.3390/s24154855
Submission received: 4 July 2024 / Revised: 21 July 2024 / Accepted: 24 July 2024 / Published: 26 July 2024
(This article belongs to the Special Issue Fault Diagnosis and Vibration Signal Processing in Rotor Systems)

Abstract

:
The manufacturing industry has been operating within a constantly evolving technological environment, underscoring the importance of maintaining the efficiency and reliability of manufacturing processes. Motor-related failures, especially bearing defects, are common and serious issues in manufacturing processes. Bearings provide accurate and smooth movements and play essential roles in mechanical equipment with shafts. Given their importance, bearing failure diagnosis has been extensively studied. However, the imbalance in failure data and the complexity of time series data make diagnosis challenging. Conventional AI models (convolutional neural networks (CNNs), long short-term memory (LSTM), support vector machine (SVM), and extreme gradient boosting (XGBoost)) face limitations in diagnosing such failures. To address this problem, this paper proposes a bearing failure diagnosis model using a graph convolution network (GCN)-based LSTM autoencoder with self-attention. The model was trained on data extracted from the Case Western Reserve University (CWRU) dataset and a fault simulator testbed. The proposed model achieved 97.3% accuracy on the CWRU dataset and 99.9% accuracy on the fault simulator dataset.

1. Introduction

At present, the manufacturing industry operates within a constantly changing and developing technology environment, and it is very important to ensure that any manufacturing processes in this environment are efficient and reliable. Rapidly detecting and responding to various failures that may occur in the manufacturing process can directly contribute to improving productivity and reducing costs. Motor-related failures occupy a large proportion of manufacturing processes, and bearing faults [1] represent one of the most common and fatal problems.
Bearings are mechanical elements that provide accurate and smooth movements, and they play an important role in mechanical equipment with shafts. Because of this importance, many studies have focused on diagnosing bearing failure [2,3,4]. However, due to the imbalance of abnormal data and the complexity of time series data, it remains difficult to diagnose bearing failures in manufacturing. This has led to the use of existing AI models such as the convolutional neural network (CNN) [5], long short-term memory (LSTM) [6], support vector machine (SVM) [7], and extreme gradient boosting (XGBoost) [8] to limit fault diagnosis.
The methods that are currently used to maintain machines in the manufacturing industry [9] can be largely divided into post-maintenance, preventive maintenance, and predictive maintenance. Post-maintenance is a method of taking measures after the machine breaks down, and it has enormous costs in terms of productivity and finances. Preventive maintenance is a method of performing maintenance at a set period, although it has the disadvantage of incurring regular maintenance costs. Pre-maintenance predicts and measures machine failure to prevent its occurrence, primarily by monitoring equipment conditions using various sensors such as vibration, temperature, and camera sensors [10], and it detects abnormal signs in advance, allowing for maintenance or repair. Pre-maintenance can minimize production interruptions or cost increases stemming from unexpected failures.
However, several problems must be solved before prediction and maintenance can be effectively implemented. The first is the imbalance between normal and abnormal data. In actual manufacturing sites, there is generally much less abnormal data than normal data, which makes it difficult to learn the model. The current paper solved this problem by generating various bearing fault data using a directly constructed fault simulator testbed. The second is the accuracy problem that artificial intelligence models struggle with. To diagnose bearing faults with high accuracy, the graph convolutional network (GCN)-based self-attention LSTM autoencoder model was proposed. An experiment using this model on the Case Western Reserve University (CWRU) open dataset [11] showed 97.3% accuracy. Moreover, the accuracy of the proposed model was 99.9% when tested on the dataset extracted from the directly constructed fault simulator testbed. These results verified the reliability and excellence of the proposed model.
The major contributions of this paper can be summarized as follows:
  • A fault simulator testbed was constructed to extract and utilize various bearing data.
  • A bearing fault diagnosis model that can be used in various situations was developed by using the CWRU dataset, which is an open dataset, along with directly extracted data.
  • The proposed model achieves 97.3% accuracy using the CWRU dataset, which is an open dataset, while it achieves 99.9% accuracy using the data that have been extracted through the fault simulator testbed.
The rest of this paper is organized as follows: Section 1 contains an introduction to detail the research purpose and research methodology. Section 2 reviews related work and explains the time series data analysis techniques and various deep learning models used therein. Section 3 elucidates the bearing failure diagnosis method proposed in this paper. Section 4 presents the experimental environment and experimental results. Finally, Section 5 synthesizes the experimental results and discusses future research plans.

2. Related Work

2.1. Types of Failures in Rotating Machines

Rotating machines are used as core pieces of equipment in various industrial fields, and their efficient operation is one of the important factors that is directly connected to the productivity of a product. Rotating machines play an essential role in converting energy into mechanical power, and they are used in various industrial environments such as turbines, generators, fans [12], and motors [13]. Failure of a rotating machine is a factor that hinders the growth of the industry, as it causes economic losses to the enterprise due to problems such as serious productivity degradation, increased maintenance costs, and equipment shutdown [14].
There has recently been a significant increase in researchers [15] examining the diagnosis of failures in rotating machines using artificial intelligence. Such research focuses on the development of artificial intelligence models that can systematically analyze various types of failures that may occur in rotating machines such as rotors [16], gears [17], shafts [18], bearings [19], stators [20], belt pulleys [21], seals [22], etc. This makes it possible to quickly diagnose and predict failures. These models enable the early detection of failures, precise analysis of their causes, and efficient response strategies—all of which contribute to prolonging the machine’s lifespan and minimizing maintenance costs. Figure 1 shows the various types of failures that can occur in rotating machines; these types of failures are highly diverse, and their characteristics vary depending on the unique role and operating environment of each component. For example, faults in bearings are mainly caused by inner races, outer races [23], and damage to carriages [24], and these failures lead to poor overall performance of the machine. Meanwhile, imbalances or cracks in the rotor increase the vibration and noise of the system, thereby lowering the efficiency and safety of the machine. This paper aims to precisely analyze the types of faults in bearings, particularly in rotating machines, and develop an effective fault diagnosis model.

2.2. Bearing Fault Diagnosis

Bearings are some of the key components of a rotating machine, and the failure of a bearing can significantly affect the performance of the entire machine. It is essential to diagnose bearing faults to detect these failures early, and it is also crucial to accordingly perform proper maintenance to improve the reliability and lifespan of the machine. There are various causes of bearing faults, such as misalignment [25], inner races, outer races, high-speed damages [26], damage to carriages, foreign body insertion, lubrication, and oxidation [27]. Early detection of the causes of these various faults is a very important part of ensuring continuous operation of the machine and maintenance of productivity. Various methods can be used to diagnose bearing faults, and typical methods include vibration analysis [28], sound analysis [29], and temperature monitoring [30].
Vibration analysis is the most commonly used method to monitor the state of a bearing. Vibration analysis involves evaluating the state of a bearing by analysis conducted in the frequency domain, and it captures changes occurring in vibration signals according to major causes of failure. The primarily used techniques include fast Fourier transform (FFT) [31], short-time Fourier transform (STFT) [32], and wavelet transform (WT) [33]. These methods can be used to diagnose failure by analyzing frequency components, and they can be used to accurately determine the normal and abnormal states of bearings while considering data continuity and periodicity.
Sound analysis is a method in which failures are diagnosed by analyzing the acoustic signals generated from a bearing. This is achieved utilizing noise analysis and ultrasonic analysis using an acoustic sensor, and abnormal noise such as abrasion or cracks generated inside the bearing can be detected through this process. In particular, acoustic analysis is more effective when it is used together with vibration analysis.

2.3. Deep Learning Techniques for Bearing Fault Diagnosis

Bearing fault diagnosis—which uses deep learning, a technique in the field of machine learning—is a method of monitoring the condition of a bearing and identifying a failure by analyzing large amounts of data obtained from a sensor. This method plays an important role in detecting and preventing problems before a failure occurs. Throughout the data collection process, sensors collect various physical signals that are generated by the bearing, such as vibration, temperature, and sound. The data collected in this way are used as basic data for bearing fault diagnosis and play an important role in the next step (the data pre-processing process).
Next, the data that have been collected from the sensor are converted into forms that are suitable for analysis through a pre-processing process before being used for further analysis. The process includes noise removal [34], normalization [35], and feature extraction in the time-frequency domain. The characteristics of the data that have undergone such pre-processing are reinforced, and unnecessary information is removed, thus increasing the accuracy of model learning. After that, the deep learning model is designed and trained. Complex data patterns are learned using various deep learning models, such as CNN, LSTM, and autoencoder. These models learn the features that are necessary to distinguish between normal and fault states in a large amount of data and to diagnose the state of the bearing based on this. The learned model can be applied to new sensor data to detect anomalies and predict the type of faults in the bearing.
Recent studies have utilized various deep learning techniques to improve the performance of bearing fault diagnosis. Z. Dong et al. [36] proposed a bearing fault diagnosis method by combining empirical wavelet transform (EWT) and one-dimensional improved self-attention-enhanced convolutional neural network (1D-ISACNN). The method was experimented on three different bearing test benches and achieved 100% accuracy in all cases. This study combined feature extraction in the frequency domain with a self-attention mechanism, which greatly improved the accuracy of bearing failure diagnosis. W. Li et al. [37] combined bidirectional long short-term memory (BiLSTM) and CNN to perform the tool residual life prediction. The model can learn bidirectional information from time-ordered data, and can accurately predict the remaining life by monitoring the status of the tool in real time. This study shows that combining the strengths of LSTM and CNN can effectively learn patterns in complex time series data.
In this paper, we propose a GCN-based LSTM autoencoder model utilizing self-attention. Unlike previous studies, the proposed model utilizes STFT instead of the wavelet transform, uses graph convolutional network (GCN) to extract spatial and structural features of the data, and combines the LSTM autoencoder with a self-attention mechanism to learn important time series patterns. This approach is advantageous for effectively detecting and diagnosing complex bearing fault patterns. The following deep learning models are utilized for bearing fault diagnosis.

2.3.1. LSTM Autoencoder

The LSTM autoencoder [38] is a neural network model that is designed to effectively encode and decode the features of time series data or sequential data. This model combines the time-dependent capture capability of LSTM with the data compression and reconstruction capabilities of the autoencoder, as shown in Figure 2. The LSTM autoencoder consists of an encoder and a decoder, which converts the input data into latent representations, and the decoder restores this latent representation to the original data. This allows the model to learn the important characteristics of the input data and minimize reconstruction errors.
This model is used for tasks such as anomaly detection, data compression, and sequence regeneration. In particular, the LSTM autoencoder has excellent performance in various applications, as it can effectively extract and restore important features of time series data. For example, in the anomaly detection task, the LSTM autoencoder can detect abnormal signs based on the difference between reconstruction errors between normal and abnormal data. In the data compression task, the data can be efficiently compressed while maintaining important information on the input data, and in the sequence regeneration task, the sequence of the original data can be accurately restored. These advantages have allowed the LSTM autoencoder to be effectively utilized in various fields, including financial data analysis, health monitoring, and predictive maintenance. Its ability to learn complex patterns of time series data and extract important features has also allowed this model to perform well in applications that require high degrees of both reliability and accuracy.

2.3.2. Attention Mechanism

The attention mechanism technique [39] is an important technology used for processing sequence data in deep learning; it was first used in natural language processing (NLP) and has since come to be widely used in various fields. This mechanism helps the model learn important information better by weighting specific parts of the input sequence. There are forms such as self-attention [40], global attention, local attention, etc. In self-attention, all elements within the input sequence interact to capture important patterns, which are implemented as scaled dot-product attention and extended to multi-head attention. Global attention considers the entire sequence, while local attention only considers a few sections to increase computational efficiency.
The attention mechanism is effectively used in the Transformer model [41]. The Transformer captures important information about the input sequence through self-attention, and it consists of an encoder and decoder structure. This structure extracts information from various perspectives and generates the final output. The advantages of the attention mechanism are that it focuses on the important parts when processing long sequence data, thereby increasing computational efficiency, which ultimately allows it to be easily integrated into various model structures, and to effectively learn important patterns of the input sequence.
In conclusion, the attention mechanism provides excellent performance for sequence data processing in deep learning, and it is used in various fields such as natural language processing, speech recognition, and time series data analysis. The use of this mechanism improves the performance of the model and enables more effective learning of important information.

2.3.3. GCN

A graph is a data structure that is composed of nodes and edges, where the nodes represent entities and the edges represent relationships between nodes. Graphs are useful for representing various types of data, and they are particularly effective in modeling correlations between data points. The K-nearest neighbor (K-NN) [42] algorithm is a method that can be used to find and connect K closest neighbors for each data point, and ultimately construct a graph based on the similarity between data points. This makes it possible to effectively model the spatial and structural relationship of data. Graph composition using K-NN is particularly advantageous for visualizing and analyzing correlations between unstructured data or time series data.
GCN [43] is a model that specializes in learning the node characteristics of graph data, and it updates the characteristics of each node by combining them with the characteristics of that node’s neighboring nodes. GCN utilizes the structural information of the graph to enhance the representation of nodes, through which it learns the patterns and interactions in the graph. The main idea of GCN is to define the convolution operation in the structure of the graph. This convolution operation is similar to the convolution operation that is used in image processing, but it reflects the structural characteristics of the graph. The formula for the GCN layer is as follows:
H ( l + 1 ) = σ D ˜ 1 / 2 A ˜ D ˜ 1 / 2 H l W l
Here, H l is the node characteristic matrix of layer l. A ˜ = A + I is the adjacent matrix of the graph, which includes connections to the node itself. D ˜ is the diagonal matrix of A ˜ . W l is the learnable weight matrix of layer l. σ is an activation function (e.g., rectified linear unit (ReLU)).
The GCN updates the node characteristics in each layer by combining them with the characteristics of its neighboring nodes. In the first layer, the initial node characteristics are accepted as input, while in the subsequent layer, the node characteristics are gradually updated using the output of the previous layer as input. Through this process, the global structure information of the graph is gradually learned.
Since GCN operates on graph data, if the input data are not in the form of a graph, it is necessary to first convert them into a graph; the K-NN algorithm can be used for this process. For each data point, K-NN is used to find K neighbors, and each data point is set as a node of the graph, while the connection of neighbors found through K-NN is set as an edge of the graph.
The graph generation process using K-NN is as follows: First, the distance between each data point is calculated. In general, the Euclidean distance or cosine similarity is used. Then, for each data point, the nearest K neighbors are found. Finally, a graph is generated by setting each data point as a node and neighbors found through K-NN as edges. The graph and node characteristic matrices generated in this way are input to the GCN model, while the characteristics of the node are updated and learned through the GCN layer.
GCN can be used in a variety of applications. For example, it can be used to analyze the relationship between users in social networks, to predict the properties of compounds by analyzing molecular structures, or model the interaction between users and items in recommendation systems. These properties of GCN ultimately make it possible to perform pattern recognition and prediction in graph data in a more efficient manner.

2.4. STFT

STFT is a technology that has been developed to overcome the limitations of FFT, and it involves dividing a signal into short time intervals and then applying the Fourier transform to each interval. This method can obtain the frequency components of the signal in each time interval, and by arranging them along the time axis, it is possible to observe the frequency change over time. STFT divides the signal into several short intervals, which generally have overlapping parts to maintain the continuity of the signal. Figure 3 shows the STFT transformation. Since it can select and analyze time intervals of various lengths, it has the advantage that it can be applied in a flexible manner according to the characteristics of the signal to be analyzed. STFT is very useful when the characteristics of the signal change over time, as is the case in real-time signal processing or music and audio signal analysis. However, depending on how the length of the interval is set, there may be a trade-off between the time resolution and the frequency resolution.

3. GCN-Based LSTM Autoencoder with Self-Attention Model

This chapter proposes a GCN-based LSTM autoencoder with self-attention model to bear fault diagnosis using multivariate time series data. The proposed model consists of two steps, as shown in Figure 4. This process can successfully diagnose the failure of bearing data.

3.1. Step 1

Figure 5 shows step 1, which visually represents the data pre-processing step. In this step, data including various fault states and steady states are loaded, and standardization and normalization are performed as well. Based on various types of bearing equipment, as shown in Figure 6, the data extracted from the three-channel vibration sensor and the sound sensor having the X, Y, and Z axes are normalized to make it suitable for analysis.
Normalized data are divided into certain sizes using the sliding window technique. The sliding window technique divides time series data into fixed-length windows and uses each window as one sample, while the fixed window size and slide are adjusted to divide the data, thus allowing the model to learn data patterns in various time intervals.
The K-NN algorithm is then used to find the k nearest neighbors for each data point, and a graph is generated based on the similarity between the data points. This process makes it possible to effectively model the spatial and structural relationship of the data. The generated graph is learned through the GCN layer.
GCN is a neural network structure that is optimized for processing graph data, and it extracts new characteristics by considering the characteristics of each node and its neighboring nodes together. This makes it possible to learn complex patterns and structural information of data and to effectively extract important features from the frequency domain of time series data. The main advantage of GCN is that it can be learned while maintaining the relationship between each data point by utilizing the graph structure, which makes a substantial contribution to increasing the accuracy of fault diagnosis. The frequency domain data obtained through this process are then transferred to step 2.

3.2. Step 2

Step 2 is shown in Figure 7, which represents the model experiment stage. It begins with the process of converting the features extracted through GCN, the last step of step 1, into STFT for analysis in the frequency domain. In this step, the data are analyzed by converting the temporal characteristics of the time series data into a frequency domain. STFT can analyze the frequency change pattern over time by dividing the time series data into short time intervals and then performing the Fourier transform for each section. Using this process, the spatial and structural characteristics obtained through GCN can be expressed in terms of changes in time and frequency. GCNs are used to extract spatial and structural characteristics of time series data. The network learns the features of each node through its relationship with its neighbors, which is advantageous for identifying multidimensional patterns in the data. GCNs are powerful tools that reflect the inertia and interactions of data based on their graph structure, which is especially useful for complex network structures.
The LSTM autoencoder with a self-attention layer simultaneously learns the long-term and short-term dependencies of the time series data. In this model, self-attention plays a role in assigning weights to important parts of the encoder’s output sequence [44]. The LSTM encoder creates a hidden state of each sequence while processing time series data, and it applies self-attention to assign weights to important parts. The self-attention mechanism calculates the importance of all time steps that are different from itself for each time step in the sequence, and it then weights them based on these calculations to emphasize important information in order of relative importance. These weights are determined through learning, and they are gradually adjusted while the model learns. The self-attention mechanism applies a probability function to estimate a score from the bottleneck vector representation. The estimated score is multiplied by the bottleneck vector representation to obtain a context vector. The context vector is then input to the decoder and reconstructed into the original encoder input data dimension. The computation method uses M. Luong et al.’s global attention concatenation multiplicative technique [45], which allows the model to emphasize important information and learn important features of time series data more effectively. The self-attention layer is defined in the following way:
e i = tanh ( W · x i + b )
a i = exp ( e i ) j exp ( e j )
o i = j a j · x j
In the above equation, x i represents a vector corresponding to a time step i of an input sequence, while W and b are learnable weights and biases. e i is the score representing the importance of each time step, and a i is the weight that is normalized through the softmax function. Finally, o i is the sum of weighted input vectors, which represents an output vector emphasizing important information. By defining the self-attention layer in this way, an abnormal state may be detected through the process of assigning weights to important portions of the output sequence of the encoder and transferring them to the decoder to reconstruct the original data.
The important features weighted by self-attention are converted into repeat vectors and passed to the decoder. The decoder is responsible for reconstructing the original data based on the features extracted from the encoder. The decoder uses the last state of the encoder as the initial state to generate a sequence, which is used to calculate the reconstruction error and detect anomalies. The main role of the decoder is to restore the original shape of the input data based on the information learned from the encoder.
The output layer of the model is responsible for the classification of failure types. The output layer consists of a fully connected (dense) layer, which uses CrossEntropyLoss as the loss function to train the model. For multi-class classification, the LeakyReLU and Softmax activation functions are used to predict the probability for each type. This allows the model to distinguish between normal and abnormal data and accurately classify different defect types. While many previous studies mainly used the ReLU activation function, the proposed model uses the LeakyReLU activation function to perform multi-class classification. To enable the model to learn complex patterns, we set LeakyReLU in the hidden layer to add nonlinearity. LeakyReLU solves the dead ReLU problem by maintaining a small gradient when the output of a particular node is negative. This makes learning more stable. Softmax is set on the output layer of a multi-class classification to convert the predicted values for each class into a probability distribution by making the output value equal to 1. The Softmax activation function calculates the log-likelihood of each class, converts it to an exponential function, and normalizes the sum to 1 to generate the final probability distribution. This process allows the model to learn the features of the input data and accurately predict the probability for each class. Finally, the CrossEntropy Loss function is used to train the model and optimize the classification performance.
Through the feature learning and self-attention mechanism of GCN, the important parts of the time series data can be highlighted, which greatly improves the accuracy of fault diagnosis. The proposed model learns the patterns of normal and abnormal data and detects anomalies in the test data through reconstruction errors to provide an early diagnosis of bearing failure in multivariate time series data. This model has the advantage of achieving significantly improved fault diagnosis accuracy by highlighting the important parts of time series data through graph-based spatial characteristic learning and attention mechanisms. This helps effectively extract and learn important information, particularly from complex time series data.
The proposed model also evaluates the performance by comparing the reconstruction loss with mean squared error (MSE) and mean absolute error (MAE). MSE is a value that is averaged by squaring the difference between predicted and actual values, and it gives a greater penalty to the large error. MAE is the average of the absolute difference between the predicted and actual values, and it deals with all errors equally. By comparing the two indicators, the reconstruction capability and fault detection performance of the model can be evaluated in a more detailed manner. Through this process, the reliability and superiority of the proposed model are verified.

4. Experiment and Results

4.1. Experimental Setup and Datasets

Table 1 presents the experimental environment. The experiment described in this paper was conducted on a Windows workstation with an Intel Core i7-8700K CPU and an NVIDIA GeForce RTX 3080 GPU. Python version 3.7.7, TensorFlow 2.11.0, and CUDA version 11.1 were used.

4.1.1. CWRU Dataset

Data provided by CWRU were used to verify the effectiveness of the proposed model. The CWRU dataset is a dataset that is widely used for bearing fault diagnosis and consists of a steady state, internal race failure, external race failure, and ball failure. The data were measured at locations such as the drive end, fan end, and base, and the depth of the defect varies between 0.007 inches, 0.014 inches, and 0.021 inches. The CWRU dataset consists of 12 kHz and 48 kHz drive-end-bearing datasets and 12 kHz fan-end-bearing data; this paper used drive-end data that were collected at 48 kHz. Specifically, data that were collected when the load of the motor was 2 horsepower (HP) was used. At this time, the shaft rotation speed of the motor was 1750 revolutions per minute (RPM).
The data used consist of 10 types: normal state, three types of ball defects (0.007 inches, 0.014 inches, and 0.021 inches), three types of internal lace failures (0.007 inches, 0.014 inches, and 0.021 inches), and three types of external lace failures (0.007 inches, 0.014 inches, and 0.021 inches) in the 6 o’clock position. This dataset contains bearing fault data under various conditions, which makes it suitable for evaluating the generalization performance of fault diagnosis models. The CWRU testbed is depicted in Figure 8.
The data were processed, made into a comma-separated variable (CSV) file format, and used. For failure identification prediction, nine feature values were calculated: maximum value, minimum value, average value, standard deviation, root mean square (RMS), skewness, kurtosis, crest factor, and shape factor. Each feature value was calculated in a time interval of 2048 points. These feature values were calculated for each segment and ultimately stored in a CSV file format.
Table 2 lists information on the dataset used in this experiment. Classifications for 10 types were diagnosed, and the train, validation, and test datasets were used at a 6:2:2 ratio for each type, respectively, while the data type numbers after STFT conversion were 340, 110, and 110, respectively. The hyperparameters used in this experiment are presented in Table 3.

4.1.2. Fault Simulator Dataset

In this paper, the fault simulator test bed shown in Figure 9 was constructed to collect data for bearing fault diagnosis. This test bed was designed to simulate various fault conditions and normal conditions. The message queuing telemetry transport (MQTT) protocol and data acquisition (DAQ) system shown in Figure 10 were used for data collection.
The test bed consists of equipment that can simulate various failure conditions that might affect bearings. This test bed can reproduce both the normal state and the failure state by replacing bearings, and it can collect vibration and sound data that is generated in each state in real time. To construct the test bed, three-axis vibration data were collected using a vibration sensor called AC115 from the Connection Technology Center company in Figure 11. Sound data were also collected using MP41, a microphone sensor from the Korean SI company, as shown in Figure 12. This sensor is capable of measuring high-precision sound, and it is used in the present work to measure the sound that is generated during bearing operation.
The data collected through the vibration and sound sensors were transmitted to the central server in real time through the MQTT protocol; MQTT is a lightweight message protocol that can efficiently transmit data in a limited environment such as a sensor network. Data transmission was performed by posting data from each sensor to the MQTT broker and receiving it from a central server. The DAQ system played a role in converting analog signals into digital data and transmitting it to a computer. Vibration and sound data were collected at a high speed through the DAQ system, while the data were collected at a sampling rate of 5000 Hz for 10 min, and the vibration and sound sensor values were stored in the CSV form with labels such as motorx, motory, motorz, sound, and time.
In this paper, bearing data were collected for 10 min by setting the rotational speed of the motor to 1500 RPM. To accurately grasp the defective sound of the bearing equipment, a sound sensor was attached to the bearing equipment and used at a position of 20 cm; when extracting the data, the data were extracted in an environment without noise other than the sound generated by the bearing. The collected bearing data were divided into five types: a normal state, high-speed damage, carriage damage, lack of lubricant, and oxidized state. Normal state data indicate a situation in which the bearing operates smoothly without failure, while high-speed damage was conducted with a bearing that has failed by rotating at high speed. Carriage damage was a state in which internal components of the bearing were damaged, and lack of lubricant was simulated as a state in which abrasion and heat were generated due to the absence of lubricant in the bearing. Finally, oxidized and corroded bearings were used in the oxidized state. Table 4 presents detailed information on the dataset used in this experiment. Each data type was used for training, validation, and testing in a respective ratio of 6:2:2, and the data numbers after STFT conversion are as follows:
The collected data were classified by each state, and they were treated using a pre-processing process. Sensor values (motorx, motory, motorz, and sound) were standardized and normalized, which had the effect of adjusting the scale of the data and improving the accuracy of the analysis. Time series data were divided into fixed-length windows through sliding windows, each window was used as one sample, and the data were divided by adjusting the window size and the slide. This process allowed the model to learn data patterns in various time intervals. The data were divided into window-extracted features in the frequency domain through STFT conversion, while the frequency domain data obtained through this process was used as the input of the model.
Table 5 presents the hyperparameters of the GCN-based LSTM autoencoder with a self-attention model. Various hyperparameters were tested using grid search, and the following settings showed the best performance. The hyperparameter values of the model are as follows.

4.2. Evaluation Metrics and Visualization

4.2.1. Evaluation Factors

This paper used various indicators to evaluate the performance of the model in terms of bearing fault diagnosis. In particular, the prediction performance of the model was confirmed in an intuitive manner using the confusion matrix. The confusion matrix consists of the following four elements: true positive ( T P ) means the case where the actual value is true and is correctly predicted as true by the model. True negative ( T N ) means the case where the actual value is false and is correctly predicted as false by the model. False positive ( F P ) means the case where the actual value is false but the model incorrectly predicts it as true. Finally, false negative ( F N ) is the case where the actual value is false but the model incorrectly predicts it as true. Based on these evaluation factors, various important evaluation indicators may be calculated. First, accuracy represents the percentage of samples correctly predicted by the model. It is calculated as the ratio of samples accurately predicted by the model to the total number of samples, and it is expressed using the following formula:
Accuracy = T P + T N T P + F N + F P + T N
Recall refers to the percentage of true positive values that the model correctly identifies as such. Reproducibility refers to how well the model finds the case where it is actually positive. High reproducibility means that the model does not miss the case where it is true; it is very important to increase reproducibility for use in certain applications. Recall is calculated using the following equation:
Recall = T P T P + F N
Precision refers to the ratio of predicted values that are correctly identified as true. High precision means that there are many cases where the model is actually true among the predictions that have been made. Reproducibility and precision exist in a trade-off relationship, and increasing one can lower the other. Precision indicates how accurate the model’s prediction is, and it is expressed by the following formula:
Precision = T P T P + F P
The F1 score is a harmonized average of precision and reproducibility, and if the data label is unbalanced, the performance of the model can be evaluated more accurately. This helps reliably evaluate the predictive performance of the model on an unbalanced dataset. The F1 score is calculated using the following formula:
F 1 - Score = 2 × Precision × Recall Precision + Recall

4.2.2. Reconstruction Loss

In this paper, we also evaluate the reconstruction loss of the autoencoder using MSE and MAE. MSE is the squared average of the difference between the predicted and actual values. Since the error is squared, it is more sensitive to large errors. The formula for MSE is as follows, where n is the total number of samples, y i is the actual value, and y i ^ is the predicted value.
M S E = 1 n i = 1 n ( y i y i ^ ) 2
MAE is the average of the absolute values of the differences between the predicted and actual values. It gives equal weight to all errors, and it is calculated using the following formula, where n is the total number of samples, y i is the actual value, and y i ^ is the predicted value.
M A E = 1 n i = 1 n y i y i ^
These metrics are used to evaluate how well a model’s predictions match the true value, and each metric can help analyze a model’s performance in different ways according to its particular characteristics. MSE is more sensitive to large errors, which is useful when it is important to minimize large errors, while MAE treats all errors equally, which is useful when the aim is to evaluate overall prediction accuracy.

4.2.3. PCA

High-dimensional data were visualized in two and three dimensions using principal component analysis (PCA). Through this, the embedding vector of the data learned by the model can be visually checked and the data distribution for each class can be identified. PCA is a technique that is used to visualize the structure of data by converting high-dimensional data into a low-dimensional space. The two-dimensional PCA visualization is shown in Figure 13, and the data distribution for each class can be visually checked in a two-dimensional space. This allows one to see how well the model separates each class.
The 3D PCA visualization is shown in Figure 14, and one can visually check the data distribution for each class in a 3D space. This makes it easier to grasp the structure of the data and the clustering performance of the model.
Through this PCA visualization, it is possible to analyze the embedding vector of the data learned by the model, and the performance of the model may be evaluated from various angles. Through this, the performance of the bearing fault diagnosis model may be comprehensively evaluated and improved.

4.3. Results

In this paper, performance comparisons with various comparative models were conducted to evaluate the performance of the proposed GCN-based LSTM autoencoder with a self-attention model. Accuracy and the F1 score were used as evaluation indicators, and the performance of each model was compared. As comparison models, LSTM, ConvLSTM, and LSTM autoencoder were used.
The performance of the proposed GCN-based LSTM autoencoder with a self-attention model was compared with other models using CWRU data. Table 6 presents the performance of the proposed model and other models. Figure 15 shows the confusion matrix for the suggested model.
The performances of the proposed model and several models were compared using data that was extracted from the directly constructed fault simulator testbed. Table 7 presents the performance of each model. The proposed model also showed high performance in directly extracted data, particularly when GCN and STFT were used together. Figure 16 shows the confusion matrix for the proposed model. As can be seen from the confusion matrix, the proposed model accurately classified most fault states and normal states.
The reconstruction loss of the autoencoder was evaluated using MSE and MAE according to the K-neighbor value. Both indicators play an important role in evaluating the reconstruction performance of the model, and the results are shown in Figure 17 and Figure 18. Comparing MSE and MAE for each K-neighbor value, the minimum F1 score splits the resulting value of 0.864 and performs better with MAE than with MSE. It performs best when the K-neighbor is 6.
Moreover, in this paper, high-dimensional data were visualized in both two and three dimensions using PCA. The two-dimensional PCA visualization is shown in Figure 19, and the data distribution for each class can be visually confirmed in a two-dimensional space. This made it possible to confirm how well the model separates each class.
As shown in Figure 20, the 3D PCA visualization made it possible to visually check the data distribution for each class in a 3D space. This allowed for a clearer understanding of the data structure and the clustering performance of the model.
As a result of the experiment, the proposed GCN-based LSTM autoencoder with the self-attention model showed superior performance compared to other comparison models on the CWRU dataset, which is an open dataset, and the directly constructed fault simulator data. In particular, the proposed model exhibited high acuity and F1 score through reconstruction loss (MSE, MAE) evaluation and confusion matrix analysis. Moreover, by visually analyzing the embedding vector of the data learned by the model through 2D PCA visualization and 3D PCA visualization, the clustering performance of the model could be evaluated. This confirmed how effectively the proposed model separated the data of each class.
The proposed model recorded 99.9% accuracy in early diagnosis of bearing fault in multivariate time series data, thus outperforming the other models. In particular, compared to the accuracy of 97.3% achieved on the CWRU dataset, it has shown the potential to increase the applicability in real industrial sites by maintaining high performance in fault simulator data that is collected in real industrial environments.
These results may contribute to improving the reliability and lifespan of the machine. The proposed model greatly improves the accuracy and reliability of bearing fault diagnosis, and these results are expected to play an important role in establishing an early failure diagnosis system in industrial sites.

5. Conclusions

In this paper, a GCN-based LSTM autoencoder with a self-attention model for bearing fault diagnosis was proposed and evaluated using multivariate time series data. The proposed model was found to increase the accuracy of fault diagnosis by combining the GCN layer and the LSTM layer to extract important features from the frequency domain. In the data pre-processing step, data including various fault states and steady states were standardized, while features in the frequency domain were extracted through STFT conversion. Moreover, time series data were divided into fixed-length windows using the sliding window technique, which was used as input for the model.
The GCN-based LSTM autoencoder with self-attention model using multivariate time series data performed fault diagnosis by synthesizing various data obtained from multiple sensors. This model learned the spatial and structural relationship of the data using the GCN layer, and it simultaneously learned both long-term and short-term dependencies through the LSTM layer. The self-attention mechanism was added between the encoder and the decoder to highlight important features and synthesize various features to ultimately classify several types of bearings.
The comparative model performance evaluation indicated that the GCN-based LSTM autoencoder with self-attention model showed superior performance over other models in accuracy and F1 score in both the open CWRU dataset and the directly constructed fault simulator dataset. In particular, it was possible to detect abnormal conditions more accurately through reconstruction errors, which is an important advantage in quickly detecting and responding to various failures that occur in the manufacturing process. As a result of visually checking the classification performance of the model through the confusion matrix, the proposed model accurately classified most fault states and normal states, while there was a small number of incorrectly classified samples.
These findings can make an important contribution to early diagnosis and response to various failures that may occur in manufacturing sites, and they can be expected to substantially contribute to increasing the practicality of the model.
To further enhance the performance of the GCN-based LSTM autoencoder with the self-attention model proposed in this study, future research is planned in the following directions: Firstly, the development of a real-time fault diagnosis system based on the proposed model is planned. This is expected to improve the model’s real-time inference capabilities through real-time data collection, processing, and online learning. The real-time fault diagnosis system can quickly detect and respond to failures occurring in the manufacturing process, which can substantially contribute to improving productivity and reducing costs.
Next, the experiment in the current research was only conducted for specific bearing fault types, but in future research, we will conduct a study to increase the versatility of the model while including various failure types other than bearings. Through this, we will develop a general-purpose fault diagnosis model that can be used in various industrial environments as well as build a fault diagnosis system that can exhibit high performance even in various situations. We will also apply additional data augmentation techniques to increase the diversity of data. The data augmentation technique that is selected for use plays an important role in improving the generalizability of the model and enhancing the fault diagnosis performance in various situations. This will enable the model to learn various data patterns and facilitate more accurate fault diagnosis.
The practical implications of the proposed model include failure prevention and maintenance cost reduction, productivity improvement, reliable data-driven decision-making, industrial applicability, and implementation of smart manufacturing systems. By monitoring machine health in real time and diagnosing faults at an early stage, unexpected machine failures can be prevented, maintenance costs can be reduced, and overall productivity can be improved by minimizing production downtime. Reliable decisions can be made based on the data provided by the fault diagnosis system, which is very helpful for machine maintenance planning, resource allocation, production schedule adjustment, etc. Furthermore, the proposed model is not limited to a specific industry but can be applied in various manufacturing sectors, which can be utilized as a bearing fault diagnosis and preventive maintenance system in various industries such as automotive, aviation, energy, marine, etc. In this way, the proposed model can contribute to optimizing manufacturing processes and achieving innovative productivity by leveraging IoT, big data, and AI, which are key elements of Industry 4.0. These practical implications show that the proposed model can serve as an efficient and reliable fault diagnosis solution in various fields of manufacturing in the future.
Also, we plan to introduce a transfer learning technique to apply the model learned from one dataset to another. Through this, we intend to develop a model that can maintain high performance in various environments and conditions. Using this method will have the following advantages:
  • Improved generalization: transfer learning allows the model to consistently perform highly on different datasets.
  • Reducing training time: Applying already trained models to other datasets can reduce training time.
  • Improved adaptability: allows the model to better adapt to new datasets or domains. This will be good for application to new manufacturing sites.
We plan to review practical applicability by applying the research results to actual industrial sites. Through such research, the performance of the model in the actual environment will be evaluated, and the model can be improved as necessary. This will increase the practicality of the model through verification in the actual industrial environment and contribute to solving problems in the actual field.
Finally, in addition to the GCN-based LSTM autoencoder with the self-attention model, we will conduct a fault diagnosis study using other recent deep learning models. We plan to compare the performance of various models and derive the optimal model. We expect that comparative studies with various deep learning models will make it possible to comprehensively evaluate the performance of the model and develop an optimal fault diagnosis model.
With these future research directions, we plan to further improve the performance of the models proposed in this study and eventually develop a practical fault diagnosis system that can be applied in various industrial environments. We believe that this will ultimately help increase the efficiency and reliability of the manufacturing industry, thereby contributing to productivity improvement and cost reductions.

Author Contributions

Conceptualization, D.L.; methodology, D.L.; software, D.L.; validation, D.L.; formal analysis, D.L.; investigation, D.L.; resources, D.L.; data curation, D.L.; writing—original draft preparation, D.L.; writing—review and editing, H.C. and J.J.; visualization, D.L.; supervision, H.C. and J.J.; project administration, D.L.; funding acquisition, D.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the SungKyunKwan University and the BK21 FOUR (Graduate School Innovation) funded by the Ministry of Education (MOE, Korea) and National Research Foundation of Korea (NRF).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the following repositories: The CWRU dataset is openly available at the Case Western Reserve University Bearing Data Center website at https://engineering.case.edu/bearingdatacenter (accessed on 1 June 2024). The dataset extracted by the authors is available on GitHub at https://github.com/daeheetwo/Fault-Simulator-Data (accessed on 1 June 2024).

Acknowledgments

The authors are grateful to the support from the SungKyunKwan University and the BK21 FOUR (Graduate School Innovation) funded by the Ministry of Education (MOE, Korea) and National Research Foundation of Korea (NRF). Moreover, this research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ICT Creative Consilience Program (IITP-2024-2020-0-01821) supervised by the IITP (Institute for Information & communications Technology Planning & Evaluation). And this research was supported by the Gyeonggi Techno Park grant funded by the Gyeonggi-Do government (No. D2323045).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, D.; Lee, J.; Park, J.; Choi, J.; Choe, T. Anomaly Detection in Rotating Motor using Two-level LSTM. In Proceedings of the KIIT Conference, Jeju Island, Republic of Korea, 25–27 November 2021; pp. 425–428. [Google Scholar]
  2. Jin, X.; Sun, Y.; Que, Z.; Wang, Y.; Chow, T.W. Anomaly detection and fault prognosis for bearings. IEEE Trans. Instrum. Meas. 2016, 65, 2046–2054. [Google Scholar] [CrossRef]
  3. Mian, T.; Choudhary, A.; Fatima, S.; Panigrahi, B.K. Artificial intelligence of things based approach for anomaly detection in rotating machines. Comput. Electr. Eng. 2023, 109, 108760. [Google Scholar] [CrossRef]
  4. Panza, M.A.; Pota, M.; Esposito, M. Anomaly Detection Methods for Industrial Applications: A Comparative Study. Electronics 2023, 12, 3971. [Google Scholar] [CrossRef]
  5. O’shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  6. Staudemeyer, R.C.; Morris, E.R. Understanding LSTM–A tutorial into long short-term memory recurrent neural networks. arXiv 2019, arXiv:1909.09586. [Google Scholar]
  7. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef]
  8. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  9. Arslankaya, S.; Atay, H. Maintenance management and lean manufacturing practices in a firm which produces dairy products. Procedia-Soc. Behav. Sci. 2015, 207, 214–224. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Chen, Y.; Wang, J.; Pan, Z. Unsupervised deep anomaly detection for multi-sensor time-series signals. IEEE Trans. Knowl. Data Eng. 2021, 35, 2118–2132. [Google Scholar] [CrossRef]
  11. Neupane, D.; Seok, J. Bearing fault detection and diagnosis using case western reserve university dataset with deep learning approaches: A review. IEEE Access 2020, 8, 93155–93178. [Google Scholar]
  12. Wang, M.; Zhou, D.; Chen, M.; Wang, Y. Anomaly detection in the fan system of a thermal power plant monitored by continuous and two-valued variables. Control Eng. Pract. 2020, 102, 104522. [Google Scholar]
  13. Lu, H.; Li, Y.; Mu, S.; Wang, D.; Kim, H.; Serikawa, S. Motor anomaly detection for unmanned aerial vehicles using reinforcement learning. IEEE Internet Things J. 2017, 5, 2315–2322. [Google Scholar] [CrossRef]
  14. Das, O.; Das, D.B.; Birant, D. Machine learning for fault analysis in rotating machinery: A comprehensive review. Heliyon 2023, 9, e17584. [Google Scholar] [CrossRef] [PubMed]
  15. Chalapathy, R.; Chawla, S. Deep learning for anomaly detection: A survey. arXiv 2019, arXiv:1901.03407. [Google Scholar]
  16. Ahmad, S.; Styp-Rekowski, K.; Nedelkoski, S.; Kao, O. Autoencoder-based condition monitoring and anomaly detection method for rotating machines. In Proceedings of the 2020 IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 4093–4102. [Google Scholar]
  17. Wang, W.; Galati, F.A.; Szibbo, D. Gear diagnostics based on LSTM anomaly detection. Int. J. Comadem 2021, 24, 3–14. [Google Scholar]
  18. Lee, S.; Park, J.W.; Kim, D.S.; Jeon, I.; Baek, D.C. Anomaly detection of tripod shafts using modified Mahalanobis distance. J. Mech. Sci. Technol. 2018, 32, 2473–2478. [Google Scholar] [CrossRef]
  19. Lee, D.; Choo, H.; Jeong, J. Anomaly Detection based on 1D-CNN-LSTM Auto-Encoder for Bearing Data. WSEAS Trans. Inf. Sci. Appl. 2023, 20, 1–6. [Google Scholar] [CrossRef]
  20. Samsi, R.; Ray, A.; Mayer, J. Early detection of stator voltage imbalance in three-phase induction motors. Electr. Power Syst. Res. 2009, 79, 239–245. [Google Scholar] [CrossRef]
  21. Pollak, A.; Temich, S.; Ptasiński, W.; Kucharczyk, J.; Gąsiorek, D. Prediction of belt drive faults in case of predictive maintenance in industry 4.0 platform. Appl. Sci. 2021, 11, 10307. [Google Scholar] [CrossRef]
  22. Bao, J.; Adcock, J.; Li, S.; Jiang, Y. Enhancing Quality Control of Chip Seal Construction through Machine Learning-Based Analysis of Surface Macrotexture Metrics. Lubricants 2023, 11, 409. [Google Scholar] [CrossRef]
  23. Zhang, F.; Huang, J.; Chu, F.; Cui, L. Mechanism and method for the full-scale quantitative diagnosis of ball bearings with an inner race fault. J. Sound Vib. 2020, 488, 115641. [Google Scholar] [CrossRef]
  24. Jírová, R.; Pešík, L.; Žul’ová, L.; Grega, R. Method of failure diagnostics to linear rolling guides in handling machines. Sensors 2023, 23, 3770. [Google Scholar] [CrossRef]
  25. Xu, H.; Yang, Y.; Ma, H.; Luo, Z.; Li, X.; Han, Q.; Wen, B. Vibration characteristics of bearing-rotor systems with inner ring dynamic misalignment. Int. J. Mech. Sci. 2022, 230, 107536. [Google Scholar] [CrossRef]
  26. Xu, G.; Hou, D.; Qi, H.; Bo, L. High-speed train wheel set bearing fault diagnosis and prognostics: A new prognostic model based on extendable useful life. Mech. Syst. Signal Process. 2021, 146, 107050. [Google Scholar] [CrossRef]
  27. Xu, F.; Ding, N.; Li, N.; Liu, L.; Hou, N.; Xu, N.; Guo, W.; Tian, L.; Xu, H.; Wu, C.M.L.; et al. A review of bearing failure Modes, mechanisms and causes. Eng. Fail. Anal. 2023, 152, 107518. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Lei, Y. Data anomaly detection of bridge structures using convolutional neural network based on structural vibration signals. Symmetry 2021, 13, 1186. [Google Scholar] [CrossRef]
  29. Nunes, E.C. Anomalous sound detection with machine learning: A systematic review. arXiv 2021, arXiv:2102.07820. [Google Scholar]
  30. Liu, W.; Jiang, H.; Che, D.; Chen, L.; Jiang, Q. A Real-time Temperature Anomaly Detection Method for IoT Data. In Proceedings of the IoTBDS, Prague, Czech Republic, 7–9 May 2020; pp. 112–118. [Google Scholar]
  31. Kao, J.B.; Jiang, J.R. Anomaly detection for univariate time series with statistics and deep learning. In Proceedings of the 2019 IEEE Eurasia Conference on IOT, Communication and Engineering (ECICE), Yunlin, Taiwan, 3–6 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 404–407. [Google Scholar]
  32. Wan, T.H.; Tsang, C.W.; Hui, K.; Chung, E. Anomaly detection of train wheels utilizing short-time Fourier transform and unsupervised learning algorithms. Eng. Appl. Artif. Intell. 2023, 122, 106037. [Google Scholar] [CrossRef]
  33. Wang, J.; Shao, S.; Bai, Y.; Deng, J.; Lin, Y. Multiscale wavelet graph autoencoder for multivariate time-series anomaly detection. IEEE Trans. Instrum. Meas. 2022, 72, 1–11. [Google Scholar] [CrossRef]
  34. Goyal, V.; Yadav, A.; Kumar, S.; Mukherjee, R. Lightweight LAE for Anomaly Detection with Sound based Architecture in Smart Poultry Farm. IEEE Internet Things J. 2023, 11, 5. [Google Scholar] [CrossRef]
  35. Yu, J.; Gao, X.; Li, B.; Zhai, F.; Lu, J.; Xue, B.; Fu, S.; Xiao, C. A filter-augmented auto-encoder with learnable normalization for robust multivariate time series anomaly detection. Neural Netw. 2024, 170, 478–493. [Google Scholar] [CrossRef]
  36. Dong, Z.; Zhao, D.; Cui, L. An intelligent bearing fault diagnosis framework: One-dimensional improved self-attention-enhanced CNN and empirical wavelet transform. Nonlinear Dyn. 2024, 112, 6439–6459. [Google Scholar] [CrossRef]
  37. Li, W.; Zhang, L.C.; Wu, C.H.; Wang, Y.; Cui, Z.X.; Niu, C. A data-driven approach to RUL prediction of tools. Adv. Manuf. 2024, 12, 6–18. [Google Scholar] [CrossRef]
  38. Lee, D.; Choo, H.; Jeong, J. Leak Detection and Classification of Water Pipeline Data Using LSTM Auto-Encoder with Xavier Initialization. In Proceedings of the 2023 IEEE/ACIS 8th International Conference on Big Data, Cloud Computing, and Data Science (BCD), Hochimin City, Vietnam, 14–16 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 69–74. [Google Scholar]
  39. Dong, F.; Chen, S.; Demachi, K.; Yoshikawa, M.; Seki, A.; Takaya, S. Attention-based time series analysis for data-driven anomaly detection in nuclear power plants. Nucl. Eng. Des. 2023, 404, 112161. [Google Scholar] [CrossRef]
  40. Jiang, K.; Liu, H.; Ruan, H.; Zhao, J.; Lin, Y. ALAE: Self-attention reconstruction network for multivariate time series anomaly identification. Soft Comput. 2023, 27, 10509–10519. [Google Scholar] [CrossRef]
  41. Xu, J.; Wu, H.; Wang, J.; Long, M. Anomaly transformer: Time series anomaly detection with association discrepancy. arXiv 2021, arXiv:2110.02642. [Google Scholar]
  42. Wu, G.; Zhao, Z.; Fu, G.; Wang, H.; Wang, Y.; Wang, Z.; Hou, J.; Huang, L. A Fast k NN-Based Approach for Time Sensitive Anomaly Detection over Data Streams. In Proceedings of the International Conference on Computational Science, Computational Science in the Interconnected World, Faro, Portugal, 12–14 June 2019; Springer: Heidelberg, Germany, 2019; pp. 59–74. [Google Scholar]
  43. Hu, Z.; Wu, T.; Zhang, Y.; Li, J.; Jiang, L. Time series anomaly detection based on graph convolutional networks. In Proceedings of the 2020 2nd International Conference on Applied Machine Learning (ICAML), Changsha, China, 16–18 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 138–145. [Google Scholar]
  44. Oluwasanmi, A.; Aftab, M.U.; Baagyere, E.; Qin, Z.; Ahmad, M.; Mazzara, M. Attention autoencoder for generative latent representational learning in anomaly detection. Sensors 2021, 22, 123. [Google Scholar] [CrossRef]
  45. Luong, M.T.; Pham, H.; Manning, C.D. Effective approaches to attention-based neural machine translation. arXiv 2015, arXiv:1508.04025. [Google Scholar]
Figure 1. Types of failures in rotating machines.
Figure 1. Types of failures in rotating machines.
Sensors 24 04855 g001
Figure 2. LSTM autoencoder model.
Figure 2. LSTM autoencoder model.
Sensors 24 04855 g002
Figure 3. STFT model.
Figure 3. STFT model.
Sensors 24 04855 g003
Figure 4. GCN-based LSTM autoencoder with self-attention model.
Figure 4. GCN-based LSTM autoencoder with self-attention model.
Sensors 24 04855 g004
Figure 5. Step 1.
Figure 5. Step 1.
Sensors 24 04855 g005
Figure 6. Types of bearing faults.
Figure 6. Types of bearing faults.
Sensors 24 04855 g006
Figure 7. Step 2.
Figure 7. Step 2.
Sensors 24 04855 g007
Figure 8. CWRU testbed.
Figure 8. CWRU testbed.
Sensors 24 04855 g008
Figure 9. Fault simulator testbed.
Figure 9. Fault simulator testbed.
Sensors 24 04855 g009
Figure 10. MQTT protocol and DAQ system.
Figure 10. MQTT protocol and DAQ system.
Sensors 24 04855 g010
Figure 11. Vibration sensor.
Figure 11. Vibration sensor.
Sensors 24 04855 g011
Figure 12. Sound sensor.
Figure 12. Sound sensor.
Sensors 24 04855 g012
Figure 13. The 2D PCA.
Figure 13. The 2D PCA.
Sensors 24 04855 g013
Figure 14. The 3D PCA.
Figure 14. The 3D PCA.
Sensors 24 04855 g014
Figure 15. Confusion matrix for the CWRU dataset.
Figure 15. Confusion matrix for the CWRU dataset.
Sensors 24 04855 g015
Figure 16. Confusion matrix for the fault simulator dataset.
Figure 16. Confusion matrix for the fault simulator dataset.
Sensors 24 04855 g016
Figure 17. MSE values for different K-neighbors.
Figure 17. MSE values for different K-neighbors.
Sensors 24 04855 g017
Figure 18. MAE values for different K-neighbors.
Figure 18. MAE values for different K-neighbors.
Sensors 24 04855 g018
Figure 19. PCA of the fault simulator test dataset (2D).
Figure 19. PCA of the fault simulator test dataset (2D).
Sensors 24 04855 g019
Figure 20. PCA of fault simulator test dataset (3D).
Figure 20. PCA of fault simulator test dataset (3D).
Sensors 24 04855 g020
Table 1. System specifications.
Table 1. System specifications.
Hardware EnvironmentSoftware Environment
CPU: Intel Core i7-8700KWindow 10, TensorFlow 2.11.0
GPU: NVIDIA GeForce RTX 3080Python 3.7, CUDA 11.1
Table 2. CWRU dataset.
Table 2. CWRU dataset.
Fault Type (2HP)NormalInner FaultBall FaultOuter Fault
Diameter (inch)-0.0070.0140.0210.0070.0140.0210.0070.0140.021
Train Data34343434343434343434
Validation Data11111111111111111111
Test Data11111111111111111111
Table 3. Hyperparameters for the CWRU dataset.
Table 3. Hyperparameters for the CWRU dataset.
LayersConfigurationsInput ShapeOutput Shape
Sliding Window & STFT
Sliding WindowWindow_Size = 8, Step_Size = 4(230, 9)(56, 8, 9)
STFTFS = 5000, Nperseg = 8(56, 8, 9)(56, 8, 5, 9)
K-NN Layer
K-NNK_Neighbors = 2(56, 8, 5, 9)(56, 8, 5, 9)
GCN Layer
GCNConv_1In_Channels = 9, Out_Channels = 16(56, 8, 5, 9)(56, 8, 5, 16)
GCNConv_2In_Channels = 16, Out_Channels = 9(56, 8, 5, 16)(56, 8, 5, 9)
Concatenate Data
Concatenate-(56, 8, 5, 9)(56, 8, 45)
LSTM autoencoder with Self-Attention
LSTM Encoder_1Units = 64(56, 8, 45)(56, 8, 64)
LSTM Encoder_2Units = 32(56, 8, 64)(56, 8, 32)
Attention Layer-(56, 8, 32)(56, 32)
Repeat VectorN = 8(56, 32)(56, 8, 32)
LSTM Decoder_1Units = 32(56, 8, 32)(56, 8, 32)
LSTM Decoder_2Units = 64(56, 8, 32)(56, 8, 64)
LSTM Decoder_3Units = 128(56, 8, 64)(56, 8, 128)
TimeDistributedUnits = 10(56, 8, 128)(56, 8, 10)
Flatten Layer-(56, 8, 10)(56, 80)
Classifier
Dense_1Units = 64, ReLU(56, 80)(56, 64)
Dropout_1Rate = 0.5(56, 64)(56, 64)
Dense_2Units = 32, ReLU(56, 64)(56, 32)
Dropout_2Rate = 0.2(56, 32)(56, 32)
Output DenseUnits = 10, Softmax(56, 32)(56, 10)
Table 4. Fault simulator dataset.
Table 4. Fault simulator dataset.
Fault TypeNormalCarriageHigh-SpeedLubricationCorrosion
Data Label01234
Train Data28,12428,12428,12428,12428,124
Validation Data93749374937493749374
Test Data93759375937593759375
Table 5. Hyperparameters for the fault simulator dataset.
Table 5. Hyperparameters for the fault simulator dataset.
LayersConfigurationsInput ShapeOutput Shape
Sliding Window & STFT
Sliding WindowWindow_Size = 128, Step_Size = 64(3,000,000, 4)(46,875, 128, 4)
STFTFS = 5000, Nperseg = 128(46,875, 128, 4)(46,875, 128, 65)
K-NN Layer
K-NNK_Neighbors = 2~8(46,875, 128, 65, 4)(46,875, 128, 65, 4)
GCN Layer
GCNConv_1In_Channels = 4, Out_Channels = 16(46,875, 128, 65, 4)(46,875, 128, 65, 16)
GCNConv_2In_Channels = 16, Out_Channels = 4(46,875, 128, 65, 16)(46,875, 128, 65, 4)
Concatenate Data
Concatenate-(46,875, 128, 65, 4)(46,875, 128, 260)
LSTM autoencoder with self-attention
LSTM Encoder_1Units = 128(46,875, 128, 260)(46,875, 128, 128)
LSTM Encoder_2Units = 64(46,875, 128, 128)(46,875, 128, 64)
LSTM Encoder_3Units = 32(46,875, 128, 64)(46,875, 128, 32)
Attention Layer-(46,875, 128, 32)(46,875, 32)
Repeat VectorN = 128(46,875, 32)(46,875, 128, 32)
LSTM Decoder_1Units = 32(46,875, 128, 32)(46,875, 128, 32)
LSTM Decoder_2Units = 64(46,875, 128, 32)(46,875, 128, 64)
LSTM Decoder_3Units = 128(46,875, 128, 64)(46,875, 128, 128)
TimeDistributedUnits = 5(46,875, 128, 128)(46,875, 128, 5)
Flatten-(46,875, 128, 5)(46,875, 640)
Classifier
Dense_1Units = 64, LeakyReLU(46,875, 640)(46,875, 64)
Dropout_1Rate = 0.5(46,875, 64)(46,875, 64)
Dense_2Units = 32, LeakyReLU(46875, 64)(46,875, 32)
Dropout_2Rate = 0.2(46,875, 32)(46,875, 32)
Output DenseUnits = 5, Softmax(46,875, 32)(46,875, 5)
Table 6. Model comparison using the CWRU dataset.
Table 6. Model comparison using the CWRU dataset.
ModelAccuracyPrecisionRecallF1 Score
LSTM0.9310.9320.9300.927
ConvLSTM0.9660.9630.9630.963
Autoencoder0.9000.9050.9000.899
LSTM Autoencoder0.9550.9570.9550.955
LSTM Autoencoder with Self-Attention0.9730.9730.9730.972
Table 7. Model comparison using the fault simulator dataset.
Table 7. Model comparison using the fault simulator dataset.
ModelAccuracyF1 Score
LSTMw/o STFTw/o GCN0.8740.873
with GCN0.8730.872
with STFTw/o GCN0.8680.867
with GCN0.9810.980
ConvLSTMw/o STFTw/o GCN0.8930.891
with GCN0.8380.837
with STFTw/o GCN0.9910.992
with GCN0.9970.997
Autoencoderw/o STFTw/o GCN0.4940.494
with GCN0.8040.804
with STFTw/o GCN0.7220.723
with GCN0.9940.994
LSTM Autoencoderw/o STFTw/o GCN0.4370.414
with GCN0.7930.792
with STFTw/o GCN0.8000.798
with GCN0.9490.949
LSTM Autoencoder with Self-Attentionw/o STFTw/o GCN0.7100.703
with GCN0.8010.800
with STFTw/o GCN0.8300.820
with GCN0.9990.999
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, D.; Choo, H.; Jeong, J. GCN-Based LSTM Autoencoder with Self-Attention for Bearing Fault Diagnosis. Sensors 2024, 24, 4855. https://doi.org/10.3390/s24154855

AMA Style

Lee D, Choo H, Jeong J. GCN-Based LSTM Autoencoder with Self-Attention for Bearing Fault Diagnosis. Sensors. 2024; 24(15):4855. https://doi.org/10.3390/s24154855

Chicago/Turabian Style

Lee, Daehee, Hyunseung Choo, and Jongpil Jeong. 2024. "GCN-Based LSTM Autoencoder with Self-Attention for Bearing Fault Diagnosis" Sensors 24, no. 15: 4855. https://doi.org/10.3390/s24154855

APA Style

Lee, D., Choo, H., & Jeong, J. (2024). GCN-Based LSTM Autoencoder with Self-Attention for Bearing Fault Diagnosis. Sensors, 24(15), 4855. https://doi.org/10.3390/s24154855

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop