Next Article in Journal
Depth-Dependent Control in Vision-Sensor Space for Reconfigurable Parallel Manipulators
Previous Article in Journal
Cross-Layer Federated Learning for Lightweight IoT Intrusion Detection Systems
Previous Article in Special Issue
Phishing URLs Detection Using Sequential and Parallel ML Techniques: Comparative Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CaFANet: Causal-Factors-Aware Attention Networks for Equipment Fault Prediction in the Internet of Things

1
The 7th Rescarch Institute of Electronics Technology Group Corporation, Guangzhou 510310, China
2
School of Information Science and Technology, Northwest University, Xi’an 710127, China
3
Department of Data Science, New Jersey Institute of Technology, Newark, NJ 07102, USA
*
Author to whom correspondence should be addressed.
Co-first author, these authors contributed equally to this work.
Sensors 2023, 23(16), 7040; https://doi.org/10.3390/s23167040
Submission received: 11 July 2023 / Revised: 2 August 2023 / Accepted: 7 August 2023 / Published: 9 August 2023

Abstract

:
Existing fault prediction algorithms based on deep learning have achieved good prediction performance. These algorithms treat all features fairly and assume that the progression of the equipment faults is stationary throughout the entire lifecycle. In fact, each feature has a different contribution to the accuracy of fault prediction, and the progress of equipment faults is non-stationary. More specifically, capturing the time point at which a fault first appears is more important for improving the accuracy of fault prediction. Moreover, the progress of the different faults of equipment varies significantly. Therefore, taking feature differences and time information into consideration, we propose a Causal-Factors-Aware Attention Network, CaFANet, for equipment fault prediction in the Internet of Things. Experimental results and performance analysis confirm the superiority of the proposed algorithm over traditional machine learning methods with prediction accuracy improved by up to 15.3%.

1. Introduction

Equipment fault prediction is extremely important for the maintenance of large equipment in the Industrial Internet of Things. Compared with traditional signal-analysis-based fault diagnosis methods [1,2,3], deep-learning-based prediction algorithms [4,5,6] can achieve better performance in equipment fault prediction.
In the industrial Internet of Things, a large device is composed of multiple components, each of which has different physical characteristics during machine operation. Sensors can be deployed to collect measurements on the operation of specific components, e.g., we may deploy vibration sensors at the bearings to monitor their operation and extract features from these monitoring data, which reflect the physics of the detected component, as inputs to the deep learning framework.
However, feature extraction and interaction are calculated in a black box in a deep learning framework [7,8]. In fact, feature selection is crucial to the accuracy of equipment fault prediction. Moreover, the features extracted from the monitoring data of the equipment operational status can characterize the operational status of the equipment from different perspectives. Essentially, the representation abilities of these extracted features are different, and their contributions to equipment fault prediction are also different. Therefore, we should treat the features extracted from monitoring data accordingly when designing fault prediction algorithms.
Typically, equipment can operate normally for months or even years. Moreover, the progress of equipment faults is often non-stationary. Compared with normal operation mode, the data collected by sensors would undergo significant changes when the equipment malfunctions. What kind of data is more meaningful and how to capture these valuable data are crucial for equipment fault prediction. Therefore, we should identify the most influential parts of the monitoring data and infer the key data points closely related to equipment faults.
In this paper, we propose a Causal-Factors-Aware Attention Networks (CaFANet) for equipment fault prediction. Different from traditional deep learning, we first conduct causal analysis of extracted signal features to quantify the impact of each feature on the prediction accuracy, and embed the influence weight of the features on the prediction performance into the feature representation for equipment fault diagnosis. Then, we calculate local causal-factors-aware weights based on a single-layer transformer and global attention weights via a time-aware key-query attention scheme. Finally, we integrate local attention with global attention using a dynamic attention fusion strategy. Compared with state-of-the-art fault prediction algorithms, our proposed algorithm not only considers the impact of features on prediction performance but also deliberates the time information of faults in both local and global levels.
Our contributions are summarized as follows.
  • We quantify the influence of features on the prediction accuracy via causal analysis and assign a causal influence weight to each feature according to its contributions to equipment fault prediction performance.
  • We investigate the influence of features and time information on equipment fault prediction using a single-layer transformer, compute local weights and global weights accordingly, and finally obtain an aggregated attention weight for each sequence to achieve better equipment fault prediction performance.
  • We evaluate the performance of CaFANet using a publicly available equipment fault prediction dataset. Compared with eleven classical baselines, the experimental results validate the effectiveness and efficiency of CaFANet.
The remainder of this paper is organized as follows. We survey the related works in Section 2. We present the system model and the target problem in detail in Section 3. The equipment fault prediction framework is described in Section 4. The experimental results and performance analysis are presented in Section 5. Finally, we conclude this paper in Section 6.

2. Related Work

Accurate equipment fault prediction can effectively reduce industrial accidents in the Industrial Internet of Things. Both traditional and modern methods can be utilized for predicting equipment faults.
The traditional methods for equipment fault diagnosis include Bayesian theory, maximum likelihood estimation, expert systems, Kalman filtering, etc. Hmida et al. [9] investigated a robust equipment fault diagnosis scheme by using a third-order Kalman filter. By taking a directed graph model into account, Jaise et al. [10] introduced a fault tree diagnosis strategy for vehicle systems. Wang et al. [11] enhanced the equipment fault signal by taking advantage of the time-frequency reduction and short-time Fourier transform and improved the performance of bearing fault diagnosis. Chen et al. [12] improved the accuracy of fault classification based on short-time Fourier transform. More importantly, this model achieved a good fault classification performance with a smaller feature scale. Shi et al. [13] designed an efficient diagnosis system by integrating an expert system and a fuzzy neural network to establish a diagnosis rule library. Zhang et al. [14] studied the stochastic resonance behavior of a coupled stochastic resonance system with time delay under the fluctuation of mass and frequency. Yan et al. [15] proposed an auxiliary indicator-based fault diagnosis system, which overcame the drawback of manually setting the modal parameters in the original singular spectral decomposition. Ding et al. [16] combined an expert system and an agent model framework to build a collaborative power grid fault diagnosis system. Although these methods based on old-fashion patterns can assist in equipment fault diagnosis, they highly rely on the professional knowledge and experience of designers. Moreover, these methods treat the extracted features equally and cannot handle the nonstationary progression of equipment faults.
There are many new technologies available for equipment fault diagnosis and prediction, such as multimodal learning, attention mechanisms, causal analysis, etc. Nan et al. [6] introduced a multimodal learning framework to combine low-quality monitoring data and high-quality monitoring data for better equipment fault prediction performance. Zhang et al. [5] realized a multidimensional neural network fusion diagnosis system with the help of a fuzzy inference system based on the data collected by multiple sensors. Zhao et al. [17] proposed a blind source extraction (BSE) method for bearing fault diagnosis based on empirical mode decomposition (EMD) and time dependence. Li et al. [18] proposed a document-level novelty detection algorithm based on attention mechanisms. Ruikun et al. [4] built a quantum neural-based network to predict fault type. Wang et al. [19] proposed an approach to discover causal structures from incomplete data by using a novel encoder and reinforcement learning. Aviles-Cruz et al. [20] designed a novel framework that used Granger-causality on a smartphone to classify and analyze human activities. In this study, to improve the performance of prediction, we design a novel equipment fault prediction framework inspired by causal analysis and an attention mechanism. We attempt to explore the intrinsic differences of features and quantify their contribution to device fault prediction performance through causal analysis. At the same time, based on an attention mechanism, we attempt to handle the non-stationary fault progression and assign different weights to serialized data points.

3. System Model

We consider an equipment fault prediction system mounted with I vibration sensors, each of which gathers K serialized data points about the operating status of target equipments. In longitudinal monitoring data, the monitoring data from each sensor corresponding to a specific fault is a time-ordered monitoring sequence. We obtain the time-domain characteristics of the initial sequences based on time-domain signal analysis. We extract J features from every data point. In summary, we have K serialized data points with J features from I sensors in the monitoring system.
Given the collected sequences throughout the entire lifecycle, the goal of equipment fault prediction system is to predict if the target equipment would fail with one certain fault in the near future. Taking a bearing as an example, as shown in Figure 1, it has three operating states, i.e., the normal operation, the early fault, and the fault. Generally, the bearing would operate in normal mode for a long period of time. Once a fault occurs, there is a significant change in the vibration signal. However, what kind of data points are more valuable for predicting equipment faults? More importantly, how should we evaluate the contribution of these data points to the prediction performance? Due to the inherent differences between causal and non-causal features, the features that affect prediction accuracy should also be treated differently. Therefore, the objective of this study is to achieve better prediction performance by taking advantage of features and time information.

4. The Proposed Prediction Framework

Our prediction framework, CaFANet (Causal-Factors-Aware Attention Networks for equipment fault prediction), as shown in Figure 2, includes four components: (1) Resampling. For longitudinal monitoring data, we first preprocess the classified equipment fault samples via time-domain analysis and encode the time-domain characteristics to obtain a time-ordered monitoring sequence N, N = [ n 1 , n 2 , , n K ] . (2) Causal analysis. We perform causal analysis on the extracted signal features to quantify the impact of each feature on the prediction accuracy. More specifically, we evaluate the model accuracy error by removing any feature and then obtain the corresponding causal contribution weight for each feature W F , W F = [ W 11 , W 12 , , W I J 1 , W I J ] . (3) Time attention analysis. We recalculate the feature embedding to obtain a new embedding sequence C, C = [ c 1 , c 2 , , c K , c T ] by taking causal weights and the time information into account, where C contains the time information and causal weights, and c T is the feature information under failure conditions. We use transformer to obtain the hidden layer representation H = [ h 1 , h 2 , , h k , h T ] of the embedded sequence C and then calculate the local attention score W l . To simulate equipment faults over time, we first embed h T into a query vector and then embed the time difference between the occurrence of the failure and the specific data point into a key vector. We use an attention mechanism with time information to obtain the global attention score W g for each embedded sequence. (4) Time-aware dynamic attention fusion. We fuse the local attention score W l and the temporal attention score W g to generate the aggregated attention score φ ¯ k .

4.1. Preprocessing

Based on the equipment state monitoring information throughout the lifecycle, which includes device failure data collected by multiple deployed sensors, we extract features and predict the device operation status. The data of each device are firstly organized, classified into different fault categories, and manually annotated to form a dataset.
We resample the labeled equipment fault data and stamp each sampling point with a timestamp. The collected signal samples are organized according to the device they come from. Using the signal time-domain analysis (here, we consider the variance, root mean square difference, and mean, etc.), we obtain the time-domain characteristics of the sorted samples and make corresponding time-domain feature samples. We then standardize the extracted time-domain feature samples to generate a unified feature code as:
s i j k = f i j k M i n ( f i j k ) M a x ( f i j k ) M i n ( f i j k ) × γ × ( i × J ) × j ,
where s i j k is the feature code of the kth serialized point of the jth feature gathered by sensor i, i = 1 , 2 , , I , j = 1 , 2 , , J , k = 1 , 2 , , K , M a x ( f i j k ) and M i n ( f i j k ) are the corresponding maximum and minimum values, respectively, and γ is a coefficient related to the size of the feature space. Note that I is the total number of deployed sensors, J is the number of time domain features extracted from monitoring data, and K is the total number of monitoring sample sequences for each device.
We discretize the continuous values of feature codes obtained by Equation (1), each feature code s i j k can be converted into a corresponding b i j k in a binary form represented by 0 , 1 o , o = γ × I × J , and all s i j k are rounded down. Let N denote the sequence of samples, and we have
N = [ n 1 , n 2 , , n k , , n K ] ,
where n k = [ b 11 k , b 12 k , , b I J k ] , and n k is the k-th sequence.

4.2. Causal Analysis

Traditional fault prediction methods for large-scale equipment treat all features equally [6]. In fact, each feature actually makes different contribution to the prediction accuracy because of its unique representation ability. More specifically, the performance of prediction depends on the selection of features. Therefore, we use causal analysis to quantify the influence of characteristics on prediction performance. The function to quantify the contribution of each feature to the accuracy of equipment fault prediction is expressed as follows:
Δ ε , f i j = ε F f i j ε F ,
where Δ ε , f i j is the effect of feature f i j on fault prediction, ε F is the error in fault prediction, and ε F f i j is the error in fault prediction without feature f i j . We use M to denote the embedding sequence of the model and M f i j to denote the embedding sequence without feature f i j . M = [ m 1 , m 2 , , m k ] , M f i j = [ m 1 * , m 2 * , and , m K * ] , m k can be calculated as follows:
m k = W m n k + b m ,
where m k is the embedding data of the kth sequence of the embedded sequence M, n k is the binary input of the feature code f i j k of the kth sequence, W m is the initialization weight matrix, W m R v , b m is the bias vector, b m R v , and v is the dimension of w m and b m .
The traditional transformer-based schemes are mainly applied to solve natural language processing problems but are rarely used in the Internet of Things. There are two advantages to applying transformer to the fault prediction of large equipment. Firstly, the transformer allows the vector to use the self-attention mechanism every time to learn its relationship with other samples from other sensors deployed in the same equipment. Secondly, the structure of the transformer provides the ability to calculate multiple samples in parallel. Thus, we use an attention-mechanism-based single-layer transformer as a model for causal analysis, denoted by T F ( ) , and the label of the prediction result denoted by e ^ , and we have:
e ^ ε F = T F ( M ) ,
e ^ ε F f i j = T F ( M f i j ) ,
where e ^ ε F and e ^ ε F f i j denote the model prediction labels with and without feature f i j , respectively.
Let e denote the true label and the cross-entropy loss function L denote the predicted error. The fault prediction error of Equation (3) can be expressed as:
ε F = L ( e , e ^ ε F ) ,
ε F f i j = L ( e , e ^ ε F f i j ) .
We can measure the contribution of each feature to the prediction accuracy via Equations (3)–(8). Then, each feature is assigned a weight according to its influence on prediction performance as follows:
W f i j = Δ ε , f i j | Δ ε , f i j | .
Then, we obtain the causal weight W F = [ W f 11 , W f 12 , , W f i j ] .

4.3. Time Attention Analysis

We recalculate the feature code using the causal influence weights obtained in Section 4.2 as follows:
s i j k W = W f i j f i j k min f i j k max f i j k min f i j k × γ × ( i × J ) × j ,
where s i j k W is the updated feature code of the k-th serialized point of the j-th feature from sensor i, i = 1 , 2 , , I , j = 1 , 2 , , J , k = 1 , 2 , , K .
Similarly, we can obtain the sample sequence N W = [ N 1 W , N 2 W , , N K 1 W , N K W , N T W ] that combines the causal influence weights, where N K W is the binary input of the updated feature code s i j k W , and N T W is the feature representation of the specific fault of the equipment in the failure state. Note that the N T W is the same for all samples when predicting a particular type of fault.
The moment when the operating state of the equipment transitions from normal to abnormal is crucial. To improve the performance of equipment fault prediction, it is necessary to capture this moment. The closer the sampling point is to the fault point, the more attention it deserves and the greater the weight it should be given. Therefore, we should take time information into account. More specifically, we embed time information and then use the updated vectors also as inputs. Firstly, since the time information and the feature vector are not in the same latent space, to characterize the importance of features associated with the time information, we need to embed the time information into the feature vector space as follows:
z k = W z ( 1 t a n h ( ( 2 p k T ) 2 ) ) + b z ,
where z k is the embedded time vector, t a n h is the hyperbolic tangent function, T is the sampling time of the last serialized point in the sequence, p k is the time difference between the current sampling point and the last sampling point, p k = T t k , W z and b z are the initialization weight matrix and bias vector, respectively, w z R v , and b z R v .
The sequence can be generated by embedding with time information and causal influence weights as follows:
c k = z k + ( W c N k W + b c ) ,
where c k is the embedded conterpart of the kth sequence. The sequence C = [ c 1 , c 2 , , c K , c T ] can be obtained by Equation (12). Note that the c T is the same for all samples.
We use a single-layer transformer (denoted by T F ( ) ) to learn the relationship between the embedding sequences and the equipment faults as follows:
h 1 , h 2 , , h K , h T = T F ( c 1 , c 2 , , c K , c T ) ,
where h k is the hidden representation of c k via the transformer, k = 1 , 2 , , K , and h T is the representation of c T , which is the embedding sequence of a specific device fault.
Due to non-stationary fault progression, we should review all sample points, rather than only focusing on the current sample point, when we analyze equipment faults and identify the turning point of the equipment’s operating status, i.e., the key point when it transitions from the normal operating state to the abnormal state. To carry out this fault diagnosis, we calculate a local attention score for each sample point as follows:
u k = W u h k + b u ,
where W u and b u are the weight matrix and bias vector, respectively, W u R l , and b u R . We use a softmax layer to generate local attention weights based on the local attention score U = [ u 1 , u 2 , , u K ] . We have
W l = S o f t m a x ( u 1 , u 2 , , u K ) = l 1 , l 2 , , l K .
We learn a local attention weight, which reflects the importance of sample points on the accuracy of equipment fault prediction, for each sample via the aforementioned time-aware transformer. Not only does a single sample point need to be focused on, but attention to the progress of the equipment operation status is more important for equipment fault prediction. Compared to existing samples, if the gathered data show only slight changes over a long period of time, it is not necessary to collect and analyze the data at regular intervals. Only when there are significant fluctuations in the monitoring data do more samples need to be collected, which helps to predict equipment failures more accurately. Therefore, the time interval between samples is crucial for fault prediction. We introduce a key-query attention model to simulate the process. Firstly, we convert the hidden representation h T obtained by Equation (13) into a query vector q as follows:
x = R e L U ( W x h T + b x ) ,
where W x R l × q and b x R q are the weights and bias matrix, respectively, l is the dimension of h k , q is the dimension of b x , and R e L U is an activation function that only maintains positive values.
To capture the importance of the time information itself during the equipment operation process, we use the time difference p k between the fault point and the sample as the key vector e k for the attention mechanism as follows:
e k = t a n h ( W e ( 1 t a n h ( ( 2 p k T ) 2 ) ) + b e ) ,
where W e R l × q and b e R q are parameters to be learned, e k is the key vector of the kth sequence, l is the dimension of h k , q is the dimension of b e , and E = [ e 1 , e 2 , , e K ] is the set of key vectors.
Combined with the query vector x derived from Equation (16), we can obtain the global attention score r k as follows:
r k = x e k r .
We can obtain global attention weights W g based on the global attention score r k via a softmax layer as follows:
W g = S o f t m a x ( r 1 , r 2 , , r K ) = g 1 , g 2 , , g K ,
where g K is the global time-aware attention weight of the Kth sequence.

4.4. Attention Fusion Strategy

The two attention vectors obtained earlier pay attention to different perspectives for equipment fault prediction. W l focuses on the characteristics of each embedded sequence, while W g concentrates on the time information of all samples. To obtain more accurate prediction results, we should take feature and time information into account. Therefore, we introduce a dynamic attention fusion strategy that integrates local attention with global attention. More specifically, we embed h T into a new space and normalize it as follow:
v = S o f t m a x ( W v h T + b v ) = a l , a g ,
where W v R 2 × l , b v R 2 , l is the dimension of h K . We then obtain an aggregated attention weight for each sequence based on local weights and global weights as follows:
φ k = l k a l + g k a g .
Finally, we can obtain the final attention score γ by normalizing the aggregated attention weights for each vector as follows:
γ = φ k k φ k .
The final attention score combines the influence of the features and time information on the device fault prediction. The feature analysis focuses on the signal values of the dataset throughout the entire lifecycle, while the time information analysis pays attention to the changes in the time interval between samples over a long period. Integrating signal features and time information can better capture the contribution of features to the prediction results.
We obtain the representation of a sample based on the hidden layer representation obtained by Equation (13) and the final attention score obtained by Equation (22) as follows:
h = γ h k .
We perform a binary prediction of equipment faults based on the the representation obtained by Equation (23) via a softmax layer as follows:
d = S o f t m a x ( W d h + b d ) ,
where W d and b d are parameters to be learned, W d R 2 × l , b d R 2 .
We define a loss function to learn all parameters, let θ denote all the parameters and d denote the ground truth, and use the cross entropy between the predicted probabilities d and the actual value d as the loss function as follows:
L ( θ ) = 1 N n = 1 N ( d l o g ( d ) + ( 1 d ) l o g ( 1 d ) ) ,
where | N | is the total number of samples.

5. Experimental Evaluation

In this section, we describe the experiments we conducted and compare the performance of the proposed prediction model and baselines on a publicly available equipment failure prediction dataset.

5.1. Dataset and Data Preprocessing

We performed experiments on the XJTU-SY bearing dataset [21], which measures the operating conditions of bearings and is widely applied to the fault prediction and remaining life estimation of the bearings. The data are collected by vibration sensors deployed on the adjustable speed generator shaft, supporting the transmission bearings, the hydraulic drive load control system, and the testing transmission bearings, etc. The monitored data can be used to evaluate if the equipment has malfunctioned.
This dataset, XJTU-SY, contains information on five bearings, each with three operating modes, and a total of 15 sets of samples throughout the entire lifecycle, each of which serves as a sub-dataset. More specifically, the first bearing has three sub-datasets, denoted as Bearing 1_1, Bearing 1_2 and Bearing 1_3. Meanwhile, we set the sampling frequency as 25.6 kHz, the average sampling interval as 1 min, and each sampling time as 1.28 s. Therefore, we obtained 32,769 signals from each sensor per minute.
Given the limited number of operating modes in the XJTU-SY bearing dataset, we enhanced the dataset by resampling because of the periodicity of the original samples. The information about each operating mode of each bearing can be divided into a normal operation stage, an early fault stage, and a fault stage. Due to the limited value of the equipment fault data for fault prediction, signals at the fault stage should not be considered. At the same time, the information entropy contained in the signal during the normal operation stage is very small; so, the information at this stage should be considered as little as possible. On the other hand, the signals corresponding to the period near the occurrence of the fault are extremely important for fault prediction and require more attention. Taking the samples at a speed of 2100 r/min and a radial force of 12 kN as an example, starting from the time when the equipment is turned on, every 2000 signals were treated as one data point and then labeled. The characteristics of these data points were obtained through numerical analysis, and thus the sequential points of each operating mode for each bearing during the whole life cycle were obtained. We sampled 20 random data points during the normal operation stage and 40 data points during the early fault stage to create a sequence, as shown in Figure 3. One thousand sequences for each operating mode of each bearing were generated as a dataset, 60% of which were used as the training set, 20% as the validation set, and 20% as the testing set.
Obviously, each sub-dataset contained the vibration signal of a bearing under one certain operating mode throughout the entire lifecycle. To verify the validity of the proposed model, we merged the sub-datasets belonging to different fault types to obtain four combined datasets. We created Dataset 1 by merging the sub-datasets Bearing 2_1 and Bearing 2_3, with inner race wear as the prediction target. We created Dataset 2 by merging the sub-datasets Bearing 2_3 and Bearing 2_5, with cage fracture as the prediction target. We created Dataset 3 by merging the sub-datasets Bearing 1_1, Bearing 2_3, and Bearing 2_5 datasets, with outer race wear as the prediction target. We created Dataset 4 by merging all the sub-datasets, with outer race fracture as the prediction target.

5.2. Time-Domain Signal Features

We selected seven time-domain features for equipment fault prediction and conduct causal analysis to measure the influence of these features on the model. The features are described as follows:
  • Variance (VAR). The VAR is used to measure the statistical dispersion of the signal. The larger the variance, the greater the signal variation. The smaller the variance, the smaller the signal fluctuation.
  • Root Mean Square (RMS). The RMS is not sensitive to early vibration signals but has good stability.
  • Average Value (AV). The AV can be used to measure the stability of signals and reflect the static properties of signal fluctuations.
  • Kurtosis (KU). KU can be used to measure the probability distribution of random variables. KU has good performance for faults with pulse signals. However, KU fails and have poor stability when a fault occurs.
  • Skewness (SK). SK can be used to measure the degree and direction of data distribution deviation and can characterize the degree of numerical asymmetry distribution. SK has good performance in the early fault stage but fails after a fault occurs.
  • Crest Factor (CF). The CF is defined as the ratio of the peak to the rectified average value and can be used to judge whether there are pulses in the signal.
  • Margin Factor (MF). The MF is defined as the ratio of the signal peak to the root square amplitude and is more sensitive to changes in the signal.

5.3. Experiments Settings

All our experiments were based on PyTorch, implemented on the Windows 10 operating system, run on 32 GB of RAM and GeForce RTX 3070 Ti GPU. The batch size was set as 50, the learning rate was set as 0.01, the space coefficient γ was set as 1400, the dropout was set as 0.5, the dimension of the hidden space I for prediction was 256, and Adam [22] was chosen as the optimizer. We rnm each prediction algorithm 10 times on each dataset and obtained the average of these 10 results as the final result.

5.4. Metrics

To evaluate the effectiveness and efficiency of the prediction model, we used accuracy (Acc), precision (Pre), and recall ( R e c a l l ) as metrics as follows:
A c c = i = 0 N P i , i i = 0 , j = 0 N P i , j ,
P r e = i = 0 N P i , i j = 0 N P j , i ,
R e c a l l = i = 0 N P i , i j = 0 N P i , j ,
where N is the total number of samples, and P i , j indicates that the predicted label is i and the actual label is j. More specifically, the prediction is correct only when i equals j.

5.5. Baselines

To demonstrate the superiority of the proposed prediction algorithm, we chose three types of algorithms as baselines.
  • The first category of baselines was classical machine learning algorithms, including SVM (Support Vector Machine) [23], LR (Linear Regression) [24] and RF (Random Forest) [25].
  • The second category of baselines was RNN-based algorithms, including LSTM (Long Short-Term Memory) [26], GRU (Gated Recurrent Unit) [27], and DA-RNN (Dual-stage Attention-based Recurrent Neural Network) [28]. These algorithms are the basic framework of most prediction algorithms.
  • The third category of baselines was algorithms that have achieved good performance in bearing fault prediction in recent years. CNN (Convolutional Neural Network) [29] is the classical and effective classification algorithm. DFC-CNN (Deep Fully Convolutional Neural Network) [30] is based on CNN and spectrogram transform for prediction. CNN-LSTM (multiscale CNN and LSTM) [31] can learn the original signal and encode it directly. GRU-HA (Gate Recurrent Unit and Hybrid Autoencoder) [32] can automatically learn the features of sequences. DA-AE (Deep Wavelet Autoencoder) [33] is an unsupervised learning algorithm and uses the original vibration signal for training.

5.6. Performance Analysis

Prediction Accuracy Comparison. As shown in Table 1, compared with the 11 benchmark algorithms, our proposed algorithm achieved the highest fault prediction accuracy. Compared with LR, our algorithm had an accuracy improvement of 15.3% on Dataset 4. Our algorithm assigned different weights to features that had different contributions to the prediction results and considered the impact of the time information on fault prediction. Therefore, our algorithm achieved optimal prediction performance on Dataset 4.
Prediction Performance Comparison. A comparison of 12 algorithms on Dataset 2, as shown in Table 2. Our method outperformed all the benchmarks on dataset 2 in terms of accuracy, precision, and recall. Compared with LR, our algorithm improved the accuracy by 7%, the accuracy ratio by 8%, and the recall rate by 6% on Dataset 2.
Effect of the Time-domain Features. Table 3 lists the impact of each time-domain feature on the accuracy of fault prediction under different fault types. CaFANet explores features that have the higher correlation with the predicted target. As expected, KU, CF and MF were more sensitive to the early numerical changes in bearing faults, thus having a greater impact on fault prediction tasks.
Effect of the Space Coefficients. The results shown in Figure 4 confirmed the impact of the space coefficient sizes on the fault prediction accuracy, which increased with the increase in the space coefficients. The larger the space coefficients, the higher the complexity of the encoding process and the higher the accuracy of the fault prediction. The running time increases linearly accordingly, as shown in Figure 5. However, if we chose smaller spatial coefficients, it is difficult to effectively express the relationship between features, resulting in low fault prediction accuracy.
Effect of the Sequence Length. Figure 6 illustrates the impact of the sequence length on the accuracy of fault prediction. Overall, the prediction accuracy increased with the increase in the sequence length. Actually, the impact of the sequence length on the prediction accuracy depended on the number of data points in the early fault stage.
Ablation Study. An ablation study on the impact of different components of the prediction model on accuracy was conducted by removing the causal analysis, global time attention, and time information embedding, respectively. As shown in Table 4, the prediction accuracy variation denoted the gap between the performance with and without the specific component. From Table 4, we observe that time embedding had the strongest impact on the prediction accuracy.

6. Conclusions

In this paper, we designed a causal-factors-aware attention network for predicting the fault types of bearings in the Internet of Things. Firstly, we measured the contribution of the time-domain features to the prediction model using causal analysis. Focusing on time-domain features with larger contributions, we then introduced a local attention mechanism to assign weights to the embedding sequence. After embedding time information, the features of the key stages were amplified, and the transformer with time-aware information was used to find the data points closely related to the faults. Finally, we reassigned weights to each signal sequence to predict the equipment faults effectively and efficiently. In this paper, we confirmed the importance of time information to prediction accuracy. Moreover, the proposed algorithm can provide guidance on the feature selection of equipment failure prediction in the industrial Internet of Things. Good prediction performance depends on high-quality data. However, it is usually difficult to obtain the monitoring data of the large equipment due to its confidential nature in the industrial Internet of Things. Even though device monitoring data is available, obtaining device fault information is still very challenging. Provided with samples and corresponding monitoring sequences, CaFaNet is able to perform interpretable equipment fault prediction and has potentials for practical use in production settings.

Author Contributions

Z.G.: Investigation, Conceptualization, Formal analysis, Methodology. S.H.: Conceptualization, Writing—original draft preparation. Y.L.: Data curation, Software, Visualization. X.N.: Data curation, Validation. X.Y.: Resources, Project administration, Supervision, Funding acquisition, Writing—review and editing. C.Q.W.: Supervision, Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was sponsored by the National Key Research and Development Program of China under Grant 2018YFB1802401.

Institutional Review Board Statement

Not applicable. This study did not involve humans or animals.

Informed Consent Statement

Not applicable. This study did not involve humans.

Data Availability Statement

The public dataset, XJTU-SY bearing datasets, can be downloaded at the following link: https://share.weiyun.com/5zLCgBL (accessed on 1 January 2021).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

In this study, we used the following abbreviations:
CaFANetCausal-Factor-Aware Attention Networks for Equipment Fault Prediction
SVMSupport Vector Machine
LRLinear Regression
RFRandom Forest
LSTMLong Short-Term Memory
GRUGated Recurrent Unit
DA-RNNDual-stage Attention-based Recurrent Neural Network
CNNConvolutional Neural Network
DEF-CNNDeep Fully Convolutional Neural Network
CNN-LSTMMultiscale CNN and LSTM
GRU-HAGate Recurrent Unit and Hybrid Autoencoder
DA-AEDeep Wavelet Autoencoder

References

  1. Peng, Y.; Liu, D.; Peng, X. A Review: Prognostics and Health Management. J. Electron. Meas. Instrum. 2010, 24, 1–9. [Google Scholar]
  2. Lei, Y.; Lin, J.; He, Z.; Zuo, M.J. A Review on Empirical Mode Decomposition in Fault Diagnosis of Rotating Machinery. Mech. Syst. Signal Process. 2013, 35, 108–126. [Google Scholar] [CrossRef]
  3. Li, S.; Xin, Y.; Li, X.; Wang, J.; Xu, K. A Review on the Signal Processing Methods of Rotating Machinery Fault Diagnosis. In Proceedings of the 8th Joint International Information Technology and Artificial Intelligence Conference (ITAIC 2019), Chongqing, China, 24–26 May 2019; pp. 1559–1565. [Google Scholar]
  4. Gong, R.K.; Ma, L.; Zhao, Y.J.; Yang, P.P.; Chen, L. Fault Diagnosis for Power Transformer based on Quantum Neural Network Information Fusion. Power Syst. Prot. Control 2011, 39, 79–84. [Google Scholar]
  5. Zhang, Q.; Wang, H.; Jing, W.; Mao, J.; Yuan, Z.; Hu, D. Shearer’s Coal-rock Recognition System based on Fuzzy Neural Network Information Fusion. China Mech. Eng. 2016, 27, 201. [Google Scholar]
  6. Nan, X.; Zhang, B.; Liu, C.; Gui, Z.; Yin, X. Multi-Modal Learning-Based Equipment Fault Prediction in the Internet of Things. Sensors 2022, 22, 6722. [Google Scholar] [CrossRef] [PubMed]
  7. Khan, S.; Yairi, T. A Review on the Application of Deep Learning in System Health Management. Mech. Syst. Signal Process. 2018, 107, 241–265. [Google Scholar] [CrossRef]
  8. Guo, L.; Lei, Y.; Xing, S.; Yan, T.; Li, N. Deep Convolutional Transfer Learning Network: A New Method for Intelligent Fault Diagnosis of Machines with Unlabeled Data. IEEE Trans. Ind. Electron. 2018, 66, 7316–7325. [Google Scholar] [CrossRef]
  9. Hmida, F.B.; Khémiri, K.; Ragot, J.; Gossa, M. Three-stage Kalman Filter for State and Fault Estimation of Linear Stochastic Systems with Unknown Inputs. J. Frankl. Inst. 2012, 349, 2369–2388. [Google Scholar] [CrossRef]
  10. Jaise, J.; Ajay Kumar, N.B.; Siva Shanmugam, N.; Sankaranarayanasamy, K.; Ramesh, T. Power System: A Reliability Assessment Using FTA. Int. J. Syst. Assur. Eng. Manag. 2013, 4, 78–85. [Google Scholar] [CrossRef]
  11. Wang, X.; He, Q. Machinery Fault Signal Reconstruction Using Time-Frequency Manifold. In Engineering Asset Management-Systems, Professional Practices and Certification, Proceedings of the 8th World Congress on Engineering Asset Management (WCEAM 2013) & the 3rd International Conference on Utility Management & Safety (ICUMAS), Hong Kong, China, 30 October–1 November 2013; Springer: Berlin/Heidelberg, Germany, 2014; pp. 777–787. [Google Scholar]
  12. Cheng, Y.; Yuan, H.; Liu, H.; Lu, C. Fault Diagnosis for Rolling Bearing based on SIFT-KPCA and SVM. Eng. Comput. 2017, 34, 53–65. [Google Scholar] [CrossRef]
  13. Shi, J.; Ma, J.; Niu, J.; Wang, E. An Intelligent Fault Diagnosis Expert System based on Fuzzy Neural Network. Vib. Shock 2017, 36, 164–171. [Google Scholar]
  14. Zhang, G.; Wang, H.; Zhang, T.Q. Stochastic Resonance of Coupled Time-delayed System with Fluctuation of Mass and Frequency and its Application in Bearing Fault Diagnosis. J. Cent. South Univ. 2021, 28, 2931–2946. [Google Scholar] [CrossRef]
  15. Yan, X.; Zhang, W.; Jia, M. A Bearing Fault Feature Extraction Method based on Optimized Singular Spectrum Decomposition and Linear Predictor. Meas. Sci. Technol. 2021, 32, 115023. [Google Scholar] [CrossRef]
  16. Ding, Y.; Zhuang, J.; Ding, P.; Jia, M. Self-supervised Pretraining via Contrast Learning for Intelligent Incipient Fault Detection of Bearings. Reliab. Eng. Syst. Saf. 2022, 218, 108126. [Google Scholar] [CrossRef]
  17. Zhao, X.; Qin, Y.; Fu, H.; Jia, L.; Zhang, X. Blind Source Extraction based on EMD and Temporal Correlation for Rolling Element Bearing Fault Diagnosis. Smart Resilient Transp. 2021, 3, 52–65. [Google Scholar] [CrossRef]
  18. Li, J.; Wu, C.; Song, R.; Li, Y.; Xie, W.; He, L.; Gao, X. Deep Hybrid 2-D-3-D CNN based on Dual Second-order Attention with Camera Spectral Sensitivity Prior for Spectral Super-resolution. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 623–634. [Google Scholar] [CrossRef]
  19. Wang, Y.; Menkovski, V.; Wang, H.; Du, X.; Pechenizkiy, M. Causal Discovery from Incomplete Data: A Deep Learning Approach. arXiv 2020, arXiv:2001.05343. [Google Scholar]
  20. Carlos, A.C.; Eduardo, R.M.; Juan, V.C.; Andrés, F.R. Granger-causality: An Efficient Single User Movement Recognition Using a Smartphone Accelerometer Sensor. Pattern Recognit. Lett. 2019, 125, 576–583. [Google Scholar]
  21. Wang, B.; Lei, Y.; Li, N.; Li, N. A Hybrid Prognostics Approach for Estimating Remaining Useful Life of Rolling Element Bearings. IEEE Trans. Reliab. 2018, 69, 401–412. [Google Scholar] [CrossRef]
  22. Diederik, P.; Jimmy, B. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  23. Wang, L. Support Vector Machines: Theory and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2005; Volume 177. [Google Scholar]
  24. Seber, G.A.; Lee, A.J. Linear Regression Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2003; Volume 330. [Google Scholar]
  25. Liaw, A.; Wiener, M. Classification and Regression by RandomForest. R News 2002, 2, 18–22. [Google Scholar]
  26. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  27. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encoder-decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  28. Li, J.; Liu, Y.; Li, Q. Intelligent Fault Diagnosis of Rolling Bearings Under Imbalanced Data Conditions Using Attention-based Deep Learning Method. Measurement 2022, 189, 110500. [Google Scholar] [CrossRef]
  29. Zhang, J.; Yi, S.; Liang, G.; Hongli, G.; Xin, H.; Hongliang, S. A New Bearing Fault Diagnosis Method based on Modified Convolutional Neural Networks. Chin. J. Aeronaut. 2020, 33, 439–447. [Google Scholar] [CrossRef]
  30. Zhang, W.; Zhang, F.; Chen, W.; Jiang, Y.; Song, D. Fault State Recognition of Rolling Bearing based Fully Convolutional Network. Comput. Sci. Eng. 2018, 21, 55–63. [Google Scholar] [CrossRef]
  31. Chen, X.; Zhang, B.; Gao, D. Bearing Fault Diagnosis base on Multi-scale CNN and LSTM Model. J. Intell. Manuf. 2021, 32, 971–987. [Google Scholar] [CrossRef]
  32. Che, C.; Wang, H.; Fu, Q.; Ni, X. Intelligent Fault Prediction of Rolling Bearing based on Gate Recurrent Unit and Hybrid Autoencoder. Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci. 2021, 235, 1106–1114. [Google Scholar] [CrossRef]
  33. Shao, H.; Hongkai, J.; Li, X.; Wu, S. Intelligent Fault Diagnosis of Rolling Bearing Using Deep Wavelet Auto-encoder with Extreme Learning Machine. Knowl.-Based Syst. 2018, 140, 1–14. [Google Scholar] [CrossRef]
Figure 1. The three stages of equipment running status.
Figure 1. The three stages of equipment running status.
Sensors 23 07040 g001
Figure 2. Overview of our proposed prediction framework.
Figure 2. Overview of our proposed prediction framework.
Sensors 23 07040 g002
Figure 3. The resampling of the signal sequences.
Figure 3. The resampling of the signal sequences.
Sensors 23 07040 g003
Figure 4. The effect of the space coefficient on accuracy.
Figure 4. The effect of the space coefficient on accuracy.
Sensors 23 07040 g004
Figure 5. Running time comparison of the different space coefficients.
Figure 5. Running time comparison of the different space coefficients.
Sensors 23 07040 g005
Figure 6. The effect of the sequence length on the accuracy.
Figure 6. The effect of the sequence length on the accuracy.
Sensors 23 07040 g006
Table 1. Accuracy comparison of Dataset 4.
Table 1. Accuracy comparison of Dataset 4.
MethodsDataset 1Dataset 2Dataset 3Dataset 4
SVM0.9010.8930.8610.855
RF0.8830.8760.8560.847
LR0.8960.8470.7680.749
LSTM0.8870.9100.8580.839
GRU0.8700.9020.8470.826
CNN0.8850.8820.8870.829
DFC-CNN0.8970.8960.8920.852
DA-RNN0.8920.8850.8890.876
CNN-LSTM0.8990.8940.8990.868
GRU-HA0.9160.9110.9000.892
DW-AE0.9220.9150.9040.896
CaFANet0.9300.9240.9130.902
Table 2. Performance comparison on Dataset 2.
Table 2. Performance comparison on Dataset 2.
MethodsAccPreReacll
SVM0.8930.8840.899
RF0.8760.8950.864
LR0.8470.8610.839
LSTM0.9100.9240.898
GRU0.9020.9350.877
CNN0.8820.9010.869
DFC-CNN0.8960.9060.892
DA-RNN0.8850.8960.872
CNN-LSTM0.8940.8990.886
GRU-HA0.9110.9340.890
DW-AE0.9150.9410.895
CaFANet0.9240.9490.905
Table 3. Causal analysis comparison of the time-domain features.
Table 3. Causal analysis comparison of the time-domain features.
FeatureDataset 1Dataset 2Dataset 3Dataset 4Overall
Var0.04130.03980.04210.04020.0409
RMS0.02790.02830.02650.02880.0279
AV0.03790.03690.03570.03700.0369
KU0.09570.09680.09120.08990.0934
SK0.02840.02370.02560.02740.0263
CF0.07150.07680.07420.07390.0741
MF0.07340.07960.07220.07850.0759
Table 4. Influence of components of the prediction model on accuracy.
Table 4. Influence of components of the prediction model on accuracy.
ModelsAcc
No causal analysis0.883
No global time attention0.866
No embedding time information0.806
Full model0.924
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gui, Z.; He, S.; Lin, Y.; Nan, X.; Yin, X.; Wu, C.Q. CaFANet: Causal-Factors-Aware Attention Networks for Equipment Fault Prediction in the Internet of Things. Sensors 2023, 23, 7040. https://doi.org/10.3390/s23167040

AMA Style

Gui Z, He S, Lin Y, Nan X, Yin X, Wu CQ. CaFANet: Causal-Factors-Aware Attention Networks for Equipment Fault Prediction in the Internet of Things. Sensors. 2023; 23(16):7040. https://doi.org/10.3390/s23167040

Chicago/Turabian Style

Gui, Zhenwen, Shuaishuai He, Yao Lin, Xin Nan, Xiaoyan Yin, and Chase Q. Wu. 2023. "CaFANet: Causal-Factors-Aware Attention Networks for Equipment Fault Prediction in the Internet of Things" Sensors 23, no. 16: 7040. https://doi.org/10.3390/s23167040

APA Style

Gui, Z., He, S., Lin, Y., Nan, X., Yin, X., & Wu, C. Q. (2023). CaFANet: Causal-Factors-Aware Attention Networks for Equipment Fault Prediction in the Internet of Things. Sensors, 23(16), 7040. https://doi.org/10.3390/s23167040

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop