Next Article in Journal
Development of a Sinusoidal Corrugated Dual-Axial Flexure Mechanism for Planar Nanopositioning
Previous Article in Journal
An Intelligent Tracking System for Moving Objects in Dynamic Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ECNN: Intelligent Fault Diagnosis Method Using Efficient Convolutional Neural Network

1
Department of Integrated Technology and Control Engineering, School of Aeronautics, Northwestern Polytechnical University, Xi’an 710072, China
2
School of Civil Aviation, Northwestern Polytechnical University, Xi’an 710072, China
3
Beijing Aerospace Systems Engineering Research Institute, Beijing 100076, China
4
The Fifth Electronics Research Institute of Ministry of Industry and Information Technology, Guangzhou 510610, China
5
China Institute of Marine Technology & Economy, Beijing 100081, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Actuators 2022, 11(10), 275; https://doi.org/10.3390/act11100275
Submission received: 20 August 2022 / Revised: 16 September 2022 / Accepted: 23 September 2022 / Published: 26 September 2022
(This article belongs to the Section Control Systems)

Abstract

:
With outstanding deep feature learning and nonlinear classification abilities, Convolutional Neural Networks (CNN) have been gradually applied to deal with various fault diagnosis tasks. Affected by variable working conditions and strong noises, the empirical datum always has different probability distributions, and then different data segments may have inconsistent contributions, so more attention should be assigned to the informative data segments. However, most of the CNN-based fault diagnosis methods still retain black-box characteristics, especially the lack of attention mechanisms and ignoring the special contributions of informative data segments. To address these problems, we propose a new intelligent fault diagnosis method comprised of an improved CNN model named Efficient Convolutional Neural Network (ECNN). The extensive view can cover the special characteristic periods, and the small view can locate the essential feature using Pyramidal Dilated Convolution (PDC). Consequently, the receptive field of the model can be greatly enlarged to capture the location information and excavate the remarkable informative data segments. Then, a novel residual network feature calibration and fusion (ResNet-FCF) block was designed, which uses local channel interactions and residual networks based on global channel interactions for weight-redistribution. Therefore, the corresponding channel weight is increased, which puts more attention on the information data segment. The ECNN model has achieved encouraging results in information extraction and feature channel allocation of the feature. Three experiments are used to test different diagnosis methods. The ECNN model achieves the highest average accuracy of fault diagnosis. The comparison results show that ECNN has strong domain adaptation ability, high stability, and superior diagnostic performance.

1. Introduction

In the era of Big Data, fault diagnosis has always played an important role in industrial production [1,2,3]. In practical environments, the equipment works with complex operating conditions and strong noise. Its core components (such as bearings, gears, and motors) occasionally fail, and some faults are challenging to locate using traditional methods, so intelligent fault diagnosis is essential to ensuring their safety and reliability [4,5,6,7,8,9].
Convolutional Neural Network (CNN) [10,11,12,13,14,15], one of the representative deep learning models, is increasingly applied to deal with fault diagnosis tasks relying on its outstanding advances in deep features learning and nonlinear classification. Zhang et al. [16] proposed a fault diagnosis method. The method uses multiple parallel convolutional layers to extract rich and complementary fault features effectively and then transforms the one-dimension signal into a two-dimension signal by continuous wavelet transforms. Huo et al. [17] extended a convolutional neural network with transfer learning, which can adaptively process one-dimensional vibration signals into two-dimension matrices and reduce the data distribution distance between the source and the target domains. Wang et al. [18] proposed a new multi-sensor information fusion method constructing the time-domain signals into a rectangular two-dimension matrix and then used an improved 2D-CNN to realize signal classification. Wen et al. [19] employed LeNet-5 to develop a new CNN and convert signals into 2D images through a conversion method. The proposed method easily captures the features. S. et al. [20] applied a one-dimensional convolutional neural network to motor fault diagnosis and proposed adding feature classification and extraction to a single learning body. Zhao et al. [21] proposed a normalized CNN for rolling bearing diagnosis with different severity and fault directions under data imbalance and variable conditions. Abdeljaber et al. [22] presented a novel, fast and accurate structural damage detection system using 1D Convolutional Neural Networks (CNNs) that has an inherent adaptive design to fuse feature extraction and classification blocks into a single and compact learning body. Dibaj et al. [23] proposed a fault diagnosis approach based on variational mode decomposition and CNN for rotating machinery. Janssens et al. [24] proposed a DL model for condition monitoring using CNN and proved that the feature learning method was significantly better than the feature extraction method in fault diagnosis of rotating machines. A Convolutional Neural Network (CNN) machine learning algorithm is proposed to classify gearbox faults in [25], and the learning features of the CNN filters are visualized to understand the physical fault diagnosis phenomena. Cheng et al. [26] implied an intelligent fault diagnosis method for the rotation machine based on a new continuous wavelet transforming the local binary convolutional neural network. Therefore, in the field of fault diagnosis, scholars have utilized a lot of convolutional neural networks to improve the accuracy of diagnosis and finally achieved excellent results. Although these studies verify the efficiency of CNN in the fault diagnosis field, the two following problems remain.
In the course of fault diagnosis, firstly, converting vibration signals to spectrogram requires a quantity of computation. Secondly, spectrograms are resized as small images before the training model to decrease computation cost time; some signal nature will be lost during the compression process. Therefore, the proposed method is dedicated to one-dimension data [27,28,29]. In addition, the traditional 1D-CNN uses the pooling layer to obtain the receptive field. Information on the Avg/Max pooling region is insufficient to capture the importance of the pooling feature. Although stride convolution could learn from neighbor features, it fails to model the importance of down sampling procedures adaptively and limits shift-invariance because it focuses only on one fixed location within each sliding window and discards the rest [30].
Since the equipment is usually affected by load changes and environmental noises, some scholars use dilated convolution to replace the pooling layer in standard convolution. This process can retain the timing relation of original signals and obtain larger-scale feature information, which helps improve the feature learning ability of neural networks. In [31], Su et al. used a dilated convolution deep belief network dynamic multilayer perceptron for bearing fault recognition under alterable running states. Zhao et al. [32] proposed a novel transfer learning framework based on a Deep Multi-Scale Convolutional Neural Network (MSCNN), in which dilated convolution and global average pooling were introduced to realize intelligent fault diagnosis of rolling bearings. Han et al. [33] used dilated convolutions to construct a Novel Multi-scale Dilated Convolutional Neural Network (NMDCNN) to enrich the field of view coverage. Meanwhile, piecewise cross-entropy is used to balance the misclassification cost between healthy samples and faulty samples. Chu et al. [34] advanced a novel multi-scale convolution model based on Multi-Dilation Rates And Multi-Attention Mechanism (MDRMA-MSCM) for mechanical fault diagnosis. Wang et al. [35] proposed Cascade Convolutional Neural Network (C-CNN) for fault diagnosis, and a cascade structure was built to avoid information loss. In conclusion, dilated convolution has made significant achievements in the field of fault diagnosis, and its advantages are as follows:
(1)
Dilated convolution replaces the pooling layer operation in feature processing. This process not only cannot reduce the receptive field of the network but also can leave the temporal relationship of the original signal completely, which is of great significance for mining the domain invariant characteristics of the signal.
(2)
The use of large convolution kernels in standard convolution will increase the computation amount. Compared with larger convolution kernels, dilated convolution has significantly less computation making the model more accurate in processing fault features.
However, dilated convolution still maintains the black-box feature, which assigns the same weight to different feature channels, and cannot adaptively adjust the corresponding weight according to the importance of the channel. In particular, the lack of attention mechanism ignores the exceptional contribution of the information data segment.
In recent years, the attention mechanism has been experimentally implied to solve this problem. It is inspired by the method that the human brain uses limited computation to obtain high-value information while processing information. Based on the idea, it has been applied in plenty of deep learning methods. Extracting the features of the original vibration signals, the attentional mechanism can dynamically enhance the weight of significant feature channels, thus improving diagnostic accuracy. Zhang et al. [36] proposed a fault diagnosis method, which is utilized to realize spatiotemporal feature fusion, where vibration signal fused features with attention weight. Li et al. [37] constructed a rolling bearing fault diagnosis model, which combines a Dual-stage Attention-based Recurrent Neural Network (DA-RNN) and Convolutional Block Attention Module (CBAM). Yang et al. [38] developed a method based on multilayer bidirectional gated recurrent units with an attention mechanism to access the interpretability of neural networks in fault diagnosis. Zhang et al. [39] proposed a Hybrid Attention improved Residual Network (HA-ResNet) based method to diagnose the fault of the wind turbines’ gearbox. This method highlights the essential frequency bands of wavelet coefficients and the fault features of convolution channels. Cao et al. [40] built a deep domain-adaptive multi-task learning model Y-Net, which is exploited to enable domain-adaptive diagnosis of faults in planetary gearboxes. The Squeeze and Excitation Residual (SE-Res) modules are utilized to reduce the redundancy of the model and improve the separability of deep features.
Attention mechanisms can be used to solve the problem that dilated convolution cannot adaptively obtain the particular weight of informative data segments. At the same time, the dilated convolution is introduced into the attention mechanism so that the model can gather a larger receptive field and extract more features. Inspired by the dilated convolution and attention mechanism, an ECNN model was proposed in this study. The model employs a novel Pyramidal Dilated Convolution to extract more valuable features, and the Residual Network Feature Calibration and Fusion (ResNet-FCF) block is implied to assign different weights to different feature channels. This article contributes:
(1)
In this model, the pyramidal dilatated convolution is designed to extract the features of data segments, which dramatically improves the receiving field and captures more features.
(2)
The ResNet-FCF is designed as a novel attention network architecture. Based on the local interaction learning scheme, ResNet-FCF introduces a residual network with one-dimensional convolution for global cross-channel interaction. The module assigns precise weight to the informative data segments.
(3)
The model has a good diagnosis effect under three practical examples.
The rest of this paper is organized as follows. The basic theory of the convolutional neural networks and residual networks are briefly introduced in Section 2. Moreover, it introduces an ECNN model and a novel fault diagnosis method. In Section 3, the experiment results are analyzed and discussed. Ultimately, the conclusions are presented in Section 4.

2. Materials and Methods

2.1. Convolutional Neural Networks

The CNN model is used for fault diagnosis, which can not only automatically extract fault features but also process a large amount of real-time data generated by the equipment.

2.1.1. Convolutional Layer

The main module of CNN [41] is the convolutional layer, which uses filters to perform a series of convolutional operations on the input data and output the corresponding features. Its mathematical model can be expressed as
y j k = X W = g W i j k · X i k 1 + b j k
where X represents the vibration signal, ∗ represents the convolution operation, k represents the k-layer network, W represents the convolution kernel, and y j k is the k-th layer output. g stands for nonlinear activation function, and b j k represents the bias vector.
When F ( x ) (the output of the residual block) is 0, the basic mapping function of the residual block becomes H ( x ) = x , which accomplishes the identity mapping of the input and output. At the same time, according to the rule of backpropagation, it can be known that the backpropagation gradient of the residual part is ( H ) / ( x ) + 1 . The model is merely required to minimize the residual map for approximating the identity map, which aims to ensure that the backpropagation gradient of the bottom layer is non-zero. Furthermore, the increase in the smoothness of interaction takes complete advantage of deep neural networks. The structure of the residual network is shown in Figure 1.

2.1.2. Pooling Layer

The pooling operation aims to reduce the dimension of the output of the previous layer. The essence is to effectively reduce the amount of calculation and retain important information, so the calculation resources and time are reduced in order to better preserve the data feature and make the model obtain a higher convergence speed. The pooling methods mainly include maximum pooling and average pooling. Maximum pooling p m a x and average pooling p a v g refer to computing the maximum value and average pooling in the pool window, respectively. The mathematical description of the two equations is shown as
p a v g l ( i , j ) = 1 W t = ( j 1 ) W + 1 j W a l ( i , t )
p max l ( i , j ) = max ( j 1 ) W + 1 i j w a l ( i , t )
where a l ( i , t ) is the activation value of the t neuron in frame i of layer l; W is the width of the pool window.

2.1.3. Fully Connected Layer

Through a series of convolution and pooling, depth feature extraction is performed layer by layer, then the features are flattened and input into the fully connected layer, and the equation is
y k = σ ( ( w f ) s m T + b f )
where w f and b f are the weight matrix and bias of the k-layer, respectively; y is the output of the fully connected layer. Usually, for obtaining the predicted output of the model, the softmax function is connected to realize the category classification after the fully connected layer, namely
f k ( y ) = exp y k / k = 1 C exp y k
where f k ( y ) is the predicted value of the softmax function for each category; C is the number of categories.

2.2. Residual Neural Network

As the deep neural network training process may lead to network degradation, He et al. [42] proposed a Residual Neural Network (ResNet), which builds a model by stacking residual blocks. The input of the residual block is x. After two convolutional layers, the basic mapping function H ( x ) is acquired. The residual block introduces a skip layer connection to reconstruct the learning process of stacked network layers, so that the residual mapping function F ( x ) of the network layer can fit H ( x ) x , where the expression of F ( x ) is as follows:
F ( x ) = W 2 ( R e L U ( W 1 ( x ) ) )
where W 1 and W 2 in represent the convolutional layer 1 and convolutional layer 2 in the residual block, and R e L U represents the nonlinear activation function.

2.3. ECNN Model

As the standard CNN model still maintains the black-box feature, it not only cannot capture the features of different scales but also ignores the unique contribution of special data segments. In order to solve the problem, we start from three aspects: raw vibration signal using Pyramidal Dilated Convolution, ResNet-FCF block construction, and the general procedure of the proposed method.
Compared with CNN, this model adopts different architectural approaches in processing original signals, feature extraction and weight allocation of signal channels, as shown in Figure 2. The ECNN model uses dilated convolution to replace the pooling layer to avoid the loss of important features and has a larger receptive field under the same parameters with more useful features extracted. After the dilated convolution operation, the global average pooling (GAP) is used to obtain the descriptor of each feature channel. Then the local channel interaction and the global channel interaction based on the residual network are fused for weight redistribution, strengthening the crucial features and suppressing the irrelevant features.

2.3.1. Pyramidal Dilated Convolution

Dilated convolution has huge advantages over standard convolution [43]. In this study, Pyramidal Dilated Convolution (PDC) is proposed, which has an adaptive receptive field with the same parameters. When a large amount of information is acquired, the parameters can be kept unchanged, resulting in a decrease in computing resources. Assuming that x ( a ) represents the one-dimensional input, A shows the data feature size, the dilation rate is r, the 1D convolution kernel is w ( i ) , and y ( a ) represents the output feature of data operation. The relationship equation between them is as follows:
y ( a ) = i = 1 A x ( a + r × i ) × w ( i ) .
The Pyramidal Dilated Convolution can not only expand the reception field of the original network but also compensate for the loss caused by under-sampling, as demonstrated in Figure 3. Therefore, it is feasible to replace standard convolution in neural networks with Pyramidal Dilated Convolution. Depending on the PDC with different dilation rates, the receptive field becomes enlarged, and in the meantime, features of data segments permit more access.
Under the condition that the size of the convolution kernel is the same, the receptive field of the Pyramidal Dilated Convolution in the i-th convolutional layer rises exponentially. The dilation rate d and the size of the convolution kernel jointly determine the receptive field size of the i-th Pyramidal Dilated Convolutional network. Equation (8) is the receptive field of dilation:
S 1 = kernel _ size
S i = S i 1 + ( kernel _ size i 1 ) d i
where S i stands for the receptive field of the i-th layer, and kernel _ size i and d i represent the size of the convolution kernel and dilation rate of the i-th layer network, respectively.

2.3.2. ResNet-FCF Block

In recent years, it has been proved that channel attention exists with great potential in improving the performance of convolutional neural networks. In this study, we propose a novel ResNet-FCF block with the structure shown in Figure 4.
The specific working process of the module is initially to perform GAP [44] after convolution transformation, whose purpose is to obtain the global information of each channel. The statistic z R C is generated by shrinking the dimension H × W of the input signal, and the p-th element of z can be expressed as:
z p = F s q u p = 1 H × W i = 1 H j = 1 W u p ( i , j )
where z p can be interpreted as a collection of local features, the statistical information of these local features can express the entire signal and have a global receptive field. The statistic z does not require dimensionality reduction, and the attention weight of each channel can be obtained in the following way:
ρ = σ ( W k z )
where ρ is the weight, and W k contains k × C parameters, which is defined as:
w 1 1 w 1 k 0 0 0 0 w 2 2 w 2 k + 1 0 0 0 0 0 w C C k + 1 w C C
σ represents the sigmoid [45] nonlinear activation function:
σ ( x ) = 1 / ( t ) x e
Equation (11) avoids complete independence between different channels and ensures efficiency and effectiveness while realizing partial cross-channel interaction. The weight of z j is calculated only by considering the interaction between z j and its k neighboring elements:
ρ i = σ j = 1 k w i j y i j , y i j Ω i k
where Ω i k represents the set of k channels adjacent to y i . In order to reduce the complexity of the model, all channels can share the same parameters, namely:
ρ i = σ j = 1 k w i y i j , y i j Ω i k
Equation (11) can be further simplified as a one-dimensional convolution operation:
φ = σ Conv 1 k ( z )
where Conv 1 k represents the one-dimensional convolution, and k is the size of the corresponding convolution kernel. It is verified through experiments that the value of k is related to the number of channels C. The simplest mapping is a linear function.
However, the relationships characterized by linear functions are excessively limited. Exponential functions are often used to deal with this nonlinear mapping relationship. The expression is as follows:
C = 2 ( γ n k b )
the number of channels is C, the expression of k can be inversely deduced as
k = log 2 C γ + b γ
in the equation: γ , b are adjustment parameters, which can be adjusted according to specific conditions.
F s c ( · ) multiplies the input x by ρ and maps it to the output with the same dimension. The output is defined as follows:
F s c ( x , ρ ) = Y
The process uses a local feature learning scheme to assign weights, then the skip part of the residual network is changed to convolution, which builds the global feature interaction. The output U of the residual block is defined as follows:
U = Conv 2 1 ( X ) + Y
where Conv 2 1 represents the convolution whose kernel size is 1.

2.4. Intelligent Fault Diagnosis Process

In this study, a novel ECNN for intelligent fault diagnosis is developed. The specific framework of diagnosis is shown in Figure 5. The general process is summarized as follows:
  • Step 1: Collect vibration signal data from equipment.
  • Step 2: The raw signal is divided. Firstly, a certain proportion of the former is the training signal, and the rest is the test signal. Then, several samples are randomly selected from the corresponding signals to form a training set and a test set.
  • Step 3: Construct an ECNN network model and select appropriate parameters.
  • Step 4: Bring the training samples into the model for training. If the model does not converge, return to the previous step to redesign.
  • Step 5: The performance of the proposed method is verified using test samples.

3. Results

In this section, the Windows10 operating system is used. The Random Access Memory (RAM) and graphics process unit are 16GB and GeForce GTX 1060, respectively. The ECNN test model was built and trained under the Keras framework. Keras is an open-source artificial neural network library written in Python, which can be used as a high-level application interface for Tensorflow, Microsoft-CNTK and Theano to design, debug, evaluate, apply and visualize deep learning models. Three examples are given to verify that the ECNN fault diagnosis model can extract a considerable number of data features with equivalent calculations and assign much weight to essential features and little to irrelevant features. In the experiment, three different data sets are used to verify its effectiveness: Case Western Reserve University (CWRU) rolling bearing, National Aeronautics and Space Administration (NASA) rolling bearing, and the Permanent Magnet Synchronous Motor (PMSM) of our lab, respectively.

3.1. The CWRU Rolling Bearing Example

The CWRU bearing data set is recognized as an authoritative fault diagnosis standard data set. In order to objectively compare the method with others, this paper selects the CWRU bearing data set for algorithm verification. In the experiment, the drive end data are selected. The bearing has three types of fault conditions: the inner ring, outer ring and rolling element. The depth and location of each type of fault are different, as shown in Table 1. There are four hundred samples collected for each condition. The load of this experiment is variable and can be divided into HP1, HP2 and HP3 load data. The experimental data contains 16 kinds of conditions. The signals are shown in Figure 6.

3.1.1. Model Structure

In the experiment, six traditional methods and four convolutional neural networks were selected for the comparison. The conventional method extracts 11 time-domain features and 13 frequency-domain features from the samples and then classifies them using BPNN and SVM. More details about the eleven TD features and thirteen FD features can be seen in [46,47], respectively. CNN, dilated convolution and SE (Squeeze-and-Excitation) CNN are taken as the comparative experimental models of deep convolutional neural networks. The first layer of the ECNN model uses a one-dimensional way for data processing, and then different dilated rates are set for the other eight layers. This process can replace the pooling layer in standard CNN so more valuable features can be extracted while the receptive field increases. The designed ResNet-FCF block assigns different weights to feature channels after each layer. The model parameters of the deep learning network ECNN are shown in Table 2.

3.1.2. Diagnosis Results of Different Methods

In order to obtain more accurate results, the experiment was divided into ten groups. The results show that the performance of BPNN and SVM depends on feature extraction to a large extent. When the selected input becomes sensitive features, the diagnostic results are further improved to 79.63% and 87.14%, respectively, but this is expensive and time-consuming. The accuracy of the proposed method is 99.12% higher than that of the nine methods, as shown in Table 3 and Figure 7.

3.1.3. Visualization of the Fault Diagnosis Process

ECNN is set for fault diagnosis, in which nine layers exist. Each layer initially uses different dilation rates for feature extraction and then utilizes the ResNet-FCF block to assign appropriate weights for the extracted features effectively. Thereby the informative data segments will receive more attention and their corresponding channel weights increase. This paper takes advantage of the t-SNE dimensionality reduction technique to visualize the features in each layer. The diagram of the structure is shown in Figure 8.

3.1.4. Weight Comparison of ECNN and SECNN

To explore the role of ResNet-FCF in every convolutional layer of ECNN, we visualize the weights generated by each convolutional layer and then compare the effects of ResNet-FCF and SE blocks at each layer in detail. As shown in Figure 9, the results indicate that when different types of samples are input, the weight generated by the fault diagnosis model is more related to the category information. The combination of local cross-channel and global cross-channel interactions in the ResNet-FCF block significantly improves the ability of channel attention.

3.1.5. The Influence of Hyperparameters

The structural parameters of the model are determined by experiments. Two key parameters k (convolution kernel size) and C (number of channels) are selected through cross-validation, as shown in Figure 10. The results of the 3D hyperparametric graph show that the accuracy of the model is the highest when the size of the convolution kernel is sixty-four, and the number of channels is three.

3.1.6. The Influence of Different Segmentation Ratios

In the experiment, with the increased number of training samples, the accuracy and experiment time of each network structure also increase. As seen in Table 4, under the conditions of different sample segmentation ratios, the ECNN fault diagnosis model has achieved the greatest results.

3.1.7. The Influence of Different SNR

To further research the performance of the model, various levels of white noise were added to the traditional learning model and the deep learning model, respectively. The results show that the traditional methods are less competitive compared with CNN in terms of noise-resistance. The model behaves brilliantly in terms of robustness compared with other deep learning networks. Table 5 shows the accuracy of each model under noise conditions. Figure 11 shows the noise-resistance curve.

3.1.8. The Influence of Variable Load

In order to analyze the generalization ability of these models, every method is required to be trained under different loads, and another load will be used as a test set. The experiment results are shown in Figure 12 and Table 6.

3.2. The NASA Rolling Bearing Example

In this section, the NASA dataset is further used to demonstrate the superiority of the proposed method in feature extraction and contributions from different data segments. From the beginning to the end of the data collection, there are four cases in the data set: normal, inner ring failure, outer ring failure and rolling element failure. The test time of normal, inner ring and rolling elements are all thirty-five days, with the exception of thirty days for the test of the outer ring. The data set extracted in this experiment was collected after frequent failures, and the data of the last five days were taken as samples shown in Figure 13 and Table 7.
This experiment obtained 300 training samples and 100 testing samples from each condition, and the length of each sample was set to 1024. Ten experiments were conducted for each one of the four deep learning networks. The experiment results are shown in Figure 14 and Table 8. Compared with other deep learning methods, the proposed method has obvious advantages and enables to accurately identify different types of faults.
The following conclusions can be drawn: different changes in load have different influences on signals. Compared with other deep learning methods, the ECNN fault diagnosis model has a high average accuracy rate that reaches 97.19% under variable working conditions. The results show that the model has a strong generalization ability and domain adaptability.

3.3. The Example of PMSM

Further study on the electrical systems of Permanent Magnet Synchronous Motors (PMSM) will be carried out to verify the transferability and extensibility of the ECNN-based method.
The experimental platform comprises the following parts: upper computer, electromechanical actuator and its controller, sensor, data acquisition system, loading system and power supply system. The structure of the experimental platform is shown in Figure 15.

3.3.1. The Faults of Motor

In order to guarantee the safety of the test and prevent irreversible damage to the motor, mainly the inter-turn short circuit fault of the motor is simulated in this study, as shown in Figure 16. The inter-turn short circuit fault is mainly reflected in the change in the winding resistance value. Therefore, the three-phase winding of the motor is externally connected, in which the resistance of the two-phase windings has the same value (the other phase is not a series of resistance or a different resistance value of the resistance) to simulate the three-phase winding asymmetry. At the same time, the three-phase current and speed signals of the motor under different fault depths are collected.
As shown in Figure 16, number 1 represent the controller of the output bus, numbers 2, 3, and 4 represent the three current sensor output signal, number 5 on behalf of the 5V DC power supply, numbers 6, 7 and 8 are used to measure the a, b, and c three-phase current of the current sensor respectively, number 9, 10 and 11 are concatenated in an a, b, and c three-phase resistor respectively, and number 12 represents the input bus at the end of the motor end.

3.3.2. Influence of Short Circuit on Three-Phase Current of Motor

The 1%, 2%, 4%, 8% and 16% inter-turn short circuit faults were injected for a PMSM stator winding. (a), (b), (c), (d), (e) and (f) are the three-phase current curves of the motor under different working conditions. The red line is the a-phase current curve, the purple line is the B-phase current curve, and the blue line is the C-phase current curve. As shown in Figure 17.

3.3.3. Verification of Different Deep Learning Algorithms

Four deep learning models were trained ten times to obtain average accuracy. As shown in Figure 18. The results show that ECNN can improve the diagnosis ability of small faults and has a good recognition ability for inter-turn short circuit faults with the same fault type and different fault depths. In the experiment, the accuracy of CNN is 93.38%, while the accuracy of optimized DCNN is improved to 96.61%. The recognition accuracy of the algorithm combining SECNN is 97.25%, while the accuracy of the ECNN network is the highest, reaching 99.87%. Deep network ECNN has great advantages in feature extraction and state recognition. The confusion matrix of the four deep learning models is shown in Figure 19.

4. Conclusions

This paper presents a new ECNN model for fault diagnosis. Firstly, a novel pyramidal dilated convolution method is proposed. It not only expands the receptive field of the model but also plays an important role in feature extraction. Secondly, aiming at the problem that all feature channels of fault information after pyramid dilated convolution are treated equally, a novel ResNet-FCF block is proposed. The ResNet-FCF block can effectively assign appropriate weights to each channel through local cross-channel interaction and global cross-channel interaction learning. The three examples’ results show the proposed method’s effectiveness and superiority. However, there is still room for improvement in the parameter tuning of this model, which can be further optimized on the number of convolution kernels on the convolution layer. Meanwhile, the recognition rate from the “3” domain to the “1” domain needs to be improved. In the future, other algorithms to improve the adaptive ability of the domain can be added to further improve the fault identification ability under complicated conditions.

Author Contributions

Conceptualization, Q.H.; methodology, Q.H.; software, Q.H.; validation, C.Z. (Chao Zhang), L.C. and Z.L.; formal analysis, C.Z. (Chao Zhang); investigation, Q.H.; data curation, C.Z. (Chaoyi Zhang) and K.Y.; writing—original draft preparation, Q.H.; writing—-review and editing, C.Z. (Chaoyi Zhang) and K.Y.; visualization, Q.H.; supervision, C.Z. (Chao Zhang), L.C. and Z.L.; project administration, C.Z. (Chao Zhang); funding acquisition, C.Z. (Chao Zhang). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Projects of China (JSZL2020203B004, JSZL202160113001 and JCKY2021607B018).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, W.; Chen, Z.; He, G. A novel weighted adversarial transfer network for partial domain fault diagnosis of machinery. IEEE Trans. Ind. Inform. 2020, 17, 1753–1762. [Google Scholar] [CrossRef]
  2. Yan, X.; Liu, Y.; Xu, Y.; Jia, M. Multichannel fault diagnosis of wind turbine driving system using multivariate singular spectrum decomposition and improved Kolmogorov complexity. Renew. Energy 2021, 170, 724–748. [Google Scholar] [CrossRef]
  3. Valizadeh Yaghmourali, Y.; Ahmadi, N.; Abbaspour-sani, E. A thermal-calorimetric gas flow meter with improved isolating feature. Microsyst. Technol. 2017, 23, 1927–1936. [Google Scholar] [CrossRef]
  4. Shashank, R.; Rai, A.A.; Rao, R.B. Intelligent fault diagnosis of Rolling-element bearing using 1-D Deep Convolutional Neural Networks. Int. J. COMADEM 2021, 24, 11–18. [Google Scholar]
  5. Eren, L.; Ince, T.; Kiranyaz, S. A generic intelligent bearing fault diagnosis system using compact adaptive 1D CNN classifier. J. Signal Process. Syst. 2019, 91, 179–189. [Google Scholar] [CrossRef]
  6. Ahmed, Q.; Khan, F.; Ahmed, S. Improving safety and availability of complex systems using a risk-based failure assessment approach. J. Loss Prev. Process. Ind. 2014, 32, 218–229. [Google Scholar] [CrossRef]
  7. Khorram, A.; Khalooei, M.; Rezghi, M. End-to-end CNN+ LSTM deep learning approach for bearing fault diagnosis. Appl. Intell. 2021, 51, 736–751. [Google Scholar] [CrossRef]
  8. Vyas, N.S.; Satishkumar, D. Artificial neural network design for fault identification in a rotor-bearing system. Mech. Mach. Theory 2001, 36, 157–175. [Google Scholar] [CrossRef]
  9. Arunthavanathan, R.; Khan, F.; Ahmed, S.; Imtiaz, S.; Rusli, R. Fault detection and diagnosis in process system using artificial intelligence-based cognitive technique. Comput. Chem. Eng. 2020, 134, 106697. [Google Scholar] [CrossRef]
  10. Zhao, B.; Zhang, X.; Li, H.; Yang, Z. Intelligent fault diagnosis of rolling bearings based on normalized CNN considering data imbalance and variable working conditions. Knowl.-Based Syst. 2020, 199, 105971. [Google Scholar] [CrossRef]
  11. Miao, M.; Sun, Y.; Yu, J. Deep sparse representation network for feature learning of vibration signals and its application in gearbox fault diagnosis. Knowl.-Based Syst. 2022, 240, 108116. [Google Scholar] [CrossRef]
  12. Su, K.; Liu, J.; Xiong, H. Hierarchical diagnosis of bearing faults using branch convolutional neural network considering noise interference and variable working conditions. Knowl.-Based Syst. 2021, 230, 107386. [Google Scholar] [CrossRef]
  13. Wang, X.; Mao, D.; Li, X. Bearing fault diagnosis based on vibro-acoustic data fusion and 1D-CNN network. Measurement 2021, 173, 108518. [Google Scholar] [CrossRef]
  14. He, C.; Ge, D.; Yang, M.; Yong, N.; Wang, J.; Yu, J. A data-driven adaptive fault diagnosis methodology for nuclear power systems based on NSGAII-CNN. Ann. Nucl. Energy 2021, 159, 108326. [Google Scholar] [CrossRef]
  15. Long, Y.; Zhou, W.; Luo, Y. A fault diagnosis method based on one-dimensional data enhancement and convolutional neural network. Measurement 2021, 180, 109532. [Google Scholar] [CrossRef]
  16. Zhang, K.; Wang, J.; Shi, H.; Zhang, X.; Tang, Y. A fault diagnosis method based on improved convolutional neural network for bearings under variable working conditions. Measurement 2021, 182, 109749. [Google Scholar] [CrossRef]
  17. Huo, C.; Jiang, Q.; Shen, Y.; Qian, C.; Zhang, Q. New transfer learning fault diagnosis method of rolling bearing based on ADC-CNN and LATL under variable conditions. Measurement 2022, 188, 110587. [Google Scholar] [CrossRef]
  18. Wang, J.; Wang, D.; Wang, S.; Li, W.; Song, K. Fault diagnosis of bearings based on multi-sensor information fusion and 2D convolutional neural network. IEEE Access 2021, 9, 23717–23725. [Google Scholar] [CrossRef]
  19. Wen, L.; Li, X.; Gao, L.; Zhang, Y. A new convolutional neural network-based data-driven fault diagnosis method. IEEE Trans. Ind. Electron. 2017, 65, 5990–5998. [Google Scholar] [CrossRef]
  20. Yang, S.; Sun, X.; Chen, D. Bearing fault diagnosis of two-dimensional improved Att-CNN2D neural network based on Attention mechanism. In Proceedings of the 2020 IEEE International Conference on Artificial Intelligence and Information Systems (ICAIIS), Dalian, China, 20–22 March 2020; pp. 81–85. [Google Scholar]
  21. Zhao, D.; Zhang, H.; Liu, S.; Wei, Y.; Xiao, S. Deep rational attention network with threshold strategy embedded for mechanical fault diagnosis. IEEE Trans. Instrum. Meas. 2021, 70, 1–15. [Google Scholar] [CrossRef]
  22. Abdeljaber, O.; Avci, O.; Kiranyaz, S.; Gabbouj, M.; Inman, D.J. Real-time vibration-based structural damage detection using one-dimensional convolutional neural networks. J. Sound Vib. 2017, 388, 154–170. [Google Scholar] [CrossRef]
  23. Dibaj, A.; Ettefagh, M.M.; Hassannejad, R.; Ehghaghi, M.B. A hybrid fine-tuned VMD and CNN scheme for untrained compound fault diagnosis of rotating machinery with unequal-severity faults. Expert Syst. Appl. 2021, 167, 114094. [Google Scholar] [CrossRef]
  24. Janssens, O.; Slavkovikj, V.; Vervisch, B.; Stockman, K.; Loccufier, M.; Verstockt, S.; Van de Walle, R.; Van Hoecke, S. Convolutional neural network based fault detection for rotating machinery. J. Sound Vib. 2016, 377, 331–345. [Google Scholar] [CrossRef]
  25. Senanayaka, J.S.L.; Van Khang, H.; Robbersmvr, K.G. CNN based Gearbox Fault Diagnosis and Interpretation of Learning Features. In Proceedings of the 2021 IEEE 30th International Symposium on Industrial Electronics (ISIE), Kyoto, Japan, 20–23 June 2021; pp. 1–6. [Google Scholar]
  26. Cheng, Y.; Lin, M.; Wu, J.; Zhu, H.; Shao, X. Intelligent fault diagnosis of rotating machinery based on continuous wavelet transform-local binary convolutional neural network. Knowl.-Based Syst. 2021, 216, 106796. [Google Scholar] [CrossRef]
  27. Liu, P.Y.; Chen, C.C.; Liong, S.T.; Tsai, M.H.; Hsieh, P.C.; Wang, K.C. Intelligent Fault Diagnosis Based on Multi-Resolution and One-Dimension Convolutional Neural Networks. In Proceedings of the 2021 International Conference on System Science and Engineering (ICSSE), Ho Chi Minh City, Vietnam, 26–28 August 2021; pp. 319–322. [Google Scholar]
  28. Gong, W.; Wang, Y.; Zhang, M.; Mihankhah, E.; Chen, H.; Wang, D. A Fast Anomaly Diagnosis Approach Based on Modified CNN and Multisensor Data Fusion. IEEE Trans. Ind. Electron. 2022, 69, 13636–13646. [Google Scholar] [CrossRef]
  29. Song, X.; Cong, Y.; Song, Y.; Chen, Y.; Liang, P. A bearing fault diagnosis model based on CNN with wide convolution kernels. J. Ambient. Intell. Humaniz. Comput. 2022, 13, 4041–4056. [Google Scholar] [CrossRef]
  30. Li, X.; Wu, Y.; Fu, Y.; Tang, C.; Zhang, L. Neighbour Feature Attention-Based Pooling. Neurocomputing 2022, 501, 285–293. [Google Scholar] [CrossRef]
  31. Su, H.; Yang, X.; Xiang, L.; Hu, A.; Xu, Y. A novel method based on deep transfer unsupervised learning network for bearing fault diagnosis under variable working condition of unequal quantity. Knowl.-Based Syst. 2022, 242, 108381. [Google Scholar] [CrossRef]
  32. Zhao, B.; Zhang, X.; Zhan, Z.; Pang, S. Deep multi-scale convolutional transfer learning network: A novel method for intelligent fault diagnosis of rolling bearings under variable working conditions and domains. Neurocomputing 2020, 407, 24–38. [Google Scholar] [CrossRef]
  33. Han, S.; Shao, H.; Zhao, R. Gearbox Fault Diagnosis using Novel Dilated CNN and Piecewise Loss Function under Unbalanced Data. In Proceedings of the 2021 Global Reliability and Prognostics and Health Management (PHM-Nanjing), Nanjing, China, 15–17 October 2021; pp. 1–5. [Google Scholar]
  34. Chu, C.; Ge, Y.; Qian, Q.; Hua, B.; Guo, J. A novel multi-scale convolution model based on multi-dilation rates and multi-attention mechanism for mechanical fault diagnosis. Digit. Signal Process. 2022, 122, 103355. [Google Scholar] [CrossRef]
  35. Wang, F.; Liu, R.; Hu, Q.; Chen, X. Cascade convolutional neural network with progressive optimization for motor fault diagnosis under nonstationary conditions. IEEE Trans. Ind. Inform. 2020, 17, 2511–2521. [Google Scholar] [CrossRef]
  36. Zhang, X.; He, C.; Lu, Y.; Chen, B.; Zhu, L.; Zhang, L. Fault diagnosis for small samples based on attention mechanism. Measurement 2022, 187, 110242. [Google Scholar] [CrossRef]
  37. Li, J.; Liu, Y.; Li, Q. Intelligent fault diagnosis of rolling bearings under imbalanced data conditions using attention-based deep learning method. Measurement 2022, 189, 110500. [Google Scholar] [CrossRef]
  38. Yang, Z.B.; Zhang, J.P.; Zhao, Z.B.; Zhai, Z.; Chen, X.F. Interpreting network knowledge with attention mechanism for bearing fault diagnosis. Appl. Soft Comput. 2020, 97, 106829. [Google Scholar] [CrossRef]
  39. Zhang, K.; Tang, B.; Deng, L.; Liu, X. A hybrid attention improved ResNet based fault diagnosis method of wind turbines gearbox. Measurement 2021, 179, 109491. [Google Scholar] [CrossRef]
  40. Cao, X.; Chen, B.; Zeng, N. A deep domain adaption model with multi-task networks for planetary gearbox fault diagnosis. Neurocomputing 2020, 409, 173–190. [Google Scholar] [CrossRef]
  41. Zhang, C.; Feng, J.; Hu, C.; Liu, Z.; Cheng, L.; Zhou, Y. An Intelligent Fault Diagnosis Method of Rolling Bearing Under Variable Working Loads Using 1-D Stacked Dilated Convolutional Neural Network. IEEE Access 2020, 8, 63027–63042. [Google Scholar] [CrossRef]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. He, J.; Wu, P.; Tong, Y.; Zhang, X.; Lei, M.; Gao, J. Bearing Fault Diagnosis via Improved One-Dimensional Multi-Scale Dilated CNN. Sensors 2021, 21, 7319. [Google Scholar] [CrossRef]
  44. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  45. Liu, Y.; Yang, Q.; An, D.; Nai, Y.; Zhang, Z. An improved fault diagnosis method based on deep wavelet neural network. In Proceedings of the 2018 Chinese Control And Decision Conference (CCDC), Shenyang, China, 9–11 June 2018; pp. 1048–1053. [Google Scholar]
  46. Lei, Y.; He, Z.; Zi, Y. A new approach to intelligent fault diagnosis of rotating machinery. Expert Syst. Appl. 2008, 35, 1593–1600. [Google Scholar] [CrossRef]
  47. Lei, Y.; He, Z.; Zi, Y. EEMD method and WNN for fault diagnosis of locomotive roller bearings. Expert Syst. Appl. 2011, 38, 7334–7341. [Google Scholar] [CrossRef]
Figure 1. Residual neural network.
Figure 1. Residual neural network.
Actuators 11 00275 g001
Figure 2. ECNN model of structures.
Figure 2. ECNN model of structures.
Actuators 11 00275 g002
Figure 3. Comparison of two convolutional networks.
Figure 3. Comparison of two convolutional networks.
Actuators 11 00275 g003
Figure 4. The construction of a ResNet-FCF block.
Figure 4. The construction of a ResNet-FCF block.
Actuators 11 00275 g004
Figure 5. The structure of intelligent fault diagnosis.
Figure 5. The structure of intelligent fault diagnosis.
Actuators 11 00275 g005
Figure 6. Raw waveforms of the bearing conditions.
Figure 6. Raw waveforms of the bearing conditions.
Actuators 11 00275 g006
Figure 7. Diagnosis results of the ten trials using different methods.
Figure 7. Diagnosis results of the ten trials using different methods.
Actuators 11 00275 g007
Figure 8. The proposed method fault classification results.
Figure 8. The proposed method fault classification results.
Actuators 11 00275 g008
Figure 9. Visual comparison of weights for ResNET-FCF and SE blocks.
Figure 9. Visual comparison of weights for ResNET-FCF and SE blocks.
Actuators 11 00275 g009
Figure 10. The effect of C and k on accuracy.
Figure 10. The effect of C and k on accuracy.
Actuators 11 00275 g010
Figure 11. Accuracy curve of each method under different SNRs.
Figure 11. Accuracy curve of each method under different SNRs.
Actuators 11 00275 g011
Figure 12. Average accuracy under different loads.
Figure 12. Average accuracy under different loads.
Actuators 11 00275 g012
Figure 13. Collection of NASA datasets.
Figure 13. Collection of NASA datasets.
Actuators 11 00275 g013
Figure 14. Diagnosis results of the ten trials using different methods.
Figure 14. Diagnosis results of the ten trials using different methods.
Actuators 11 00275 g014
Figure 15. PMSM experimental setup.
Figure 15. PMSM experimental setup.
Actuators 11 00275 g015
Figure 16. Failure testing of PMSM.
Figure 16. Failure testing of PMSM.
Actuators 11 00275 g016
Figure 17. Three phase current curve of the motor under different conditions.
Figure 17. Three phase current curve of the motor under different conditions.
Actuators 11 00275 g017aActuators 11 00275 g017b
Figure 18. Diagnosis results of the four trials using different methods.
Figure 18. Diagnosis results of the four trials using different methods.
Actuators 11 00275 g018
Figure 19. The confusion matrix of the four deep learning models.
Figure 19. The confusion matrix of the four deep learning models.
Actuators 11 00275 g019
Table 1. Description of the bearing operation conditions.
Table 1. Description of the bearing operation conditions.
Bearing Operation ConditionsFault DiameterSize of TrainingTesting SamplesLabel
Normal-3001000
Ball fault0.0073001001
Ball fault0.0143001002
Ball fault0.0213001003
Ball fault0.0283001004
Inner race fault0.0073001005
Inner race fault0.0143001006
Inner race fault0.0213001007
Inner race fault0.0283001008
Outer race fault 6 o’clock0.0073001009
Outer race fault 6 o’clock0.01430010010
Outer race fault 6 o’clock0.02130010011
Outer race fault 3 o’clock0.00730010012
Outer race fault 3 o’clock0.02130010013
Outer race fault 12 o’clock0.00730010014
Outer race fault 12 o’clock0.02130010015
Table 2. Parameters of the ECNN model.
Table 2. Parameters of the ECNN model.
Network LayerSize/StrideKernelsDilate FactorOutput ShapeReceptive FieldPadding
Input1024 × 11
Conv 13 × 1/1 × 16411024 × 643Yes
ResNet-FCF---1024 × 643-
Conv 23 × 1/1 × 16421020 × 647No
ResNet-FCF---1020 × 647-
Conv 33 × 1/1 × 16441012 × 6415No
ResNet-FCF---1012 × 6415-
Conv 43 × 1/1 × 1648996 × 6431No
ResNet-FCF---996 × 6431-
Conv 53 × 1/1 × 16416964 × 6463No
ResNet-FCF---964 × 6463-
Conv 63 × 1/1 × 16432900 × 64127No
ResNet-FCF---900 × 64127-
Conv 73 × 1/1 × 16464772 × 64255No
ResNet-FCF---772 × 64255-
Conv 83 × 1/1 × 164516 × 64511No
ResNet-FCF---516 × 64511-
Conv 93 × 1/1 × 1642564 × 641023No
ResNet-FCF---4 × 641023-
Flatten128
FC128128
SoftMax1616
Table 3. Parameters of the ECNN model.
Table 3. Parameters of the ECNN model.
Network LayerSize/StrideKernelsDilate FactorOutput ShapeReceptive FieldPadding
Input1024 × 11
Conv 13 × 1/1 × 16411024 × 643Yes
ResNet-FCF---1024 × 643-
Conv 23 × 1/1 × 16421020 × 647No
ResNet-FCF---1020 × 647-
Conv 33 × 1/1 × 16441012 × 6415No
ResNet-FCF---1012 × 6415-
Conv 43 × 1/1 × 1648996 × 6431No
ResNet-FCF---996 × 6431-
Conv 53 × 1/1 × 16416964 × 6463No
ResNet-FCF---964 × 6463-
Conv 63 × 1/1 × 16432900 × 64127No
ResNet-FCF---900 × 64127-
Conv 73 × 1/1 × 16464772 × 64255No
ResNet-FCF---772 × 64255-
Conv 83 × 1/1 × 164516 × 64511No
ResNet-FCF---516 × 64511-
Conv 93 × 1/1 × 1642564 × 641023No
ResNet-FCF---4 × 641023-
Flatten128
FC128128
SoftMax1616
Table 4. Accuracy of different segmentation ratios.
Table 4. Accuracy of different segmentation ratios.
MethodsBP-rBP-TBP-FSVM-rSVM-TSVM-FCNNDCNNSECNNECNN
350/50Correct samples1281611410150379787137951571715782156401581815818
Accuracy82.96%75.67%95.33%71.88%88.56%98.36%98.17%99.02%98.98%99.34%
Accuracy std1.05%2.81%0.36%0.55%1.11%0.27%0.38%0.32%0.70%0.40%
Time6.956.297.2912.372.561.8630.8359.9744.6971.68
325/75Correct samples1281611410150379787137951571715782156401581815818
Accuracy80.72%74.58%94.61%71.25%87.58%98.33%97.75%98.87%98.79%99.27%
Accuracy std0.84%3.50%0.34%1.47%1.00%0.31%1.21%0.60%0.58%0.42%
Time6.316.216.9312.172.51.8128.3858.6941.770.87
300/100Correct samples1281611410150379787137951571715782156401581815818
Accuracy79.63%71.91%93.17%71.20%87.14%98.15%95.47%97.07%98.72%99.12%
Accuracy std1.38%1.99%2.84%1.05%1.17%0.29%2.08%1.50%0.41%0.37%
Time6.16.146.812.092.41.7828.0552.5638.3662.89
275/125Correct samples1281611410150379787137951571715782156401581815818
Accuracy78.48%69.56%92.98%71.07%86.12%98.13%95.40%97.01%96.95%98.79%
Accuracy std1.27%2.00%1.66%0.56%1.28%0.17%2.64%0.49%2.83%0.33%
Time5.895.636.6511.52.381.7625.1952.1437.7862.46
250/150Correct samples1281611410150379787137951571715782156401581815818
Accuracy77.85%69.02%92.42%70.60%86.08%98.10%95.02%96.82%96.72%98.58%
Accuracy std0.63%2.70%3.16%1.00%2.08%0.19%5.68%2.07%1.03%1.02%
Time5.745.596.3211.262.311.7424.1645.5934.9354.35
225/175Correct samples1281611410150379787137951571715782156401581815818
Accuracy72.34%67.82%89.83%70.37%86.08%98.07%93.61%96.66%96.56%97.87%
Accuracy std1.50%0.57%3.48%0.50%1.57%0.23%3.70%0.58%2.25%0.41%
Time5.625.486.2710.812.191.7122.8944.7433.3454.22
200/100Correct samples1281611410150379787137951571715782156401581815818
Accuracy71.35%66.72%88.43%69.07%85.37%98.04%93.48%96.24%95.67%97.23%
Accuracy std0.94%3.74%3.29%1.59%0.31%0.38%6.77%1.33%5.45%1.82%
Time5.375.16.141.182.161.6819.5738.6432.8145.94
In this table, the accuracy reflects the average accuracy; the time represents the average time and units in seconds (s).
Table 5. The accuracy of each method under different SNRs.
Table 5. The accuracy of each method under different SNRs.
MethodSNR(dB)
3210−1−2−3−4
BPNN with raw data64.13%60.81%57.25%54.25%45.63%43.81%34.43%30.31%
BPNN with TD features77.69%77.63%76.37%76.00%75.88%74.63%74.18%74.06%
BPNN with FD features90.94%90.81%87.88%83.63%82.81%79.00%73.87%72.68%
SVM with raw data69.63%68.25%66.56%66.25%64.88%64.06%62.93%62.68%
SVM with TD features84.75%84.69%84.25%84.19%84.00%83.31%81.18%80.87%
SVM with FD features96.44%95.56%94.38%92.44%92.00%90.12%85.31%83.43%
CNN96.75%95.38%93.56%87.93%85.25%83.06%79.25%77.87%
DCNN97.87%97.75%97.37%96.81%95.69%92.31%89.37%80.87%
SECNN97.69%97.31%97.06%96.75%95.50%92.06%88.58%78.25%
ECNN98.56%98.06%98.00%97.75%96.74%94.44%92.12%87.43%
Table 6. Variable load accuracy rate of each method.
Table 6. Variable load accuracy rate of each method.
MethodsBP-rBP-TBP-FSVM-rSVM-TSVM-FCNNDCNNSECNNECNN
1→2145.77%78.30%96.58%50.73%85.20%94.53%96.80%98.42%99.64%99.23%
242.95%76.77%96.69%48.85%86.24%94.93%96.12%98.91%98.83%99.98%
344.03%77.38%96.83%49.29%86.02%94.86%93.32%98.83%96.95%99.97%
443.39%79.28%96.67%51.25%85.23%94.64%94.85%99.94%97.17%99.98%
542.47%80.92%97.68%50.57%84.94%94.66%93.45%99.65%94.74%99.81%
Avg43.72%78.53%96.89%50.14%85.52%94.72%94.91%99.15%97.47%99.79%
Std1.28%1.64%0.45%1.02%0.57%0.16%1.55%0.62%1.89%0.33%
2→1138.83%79.70%94.97%56.83%83.51%94.39%95.00%97.35%92.77%97.90%
239.80%79.94%94.59%54.23%84.01%94.43%90.03%98.60%94.59%98.15%
338.51%78.41%94.16%51.10%84.39%94.32%95.85%98.44%94.85%98.69%
438.53%82.01%94.68%52.41%85.39%94.37%92.03%97.38%94.51%98.56%
539.43%80.26%95.06%53.72%80.98%94.18%92.90%96.94%93.97%97.95%
Avg39.02%80.06%94.69%53.66%83.65%94.34%93.16%97.74%94.14%98.25%
Std0.57%1.29%0.35%2.15%1.65%0.10%2.33%0.73%0.83%0.36%
1→3138.79%80.07%85.59%55.02%83.85%91.23%87.61%90.83%91.35%98.94%
240.56%78.30%88.83%52.90%83.44%91.09%92.39%95.94%90.17%98.97%
340.20%77.68%85.34%57.02%82.73%90.33%83.51%98.33%95.06%98.04%
437.86%83.54%83.08%55.59%83.00%92.29%89.66%97.53%88.27%99.29%
540.03%80.76%87.72%54.14%87.12%91.12%92.77%97.06%90.99%99.15%
Avg39.49%80.07%86.11%54.93%84.03%91.21%89.19%95.94%91.17%98.88%
Std1.13%2.31%2.24%1.55%1.78%0.70%3.81%2.98%2.48%0.49%
3→1134.26%77.78%91.22%51.86%83.49%88.35%82.34%84.30%81.14%91.83%
234.83%76.33%87.23%50.11%81.99%88.59%82.34%83.68%87.12%91.70%
334.09%70.04%90.26%49.89%81.46%91.18%80.57%85.81%80.22%90.93%
433.46%77.97%92.82%48.01%83.41%92.96%80.74%87.09%79.99%92.19%
533.84%77.35%87.58%49.83%82.95%89.17%81.64%86.74%85.56%91.97%
Avg34.10%75.89%89.82%49.94%82.66%90.05%81.53%85.52%82.81%91.72%
Std0.51%3.33%2.39%1.37%0.90%1.97%0.85%1.49%3.30%0.48%
2→3139.27%79.50%89.14%58.38%85.37%94.92%85.15%90.42%90.61%98.93%
241.24%80.24%89.41%55.21%85.73%94.72%89.98%93.07%91.10%99.62%
341.37%77.46%87.18%53.96%85.42%94.34%86.66%97.63%92.00%99.86%
441.05%82.84%87.59%55.54%84.66%94.21%86.27%91.68%93.23%98.86%
541.93%82.73%90.56%52.73%85.43%95.10%90.29%87.17%89.76%98.55%
Avg40.97%80.56%88.78%55.16%85.32%94.66%87.67%91.99%91.34%99.16%
Std1.01%2.28%1.38%2.11%0.40%0.38%2.32%3.83%1.33%0.55%
3→2135.31%81.12%94.08%47.52%81.98%91.42%82.99%87.26%81.12%95.47%
235.05%74.39%94.13%45.11%81.14%91.09%83.62%86.66%85.70%95.36%
335.42%71.81%96.18%43.88%81.19%90.69%84.34%87.47%81.27%95.11%
435.45%77.67%96.28%44.65%80.22%92.14%84.50%87.36%86.92%95.33%
535.67%76.22%95.11%44.73%79.40%91.02%83.87%92.68%87.11%95.48%
Avg35.38%76.24%95.15%45.18%80.79%91.27%83.86%88.29%84.42%95.35%
Std0.22%3.50%1.06%1.39%0.99%0.55%0.61%2.47%3.00%0.15%
Table 7. Data description.
Table 7. Data description.
ConditionsDaysTrain SamplesTest SamplesDatasetBearingLabel
Normal25–35300100No.1No.10
Inner face25–35300100No.1No.31
Outer face20–30300100No.3No.32
Roller element25–35300100No.1No.43
Table 8. Diagnosis results of the ten trials using different methods.
Table 8. Diagnosis results of the ten trials using different methods.
MethodsTotal SampleCorrect SampleAverage Testing Accuracy
CNN4000381595.38% ± 0.54%
DCNN4000392498.10% ± 0.42%
SECNN4000391397.83% ± 0.50%
ECNN4000398299.55% ± 0.33%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, C.; Huang, Q.; Zhang, C.; Yang, K.; Cheng, L.; Li, Z. ECNN: Intelligent Fault Diagnosis Method Using Efficient Convolutional Neural Network. Actuators 2022, 11, 275. https://doi.org/10.3390/act11100275

AMA Style

Zhang C, Huang Q, Zhang C, Yang K, Cheng L, Li Z. ECNN: Intelligent Fault Diagnosis Method Using Efficient Convolutional Neural Network. Actuators. 2022; 11(10):275. https://doi.org/10.3390/act11100275

Chicago/Turabian Style

Zhang, Chao, Qixuan Huang, Chaoyi Zhang, Ke Yang, Liye Cheng, and Zhan Li. 2022. "ECNN: Intelligent Fault Diagnosis Method Using Efficient Convolutional Neural Network" Actuators 11, no. 10: 275. https://doi.org/10.3390/act11100275

APA Style

Zhang, C., Huang, Q., Zhang, C., Yang, K., Cheng, L., & Li, Z. (2022). ECNN: Intelligent Fault Diagnosis Method Using Efficient Convolutional Neural Network. Actuators, 11(10), 275. https://doi.org/10.3390/act11100275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop