Next Article in Journal
α-Geodesical Skew Divergence
Next Article in Special Issue
Generalized Ordinal Patterns and the KS-Entropy
Previous Article in Journal
Cross-Scale Causality and Information Transfer in Simulated Epileptic Seizures
Previous Article in Special Issue
Image Statistics Preserving Encrypt-then-Compress Scheme Dedicated for JPEG Compression Standard
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Fault Identification for Rolling Bearings Fusing Average Refined Composite Multiscale Dispersion Entropy-Assisted Feature Extraction and SVM with Multi-Strategy Enhanced Swarm Optimization

1
College of Electrical Engineering & New Energy, China Three Gorges University, Yichang 443002, China
2
Hubei Provincial Key Laboratory for Operation and Control of Cascaded Hydropower Station, China Three Gorges University, Yichang 443002, China
3
Hubei Key Laboratory of Hydroelectric Machinery Design & Maintenance, China Three Gorges University, Yichang 443002, China
*
Authors to whom correspondence should be addressed.
Entropy 2021, 23(5), 527; https://doi.org/10.3390/e23050527
Submission received: 28 March 2021 / Revised: 20 April 2021 / Accepted: 22 April 2021 / Published: 25 April 2021

Abstract

:
Rolling bearings act as key parts in many items of mechanical equipment and any abnormality will affect the normal operation of the entire apparatus. To diagnose the faults of rolling bearings effectively, a novel fault identification method is proposed by merging variational mode decomposition (VMD), average refined composite multiscale dispersion entropy (ARCMDE) and support vector machine (SVM) optimized by multistrategy enhanced swarm optimization in this paper. Firstly, the vibration signals are decomposed into different series of intrinsic mode functions (IMFs) based on VMD with the center frequency observation method. Subsequently, the proposed ARCMDE, fusing the superiorities of DE and average refined composite multiscale procedure, is employed to enhance the ability of the multiscale fault-feature extraction from the IMFs. Afterwards, grey wolf optimization (GWO), enhanced by multistrategy including levy flight, cosine factor and polynomial mutation strategies (LCPGWO), is proposed to optimize the penalty factor C and kernel parameter g of SVM. Then, the optimized SVM model is trained to identify the fault type of samples based on features extracted by ARCMDE. Finally, the application experiment and contrastive analysis verify the effectiveness of the proposed VMD-ARCMDE-LCPGWO-SVM method.

1. Introduction

The operating conditions of industrial equipment are complicated, and rolling bearings are widely employed in the types of machinery that play important roles in industrial systems, such as coal, petrochemical, electric power and other industries [1,2]. Rolling bearings will inevitably cause damage to different degrees when running for a long time. What is worse, a fault in the rolling bearings may result in mechanical failure, causing economic loss and personal injury, and even inducing catastrophic accidents. However, monitoring the health condition of the rolling bearings through appropriate indicators and providing information can greatly reduce the occurrence of failures, and avoid major accidents [3,4].
Generally, once the rolling bearings fail, it is accompanied by vibration and sound. Therefore, using appropriate technology to process the collected vibration or acoustic signals could well detect potential failures [5,6,7]. Feature extraction is a committed step in identifying rolling bearing faults, but the vibration signals present nonlinear and nonstationary characteristics resulting in a limitation in the ability of feature extraction. Since signals are generally affected by noise information, an excellent signal processing method is necessary to eliminate the negative effects of these interferences [8,9,10]. Therefore, various time-frequency signal analysis approaches have been widely employed to extract the features in the fault identification of rolling bearings, including empirical mode decomposition (EMD) [11,12], local mean decomposition (LMD) [13], ensemble empirical mode decomposition (EEMD) [14] and variational mode decomposition (VMD) [15,16]. Many scholars have conducted much research into these methods: EMD is efficient for dealing with nonstationary signals, decomposing complex signals into a series of intrinsic mode functions (IMFs) adaptively, while endpoint effect defects remain. Compared to EMD, LMD has advantages in reducing iteration times and suppressing endpoint effect, which can adaptively decompose signal into a sum of subcomponents [17]. Nevertheless, LMD is complex in calculation and susceptible to sampling frequency, which affects the decomposition errors. To overcome these limitations, EEMD adds a set of white noise to help analyze the original signal [18]. However, the result impairs the purity of the original signal in the feature extraction process. In contrast to the above methods, VMD has excellent performance in signal processing, which avoids the mode mixing problem in EMD, the influence of sampling frequency in LMD and noise effect in EEMD. Furthermore, the capability and advancement of VMD has already been confirmed by preceding studies in engineering applications [19]. Thus, VMD was employed to decompose the nonstationary fault signals here, which laid the foundation for fault pattern recognition in rolling bearings.
It is key to extract fault features from the vibration signals in order to realize machinery equipment fault identification [20,21]. On account of nonstationary signals being decomposed by VMD, fault feature extraction would be successful and effective in this study. Entropy is a physical quantity, representing the regularity and complexity of a system which can reflect the nonlinear characteristics of a vibration signal. For example, permutation entropy (PE) [22], sample entropy (SampEn) [23] and fuzzy entropy (FE) [24,25] are all familiar entropies in the feature extraction of rolling bearing fault identification. The PE concept computes simply and quickly, but the disparity between the signal amplitude values is not adequately taken into account [26]. However, dispersion entropy (DE) [27,28] has the advantages of less influence by mutation signals, which can solve the shortages of slowing the calculation in SampEn and FE. Nevertheless, DE does not take into account sufficiently the relationship information between neighboring amplitudes. Meanwhile, DE is adept in analyzing time series at a single scale, but it may ignore the hidden valuable fault information at other scales. To overcome the drawbacks, previous researchers have made improvements, for example, GRCMMFDE was proposed by Zheng et al. [20] to extract fault features. In this paper, a modified DE, namely average refined composite multiscale dispersion entropy (ARCMDE) is put forward, which can not only preserve the original data effectively, but also enhance the ability of multiscale fault feature extraction of the IMFs by fusing average refined composite multiscale procedures.
It is virtually a pattern recognition issue for identifying rolling bearing faults. Therefore, many pattern recognition methods have been employed in various engineering application problems. For instance, artificial neural network (ANN) [29,30], Bayesian decision [31] and support vector machine (SVM) [32] have been employed in identification issues. Among the above methods, ANN has a strong capacity to deal with pattern recognition problems, while it requires abundant samples and is time-consuming to adjust the network structure parameters. Bayesian decision performs with notable capacity by considering prior probability, yet good accuracy is premised on a prior model with appropriate assumptions. Compared with above methods, SVM requires a small number of samples for training, and has good generalization ability. What is more, it has particular advantages in dealing with nonlinear and multidimensional pattern recognition problems [33]. It can satisfy the classification performance by means of finding an optimal hyperplane. Meanwhile, SVM has been applied for pattern recognition combining with feature extraction in rolling bearing fault identification. Therefore, SVM is explored to implement fault identification here.
The SVM model is easily affected by penalty factor C and kernel parameter g when performing pattern recognition. To address this issue, many optimization algorithms have been used to optimize the SVM model, for instance, Harris hawks optimization (HHO) [34,35], whale optimization algorithm (WOA) [36], particle swarm optimization (PSO) [37], moth−flame optimization (MFO) [38], differential evolution (DE) [39], sine cosine algorithm (SCA) [40] and grey wolf optimization (GWO) [41]. Although these intelligent optimization algorithms have achieved some favorable results, there are still problems of premature convergence of different degrees. In order to improve convergence precision, an enhanced GWO algorithm (LCPGWO) coupled with levy flight [42], cosine factor and polynomial mutation [43] is proposed in this paper. Compared with PSO, GWO, SCA, WOA, MFO and DE algorithms on 12 well-known benchmark functions, the results show that the LCPGWO has greater advantages in finding the optimal solution, which is employed to optimize the penalty factor C and kernel parameter g of SVM in this study.
In conclusion, firstly, the nonstationary original vibration signals were decomposed into several IMFs by means of VMD. Afterwards, ARCMDE was proposed to construct the feature vectors of different fault samples. Subsequently, the LCPGWO was explored to optimize the SVM model, which was employed to carry out the classification of different fault samples. Lastly, VMD-ARCMDE-LCPGWO-SVM method was applied to compare with other methods in terms of different locations and the motor speeds of rolling bearing faults. The performance of the proposed method was proved to be perfect for the engineering application problem. This study has the following contributions:
(1)
Average refined composite multiscale dispersion entropy (ARCMDE) was proposed to enhance the ability of fault feature extraction.
(2)
A novel multistrategy enhanced swarm optimizer (LCPGWO) was proposed to calibrate the parameters of SVM, which made it an excellent fault identification model.
(3)
The effectiveness of LCPGWO was verified by performance analysis with 12 well-known benchmark functions.
(4)
The superiority of the proposed fault identification method was ascertained by engineering experiment and comparative analysis.
The rest of this paper is arranged like this. Section 2 contributes to the fundamental theories about VMD and SVM. The proposed fault identification method according to ARCMDE and LCPGWO optimization approach is presented in Section 3. Section 4 is devoted to demonstrate the superiority of the proposed method in engineering application. The conclusions are in Section 5.

2. Fundamental Theories

2.1. Variational Mode Decomposition

VMD is a nonrecursive signal preprocessing, which can adaptively decompose the nonstationary signal into K band-limited intrinsic mode functions (IMFs) by setting the mode number K previously. The core of the VMD method is employed to construct and solve variational problems, which is established below:
min m k , ω k { k t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 } s . t . k = 1 K m k ( t ) = f ( t ) , k = 1 , 2 , , K
where m k = { m 1 , m 2 , , m k } and ω k = { ω 1 , ω 2 , , ω k } represent the set of K mode functions and central frequencies, respectively. The t is the partial derivative of time t , δ ( t ) is the unit pulse function and f ( t ) is the input signal of the given real value.
The above variational problem can transform into an unconstrained problem, which can be expressed as:
L ( m k , ω k , β ) = α k t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 + f ( t ) k m k ( t ) 2 2 + β ( t ) , f ( t ) k m k ( t )
where α represents the penalty factor and β ( t ) is the Lagrange multiplier [44].
Then m k and ω k can be optimized by Equations (3) and (4), respectively.
m k n + 1 = min { α t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 + f ( t ) i m i ( t ) + β ( t ) 2 2 2 }
ω k n + 1 = min { t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 }
The iterative equations in frequency domain are derived as follows:
m ^ k n + 1 ( ω ) = f ^ ( ω ) i k m ^ i ( ω ) + β ^ ( ω ) 2 1 + 2 α ( ω ω k ) 2
ω k n + 1 = 0 ω | m ^ k ( ω ) | 2 d ω 0 | m ^ k ( ω ) | 2 d ω
The Lagrange multipliers are expressed in Equation (7).
β ^ n + 1 ( ω ) = β ^ n ( ω ) + γ 1 ( f ( ω ) k m ^ k n + 1 ( ω ) )
where γ 1 represents an updating parameter.
The processes of VMD are the following:
Step 1: Initialize m k 1 , ω k 1 , β 1 , n = 1;
Step 2: Start loop, n = n + 1;
Step 3: Update m k and ω k on the basis of Equations (5) and (6);
Step 4: Update β according to Equation (7);
Step 5: If k m ^ k n + 1 m ^ k n 2 2 / m ^ k n 2 2 < ε , stop the loop, else turn to Step 2 for next iteration.

2.2. Support Vector Machine

SVM is designed for the two-classification issue, which can solve learning problems with limited samples. For a given sample set { ( x i , y i ) | i = 1 ,   2 , , n } , it maps the sample space to higher dimensions, and then a hyperplane is constructed, which can transform the nonlinear problem of the sample space into the linear problem of the feature space to solve. The hyperplane function is defined as follow:
ϖ · x + b = 0
where ϖ and b are weight vector and bias parameter, respectively. Equation ϖ · x is the inner product.
For example, in order to correctly identify samples of a binary classification issue, all samples demand the following conditions:
ϖ · x i + b { > 1 f o r y i = 1 < 1 f o r y i = 1
The maximizing classification interval is 2 / ϖ 2 , which is obtained by minimizing ϖ 2 . Then the slack term ξ and penalty factor C are brought into Equation (9) to solve the linear indivisibility problem of SVM model.
{ min f = 1 2 w 2 + C i = 1 n ξ i s . t . y i ( ϖ T · x i + b ) 1 ξ i , i = 1 , 2 , , n
Lagrange function is introduced in Equation (10), which can be described as:
max L = i = 1 n μ i 1 2 i , j = 1 n μ i μ j y i y j K ( x i , x j ) s . t . i = 1 n μ i y i = 0 , μ i 0 , i = 1 , 2 , , n
where μ i means the Lagrange multiplier, K ( x i , x j ) is the kernel function of SVM.
In this paper, the radial basis function (RBF) is selected as the kernel function of SVM. Solving the dual problem of Equation (11), the optimal classification discriminant function with RBF is defined as:
f ( x ) = s g n ( i = 1 n μ i K ( x i , x ) + b )
Among them, RBF function is represented as:
K ( x i , x j ) = ϕ ( x i ) · ϕ ( x j ) = exp ( g | | x i x j | | 2 )
where g represents the kernel parameter, ϕ ( x ) is the nonlinear vector function.

3. Intelligent Fault Identification for Rolling Bearings Fusing the Proposed Method

3.1. Average Refined Composite Multiscale Dispersion Entropy

3.1.1. Dispersion Entropy

For a given time series r = { r i , i = 1 , 2 , N } , which length is N. r i is normalized by mean of employing a mapping function [27].
y i = 1 σ 2 π r i e ( s μ ) 2 2 σ 2 d s
where σ represents the variance of the normal distribution and μ represents the expectation value. The time series r is normalized to y = { y 1 , y 2 , y n } , y i ( 0 , 1 ) . Subsequently, the phase space is reconstructed into a matrix for y:
y j m = [ y j , y j + t d , , y j + ( m 1 ) t d ]
where j = 1 , 2 , , N ( m 1 ) t d , m is embedding dimension, t d is time delay, y j m is mapped to the scope [ 1 , c ] :
z i c = r o u n d ( c · y i + 0.5 )
z j m , c = [ z j c , z j + t d c , , z j + ( m 1 ) t d c ]
where z j c represents the j-th member of class sequence z j m , c , and round is rounding. Each z j m , c corresponds to a dispersion pattern π v 0 v 1 v m 1 with z j c = v 0 ,   z j + d c = v 1 , ,   z j + ( m 1 ) τ c = v m 1 .
The frequency of π v 0 v 1 v m 1 can be deduced as:
p = N u m b e r { j | j N ( m 1 ) t d , π v 0 v 1 v m 1 } N ( m 1 ) t d
where N u m b e r { j | j n ( m 1 ) t d , π v 0 v 1 v m 1 } is the emergence number of each π v 0 v 1 v m 1 that corresponding to z j m , c :
The dispersion entropy is defined as:
D E ( r , m , c , t d ) = π = 1 c m p · ln ( p )
There is a linear negative correlation between DE value and time series. The larger the DE value, the more irregular the time series.

3.1.2. Average Refined Composite Multiscale Dispersion Entropy

As a single scale method, DE may result, in that much useful and significant information hides in multiple scales, which would limit the representation precision of nonstationary fault signals. To solve the disadvantage, average refined composite multiscale dispersion entropy (ARCMDE) is proposed, which is utilized to extract multiscale fault features from IMFs. Given time series r = { r i , i = 1 , 2 , N } with length N, the k-th composite multiscale coarse-grained sequence u k ( τ ) = { u k , 1 ( τ ) , x k , 2 ( τ ) , } is defined as:
u k , j ( τ ) = 1 τ i = k + ( j 1 ) τ k + j τ 1 u i , 1 j | N τ | , 1 k τ
where τ is scale factor.
For each scale factor, refined composite multiscale dispersion entropy (RCMDE) is expressed in Equation (21).
R C M D E ( u , m , c , t d , τ ) = π = 1 c m p ¯ ( π v 0 v 1 v m 1 ) · ln ( p ¯ ( π v 0 v 1 v m 1 ) ) p ¯ ( π v 0 v 1 v m 1 ) = 1 τ k 1 τ p k ( τ )
where p ¯ ( π v 0 v 1 v m 1 ) is mean probability of the dispersion pattern π of coarse-grained sequence r k ( τ ) .
Finally, ARCMDE is expressed as average value of all RCMDE in the scale τ , such that:
A R C M D E ( r , m , c , t d , τ ) = 1 τ k = 1 τ R C M D E ( u , m , c , t d , τ )

3.2. GWO Coupled with Multiple Enhancement Strategies

3.2.1. Grey Wolf Optimization

Grey wolves are in a dominant position in the competitive natural environment which have a strict social hierarchy and ingenious cooperative predation. According to the behavior of grey wolves when they hunt, GWO is proposed to solve optimization problems, where the grey wolves are graded into four levels [45]. The first level, called α wolf, may not be the strongest wolf, but it is the best manager in the system and responsible for overall planning. The β wolf on the second level is the best substitute for the α wolf. The δ wolf on the third level acts as the suboptimal solution. The ω wolf is the candidate solution at the bottom and is responsible for balancing the internal relations of the wolf population.
The mathematical expression of grey wolves’ predation can be expressed as follows:
D = | C · X p ( t ) X ( t ) |
X ( t + 1 ) = X p ( t ) A · D
where D represents the distance between the wolf and the prey, X and X p denote the position of the grey wolf and prey, respectively, t is current iterations time.
The coefficient vectors A in Equation (23) and C in Equation (24) are expressed as follows:
A = 2 a · h 1 a
C = 2 · h 2
a = 2 t 2 max
where a is convergence factor, which decreases linearly from 2 to 0, h 1 and h 2 are random vectors in [0, 1], max is the maximum number of iterations.
In GWO algorithm, α , β and δ wolves to approach and surround the prey when the prey is identified by the grey wolves. Therefore, the position of the prey can be determined by the position of the grey wolves. The mathematical model for updating the position of each wolf is as follows:
{ D α = | C 1 · X α ( t ) X ( t ) | D β = | C 2 · X β ( t ) X ( t ) | D δ = | C 3 · X δ ( t ) X ( t ) |
{ X 1 = X α A 1 · D α X 2 = X β A 2 · D β X 3 = X δ A 3 · D δ
where D α , D β and D δ represent distances of α , β and δ wolves from other individuals respectively, X 1 , X 2 and X 3 denote the current position of α , β and δ wolves respectively.
The positional relationship between the grey wolf individual ω and the prey can be determined as follows:
X ( t + 1 ) = X 1 + X 2 + X 3 3
If | A | < 1 , the wolves will attack the prey; otherwise, the wolves will search for the prey. To sum up, the pseudocode of GWO algorithm is shown in Algorithm 1.
Algorithm 1. The algorithm pseudocode of GWO.
  • Initialize grey wolf population X i ( i = 1 , 2 , 3 , n )
  • Initialize the parameters a, A and C
  • Evaluate the fitness of each wolf
  • Assign the best three grey wolves to X α , X β , X δ
  • whilet < max iteration
  • for each search agent
  • Update the position of current grey wolves by Equation (30)
  • end for
  • Update a, A and C
  • Evaluate the fitness of each wolf
  • Update X α , X β , X δ
  • t = t + 1
  • end while
  • return X α

3.2.2. Grey Wolf Optimization Coupled with Multiple Enhancement Strategies

GWO is slow in convergence speed, resulting in an easy fall into local optimum in the later iteration. In this section, the improved GWO coupled with multiple enhancement strategies (LCPGWO) is explored to solve the shortcomings of GWO, which can improve the ability of global search, accelerate convergence and enhance the capacity for preventing local optimum during the later iteration. The realization processes of LCPGWO are described below in detail.
As shown in Equation (25), parameter a influences the change of coefficient vectors A, which coordinates the local and global explorations. The larger a is, the stronger the global exploration ability; the smaller a is, the stronger the local exploration ability. To promote the adaptation during both local and global explorations, the linearly decreasing a in Equation (27) is substituted with cosine factor as shown in Equation (31). Thus, a is large and reduces slowly for global exploration in the early iteration stage, while it will reduce rapidly for the local search in the later iteration stage.
a = 2 · cos ( π 2 · t max )
Additionally, inertial weight based on cosine factor is introduced in this paper to enhance the global exploration, which can be seen in Equation (32).
W = 2 · cos ( π 2 · t max ) 1
With inertial weight, the positions of α , β and δ wolves are reformulated in Equation (33).
{ X 1 = X α W · A 1 · D α X 2 = X β W · A 2 · D β X 3 = X δ W · A 3 · D δ
In iterative process, it falls into local optimum easily when the ω wolf approaches the other three wolves. By introducing inertial weight [46], the positional relationship between the grey wolf individual ω and the prey can be redefined as follows:
X ( t + 1 ) = W 1 · X 1 + W 2 · X 2 + W 3 · X 3 3
{ W 1 = | X 1 | | X 1 | + | X 2 | + | X 3 | W 2 = | X 2 | | X 1 | + | X 2 | + | X 3 | W 3 = | X 3 | | X 1 | + | X 2 | + | X 3 |
where W 1 , W 2 and W 3 represent the learning rate of ω to α , β and δ wolves, respectively.
In this paper, α wolf is searched globally by levy flight strategy for preventing local optimum, where the flight step is stable expansion distribution. The next generation of α wolf is calculated as follows:
X ( t + 1 ) = X ( t ) + d L e v y ( θ )
where X(t) is position of α wolf at t-th iteration, operator is entry-wise multiplications, d and L e v y ( θ ) are random numbers and step of α wolf respectively, which are determined by Equations (37) and (38).
d = d 0 · ( X ( t ) X α ( t ) )
L e v y ( θ ) u = t 1 θ
where d 0 is a constant, θ is levy index, a random number between 0 and 2, whose value is set at 1.5 here, the flight step of α wolf is a power-law formula.
A more detailed description about levy flight can be summarized as Equation (39).
d L e v y ( θ ) 0.01 u | v | 1 θ ( X ( t ) X α ( t ) )
where s and v are both normal distributions:
{ s N ( 0 , σ s 2 ) v N ( 0 , σ v 2 )
σ s = { γ ( 1 + θ ) sin ( π θ / 2 ) γ θ [ ( 1 + θ ) / 2 ] 2 ( θ 1 ) / 2 } 1 / θ , σ v = 1
where parameter γ is the standard gamma function.
In swarm intelligence optimization algorithm, it traps in local optimum easily. For this purpose, a polynomial mutation operator for GWO is introduced in this section to promote the exploring ability within the whole situation space, thus to avoid from trapping in local optimum as well as maintain the diversity of solution in the later iteration stage. The mathematical formula of the polynomial mutation is written in Equation (42).
X ( t + 1 ) = X ( t ) + ξ ( u k l k )
where X(t) is the original optimal individual position, X(t + 1) is the mutated optimal individual position, u k represents the upper limit of the position and l k is the lower limit of the position.
The parameter ξ is calculated as follows:
ξ = { [ 2 u + ( 1 2 s ) ( 1 ξ 1 ) η + 1 ] 1 η + 1 1 , s 0.5 1 [ 2 ( 1 s ) + 2 ( s 0.5 ) ( 1 ξ 2 ) η + 1 ] 1 η + 1 , s > 0.5
where parameter s is a random number in [0, 1], η is also [0, 1].
The parameters ξ 1 and ξ 2 are deduced in Equation (44).
{ ξ 1 = ( X ( t ) l k ) / ( u k l k ) ξ 2 = ( u k X ( t ) ) / ( u k l k )
To sum up, the pseudocode of LCPGWO algorithm is displayed in Algorithm 2.
Algorithm 2. The algorithm pseudocode of LCPGWO.
  • Initialize grey wolf population X i ( i = 1 , 2 , 3 , n )
  • Initialize the parameters a by Equation (31), and initialize A and C
  • Evaluate the fitness of each wolf
  • Assign the best three grey wolves to X α , X β , X δ
  • whilet < max iteration
  • for each search agent
  • Update the position of current grey wolves by Equation (34)
  • Calculate the new positions of grew wolves employing the levy flight and polynomial mutation by Equations (36) and (42)
  • end for
  • Update a, A and C
  • Evaluate the fitness of each wolf
  • Update X α , X β , X δ
  • t = t + 1
  • end while
  • return X α

3.2.3. Experimental Study and Results Analysis

Benchmark Functions

To prove the effectiveness of the proposed LCPGWO algorithm, six well-known nature-inspired optimization algorithms, including PSO, GWO, SCA, WOA, MFO and DE were applied for comparison. Meanwhile, 12 benchmark functions were selected for optimization experiments as listed in Table 1 and divided into two categories, where F1-F7 were unimodal functions and F8-F12 were multimodal functions [47,48,49]. In Table 1, Fmin was the minimum value of each benchmark function. The unimodal functions were mainly employed to test the convergence rate of the algorithms, while the multimodal functions were carried out to test the global exploration ability of the algorithms.

Comparison and Analysis with Different Algorithms

The experiment was on a personal computer, which was equipped with Windows 10 system and an Intel(R) Core (TM) CPU at 2.89 GHz and 4 GB memory. The simulation software was MATLAB R2016a.
Each benchmark function was run 10 times independently to obtain an objective result. The iteration number and searching agents in the experiment were set at 200 and 40, respectively. The detailed parameter settings are shown in Table 2. During iterations of all algorithms, the optimal fitness values were recorded every time. The average fitness values were obtained to draw a curve reflecting the convergence trend of the algorithms. The convergence curves of PSO, GWO, SCA, WOA, MFO, DE and LCPGWO algorithms on the 12 well-known benchmark functions are listed in Figure 1. At the same time, the maximum value, minimum value, mean values and standard deviations of the optimal solution obtained by all algorithms are displayed in Table 3, where a lower value means better search ability and stability.
From Figure 1, it was observed that the proposed LCPGWO algorithm converged better than PSO, GWO, SCA, WOA, MFO, DE algorithms for all F1−F12, indicating that the proposed LCPGWO algorithm was able to prevent local optimum and converge to the optimal value at a faster speed. From Table 3, it can be concluded that the proposed LCPGWO algorithm achieved the lowest value in the maximum value, minimum value, mean value and standard deviation for both unimodal and multimodal functions. In particular, the results of the LCPGWO algorithm were superior to PSO, GWO, SCA, WOA, MFO, DE algorithms, especially on functions F6, F8, F9 and F10. On the whole, the proposed LCPGWO algorithm is more effective and feasible than contrastive methods.

3.3. SVM Optimized by LCPGWO

In order to obtain a good generalization performance in dealing with fault identification issues, it is necessary to assign appropriate parameters C and g of SVM. Thus, the proposed LCPGWO was used to optimize SVM model. The main procedures of classification machine learning with SVM optimized by the proposed LCPGWO algorithm are in below:
Step 1: Initialize the population and set relevant parameters;
Step 2: Update individual’s status according to Equations (36) and (42);
Step 3: Calculate the fitness value, which is the cross-validation accuracy of SVM;
Step 4: Update individual’s new position;
Step 5: Repeat Steps 2–4 until the maximum time of iterations is reached or the convergence condition is met;
Step 6: Choose the maximal cross-validation accuracy as the optimal parameters C and g of SVM;
Step 7: Train the optimal SVM model according to the training set;
Step 8: Recognize the testing set and finish the identification.

3.4. Intelligent Fault Identification for Rolling Bearings Fusing the Proposed Method

A novel fault identification method is proposed according to VMD, ARCMDE as the feature extraction and SVM optimized by LCPGWO as classification model in this paper. The flowchart of the fault identification with the proposed method is illustrated in Figure 2. To be specific, firstly, vibration signals were decomposed into four IMFs of different series by VMD. Afterwards, the feature vectors were constructed by means of the proposed ARCMDE, which extracted fault features from IMFs. Finally, the parameters of SVM model were optimized by multistrategy enhanced swarm optimization algorithm LCPGWO, thus achieving the fault pattern recognition.

4. Engineering Application

4.1. Data Collection

To attest the effectiveness of the proposed method in this paper, the machinery fault simulator (MFS) manufactured by SQI company was used to measure the relevant experimental data for bearings. The detailed information of machinery fault simulator is shown in Figure 3. The type of rolling bearings selected in the experiment was ER12KCL. Meanwhile, the motor speeds of the bearings were 1800 rpm and 2200 rpm when collecting experimental data. The time of sample data collection was set as 10 s. The bearings’ state types were also divided into four, which were inner race fault, ball fault, outer race fault and combination fault, which are displayed in Figure 4. The diameter of all the experimental bearings was 3/4 inches. The vibration signals of the rolling bearings were collected by employing the acceleration sensor mounted on the bearing seat of the motor drive end, and the sampling frequency was 12.8 kHz. Further, there were 61 samples of the vibration signals for each type, and one sample possessed 2048 sample points. The detailed data about the experiments are shown in Table 4.

4.2. Application to Fault Identification of Rolling Bearings

To fully prove that the proposed fault identification method was effective, other relevant methods were applied to compare with the proposed VMD-ARCMDE-LCPGWO-SVM method. More specifically speaking, FE, DE and RCMDE were employed to compare with ARCMDE at the feature extraction stage; GWO was applied for comparison at the parameter optimization stage. The settings of the same parameters were uniform in all the comparison experiments.
Feature extraction is the main problem in the process of identifying rolling bearing faults. VMD was selected to decompose the fault signal into a set of IMFs. The parameter of decomposing mode number K was decided in advance, where it was determined by the center frequency observation method according to a previous study [50]. In this paper, the K value was obtained by experiment using sample data under a motor speed of 1800 rpm. As shown in Figure 5, if K is too large, the center frequencies of adjacent IMFs are too close, resulting in mode mixing, which means excessive decomposition. However, if K is too small, the fault signal cannot be effectively decomposed, which leads to more valuable information being ignored. Therefore, the K value in the paper was set as 4.
The waveforms of the original signals with different fault positions (L1, L2, L3, L4, L5) and different motor speeds (L3, L4, L8, L9) are illustrated in Figure 6. With VMD decomposition, all the vibration signals were decomposed into four subcomponents including IMF1, IMF2, IMF3, IMF4, as shown in Figure 7. IMFs decomposed from original signals have quite different fluctuation characteristics.
After the IMFs of all samples were obtained through signal decomposition, fault feature vectors were constructed by calculating ARCMDE values, where parameters should be chosen properly beforehand [51,52]. Here, four parameters were set in advance, which were embedding dimension m, number of class c, maximum scale factor τ max , and time delay t d . By referring to previous papers [53], the parameter settings of ARCMDE are displayed in Table 5.
To verify the performance of the proposed method, 61 feature vectors belonging to different fault types were selected for the contrast experiment. They were randomly divided into two parts, where 40 vectors were selected for training and the remaining 21 vectors were for testing. After that, the proposed LCPGWO method was utilized to enhance the classification identification performance of SVM by searching the optimal values of parameters C and g of SVM model. The searching range of C and g were set in [2−10, 210], meanwhile, the optimization experiments were accomplished by 100 iterations and 20 searching agents. The five-fold cross-validation was applied for calculating the fitness values of training samples in this experiment. Hence, according to the obtained optimal parameters C and g by the proposed LCPGWO method, the SVM model was trained and employed to achieve fault identification. For a dependable verification of the proposed method about the effectiveness and superiority, each of these comparative fault identification methods was run 10 times, on average, independently, and training samples were randomly selected. Moreover, accuracy (ACC), adjusted rand index (ARI), F-measure (F) and normalized mutual information (NMI) [54,55] were applied to evaluate the capability of these different approaches. The higher values of the four metrics mean that the matching degree between fault identification result and real samples distribution information is better. The calculation methods of the four metrics are shown in the Table 6, where the range of ARI is set in [−1, 1], and the rest of the metrics are in [0, 1].
The following notations in Table 6 are adopted: TP, TN, FP, FN represent true positive, true negative, false positive, false negative on the basis of the result of fault identification and actual label; Φ and Ω are the sets of given actual label classified result, respectively; n11 is the number of sample pairs for the same label in both Φ and Ω, while n00 is for different label, C n 2 is all possible sample pairs combinations; P(𝜑) and P(𝜔) represent the probability functions of Φ and Ω respectively; P(𝜑,𝜔) is the joint probability functions of Φ with Ω.
Eight relevant methods were employed to compare for illustration of the advantages of the proposed approach in this study. The four evaluation values of fault identification results are shown in Table 7. By comparing with the results of different methods, it proves that the proposed VMD-ARCMDE-LCPGWO-SVM method is the best of four evaluation metrics. It has the highest values at 0.9597, 0.9627, 0.9838, 0.9838 under a motor speed of 1800 rpm and 0.9303, 0.9381, 0.9712, 0.9714 under a motor speed of 2200 rpm. Although the performance of NMI under 2200 rpm was not optimal, it was still very good, which could be considered a desirable result. The evaluation value deviations were also very low. In order to better analyze the results, the results when the motor speed was 1800 rpm were taken as the analysis. For feature extraction, VMD-FE-GWO-SVM, VMD-DE-GWO-SVM, VMD-RCMDE-GWO-SVM and VMD-ARCMDE-GWO-SVM methods were compared, respectively. The ACC of the VMD-ARCMDE-GWO-SVM was 0.9781, which was superior to the VMD-FE-GWO-SVM, VMD-DE-GWO-SVM and VMD-RCMDE-GWO-SVM methods. Similarly, comparing with VMD-FE-LCPGWO-SVM, VMD-DE-LCPGWO-SVM and VMD-RCMDE-LCPGWO-SVM methods, the proposed VMD-ARCMDE-LCPGWO-SVM method also had the highest accuracy. The results revealed that the proposed ARCMDE method was superior for feature extraction.
For the parameter optimization of the SVM model, it can be observed that the ACC of the proposed VMD-ARCMDE-LCPGWO-SVM method was far better than VMD-ARCMDE-GWO-SVM method. Additionally, the VMD-FE-LCPGWO-SVM and VMD-DE-LCPGWO-SVM methods also performed better than the VMD-FE-GWO-SVM and VMD-DE-GWO-SVM methods, respectively, which proved the effectiveness of LCPGWO method for parameter optimization. Based on the above experimental analyses, the conclusion can be drawn that the proposed VMD-ARCMDE-LCPGWO-SVM method achieves stable competitiveness when compared with other methods.
In order to show the fault identification evaluation results comparison of different methods more intuitively, the comparison of evaluation values of different methods under 1800 rpm are shown in Figure 8, illustrating that the proposed method has obvious advantage in fault identification. As shown in Figure 9, the proposed VMD-ARCMDE-LCPGWO-SVM method achieved more outstanding results than the other methods on the whole, where the evaluation values of the data had a strong performance at 2200 rpm. Furthermore, the boxplots of the four evaluation values are displayed in Figure 10, it demonstrates the performances of the different methods. The proposed method possesses better stability and overall performance. Therefore, with the experiments on various fault locations, motor speeds and the detailed comparative analysis given above, the superiority of the proposed identification model is effectively demonstrated.

5. Conclusions

Increasingly complex rotating machinery equipment must have excellent mechanical fault identification technology to ensure its safe and effective operation. In this paper, a novel fault identification approach is proposed by fusing VMD, ARCMDE and SVM with LCPGWO optimization. Firstly, VMD was employed to decompose the nonstationary fault signals into several IMFs by the center frequency observation method. Afterwards, ARCMDE fusing the superiorities of DE and average refined composite multiscale procedure, was proposed to construct the feature vectors of different fault samples, which performed excellently in multiscale fault feature extraction from the IMFs. Subsequently, LCPGWO, which was GWO enhanced by multistrategy, including levy flight, cosine factor and polynomial mutation strategies, was compared with the other algorithms on different benchmark functions. The results demonstrated that the use of LCPGWO was explored to improve the ability of global search, accelerate convergence and enhance the capacity for jumping out of local optimum in the later iteration. Thus, LCPGWO was applied to optimize penalty factor C and kernel parameter g of SVM model, which was employed to realize the fault classification for different fault samples. Lastly, the proposed VMD-ARCMDE-LCPGWO-SVM method was applied to compare with other methods for rolling bearing fault identification. Meanwhile, the experiment results were measured by four evaluation metrics named ACC, ARI, F and NMI. The proposed fault identification method has smaller error, better stability and higher reliability than the other contrastive methods. Particularly, under motor speed 1800 rpm, the identification accuracy of the proposed method was 9.33, 3.62, 1.71 and 0.57% higher than the VMD-FE-GWO-SVM, VMD-DE-GWO-SVM, VMD-RCMDE-GWO-SVM and VMD-ARCMDE-GWO-SVM methods; and also 8.09, 3.43 and 0.67 higher than the VMD-FE-LCPGWO-SVM, VMD-DE-LCPGWO-SVM and VMD-RCMDE-LCPGWO-SVM methods. Meanwhile, the evaluation metrics were also outstanding under 2200 rpm. Therefore, it can be expected to provide a new way for rolling bearing fault identification.

6. Discussion

The generation and development of rolling bearing faults are caused by the coupling of many factors, which contain a large number of uncertain factors. Conventional diagnostic methods have difficulty in obtaining satisfactory results. The SVM method is a relatively novel method in the field of rolling bearing fault identification. Although some research has been done in this paper, the author believes that there are still some issues worthy of further research: (1) in practical engineering applications, different components in the unit influence each other, complex rolling bearing combinations may have multifactor failures in the future. Therefore, it is still necessary to conduct in depth failure mechanism research to make the identification work more targeted, more accurate and reliable. (2) The research of a fault identification classifier is only one aspect of the problem of fault identification. The premise of fault identification is to apply an advanced signal analysis method to extract more effective and more capable features from the rolling bearings’ operating state. Therefore, it is necessary to extract fault feature information from multiple angles according to the bearing fault signal characteristics and combined with new signal processing technology, so as to lay a foundation for the SVM to provide more effective fault features. (3) The occurrence of rolling bearing faults is a gradual process, and minor failures have little impact, but parts must be replaced after reaching a certain degree of severity. Therefore, real-time monitoring of rolling bearings is very important. This paper only analyzes the bearing vibration signals collected on the experimental platform, and does not realize real-time monitoring, which is also the direction of author’s next research.

Author Contributions

H.S. performed the experiments and contributed to paper writing, W.F. designed the research, B.L. participated in revision process, K.S. provided recommendations for this research and D.Y. participated in the discussion. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Open Fund of Hubei Key Laboratory of Hydroelectric Machinery Design & Maintenance (2020KJX03).

Data Availability Statement

The data included in this study are all owned by the research group and will not be transmitted.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, X.; Qin, Y.; He, C.; Jia, L.; Kou, L. Rolling element bearing fault diagnosis under impulsive noise environment based on cyclic correntropy spectrum. Entropy 2019, 21, 50. [Google Scholar] [CrossRef] [Green Version]
  2. Wang, Z.; Zhou, J.; Lei, Y.; Du, W. Bearing fault diagnosis method based on adaptive maximum cyclostationarity blind deconvolution. Mech. Syst. Signal Process. 2021, in press. [Google Scholar]
  3. Xu, L.; Chatterton, S.; Pennacchi, P. Condition monitoring of rolling element bearing based on moving average cross-correlation of power spectral density. In Proceedings of the IFToMM World Congress on Mechanism and Machine Science, Krakow, Poland, 15–18 July 2019; Springer: Cham, Switzerland, 2019; pp. 3411–3418. [Google Scholar]
  4. Xu, L.; Chatterton, S.; Pennacchi, P. A novel method of frequency band selection for squared envelope analysis for fault diagnosing of rolling element bearings in a locomotive powertrain. Sensors 2018, 18, 4344. [Google Scholar] [CrossRef] [Green Version]
  5. Glowacz, A. Recognition of acoustic signals of induction motor using fft, smofs-10 and isvm. Eksploat. Niezawodn. 2015, 17, 569–574. [Google Scholar] [CrossRef]
  6. Sun, S.; Przystupa, K.; Wei, M.; Yu, H.; Ye, Z.; Kochan, O. Fast bearing fault diagnosis of rolling element using Lévy Moth-Flame optimization algorithm and Naive Bayes. Eksploat. Niezawodn. 2020, 22, 730–740. [Google Scholar] [CrossRef]
  7. Xu, L.; Pennacchi, P.; Chatterton, S. A new method for the estimation of bearing health state and remaining useful life based on the moving average cross-correlation of power spectral density. Mech. Syst. Signal Process. 2020, 139, 106617. [Google Scholar] [CrossRef] [Green Version]
  8. Hernandez-Muriel, J.A.; Bermeo-Ulloa, J.B.; Holguin-Londono, M.; Alvarez-Meza, A.M.; Orozco-Gutierrez, A.A. Bearing health monitoring using relief-F-based feature relevance analysis and HMM. Appl. Sci. 2020, 10, 5170. [Google Scholar] [CrossRef]
  9. Gradzki, R.; Kulesza, Z.; Bartoszewicz, B. Method of shaft crack detection based on squared gain of vibration amplitude. Nonlinear Dyn. 2019, 98, 671–690. [Google Scholar] [CrossRef] [Green Version]
  10. Gradzki, R.; Lindstedt, P.; Kulesza, Z.; Bartoszewicz, B. Rotor blades diagnosis method based on differences in phase shifts. Shock Vib. 2018, 2018. [Google Scholar] [CrossRef]
  11. Fu, W.; Wang, K.; Tan, J.; Zhang, K. A composite framework coupling multiple feature selection, compound prediction models and novel hybrid swarm optimizer-based synchronization optimization strategy for multi-step ahead short-term wind speed forecasting. Energy Convers. Manag. 2020, 205, 112461. [Google Scholar] [CrossRef]
  12. Xiong, D.; Fu, W.; Wang, K.; Fang, P.; Chen, T.; Zou, F. A blended approach incorporating TVFEMD, PSR, NNCT-based multi-model fusion and hierarchy-based merged optimization algorithm for multi-step wind speed prediction. Energy Convers. Manag. 2021, 230, 113680. [Google Scholar] [CrossRef]
  13. Smith, J.S. The local mean decomposition and its application to EEG perception data. J. R. Soc. Interface 2005, 2, 443–454. [Google Scholar] [CrossRef]
  14. Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
  15. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  16. Wang, R.; Li, C.; Fu, W.; Tang, G. Deep learning method based on gated recurrent unit and variational mode decomposition for short-term wind power interval prediction. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 3814–3827. [Google Scholar] [CrossRef] [PubMed]
  17. Cheng, J.; Yang, Y.; Yang, Y. A rotating machinery fault diagnosis method based on local mean decomposition. Digit. Signal Process. 2012, 22, 356–366. [Google Scholar] [CrossRef]
  18. Wu, Z.; Huang, N.E. A study of the characteristics of white noise using the empirical mode decomposition method. Proc. R. Soc. Lond. Ser. A Math. Phys. Eng. Sci. 2004, 460, 1597–1611. [Google Scholar] [CrossRef]
  19. Zhang, M.; Jiang, Z.; Feng, K. Research on variational mode decomposition in rolling bearings fault diagnosis of the multistage centrifugal pump. Mech. Syst. Signal Process. 2017, 93, 460–493. [Google Scholar] [CrossRef] [Green Version]
  20. Zheng, J.; Pan, H. Use of generalized refined composite multiscale fractional dispersion entropy to diagnose the faults of rolling bearing. Nonlinear Dyn. 2020, 101, 1417–1440. [Google Scholar] [CrossRef]
  21. Wang, Z.; Yang, N.; Li, N.; Du, W.; Wang, J. A new fault diagnosis method based on adaptive spectrum mode extraction. Struct. Health Monit. 2021. [Google Scholar] [CrossRef]
  22. Zhang, W.; Zhou, J. Fault diagnosis for rolling element bearings based on feature space reconstruction and multiscale permutation entropy. Entropy 2019, 21, 519. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Yang, F.; Kou, Z.; Wu, J.; Li, T. Application of mutual information-sample entropy based MED-ICEEMDAN de-noising scheme for weak fault diagnosis of hoist bearing. Entropy 2018, 20, 667. [Google Scholar] [CrossRef] [Green Version]
  24. Zheng, J.; Cheng, J.; Yang, Y. A rolling bearing fault diagnosis approach based on LCD and fuzzy entropy. Mech. Mach. Theory 2013, 70, 441–453. [Google Scholar] [CrossRef]
  25. Wang, K.; Fu, W.; Chen, T.; Zhang, B.; Xiong, D.; Fang, P. A compound framework for wind speed forecasting based on comprehensive feature selection, quantile regression incorporated into convolutional simplified long short-term memory network and residual error correction. Energy Convers. Manag. 2020, 222, 113234. [Google Scholar] [CrossRef]
  26. Yan, R.; Liu, Y.; Gao, R.X. Permutation entropy: A nonlinear statistical measure for status characterization of rotary machines. Mech. Syst. Signal Process. 2012, 29, 474–484. [Google Scholar] [CrossRef]
  27. Rostaghi, M.; Azami, H. Dispersion entropy: A measure for time-series analysis. IEEE Signal Process. Lett. 2016, 23, 610–614. [Google Scholar] [CrossRef]
  28. Shao, K.; Fu, W.; Tan, J.; Wang, K. Coordinated approach fusing time-shift multiscale dispersion entropy and vibrational Harris hawks optimization-based SVM for fault diagnosis of rolling bearing. Measurement 2021, 173, 108580. [Google Scholar] [CrossRef]
  29. Yu, Y.; Yu, D.; Cheng, J. A roller bearing fault diagnosis method based on EMD energy entropy and ANN. J. Sound Vib. 2006, 294, 269–277. [Google Scholar] [CrossRef]
  30. Raj, N.; Jagadanand, G.; George, S. Fault detection and diagnosis in asymmetric multilevel inverter using artificial neural network. Int. J. Electron. 2018, 105, 559–571. [Google Scholar] [CrossRef]
  31. Cai, B.; Huang, L.; Xie, M. Bayesian Networks in Fault Diagnosis. IEEE Trans. Ind. Inform. 2017, 13, 2227–2240. [Google Scholar] [CrossRef]
  32. Fu, W.; Wang, K.; Zhang, C.; Tan, J. A hybrid approach for measuring the vibrational trend of hydroelectric unit with enhanced multi-scale chaotic series analysis and optimized least squares support vector machine. Trans. Inst. Meas. Control 2019, 41, 4436–4449. [Google Scholar] [CrossRef]
  33. Chen, S.; Samingan, A.K.; Hanzo, L. Support vector machine multiuser receiver for DS-CDMA signals in multipath channels. IEEE Trans. Neural Netw. 2001, 12, 604–611. [Google Scholar] [CrossRef]
  34. Fu, W.; Lu, Q. Multiobjective optimal control of FOPID controller for hydraulic turbine governing systems based on reinforced multiobjective Harris Hawks optimization coupling with hybrid strategies. Complexity 2020, 2020, 9274980. [Google Scholar] [CrossRef]
  35. Fu, W.; Zhang, K.; Wang, K.; Wen, B.; Fang, P.; Zou, F. A hybrid approach for multi-step wind speed forecasting based on two-layer decomposition, improved hybrid DE-HHO optimization and KELM. Renew. Energy 2021, 164, 211–229. [Google Scholar] [CrossRef]
  36. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  37. Xiao, Y.; Kang, N.; Hong, Y.; Zhang, G. Misalignment fault diagnosis of DFWT based on IEMD energy entropy and PSO-SVM. Entropy 2017, 19, 6. [Google Scholar] [CrossRef]
  38. Mirjalili, S. Moth-Flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl. Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  39. Das, S.; Suganthan, P.N. Differential evolution: A survey of the state-of-the-art. IEEE Trans. Evol. Comput. 2010, 15, 4–31. [Google Scholar] [CrossRef]
  40. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl. Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  41. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  42. Haklı, H.; Uğuz, H. A novel particle swarm optimization algorithm with Levy flight. Appl. Soft Comput. 2014, 23, 333–345. [Google Scholar] [CrossRef]
  43. Zeng, G.-Q.; Chen, J.; Li, L.-M.; Chen, M.-R.; Wu, L.; Dai, Y.-X.; Zheng, C.-W. An improved multi-objective population-based extremal optimization algorithm with polynomial mutation. Inf. Sci. 2016, 330, 49–73. [Google Scholar] [CrossRef]
  44. Chan, R.H.; Tao, M.; Yuan, X. Constrained total variation deblurring models and fast algorithms based on alternating direction method of multipliers. SIAM J. Imaging Sci. 2013, 6, 680–697. [Google Scholar] [CrossRef]
  45. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  46. Rodríguez, L.; Castillo, O.; Soria, J.; Melin, P.; Valdez, F.; Gonzalez, C.I.; Martinez, G.E.; Soto, J. A fuzzy hierarchical operator in the grey wolf optimizer algorithm. Appl. Soft Comput. 2017, 57, 315–328. [Google Scholar] [CrossRef]
  47. Zhang, X.; Kang, Q.; Tu, Q.; Cheng, J.; Wang, X. Efficient and merged biogeography-based optimization algorithm for global optimization problems. Soft Comput. 2019, 23, 4483–4502. [Google Scholar] [CrossRef]
  48. Sun, L.; Chen, S.; Xu, J.; Tian, Y. Improved monarch butterfly optimization algorithm based on opposition-based learning and random local perturbation. Complexity 2019, 2019, 1–20. [Google Scholar] [CrossRef] [Green Version]
  49. Zhang, X.; Kang, Q.; Cheng, J.; Wang, X. A novel hybrid algorithm based on biogeography-based optimization and grey wolf optimizer. Appl. Soft Comput. 2018, 67, 197–214. [Google Scholar] [CrossRef]
  50. Fu, W.; Shao, K.; Tan, J.; Wang, K. Fault diagnosis for rolling bearings based on composite multiscale fine-sorted dispersion entropy and SVM with hybrid mutation SCA-HHO algorithm optimization. IEEE Access 2020, 8, 13086–13104. [Google Scholar] [CrossRef]
  51. Zhang, W.; Zhou, J. A comprehensive fault diagnosis method for rolling bearings based on refined composite multiscale dispersion entropy and fast ensemble empirical mode decomposition. Entropy 2019, 21, 680. [Google Scholar] [CrossRef] [Green Version]
  52. Azami, H.; Rostaghi, M.; Abasolo, D.; Escudero, J. Refined composite multiscale dispersion entropy and its application to biomedical signals. IEEE Trans. Biomed. Eng. 2017, 64, 2872–2879. [Google Scholar] [PubMed] [Green Version]
  53. Cheng, X.; Wang, P.; She, C. Biometric identification method for heart sound based on multimodal multiscale dispersion entropy. Entropy 2020, 22, 238. [Google Scholar] [CrossRef] [Green Version]
  54. Fahad, A.; Alshatri, N.; Tari, Z.; Alamri, A.; Khalil, I.; Zomaya, A.Y.; Foufou, S.; Bouras, A. A survey of clustering algorithms for big data: Taxonomy and empirical analysis. IEEE Trans. Emerg. Top. Comput. 2014, 2, 267–279. [Google Scholar] [CrossRef]
  55. Liu, Z.; Cao, H.; Chen, X.; He, Z.; Shen, Z. Multi-Fault classification based on wavelet SVM with PSO algorithm to analyze vibration signals from rolling element bearings. Neurocomputing 2013, 99, 399–410. [Google Scholar] [CrossRef]
Figure 1. The convergence curves of the seven algorithms on benchmark functions.
Figure 1. The convergence curves of the seven algorithms on benchmark functions.
Entropy 23 00527 g001
Figure 2. The flowchart of the fault identification with the proposed method.
Figure 2. The flowchart of the fault identification with the proposed method.
Entropy 23 00527 g002
Figure 3. The machinery fault simulator of the rolling bearings.
Figure 3. The machinery fault simulator of the rolling bearings.
Entropy 23 00527 g003
Figure 4. The fault state types of the bearings.
Figure 4. The fault state types of the bearings.
Entropy 23 00527 g004
Figure 5. The variation of central frequency with iteration under different K values.
Figure 5. The variation of central frequency with iteration under different K values.
Entropy 23 00527 g005
Figure 6. Time and frequent domain waveforms of different signals.
Figure 6. Time and frequent domain waveforms of different signals.
Entropy 23 00527 g006aEntropy 23 00527 g006b
Figure 7. The VMD decomposition results of different signals.
Figure 7. The VMD decomposition results of different signals.
Entropy 23 00527 g007aEntropy 23 00527 g007b
Figure 8. Comparison of evaluation values of different methods under 1800 rpm.
Figure 8. Comparison of evaluation values of different methods under 1800 rpm.
Entropy 23 00527 g008
Figure 9. The radar chart of evaluation results with different methods under 2200 rpm.
Figure 9. The radar chart of evaluation results with different methods under 2200 rpm.
Entropy 23 00527 g009
Figure 10. Boxplots of identification results with different methods: the x-axis tick labels correspond to: 1: VMD-FE-GWO-SVM; 2: VMD-FE-LCPGWO-SVM; 3: VMD-DE-GWO-SVM; 4: VMD-DE-LCPGWO-SVM; 5: VMD-RCMDE-GWO-SVM; 6: VMD-RCMDE-LCPGWO-SVM; 7: VMD-ARCMDE-GWO-SVM; 8: VMD-ARCMDE-LCPGWO-SVM.
Figure 10. Boxplots of identification results with different methods: the x-axis tick labels correspond to: 1: VMD-FE-GWO-SVM; 2: VMD-FE-LCPGWO-SVM; 3: VMD-DE-GWO-SVM; 4: VMD-DE-LCPGWO-SVM; 5: VMD-RCMDE-GWO-SVM; 6: VMD-RCMDE-LCPGWO-SVM; 7: VMD-ARCMDE-GWO-SVM; 8: VMD-ARCMDE-LCPGWO-SVM.
Entropy 23 00527 g010
Table 1. Overview of 12 benchmark functions.
Table 1. Overview of 12 benchmark functions.
No.FunctionRangeFmin
1 F 1 ( x ) = i = 1 n x i 2 [−100, 100]0
2 F 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | [−10, 10]0
3 F 3 ( x ) = i = 1 n ( j = 1 i x j ) 2 [−100, 100]0
4 F 4 ( x ) = max ( | x i | , 1 i n ) [−100, 100]0
5 F 5 ( x ) = i = 1 n 1 [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] [−30, 30]0
6 F 6 ( x ) = i = 1 n i x i 2 [−10, 10]0
7 F 7 ( x ) = i = 1 n ( 10 6 ) ( i 1 ) / ( n 1 ) x i 2 [−100, 100]0
8 F 8 ( x ) = 20 e ( 0.2 1 n i = 1 n x i 2 ) e ( 1 n i = 1 n cos ( 2 π x i ) ) + 20 + e [−32, 32]0
9 F 9 ( x ) = 1 4000 i = 1 n x i 2 + i = 1 n cos ( x i i ) + 1 [−100, 100]0
10 F 10 ( x ) = i = 1 n ( x i 2 10 cos ( 2 π x i ) ) + 10 n [−5.12, 5.12]0
11 F 11 ( x ) = i = 1 n | x i sin ( x i ) + 0.1 x i | [−10, 10]0
12 F 12 ( x ) = p n { 10 s i n ( p y 1 ) + i = 1 n 1 ( y i 1 ) 2 [ 1 + 10 s i n 2 ( p y i + 1 ) ] + ( y n 1 ) 2 } + i = 1 n u ( x i , 10 , 100 , 4 ) y i = 1 + x i + 1 4 u ( x i , a , m ) = { k ( x i a ) m x i > a 0 a < x i < a k ( x i a ) m x i < a [−50, 50]0
Table 2. Parameter settings of different optimization algorithms.
Table 2. Parameter settings of different optimization algorithms.
ModelsParameterDetermination ApproachRange Determined Value
PSOiteration numberpreset200
searching agentspreset40
dimensionspreset30
GWOiteration numberpreset200
searching agentspreset40
dimensionspreset30
SCAiteration numberpreset200
searching agentspreset40
dimensionspreset30
WOAiteration numberpreset200
searching agentspreset40
dimensionspreset30
MFOiteration numberpreset200
searching agentspreset40
dimensionspreset30
DEiteration numberpreset200
searching agentspreset40
dimensionspreset30
LCPGWOiteration numberpreset200
searching agentspreset40
dimensionspreset30
Table 3. The comparison results of the seven algorithms on benchmark functions.
Table 3. The comparison results of the seven algorithms on benchmark functions.
FunctionPSOGWOSCAWOAMFODELCPGWO
F1Max2.19 × 10−11.69 × 10−91.21 × 1039.99 × 10−61.86 × 1036.17 × 1011.29 × 10−181
Min2.14 × 10−26.60 × 10−114.32 × 1013.71 × 10−74.87 × 1022.28 × 1016.51 × 10−183
Mean1.17 × 10−13.62 × 10−104.33 × 1023.12 × 10−61.14 × 1033.46 × 1013.04 × 10−182
Std6.24 × 10−24.87 × 10−104.43 × 1023.07 × 10−64.22 × 1021.18 × 1010.00
F2Max2.22 × 1001.34 × 10−63.79 × 1001.28 × 10−45.33 × 1012.22 × 1001.82 × 10−83
Min4.04 × 10−17.07 × 10−72.46 × 10−12.19 × 10−51.03 × 1011.52 × 1001.20 × 10−84
Mean9.68 × 10−19.95 × 10−71.42 × 1004.85 × 10−52.99 × 1011.75 × 1007.58 × 10−84
Std5.24 × 10−11.88 × 10−71.07 × 1003.25 × 10−51.39 × 1012.10 × 10−15.04 × 10−84
F3Max5.36 × 1025.26 × 1003.16 × 1041.03 × 1024.52 × 1044.83 × 1041.01 × 10−180
Min2.45 × 1023.78 × 10−24.76 × 1032.36 × 1001.47 × 1043.31 × 1041.34 × 10−182
Mean3.48 × 1021.63 × 1001.70 × 1043.51 × 1012.50 × 1044.28 × 1042.42 × 10−181
Std1.02 × 1021.68 × 1007.81 × 1033.39 × 1019.84 × 1034.83 × 1030.00
F4Max2.75 × 1005.87 × 10−26.53 × 1018.65 × 10−18.06 × 1014.08 × 1018.75 × 10−97
Min1.88 × 1005.03 × 10−32.88 × 1011.11 × 10−15.16 × 1013.49 × 1012.56 × 10−97
Mean2.10 × 1001.80 × 10−25.27 × 1013.37 × 10−16.50 × 1013.79 × 1015.05 × 10−97
Std2.58 × 10−11.77 × 10−21.32 × 1012.68 × 10−19.07 × 1001.98 × 1002.11 × 10−97
F5Max5.11 × 1022.88 × 1011.38 × 1072.86 × 1011.20 × 1061.05 × 1042.24 × 101
Min6.75 × 1012.62 × 1015.75 × 1042.61 × 1011.02 × 1053.57 × 1031.00 × 101
Mean1.86 × 1022.76 × 1012.86 × 1062.75 × 1015.73 × 1056.52 × 1031.67 × 101
Std1.33 × 1029.23 × 10−14.50 × 1068.08 × 10−13.68 × 1052.24 × 1034.11 × 100
F6Max7.95 × 1002.33 × 10−101.47 × 1027.43 × 10−61.95 × 1036.83 × 1001.30 × 10−188
Min5.82 × 10−19.51 × 10−123.23 × 1001.64 × 10−88.74 × 1013.47 × 1001.22 × 10−189
Mean1.77 × 1007.16 × 10−115.54 × 1011.17 × 10−66.96 × 1024.72 × 1006.08 × 10−189
Std2.19 × 1006.86 × 10−114.91 × 1012.26 × 10−65.81 × 1029.98 × 10−10.00
F7Max4.26 × 1048.55 × 10−73.90 × 1051.15 × 10−21.54 × 1088.14 × 1043.81 × 10−166
Min1.16 × 1031.80 × 10−74.57 × 1036.78 × 10−41.42 × 1063.80 × 10441.25 × 10−167
Mean7.97 × 1035.56 × 10−71.09 × 1054.09 × 10−32.93 × 1076.39 × 1041.49 × 10−166
Std1.26 × 1042.28 × 10−71.11 × 1053.72 × 10−34.52 × 1071.56 × 1040.00
F8Max1.66 × 1004.62 × 10−62.04 × 1012.04 × 1011.99 × 1013.75 × 1007.99 × 10−15
Min1.75 × 10−11.57 × 10−63.45 × 1004.58 × 10−57.53 × 1002.95 × 1004.44 × 10−15
Mean1.12 × 1003.36 × 10−61.34 × 1016.07 × 1001.50 × 1013.38 × 1006.57 × 10−15
Std4.63 × 10−11.03 × 10−67.40 × 1009.78 × 1005.24 × 1002.46 × 10−11.83 × 10−15
F9Max4.96 × 10−27.78 × 10−22.06 × 1002.77 × 10−21.55 × 1009.73 × 10−10.00
Min5.88 × 10−32.21 × 10−128.26 × 10−13.56 × 10−81.11 × 1007.73 × 10−10.00
Mean1.94 × 10−21.08 × 10−21.13 × 1009.18 × 10−31.29 × 1008.71 × 10−10.00
Std1.22 × 10−22.44 × 10−23.65 × 10−11.06 × 10−21.21 × 10−16.18 × 10−20.00
F10Max1.63 × 1023.02 × 1011.61 × 1024.41 × 1012.55 × 1021.44 × 1020.00
Min6.25 × 1016.38 × 1002.92 × 1018.63 × 1001.17 × 1021.22 × 1020.00
Mean9.41 × 1011.60 × 1016.56 × 1012.09 × 1011.76 × 1021.32 × 1020.00
Std3.17 × 1017.64 × 1003.79 × 1011.14 × 1014.04 × 1017.99 × 1000.00
F11Max2.76 × 1005.67 × 10−31.18 × 1012.36 × 1001.63 × 1019.46 × 1002.54 × 10−2
Min6.25 × 10−12.34 × 10−33.03 × 10−12.40 × 10−34.10 × 1006.46 × 1007.47 × 10−69
Mean1.55 × 1003.58 × 10−34.85 × 1005.15 × 10−19.40 × 1007.99 × 1002.54 × 10−3
Std7.38 × 10−11.11 × 10−34.44 × 1007.79 × 10−14.08 × 1001.09 × 1008.03 × 10−3
F12Max−4.61 × 102−5.78 × 102−4.86 × 102−7.67 × 102−9.95 × 102−1.04 × 103−1.06 × 103
Min−9.78 × 102−7.16 × 102−5.88 × 102−8.71 × 102−1.06 × 103−1.06 × 103−1.06 × 103
Mean−6.96 × 102−6.39 × 102−5.29 × 102−8.24 × 102−1.05 × 103−1.06 × 103−1.06 × 103
Std1.39 × 1024.53 × 1013.40 × 1013.36 × 1012.14 × 1017.94 × 1002.40 × 10−13
Table 4. Description of the experimental data.
Table 4. Description of the experimental data.
Motor SpeedFault PositionNumber of Total SamplesNumber of Training SamplesNumber of Testing SamplesLabel
1800 rpmNormal614021L1
Inner race614021L2
Outer race614021L3
Ball fault614021L4
Combination fault614021L5
2200 rpmNormal614021L6
Inner race614021L7
Outer race614021L8
Ball fault614021L9
Combination fault614021L10
Table 5. The parameter settings of ARCMDE.
Table 5. The parameter settings of ARCMDE.
Parameter τ max mc t d
Value20461
Table 6. The calculation method of the four metrics.
Table 6. The calculation method of the four metrics.
AbbreviationExpression
ACC A C C = T P + T N T P + F N + F P + F N
ARI A R I = n 11 + n 00 C n 2
F F = 2 ( T P / ( T P + F P ) ) ( T P / ( T P + F N ) ) T P / ( T P + F P ) + T P / ( T P + F N )
NMI N M I = φ Φ ω Ω P ( φ , ω ) l o g ( P ( φ , ω ) / P ( φ ) P ( ω ) ) ( φ Φ P ( φ ) l o g ( P ( φ ) ) ) ( ω Ω P ( ω ) l o g ( P ( ω ) ) )
Table 7. Comparison results with different methods under variable sampling speeds.
Table 7. Comparison results with different methods under variable sampling speeds.
Motor SpeedMethodsBest CBest gEvaluation Metrics
ARINMIFACC
1800 rpmVMD-FE-GWO-SVM334.360813.26050.7697
[−0.0439, 0.0532]
0.8142
[−0.0477, 0.0545]
0.8885
[−0.0286, 0.0264]
0.8905
[−0.0238, 0.0238]
VMD-FE-LCPGWO-SVM14.7925.82590.7919
[−0.1008, 0.0650]
0.8360
[−0.0202, 0.0620]
0.9004
[−0.06130, 0.0320]
0.9029
[−0.0553, 0.0304]
VMD-DE-GWO-SVM2.393421.59670.8748
[−0.0544, 0.0775]
0.8845
[−0.0568, 0.0678]
0.9473
[−0.0240, 0.0334]
0.9476
[−0.0238, 0.0340]
VMD-DE-LCPGWO-SVM166.104138.88570.8791
[−0.0868, 0.0508]
0.8906
[−0.0732, 0.0382]
0.9490
[−0.0335, 0.0222]
0.9495
[−0.0352, 0.0219]
VMD-RCMDE-GWO-SVM445.52780.03530.9202
[−0.0716, 0.0327]
0.9298
[−0.0563, 0.0305]
0.9658
[−0.0371, 0.0151]
0.9667
[−0.0334, 0.0143]
VMD-RCMDE-LCPGWO-SVM708.82850.26940.9439
[−0.0348, 0.0319]
0.9488
[−0.0246, 0.0273]
0.9769
[−0.0159, 0.0136]
0.9771
[−0.0152, 0.0134]
VMD-ARCMDE-GWO-SVM683.770.250.9458
[−0.0639, 0.0300]
0.9500
[−0.0533, 0.0261]
0.9780
[−0.0248, 0.0125]
0.9781
[−0.0257, 0.0124]
VMD-ARCMDE-LCPGWO-SVM5.61240.24510.9597
[−0.0310, 0.0403]
0.9627
[−0.0342, 0.0373]
0.9838
[−0.0124, 0.0162]
0.9838
[−0.0124, 0.0162]
2200 rpmVMD-FE-GWO-SVM43.60130.44890.7216
[−0.1137,0.0993]
0.7573
[−0.0814,0.0705]
0.8725
[−0.0512,0.0509]
0.8733
[−0.0543,0.0505]
VMD-FE-LCPGWO-SVM72.13920.88060.7327
[−0.0991,0.0919]
0.7640
[−0.0868,0.0831]
0.8761
[−0.0472,0.0458]
0.8781
[−0.0495,0.0457]
VMD-DE-GWO-SVM20.57327.45120.8604
[−0.0651, 0.0447]
0.8732
[−0.0410, 0.0392]
0.9417
[−0.0263, 0.0202]
0.9419
[−0.0276, 0.0200]
VMD-DE-LCPGWO-SVM5.569.69020.8652
[−0.0663, 0.0877]
0.8815
[−0.0544, 0.0788]
0.9439
[−0.0316, 0.0370]
0.9438
[−0.0295, 0.0371]
VMD-RCMDE-GWO-SVM653.10940.46380.9220
[−0.0535, 0.0538]
0.9330
[−0.0346, 0.0430]
0.9675
[−0.0249, 0.0230]
0.9676
[−0.0248, 0.0229]
VMD-RCMDE-LCPGWO-SVM408.94630.20120.9242
[−0.0557, 0.0287]
0.9337
[−0.0352, 0.0267]
0.9684
[−0.0258, 0.0125]
0.9686
[−0.0257, 0.0124]
VMD-ARCMDE-GWO-SVM767.92400.00130.9271
[−0.0401, 0.0487]
0.9385
[−0.0377, 0.0376]
0.9693
[−0.0176, 0.0212]
0.9695
[−0.0171, 0.0210]
VMD-ARCMDE-LCPGWO-SVM172.45960.01850.9303
[−0.0461, 0.0455]
0.9381
[−0.0380, 0.0380]
0.9712
[−0.0207, 0.0193]
0.9714
[−0.0190, 0.0191]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shi, H.; Fu, W.; Li, B.; Shao, K.; Yang, D. Intelligent Fault Identification for Rolling Bearings Fusing Average Refined Composite Multiscale Dispersion Entropy-Assisted Feature Extraction and SVM with Multi-Strategy Enhanced Swarm Optimization. Entropy 2021, 23, 527. https://doi.org/10.3390/e23050527

AMA Style

Shi H, Fu W, Li B, Shao K, Yang D. Intelligent Fault Identification for Rolling Bearings Fusing Average Refined Composite Multiscale Dispersion Entropy-Assisted Feature Extraction and SVM with Multi-Strategy Enhanced Swarm Optimization. Entropy. 2021; 23(5):527. https://doi.org/10.3390/e23050527

Chicago/Turabian Style

Shi, Huibin, Wenlong Fu, Bailin Li, Kaixuan Shao, and Duanhao Yang. 2021. "Intelligent Fault Identification for Rolling Bearings Fusing Average Refined Composite Multiscale Dispersion Entropy-Assisted Feature Extraction and SVM with Multi-Strategy Enhanced Swarm Optimization" Entropy 23, no. 5: 527. https://doi.org/10.3390/e23050527

APA Style

Shi, H., Fu, W., Li, B., Shao, K., & Yang, D. (2021). Intelligent Fault Identification for Rolling Bearings Fusing Average Refined Composite Multiscale Dispersion Entropy-Assisted Feature Extraction and SVM with Multi-Strategy Enhanced Swarm Optimization. Entropy, 23(5), 527. https://doi.org/10.3390/e23050527

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop