Next Article in Journal
Seismic Response of Aeolian Sand High Embankment Slopes in Shaking Table Tests
Previous Article in Journal
An Improved SPEA2 Algorithm with Local Search for Multi-Objective Investment Decision-Making
Previous Article in Special Issue
A Fault Diagnosis Approach for Rolling Bearing Based on Convolutional Neural Network and Nuisance Attribute Projection under Various Speed Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Diagnosis for Rolling Bearing Based on Semi-Supervised Clustering and Support Vector Data Description with Adaptive Parameter Optimization and Improved Decision Strategy

1
College of Electrical Engineering & New Energy, China Three Gorges University, Yichang 443002, China
2
Hubei Provincial Key Laboratory for Operation and Control of Cascaded Hydropower Station, China Three Gorges University, Yichang 443002, China
3
Faculty of Mechanical and Material Engineering, Huaiyin Institute of Technology, Huai’an 223003, China
4
School of Hydropower and Information Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(8), 1676; https://doi.org/10.3390/app9081676
Submission received: 28 March 2019 / Revised: 16 April 2019 / Accepted: 19 April 2019 / Published: 23 April 2019
(This article belongs to the Special Issue Fault Detection and Diagnosis in Mechatronics Systems)

Abstract

:
Rolling bearing is of great importance in modern industrial products, the failure of which may result in accidents and economic losses. Therefore, fault diagnosis of rolling bearing is significant and necessary and can enhance the reliability and efficiency of mechanical systems. Therefore, a novel fault diagnosis method for rolling bearing based on semi-supervised clustering and support vector data description (SVDD) with adaptive parameter optimization and improved decision strategy is proposed in this study. First, variational mode decomposition (VMD) was applied to decompose the vibration signals into sets of intrinsic mode functions (IMFs), where the decomposing mode number K was determined by the central frequency observation method. Next, fuzzy entropy (FuzzyEn) values of all IMFs were calculated to construct the feature vectors of different types of faults. Later, training samples were clustered with semi-supervised fuzzy C-means clustering (SSFCM) for fully exploiting the information inside samples, whereupon a small number of labeled samples were able to provide sufficient data distribution information for subsequent SVDD algorithms and improve its recognition ability. Afterwards, SVDD with improved decision strategy (ID-SVDD) that combined with k-nearest neighbor was proposed to establish diagnostic model. Simultaneously, the optimal parameters C and σ for ID-SVDD were searched by the newly proposed sine cosine algorithm improved with adaptive updating strategy (ASCA). Finally, the proposed diagnosis method was applied for engineering application as well as contrastive analysis. The obtained results reveal that the proposed method exhibits the best performance in all evaluation metrics and has advantages over other comparison methods in both precision and stability.

1. Introduction

Rolling bearing is one of the most commonly used components in mechanical equipment, whose running state directly affects the accuracy, reliability, and service life of the whole machine [1]. Hence, how to recognize and diagnose rolling bearing faults remains one of the main concerns in preventing failures of mechanical systems [2]. However, rolling bearing is prone to failure due to the complex operating conditions, such as improper assembly, poor lubrication, water and foreign body invasion, corrosion or overload [3]. Therefore, effective methods need to be proposed to diagnosis a fault of rolling bearing, which can promote the reliability of industrial manufacture.
Owing to the rich information carried by vibration signals, most of fault diagnosis methods for rolling bearings rely on analyzing vibration signals [4,5], the resistance of the bearing can be measured using electrodynamic sensors or laser vibrometers [6]. Considering that vibration signals are commonly non-stationary, it is difficult to extract critical features directly from fault signals. To solve this problem, various non-stationary signal processing methods are proposed in previous studies. For instance, wavelet transform (WT) is a self-adaptive signal decomposition method [7], which inherits and develops the idea of localization of short-time Fourier transform and is an ideal tool for time-frequency analysis, but its generalization will be restricted once a wavelet basis is chosen. Empirical mode decomposition (EMD) has a strong ability to deal with non-stationary signals and does not need to pre-set any basis function [8], but its performance is affected by end effects and mode mixing. To overcome these defects, an improved version of EMD, ensemble empirical mode decomposition (EEMD), is proposed by introducing a noise-assisted analysis method [9], which is also increases the computational cost and cannot completely neutralize the added noise. Unlike the above methods, variational mode decomposition (VMD) is a novel quasi-orthogonal signal decomposition method which overcomes the generalization problem of WT or mode mixing and end effects of EMD and EEMD [10]. Its advancement and effectiveness have been demonstrated in previous studies [11,12]. With signals processed by non-stationary analytical method, the non-stationarity of signals would be weakened to some extent, and fault features can be further extracted. Due to the smoothness and robustness, as well as fewer samples required, fuzzy entropy (FuzzyEn) [13] is introduced to measure the features of vibration data and construct fault samples, which has been widely applied for feature extraction and shows superior performance in the field of fault diagnosis [14,15].
Fault diagnosis of rolling bearings is actually a problem of pattern recognition [16], and many machine learning methods have been developed for this problem [17]. Traditional machine learning methods require a large amount of labeled data to ensure the generalization of the algorithm. However, in practical problems, most of the samples are unlabeled, which results in unsatisfactory results of trained learning models in practical applications and makes a large number of unlabeled samples useless. Hence, semi-supervised methods are applied to fully exploit the information inside samples and expand the training set of machine learning algorithms. Semi-supervised fuzzy C-means clustering (SSFCM) [18] is a “semi-supervised version” of the fuzzy C-means clustering (FCM) algorithm. By introducing semi-supervised learning theory into FCM, training samples can be labeled under the supervision of the partly labeled samples, which makes up for the deficiency of labeled samples and improving the accuracy of pattern recognition [19,20]. In engineering application, several machine learning methods, including k-nearest neighbor (KNN) [21], artificial neural network (ANN) [22], support vector machine (SVM) [23] and support vector data description (SVDD) [24] are widely applied to solve pattern recognition problems. KNN is easy to realize and susceptible to the distribution of samples. ANN has good performance in dealing with non-linear problems under the circumstance of a large number of samples. Based on structural risk minimization and statistical learning theory, SVM proposed by Vapnik finds an optimal hyperplane that meets the classification requirements, thus the hyperplane maximizes the interval between two classes while ensuring classification effect. Correspondingly, SVDD is a kernel method inspired by the idea of SVM theory [25], which possesses a prominent recognition ability through establishing a minimum hypersphere that contains as many samples as possible [26]. Currently, combined with relative distance decision strategy, SVDD has been successfully employed in the field of pattern recognition [27,28]. However, traditional relative distance decision strategy is difficult to accurately classify unknown samples locating in overlap regions of the hyperspheres and regions outside all the hyperspheres. To solve this problem, SVDD based on an improved decision strategy (ID-SVDD) with the fusion of KNN method is proposed. To be specific, support vectors (SVs) are firstly extracted by establishing SVDD models. Next, the Euclidean distances between the testing sample and SVs are calculated. Whereupon, a fault type of the testing sample can be determined according to the proposed decision strategy.
Although SVDD has excellent effect in pattern recognition, its performance is affected by the parameters. In view of this, various optimization algorithms are proposed and applied to search the best parameters [29,30], including genetic algorithm (GA) [31], particle swarm optimization (PSO) [32], bacterial foraging algorithm (BFA) [33] and artificial sheep algorithm (ASA) [34]. As a novel optimization algorithm, sine cosine algorithm (SCA) proposed by Mirjalili [35] is proved to be effective in many studies [36,37]. To achieve better performance in convergence precision, a reformed sine cosine algorithm with adaptive strategy (ASCA) is developed in this paper. On the whole, a novel diagnosis method based on semi-supervised clustering and SVDD with adaptive parameter optimization and improved decision strategy is proposed in this study. Firstly, VMD is applied to decompose the vibration signals into sets of IMFs, and the decomposing mode number K is preset with central frequency observation method. Then FuzzyEn values of all IMFs are calculated to construct the feature vectors of corresponding samples with different fault types. Afterwards, training samples are clustered by SSFCM for information mining. Subsequently, SVDD with improved decision strategy (ID-SVDD) that combines with k-nearest neighbor is proposed to establish diagnostic model. Meanwhile, the optimal parameters C and σ for ID-SVDD is searched by the proposed adaptive sine cosine algorithm (ASCA). Finally, the engineering application and contrastive analysis indicate the availability and superiority of the proposed method.
The paper is organized as follows: Section 2 is dedicated to the basic knowledge of VMD, FuzzyEn, SSFCM, and SVDD. Section 3 introduces the proposed fault diagnosis method based on SSFCM and ASCA-optimized ID-SVDD. Section 4 presents the engineering application and comparative analysis, the experimental results of which demonstrate the superiority of the proposed method. Some discussion about the method presented in this paper is in Section 5. The conclusion is summarized in Section 6.

2. Fundamental Theories

2.1. Variational Mode Decomposition

Variational mode decomposition (VMD) is a novel non-recursive signal decomposition method [38], whose essence is to solve a variational optimization problem. Setting a scale K in advance, a given signal can be decomposed into K band-limited intrinsic mode functions (IMFs), among which each sub-signal is related to a certain center frequency. The constrained variation problem can be described as follows:
min m k , ω k { k = 1 K t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 } s . t . k = 1 K m k = f , k = 1 , 2 , , K
where m k = { m 1 , m 2 , , m k } represents the set of all mode functions, and ω k = { ω 1 , ω 2 , , ω k } represents the set of center frequencies, while t and δ ( t ) are the partial derivative of time t for the function and unit pulse function, respectively. f is the given real valued input signal.
Lagrange multipliers are introduced to obtain the optimal solution of above constrained variational problem. Then problem (1) can be specified as follows:
L ( m k , ω k , β ) = α k = 1 K t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 + f ( t ) k K m k ( t ) 2 2 + β ( t ) , f ( t ) k K m k ( t )
where α and β(t) represent the penalty factor and Lagrange multiplier respectively.
Next, the alternate direction method of multipliers (ADMM) and iterative search are utilized to obtain the saddle point of Lagrange multipliers [39]. The optimization problems of mk and 𝜔k are formulated as (3) and (4) respectively.
m k n + 1 = min { α t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 + f ( t ) i m i ( t ) + β ( t ) 2 2 2 }
ω k n + 1 = min { t [ ( δ ( t ) + j π t ) m k ( t ) ] e j ω k t 2 2 }
Solving problems (3) and (4), the iterative equations are deduced as follows:
m ^ k n + 1 ( ω ) = f ^ ( ω ) i k m ^ i ( ω ) + β ^ ( ω ) 2 1 + 2 α ( ω ω k ) 2
ω k n + 1 = 0 ω | m ^ k ( ω ) | 2 d ω 0 | m ^ k ( ω ) | 2 d ω
The Lagrange multipliers can be iterated with Equation (7).
β ^ n + 1 ( ω ) = β ^ n ( ω ) + τ ( f ( ω ) k m ^ k n + 1 ( ω ) )
where τ is an updating parameter.
The main steps of VMD can be summarized as follows:
Step 1: Initialize m k 1 , ω k 1 , β 1 , n = 1;
Step 2: Execute loop, n = n + 1;
Step 3: Update mk and 𝜔k based on Equations (5) and (6);
Step 4: Update β based on Equation (7);
Step 5: If k m ^ k n + 1 m ^ k n 2 2 / m ^ k n 2 2 < ε loop end, else turn to step 2 for next iteration.

2.2. Fuzzy Entropy

To solve the defect that sample entropy [40] adopts a hard threshold criterion which may result in an unstable discrimination result, fuzzy entropy [41] introduces an exponential function from fuzzy theory as the fuzzy entropy function. For the time series H = [h(1), h(2), ..., h(N)], set the fractal dimension m and then construct m-dimensional vector. The obtained m-dimensional vector is as follows:
H m ( i ) = [ h ( i ) , h ( i + 1 ) , , h ( i + m 1 ) ] g ( i )
where i = 1, 2, …, N − m + 1, and g ( i ) = 1 m j = 0 m 1 h ( i + j ) .
Define the distance dij between vector Hm(i) and Hm(j) as maximum absolute value of difference value between the corresponding element of them, that is:
d i j m = max | h ( i + k ) g ( i ) [ h ( j + k ) g ( j ) ] |
where i, j = 1, 2, …, N − m + 1 and i ≠ j; while k = 0, 1, …, m − 1.
Introducing fuzzy membership function, the similarity between vector Hm(i) and Hm(j) can be defined as:
A i j m = { 1 , d i j m = 0 exp [ l n 2 ( d i j m r ) 2 ] , d i j m > 0 }
where r is a positive real number, and r = R·std(H), std(H) is the standard deviation of the given time series.
Define function Φ m ( r ) as:
Φ m ( r ) = 1 N m + 1 i = 1 N m + 1 C i m
where C i m ( r ) = 1 N m j = 1 , j i N m + 1 A i j m .
Finally, FuzzyEn (with parameters m, N, r) of the time series can be deduced as the deviation of natural logarithm of Φ m from Φ m + 1 , which means:
F E ( m , N , r ) = ln Φ m ( r ) ln Φ m + 1 ( r )

2.3. Semi-Supervised Fuzzy C-Means Clustering

Semi-supervised fuzzy C-means clustering (SSFCM) [18] is an semi-supervised clustering algorithm based on the fuzzy C-means clustering (FCM) algorithm [42]. Specifically, supposing that the sample to be clustered is S = [s1, s2, , sn] and the number of clusters is c (2 ≤ cn), then the objective function of SSFCM is given as:
J w , α ( U , V ; S , F ) = i = 1 c j = 1 n u i j w d i j 2 + ρ i = 1 c j = 1 n ( u i j f i j b j ) w d i j 2
where Ucxn = [uij], uij represents the fuzzy membership of the jth sample belonging to the ith cluster, and the matrix formed by cluster centers is V = [v1, v2, …, vn]T. Fcxn = [fij], fij denotes the membership value of labeled samples, and bj is a two-valued (Boolean) indicator to distinguish labeled and unlabeled samples. Moreover, dij = ||sj-vi|| is the Euclidean distance between the jth sample and the ith cluster center, while w (w ≥ 1) is weighting parameter, and 𝜌 (𝜌 ≥ 0) is the balance coefficient.
With Lagrange multipliers introduced, the optimization problem of SSFCM can be converted into the following constrained minimization problem:
J w , α ( U , V , λ ; S , F ) = i = 1 c j = 1 n u i j w d i j 2 + ρ j = 1 n i = 1 c ( u i j f i j b j ) w d i j 2 j = 1 n λ j [ i = 1 c ( u i j 1 ) ]
where λj, j = 1, 2, …, n, is the Lagrange multiplier. In the sequel, fuzzy membership uij and cluster center vi can be iterated as:
u i j = 1 1 + ρ [ 1 + ρ ( 1 b j k = 1 c f k j ) k = 1 c d i j 2 / d k j 2 + ρ f i j b j ]
v i = j = 1 n u i j 2 s j / j = 1 n d i j 2
where i = 1, 2, …, c, j = 1, 2, …, n.
On the whole, primary procedures of SSFCM can be summarized as follows:
Step 1: Initialize membership matrix U and give values of c, b, F;
Step 2: Calculate prototypical cluster center V;
Step 3: Update U based on Equation (15);
Step 4: Compare U′ to U, if ||U-U′|| < δ (U′ is the last iteration result and δ is a certain tolerance) then loop end, else go to step 2.

2.4. Support Vector Data Description

Support vector data description (SVDD) [25] maps data set into a high-dimensional feature space and deduces a minimum hypersphere which contains as many target objects as possible. Given a target object set xi, i = 1, 2, …, n, a hypersphere of SVDD can be established. Accordingly, the sphere of the hypersphere can be described by center a and radius R of the sphere. The expression for the hypersphere is as follows:
min f = R 2 + C i ξ i s . t . | | x i a | | 2 R 2 + ξ i ,   ξ i 0 , i
where ξi is the slack variable, C is the penalty factor to reach a compromise between the size of hypersphere and number of misclassified samples. By introducing Lagrange multipliers, Equation (17) can be transformed into the following form:
L = R 2 + C i ξ i i α i ( R 2 + ξ i | | x i a | | 2 ) i β i ξ i
where α (α ≥ 0) and β (β ≥ 0) are Lagrange multipliers. Finally, a Gaussian radial basis function is applied to replace the inner product, transforming the optimization problem into the following dual problem:
max L = i α i K ( x i , x i ) i , j α i α j K ( x i , x j ) K ( x i , x j ) = exp ( | | x i x j | | 2 σ 2 ) s . t . i α i = 1 , a = i α i x i ,   0 α i C , i
where σ is the kernel parameter.
Solving programming problem (19), αi is obtained. To be specific, if 0 < αi < C, the target sample is on the hypersphere, which is called the support vector (SV); if ai = 0, the sample is in the hypersphere; if αi = C, the sample is outside the hypersphere, which is called bounded support vector (BSV). For a certain SV xp, radius of the hypersphere can be calculated as R. For any sample y to be tested, its distance from the center of the hypersphere is D:
R 2 = K ( x p , x p ) 2 i = 1 n α i K ( x i , x p ) + i , j = 1 n α i α j K ( x i x j )
D 2 = K ( y , y ) 2 i = 1 n α i K ( x i , y ) + i , j = 1 n α i α j K ( x i x j )

3. Fault Diagnosis Based on Semi-Supervised Clustering and Support Vector Data Description with Adaptive Parameter Optimization and Improved Decision Strategy

3.1. Improved Decision Strategy

SVDD with relative distance decision strategy (RD-SVDD) is commonly applied for pattern recognition. However, it is difficult for relative distance decision strategy to achieve precise identification, especially for unknown samples within overlap regions of the hyperspheres and regions outside all the hyperspheres. The schematic diagram of positional relationship between samples and hyperspheres is shown in Figure 1. Therefore, further development of decision strategy is required. To this end, SVDD with improved decision strategy (ID-SVDD) that fuses with the k-nearest neighbor (KNN) method is proposed in this research.
The KNN method proposed by Cover et al. [43] is one of the most widely used algorithms for pattern recognition, which is easy to understand and realize. The basic idea of KNN is to measure the differences between samples and their neighbors, and Euclidean and Manhattan distances are usually employed to achieve the measurement. Let p, q be two samples with m terms, the distances between them can be defined as:
Euclidean   distance :   D ( p , q ) = i = 1 m ( p i q i ) 2
Manhattan   distance :   D ( p , q ) = i = 1 m | p i q i |
where pi and qi are the ith component of p and q respectively.
With training samples labeled by semi-supervised clustering, SVDD models of different types of faults are established, from which hypersphere radiuses Rs and support vectors SVs can also be obtained. Furthermore, by comparing distance D from testing sample to the centers of hyperspheres, it can be judged whether the sample is within the hypersphere. There are two cases after the comparison: If the sample to be classified is only in one hypersphere, the fault type of the sample is directly ascertained; otherwise, if the sample is in the overlap regions of the hyperspheres or regions outside all the hyperspheres, KNN is needed for further classification. To be specific, KNN firstly calculates the distances between testing sample and SVs, finding k-nearest neighbors among SVs whose distances are closest to the testing sample, then the type of the testing sample can be determined according to majority fault types of k-nearest neighbors.
The main steps of ID-SVDD are shown below:
Step 1: Establish SVDD models for different types of faults based on Equations (19) and (20), thus Rs and SVs are obtained;
Step 2: Calculate distance D on the basis of Equation (21);
Step 3: Compare distance D to the radius R of each hypersphere, if testing sample is only in one hypersphere then go to step 6, else go to next step;
Step 4: Set KNN coefficient k, and find k-nearest neighbors according to Equation (22) or (23);
Step 5: Determine the type of testing sample with KNN decision;
Step 6: Decision completed.
The flowchart of ID-SVDD is shown in Figure 2.

3.2. Adaptive Parameter Optimization

3.2.1. Sine Cosine Algorithm

Exploration and exploitation [35] are two phases in which the sine cosine algorithm (SCA) processes the optimization problem. In the phase of exploration, a set of random solutions are initiated as the outset of optimization process. With a strong randomness, SCA is able to search for feasible solutions quickly in the searching space. In the phase of exploitation, the random solutions change gradually, and the randomness is distinctly weaker than that of exploration phase, which is conducive to a better local searching.
Let Z i = ( Z i 1 , Z i 2 , , Z i l ) T be the position of ith (i = 1, 2,…, M) individual, whereupon each solution of the optimization problem corresponds to the position of corresponding individual in the searching space, where l is the dimension of individuals. The best position of individual i is P i = ( P i 1 , P i 2 , , P i l ) T . During iterations, the position of individual i will be updated by the following Equations [35]:
Z i k + 1 = Z i k + r 1 × sin ( r 2 ) × | r 3 P i k Z i k | Z i k + 1 = Z i k + r 1 × cos ( r 2 ) × | r 3 P i k Z i k |
where Z i k is the position of individual i in kth iteration, r1, r2, and r3 are all random values.
The above equations are synthesized to:
Z i k + 1 = { Z i k + r 1 × sin ( r 2 ) × | r 3 P i k Z i k | , r 4 < 0.5 Z i k + r 1 × cos ( r 2 ) × | r 3 P i k Z i k | , r 4 0.5
where r4 is a random value within [0, 1].
As the above equations show, the updating equation has four main parameters, i.e., r1, r2, r3 and r4. The parameter r1 contributes to determine the position’s region of individual i at next iteration. The parameter r2 is a random number in the scope of [0, 2π], which defines distance that next movement should be towards. To stochastically emphasize (r3 > 1) or deemphasize (r3 < 1) the effect of the best position of individual, a random weight r3 within [0, 2] is brought in the equations. Finally, the parameter r4 is a random number in the range of [0, 1] to switch fairly between sine and cosine components, specifically when r4 < 0.5, sine component dominates iterations of the position of individual i, otherwise cosine component dominates iterations.
To seek feasible solutions in the searching space and eventually find the global optimum, exploration and exploitation phases should be in a balanced state. For this reason, the amplitudes of the sine and cosine functions are adaptively adjusted by changing r 1 as follows [35]:
r 1 = a t a T
where T, t and a are the maximum number of iterations, the current number of iterations and a constant respectively.

3.2.2. Adaptive Sine Cosine Algorithm

As mentioned in SCA, r1 is a crucial parameter connecting with the sine and cosine functions r1sin(r2) or r1cos(r2). On the basis of the design principle of SCA, the algorithm firstly explores different regions and then exploits promising regions. Accordingly, the value of r1 should be monotone decreasing during iterations. However, the updating strategy of r1 as shown in Equation (26) is a linear decrease with equal-step, which will restrict the convergence and accuracy. For this purpose, an improved strategy for updating r1 which is called adaptive sine cosine algorithm (ASCA) is proposed in this paper.
Considering that the fitness values of all individuals would change among the iterations, a hierarchical strategy based on the fitness values of individuals is introduced to divide all individuals into three subgroups [44]. More specifically, let fi be the fitness value of individual i, fa be the average fitness value of all individuals, and fe represents the average fitness value of individuals that fitness value better than fa. If fi is better than fe, individual i is defined as elite individual; if fi is better than fa but worse than fe, individual i is defined as ordinary individual; if fi is worse than fa, then define individual i as inferior individual. The hierarchical strategy is illustrated in Figure 3a. After the above process, the number of individuals in different subgroups can be adaptively adjusted according to the fitness values during iterations, and different subgroups implement different adaptive operations. The updating strategy for r1 is defined as follows and changes of r1 for individuals of different subgroups is shown in Figure 3b:
(1) Elite individuals: Are the best individuals among all ones, and close to the optimal value. A smaller r1 is given with a cubic function to enhance local searching:
r 1 = ( r max r min ) ( T t T ) 3 + r min
where rmax and rmin are the maximum and minimum value of r1 respectively, while T and t are the maximum and current number of iterations respectively.
(2) Ordinary individuals: Possess good global and local searching ability. r1 maintains the original linear decreasing trend:
r 1 = a t a T
(3) Inferior individuals: Are far from the optimal value. A bigger r1 is needed to enhance global searching in early iteration, and to jump out the local optimum later:
r 1 = ( r min r max ) e b t T + r max
where b is a positive constant to adjust exponential curve.

3.3. Fault Diagnosis Based on SSFCM and ID-SVDD Optimized by ASCA

In this section, a novel model based on semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with improved decision strategy (SSFCM-ASCA-ID-SVDD) is proposed to diagnose faults for rolling bearing. The specific steps are detailed as follows:
Step 1: Collect vibration signals;
Step 2: Determine the mode number K by central frequency observation method;
Step 3: Decompose the signals into sets of IMFs with VMD;
Step 4: Calculate FuzzyEn values of all IMFs, and construct the fault feature vectors of training and testing samples;
Step 5: Cluster training samples by SSFCM, thus the training samples are all labeled;
Step 6: Optimize the parameters C and σ for SVDD with the proposed ASCA;
Step 7: Train the ID-SVDD model with the optimal parameters C and σ;
Step 8: Apply the optimal ID-SVDD model to classify different types of faults and evaluate the performance of the model.
The flowchart of the proposed fault diagnosis model is shown in Figure 4.

4. Engineering Application

4.1. Data Collection

To validate the performance of the proposed method, a series of vibration signals with different fault locations and sizes were gathered from Bearings Data Center of Case Western Reserve University [45]. The experiment stand was mainly composed of a motor, an accelerometer, a torque sensor/encoder, and a dynamometer. The bearing was a deep groove ball bearing with model SKF6205-2RS. By using electro-discharge machining (EDM), the experiment device simulated three fault states of the rolling bearing: The inner race fault, ball element fault, and outer race fault. The depth of faults was 0.011 inches. Vibration signals collected from the drive end (DE) were taken as the research objects. In the experiment, the sample frequency was 12,000 Hz, and the rotation speed was 1750 rpm under the rated load of 2 hp. To fully verify the validity of the proposed fault diagnosis method, 9 types of samples were used in this paper, namely, the inner race fault, ball fault, and outer race fault with diameters of 0.007, 0.014, and 0.021 inches (i.e., each of the three types of faults has three defect sizes). Further, the vibration signal of each type of fault was divided into 59 segments containing 2048 sampling points. An image of the experiment device is shown in Figure 5. The experimental signals are listed in Table 1.

4.2. Application to Fault Diagnosis of Rolling Bearing

In order to demonstrate the effectiveness of the proposed SSFCM-ASCA-ID-SVDD method, experiments were implemented by contrast with constrained-K-means (CK-means) [46] in the clustering phase and with SCA in the parameter optimization phase. Similarly, support vector machine (SVM) and SVDD with relative distance decision strategy (RD-SVDD) were applied for comparison during decision making phase. In the proposed method, the first step was to decompose the signal of each type of fault into a set of components with VMD, whose mode number K needed to be determined in advance with central frequency observation method. If the value of K is too large, center frequencies of adjacent IMFs may be too close, and mode mixing will emerge. The center frequencies of all IMFs when taking different K are listed in Table 2, where K was ascertained by the sample of inner race fault with diameter of 0.007 inches (L1). As illustrated in Table 2, when K was set to 5, the first two center frequencies were relatively close, which meant that excessive decomposition occurred. The same conclusion can also be seen in Figure 6, during the iterative calculation of VMD, if K was 5 or greater, some center frequencies of IMFs would be relatively close to each other. For example, when K was 5, IMF1 and IMF2 were relatively close, while K was 6, IMF2 and IMF3 were relatively close. Therefore, parameter K was set to 4 in this research.
The decomposition results of samples with different fault types are illustrated in Figure 7, from which it can be seen that waveforms of different fault samples are different to some extent. After VMD decomposition, FuzzyEn values were calculated to construct the fault feature vectors, and the fractal dimension m was set 2 while the positive real number r was set as 0.2. The first three FuzzyEn vectors of different fault samples (L1–L9) are reported in Table 3.
The feature vectors of samples from different fault types were divided randomly into two parts with 40 and 19 vectors, respectively, the first part was used for training and the second one for testing. In order to better illustrate the characteristics and effects of SSFCM, training samples with ratio of 0.3 and 0.5 were randomly selected as labeled samples, while the rest were samples with unknown fault types. The weighting parameter w of SSFCM was set to 2 while the balance coefficient α was set to 1/0.3 and 1/0.5, respectively, according to the ratio. Figure 8 depicts the 2D projection of training samples after clustering. The figure shows that training samples are gathered into nine clusters under the supervision of labeled samples and distributed near the center of each cluster. The contour-map displays the samples with the same partitioning values.
In the optimization phase, the proposed ASCA approach was applied to optimize the penalty factor C and the kernel parameter σ, where the individual number was set to 30 and iteration times were set to 100. Considering the constraints of the optimization problem of SVDD, C should subject to C ≥ 1/n, where n is the number of target objects. Likewise, C should be subject to C < 1 or else it loses constraint on Lagrange multipliers. Taken together, searching ranges of C and σ were [1/n, 1] and [2−10, 210], respectively. The coefficient k was set to 3 for KNN decision. During optimization process, the fitness value was calculated by five-fold cross validation on the training samples.
The optimization effect with the given (C, σ) was measured by average accuracy of cross validation, then the optimal (C, σ) was obtained. The convergence procedure of the adaptive parameter optimization is depicted in Figure 9a. It can be seen that the absolute average fitness value of all individuals increases rapidly at early stage of iterations and then tends to be stable, which means the individuals were moving closer to the global optimal solution. The shaded part shows the distribution of the convergence curve in 10 optimization experiments. The comparison of ASCA and SCA is shown in Figure 9b, where all convergence experiments use the same FuzzyEn values and each convergence curve is the average of 10 experiments. The figure shows that both ASCA and SCA have a good convergence effect, but ASCA keeps a better convergence effect than SCA in the whole iterations. Besides, in the middle and later stages of iterations, the convergence curve of SCA decreases slightly, while ASCA remains stable.
With the optimal parameters (C, σ) and labeled training samples, the ID-SVDD model was trained and applied to recognize testing samples. For an in-depth verification of the proposed method, diagnosis results were averaged over 10 runs and the training samples were chosen at random in every run. In this application, normalized mutual information (NMI) as well as accuracy (ACC) as two widely used evaluation metrics are employed to evaluate the diagnosis results [47], which reflects the matching degree between classification results and real sample distribution information. The scope of two metrics is both [0, 1]. The value closer to 1 indicates a better classification performance of the model. In addition, corresponding deviation scope for each result was also calculated as a reference for evaluation. The optimal (C, σ) is presented corresponding to the best accuracy among 10 runs.
In contrastive experiments, parameters of CK-means, SCA, and RD-SVDD were set similar to the proposed method. That is, training samples with a ratio of 0.3 and 0.5 were randomly selected as labeled samples for CK-means clustering; SCA had 30 individuals and iterated 100 times; the ranges of C and σ of RD-SVDD were [1/n, 1] and [2−10, 210] respectively. Considering the difference between SVDD and SVM, the searching scopes of parameter C and g of SVM were both set as [2−10, 210] respectively. Moreover, the optimal parameters C and σ(g) of all contrastive methods were selected in the same way the proposed method done. The performance of all contrastive methods was evaluated by NMI and ACC as well. A list of all important parameters in the experiment is shown in Table 4.
The fault diagnosis results with different methods are shown in Table 5 and Figure 10. It can be seen from Table 5 that the proposed SSFCM-ASCA-ID-SVDD method achieves the best performance in terms of NMI and ACC metrics among all methods, i.e., 0.0868 and 0.9333 when the ratio is 0.3, 0.9254 and 0.09649 when the ratio is 0.5. To be specific, in the case of the ratio of 0.5: The comparison of FCM-ASCA-ID-SVDD, CK-means-ASCA-ID-SVDD and the proposed method indicates that NMI value of proposed method is 0.2913 and 0.0377 higher than other two methods, as well as that ACC value is 0.1228 and 0.0263 higher than other two methods, which shows SSFCM possess a better clustering precision. Similarly, the comparison analysis between SSFCM-SCA-ID-SVDD and the proposed method shows that ASCA-optimized ID-SVDD improves NMI value by 0.0152 and ACC value by 0.0099 than SCA-optimized ID-SVDD, indicating the effectiveness of the proposed adaptive parameter optimization strategy. Furthermore, the contrast among SSFCM-ASCA-SVM, SSFCM-ASCA-RD-SVDD and the proposed method shows that the NMI evaluation of proposed method is 0.0335 and 0.0568 higher than that of other two methods, while the ACC evaluation of proposed method is 0.0263 and 0.0409 higher than that of other two methods. Thus, it can be concluded that the proposed ID-SVDD model with improved decision strategy has better diagnosis performance.
Additionally, through the comprehensive analysis with a labeled sample ratio of 0.3 and 0.5, it showed that the proposed method achieved the best results under both the two ratio conditions. Meanwhile, the metrics under a ratio of 0.5 was better than that under a ratio of 0.3, which manifests that the clustering performance of SSFCM would be improved with the increasing of the labeled ratio, thus improving the diagnostic accuracy. Figure 11 is the boxplots of evaluation values, illustrating the performance of different diagnosis methods. As shown in the figure, the proposed SSFCM-ASCA-ID-SVDD method achieves a higher precision and possess better stability than other contrastive methods.

5. Discussion

The superiority of the proposed ASCA and ID-SVDD methods have been effectively demonstrated by the above analysis. Likewise, the accuracy of the diagnosis model has also been verified by contrastive experiments on various fault sizes and locations. This preliminary study mainly deals with the fault diagnosis of vibration signals on the rated load. For the practical application in real industrial operating conditions, variable load conditions should also be considered. Those application situations will be researched in our future work. Besides, the proposed ASCA algorithm can be optimized through better strategies to further improve the convergence speed and global search ability. Moreover, the multi-objective optimization that has been widely utilized in the field of controlling [48,49,50] could be implemented in fault diagnosis, which is expected to improve the accuracy and reduce the variance of the outputs of the model [51].

6. Conclusions

A novel fault diagnosis method for rotating bearing based on semi-supervised clustering and support vector data description with adaptive parameter optimization and improved decision strategy is presented in this study. Due to the non-stationarity of the original signals, signals collected from different types of faults were firstly split by VMD into sets of IMFs, before which process, the decomposing mode number K was determined by central frequency observation method. Next, the FuzzyEn values of all IMFs were calculated to construct the feature vectors of different types of faults. Subsequently, all training samples were labeled by using SSFCM under the supervision of partly labeled samples. Finally, SVDD with adaptive parameter optimization and improved decision strategy, i.e., ASCA-ID-SVDD, was proposed to diagnose different fault samples. In an engineering application, the contrastive experiments were implemented between the proposed SSFCM-ASCA-ID-SVDD method and other relevant methods. In the sequel, nine types of faults with different locations and sizes were successfully diagnosed, and the results show that the proposed method exhibited the best performance in both NMI and ACC evaluations. Particularly, when training samples with the ratio of 0.5 that are randomly selected as the labeled samples, the NMI value of the proposed method is 0.0335, 0.2913, 0.0377, 0.0568, and 0.0152 higher than that of SSFCM-ASCA-SVM, FCM-ASCA-ID-SVDD, CK-means-ASCA-ID-SVDD, SSFCM-ASCA-RD-SVDD, and SSFCM-SCA-ID-SVDD respectively. Likewise, the ACC value of the proposed method is 0.0263, 0.1228, 0.0263, 0.0409, and 0.0099 higher than that of other methods respectively. Moreover, the boxplots of different methods illustrate the precision and stability of the proposed method. In summary, the proposed method shows better superiority to other methods and achieves preferable diagnostic performance.

Author Contributions

J.T. was responsible for experiments and paper writing, W.F. conceived the research and designed methodology, K.W. helped analyze results, X.X. and W.H. provided recommendations for this research, Y.S. participated in the discussion.

Funding

This work was funded by the National Natural Science Foundation of China (NSFC) (51741907, 51709122), Hubei Provincial Major Project for Technical Innovation (2017AAA132), the Open Fund of Hubei Provincial Key Laboratory for Operation and Control of Cascaded Hydropower Station (2017KJX06), Research Fund for Excellent Dissertation of China Three Gorges University (2019SSPY070).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yi, C.; Lv, Y.; Xiao, H.; You, G.; Dang, Z. Research on the blind source separation method based on regenerated phase-shifted sinusoid-assisted EMD and its application in diagnosing rolling-bearing faults. Appl. Sci. 2017, 7, 414. [Google Scholar] [CrossRef]
  2. Riaz, S.; Elahi, H.; Javaid, K.; Shahzad, T. Vibration feature extraction and analysis for fault diagnosis of rotating machinery-a literature survey. Asia Pac. J. Multidiscip. Res. 2017, 5, 103–110. [Google Scholar]
  3. Mbo’O, C.P.; Hameyer, K. Fault diagnosis of bearing damage by means of the linear discriminant analysis of stator current features from the frequency selection. IEEE Trans. Ind. Appl. 2016, 52, 3861–3868. [Google Scholar] [CrossRef]
  4. Yuan, R.; Lv, Y.; Song, G. Fault diagnosis of rolling bearing based on a novel adaptive high-order local projection denoising method. Complexity 2018, 2018, 3049318. [Google Scholar] [CrossRef]
  5. Fu, W.; Tan, J.; Li, C.; Zou, Z.; Li, Q.; Chen, T. A hybrid fault diagnosis approach for rotating machinery with the fusion of entropy-based feature extraction and SVM optimized by a chaos quantum sine cosine algorithm. Entropy 2018, 20, 626. [Google Scholar] [CrossRef]
  6. Adamczak, S.; Stepien, K.; Wrzochal, M. Comparative study of measurement systems used to evaluate vibrations of rolling bearings. Procedia Eng. 2017, 192, 971–975. [Google Scholar] [CrossRef]
  7. Kankar, P.K.; Sharma, S.C.; Harsha, S.P. Rolling element bearing fault diagnosis using wavelet transform. Neurocomputing. 2011, 74, 1638–1645. [Google Scholar] [CrossRef]
  8. Ali, J.B.; Fnaiech, N.; Saidi, L.; Chebel-Morello, B.; Fnaiech, F. Application of empirical mode decomposition and artificial neural network for automatic bearing fault diagnosis based on vibration signals. Appl. Acoust. 2015, 89, 16–27. [Google Scholar]
  9. Fu, W.; Zhou, J.; Zhang, Y.; Zhu, W.; Xue, X.; Xu, Y. A state tendency measurement for a hydro-turbine generating unit based on aggregated EEMD and SVR. Meas. Sci. Technol. 2015, 26, 125008. [Google Scholar] [CrossRef]
  10. Li, K.; Su, L.; Wu, J.; Wang, H.; Chen, P. A rolling bearing fault diagnosis method based on variational mode decomposition and an improved kernel extreme learning machine. Appl. Sci. 2017, 7, 1004. [Google Scholar] [CrossRef]
  11. Fu, W.; Wang, K.; Li, C.; Li, X.; Li, Y.; Zhong, H. Vibration trend measurement for a hydropower generator based on optimal variational mode decomposition and an LSSVM improved with chaotic sine cosine algorithm optimization. Meas. Sci. Technol. 2019, 30, 015012. [Google Scholar] [CrossRef]
  12. Fu, W.; Wang, K.; Zhou, J.; Xu, Y.; Tan, J.; Chen, T. A hybrid approach for multi-step wind speed forecasting based on multi-scale dominant ingredient chaotic analysis, KELM and synchronous optimization strategy. Sustainability 2019, 11, 1804. [Google Scholar] [CrossRef]
  13. Xie, H.; Sivakumar, B.; Boonstra, T.W.; Mengersen, K. Fuzzy entropy and its application for enhanced subspace filtering. IEEE Trans. Fuzzy Syst. 2018, 26, 1970–1982. [Google Scholar] [CrossRef]
  14. Li, Y.; Xu, M.; Wang, R.; Huang, W. A fault diagnosis scheme for rolling bearing based on local mean decomposition and improved multiscale fuzzy entropy. J. Sound Vib. 2016, 360, 277–299. [Google Scholar] [CrossRef]
  15. Zhu, X.; Zheng, J.; Pan, H.; Bao, J.; Zhang, Y. Time-shift multiscale fuzzy entropy and laplacian support vector machine based rolling bearing fault diagnosis. Entropy 2018, 20, 602. [Google Scholar] [CrossRef]
  16. Fu, W.; Tan, J.; Zhang, X.; Chen, T.; Wang, K. Blind parameter identification of MAR model and mutation hybrid GWO-SCA optimized SVM for fault diagnosis of rotating machinery. Complexity 2019, 2019, 3264969. [Google Scholar] [CrossRef]
  17. Cerrada, M.; Sánchez, R.; Li, C.; Pacheco, F.; Cabrera, D.; José, V.D.O.; Vásquez, R.E. A review on data-driven fault severity assessment in rolling bearings. Mech. Syst. Signal Process. 2018, 99, 169–196. [Google Scholar] [CrossRef]
  18. Pedrycz, W. Algorithms of fuzzy clustering with partial supervision. Pattern Recognit. Lett. 1985, 3, 13–20. [Google Scholar] [CrossRef]
  19. Xu, C.; Zhang, P.; Ren, G.; Fu, J. Engine wear fault diagnosis based on improved semi-supervised fuzzy C-means clustering. J. Mech. Eng. 2011, 47, 55. [Google Scholar] [CrossRef]
  20. Arshad, A.; Riaz, S.; Jiao, L. Semi-supervised deep fuzzy c-mean clustering for imbalanced mulit-class classification. IEEE Access 2019, 7, 28100–28112. [Google Scholar] [CrossRef]
  21. Yao, B.; Su, J.; Wu, L.; Guan, Y. Modified local linear embedding algorithm for rolling element bearing fault diagnosis. Appl. Sci. 2017, 7, 1178. [Google Scholar] [CrossRef]
  22. Raj, N.; Jagadanand, G.; George, S. Fault detection and diagnosis in asymmetric multilevel inverter using artificial neural network. Int. J. Electron. 2017, 105, 559–571. [Google Scholar] [CrossRef]
  23. Chen, H.; Wang, J.; Li, J.; Tang, B. A texture-based rolling bearing fault diagnosis scheme using adaptive optimal kernel time frequency representation and uniform local binary patterns. Meas. Sci. Technol. 2017, 28, 035903. [Google Scholar] [CrossRef]
  24. Wang, H.; Chen, J. Performance degradation assessment of rolling bearing based on bispectrum and support vector data description. J. Vib. Control 2014, 20, 2032–2041. [Google Scholar] [CrossRef]
  25. Tax, D.M.J.; Duin, R.P.W. Support vector data description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
  26. Zhou, J.; Fu, W.; Zhang, Y.; Xiao, H.; Xiao, J.; Zhang, C. Fault diagnosis based on a novel weighted support vector data description with fuzzy adaptive threshold decision. Trans. Inst. Measure. Control 2018, 40, 71–79. [Google Scholar] [CrossRef]
  27. Zhou, J.; Guo, H.; Zhang, L.; Xu, Q.; Li, H. Bearing performance degradation assessment using lifting wavelet packet symbolic entropy and SVDD. Shock Vib. 2016, 2016, 1–10. [Google Scholar] [CrossRef]
  28. Wang, H. Fault diagnosis of rolling element bearing compound faults based on sparse no-negative matrix factorization-support vector data description. J. Vib. Control 2018, 24, 272–282. [Google Scholar] [CrossRef]
  29. Liu, Y.; Qin, H.; Zhang, Z.; Yao, L.; Wang, C.; Mo, L.; Ouyang, S.; Li, J. A region search evolutionary algorithm for many-objective optimization. Inform. Sci. 2019. [Google Scholar] [CrossRef]
  30. Fu, W.; Tan, J.; Xu, Y.; Wang, K.; Chen, T. Fault diagnosis for rolling bearings based on fine-sorted dispersion entropy and SVM optimized with mutation SCA-PSO. Entropy 2019, 21, 404. [Google Scholar] [CrossRef]
  31. Zhang, C.; Peng, T.; Li, C.; Fu, W.; Xia, X.; Xue, X. Multiobjective optimization of a fractional-order PID controller for pumped turbine governing system using an improved NSGA-III algorithm under multiworking conditions. Complexity 2019, 2019, 5826873. [Google Scholar] [CrossRef]
  32. Duan, L.; Xie, M.; Bai, T.; Wang, J. A new support vector data description method for machinery fault diagnosis with unbalanced datasets. Expert Syst. Appl. 2016, 64, 239–246. [Google Scholar] [CrossRef]
  33. Xu, Y.; Zheng, Y.; Du, Y.; Yang, W.; Peng, X.; Li, C. Adaptive condition predictive-fuzzy PID optimal control of start-up process for pumped storage unit at low head area. Energy Convers. Manag. 2018, 177, 592–604. [Google Scholar] [CrossRef]
  34. Xu, Y.; Li, C.; Wang, Z.; Zhang, N.; Peng, B. Load frequency control of a novel renewable energy integrated micro-grid containing pumped hydropower energy storage. IEEE Access 2018, 6, 29067–29077. [Google Scholar] [CrossRef]
  35. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  36. Fu, W.; Wang, K.; Li, C.; Tan, J. Multi-step short-term wind speed forecasting approach based on multi-scale dominant ingredient chaotic analysis, improved hybrid GWO-SCA optimization and ELM. Energy Convers. Manag. 2019, 187, 356–377. [Google Scholar] [CrossRef]
  37. Elaziz, M.A.; Oliva, D.; Xiong, S. An improved opposition-based sine cosine algorithm for global optimization. Expert Syst. Appl. 2017, 90, 484–500. [Google Scholar] [CrossRef]
  38. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
  39. Chan, R.H.; Tao, M.; Yuan, X. Constrained total variation deblurring models and fast algorithms based on alternating direction method of multipliers. SIAM J. Imaging Sci. 2013, 6, 680–697. [Google Scholar] [CrossRef]
  40. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Phys. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  41. Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of surface EMG signal based on fuzzy entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef] [PubMed]
  42. Bezdek, J.C.; Ehrlich, R.; Full, W. FCM: The fuzzy c -means clustering algorithm. Comput. Geosci. 1984, 10, 191–203. [Google Scholar] [CrossRef]
  43. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  44. Wu, H.; Zhu, C.; Chang, B.; Liu, J. Adaptive genetic algorithm to improve group premature convergence. J. Xi’an Jiaotong Univ. 1999, 33, 27–30. (In Chinese) [Google Scholar]
  45. Bearing Data Center of the Case Western Reserve University. Available online: http://csegroups.case.edu/bearingdatacenter/pages/download-data-file (accessed on 28 September 2018).
  46. Basu, S.; Banerjee, A.; Mooney, R.J. Semi-supervised clustering by seeding. In Proceedings of the Nineteenth International Conference on Machine Learning, Sydney, Australia, 8–12 July 2002; pp. 19–26. [Google Scholar]
  47. Fahad, A.; Alshatri, N.; Tari, Z.; Alamri, A.; Khalil, I.; Zomaya, A.Y.; Foufou, S.; Bouras, A. A survey of clustering algorithms for big data: Taxonomy and empirical analysis. IEEE Trans. Emerg. Top. Comput. 2014, 2, 267–279. [Google Scholar] [CrossRef]
  48. Lai, X.; Li, C.; Zhou, J.; Zhang, N. Multi-objective optimization for guide vane shutting based on MOASA. Renew. Energy 2019, 139, 302–312. [Google Scholar] [CrossRef]
  49. Li, C.; Wang, W.; Chen, D. Multi-objective complementary scheduling of hydro-thermal-RE power system via a multi-objective hybrid grey wolf optimizer. Energy 2019, 171, 241–255. [Google Scholar] [CrossRef]
  50. Liu, Y.; Qin, H.; Mo, L.; Wang, Y.; Chen, D.; Pang, S.; Yin, X. Hierarchical flood operation rules optimization using multi-objective cultured evolutionary algorithm based on decomposition. Water Resour. Manag. 2019, 33, 337–354. [Google Scholar] [CrossRef]
  51. Jiang, L.; Xuan, J.; Shi, T. Feature extraction based on semi-supervised kernel marginal fisher analysis and its application in bearing fault diagnosis. Mech. Syst. Signal Process. 2013, 41, 113–126. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of positional relationship between samples and hyperspheres.
Figure 1. Schematic diagram of positional relationship between samples and hyperspheres.
Applsci 09 01676 g001
Figure 2. The flowchart of support vector data description with improved decision strategy (ID-SVDD).
Figure 2. The flowchart of support vector data description with improved decision strategy (ID-SVDD).
Applsci 09 01676 g002
Figure 3. Schematic diagram of the proposed adaptive sine cosine algorithm (ASCA) based on hierarchical strategy: (a) Hierarchy of individuals (fitness value decreases from top down), (b) changes of r1 for individuals of different subgroups.
Figure 3. Schematic diagram of the proposed adaptive sine cosine algorithm (ASCA) based on hierarchical strategy: (a) Hierarchy of individuals (fitness value decreases from top down), (b) changes of r1 for individuals of different subgroups.
Applsci 09 01676 g003
Figure 4. The flowchart of the proposed fault diagnosis method.
Figure 4. The flowchart of the proposed fault diagnosis method.
Applsci 09 01676 g004
Figure 5. Experiment stand and its measurement sensors in bearing data center.
Figure 5. Experiment stand and its measurement sensors in bearing data center.
Applsci 09 01676 g005
Figure 6. Center frequency distribution with different K values.
Figure 6. Center frequency distribution with different K values.
Applsci 09 01676 g006
Figure 7. The variational mode decomposition (VMD) decomposition results of signals with different faults: (a) inner race (0.007 inches); (b) inner race (0.014 inches); (c) inner race (0.021 inches); (d) ball (0.007 inches); (e) ball (0.014 inches); (f) ball (0.021 inches); (g) outer race (0.007 inches); (h) outer race (0.014 inches); (i) outer race (0.021 inches).
Figure 7. The variational mode decomposition (VMD) decomposition results of signals with different faults: (a) inner race (0.007 inches); (b) inner race (0.014 inches); (c) inner race (0.021 inches); (d) ball (0.007 inches); (e) ball (0.014 inches); (f) ball (0.021 inches); (g) outer race (0.007 inches); (h) outer race (0.014 inches); (i) outer race (0.021 inches).
Applsci 09 01676 g007
Figure 8. 2D projection of the training samples (0.5 labeled).
Figure 8. 2D projection of the training samples (0.5 labeled).
Applsci 09 01676 g008
Figure 9. Convergence procedure of adaptive parameter optimization (0.5 labeled): (a) Distribution of convergence curves in 10 runs, (b) comparison of different optimization methods.
Figure 9. Convergence procedure of adaptive parameter optimization (0.5 labeled): (a) Distribution of convergence curves in 10 runs, (b) comparison of different optimization methods.
Applsci 09 01676 g009
Figure 10. Comparison of evaluation values for different methods.
Figure 10. Comparison of evaluation values for different methods.
Applsci 09 01676 g010
Figure 11. Boxplots of evaluation values for different methods, the x-axis tick labels correspond to: 1: semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector machine (SSFCM-ASCA-SVM) (0.3); 2: constrained-K-means, adaptive sine cosine algorithm and support vector data description with improved decision strategy (CK-means-ASCA-ID-SVDD) (0.3); 3: semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with relative distance decision strategy (SSFCM-ASCA-RD-SVDD) (0.3); 4: semi-supervised fuzzy C-means clustering, sine cosine algorithm and support vector data description with improved decision strategy (SSFCM-SCA-ID-SVDD) (0.3); 5: semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with improved decision strategy (SSFCM-ASCA-ID-SVDD) (0.3); 6 CK-means-ASCA-SVM (0.5); 7: SSFCM-ASCA-SVM (0.5); 8: fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with improved decision strategy (FCM-ASCA-ID-SVDD )(0.5); 9: CK-means-ASCA-ID-SVDD (0.5); 10: SSFCM-SCA-RD-SVDD (0.5); 11: SSFCM-ASCA-RD-SVDD (0.5); 12: SSFCM-SCA-ID-SVDD (0.5); 13: SSFCM-ASCA-ID-SVDD (0.5).
Figure 11. Boxplots of evaluation values for different methods, the x-axis tick labels correspond to: 1: semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector machine (SSFCM-ASCA-SVM) (0.3); 2: constrained-K-means, adaptive sine cosine algorithm and support vector data description with improved decision strategy (CK-means-ASCA-ID-SVDD) (0.3); 3: semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with relative distance decision strategy (SSFCM-ASCA-RD-SVDD) (0.3); 4: semi-supervised fuzzy C-means clustering, sine cosine algorithm and support vector data description with improved decision strategy (SSFCM-SCA-ID-SVDD) (0.3); 5: semi-supervised fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with improved decision strategy (SSFCM-ASCA-ID-SVDD) (0.3); 6 CK-means-ASCA-SVM (0.5); 7: SSFCM-ASCA-SVM (0.5); 8: fuzzy C-means clustering, adaptive sine cosine algorithm and support vector data description with improved decision strategy (FCM-ASCA-ID-SVDD )(0.5); 9: CK-means-ASCA-ID-SVDD (0.5); 10: SSFCM-SCA-RD-SVDD (0.5); 11: SSFCM-ASCA-RD-SVDD (0.5); 12: SSFCM-SCA-ID-SVDD (0.5); 13: SSFCM-ASCA-ID-SVDD (0.5).
Applsci 09 01676 g011
Table 1. List of the vibration samples used in the experiment.
Table 1. List of the vibration samples used in the experiment.
Fault LocationDiameter (inches)Fault LabelNumber of Samples
Inner race0.007L159
Inner race0.014L259
Inner race0.021L359
Ball0.007L459
Ball0.014L559
Ball0.021L659
Outer race0.007L759
Outer race0.014L859
Outer race0.021L959
Table 2. Center frequencies with different K value.
Table 2. Center frequencies with different K value.
Number of ModesCenter Frequencies (Hz)
22903.441166.73
33832.192928.051285.99
43179.092236.651107.20163.33
53945.873647.072922.651477.51675.67
63952.423672.883042.012756.441465.45668.06
74079.223835.393547.952921.581597.691263.47633.97
84088.983848.103573.753040.862762.201583.651255.25632.93
95175.264031.423821.293562.163040.242761.831583.051254.80633.18
Table 3. Fuzzy entropy values of different fault samples.
Table 3. Fuzzy entropy values of different fault samples.
Fault LabelSample NumberFuzzy Entropy for Different IMFs
IMF1IMF2IMF3IMF4
L111.79961.80521.17440.6738
21.76911.81041.16490.6638
31.82861.77671.16950.6738
L211.72421.35991.16950.5838
21.63531.23881.17760.5799
31.63771.37051.19810.6119
L312.03561.67031.83100.6125
21.91681.71861.82020.6199
31.92481.70811.83250.6094
L412.21602.02360.93340.3960
22.22822.07230.90720.3288
32.20252.01160.96500.5454
L511.71231.64511.21690.4758
21.99051.64871.19660.4806
31.91241.69211.19900.4936
L611.73382.09301.26180.6195
22.13331.98241.21640.6141
32.12291.99461.24350.6438
L711.28591.75361.58430.9124
21.23131.75081.61090.8821
31.33351.75561.60991.0993
L812.10042.07250.85020.4750
22.19272.16510.79060.4937
32.15652.12610.92070.5691
L910.74370.95960.92060.4920
20.69161.06091.25330.5486
30.92970.91850.90810.5329
Table 4. Value of parameter in the experiment.
Table 4. Value of parameter in the experiment.
ParameterDescriptionMethodValue
Kmode numberVMD4
mfractal dimensionFuzzyEn2
rpositive real numberFuzzyEn0.2
wweighting parameterSSFCM2
αbalance coefficientSSFCM1/0.3 and 1/0.5
Mnumber of individualsASCA30
Titeration timesASCA100
Cpenalty factorSVDD0.1177 and 0.1523
σkernel parameterSVDD26.7073 and 0.0011
knumber of nearest neighborsKNN3
Table 5. Fault diagnosis results with different methods.
Table 5. Fault diagnosis results with different methods.
Processing Methods (Labeled Ratio)Cσ (g)Result Evaluation
Normalized Mutual Information (NMI)Accuracy (ACC)
SSFCM-ASCA-SVM (0.3)39.85453.42620.8860, [−0.030, 0.042]0.9351, [−0.029, 0.024]
CK-means-ASCA-ID-SVDD (0.3)0.098372.28030.8534, [−0.029, 0.027]0.9035, [−0.020, 0.032]
SSFCM-ASCA-RD-SVDD (0.3)0.03231024.00000.8505, [−0.069,0.060]0.9053, [−0.063, 0.054]
SSFCM-SCA-ID-SVDD (0.3)0.956029.58150.8833, [−0.065, 0.033]0.9333, [−0.056, 0.026]
SSFCM-ASCA-ID-SVDD (0.3)0.117726.70730.8868, [−0.050, 0.055]0.9380, [−0.043, 0.033]
CK-means-ASCA-SVM (0.5)31.43511.25670.8838, [−0.035, 0.061]0.9333, [−0.056, 0.043]
SSFCM-ASCA-SVM (0.5)73.43610.28250.8919, [−0.044, 0.071]0.9386, [−0.044, 0.026]
FCM-ASCA-ID-SVDD (0.5)0.98921005.04370.6323, [−0.044, 0.071]0.8421, [−0.053, 0.047]
CK-means-ASCA-ID-SVDD (0.5)0.03030.00250.8877, [−0.041, 0.048]0.9386, [−0.044, 0.032]
SSFCM-SCA-RD-SVDD (0.5)0.51040.10060.8601, [−0.083, 0.069]0.9175, [−0.064, 0.053]
SSFCM-ASCA-RD-SVDD (0.5)0.245618.39670.8686, [−0.043, 0.070]0.9240, [−0.041, 0.047]
SSFCM-SCA-ID-SVDD (0.5)0.819551.03770.9102, [−0.020, 0.026]0.9550, [−0.013, 0.016]
SSFCM-ASCA-ID-SVDD (0.5)0.15230.00110.9254, [−0.036, 0.021]0.9649, [−0.029, 0.012]

Share and Cite

MDPI and ACS Style

Tan, J.; Fu, W.; Wang, K.; Xue, X.; Hu, W.; Shan, Y. Fault Diagnosis for Rolling Bearing Based on Semi-Supervised Clustering and Support Vector Data Description with Adaptive Parameter Optimization and Improved Decision Strategy. Appl. Sci. 2019, 9, 1676. https://doi.org/10.3390/app9081676

AMA Style

Tan J, Fu W, Wang K, Xue X, Hu W, Shan Y. Fault Diagnosis for Rolling Bearing Based on Semi-Supervised Clustering and Support Vector Data Description with Adaptive Parameter Optimization and Improved Decision Strategy. Applied Sciences. 2019; 9(8):1676. https://doi.org/10.3390/app9081676

Chicago/Turabian Style

Tan, Jiawen, Wenlong Fu, Kai Wang, Xiaoming Xue, Wenbing Hu, and Yahui Shan. 2019. "Fault Diagnosis for Rolling Bearing Based on Semi-Supervised Clustering and Support Vector Data Description with Adaptive Parameter Optimization and Improved Decision Strategy" Applied Sciences 9, no. 8: 1676. https://doi.org/10.3390/app9081676

APA Style

Tan, J., Fu, W., Wang, K., Xue, X., Hu, W., & Shan, Y. (2019). Fault Diagnosis for Rolling Bearing Based on Semi-Supervised Clustering and Support Vector Data Description with Adaptive Parameter Optimization and Improved Decision Strategy. Applied Sciences, 9(8), 1676. https://doi.org/10.3390/app9081676

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop