Next Article in Journal
Differences in Physiological Reactions Due to a Competitive Rehabilitation Game Modality
Next Article in Special Issue
Analysis of the Vibration Suppression of Double-Beam System via Nonlinear Switching Piezoelectric Network
Previous Article in Journal
Vertical Jumping for Legged Robot Based on Quadratic Programming
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Extraction Using Sparse Kernel Non-Negative Matrix Factorization for Rolling Element Bearing Diagnosis

1
School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
Key Laboratory of Education Ministry for Modern Design and Rotor-Bearing System, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3680; https://doi.org/10.3390/s21113680
Submission received: 19 April 2021 / Revised: 15 May 2021 / Accepted: 19 May 2021 / Published: 25 May 2021

Abstract

:
For early fault detection of a bearing, the localized defect generally brings a complex vibration signal, so it is difficult to detect the periodic transient characteristics from the signal spectrum using conventional bearing fault diagnosis methods. Therefore, many matrix analysis technologies, such as singular value decomposition (SVD) and reweighted SVD (RSVD), were proposed recently to solve this problem. However, such technologies also face failure in bearing fault detection due to the poor interpretability of the obtained eigenvector. Non-negative Matrix Factorization (NMF), as a part-based representation algorithm, can extract low-rank basis spaces with natural sparsity from the time–frequency representation. It performs excellent interpretability of the factor matrices due to its non-negative constraints. By this virtue, NMF can extract the fault feature by separating the frequency bands of resonance regions from the amplitude spectrogram automatically. In this paper, a new feature extraction method based on sparse kernel NMF (KNMF) was proposed to extract the fault features from the amplitude spectrogram in greater depth. By decomposing the amplitude spectrogram using the kernel-based NMF model with L1 regularization, sparser spectral bases can be obtained. Using KNMF with the linear kernel function, the time–frequency distribution of the vibration signal can be decomposed into a subspace with different frequency bands. Thus, we can extract the fault features, a series of periodic impulses, from the decomposed subspace according to the sparse frequency bands in the spectral bases. As a result, the proposed method shows a very high performance in extracting fault features, which is verified by experimental investigations and benchmarked by the Fast Kurtogram, SVD and NMF-based methods.

1. Introduction

As the rotary support component of most machinery, the fault detection and diagnosis of the rolling element bearing is crucial to prevent machinery breakdowns [1]. A variety of bearing fault detection techniques such as acoustic emission, electrostatic and vibration are used meticulously by industrial enterprises [2,3,4,5]. Among them, vibration monitoring is the most established diagnostic technique for rolling element bearing.
While a localized defect occurs on the surface of a relatively moving part, a series of impulses will be excited periodically or quasi-periodically. Due to different natural frequencies for rotating structure parts, multiple resonance regions reflecting impulse information will exist in the vibration signal of a mechanical system [6]. Inevitably, in the early stage of bearing failure, the noise energy usually contaminates the fault information of resonance regions.
Generally, an envelope-based analysis is one of the most popular technologies for bearing fault detection, because the envelope demodulation of a vibration signal can provide more valuable information [7,8,9]. Frequency band selection is substantial to get a successful envelope demodulation, because fault features cannot be extracted if the frequency band is not selected appropriately. Therefore, two algorithms were proposed, based on STFT and wavelet packet transform, respectively, to generate the Kurtogram and obtain the optimal frequency band with maximum kurtosis [10,11]. Apart from Kurtogram, another representative method is the Protugram [12], where the maximum kurtosis is extracted from the envelope signal instead of the raw signal. The Autogram [13] selects the optimal band with the highest impulsiveness also based on the maximum kurtosis, but unlike the Kurtogram, it is calculated based on the unbiased autocorrelation of the squared envelope of the demodulated signals. The Sparsogram [14] is based on the sparsity level on different bands based on the wavelet packet, and the Infogram utilizes the negentropy as a feature to detect the impulsive bands of the signal for demodulation [15]. SVD, which can adaptively extract representative features from the time–frequency distribution or Hankel matrix for the bearing fault diagnosis [16,17], has been proposed as a supplement to the envelope demodulation method in recent years. Jiang et al. utilized the ratio of singular values as an evaluation for the fault feature and further introduced the difference spectrum to select the informative singular vector due to different singular values for the signal and noise [18]. Xu et al. chose the anti-averaging method for sub-signal reconstruction and combined the SVD and squared envelope spectrum to identify the fault type. By the way, the wavelet transform matrix [19] can also be used for the input matrix of the SVD. It is worth noting that the SVD-based method is essentially the maximum variance projection, which means that the components with higher energy will be decomposed first, while the fault information is always neglected. To unravel this problem, an algorithm called RSVD was proposed based on the periodicity of the defect characteristics [20]. Mathematically, SVD-based methods only impose orthogonal constraints, so the value of the eigenvector atoms can be positive or negative. Such eigenvectors are usually associated with multiple frequency components, which will make the decomposition hard to understand and induce noise in the final results.
NMF has been widely used in audio source separation because of its ability to automatically separate components with different frequency and time information in the spectrogram [21,22]. By this virtue, researchers introduced NMF to the field of bearing fault detection to find the part-based representations of the bearing fault signal. Different from SVD, NMF imposes a non-negativity constraint on low-rank subspaces, so it has a better interpretability and can additively separate the components in frequency bands. For the amplitude spectrogram of the bearing fault signal, NMF can group the spectral components and sparsely present them in the factor matrices. Liang et al. [23] employed a deconvolution method for extracting impulse excitation using Convolution NMF and characterized the localized features of the impulse response, including the resonance frequency band and the attenuation response, effectively. Liu et al. [24] performed the Semi-NMF for the time–frequency matrix reconstruction, based on the vibration signal with a rolling element bearing defect, to extract the impulse response. Based on the amplitude spectrogram decomposition, Wodecki et al. [25] employed NMF to cluster the spectral components into the base matrix so that the optimal frequency band can be adaptively obtained. In essence, the above NMF-based developments were used to seek low-rank representations of high-dimensional time–frequency matrices for the bearing fault diagnosis based on a linear mixing model.
Considering that the kernel methods can offer a nonlinear mapping of the input space, some kernel-based NMF approaches [26,27,28] have been proposed to obtain a more trustworthy solution. Generally, as one of the outstanding properties, KNMF can provide a sparser representation of the data than NMF. Such properties can restrain the noise and help us select the frequency band more efficiently while we perform a fault diagnosis based on the amplitude spectrogram. Unfortunately, although the factor matrices obtained by KNMF are naturally sparse, it is difficult for users to control the sparse degree of representation. Hence, sparsity constraints can be imposed either with projections or L1 regularization on the KNMF model to induce a sparser solution.
We propose a novel feature extraction method in this paper, based on linear kernel NMF with a sparsity constraint, to detect fault characteristics from the amplitude spectrogram of bearing fault signal. To our knowledge, it is the first time a L1 regularization has been associated with a linear kernel NMF model, and it is the first time such models are being used in the context of a bearing fault diagnosis. With the linear kernel function, the time–frequency distribution is naturally decomposed into the superposition of subspaces by NMF. With the L1 regularization, a sparse KNMF model was established to obtain sparser factor matrices, which have energy only at the feature frequency bands presented in the input amplitude spectrogram. Thus, the optimal subspace corresponding to periodic impulses can be selected effectively by the sparse frequency band distribution of the sparse factor matrix. The proposed method outperforms the SVD method for its good interpretability and sparse decomposition results. Different from the envelope-based methods, the proposed method takes the time–frequency matrix of the short-time Fourier transform (STFT) as the input and can limit the in-band noise that perplexes the envelope-based methods by sparse representation. Several engineering applications show that the proposed method performs better than the SVD and envelope-based methods, especially in early fault detection cases.
This paper is organized as follows: Section 2 provides the fundamental knowledge about KNMF. In Section 3, the principles of the multiplicative update rule of kernel NMF with sparsity constraint are shown in detail. Section 4 presents a feature extraction strategy based on sparse KNMF; the vibration signals of the rolling bearing faults are presented to evaluate the proposed method in Section 5. At last, the conclusions are drawn in Section 6.

2. Kernel Non-Negative Matrix Factorization

2.1. NMF

As one of the low-rank decomposition tools, the goal of NMF is to factor a non-negative matrix into the product of two non-negative matrices called the basis matrix and the coefficient matrix:
X m × n U m × r V r × n
Different from SVD, the non-negativity constraints are imposed on two factor matrices, which can effectively show the concept of part-based representation. To solve matrices U and V in Equation (1), the usual Frobenius norm-based cost function is as follows:
min U , V 1 2 | | X - U V | | F 2 , s . t . U , V 0
where U(m × r) is defined as the basis matrix, V(r × n) is a coefficient matrix, | | | | F 2 is the Frobenius norm and r << min ( m , n ) . Considering Equation (2) is not convex to U and V simultaneously, besides the gradient algorithm, multiplicative and alternating least squares algorithms were also proposed.

2.2. Kernel NMF

With nonlinear mapping: X φ ( X ) . For the decomposition of X = U V , the factor matrix U can be defined by
U = X W
where W is the transformation matrix, and each column of W satisfies the constraint that the sum is one. Thus, the cost function (2) can be rewritten as follows:
min D F ( φ ( X U φ V ) ) = 1 2 φ ( X ) φ ( X ) W V F 2 , s . t . W 0 , V 0
In theory, the kernel function can improve the sparsity of the factor matrices, but the sparsity is still influenced by the initialization, kernel function type and parameter value. We will show that, in these cases [29], the sparse constraints can effectively match the features of the data. Additionally, for the case of the spectrogram of the bearing vibration signals, it was hoped that the sparsity of the factor matrices would make the impact energy more concentrated, and it is easy to select the feature frequency band.

3. Sparse KNMF and Update Rule

Although several approaches have been proposed to effectively control the sparseness of U and V in references [30,31], the sparsity constraints may result in the loss of useful information [32]. To get better sparseness of the frequency information from the time–frequency distribution, the constraints about the L1 norm of the coefficient matrix V is utilized in KNMF, and the sparse kernel non-negative matrix factorization (SKNMF) is also proposed. Taking the sparseness constraints of V into account, the Lagrangian is given by
min L F ( W , V ) = 1 2 φ ( X ) φ ( X ) W V F 2 + α V 1
where α is the constraint term parameter, and λ and μ are the Lagrange multipliers. Under the conditions of Karush-Kuhn-Tucker (KKT) optimality, (W, V) is a stationary point of Equation (5) if
W 0 V 0 , ( L F ( W , V ) / W ) i j 0 ( L F ( W , V ) / V ) i j 0 W i j ( L F ( W , V ) / W ) i j = 0 V i j ( L F ( W , V ) / V ) i j = 0
when W i j > 0 and V i j > 0 , the two partial derivatives of the objective function, from Equation (6), are denoted as follows:
( L ( W , V ) ) / W = 0 ( L ( W , V ) ) / V = 0
Further, the optimal solution of Equation (7) can be defined following function:
φ ( X ) T φ ( X ) W V V T φ ( X ) T φ ( X ) V T λ = 0 W T φ ( X ) T φ ( X ) W V W T φ ( X ) T φ ( X ) + α V V 1 μ = 0
where V V 1 is the gradient vector and means of the partial derivative with respect to V 1 . Obviously, because the L1 norm cannot be differentiated directly, the subdifferential method of the real value variable [33] is used for this problem.
Set V V = [ v 1 v 1 1 , v 2 v 2 2 , , v n v n n ] T , where the ith element is defined as
( v n ) i v n 1 = v n 1 ( v n ) i = { { + 1 } , ( v n ) i > 0 [ 1 , + 1 ] , ( v n ) i = 0 { 1 } , ( v n ) i < 0
Due to the non-negative of the elements of matrix V, the set of [+1, −1] is set to 0 to simplify the calculations, and the above equation can be rewritten as
( v n ) i v n 1 = v n 1 ( v n ) i = { { + 1 } , ( v n ) i > 0 { 0 } , ( v n ) i = 0
Let us define K = φ ( X ) T φ ( X ) and use the KKT optimality conditions. We can obtain the following update rules when λ and μ are set to 0:
W i j = W i j ( K V T ) i j ( K W V V T ) i j V i j = V i j ( W T K ) i j ( W T K W V + α V V 1 ) i j
where the kernel matrix K can be obtained with kernel functions, such as the polynomial kernel functions, Gaussian kernel functions and so on.

4. Feature Extraction Strategy Based on SKNMF

For the bearing fault detection, the periodic impulses caused by a defect are localized at the resonance frequency band and corresponding time point in the time–frequency distribution. The fault impulse feature can be extracted by the advantages of SKNMF in the parts-based representation. The scheme of the whole strategy is displayed in Figure 1, and the implementation procedure is detailed as follows:

4.1. Time–frequency Distribution Construction

We denote by TFt,f the corresponding time–frequency matrix of the measured signal, and with STFT, TFt,f is described as below:
T F t , f = | x ( t + τ ) w ( τ ) exp ( 2 j π f τ ) d τ |
where x(t) is the measured signal, w(t) is the window function (i.e., Hanning window) and t and f are the time and frequency, respectively.

4.2. Subspace Extraction with SKNMF

For the time–frequency space TFt,f, where its rows and columns represent the frequency and time location, respectively, the factor matrix V and transform matrix W are learned by the multiplicative updates of Equation (11). Then, in the feature space, according to Equations (1) and (3), we can obtain following expression:
φ ( T F t , f ) = U φ V
U φ = φ ( T F t , f ) W
where φ is mapping of the input space. U φ is the base matrix, where each column of U φ is denote by u φ k . Considering the generalization ability of the global kernel, the linear kernel function is selected and the kernel matrix K = X T X . For TFt,f, we can obtain φ ( T F t , f ) = T F t , f . Thus, according to Equation (13), the time–frequency space TFt,f is given by
T F t , f = k = 1 j u φ k v k
Naturally, Equation (15) shows that the time–frequency matrix TFt,f consists of a series of time–frequency subspace T F t , f k = u φ k v k , and it can be rewritten by
T F t , f = k = 1 j T F t , f ( k )

4.3. Subspace Selection and Waveform Reconstruction

It is clear that each column of V has energy only at the main frequencies that are present in the subspace T F t , f ( k ) . We can interpret these columns as the frequency energy contained in the spectrogram. Due to the sparsity of V, the corresponding frequencies in the subspace are more concentrated—namely, the noise interference components are less. Therefore, combining the resonance regions of the spectrum with the larger weight of V, subspaces T F t , f ( i ) ( i ( 1 , 2 , r ) ) that represent the impulse features effectively can be determined. Then, using the inverse STFT to the estimated impulse components, the extracted signal x s can be obtained.
x ( t ) = 1 2 π T F t , f S w ( t τ ) e j f t d f d τ
where T F t , f S = T F t , f ( i ) ( i ( 1 , 2 , r ) ) T is the selected time–frequency subspace.

4.4. Envelope Demodulation

Since the extracted signal x s contains the periodic impulse components, the envelope signal is obtained by Z ( t ) = | x s ( t ) | 2 + { H [ x s ( t ) ] } 2 , where H is the Hilbert transform, and the frequency information of the periodic impulses can also be derived from the envelope spectrum of Z(t).

5. Experimental Results

Two experiments were investigated to validate the proposed SKNMF, and the results are compared with not only the Kurtogram [10] and RSVD [20] but, also, NMF [34], sparse NMF (SNMF) [32] and kernel NMF (KNMF). Although a numerical experiment can be easily performed, it is difficult to accurately reflect the real working environment and complex frequency components of the vibration signal. Therefore, we used the real-world vibration data for comparison and analysis, while a measured data in practice was adopted for experiment verification.

5.1. Comparison Analysis Experiment

The first experiment was designed to illustrate the efficiency of SKNMF for detecting an incipient bearing defect. The measured data came from the specially designed bearing test rig described in Figure 2. Four bearings were installed on a rotor shaft driven by an AC motor. The rotation speed was 2000 r/min, and acceleration sensors were mounted in the vertical and horizontal directions of each bearing seat. With a normal load and 20-kHz sampling frequency, the vibration data of the bearings was collected every 10 minutes. Finally, the outer race fault occurred in bearing 1 at the end of the experiment. For detailed information about the experiment, please refer to the literature [35].
According to the RMS value of the measured signals, we selected the vibration data collected during the 88.5 h for analysis, and the waveform and its spectrum are shown in Figure 3. Due to the early bearing defect, there are no periodic impulses shown in Figure 3a. For the spectrum in Figure 3b, not only the low frequency component is complicated, but also, different resonance regions can be observed.
Figure 4 presents the time–frequency distribution obtained with the Hanning window of 20 samples. Obviously, the time–frequency space of the early fault signal presented a combination of time information and frequency energy. It was found that the frequency bands were focused at about 980 Hz, 1800 Hz, 3200 Hz and 4300 Hz. respectively, while the energy was lower for the frequency bands at about 1800 Hz and 3200 Hz.
The SKNMF method was adopted to decompose the time–frequency space, where the elements in the initial matrices W and V were randomly chosen, and the factorization rank r and sparse coefficient α were set to 7 and 0.55, respectively. The coefficient matrix V is shown in Figure 5, where the vertical axis represents the frequency, and the horizontal axis is associated with a serial number of the subspace. It is shown that SKNMF with sparsity constraint tends to provide a sparse weight distribution in Figure 5, where the corresponding subspace has a more concentrate frequency band. From the frequency distribution of Figure 5, the corresponding weights of the fifth and sixth subspaces in the main frequencies were concentrated and larger. Therefore, the union of the two subspaces was selected, and the corresponding time–frequency subspace shown in Figure 6 was calculated by Equation (16). The extracted signal was obtained by inverse STFT, as shown in Figure 7a, and it revealed impulses compared with the original signal demonstrated in Figure 3. The spectrum of Figure 7b shows that the frequency band between 3100 Hz and 4400 Hz can be extracted adaptively, while the noise and other component interference can be eliminated effectively. The envelope analysis was performed on the extracted signal, and the corresponding envelope spectrum is shown in Figure 8. It is clear that the outer race fault characteristic fBPFO (236.4 Hz) denoted by a red arrow is rather evident. Besides, the second order harmonic of fBPFO denoted by another red arrow was also distinctive.
For comparison, the Kurtogram [10] of the same measured signal is present in Figure 9, where it is found that the optimal demodulation band is from 6250 Hz to 7500 Hz. According to the obtained optimal demodulation band, a bandpass filter with a center frequency of 6875 Hz and a bandwidth of 1250 Hz is constructed and used to filter the raw vibration signal. The waveform of the filtered signal is presented in Figure 10. However, the results indicate that the filtered signal with maximum kurtosis still has noise interference, and the transient feature in the waveform is not very clear. When the envelope analysis is applied to this signal, little diagnostic information could be obtained, as shown in Figure 11. In fact, we know that kurtosis is a measure of the peakedness, so it easily tends to highlight the outliers of the signal caused by the noise.
The RSVD method [20] was then introduced to detect the impulse components. The waveform and spectrum of the reconstructed signal is shown in Figure 12, from which the fourth subspace was selected for construction according to the first 30 PMIs of the decomposed singular components. It is obvious that SVD can be decomposed by frequency information adaptively. However, compared with the above methods, except for the main frequency band, there are still other frequency components in Figure 12b. Therefore, from the envelope spectrum given in Figure 13, one can also find that the repetition frequency of the impulse is severely disturbed by the other interference components. In essence, only under orthogonality constraint, the decomposed eigenvector can be a positive or negative value. It means that the basis vector is related to many frequencies. Therefore, as the superposition of the subspace, the subspace usually includes not only multiple resonance regions but, also, noise.
To explain this drawback of the SVD method, the time–frequency matrix shown in Figure 4 is decomposed by SVD, where the decomposition equation is T F t , f = U S V , and the column vectors of U and V are the orthonormal eigenvectors. Figure 14 illustrates the first six columns of V, and it is well-known that the eigenvector of V reflects the frequency distribution of the raw signal. Since the SVD methods allow the atoms of the eigenvector as the negative value, as a result, each eigenvector mostly has positive and negative atoms in the main frequencies (i.e., 980 Hz, 1800 Hz, 3200 Hz, 4300 Hz and so on). Therefore, the subspace constructed by the eigenvectors of U and V also includes multiple frequency bands, which means that the noise cannot be limited effectively.
To further test the effectiveness of the proposed method, Figure 15 illustrates the coefficient matrix V computed by NMF [34]. Certainly, one can easily see that the weights in KNMF are sparser than that of NMF. The poor sparsity makes each subspace factorized by NMF contain a lot of frequency components. Considering that the energy distribution of the seventh subspace is larger between 2400 Hz and 4700 Hz, the seventh subspace is selected as the extracted time–frequency subspace, as given in Figure 16. Using inverse STFT, the extracted waveform and spectrum are shown in Figure 17. Known from Figure 17b, except for the 986-Hz and 2400 Hz–4700-Hz frequency bands, the noise and other components are also observed. Thus, the periodic transient components were almost swallowed by noise in the waveform of Figure 17a. As a result, the envelope spectrum in Figure 18 shows that the fault characteristic fBPFO is submerged by many high-amplitude frequency components.
SKNMF performs better than NMF, because the sparseness of the coefficient matrix is useful for the frequency information determination. It can be observed from Figure 5 that the sparsity of the decomposition result is strengthened, and SKNMF is sensitive to the different frequency band and can separate them effectively.
Figure 19 shows the coefficient matrix V of the sparse NMF method [32], where the rank r of the factorization is still set to 7, and parameters η and β are set to 0 and 0.55, respectively. Similar to the NMF, according to the spectrum and frequency distributions of the coefficient matrix of Figure 19, we chose the sixth subspace as the extracted time–frequency distribution. Thus, from the selected time–frequency subspace shown in Figure 20, one can obtain the extracted signal and spectrum. Compared with the original spectrum, the corresponding spectrum shown in Figure 21b includes not only 3600 Hz–5200-Hz frequency bands but, also, noise components. The transient signals in Figure 21a indicate that the extracted signal still has noise, and the transient feature is not very clear. Meanwhile, the fault frequency is also not easily distinguishable in the envelope spectrum, because only the first-order harmonic of fBPFO is shown in Figure 22, where the fault character interfered with unknown high-amplitude components.
As the SNMF method was improved in the sparseness of the coefficient matrix, the extracted feature frequency band of the time–frequency distribution can be more concentrated. However, compared with SKNMF, except for the main frequency band, there were still other components that resulted in the feature frequency submerged in the envelope spectrum. Based on the kernel mapping and L1 norm-oriented sparseness, the SKNMF-based method can effectively eliminate the noise in the extracted signal. Therefore, the repetition frequency of the pulse characteristic can be highlighted.
Besides the above liner NMF-based models, the KNMF method with a multiplicative update strategy was employed to analyze the same signal. Figure 23 presents the heat map of coefficient matrix V. Compared with the results of NMF, the decomposition results of KNMF had a certain sparsity, which was similar to the results of SNMF.
However, the sparseness of matrix V derived by KNMF was not as good as that of SKNMF, shown in Figure 5. Since the weight distribution of the subspace was not concentrated, a seventh subspace with the larger weight value was selected, and the extracted subspace is shown in Figure 24. Figure 25a,b shows the spectrum and the envelope spectrum of the extracted signal. Compared with Figure 7b, the extracted feature frequency bandwidth of KNMF was wider than that of SKNMF, which led to the defect feature KNMF extracted not being as clear as what SKNMF extracted.

5.2. Experimental Verification

5.2.1. Test Rig

A vibration signal, including multiple resonance regions, was also used to verify the SKNMF method in extracting the periodic impulses. The vibration data were obtained from the machinery fault simulator shown in Figure 26a, which was constructed with a motor, couplings, rotors, belt tensioning and gearbox. The rotating shaft was driven by an AC motor, and a fault simulation was conducted by replacing the fault motors. The motor bearing was SKF 6203. A constant load was applied by means of the magnetic brake. A zoomed version of the AC motor mounted with an acceleration sensor is shown in Figure 26b. The technical details of the acceleration sensor are listed in Table 1.
In the experiment, the AC motor with built-in inner race faulted bearings was installed under 1 pound of torque, and the PCB 352C68 ICP acceleration sensor was mounted on the bearing set of the motor. The rotary speed fr controlled by a speed controller was 25 Hz, and the sample frequency was 10,240 Hz. According to the structural parameters of the bearing, the frequency of the inner race fault fBPFI was 4.9fr Hz.
The waveforms and spectrum of the vibration signal of the defect bearing was plotted in Figure 27. Obviously, we can see that the many frequency bands were obviously visible. Obviously, the time–frequency distribution shown in Figure 28 presents the resonance region center focused at about 500 Hz, 2050 Hz and 3150 Hz and other high-frequency components.

5.2.2. Results and Analysis

The proposed SKNMF-based method was adopted to decompose the time–frequency space, in which the rank r of the factorization was set to 11 because of more frequency bands. The extracted coefficient matrix is shown in Figure 29. From the weight distribution of Figure 29, the sparse weight of each rank was concentrated in different frequency bands, and the corresponding weight values in the subspaces of the first, second and tenth were larger. Therefore, the union of these subspaces was selected as the extracted signal given in Figure 30, and Figure 31 shows the spectrum and envelope spectrum of the extracted signal.
In Figure 31a, the feature frequency band between 3470 Hz and 4440 Hz can be extracted adaptively. The corresponding feature frequency fBPFI that indicates the inner race defect can be observed clearly in Figure 31b, where the first three harmonics of fBPFI are highlighted by three red arrows. It can be concluded from the results that the extracted signal with SKNMF can be used as an effective extraction means for the impulse feature caused by the localized defect.
Then, the Kurtogram was adapted to analyze the same measured signal in Figure 27. According to the maximum value principle shown in Figure 32, the envelope spectrum of the signal filtered by the optimal frequency band between 3760 Hz and 3920 Hz is demonstrated in Figure 33, where it is found that the repetition frequency of the impulses is not prominent.
There is no denying that the Kurtogram-based method has a certain ability to extract the feature frequency of the impulses. However, in Figure 33, the transient feature of the extracted signal was still disturbed by the noise and other frequencies, and the feature frequency in the envelope spectrum was less clear when compared with the envelope spectrum obtained by the proposed method, as shown in Figure 31b.
The RSVD method was also adapted for comparison. By integrating 30 singular components, the spectrum and envelope spectrum of the reconstructed signal were obtained by the third subspace and illustrated in Figure 34, from which multiple frequency bands can be selected. The corresponding feature frequency of the inner race fault was not evidently displayed. Clearly, due to more frequency bands and noise contained in the vibration signal, for the adaptive decomposition of SVD, the feature signal and noise components were embedded in the same eigenvector subspace, which could affect the effect of the fault frequency extraction.
In summary, SKNMF outperforms two excellent techniques. To clarify the essence of SKNMF, the coefficient matrices obtained by NMF, SNMF and KNMF are shown in Figure 35. It is obvious that the NMF-based method has the worst sparsity, while the factor matrix of KNMF is similar to SNMF because of its natural sparsity.
Considering the distribution of subspace frequency band extracted by SKNMF, a similar frequency band was also selected in three NMF-based methods. The time–frequency subspace of NMF was selected with a union of the third, sixth and ninth subspaces. The envelope spectrum of the corresponding signal showed the fault-related signatures in Figure 36a. For the SNMF method, the union of the fifth and tenth subspaces was selected for the extracted result, and the envelope spectrum of the signal provided fault information in Figure 36b. Finally, KNMF was also adopted to process the weak signal and select the union of the first, third and tenth subspaces for analysis. The envelope spectrum of three extracted signals is plotted in Figure 36c. Although the envelope spectrum reflected the fault frequency, the signal-to-noise ratio was much lower than that of Figure 31b.

6. Conclusions

The vibration signal of a rolling bearing with a localized defect is generally a mixture of multiple components, in which the characteristic component, revealing the latent impulses, is often weak and difficult to detect. In this paper, a sparse kernel NMF model and corresponding feature extraction method were proposed to address this concern.
Two cases, one with an outer race fault and another with an inner race fault, were provided to illustrate that the proposed method can achieve the extraction of local weak signatures under the strong interference of a background noise. The proposed method was benchmarked by Fast Kurtogram, the SVD-based method, the NMF-based method, the SNMF-based method and the KNMF-based method. When the signal of interest was submerged by a strong background noise, the Fast Kurtogram showed how to keep the peak frequencies of interest below the noise level. The SVD-based method was confused by its orthogonal constraints, which often induce the eigenvectors associated with multiple frequency components, making the decomposition hard to understand and inducing noise in the final results. As for the NMF-based method, although it can extract the frequencies of interest, the final diagnosis result was not as clear as the proposed method. The comparison analysis experiment confirmed that SKNMF was superior to the other matrix decomposition methods, because it was sensitive to the different frequency bands and could separate them adaptively. As the result, the experimental verification allowed us to conclude that SKNMF is reliable enough to extract fault features and, hence, achieve early fault detection. It should be noted that the proposed method used an alternate iteration strategy, which made the algorithms not fast enough to be promoted for online monitoring.

Author Contributions

Conceptualization, L.L. and F.L.; Methodology, X.D. and Y.C.; Software, Y.C.; writing—original draft preparation, X.D. and Y.C.; Validation, F.L., H.W. and X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (52075431) and the Natural Science Foundation of Shaanxi Province China (2020JM-050).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liang, L.; Liu, F.; Li, M.; He, K.; Xu, G. Feature selection for machine fault diagnosis using clustering of non-negation matrix factorization. Meas. J. Int. Meas. Confed. 2016, 94, 295–305. [Google Scholar] [CrossRef]
  2. Immovilli, F.; Bellini, A.; Rubini, R.; Tassoni, C. Diagnosis of bearing faults in induction machines by vibration or current signals: A critical comparison. IEEE Trans. Ind. Appl. 2010, 46, 1350–1359. [Google Scholar] [CrossRef]
  3. Rzeszucinski, P.; Orman, M.; Pinto, C.T.; Tkaczyk, A.; Sulowicz, M. Bearing Health Diagnosed with a Mobile Phone: Acoustic Signal Measurements Can be Used to Test for Structural Faults in Motors. IEEE Ind. Appl. Mag. 2018, 24, 17–23. [Google Scholar] [CrossRef]
  4. Orman, M.; Rzeszucinski, P.; Tkaczyk, A.; Krishnamoorthi, K.; Pinto, C.T.; Sulowicz, M. Bearing fault detection with the use of acoustic signals recorded by a hand-held mobile phone. In Proceedings of the 2015 International Conference on Condition Assessment Techniques Electrical Systems (CATCON), Bangalore, India, 10–12 December 2015; pp. 252–256. [Google Scholar] [CrossRef]
  5. Frosini, L.; Harlişca, C.; Szabó, L. Induction machine bearing fault detection by means of statistical processing of the stray flux measurement. IEEE Trans. Ind. Electron. 2014, 62, 1846–1854. [Google Scholar] [CrossRef]
  6. Zhang, H.; Chen, X.; Du, Z.; Yan, R. Kurtosis based weighted sparse model with convex optimization technique for bearing fault diagnosis. Mech. Syst. Signal Process. 2016, 80, 349–376. [Google Scholar] [CrossRef]
  7. Sheen, Y.T. An envelope analysis based on the resonance modes of the mechanical system for the bearing defect diagnosis. Meas. J. Int. Meas. Confed. 2010, 43, 912–934. [Google Scholar] [CrossRef]
  8. Zimroz, R.; Bartelmus, W. Application of adaptive filtering for weak impulsive signal recovery for bearings local damage detection in complex mining mechanical systems working under condition of varying load. Trans Tech Publ. 2012, 180, 250–257. [Google Scholar] [CrossRef]
  9. Barszcz, T.; Randall, R.B. Application of spectral kurtosis for detection of a tooth crack in the planetary gear of a wind turbine. Mech. Syst. Signal Process. 2009, 23, 1352–1365. [Google Scholar] [CrossRef]
  10. Antoni, J. Fast computation of the kurtogram for the detection of transient faults. Mech. Syst. Signal Process. 2007, 21, 108–124. [Google Scholar] [CrossRef]
  11. Antoni, J. The spectral kurtosis: A useful tool for characterising non-stationary signals. Mech. Syst. Signal Process. 2006, 20, 282–307. [Google Scholar] [CrossRef]
  12. Barszcz, T.; Jabłoński, A. A novel method for the optimal band selection for vibration signal demodulation and comparison with the Kurtogram. Mech. Syst. Signal Process. 2011, 25, 431–451. [Google Scholar] [CrossRef]
  13. Moshrefzadeh, A.; Fasana, A. The Autogram: An effective approach for selecting the optimal demodulation band in rolling element bearings diagnosis. Mech. Syst. Signal Process. 2018, 105, 294–318. [Google Scholar] [CrossRef]
  14. Tse, P.W.; Wang, D. The sparsogram: A new and effective method for extracting bearing fault features. In Proceedings of the 2011 Prognostics and System Health Managment Conference, Shenzhen, China, 24–25 May 2011. [Google Scholar] [CrossRef]
  15. Antoni, J. The infogram: Entropic evidence of the signature of repetitive transients. Mech. Syst. Signal Process. 2016, 74, 73–94. [Google Scholar] [CrossRef]
  16. Cong, F.; Chen, J.; Dong, G.; Zhao, F. Short-time matrix series based singular value decomposition for rolling bearing fault diagnosis. Mech. Syst. Signal Process. 2013, 34, 218–230. [Google Scholar] [CrossRef]
  17. Li, H.; Liu, T.; Wu, X.; Chen, Q. Research on bearing fault feature extraction based on singular value decomposition and optimized frequency band entropy. Mech. Syst. Signal Process. 2019, 118, 477–502. [Google Scholar] [CrossRef]
  18. Jiang, H.; Chen, J.; Dong, G.; Liu, T.; Chen, G. Study on Hankel matrix-based SVD and its application in rolling element bearing fault diagnosis. Mech. Syst. Signal Process. 2015, 52–53, 338–359. [Google Scholar] [CrossRef]
  19. Jiang, Y.; Kan, J.; Zhang, Z.; Cheng, G.; Wen, J. Machine fault diagnosis based on reassigned wavelet scalogram and SVD. Int. J. Adv. Comput. Technol. 2012, 4, 10–21. [Google Scholar] [CrossRef]
  20. Zhao, M.; Jia, X. A novel strategy for signal denoising using reweighted SVD and its applications to weak fault feature enhancement of rotating machinery. Mech. Syst. Signal Process. 2017, 94, 129–147. [Google Scholar] [CrossRef]
  21. Wen, H.; Liang, L.; Liu, F. Speech enhancement using nonnegative matrix factorization with overlapping group sparsity constraint. ACM Int. Conf. Proceeding Ser. 2020, 87–91. [Google Scholar] [CrossRef]
  22. Zhou, Q.; Feng, Z.; Benetos, E. Adaptive noise reduction for sound event detection using subband-weighted NMF. Sensors 2019, 19, 3206. [Google Scholar] [CrossRef] [Green Version]
  23. Liang, L.; Shan, L.; Liu, F.; Li, M.; Niu, B.; Xu, G. Impulse Feature Extraction of Bearing Faults Based on Convolutive Nonnegative Matrix Factorization. IEEE Access 2020, 8, 88617–88632. [Google Scholar] [CrossRef]
  24. Liu, H.; Wen, H.; Liang, L.; Wang, J.; Fan, Z.; Hu, J. Impulse Response Extraction for Rolling Bearing Faults Based on Semi-NMF. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 December 2019; pp. 884–887. [Google Scholar] [CrossRef]
  25. Wodecki, J.; Kruczek, P.; Bartkowiak, A.; Zimroz, R.; Wyłomańska, A. Novel method of informative frequency band selection for vibration signal using Nonnegative Matrix Factorization of spectrogram matrix. Mech. Syst. Signal Process. 2019, 130, 585–596. [Google Scholar] [CrossRef]
  26. Pan, B.; Lai, J.; Chen, W.S. Nonlinear nonnegative matrix factorization based on Mercer kernel construction. Pattern Recognit. 2011, 44, 2800–2810. [Google Scholar] [CrossRef]
  27. Kallas, M.; Honeine, P.; Richard, C.; Francis, C.; Amoud, H. Non-negativity constraints on the pre-image for pattern recognition with kernel machines. Pattern Recognit. 2013, 46, 3066–3080. [Google Scholar] [CrossRef] [Green Version]
  28. Li, Y.; Ngom, A. A new Kernel non-negative matrix factorization and its application in microarray data analysis. In Proceedings of the 2012 IEEE Symposium on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), San Diego, CA, USA, 9–12 May 2012; pp. 371–378. [Google Scholar] [CrossRef]
  29. Hoyer, P.O. Non-negative matrix factorization with sparseness constraints. J. Mach. Learn. Res. 2004, 5, 1457–1469. [Google Scholar]
  30. Pascual-Montano, A.; Carazo, J.M.; Kochi, K.; Lehmann, D.; Pascual-Marqui, R.D. Nonsmooth nonnegative matrix factorization (nsNMF). IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 403–415. [Google Scholar] [CrossRef]
  31. Pauca, V.P.; Piper, J.; Plemmons, R.J. Nonnegative matrix factorization for spectral data analysis. Linear Algebra Its Appl. 2006, 416, 29–47. [Google Scholar] [CrossRef] [Green Version]
  32. Kim, H.; Park, H. Sparse non-negative matrix factorizations via alternating non-negativity-constrained least squares for microarray data analysis. Bioinformatics 2007, 23, 1495–1502. [Google Scholar] [CrossRef] [Green Version]
  33. Hale, E.T.; Yin, W.; Zhang, Y. Fixed-point continuation for 11-minimization: Methodology and convergence. SIAM J. Optim. 2008, 19, 1107–1130. [Google Scholar] [CrossRef]
  34. Berry, M.W.; Browne, M.; Langville, A.N.; Pauca, V.P.; Plemmons, R.J. Algorithms and applications for approximate nonnegative matrix factorization. Comput. Stat. Data Anal. 2007, 52, 155–173. [Google Scholar] [CrossRef] [Green Version]
  35. Qiu, H.; Lee, J.; Lin, J.; Yu, G. Wavelet filter-based weak signature detection method and its application on rolling element bearing prognostics. J. Sound Vib. 2006, 289, 1066–1090. [Google Scholar] [CrossRef]
Figure 1. Scheme of the feature extraction based on the SKNMF.
Figure 1. Scheme of the feature extraction based on the SKNMF.
Sensors 21 03680 g001
Figure 2. Bearing test rig.
Figure 2. Bearing test rig.
Sensors 21 03680 g002
Figure 3. Waveform and spectrum of the outer race defect: (a) waveform and (b) spectrum.
Figure 3. Waveform and spectrum of the outer race defect: (a) waveform and (b) spectrum.
Sensors 21 03680 g003
Figure 4. Time–frequency distribution of the outer race defect.
Figure 4. Time–frequency distribution of the outer race defect.
Sensors 21 03680 g004
Figure 5. Coefficient matrix obtained by SKNMF.
Figure 5. Coefficient matrix obtained by SKNMF.
Sensors 21 03680 g005
Figure 6. Selected time–frequency subspace.
Figure 6. Selected time–frequency subspace.
Sensors 21 03680 g006
Figure 7. Waveform and spectrum of the extracted signal: (a) waveform and (b) spectrum.
Figure 7. Waveform and spectrum of the extracted signal: (a) waveform and (b) spectrum.
Sensors 21 03680 g007
Figure 8. Envelope spectrum of the extracted signal (the first two order harmonics of fBPFO are highlighted by two red arrows).
Figure 8. Envelope spectrum of the extracted signal (the first two order harmonics of fBPFO are highlighted by two red arrows).
Sensors 21 03680 g008
Figure 9. The Kurtogram and the optimal frequency band highlighted by a red dash circle.
Figure 9. The Kurtogram and the optimal frequency band highlighted by a red dash circle.
Sensors 21 03680 g009
Figure 10. Filtered signal.
Figure 10. Filtered signal.
Sensors 21 03680 g010
Figure 11. Envelope spectrum of the filtered signal.
Figure 11. Envelope spectrum of the filtered signal.
Sensors 21 03680 g011
Figure 12. Waveform and spectrum of the 4th subspace obtained by the RSVD method: (a) waveform of the reconstructed signal and (b) spectrum of the reconstructed signal.
Figure 12. Waveform and spectrum of the 4th subspace obtained by the RSVD method: (a) waveform of the reconstructed signal and (b) spectrum of the reconstructed signal.
Sensors 21 03680 g012
Figure 13. Envelope spectrum of the reconstructed signal obtained by the RSVD method.
Figure 13. Envelope spectrum of the reconstructed signal obtained by the RSVD method.
Sensors 21 03680 g013
Figure 14. First six eigenvectors of Figure 4 extracted by SVD: (a) the first three eigenvectors and (b) the fourth–sixth eigenvectors.
Figure 14. First six eigenvectors of Figure 4 extracted by SVD: (a) the first three eigenvectors and (b) the fourth–sixth eigenvectors.
Sensors 21 03680 g014
Figure 15. Coefficient matrix obtained by NMF.
Figure 15. Coefficient matrix obtained by NMF.
Sensors 21 03680 g015
Figure 16. Selected time–frequency subspace for NMF.
Figure 16. Selected time–frequency subspace for NMF.
Sensors 21 03680 g016
Figure 17. Waveform and spectrum of the extracted signal: (a) waveform of the extracted signal and (b) spectrum of the extracted signal.
Figure 17. Waveform and spectrum of the extracted signal: (a) waveform of the extracted signal and (b) spectrum of the extracted signal.
Sensors 21 03680 g017
Figure 18. Envelope spectrum of the extracted signal.
Figure 18. Envelope spectrum of the extracted signal.
Sensors 21 03680 g018
Figure 19. Coefficient matrix obtained by SNMF.
Figure 19. Coefficient matrix obtained by SNMF.
Sensors 21 03680 g019
Figure 20. Selected time–frequency subspace for SNMF.
Figure 20. Selected time–frequency subspace for SNMF.
Sensors 21 03680 g020
Figure 21. Waveform and spectrum of the extracted signal: (a) waveform and (b) spectrum.
Figure 21. Waveform and spectrum of the extracted signal: (a) waveform and (b) spectrum.
Sensors 21 03680 g021
Figure 22. Envelope spectrum of the extracted signal (the fault characteristic fBPFO is highlighted by a red arrow).
Figure 22. Envelope spectrum of the extracted signal (the fault characteristic fBPFO is highlighted by a red arrow).
Sensors 21 03680 g022
Figure 23. Coefficient matrix heat map obtained by KNMF.
Figure 23. Coefficient matrix heat map obtained by KNMF.
Sensors 21 03680 g023
Figure 24. Selected time–frequency subspace for KNMF.
Figure 24. Selected time–frequency subspace for KNMF.
Sensors 21 03680 g024
Figure 25. Spectrum and envelope spectrum of the extracted signal: (a) spectrum and (b) envelope spectrum (the fault characteristic fBPFO is highlighted by a red arrow).
Figure 25. Spectrum and envelope spectrum of the extracted signal: (a) spectrum and (b) envelope spectrum (the fault characteristic fBPFO is highlighted by a red arrow).
Sensors 21 03680 g025
Figure 26. (a) Structure of the machinery fault simulator. (b) AC motor with acceleration sensor mounted.
Figure 26. (a) Structure of the machinery fault simulator. (b) AC motor with acceleration sensor mounted.
Sensors 21 03680 g026
Figure 27. Waveform and spectrum of the vibration signal with an inner race fault bearing: (a) waveform and (b) spectrum.
Figure 27. Waveform and spectrum of the vibration signal with an inner race fault bearing: (a) waveform and (b) spectrum.
Sensors 21 03680 g027
Figure 28. Time–frequency distribution of a defected bearing.
Figure 28. Time–frequency distribution of a defected bearing.
Sensors 21 03680 g028
Figure 29. Coefficient matrix heat map of SKNMF.
Figure 29. Coefficient matrix heat map of SKNMF.
Sensors 21 03680 g029
Figure 30. Extracted signal waveform.
Figure 30. Extracted signal waveform.
Sensors 21 03680 g030
Figure 31. Spectrum and envelop spectrum of the extracted signal: (a) spectrum and (b) envelope spectrum (the first three order harmonics of fBPFO are highlighted by three red arrows).
Figure 31. Spectrum and envelop spectrum of the extracted signal: (a) spectrum and (b) envelope spectrum (the first three order harmonics of fBPFO are highlighted by three red arrows).
Sensors 21 03680 g031
Figure 32. The Kurtogram and the optimal frequency band highlighted by a red dash circle.
Figure 32. The Kurtogram and the optimal frequency band highlighted by a red dash circle.
Sensors 21 03680 g032
Figure 33. Envelope spectrum of the filtered signal.
Figure 33. Envelope spectrum of the filtered signal.
Sensors 21 03680 g033
Figure 34. Spectrum and envelope spectrum of the 3rd subspace obtained by the RSVD method: (a) spectrum of the reconstructed signal and (b) envelope spectrum of the reconstructed signal.
Figure 34. Spectrum and envelope spectrum of the 3rd subspace obtained by the RSVD method: (a) spectrum of the reconstructed signal and (b) envelope spectrum of the reconstructed signal.
Sensors 21 03680 g034
Figure 35. Coefficient matrixes with other NMF-based methods: (a) NMF-based, (b) SNMF-based and (c) KNMF-based.
Figure 35. Coefficient matrixes with other NMF-based methods: (a) NMF-based, (b) SNMF-based and (c) KNMF-based.
Sensors 21 03680 g035
Figure 36. Envelope spectrums of the extracted signal with the other NMF-based methods. (a) NMF-based method, (b) SNMF-based method and (c) KNMF-based method (the fault characteristic fBPFO is highlighted by a red arrow in each method).
Figure 36. Envelope spectrums of the extracted signal with the other NMF-based methods. (a) NMF-based method, (b) SNMF-based method and (c) KNMF-based method (the fault characteristic fBPFO is highlighted by a red arrow in each method).
Sensors 21 03680 g036
Table 1. Technical details of the acceleration sensor.
Table 1. Technical details of the acceleration sensor.
Sensor ModelSensitivityMeasurement RangeBroadband ResolutionFrequency Range
PCB 352C68 ICP(±10%) 100 mV/g (10.2 mV/(m/s²))±50 g pk (±491 m/s² pk)0.00016 g rms (0.0015 m/s² rms)(±5%) 0.5 to 10,000 Hz
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liang, L.; Ding, X.; Liu, F.; Chen, Y.; Wen, H. Feature Extraction Using Sparse Kernel Non-Negative Matrix Factorization for Rolling Element Bearing Diagnosis. Sensors 2021, 21, 3680. https://doi.org/10.3390/s21113680

AMA Style

Liang L, Ding X, Liu F, Chen Y, Wen H. Feature Extraction Using Sparse Kernel Non-Negative Matrix Factorization for Rolling Element Bearing Diagnosis. Sensors. 2021; 21(11):3680. https://doi.org/10.3390/s21113680

Chicago/Turabian Style

Liang, Lin, Xingyun Ding, Fei Liu, Yuanming Chen, and Haobin Wen. 2021. "Feature Extraction Using Sparse Kernel Non-Negative Matrix Factorization for Rolling Element Bearing Diagnosis" Sensors 21, no. 11: 3680. https://doi.org/10.3390/s21113680

APA Style

Liang, L., Ding, X., Liu, F., Chen, Y., & Wen, H. (2021). Feature Extraction Using Sparse Kernel Non-Negative Matrix Factorization for Rolling Element Bearing Diagnosis. Sensors, 21(11), 3680. https://doi.org/10.3390/s21113680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop