4.2. Application of Improved Algorithm in Bearing Fault
Fault diagnosis model is constructed according to
Figure 2. This paper chooses West University of rolling bearing samples and the numbers of each state are at least 121,200. Test data and training data account for half of the total data. The detailed description of various bearing States is shown in
Table 2.
In this paper, the vibration signal is mainly processed from three aspects.
First, the feature extraction is performed by the time domain method.
The statistical characteristics of signal vibration amplitude will change with the location and the size of the fault. The time domain waveform is dynamically transformed over time. The amplitude of the vibration signal can reflect the characteristic information of the signal intuitively. The time domain waveform information can be used to diagnose the state of the bearing by analyzing the amplitude, shape and other characteristics of the waveform. The time domain characteristic parameters are different due to different fault types and different fault degree. Generally speaking, the time domain feature provides the global characteristics of bearing state, and can effectively extract the bearing fault feature.
In the actual situation, there is various information of bearing fault, and a faults are often accompanied with other faults, such as bearing deformation, corrosion and so on. In order to diagnose the fault more effectively, we need to extract the feature of bearing fault data. In this paper, 17-time domain extraction methods are used to extract the features of the signal.
In
Table 3,
is the representative of the signal sample
and
m represents the number of samples. Seventeen time domain feature attributes is:
the average value,
absolute mean,
effective value,
average power,
square amplitude,
peak,
peak-to-peak,
variance,
standard deviation,
skewness,
kurtosis,
waveform,
Crest index,
impluse index,
margin index,
skewness index and
kurtosis index.
Second, the main feature selection of feature extraction data is made by the Decision Tree.
The main description of the J48 algorithm is given in Chapter 3, and the output tree structure shown in
Figure 7. It can be seen from the diagram that the main characteristics of bearing data are
,
,
and
.
The 17 characteristic attributes obtained by feature extraction are interrelated with each other, which leads to data redundancy. The attributes with low correlation are obtained by extracting the main features with J48 so that the independence of data can be enhanced.
The description and significance of these four main time-domain features are as follows:
average value : is mainly used to reflect the trend of the bearing fault signal,
square amplitude : is mainly used to describe the energy of signals,
waveform index : is sensitive to fault signals with stable waveform,
kurtosis index : kurtosis is sensitive to bearing defects and can reliably reflect the state of rolling bearings. It is not easy to be affected by temperature, speed, etc. and comprehensive analysis of kurtosis, peak factor, and effective value.
In
Figure 7, the intermediate node represents the attribute of the decision with an ellipse, and the leaf node represents the classification result with a rectangle. The data between nodes are the classification condition. The graph is a part of the Decision Tree. Class label is a class with the highest probability in classification result when it has little effect on feature selection.
Third, the main feature of extraction is pruned with SSVM.
The J48 algorithm is mainly used to extract attribute vector so that the connection between data is reduced and the independence between data is enhanced. This paper mainly uses SSVM as mentioned above to reduce the similar attributes on the data dimension. The more similar the attribute is, the more redundant it would be. The data redundancy between the pruned data will be reduced so that the independence of the data dimension can be enhanced.
SSVM is used to select the appropriate data for pruning. When the data is removed excessively or removed too little, the classification result will be affected. Therefore, it is very important to choose the appropriate threshold. The threshold in this article is the accuracy rate of test data tested by SVM. When the accuracy is greater than a certain value, we think that these kinds of data are not redundant, so we do not prune it. Therefore, the classification data, which is below the threshold, is selected, and then remove the nearest neighbor inconsistent data.
Table 4 shows the selected data corresponding to the pruning data and the pruned training data set, and
Figure 8 is the test accuracy of the bearing data corresponding to the selection threshold. From
Table 4 and
Figure 8, it can be concluded that the data trimming is too small to make the classification effect not obvious, and too much data pruning will result in important data loss. It can be seen from
Figure 8 that, when the threshold is 0.9, the corresponding accuracy is the highest than others. Therefore, the training data with a threshold below 0.9 is selected for SSVM pruning. Only in this way can the fault diagnosis be performed effectively.
After processing, the vibration data from the three aspects above, the redundant data is removed from the feature vector and the dimension vector, respectively.
Figure 9 shows the three-dimensional data of time domain feature extractiont, three-dimensional data after J48 select feature, and three-dimensional data after J48 and SSVM trimming. The axes
x,
y, and
z in
Figure 9 are dimensional features. Among them,
Figure 9a selects three dimensions of mean, absolute mean and effective value,
Figure 9b,c select three dimensions of mean, waveform index and kurtosis index. It can be seen from
Figure 9 that each class of data has obvious overlap in
Figure 9a, the overlap ratio of each kind of data in
Figure 9b is obviously lower than of
Figure 9a, and
Figure 9c obviously separates each type of category data. Therefore, it is shown from
Figure 9 that the redundancy between the processed data is greatly reduced, so that the correlation between the data is reduced, and the influence of NB independence assumption on the fault diagnosis is finally reduced.
The processing bearing fault data correlation is low, which reduces the limitation of the independence assumption on NB fault diagnosis.
Table 5 is the confusion matrix of NB fault diagnosis for the processed data, and
Table 6 is a confusion matrix for bearing fault diagnosis using an NB model without redundant vibration data. As can be seen from the table, the model has been improved for each category after redundancy removal.
In order to verify the validity of this algorithm in bearing data, the data simulation is carried out by MATLAB (Version 8.6, The MathWorks, MA, USA).
Figure 10 and
Table 7 are bearing fault diagnosis results. In
Figure 10 and
Table 7,the meaning of NB+J48+SVM is that first data is selected by J48,then the data after feature selection is pruned by SVM and the fault diagnosis of NB is finally carried out. Compared with other experimental results, the bearing fault diagnosis experimental results on JSSVM-NB is better than removing the data redundancy by feature vector and data vector. Compared with other experiments, the accuracy of the fault diagnosis model is 99.17%.
Table 8 shows the comparison of results of about JSSVM-NB and reference [
35], which have the same data for bearing fault diagnosis. It can be seen from
Table 7 and
Table 8 that the JSSVM-NB model is effective for rolling bearing fault diagnosis.