Next Article in Journal
Advances in LiDAR Hardware Technology: Focus on Elastic LiDAR for Solid Target Scanning
Previous Article in Journal
MFMamba: A Mamba-Based Multi-Modal Fusion Network for Semantic Segmentation of Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Fault Diagnosis of Hydraulic System Based on Multiscale One-Dimensional Convolutional Neural Networks with Multiattention Mechanism

1
College of Mechanical and Vehicle Engineering, Taiyuan University of Technology, Taiyuan 030024, China
2
Shanxi Key Laboratory of Fully Mechanized Coal Mining Equipment, Taiyuan University of Technology, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(22), 7267; https://doi.org/10.3390/s24227267
Submission received: 5 October 2024 / Revised: 30 October 2024 / Accepted: 11 November 2024 / Published: 14 November 2024
(This article belongs to the Section Physical Sensors)

Abstract

:
Hydraulic systems are critical components of mechanical equipment, and effective fault diagnosis is essential for minimizing maintenance costs and enhancing system reliability. In practical applications, data from hydraulic systems are collected with varying sampling frequencies, coupled with complex interdependencies within the data, which poses challenges for existing fault diagnosis algorithms. To solve the above problems, this paper proposes an intelligent fault diagnosis of a hydraulic system based on a multiscale one-dimensional convolution neural network with a multiattention mechanism (MA-MS1DCNN). The proposed method first extracts features from multirate data samples using a parallel 1DCNN with different receptive fields. Next, a Hybrid Attention Module (HAM) is proposed, consisting of two submodules: the Correlation Attention Module (CAM) and the Importance Attention Module (IAM), which aim to meticulously and comprehensively model the complex relationships between channel features. Subsequently, to effectively utilize the feature information of different frequencies, the HAM is integrated into the 1DCNN to form the MA-MS1DCNN. Finally, the proposed method is evaluated and experimentally compared using the UCI hydraulic system dataset. The results demonstrate that, compared to existing methods such as Shapelet, MCIFM, and CNNs, the proposed method shows superior diagnostic performance.

1. Introduction

The hydraulic system is an efficient transmission device based on fluid mechanics, capable of precise control and powerful output. It has found widespread application in fields such as energy, transportation, and advanced manufacturing. However, hydraulic systems frequently experience failures because of the harsh environments and complex working conditions. If faults in the hydraulic system are not detected and repaired in a timely manner, they may further deteriorate, significantly impacting productivity and endangering production safety [1,2]. Thus, the study of fault diagnosis techniques for hydraulic systems is crucial in terms of enhancing the safety and reliability of mechanical equipment.
With the rapid development of sensor and data transmission technologies, shallow machine learning-based fault diagnosis methods that utilize large amounts of operational data have received increasing attention. For example, Guo et al. [3] proposed a method for identifying the health state of hydraulic systems based on integrated support vector machine. Lei et al. [4] combined principal component analysis (PCA) with extreme gradient boosting (XGBoost) algorithms to enhance the accuracy of fault diagnosis for hydraulic reversing valves. Joo et al. [5] used the Shapelet transform algorithm to achieve fault classification for temporal samples of hydraulic systems. Helwig et al. [6] employed linear discriminant analysis (LDA) to project time-domain and frequency-domain features into a lower-dimensional discriminant space for the predictive maintenance of hydraulic systems. Although fault diagnosis methods based on shallow machine learning have achieved significant results, they are unable to extract features on their own and require manually designed feature extraction algorithms [7]. In addition, shallow machine learning algorithms have limited expressive power to effectively handle complex operating conditions.
As an important branch of machine learning, deep learning methods have been widely used for various tasks, including image recognition [8], object detection [9], and machine translation [10]. The core of deep learning is feature learning. With the powerful modeling ability of deep networks, deep learning can automatically learn complex fault features from data, thus overcoming the shortcomings of shallow machine learning. In recent years, deep learning has gained more and more attention in the field of fault diagnosis [11,12,13]. For example, Wang et al. [14] combined the sliding window spectrum feature with a deep confidence network to realize the fault diagnosis of a hydraulic system. Tang et al. [15] utilized continuous wavelet transform to convert a one-dimensional (1D) pressure signal into a two-dimensional image, which was then input to a 2DCNN for the fault diagnosis of a hydraulic pump. Ji et al. [16] used the Dempster–Shafer theory to integrate the results of the LSTM, CNN, and RF algorithms to improve the diagnostic accuracy of an electromagnetic directional valve. Mallak et al. [17] proposed a structure that combines the LSTM autoencoder and deep learning to realize a fault monitoring and diagnostic method for a hydraulic system in stages. Deep learning algorithms based on a single signal have realized considerable progress in the field of fault diagnosis; however, they are limited to processing the data of a single sensor, which may result in extracted features that lack diversity or comprehensiveness.
Compared to single-sensor data, multisource data can provide more comprehensive information. Fault diagnosis methods based on multisource data will be the trend of the future. Shao et al. [18] proposed a stackable wavelet autoencoder structure with a Morlet wavelet function, which fuses noisy data with vibration data and improves fault diagnostic accuracy. Jing et al. [19] proposed a data fusion method based on a deep CNN that can fuse information from multiple sensors (vibration, noise, and current) adaptively. Wang et al. [20] proposed a method to convert vibration signals from multiple sensors into color images, and Huang et al. [21] designed a multirate data fusion CNN to address the problem where sensors have different sampling rates, and this method achieved fusion of multisource data at the feature level. Wang et al. [22] proposed a heterogeneous ensemble deep neural network (HEDNN) method based on a bagging strategy, which can synthesize the deep features of multisource data to improve the diagnostic accuracy. Buabeng et al. [23] fused flow, pressure, and power data using a systematic combination of hybrid clustering (CLUST) and an optimized multilayer perceptron neural network with unsupervised and supervised learning based on the grey wolf algorithm (GWO-MLP).
The attention mechanism can dynamically adjust focus on different data sources, enhancing the model’s ability to effectively utilize multisource information. Previous studies have designed various attention mechanisms [24,25,26,27] and successfully applied them in the fault diagnosis field. For example, Si et al. [28] introduces the self-attention mechanism into LSTM to achieve high diagnostic accuracy even under variable load and strong noise conditions. Yang et al. [29] presents an attention-parallel multisource data fusion residual network to solve the problems of high similarity of the fault features of different power components cascaded in H-bridge inverters. Song et al. [30] embedded a Convolutional Block Attention Module into a convolutional layer to enhance feature mapping, and good diagnostic results were obtained on limited datasets. Miao et al. [31] proposed an interactive channel attention mechanism to help the fault diagnosis model capture important information from both global and local channels. Fan et al. [32] proposed a multiattention feature fusion network to further enhance the optimization of features by the attention mechanism. Attention mechanisms can enhance the representation of feature information. However, the aforementioned studies merely address the implementation of a single attention mechanism and fail to examine the interactions among various types and scales of attention. Multiattention mechanisms can facilitate the creation of a more refined feature map, which warrants further investigation.
To solve these problems, this study proposes a multiscale 1DCNN with multiattention mechanism for hydraulic system fault diagnosis. The proposed method uses multiple convolutional layers of different scales to extract fault features in sensors with different sampling frequencies. Then, the HAM is employed to enhance the expression of related features. The primary contributions of this study are summarized as follows.
(1)
The proposed MA-MS1DCNN can extract deep features from time-series data of varying lengths, enabling end-to-end fault diagnosis. To the best of our knowledge, this is the first application of a multiattention-based multiscale 1DCNN in the fault diagnosis of hydraulic systems.
(2)
Two channel attention submodules that focus on different channel information, i.e., the IAM and CAM, are proposed. Further, the HAM is proposed by combining the submodules efficiently using residual connection. The HAM allows the adaptive optimization of features from both importance and relevance perspectives.
(3)
Based on a multiscale 1DCNN, this paper designs ablation and arrangement experiments for attention mechanisms, demonstrating that the interaction between different attention mechanisms can effectively improve the accuracy of fault diagnosis. Furthermore, compared to current mainstream methods for hydraulic system fault diagnosis, the proposed approach exhibits stronger diagnostic performance.

2. Theoretical Background

2.1. 1DCNN

A 1DCNN is a deep learning model that specializes in processing one-dimensional data (e.g., time-series and text). Similar to traditional convolutional neural networks (CNNs), 1DCNNs extract features through convolutional layers. The operation process of the convolutional layer can be expressed as follows:
x j l = f ( i x i l 1 × ω i j l + b j l )
where x j l is the j-th output feature of the l-th layer, and x i l 1 is the i-th input feature of the (l − 1)-th layer. f is the activation function, ω i j l denotes the convolution kernel filter that connects the i-th input features and the j-th output feature map, and b j l is the j-th bias of the l-th layer.

2.2. Residual Connection

To address the problem of network degradation due to increasing network depth, He et al. [33] proposed a residual network. A residual network introduces shortcut connections by adding the input directly to the output of each residual block, thereby facilitating cross-layer information propagation. The architectures of the residual block are shown in Figure 1.
When the input x has the same dimension as output y , the output of the residual block is given as follows:
y = f ( w , x ) + x
where f ( w , x ) represents the residual mapping to be fitted by the residual block. When the inputs x and y have different dimensions, a linear mapping must be added to the residual block to match the dimensions. The residual block output is expressed as follows:
y = f ( w , x ) + w x
where f ( w , x ) represents the residual mapping to be fitted by the residual block, and w represents the linear mapping. Shortcut connections provide an additional path for information transfer by adding the input signal to the output signal that has undergone multilayer transformation, thereby effectively mitigating the issue of information fading in deep networks.

2.3. Squeeze-and-Excitation Networks

Squeeze-and-Excitation Networks (SENet) are one of the most classical attention mechanisms and consists of two parts, Squeeze and Excitation. The Squeeze part compresses the features using a global average pool (GAP) to obtain statistical features, S , for each channel. The Excitation part excites the statistical features through a fully connected network to produce a set of weight coefficients, E . The structure is shown in Figure 2. The formulas of the above processes are as follows:
S = F s q ( X ) = 1 H × W i = 1 H j = 1 W x ( i , j )
E = F e x ( S ) = σ ( W 2 ( δ ( W 1 S ) ) )
X = X E

2.4. Efficient Channel Attention Network

An Efficient Channel Attention Network (ECA-Net) uses convolution to mine the correlation between channels, and its structure is shown in Figure 3. The straight lines in the figure represent convolution operations, and the different colors represent different parameters of the convolution layer. First, the ECA-Net uses a global average pooling layer to compress the input feature map to obtain a 1D feature map containing global information. Then, convolution is employed to capture local cross-channel correlation after reshaping the 1D feature map. Finally, the Sigmoid function is employed to generate the channel weights. The implementation process of ECANet is expressed as follows:
X ˜ = X σ ( F K 1 ( R S ( F g a p ( X ) ) ) )
where X is the input feature map, σ is the Sigmoid function, F K 1 is the convolution layer with a kernel size of K and kernel number of 1, R S represents the reshaping operation, X ˜ is the output feature map, and F g a p represents the global average pooling operation.

2.5. Selective Kernel Networks

By fusing multiscale features, networks can effectively utilize information from various scales to create a more comprehensive feature representation. Li et al. [34] proposed Selective Kernel Networks (SKNet), which incorporate a dynamic selection mechanism, allowing these networks to automatically choose different convolutional kernels based on the characteristics of input data, achieving more flexible and efficient feature extraction. The structure of SKNet is shown in Figure 4.
Firstly, two convolution filters with different sizes are used to generate multiscale feature U ¯ .
U ¯ = [ U 1 ; U 2 ] = [ F 3 1 ( X ) ; F 5 1 ( X ) ]
where F 3 1 represent the convolution layer with a kernel size of 3 and kernel number of 1. F 5 1 represent the convolution layer with a kernel size of 5 and kernel number of 1. The different scale features, U i , are fused by element-by-element summation to obtain the fused feature U . Subsequently, U is squeezed and excited to generate the compact feature Z.
U = U 1 U 2
Z = F f c ( F g a p ( U ) ) = δ ( β ( W ( S ) ) )
where represents element-wise summation, F g a p is the global average pooling operation, S is the channel-wise statistics, δ is the ReLU function, β represents batch normalization, W represents the linear map, and F f c represents the fully connected layer function. Finally, the weights of each feature scale are calculated using the Softmax function, and then, the weights are weighted and summed with the different scale features, U i , to achieve the dynamic fusion of features.
a = e A z e A z + e B z , b = e B z e A z + e B z
U = a U 1 + b U 2
where A and B represents the linear map, U is the output feature map, and a and b represent the soft attention vectors for U 1 and U 2 , respectively.

3. Proposed Fault Diagnosis Method

3.1. Correlation Attention Module

The ECA-Net is limited to the extraction of single-scale correlation and may not improve model performance considerably when handling multisource data with different sampling frequencies. Thus, based on the ECA model, this paper introduces a selective kernel strategy and proposed the Correlation Attention Module (CAM). The CAM can utilize adaptive weights to fuse multiscale correlation to enhance the expression ability of the attention mechanism for complex channel relationships. The CAM primarily comprises three main components, i.e., feature compression, channel correlation construction, and feature recalibration. The basic structure of the CAM is shown in Figure 5. The different stages of the feature maps in the figure are distinguished using different colors, where the attention feature map uses colors with different brightness to express the significant difference in the weights of the feature maps in the channel dimension.
Feature compression: Here, assume that the input to the CAM is X = [ x 1 , x 2 , , x C ] , x i R W × 1 . A global max pooling layer is used to compress the input X to obtain the global feature S M = [ S M 1 , S M 2 , , S M C ] R 1 × C . The i-th global feature can be expressed as follows:
S M i = M a x [ x i ( 1 ) , x i ( 2 ) , , x i ( W ) ] , S M i R 1 × 1
Then, the feature information S M is reshaped to obtain S M , S M R C × 1 .
Channel correlation construction: First, H convolution filters with different sizes are used to generate multiscale attention maps, V ¯ . In this paper, H is set as 2.
V ¯ = [ V 1 ; V 2 ] = [ F 3 1 ( S M ) ; F 5 1 ( S M ) ]
where F 3 1 represents the convolution layer with a kernel size of 3 and kernel number of 1. F 5 1 represents the convolution layer with a kernel size of 5 and kernel number of 1. The different scale attention maps, V i , are fused by element-by-element summation to obtain the channel-wise statistics, V , V R C × 1 . Then, a global average pooling layer is used to compress the statistics, V , and generate a feature map, S A .
V = V 1 V 2
S A = F g a p ( V ) = 1 1 × C j = 1 C v j
H fully connected layers are used to nonlinearly transform the feature S A to obtain final features, S A i , corresponding to H different scales. Lastly, the Softmax function is utilized to determine the weights, w i , for each scale. The process of calculating the weights can be expressed as follows:
[ S A 1 , S A 2 ] = [ F f c 1 ( S A ) , F f c 2 ( S A ) ] = [ δ ( β ( W 1 ( S A ) ) ) , δ ( β ( W 2 ( S A ) ) ) ]
w i = exp ( S A i ) j = 1 2 exp ( S A j )
where F f c i represents the i-th fully connected layer function, δ represents the ReLU activation function, β represents batch normalization, and W 1 and W 2 represent the linear mapping.
Feature calibration: The different scale attention maps, V i , are multiplied by the corresponding weights, w i , and then summed to obtain the fusion feature map. Then, the reshaping operation, together with the Sigmoid function, is used to obtain the channel correlation weights, I C A M . The input feature map X is multiplied element by element with I C A M to obtain the weighted feature map Y C A M . The feature calibration can be expressed as follows:
I C A M = R S ( σ ( w 1 V 1 + w 2 V 2 ) )
Y C A M = I C A M X
where σ represents the Sigmoid activation function, represents element-wise product, R S represents the reshaping operation, w 1 and w 2 represent the soft attention vectors for V 1 and V 2 , respectively.

3.2. Importance Attention Module

The CAM uses convolution to extract features to model the correlation between channels. However, limited by the size of the convolution kernel, the CAM can only focus on locally neighboring channels, making it difficult to effectively express global channel information. Therefore, based on the SENet model, an Importance Attention Module (IAM) is constructed to model the global importance of the channels, and the structure of the IAM is shown in Figure 6.
First, to avoid information loss, two layers of point-wise convolutions with the same channel dimension are used to capture the importance of channels. Then, the attention feature map is processed by the Sigmoid function to obtain the weight I I A M . Finally, the output of the IAM, Y I A M , is obtained by multiplying I I A M with the input feature map. Assuming that the input feature map of the module is X, the above process can be expressed as follows:
I I A M = σ ( F 1 C ( δ ( F 1 C ( F g m p ( X ) ) ) ) )
Y I A M = I I A M X
where F 1 C represents the convolution layer with a kernel size of 1 and kernel number of C , F g m p represents the global max pooling operation, δ represents the Relu activation function, and σ represents the Sigmoid activation function.

3.3. Hybrid Attention Module

To fully leverage the intricate information across channels, we devised the optimal fusion of the IAM and CAM, resulting in the introduction of the HAM. The HAM is a combination of the IAM and CAM that integrates the channel importance and correlation information, and it can optimize the input features from different perspectives to further enhance the diagnostic ability of the model. Here, the residual connection concept is introduced to establish a shortcut connection between the input features and the recalibration features, which improves the feasibility of channel information extraction while preserving the original information. The structure of the HAM is shown in Figure 7.
The output of the HAM can be expressed as follows: I have checked I have checked I have checked I have checked
Y I A M = I I A M X + X
Y H A M = I C A M Y I A M + Y I A M
where X is the input feature map of the HAM, I I A M is the importance attention weight, Y I A M is the output feature map of the IAM, I C A M is the correlation attention weight, and Y H A M is the output feature map of the HAM.

3.4. MA-MS1DCNN

The constructed attention module is embedded into a 1DCNN to propose an MA-MS1DCNN. The 1DCNN was selected because the signals collected by the sensors are time-series data represented as one-dimensional structures, which are suitable for input to a 1DCNN. In addition, a 2DCNN typically needs to convert one-dimensional sensor signals (e.g., via wavelet transform and Fourier transform). However, in this conversion process, the sequence and period information of the original signal may be lost. The 1DCNN also has a smaller number of parameters, which facilitates effective real-time diagnosis compared to the 2DCNN. For example, the input layer of the 2DCNN in [13] was 224 × 224 × 3. The MA-MS1DCNN comprises a feature extraction layer, a feature fusion layer, and a fault classification layer, and its overall structure is shown in Figure 8.
The feature extraction layer comprises multiple feature extractors, which are used to extract deep features from raw data with different sampling rates. Each feature extractor contains two convolution blocks comprising a convolution layer, a normalization layer, an activation function layer, and a pooling layer. To suppress high-frequency noise and preserve periodic information, unlike image processing methods that frequently use convolution kernels of size 3 × 3 or 5 × 5, in the first convolutional layer, the MA-MS1DCNN uses a large convolution kernel (e.g., a convolution kernel size of 100 for the 100 Hz feature extractor) to extract the features. Then, the ReLU activation function is applied after the convolutional layer. Through this activation function’s mapping, the MA-MS1DCNN can learn more complex nonlinear features. Finally, the max pooling layer is used to reduce the feature dimension while extracting the most representative features. Here, the input of the r-th feature extractor is X r = [ X 1 r , X 2 r , , X N r ] , X i r R M × 1 , where N is the number of sensors at the same sampling rate, and M is the number of samples. After the convolution operation, the timing data are converted into a multichannel feature map Z = [ Z 1 , Z 1 , , Z r , , Z R ] . The r-th multichannel feature map is expressed as Z r = [ Z 1 r , Z 2 r , , Z C r ] , Z i r R W × 1 , where W is the length of the feature map and C is the number of feature map channels.
To reduce the redundancy of the features and optimize the extracted features, in the feature fusion layer, the multichannel feature map Z undergoes processing by the HAM, where the channel weights are assigned to recalibrate the multichannel feature map Z into feature Y = [ Y 1 , Y 2 , , Y r , , Y R ] , Y r R W × C . This process does not change the size and dimension of the features. Then, feature fusion is performed via flattening and concatenation to output the fused feature map T . Using the attention mechanism to guide multisource data fusion allows the model to utilize important information from multisource data selectively to learn more complex and useful features, which improves the fault diagnosis performance. This process is expressed as follows.
Y f l r = F f l a t t e n ( Y r ) = [ y 11 r , y 12 r , , y 1 W r , , y C 1 r , y C 2 r , , y C W r ]
T = [ Y f l 1 , Y f l 2 , , Y f l r , , Y f l R ]
where F f l a t t e n represents the flattening operation, W is the feature length after recalibration, C is the number of feature map channels after recalibration, and Y f l r represents the r-th recalibration feature after the flattening operation.
The fault classification layer uses two fully connected layers to reduce the dimensionality of the fusion feature T in a layer-by-layer manner, and the ReLU activation function is implemented to enhance the model’s representation ability. In addition, the Softmax function is used to convert the output of the fully connected layer into the probability of each category of fault.
G = F f c ( δ ( F f c ( T ) ) )
P j = exp ( G j ) k = 1 K exp ( G k )
where P j is the probability of category j , F f c is the fully connected layer function, K is the number of fault categories, and G is the Softmax input vector.

4. Experiment and Discussion

4.1. Experimental Dataset

In this study, the hydraulic system dataset from the UCI database was used to verify the performance of the proposed fault diagnosis method. The structure of the hydraulic system is shown in Figure 9, which is briefly described below.
The load level of the hydraulic system is controlled by a (V11) proportional pressure valve. By adjusting the parameters of the V11, the hydraulic system can simulate either a stable working cycle with a constant load or an unstable working cycle with a uniform load change. Here, the working period was fixed to 60 s, and a total of 2205 samples were collected (1449 samples of a steady-state work cycle and 756 samples of an unstable work cycle). To verify the generalizability of the proposed method, this experiment did not differentiate the data of different states. The data were collected from seventeen sensors (fourteen real sensors and three virtual sensors), which were categorized into three different sampling frequencies: seven sensors with a frequency of 100 Hz (including six pressure sensors (PS1–PS6) and one power sensor (ESP1)), two pressure sensors at 10 Hz (FS1–FS2), and eight sensors at 1 Hz (including four temperature sensors (TS1–TS4), one vibration (VS) sensor, one cooling power (CP) sensor, one cooling efficiency (CE) sensor, and one system efficiency (SE) sensor). In addition, failure sample data were acquired by simulating the degradation of several important components (i.e., the hydraulic cooler, pumps, valves, and accumulators). The details of the component degradation are shown in Table 1. Note that component degradation may occur simultaneously, which increases the difficulty of the fault diagnosis process.

4.2. Experimental Details

To compare the experimental results, the selection of sensors was consistent with that reported in the literature [20]. Here, three types of sensors with different frequencies were used, including sensors PS1–PS5 (sampling frequency: 100 Hz), FS1 and FS2 (sampling frequency: 10 Hz), and the TS1, VS, CP, and SE sensors (sampling frequency: 1 Hz). The model parameters are shown in Table 2, taking the pump fault diagnosis method as an example. The 2205 samples were randomly disrupted and divided into training and testing sets in a ratio of 8:2. Both training and testing were based on the Pytorch framework in Python 3.9. The hardware environment included a 12th Gen Intel(R) Core(TM) i7-12650H 2.30 GHz processor (Intel, Santa Clara, CA, USA) and an NVIDIA GeForce RTX 4060 GPU (NVIDIA, Santa Clara, CA, USA). Each sample was processed by taking the standard normalization method of subtracting mean and dividing by standard deviation to accelerate the convergence of the model. The Adam optimizer was used during the training process, and the cross-entropy loss function was selected for backpropagation to update the parameters. Here, the learning rate was updated using the learning rate decay approach, which was scaled down to 10% of the original learning rate every 50 iterations. The selection of each hyperparameter was determined based on repeated experiments and manual optimization to obtain the best results. The hyperparameters of the proposed network model are shown in Table 3. It should be noted that the proposed method is robust against parameter selection, and parameter variations have little effect on the final results. The pump fault diagnosis model is used as an experimental object to conduct hyperparameter sensitivity experiments, and the specific results are shown in Table 4.

4.3. Feature Visualization Analysis

To gain an intuitive understanding of the impact of the attention mechanism on the fault diagnosis of hydraulic systems, we selected a hydraulic pump as the analysis subject and employed t-SNE [35] to visualize the feature output at different stages of the model. First, we input the test data to the trained model and extracted feature maps from the various stages of the output. Then, the extracted high-dimensional feature maps were mapped to a 2D plane using t-SNE, where the features were represented as points within a 2D coordinate system. Finally, a scatter plot was generated based on these coordinates. To verify the effectiveness of the HAM, this paper selectively extracts three kinds of features: the features learned by the feature extractor, the features optimized by the HAM, and the features learned by the classification. As shown in Figure 10, after feature extraction, the fault samples with severe leakage were nearly completely separated from the other types of samples, and significant overlap between the samples with weak leakage and no leakage can be observed. After recalibrating the channel weights using the proposed attention mechanism, the number of overlapping samples was reduced significantly, which indicates that introducing the attention mechanism can significantly improve the fault identification performance. The results of the final classifier are shown in Figure 10c. As can be seen, three types of fault samples are completely separated, the clustering boundary is clear, and there is no overlapping phenomenon. These results indicate that the proposed fault diagnosis framework can effectively realize accurate diagnosis of the hydraulic system.

4.4. Attention Mechanism and Multiscale Feature Analysis

To analyze the proposed attention mechanism, the weight vectors output from the three attention modules in the feature fusion layer are visualized in Figure 11, where the attention weight maps are arranged horizontally in channel order, and the different weights are displayed in different colors. The visualization of the weights reveals the weight values of channels vary from 0.10 to 0.96. These results indicate that the proposed attention mechanism can achieve feature information selection by suppressing irrelevant features. The observations also demonstrate that, even after sensor data filtering, the attention mechanism suppresses information from more than half of the channels, and the weights of approximately one-quarter of the channels were concentrated around 0.1, thereby exhibiting a significant suppression effect. In addition, due to the parameter-sharing nature of the convolution operation, the information between the neighboring channels may be highly correlated. However, the visualization results show significant weight differences, even among adjacent channels, indicating that the proposed mechanism accurately captures channel information and assigns appropriate weights.
Furthermore, the weights w i of different scale features in all CAMs were extracted to analyze the role of the multiscale features in terms of constructing the attention mechanism. Here, to avoid particularity and contingency, 10 experiments were conducted to obtain the average value. The corresponding results are shown in Table 5.
The results demonstrate that the weights of different scale features were roughly distributed in the range of 40–60%. Despite the differences in the importance of different branches, no extreme suppression or enhancement was observed, as in the case of channel attention weights. These findings demonstrate that the channel correlation requires the multiscale feature fusion process to capture more comprehensive information from the features at different scales.

4.5. Attention Module Ablation Experiment

Ablation experiments were performed to further verify the effectiveness and necessity of the proposed attention module. The four networks considered in these experiments were structured as follows:
(1)
1DCNN without the attention module;
(2)
1DCNN with the IAM;
(3)
1DCNN with the CAM;
(4)
MA-MS1DCNN;
The mean values obtained over 10 ablation experiments are shown in Table 6.
Figure 12 shows that the 1DCNN-IAM and 1DCNN-CAM methods obtained better fault diagnosis performance compared to the 1DCNN method, which indicates that the IAM and CAM can optimize the features based on the importance and correlation information of the channels, and this improved the accuracy of fault diagnosis. Note that the MA-MS1DCNN, which integrates two attention modules, exhibited optimal performance in all diagnostic tasks. These results suggest that fusing different levels of information in the process of constructing the channel attention can enrich and deepen the feature characterization and further improve the effectiveness of the feature extraction process.

4.6. Arrangement of IAM and CAM

Due to the differences in the functionality of the two submodules, the arrangement may affect the overall performance when they are combined. Thus, three different arrangements of the IAM and CAM were compared in this study, i.e., serial IAM-CAM, serial CAM-IAM, and parallel use of the IAM and CAM. Table 7 summarizes the fault diagnostic accuracy of these distinct arrangements. The results indicate that the accuracy of the serial structure was higher than that of the parallel structure, which may be due to the fact that the direct summing operation in the parallel structure causes the two attention modules to interfere with each other and does not synthesize the multiple units of information between channels effectively. In addition, the performance of the serial IAM-CAM arrangement was better than that of the serial CAM-IAM arrangement. By analyzing the working principle of the submodules, the following conclusion can be obtained. Through the processing of the IAM, the CAM will focus on constructing correlations between important channels, thereby generating a finer attention feature map. Finally, the performance of all three arrangements was better than using any submodule alone, which suggests that synthesizing multiple units of information about the channels is important, and a reasonable arrangement will further enhance performance.

4.7. Comparison with Existing Model

To verify the superior performance of the proposed method, it was compared with eight existing methods (the SVM [36], Shapelet [6], ETSC [37], CLUST-GWO-MLP [23], LDA [6], MCIFM [38], CNN [21], and HEDNN [22] methods) on the same experimental dataset. The SVM is a classical shallow machine learning algorithm that is frequently used as a classifier. Shapelet transformation converts time-series data to distance data and then performs fault classification. ETSC is an early classification algorithm specialized for time-series data that enables classification decisions before the time-series data are fully observed. CLUST-GWO-MLP uses unsupervised clustering (CLUST) to group features to reduce complexity and supervised learning via optimized multilayer perceptron neural networks (GWO-MLP) to classify faults. LDA reduces the dimensionality of manually selected high-dimensional features to a low-dimensional discriminant space to realize fault classification. Note that the structures of the CNN and MCIFM methods are similar in that they both adopt a parallel network structure. The difference is that the CNN method uses multiple 1DCNNs with different receptive fields in the feature extraction stage. HEDNNs use a bagging strategy to integrate different DNNs for robust diagnostic results. To reduce the influence of randomness, the above methods were evaluated over 10 repeated experiments, and the average accuracy was considered as the final result. The corresponding results are shown in Table 8.
Figure 13 shows that the SVM and ETSC methods, as shallow learning methods, were limited by the difficulty of extracting features and has relatively poor performance. In addition, the feature extraction and model training of the Shapelet, LDA, and CLUST-GWO-MLP methods cannot be optimized jointly, and their performance is not significantly better than that of the SVM and ETSC methods. The MCIFM, CNN, and HEDNN methods involve the powerful feature extraction ability of deep learning, and the diagnostic accuracy was improved considerably. However, all the methods failed to filter the extracted features effectively, and the redundant information will affect the diagnostic accuracy. The proposed method can effectively screen multichannel information and enhance the expression of useful information. Compared to the existing methods, the proposed method achieved excellent results in diagnosing four hydraulic components, showing strong stability. The accuracy for the cooler, valve, and accumulator, as well as the average accuracy across all four components, surpassed that of the compared methods. These findings indicate that the proposed method enables high-precision fault diagnosis for various components in hydraulic systems.
Figure 14 shows the confusion matrix of the fault diagnosis results for the four hydraulic components. Here, the horizontal axis represents the predicted label of the sample, the vertical axis represents the true label of the sample, and the diagonal of the matrix is the number of samples that were predicted correctly. As can be seen, all samples were classified correctly in the fault diagnosis of the cooler and hydraulic valve components. For the pump and accumulator components, only a few samples were misclassified into adjacent classes, which was primarily caused by the lack of obvious fault characteristics of the sensor signals under continuous degradation.

5. Conclusions

This article proposed the MA-MS1DCNN, which introduces an attention mechanism into the fault diagnosis method for hydraulic systems. MA-MS1DCNN employs a multiscale 1DCNN to extract features from raw data at various sampling rates. Subsequently, it develops an HAM that consists of two attention submodules designed to simultaneously capture the importance and relevance information of the channels. This approach enhances the feature selection ability and the fault detection performance of the 1DCNN. The conclusions drawn from this study are as follows: (1) The proposed attention mechanism has demonstrated its effectiveness in enabling the model to comprehend the complex relationships among multisource data, thus allowing for the learning of more salient features. (2) The multiattention mechanism can optimize features from different perspectives, and a judicious combination of these mechanisms maximizes the optimization effect. (3) The efficacy of the proposed method is validated using publicly available hydraulic system datasets, with the experimental results indicating that the diagnostic accuracy of this method in four hydraulic components—the cooler, valve, pump, and accumulator—is maintained at a high level, specifically 100%, 100%, 99.76%, and 99.69%, respectively. In comparison with eight state-of-the-art fault diagnosis methods, the MA-MS1DCNN showed significant advantages in accuracy and stability.
In this study, the attention mechanism was introduced to the fault diagnosis process for hydraulic systems, which further improved the performance of the model. However, there still exist some limitations in the current work, since the CAM and IAM employ a global pooling layer to compress the data in the time dimension when capturing channel significance, which can lead to the loss of some details and local information. In addition, the proposed method does not consider the effect of spatial features on fault diagnosis. Thus, based on the current method, designing an appropriate network structure to utilize the spatial features effectively to improve the accuracy of fault diagnosis will also be the focus of future work.

Author Contributions

Conceptualization, J.S. and H.D.; methodology, J.S.; software, N.L.; validation, J.S., X.S. and X.D.; formal analysis, J.S.; investigation, H.D.; resources, H.D.; data curation, J.S.; writing—original draft preparation, J.S.; writing—review and editing, N.L., X.S. and X.D.; visualization, N.L.; supervision, H.D.; project administration, H.D.; funding acquisition, H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 52174148), in part by the Shanxi Science and Technology Major Special Plan “Unveiling and Leading” Project (No. 202101010101018), in part by the Shanxi Province Science and Technology Cooperation and Exchange Special Project (No. 202104041101003), in part by the Shanxi Science and Technology Innovation Talent Team Project (No. 202204051001017), and in part by the Shanxi Provincial Key Research and Development Project (No. 202202100401013).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Guan, H.; Ren, Y.; Tang, H.; Xiang, J. Intelligent fault diagnosis methods for hydraulic components based on information fusion: Review and prospects. Meas. Sci. Technol. 2024, 35, 082001. [Google Scholar] [CrossRef]
  2. Dai, J.; Tang, J.; Huang, S.; Wang, Y. Signal-Based Intelligent Hydraulic Fault Diagnosis Methods: Review and Prospects. Chin. J. Mech. Eng. 2019, 32, 11–32. [Google Scholar] [CrossRef]
  3. Guo, P.; Wu, J.; Xu, X.; Cheng, Y.; Wang, Y. Health condition monitoring of hydraulic system based on ensemble support vector machine. In Proceedings of the 2019 Prognostics and System Health Management Conference (PHM-Qingdao), Qingdao, China, 25–27 October 2019; pp. 1–5. [Google Scholar]
  4. Lei, Y.; Jiang, W.; Jiang, A.; Zhu, Y.; Niu, H.; Zhang, S. Fault Diagnosis Method for Hydraulic Directional Valves Integrating PCA and XGBoost. Processes 2019, 7, 589. [Google Scholar] [CrossRef]
  5. Joo, Y.; Jeong, J.; Bae, S. Under Sampling Bagging Shapelet Transformation for Hydraulic System. In Proceedings of the International Conference on Omni-Layer Intelligent Systems, Association for Computing Machinery, Crete, Greece, 5–7 May 2019; pp. 13–18. [Google Scholar]
  6. Helwig, N.; Pignanelli, E.; Schutze, A. Condition monitoring of a complex hydraulic system using multivariate statistics. In Proceedings of the 2015 IEEE International Instrumentation and Measurement Technology Conference (I2MTC) Proceedings, Pisa, Italy, 11–14 May 2015; Volume 2015, pp. 210–215. [Google Scholar]
  7. Li, Z.B.; Jiang, W.L.; Zhang, S.; Sun, Y.; Zhang, S.Q. A Hydraulic Pump Fault Diagnosis Method Based on the Modified Ensemble Empirical Mode Decomposition and Wavelet Kernel Extreme Learning Machine Methods. Sensors 2021, 21, 2599. [Google Scholar] [CrossRef] [PubMed]
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  9. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  10. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to Sequence Learning with Neural Networks. Adv. Neural Inf. Process. Syst. 2014, 2, 3104–3112. [Google Scholar]
  11. Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep Transfer Learning for Bearing Fault Diagnosis: A Systematic Review Since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 3508221. [Google Scholar] [CrossRef]
  12. Neupane, D.; Seok, J. Bearing Fault Detection and Diagnosis Using Case Western Reserve University Dataset With Deep Learning Approaches: A Review. IEEE Access 2020, 8, 93155–93178. [Google Scholar] [CrossRef]
  13. Shao, S.; McAleer, S.; Yan, R.; Baldi, P. Highly Accurate Machine Fault Diagnosis Using Deep Transfer Learning. IEEE Trans. Ind. Inform. 2019, 15, 2446–2455. [Google Scholar] [CrossRef]
  14. Wang, X.; Huang, J.; Ren, G.; Wang, D. A hydraulic fault diagnosis method based on sliding-window spectrum feature and deep belief network. J. Vibroeng. 2017, 19, 4272–4284. [Google Scholar] [CrossRef]
  15. Tang, S.; Zhu, Y.; Yuan, S. An adaptive deep learning model towards fault diagnosis of hydraulic piston pump using pressure signal. Eng. Fail. Anal. 2022, 138, 106300. [Google Scholar] [CrossRef]
  16. Ji, X.; Ren, Y.; Tang, H.; Shi, C.; Xiang, J. An intelligent fault diagnosis approach based on Dempster-Shafer theory for hydraulic valves. Measurement 2020, 165, 108129. [Google Scholar] [CrossRef]
  17. Mallak, A.; Fathi, M. Sensor and Component Fault Detection and Diagnosis for Hydraulic Machinery Integrating LSTM Autoencoder Detector and Diagnostic Classifiers. Sensors 2021, 21, 433. [Google Scholar] [CrossRef] [PubMed]
  18. Shao, H.; Lin, J.; Zhang, L.; Galar, D.; Kumar, U. A novel approach of multisensory fusion to collaborative fault diagnosis in maintenance. Inf. Fusion 2021, 74, 65–76. [Google Scholar] [CrossRef]
  19. Jing, L.; Wang, T.; Zhao, M.; Wang, P. An Adaptive Multi-Sensor Data Fusion Method Based on Deep Convolutional Neural Networks for Fault Diagnosis of Planetary Gearbox. Sensors 2017, 17, 414. [Google Scholar] [CrossRef]
  20. Wang, H.; Li, S.; Song, L.; Cui, L.; Wang, P. An Enhanced Intelligent Diagnosis Method Based on Multi-Sensor Image Fusion via Improved Deep Learning Network. IEEE Trans. Instrum. Meas. 2020, 69, 2648–2657. [Google Scholar] [CrossRef]
  21. Huang, K.; Wu, S.; Li, F.; Yang, C.; Gui, W. Fault Diagnosis of Hydraulic Systems Based on Deep Learning Model With Multirate Data Samples. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6789–6801. [Google Scholar] [CrossRef]
  22. Wang, J.L.; Xu, H.Y.; Liu, J.; Peng, X.Y.; He, C.M. A bagging-strategy based heterogeneous ensemble deep neural networks approach for the multiple components fault diagnosis of hydraulic systems. Meas. Sci. Technol. 2023, 34, 988–1004. [Google Scholar] [CrossRef]
  23. Buabeng, A.; Simons, A.; Frempong, N.K.; Ziggah, Y.Y. A novel hybrid predictive maintenance model based on clustering, smote and multi-layer perceptron neural network optimised with grey wolf algorithm. SN Appl. Sci. 2021, 3, 593. [Google Scholar] [CrossRef]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  25. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  26. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef]
  27. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  28. Si, Y.P.; Wang, R.J.; Zhang, S.Q.; Zhou, W.T.; Lin, A.H.; Wang, Y.C. Fault Diagnosis Based on Attention Collaborative LSTM Networks for NPC Three-Level Inverters. IEEE Trans. Instrum. Meas. 2022, 71, 3512416. [Google Scholar] [CrossRef]
  29. Yang, W.M.; Gu, J.F.; Wang, X.G.; Wang, W.N. Attention-parallel multisource data fusion residual network-based open-circuit fault diagnosis of cascaded H-bridge inverters. J. Power Electron. 2024, 24, 875–886. [Google Scholar] [CrossRef]
  30. Song, S.S.; Zhang, S.Q.; Dong, W.; Li, G.C.; Pan, C.Y. Multi-source information fusion meta-learning network with convolutional block attention module for bearing fault diagnosis under limited dataset. Struct. Health Monit. 2024, 23, 818–835. [Google Scholar] [CrossRef]
  31. Miao, J.G.; Deng, C.Y.; Zhang, H.; Miao, Q. Interactive channel attention for rotating component fault detection with strong noise and limited data. Appl. Soft Comput. 2023, 138, 110171. [Google Scholar] [CrossRef]
  32. Fan, Z.X.; Xu, X.G.; Wang, R.J.; Wang, H.J. Fan Fault Diagnosis Based on Lightweight Multiscale Multiattention Feature Fusion Network. IEEE Trans. Ind. Inform. 2022, 18, 4542–4554. [Google Scholar] [CrossRef]
  33. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  34. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  35. Laurens, V.D.M.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  36. Xu, Z.; Yu, H.; Tao, J.; Liu, C. Compound fault diagnosis in hydraulic system with multi-output SVM. In Proceedings of the CSAA/IET International Conference on Aircraft Utility Systems (AUS 2020), Online, 18–21 September 2020; Volume 2020, pp. 84–89. [Google Scholar]
  37. Askari, B.; Carli, R.; Cavone, G.; Dotoli, M. Data-Driven Fault Diagnosis in a Complex Hydraulic System Based on Early Classification. In Proceedings of the 1st IFAC Workshop on Control of Complex Systems (COSY), Bologna, Italy, 24–25 November 2022; pp. 187–192. [Google Scholar]
  38. Ma, X.; Wang, P.; Zhang, B.; Sun, M. An improved Multi-channels Information Fusion Model with multi-scale signals for fault diagnosis. In Proceedings of the 2021 IEEE 2nd International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Nanchang, China, 26–28 March 2021; pp. 552–555. [Google Scholar]
Figure 1. Architecture of (a) residual block without linear mapping and (b) residual block with linear mapping.
Figure 1. Architecture of (a) residual block without linear mapping and (b) residual block with linear mapping.
Sensors 24 07267 g001
Figure 2. Structure of Squeeze-and-Excitation Networks.
Figure 2. Structure of Squeeze-and-Excitation Networks.
Sensors 24 07267 g002
Figure 3. Structure of Efficient Channel Attention Network.
Figure 3. Structure of Efficient Channel Attention Network.
Sensors 24 07267 g003
Figure 4. Structure of Selective Kernel Networks.
Figure 4. Structure of Selective Kernel Networks.
Sensors 24 07267 g004
Figure 5. Structure of Correlation Attention Module.
Figure 5. Structure of Correlation Attention Module.
Sensors 24 07267 g005
Figure 6. Structure of Importance Attention Module.
Figure 6. Structure of Importance Attention Module.
Sensors 24 07267 g006
Figure 7. Structure of Hybrid Attention Module.
Figure 7. Structure of Hybrid Attention Module.
Sensors 24 07267 g007
Figure 8. Framework of the proposed method.
Figure 8. Framework of the proposed method.
Sensors 24 07267 g008
Figure 9. Structure of hydraulic system.
Figure 9. Structure of hydraulic system.
Sensors 24 07267 g009
Figure 10. Visualization of the test samples using t-SNE: (a) features learned by the feature extractor, (b) features optimized by the HAM, and (c) features learned by the classification.
Figure 10. Visualization of the test samples using t-SNE: (a) features learned by the feature extractor, (b) features optimized by the HAM, and (c) features learned by the classification.
Sensors 24 07267 g010
Figure 11. Weight visualization results of three attention modules.
Figure 11. Weight visualization results of three attention modules.
Sensors 24 07267 g011
Figure 12. Accuracy of ablation experiments.
Figure 12. Accuracy of ablation experiments.
Sensors 24 07267 g012
Figure 13. Classification accuracy of different models.
Figure 13. Classification accuracy of different models.
Sensors 24 07267 g013
Figure 14. Confusion matrix of the proposed method: (a) cooler, (b) valve, (c) pump, and (d) accumulator components.
Figure 14. Confusion matrix of the proposed method: (a) cooler, (b) valve, (c) pump, and (d) accumulator components.
Sensors 24 07267 g014
Table 1. Component failure and failure levels.
Table 1. Component failure and failure levels.
ComponentFaultConditionSample
Number
Cooler C1Cooling
power
decrease
Full
efficiency
732
Reduced
efficiency
732
Close to total failure741
Valve V10Switching
degradation
Optimal switching behavior1125
Small lag360
Severe lag360
Close to total failure360
Pump MP1Internal
leakage
Severe
leakage
1221
Weak
leakage
492
No leakage492
Accumulators
A1–A4
Gas leakageOptimal
pressure
808
Slightly
reduced
pressure
399
Severely reduced
pressure
399
Close to total failure599
Table 2. Parameters in fault diagnosis of the pump.
Table 2. Parameters in fault diagnosis of the pump.
Extractor123
Input length600060060
Number of sensors528
Output channel (Conv1)16816
Size (Conv1)100305
Strides (Conv1)811
Output channel (Conv2)321632
Size (Conv2)1573
Strides (Conv2)311
Output nodes (fc1)32
Output nodes (fc2)3
Table 3. Training parameter settings.
Table 3. Training parameter settings.
ComponentCoolerValvePumpAcc
OptimizerAdamAdamAdamAdam
Minibatch64643232
Epochs125125125200
Learning rate0.0010.0010.00050.0005
Drop factor0.10.10.10.1
Drop epoch50505060
Table 4. Results of parameter sensitivity analysis.
Table 4. Results of parameter sensitivity analysis.
Batch SizeLearning Rate
0.010.0010.00050.00001
1699.4599.7099.7399.68
3299.5399.7299.7699.71
6499.5799.6999.7299.65
12899.4999.6799.7099.63
Table 5. Weights of different branches.
Table 5. Weights of different branches.
WeightCAM1CAM2CAM3
w 1 0.520.420.61
w 2 0.480.580.39
Table 6. Results of ablation experiments.
Table 6. Results of ablation experiments.
MethodComponent
CoolerValvePumpAcc
1DCNN99.8610098.5397.43
1DCNN-IAM99.9710099.1599.07
1DCNN-CAM99.9510099.0998.95
MA-MS1DCNN10010099.7699.69
Table 7. Combining methods of IAM and CAM.
Table 7. Combining methods of IAM and CAM.
MethodComponent
CoolerValvePumpAcc
Parallel IAM and CAM99.9810099.2399.10
Serial CAM-IAM10010099.4999.41
Serial IAM-CAM10010099.7699.69
Table 8. Performance comparison with existing methods.
Table 8. Performance comparison with existing methods.
MethodComponent
CoolerValvePumpAccMean
SVM99.8697.9396.6994.9097.35
Shapelet98.0590.1693.9590.0693.06
ETSC10010096.0084.4095.10
CLUST-
GWO-MLP
99.9499.9993.9989.8295.94
LDA99.8999.9799.4192.4897.94
MCIFM99.9899.9710097.6099.39
CNN10010098.9899.3599.58
DNN99.91100 99.7798.7099.59
Proposed method10010099.7699.6999.86
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, J.; Ding, H.; Li, N.; Sun, X.; Dong, X. Intelligent Fault Diagnosis of Hydraulic System Based on Multiscale One-Dimensional Convolutional Neural Networks with Multiattention Mechanism. Sensors 2024, 24, 7267. https://doi.org/10.3390/s24227267

AMA Style

Sun J, Ding H, Li N, Sun X, Dong X. Intelligent Fault Diagnosis of Hydraulic System Based on Multiscale One-Dimensional Convolutional Neural Networks with Multiattention Mechanism. Sensors. 2024; 24(22):7267. https://doi.org/10.3390/s24227267

Chicago/Turabian Style

Sun, Jiacheng, Hua Ding, Ning Li, Xiaochun Sun, and Xiaoxin Dong. 2024. "Intelligent Fault Diagnosis of Hydraulic System Based on Multiscale One-Dimensional Convolutional Neural Networks with Multiattention Mechanism" Sensors 24, no. 22: 7267. https://doi.org/10.3390/s24227267

APA Style

Sun, J., Ding, H., Li, N., Sun, X., & Dong, X. (2024). Intelligent Fault Diagnosis of Hydraulic System Based on Multiscale One-Dimensional Convolutional Neural Networks with Multiattention Mechanism. Sensors, 24(22), 7267. https://doi.org/10.3390/s24227267

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop