Next Article in Journal
Usefulness of OSM and BDOT10k Data for Developing Tactile Maps of Historic Parks
Previous Article in Journal
Comparison of Different Approaches of Machine Learning Methods with Conventional Approaches on Container Throughput Forecasting
Previous Article in Special Issue
Development of a Framework to Aid the Transition from Reactive to Proactive Maintenance Approaches to Enable Energy Reduction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Transformer Network for Ship HRRP Target Recognition

School of Electronic Engineering, Naval University of Engineering, Wuhan 430033, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 9728; https://doi.org/10.3390/app12199728
Submission received: 1 July 2022 / Revised: 19 September 2022 / Accepted: 25 September 2022 / Published: 27 September 2022

Abstract

:
The traditional High-Resolution Range Profile (HRRP) target recognition method has difficulty automatically extracting target deep features, and has low recognition accuracy under low training samples. To solve these problems, a ship recognition method is proposed based on the lightweight Transformer model. The model enhances the representation of key features by embedding Recurrent Neural Networks (RNN) into Transformer’s encoder. The Group Linear Transformations (GLTs) are introduced into Transformer to reduce the number of parameters in the model, and stable features are extracted through linear intergroup dimensional transformations. The adaptive gradient clipping algorithm is combined with the Stochastic Gradient Descent (SGD) optimizer to allow the gradient to change dynamically with the training process and to improve the training speed and generalization ability of the model. Experimental results on the simulated dataset show that multi-layer model stacking can effectively extract deep features of targets and raise recognition accuracy. At the same time, the lightweight Transformer model can maintain good recognition performance with low parameters and low training samples.

1. Introduction

HRRP is the vector sum of projections in the target direction after the wideband radar-emitting electromagnetic wave is scattered by the target. It contains target scattering characteristics, geometric structure and attitude, and compared to Synthetic Aperture Radar (SAR) images, HRRP has the advantages of easy acquisition, storage and fast processing [1]. As a result, target recognition using HRRP has gained widespread interest in the field of Radar Automatic Target Recognition (RATR) [2,3,4,5,6,7].
Discerning how to extract easily distinguishable deep features in HRRP is key to target recognition. Earlier, Du et al. [8] proposed a method to calculate the Euclidean distance in the higher-order spectral feature space to explore the most favourable higher-order spectral features for recognition. Zhang et al. [9] selected the bispectra with maximum interclass differentiability as a feature to avoid the effect of harmful bispectra features on recognition. Molchanov et al. [10] extracted the cepstral coefficients from the micro-Doppler signals of radar echoes and proposed a classification feature based on bicoherence estimation. However, these recognition methods require a priori knowledge of the relevant spectrum and obtaining high recognition rates relies on manual extraction of better classification features.
Since traditional recognition method requires the manual extraction of features and have a limited ability to extract feature representations, the researchers introduced deep learning to automatically extract target features. Wan et al. [11] and Guo et al. [12] applied ordinary Convolutional Neural Network (CNN) and lightweight CNN to the recognition of HRRP, respectively. Subsequently, Chen et al. [13] combined a bi-directional gated recurrent unit (Bi-GRU) with CNN and used Bi-GRU to enhance CNN to extract the valuable features for recognition, and the experimental results demonstrated that the use of CNN could effectively improve the recognition rate of ship targets.
All the above models ignore the fact that HRRP signals are temporal signals, and in order to make better use of the temporal dependence in HRRP, Liu et al. [14] applied recurrent neural networks (RNN) to HRRP recognition and achieved better results under small sample conditions. However, the RNN was also found to suffer from the problem of gradient vanishing. To solve this problem, Zhang et al. [15] proposed a VGM-RNN model to extrapolate missing samples in the sequence, which not only effectively alleviates the problem of gradient vanishing, but also has excellent noise immunity. The literature [16,17,18] proposes to combine the attention mechanism with RNN, Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRU). According to the importance of the implicit state features in target recognition, weights are assigned to the implicit state features encoded by RNN, LSTM and GRU to make the extracted features more discriminative. Du et al. [19] combined RNN with a deep clustering mechanism, using region decomposition and attention mechanism to accumulate each region weight, while optimizing the cross-entropy loss function to make the model more robust and recognition performance better. The literature [20] used attention modules and feedforward neural units to construct Transformer networks, and extract deep features by stacking attention modules and feedforward neural units [21,22].
This paper improves the encoding part of the Transformer model and proposes a lightweight Transformer model. The model is based on the Transformer encoders and introduces a local RNN module [23] at the input of each encoder to extract correlations between neighbouring distance units, and combines the global correlations extracted by the Transformer encoders to enhance the representation of feature. At the same time, intergroup dimensional transformations and feature interactions by GLTs [24] are used to fully extract the target features and reduce parameters in the model. The use of adaptive gradient clipping algorithm [25] in the SGD optimizer allows the gradient descent process to vary dynamically with model training, speeding up model training process and enhancing the stability of model.

2. Proposed Model

2.1. Lightweight Transformer Model

The encoder in the Transformer model is improved in this paper, and the structure of the three-layer lightweight Transformer model is shown in Figure 1. The lightweight Transformer model uses the pre-processed HRRP adjacent to k time steps as input through the local RNN module to extract target local temporal feature, and uses residual connection to alleviate the gradient vanishing problem during RNN training process. Combining GLTs with Transformer encoder: the number of model parameters is reduced by transforming the dimension of the input side of the self-attention unit and the feedforward neural network unit, and the features are fully extracted in different dimensions. During model optimization process, the SGD optimizer is combined with adaptive gradient clipping algorithm, so that the results of the training process can be fed back to the gradient descent process to speed up the training speed and enhance model stability.

2.2. Data Pre-Processing

The raw HRRP data of radar targets have aspect sensitivity, amplitude sensitivity, and time-shift sensitivity [26], which have a significant impact on model recognition performance. Therefore, before using the model to recognize, this paper applies Euclidean norm to the raw ship target HRRP to remove the amplitude sensitivity. Suppose the HRRP data after amplitude normalization is:
p = [ p 1 , p 2 , , p L ] T
where L is the number of distance units and pi is the normalized amplitude of the ith distance unit. When using HRRP data for target recognition, the time domain HRRP is converted into sequence data and then input into the model for target recognition.
We use the sliding window method proposed in the literature [27] to transform the HRRP data into serial data after amplitude normalization in the following steps. The amplitude normalized HRRP data is intercepted as a short sequence containing multiple range cells by means of sliding window.
X = [ x 1 , x 2 , , x T ]
x l = [ p ( l 1 ) c + 1 , p ( l 1 ) c + 2 , , p ( l 1 ) c + d ]
where x l d × 1 is the input at moment l with length d. The matrix X d × T corresponds to T time steps with data of dimension d at each time step.

2.3. Local RNN Module

The local RNN module aims at extracting local features and makes full use of the advantages of RNN in short sequence feature extraction. With a short sequence consisting of three time steps as a single RNN input, the structure of the local RNN module is shown in Figure 2. The white circles represent the zero sequence, the blue circles represent the original input sequence, and the light blue circles represent the output sequence. The white circles and blue circles form the input of the local RNN module.
Take the sequence containing k time steps in HRRP as the input of a single RNN as an example. The pre-processed HRRP sequence containing T time steps is X = [ x 1 , x 2 , , x T ] . A traditional RNN takes a sequence of T time steps and inputs them in time order to obtain a hidden state vector characterizing the target feature. Since RNN has the problem of gradient vanishing when processing long sequences, the local RNN module is used to alleviate the gradient vanishing problem while extracting local features from short sequences. The local RNN module is essentially extracting correlation features between neighbouring distance units within the HRRP sequences. The local RNN module first complements the left side of the input sequence X with k − 1 zero vectors according to the set short sequences size k to form a new input sequence X .
X = [ 0 , , k 1 0 , x 1 , x 2 , , x T ]
The sliding window of size k, sliding through the distance of 1 from left to right in turn, forms T short sequences, each containing k time steps in X . The T short sequences are fed into T RNNs, each of which outputs a hidden state vector characterizing the feature, and the T hidden state vectors are stitched into a new hidden state vector.
h t = R N N ( x t k + 1 , x t k + 2 , , x t )
H = [ h 1 , h 2 , , h T ]
Adding a local RNN module before Transformer’s encoder can make up for Transformer’s disadvantage in local feature extraction, and act as a positional encoding. The features extracted by the improved model contain more information and are more comprehensive and more generalizable.

2.4. Group Linear Transformations (GLTs)

For a model, a larger number of parameters means higher equipment requirements and longer training time. A comparison of the Transformer encoder and the Transformer encoder combined with GLTs is shown in Figure 3. Compared with the standard Transformer encoder structure, the Transformer encoder combined GLTs uses GLTs at the input to process the input data. In two stages of expansion and reduction, the input data are first mapped to a high-dimensional spatial expansion to obtain more easily distinguishable feature, and then the dimension is reduced as the input to the Self-Attention layer. This transformation allows for more efficient parameter assignment in the expansion and reduction stages, yielding better results than simple linear layer coding. This is because GLTs are able to learn a wider range of features in different dimensions. Under the condition of ensuring good recognition accuracy, Multi-head Attention can be replaced by Self-Attention, while the input dimension of Self-Attention is half of that of Multi-head Attention in the standard Transformer. So, the number of multiplication and addition operations required for calculation is reduced by half, and the corresponding computational effort is greatly reduced. The input dimension of the lightweight feedforward neural network in the Transformer encoder combined with GLTs is the same as that of the standard Transformer, but the dimension reduction process in the first layer reduces both the dimension and the size of the feedforward neural network to one-fourth of the original size, and the parameters become one-sixteenth of the original size.
The hidden vector H , which is output by the local RNN module, is used as the input to the Transformer encoder with the GLTs. Suppose the number of layers of GLTs is N. In the expansion stage, in N / 2 linear layers, the d m -dimensional input is projected into the high-dimensional space with d max = w m d m , where w m is expanded dimensional coefficients; similarly, in the reduction stage, the d max -dimensional vector is projected into the d 0 -dimensional space using the remaining N N / 2 layers, projecting the d max dimensional vector into the d 0 dimensional space.
g l = { min ( 2 l 1 , g max ) , 1 l N / 2 g N 1 , Otherwise
where g max is the set maximum number of linear transformation layers per layer. g l denotes the number of linear transformation layers contained in the lth layer.
Input the hidden vector H :
Y l = { F ( H , W l , b l , g l ) , l = 1 F ( M I X ( H , Y l 1 ) , W l , b l , g l ) , O t h e r w i s e
where W l = [ W 1 l , W 2 l , , W g l l ] and b l = [ b 1 l , b 2 l , , b g l l ] are the weights and bias of the g l group linear transformation function in the lth layer, respectively. F is a linear transformation function that divides the input H into g l non-overlapping groups so that H can be written as H = [ H 1 , H 2 , , H g l ] , and the linear layer transforms each H i by the weights W i l and b i l to obtain an output Y i l = H i W i l + b i l . The output Y i l of each group is then stitched to produce the final output Y l .
GLTs combine the output Y l 1 with the input H through the mixer MIX to mix the features and extract more stable features of the target, which can also avoid the problem of gradient vanishing.
M I X ( H , Y l 1 ) = s h u f f l e [ H , Y l 1 ]
Due to the superior performance of GLTs, Self-Attention is used instead of Multi-head Attention to obtain the same attention effect and achieve a better recognition accuracy using smaller dimensions and fewer operations to calculate the attention scores and carry out the assignment of weights:
u = s e l f _ A t t e n t i o n ( Y l ) = A t t e n t i o n ( Q , K , V ) = s o f t max ( Q K T d k ) V
where Q, K, and V are three key-value vectors, which are abbreviations of Query, Key, and Value vectors, respectively. d k is the dimension of the K vector, and Q and K are obtained by linear transformation of the output of the previous layer, and the final attention values are obtained by calculating the attention distribution between Q and K and attaching it to V.
The feedforward neural network layer contains two linear layers and a nonlinear activation function for feature transformation of the attention scores, where the nonlinear activation function is using GELU and the dimensionality of the linear layer is changed by first downscaling to one-fourth of the input dimension and then upscaling to the original dimension:
v = Feedforward ( u ) = ( ( u W 1 + b 1 ) × ϕ ( u W 1 + b 1 ) ) W 2 + b 2
where ϕ ( x ) is the cumulative distribution function of the normal distribution, W 1 , b 1 are the weights and biases of the linear layer of the reduced-dimensional part, and W 2 , b 2 are the weights and biases of the linear layer of the raised-dimensional part.
Supposing that the Transformer encoder module with GLTs has M layers, the output v M obtained after the last layer is decoded by a network containing a single linear layer, and finally normalized by the SoftMax function to obtain the final output.
y = L i n e a r ( v M )
y = s o f t max ( y )
where y denotes the probability vector that the sample belongs to each class of ship targets.

2.5. Adaptive Gradient Clipping

In this paper, an adaptive gradient clipping algorithm is used in the process of optimization. Prem Seetharaman et al. [25] proposed a simple method to adaptively select the gradient clipping threshold based on the numerical history of the gradient descent process observed during the training process, resulting in smoother loss function curves and more robust recognition performance.
Suppose the objective function is f ( X ; θ ) , where X is the input data and θ is a parameter, and given a learning rate λ , the parameter θ is updated iteratively from t − 1 to t with gradient descent as defined below.
θ t = θ t 1 λ h c θ f ( X , θ t 1 )
h c = min { η c θ f ( X , θ t 1 ) , 1 }
where η c is a clipping hyperparameter that is set according to the model. In clip by norm, the entire gradient is scaled if the norm of the gradient exceeds a threshold.
The function of the adaptive gradient clipping algorithm is to let η c follow the model training process dynamically by setting a new hyperparameter p, representing percentile cut-off for setting η c ( t ) to cut the gradient parametrization. In a training process, X t D , D represents the training dataset and t is one of the training batches. Supposing that G h stores the gradient norm of each batch and the initial G h ( 0 ) = [   ] .
g h ( t ) = θ f ( X t ; θ t 1 )
The gradient norm is added to G h for each batch computed, i.e., during the training of the tth batch, there are:
G h ( t ) = [ g h ( 0 ) , g h ( 1 ) , , g h ( t 1 ) , g h ( t ) ]  
Calculate the gradient norm of the pth percentile value in the current gradient vector.
η c ( t ) = p × G h ( t )
The gradient cropped θ t is obtained by bringing η c ( t ) into (15).

3. Experimental Results and Analysis

3.1. Introduction to the Dataset

Most of the ship targets are non-cooperative targets, and it is difficult to build an HRRP database of the targets from the actual measurement data. In this paper, we use CAD3D software to build ten kinds of 1:1 ship target models and import them into CST electromagnetic simulation software. The structural parameters of ten kinds of ship targets are shown in Table 1. CST simulation parameters are set as follows: azimuth angle is 0°–360°, pitch angle is 90°, angle step is 1°; radar centre frequency is 3 GHz, bandwidth is 150 MHz, polarization mode includes vertical polarization and horizontal polarization, frequency sampling points are 360, the default optimal grid profiling size of the software is used, and ray tracing algorithm is selected for the solution. Eventually, the HRRP data of 360 azimuths for ten types of ship targets are simulated.
The lack of training samples can lead to overfitting of the model, and the number of HRRP simulated with CST is far from sufficient for the proposed model, which needs to be expanded. In this paper, the original data are expanded by adding the clutter conforming to the K distribution in three times by the size of the signal-to-clutter ratio of 10 dB, and the data are expanded by a factor of three at each signal to clutter ratio. The HRRP of the ten-class ship target under the 10 dB signal-to-clutter ratio condition is shown in Figure 4. The HRRP of a class ten ship target under 10 dB signal-to-clutter ratio for both polarizations is shown in Figure 4.
The model parameters are set: the total number of training epochs is 300, the initial learning rate is 0.03, and the learning rate becomes one-fourth of the original one for every 80 iterations during the training process. The dimensions of the hidden layers are all set to 128. The optimizer is chosen to be Stochastic Gradient Descent (SGD) with a batch size of 32. The sliding window length of the preprocessing is d = 48 and the sliding distance l = 24. The kernel function used for Support Vector Machines (SVM) is the Radial Basis Function. The role of Principal component analysis (PCA) is to reduce the input data to two dimensions.

3.2. Recognition Performance

To explore the effect of the number of stacked layers on recognition performance, the number of layers one to eight are taken, and the relationship between recognition accuracy and the number of layers is shown in Figure 5. For recording purposes, we denote the local RNN module as LR and the adaptive gradient clipping algorithm as AC.
As can be seen in Figure 5, except for the LSTM model, the rest of the models are able to learn deeper features as the number of stacked layers increases, and the recognition performance of the models is improved. The Transformer model with the addition of the local RNN module extracts the correlation between neighbouring distance units of HRRP, which makes the feature representation more powerful and the model can achieve more than 88% accuracy even at a lower number of layers. With the addition of GLTs, more stable features are extracted by projecting the features in different dimensions. The model achieves more than 94% recognition accuracy at lower layers, which is much better than other models, and the model is no longer sensitive to layer changes. Compared with other models, the accuracy variation curve with the number of layers is smoother. To verify the best performance of the model, all models below use eight layers.
Table 2 lists the recognition accuracy of various models. As can be seen from the results in Table 2, the recognition accuracy achieved by the machine learning method SVM and the SVM method after PCA dimensionality reduction is not very satisfactory. In contrast, the recognition accuracy is higher by the method of building neural networks. Compared with the traditional RNN and LSTM models, which are not ideal for the recognition of ship target HRRP, the recognition accuracy of the Transformer model is greatly improved. Compared with traditional Recurrent Neural Networks, the attention module in Transformer can better learn the important features in the sequence data, allowing the model to adaptively search for feature regions important for recognition and amplify the significant feature regions that can distinguish different targets, instead of aimlessly matching all regions of targets. With the addition of the local RNN module, RNN is used to focus the correlation between neighbouring distance units to extract local features, which are complementary to Transformer for overall feature extraction, and the recognition accuracy is raised. With the addition of GLTs, the features extracted by the local RNN module are fed into the linear transformation of the N-layer to realize the mapping of the input vector from low-dimensional to high-dimensional, allowing more features to be represented. Then, the high-dimensional easily distinguishable feature is reduced to the input dimension of Self-Attention by linear transformation of N layers. At the same time, feature exchange between groups is performed in the linear transform group to achieve global feature learning, which enhances the data expression and effectively raises the recognition accuracy. The adaptive gradient clipping algorithm is used to optimize the gradient descent process so that the gradient can obtain the optimal solution within a reasonable range, and the recognition accuracy is raised by a small margin. Compared with Transformer, the recognition accuracy is raised by 1.61% with the addition of local RNN module, by 3.88% with the addition of LR and GLTs, and by 4.61% with the addition of LR, GLTs and AC.
Figure 6 shows the curves of the relationship between the model recognition accuracy and epoch. As the number of epochs increases, the recognition accuracy of each model subsequently raises. Compared with traditional RNN and LSTM models, Transformer has faster convergence and less fluctuation after the model reaches stability. Adding the local RNN module, the ability of the model to extract more expressive features is highlighted after 50 epochs, and the recognition accuracy is raised. On this basis, the advantage of extracting more easily distinguishable features is obvious with the addition of GLTs. With the addition of adaptive gradient clipping algorithm, the recognition accuracy raised significantly in the first few epochs, the convergence of the model was significantly faster, and the model had more stable performance.

3.3. Feature Visualization

To verify the separability of the features extracted by different models, the original HRRP data and the features extracted by the six different models were input to the two-dimensional ISOMAP for visualization, and the results of feature visualization are shown in Figure 7. The data points in ten colours in Figure 7 represent ten types of targets. In Figure 7a, the ten types of ship targets in the original HRRP data are basically overlapped together, and the separability is poor. In Figure 7b,c, the feature separability extracted by RNN and LSTM models is better than Figure 7a, and the spacing between target samples of different categories increases, but still, most of the features overlap and are still inseparable. Figure 7d,e shows significantly better feature extraction after Transformer and Transformer + LR models than Figure 7b,c, with better aggregation of the same category targets and increased spacing between samples of different category targets, making the features of each target separable. However, the distances of samples within the same category are greater than the distances of samples between different categories, resulting in fuzzy boundaries of different target features. In Figure 7f,g, the features extracted by Transformer + LR + GLTs and Transformer + LR + GLTs + AC models are separable, with clearer boundaries between classes and almost no overlap.

3.4. Effect of Training Set Size

A robust recognition model should be able to maintain good recognition performance even with a small number of training samples. To test the effect of the size of the training sample on the recognition rate, six size training sets were used in the experiments, namely 4320, 8640, 10,800, 12,960, 17,280 and 19,440, corresponding to 20%, 40%, 50%, 60%, 80% and 90% of the data set, respectively. The results of the six types of models on different training sets are shown in Table 3.
As can be seen from the table, with only 20% of the training samples, the local RNN module, GLTs combined with the Transformer model can achieve 81.83% accuracy, which is a 16% raise in accuracy compared to the Transformer. In the process of changing the training set from 90% to 20%, the accuracy of Transformer model decreased by 27.53%, while the local RNN module, GLTs combined with Transformer only decreased by 15.3%, indicating that the model is more stable in the process of changing the training set. Transformer combined with the local RNN module can raise the recognition accuracy by 1–3% when the training set is reduced compared to Transformer model. Therefore, in comparison, the GLTs can make the Transformer model more robust when the training set is reduced.

3.5. Model Complexity Analysis

The complexity of the model is intuitively reflected by the number of parameters and the amount of computation. The number of parameters and the amount of computation for six different model are shown in Table 4.
From Table 4, it can be seen that the Transformer model combined with the local RNN module and GLTs, is about the same as the RNN model in terms of parameters, compared to the Transformer model, which has only a quarter of parameters, and moreover, only a fifth of parameters of the Transformer model combined with the local RNN module. In terms of computation, the Transformer model combining local RNN modules and GLTs is about the same as the RNN and LSTM models, much smaller than the Transformer and the Transformer model combined with local RNN module, achieving the best recognition rate using fewer parameters and computation. Model complexity is proportional to the number of parameters, so combined with the local RNN module and GLTs, the lightweight Transformer model achieves a win-win situation between complexity and accuracy.

4. Conclusions

In this paper, a lightweight Transformer network is proposed for recognizing ship targets. Through experimental studies, the model is able to enhance the effect of feature extraction and improve the recognition accuracy of the model with a smaller number of parameters. Under the same experimental conditions, the number of parameters and computational effort are greatly reduced and the accuracy is improved by at least 3.8 percent compared to other models. In future research, we will extend the application of the model to speech recognition, heart sound classification, picture recognition and other field, and compare it with other classical deep learning models. At the same time, we will continue to improve the model, adding functions such as rejection of targets outside the database, alarming of abnormal targets, etc., to expand the application scope of the model.

Author Contributions

Conceptualization, Z.Y. and J.L.; methodology, Z.Y.; software, Z.Y.; validation, L.W. and J.L.; formal analysis, Z.Y.; investigation, Z.Y.; resources, J.L.; data curation, L.W.; writing—original draft preparation, Z.Y.; writing—review and editing, J.L.; visualization, L.W.; supervision, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 61501486.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are available upon reasonable request to the submitting author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Du, L.; Liu, H.; Wang, P. Noise robust radar HRRP target recognition based on multitask factor analysis with small training data size. IEEE Trans. Signal Process. 2012, 60, 3546–3559. [Google Scholar]
  2. Shi, L.; Wang, P.; Liu, H. Radar HRRP statistical recognition with local factor analysis by automatic Bayesian Ying-Yang harmony learning. IEEE Trans. Signal Process. 2010, 59, 610–617. [Google Scholar] [CrossRef]
  3. Du, L.; Wang, P.; Liu, H. Bayesian spatiotemporal multitask learning for radar HRRP target recognition. IEEE Trans. Signal Process. 2011, 59, 3182–3196. [Google Scholar] [CrossRef]
  4. Liu, H.; Feng, B.; Chen, B. Radar high-resolution range profiles target recognition based on stable dictionary learning. IET Radar Sonar Navig. 2016, 10, 228–237. [Google Scholar] [CrossRef]
  5. Feng, B.; Chen, B.; Liu, H. Radar HRRP target recognition with deep networks. Pattern Recognit. 2017, 61, 379–393. [Google Scholar] [CrossRef]
  6. Liao, K.; Si, J.; Zhu, F. Radar HRRP target recognition based on concatenated deep neural networks. IEEE Access 2018, 6, 29211–29218. [Google Scholar] [CrossRef]
  7. Guo, C.; Jian, T.; Xu, C.A. Radar HRRP target recognition based on deep multi-Scale 1D convolutional neural network. J. Electron. Inf. Technol. 2019, 41, 1302–1309. [Google Scholar]
  8. Du, L.; Liu, H.; Bao, Z. Radar HRRP target recognition based on higher order spectra. IEEE Trans. Signal Process. 2005, 53, 2359–2368. [Google Scholar]
  9. Zhang, X.D.; Shi, Y.; Bao, Z. A new feature vector using selected bispectra for signal classification with application in radar target recognition. IEEE Trans. Signal Process. 2001, 49, 1875–1885. [Google Scholar] [CrossRef]
  10. Molchanov, P.; Egiazarian, K.; Astola, J. Classification of aircraft using micro-Doppler bicoherence-based features. IEEE Trans. Aerosp. Electron. Syst. 2014, 50, 1455–1467. [Google Scholar] [CrossRef]
  11. Wan, J.; Chen, B.; Xu, B.; Liu, H.; Jin, L. Convolutional neural networks for radar HRRP target recognition and rejection. EURASIP J. Adv. Signal Process. 2019, 2019, 5. [Google Scholar] [CrossRef]
  12. Guo, C.; Wang, H.; Jian, T.; Het, Y.; Zhalng, X. Radar target recognition based on feature Pyramid fusion lightweight CNN. IEEE Access 2019, 7, 51140–51149. [Google Scholar] [CrossRef]
  13. Chen, J.; Du, L.; Guo, G.; Yin, L.; Weti, D. Target-attentional CNN for Radar Automatic Target Recognition with HRRP. Signal Process. 2022, 196, 108497. [Google Scholar] [CrossRef]
  14. Liu, J.; Chen, B.; Chen, W. Radar HRRP target recognition with target aware two-dimensional recurrent neural network. In Proceedings of the IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Dalian, China, 20–23 September 2019; pp. 1–5. [Google Scholar]
  15. Zhang, Y.; Xiao, F.; Qian, F. VGM-RNN: HRRP sequence extrapolation and recognition based on a novel optimized RNN. IEEE Access 2020, 8, 70071–70081. [Google Scholar] [CrossRef]
  16. Xu, B.; Chen, B.; Wan, J. Target-aware recurrent attentional network for radar HRRP target recognition. Signal Process. 2019, 155, 268–280. [Google Scholar] [CrossRef]
  17. Liu, J.; Chen, B.; Jie, X. Radar high-resolution range profile target recognition based on attention mechanism and bidirectional gated recurrent. J. Radars 2019, 8, 589–597. [Google Scholar]
  18. Zhang, Y.F.; Zhang, S.H.; Liu, Y.X. Radar HRRP sequence target recognition method of attention mechanism based stacked LSTM network. Syst. Eng. Electron. 2021, 43, 2775–2781. [Google Scholar]
  19. Du, C.; Tian, L.; Chen, B. Region-factorized recurrent attentional network with deep clustering for radar HRRP target recognition. Signal Process. 2021, 183, 108010. [Google Scholar] [CrossRef]
  20. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 8, 5598–6008. [Google Scholar]
  21. Hsu, W.N.; Bolte, B.; Tsai, Y.H. Hubert: Self-supervised speech representation learning by masked prediction of hidden units. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 3451–3460. [Google Scholar] [CrossRef]
  22. Li, X.; Cooper, S.A.; Tang, Y. Deep transformers with latent depth. Adv. Neural Inf. Process. Syst. 2020, 33, 1736–1746. [Google Scholar]
  23. Wang, Z.; Ma, Y.; Liu, Z. R-transformer: Recurrent neural network enhanced transformer. arXiv 2019, arXiv:1907.05572. [Google Scholar]
  24. Mehta, S.; Ghazvininejad, M.; Iyer, S. Delight: Deep and light-weight transformer. arXiv 2020, arXiv:2008.00623. [Google Scholar]
  25. Seetharaman, P.; Wichern, G.; Pardo, B. Autoclip: Adaptive gradient clipping for source separation networks. In Proceedings of the IEEE 30th International Workshop on Machine Learning for Signal Processing (MLSP), Espoo, Finland, 21–24 September 2020; pp. 1–25. [Google Scholar]
  26. Chen, W.; Chen, B.; Liu, Y.; Petng, X.; Fan, H.; Yu, F.; Liu, H. Bidirectional recurrent gamma belief network for HRRP target recognition. Signal Process. 2021, 188, 108213. [Google Scholar] [CrossRef]
  27. Xu, B.; Chen, B.; Liu, H. Attention-based recurrent neural network model for radar high-resolution range profile target recognition. J. Electron. Inf. Technol. 2016, 38, 2988–2995. [Google Scholar]
Figure 1. Lightweight Transformer model structure. The feature extraction module consists of a local RNN module and a Transformer encoder modified by GLTs. The classifier consists of a linear layer and a SoftMax function.
Figure 1. Lightweight Transformer model structure. The feature extraction module consists of a local RNN module and a Transformer encoder modified by GLTs. The classifier consists of a linear layer and a SoftMax function.
Applsci 12 09728 g001
Figure 2. Local RNN module.
Figure 2. Local RNN module.
Applsci 12 09728 g002
Figure 3. Comparison of Transformer encoder and Transformer encoder combined with GLTs. The left is the standard Transformer encoder. The right is the Transformer encoder combined with GLTs.
Figure 3. Comparison of Transformer encoder and Transformer encoder combined with GLTs. The left is the standard Transformer encoder. The right is the Transformer encoder combined with GLTs.
Applsci 12 09728 g003
Figure 4. HRRP of class ten ship targets. (aj) represent ships one–ten.
Figure 4. HRRP of class ten ship targets. (aj) represent ships one–ten.
Applsci 12 09728 g004
Figure 5. Relationship between the number of layers and the recognition accuracy.
Figure 5. Relationship between the number of layers and the recognition accuracy.
Applsci 12 09728 g005
Figure 6. Relationship between recognition accuracy and epoch.
Figure 6. Relationship between recognition accuracy and epoch.
Applsci 12 09728 g006
Figure 7. Two-dimensional ISOMAP projection. a-h are the two-dimensional visualization results of the extracted features of several models. (a) original data, (b) RNN, (c) LSTM, (d) Transformer, (e) R-Transformer, (f) R-DeLighT, (g) R-DeLighT + Autoclip.
Figure 7. Two-dimensional ISOMAP projection. a-h are the two-dimensional visualization results of the extracted features of several models. (a) original data, (b) RNN, (c) LSTM, (d) Transformer, (e) R-Transformer, (f) R-DeLighT, (g) R-DeLighT + Autoclip.
Applsci 12 09728 g007
Table 1. Structural parameters of ten types of ship targets.
Table 1. Structural parameters of ten types of ship targets.
Ship NumberLength (m)Width (m)
1182.824.1
2153.820.4
3162.921.4
499.615.2
5332.876.4
6337.277.2
717.13.6
8143.415.2
917.64.5
1016.34.7
Table 2. Recognition performance of six different models.
Table 2. Recognition performance of six different models.
ModelsAccuracy (%)
SVM46.52
PCA + SVM49.21
RNN77.69
LSTM64.26
Transformer93.25
Transformer + LR94.86
Transformer + LR + GLTs97.13
Transformer + LR + GLTs + AC97.86
Table 3. Accuracy with different training set sizes (%).
Table 3. Accuracy with different training set sizes (%).
The Proportion of Training Set20%40%50%60%80%90%
RNN50.0564.9468.1770.2575.5777.69
LSTM31.2943.7949.3254.2961.3464.26
Transformer65.7278.7183.1285.3391.7193.25
Transformer + LR69.0382.4385.8688.4092.6994.86
Transformer + LR + GLTs81.8387.4690.1092.6095.6797.13
Transformer + LR + GLTs + AC82.5387.9790.8893.2396.1397.86
Table 4. Model parameters and flops.
Table 4. Model parameters and flops.
ModelParams (K)Flops (G)Accuracy (%)
RNN255.27<0.0177.69
LSTM954.23<0.0164.26
Transformer1130.210.0193.25
Transformer + LR1531.670.0394.86
Transformer + LR + GLTs292.94<0.0197.13
Transformer + LR + GLTs + AC292.94<0.0197.86
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yue, Z.; Lu, J.; Wan, L. Lightweight Transformer Network for Ship HRRP Target Recognition. Appl. Sci. 2022, 12, 9728. https://doi.org/10.3390/app12199728

AMA Style

Yue Z, Lu J, Wan L. Lightweight Transformer Network for Ship HRRP Target Recognition. Applied Sciences. 2022; 12(19):9728. https://doi.org/10.3390/app12199728

Chicago/Turabian Style

Yue, Zhibin, Jianbin Lu, and Lu Wan. 2022. "Lightweight Transformer Network for Ship HRRP Target Recognition" Applied Sciences 12, no. 19: 9728. https://doi.org/10.3390/app12199728

APA Style

Yue, Z., Lu, J., & Wan, L. (2022). Lightweight Transformer Network for Ship HRRP Target Recognition. Applied Sciences, 12(19), 9728. https://doi.org/10.3390/app12199728

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop