Next Article in Journal
A CMOS W-Band Amplifier with Tunable Neutralization Using a Cross-Coupled MOS–varactor Pair
Next Article in Special Issue
915-MHz Continuous-Wave Doppler Radar Sensor for Detection of Vital Signs
Previous Article in Journal
Enabling Green Wireless Sensor Networks: Energy Efficient T-MAC Using Markov Chain Based Optimization
Previous Article in Special Issue
Multi-Sensor Optimization Scheduling for Target Tracking Based on PCRLB and a Novel Intercept Probability Factor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Multi-Input Bidirectional LSTM and HMM Based Approach for Target Recognition from Multi-Domain Radar Range Profiles

1
School of Electronic and Information Engineering, Beihang University, Beijing 100191, China
2
School of Computing, Edinburgh Napier University, Edinburgh EH10 5DT, Scotland, UK
3
Department of Informatics, University of Leicester, Leicester LE1 7RH, UK
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2019, 8(5), 535; https://doi.org/10.3390/electronics8050535
Submission received: 4 April 2019 / Revised: 4 May 2019 / Accepted: 8 May 2019 / Published: 13 May 2019
(This article belongs to the Special Issue Radar Sensor for Motion Sensing and Automobile)

Abstract

:
Radars, as active detection sensors, are known to play an important role in various intelligent devices. Target recognition based on high-resolution range profile (HRRP) is an important approach for radars to monitor interesting targets. Traditional recognition algorithms usually rely on a single feature, which makes it difficult to maintain the recognition performance. In this paper, 2-D sequence features from HRRP are extracted in various data domains such as time-frequency domain, time domain, and frequency domain. A novel target identification method is then proposed, by combining bidirectional Long Short-Term Memory (BLSTM) and a Hidden Markov Model (HMM), to learn these multi-domain sequence features. Specifically, we first extract multi-domain HRRP sequences. Next, a new multi-input BLSTM is proposed to learn these multi-domain HRRP sequences, which are then fed to a standard HMM classifier to learn multi-aspect features. Finally, the trained HMM is used to implement the recognition task. Extensive experiments are carried out on the publicly accessible, benchmark MSTAR database. Our proposed algorithm is shown to achieve an identification accuracy of over 91% with a lower false alarm rate and higher identification confidence, compared to several state-of-the-art techniques.

1. Introduction

Radar is an active detection sensor. It is not disturbed by natural light, fog, or rainy weather [1]. It can recognize targets of interest by transmitting and receiving electromagnetic waves. Therefore, radar automatic target recognition technology plays an important role in human–machine interaction designs of various intelligent devices such as autonomous driving [2,3,4], intelligent wheelchair [5,6], and so on.
High-resolution range profile (HRRP) is a one-dimensional signal that can represent the geometric shape of the target in the direction of the radar line of sight. Recognition of the target represented in HRRP is an important approach for radars to monitor targets of interest. Compared with two-dimensional imaging of radars [7,8,9], HRRP has the advantages of easy acquisition, processing, and storage; hence using HRRP to achieve radar target recognition is more suitable for human–machine interaction designs of intelligent machine. However, HRRP has shortcomings such as target aspect, translation and amplitude sensitivity [10,11]. Therefore, radar target recognition based on HRRP is a complex and nonlinear classification problem.
As one-dimensional signals, the target information provided by HRRP is very limited. In this paper, we are interested to construct a new classifier based on the limited information. In response to this problem, many researchers have conducted extensive research and proposed many landmark methods. These are usually based on: (1) the distribution features of HRRP; (2) the data domain of HRRP; (3) different types of HRRP.
The recognition algorithms based on HRRP distribution features use statistical theory as the main analysis tool to identify the target of interest. For example, it was reported that the HRRP signal has specific distributions under some conditions. These distributions included β distribution [12], Gaussian distribution [13], γ distribution [14], double distribution composite model [15] and so on. Subsequently, the target of interest can be identified by extracting statistical features such as center distance and invariant moment. In addition, some researchers attempted to construct statistical modeling of HRRP from the target. When the position of the radar observation is unchanged, the aspect changes of the target were considered to be a random non-stationary process, given that the angular changes of these aspects exceed a certain range. Otherwise, they can be approximated as a stationary process. These processes can be described by a Hidden Markov Model (HMM), which has been widely used in target recognition based on multi-aspect HRRP sequences [16,17,18]. The recognition algorithms based on HRRP data domain mainly explore their characteristics from different data domains of HRRP. These data domains include time, frequency, and time-frequency domain, and result in different classification outcomes. For example, Liao et al. [16] used the time-domain amplitude features of HRRP to train HMMs, and obtained 82% recognition accuracy in 3° aperture data, and 92% recognition accuracy in 6° aperture data. Albrecht et al. [17] used the frequency-domain power spectrum features derived from the time-domain HRRP to train HMMs. Due to the translation invariance property of the power spectrum features, 94% recognition accuracy was obtained in [17]. Zhang et al. [18] used time-frequency domain (T-F) features to train HMMs. With more abundant features, 95% recognition accuracy was obtained in [18]. The recognition algorithms based on multi-look HRRP are also constructed from the aspect angle of the target of interests. However, unlike the recognition algorithms based on HRRP distribution features, they mainly perform non-coherent superposition of multiple HRRPs within a certain aspect angle range of the target, to suppress noise and fuse multi-look aspect features to improve the recognition performance. There are many strategies for multi-look processing, which can generally be performed on the sample level [19,20,21] or the feature level [22,23]. Sample level multi-look processing generally uses the method of averaging multiple samples, while feature level multi-look processing first needs to extract features from individual samples separately, and then fuse these features. Compared with single-look algorithms, multi-look algorithms improve the recognition accuracy to a certain extent, in both sample and feature levels. For example, in [20], a 97% correct identification rate was achieved using the multi-look method, while only a 37% accuracy rate was obtained using the single-look method.
The above traditional recognition algorithms explore HRRP signals from different perspectives and have achieved promising success. However, these algorithms usually produce single features, which makes it difficult to further improve the recognition performance. In state-of-the-art recognition algorithms based on HRRP data domain, combining the advantages of several data domains has not been well investigated. On the other hand, in recognition algorithm based on multi-look HRRP, sample level multi-look processing makes it difficult control the number that needs to be averaged. If these quantities are too large, the processing time of the recognition algorithms will increase, and conversely, it is difficult to improve the recognition performance. Feature level multi-look processing has the potential to improve recognition performance, but the implementation of such algorithms is difficult.
“End-to-End” feature learning using deep learning provides a solution to the problems of feature extraction and fusion in conventional algorithms mentioned above. The essence is to use multi-layer neural networks to automatically extract basic features of the original data, or use multi-branch neural networks to automatically fuse multiple features [24]. In recent years, many researchers have proposed several deep learning models based on different applications, such as Deep Convolutional Neural Network (CNN) [25], Stacked Autoencoders (SAE) [26], Restricted Boltzmann machines(RBMs) [27], and Long Short-Term Memory (LSTM) Recurrent Neural Network [28]. CNN is considered one of the best models to solve the “perception” problem, but is mainly used to extract two-dimensional data features [29]. The SAE and RBMs models are unsupervised learning methods, which have been used for HRRP target recognition [30,31]. However, due to lack of prior knowledge, these unsupervised learning methods are unable to maintain high recognition accuracy when dealing with multi-class recognition problems. LSTM is an improved recurrent neural network. It solves the problem of gradient disappearance of traditional recurrent neural networks using long-span prior information. Thus, LSTM has unique advantages in processing sequential data [32]. At present, LSTM is mainly used in speech recognition [33] and natural language processing [34]. Some researchers [35,36] used HRRP as a one-dimensional signal, and applied LSTM to HRRP target recognition. In dealing with different types of targets, Jithesh et al. and Bin et al. [35,36] have achieved good classification performance. These algorithms provide insights into exploring the sequence feature of HRRP. Further, some researchers combined LSTM with CNN by a fully connected layer with sequence characteristics [37,38]. Such combined algorithms can, on the one hand, exploit advantages of both LSTM and CNN; and on the other hand, effectively and automatically fuse the features that are irrelevant in a physical sense, thereby greatly improving the recognition performance. These algorithms also provide ideas for exploring the feature fusion of different data domains in HRRP. Furthermore, in recent years, some researchers e.g., [39] have proposed a bidirectional LSTM (BLSTM) algorithm based on the LSTM. The current features of sequential data were not only dependent on the past information, but also related to future information. This context-dependent relationship can further improve the ability of LSTM to process sequential data.
In summary, the construction of an optimal classifier based on limited HRRP signals to correctly extract, learn, and fuse HRRP features for improving target recognition, is still an open question. In contrast to traditional recognition algorithms, we first extract 2-D sequence features from the formation process of HRRP, and then generate different data domains from these 2-D sequence features. Based on these sequence features, a novel target identification method is presented by combining bidirectional Long Short-Term Memory (BLSTM) and a Hidden Markov Model (HMM). The proposed algorithm first learns and fuses the sequence features of different data domains, and then combines these with the sequence features of the target multi-aspect, to achieve recognition. It consists of a shallow CNN, a multi-input BLSTM, and multiple HMMs. The shallow CNN is used for learning and dimensionality reduction of two-dimensional features in the time-frequency domain; the multi-input BLSTM is used for feature fusion of different data domains; and HMMs are used for target multi-aspect sequence feature learning and final target recognition.
The rest of this paper is organized as follows. Section 2 gives a brief review of the approaches involved. Section 3 describes the proposed method in detail. Section 4 presents a pipeline of transforming the real synthetic aperture radar (SAR) image to HRRP and then generating multi-domain HRRP sequences. Section 5 reports comparative experimental results for radar target recognition. Finally, concluding remarks are given in Section 6.

2. LSTM and BLSTM

LSTM is a special recurrent neural network model. Through a special gate structure, it can store and retrieve information for a long time. Figure 1a shows an LSTM storage module. It comprises three gate structures (input i n , forget f n , and output o n ), a memory unit controller cell, two input and output activation units, and three peepholes connections. The input and output gates are used to control the block input and output of the cell, and the forgetting gates are used to control the memory and forgetting state of the cell. The peephole is connected with status information before all the doors, to obtain Constant Error Carousel information [40] that allows the cell to record more sequential information. Finally, the block output information is recurrent, and connects to the block input and all other gates, which enables LSTM to model complex and long-term dynamic features, and solves the gradient disappearance problem caused by long sequences in traditional recurrent neural networks [41]. The forward mechanism of the LSTM can be expressed by the following equation [42]:
I n = h ( W I x n + R I O n 1 + b I ) i n = σ ( W i x n + R i O n 1 + p i c n 1 + b i ) f n = σ ( W f x n + R f O n 1 + p f c n 1 + b f ) c n = i n I n + f n c n 1 o n = σ ( W o x n + R o y n 1 + p o c n + b o ) O n = o n h ( c n )
where I n , i n , f n , c n , o n , and O n represent the block input, input gate, forget gate, memory cells, output gate, and block output respectively. n is the number of sequential data. x n is the input feature at the nth sequence. W is the weight matrix. R is the recurrent weight matrix. b is the bias vector. p is the peephole weight vector, and the subscript I, i, f, o respectively represent the block input, input gate, forget gate, and output gate. σ is the logistic sigmoid activation function. h is the hyperbolic tangent activation function, and ⊙ denotes point-wise product with the gate value.
The LSTM has a disadvantage that it can only get past information but not future information. Compared with the LSTM, BLSTM solves this problem. As shown in Figure 1b, there are two independent LSTM networks in a BLSTM module. These two LSTM networks have different directions, one is a forward LSTM, and the other is a reverse LSTM. The forward LSTM is mainly used to extract the future information of sequence data, while the reverse LSTM is mainly used to extract past information of the sequence data. Finally, their results are connected to the same output unit, and future and past features are fused to produce the output. In this way, the BLSTM is able to extract and fuse future and past features of the sequence data, which can be expressed by the following formulas [43]:
O n = Γ ( W I x n + W O O n 1 + b O ) O n = Γ ( W I x n + W O O n + 1 + b O ) y n = W y O n + W y O n + b y
where O n is the forward hidden sequence, O n is the backward hidden sequence, and Γ is implemented by Equation (1).
In this paper, the BLSTM is used to extract future and past features of the sequence data to implement learning of multi-domain HRRP sequence features. However, in high-dimensional sequence data, the BLSTM is computationally inefficient and has limited ability to extract features. Therefore, choice of appropriate feature extraction and dimensionality reduction algorithms, combined with the BLSTM, are key to solving this problem.

3. Proposed Methods

The proposed target recognition framework, termed MIBL-HMM, is illustrated in Figure 2, including four processing steps, specifically: generating a multi-domain HRRP sequence, reducing feature dimensionality using a shallow CNN, fusing multi-domain sequence features with the multi-input BLSTM, and determining the category of each target sample via a HMM classifier.

3.1. Multi-Domain HRRP Sequence Generation

The generation of a multi-domain HRRP sequence is divided into two steps: (1) extraction of the HRRP sequence; (2) Based on the HRRP sequence, multiple data domains are generated, termed the multi-domain HRRP sequence. Details are given next.
(1) Extraction of the HRRP sequence:
HRRP is a one-dimensional signal that reflects different scatters of the target in the direction of the line of sight radar. One HRRP represents the target’s response to one high range resolution radar pulse. In one HRRP, the amplitude of each range unit represents the intensity of the electromagnetic wave reflected by different scatters of the target. In a certain range of observation angles, the order of these intensities reflects the order of different scatters of the target in the direction of line of sight radar. Figure 3a shows one HRRP generated by a tank in response to one pulse of the radar. In this HRRP, these range units present the features of tank components in the order of the line of sight radar, specifically: first, the head of the tank; secondly, the body of the tank; and finally, the tail of the tank. Therefore, the range units contained in one HRRP have obvious sequence characteristics. This conclusion is consistent with [35,36]. In contrast to the literature, as shown in Figure 3b, we believe multiple HRRPs also possess timing sequence characteristics. This is because the electromagnetic pulses emitted by the radar have obvious timing sequence characteristics, so the HRRPs collected continuously also exhibit time-series characteristics. Subsequently, integrated with the sequence characteristics of a single HRRP, multiple HRRPs continuously acquired by the radar, will have obvious 2-D sequence characteristics. In this paper, we will extract HRRP sequences from these multiple HRRPs with 2-D sequence characteristics.
For a period of time, there is little difference, in theory, between multiple HRRPs continuously collected by the radar in a certain pose of the same target. However, due to various factors such as the environment, the difference between these HRRPs is relatively large. To alleviate these gaps and remove noise, the HRRPs collected during this period are arbitrarily divided into n groups which are not intersecting with each other. Next, each group is averaged by the multi-look processing strategy at the sample level. Under the condition that the original order is unchanged, a new set of n HRRPs is obtained, which is called a HRRP sequence. Such a HRRP sequence contains n elements, and the amount of information contained in each element is related to the average number of HRRPs. The larger the number, the more the amount of information. The information contained in each HRRP can be expressed by the radar aperture angle. For example, the radar aperture angle corresponding to each HRRP is 0.01 . When the average number of HRRPs is 10, the information contained in each element in the HRRP sequence is 0.1 aperture. It should be noted that this average number is limited for different purposes of HRRP. When HRRPs are used as training sets, the average number can take a large value, which is beneficial for the learning of target features. However, when HRRPs are used as test sets, this average number generally does not take a larger value. In this case, it is only necessary to select the corresponding number in each group of the split, to generate a test HRRP sequence. Currently, although each sequence of the training set and the test set contain the same elements, each element of the test set sequence contains less information than each element of the training set sequence. This is to save time of data processing in the recognition algorithm and speed up recognition speed.
(2) Generation of multi-domain HRRP sequence:
The various data domains considered in this paper include time domain, frequency domain, and time-frequency domain. Therefore, it is only necessary to generate these three data domains for each element of the HRRP sequence to obtain the multi-domain HRRP sequence. For the time domain and frequency domain features, the amplitude, value and power spectrum of the HRRP signal are used, respectively. For the time-frequency domain features, there are many extraction methods, such as short-time Fourier transform (STFT), Wigner-Ville distribution (WVD), and Adaptive Gaussian representation (AGR). STFT is the most commonly used time-frequency domain analysis method. It represents the signal characteristics at a certain moment, by a segment of the signal in the time window. However, the precision of time-frequency domain features obtained by the STFT method in time resolution and frequency resolution, is not available at the same time. The WVD method makes up for the shortcomings of the STFT method, but it leads to the problem of “cross-term interference”. Compared with these time-frequency domain feature extraction methods, the AGR method can decompose the echo signal into time-frequency center and local resonance of time-frequency domain resolution, and adjust the corresponding parameters using a Gaussian basis function to optimize the relationship between them. It can not only solve the “cross-term interference” problem faced by the WVD method, but also extract the adaptive spectrogram features effectively in the time-frequency plane. To-date, it has been successfully applied to ISAR imaging and radar target recognition [44,45]. Therefore, in this work, we use the AGR method to extract the time-frequency features of HRRP, as in [18].

3.2. Feature Dimensionality Reduction Using a Shallow CNN

A key step in radar target recognition based on time-frequency data is the dimensionality reduction of time-frequency features. Dimensionality reduction of data generally faces such dilemma: on the one hand, we hope to reduce the dimensionality of data as much as possible to avoid redundancy and complexity; on the other hand, we hope to retain as many features of data as possible to expand the learning space for extracting effective features. To alleviate this problem, this paper uses a shallow CNN to reduce the dimensionality of time-frequency features.
The shallow CNN comprises an input layer, two convolution layers, and a fully connected layer. In the input layer, the number of neurons is equal to the dimension of time-frequency features. In the fully connected layer, the number of neurons is equal to the dimension of the BLSTM’s neural network input layer. To train the weights between the input layer and the hidden layer, we add the s o f t m a x classifier on the fully connected layer. The number of the classifiers is equal to the number of target categories. The shallow CNN model can be expressed as follows:
f ( y i | X ) = s o f t m a x [ b ( 3 ) + W ( 3 ) ( f l a t t e n ( R ( b ( 2 ) + W ( 2 ) ( R ( b ( 1 ) + W ( 1 ) X ) ) ) ) ) ]
where X is the input feature vector of a time-frequency matrix, y i indicates class i, f ( · ) is the probability of X belonging to class i, s o f t m a x represents the softmax function, b ( k ) is the bias of the kth layer, W ( k ) represents weights between the k t h and (k + 1)th layers, ⊗ represents the convolution operation, f l a t t e n ( · ) is a function which connects the convolution layer and the fully connected layer by collapsing an array into one dimension, and R indicates the Rectified Linear Unit (ReLU) activation function.
In the training stage, the weights of the convolution layer and the fully connected layer are initialized with random values, and then continuously adjusted by the back-propagation algorithm, under the condition of decreasing the cross-entropy loss. The training process is repeated till the learning error falls below a moderate tolerance level. Then, the dimensionality-reduced features x can be obtained using the feed-forward CNN network:
x ( W ( 1 ) , X ) = R ( b ( 1 ) + W ( 1 ) X )
Unlike common dimensionality reduction methods (such as principal component analysis, linear discriminant analysis, etc.), the shallow CNN has obvious advantages. It can not only reduce the dimensionality of data, but also extract spatial features of data by connecting with the fully connected layer in a non-linear way, which effectively enhances the feature preservation of data after dimensionality reduction.

3.3. Multi-Domain Features Fusing with Multi-Input BLSTM

The aim of the multi-input BLSTM is to learn and fuse the features of multi-domain HRRP sequences. It consists of two parts:
(1) Multi-input, i.e., the algorithm has multiple input interfaces. These are divided into a main input and branch inputs. There is only one main input, which is located in the first layer, and there are many branch inputs, which are in subsequent layers other than the first layer. In terms of the importance level of features, the most important features are input from the main input interface, and secondary features are input from the branch input interface. The deeper the branch interface, the less important feature input is received. The design of the multi-input, on the one hand side, ensures the most important features are learned using the whole neural network, i.e., by extracting the most abstract and easily recognizable features; and on the other hand, in the process of learning the most important features, it gradually fuses other features in order to suppress the neural network over-fitting, and finally improves the recognition performance of the algorithm.
Figure 4 shows the multi-input BLSTM used in our algorithm. It contains three input interfaces: one main input and two branch inputs. The main input interface receives time-frequency sequence features after dimensionality reduction, the first branch interface receives the time-domain sequence features, and the second branch interface receives the frequency-domain features. The important levels of these three different data domain features will be discussed in the experimental section.
(2) Features combination and fusion. In the process of learning the main input features, the multi-input BLSTM needs to gradually fuse features. At this time, it will face the problem of combining and fusing the features of two different data domains. There are many ways to combine sequence data, such as element-wise addition, subtraction, multiplication, or concatenation. Here, we use element-wise addition, which is the addition of the corresponding elements, as shown in Figure 4. In this way, simple linear addition does not destroy the features of different data domain sequences; and, the value of each range bin in HRRP is small, and can be appropriately increased by the addition with highlighted features. After the combination, the sequence features will go into the BLSTM network to automate the fusion of different data domain sequence features. Since each BLSTM network consists of two recurrent neural networks, one is forward processing data, the other is the backward processing data, both of which are connected to the same output layer so that each BLSTM network has the ability to learn the past and future information from data. Therefore, in the process of sequence feature fusion, the multi-input BLSTM cannot only learn and fuse the sequence features of different data domains at the current time, but also fuse the past and future information of these sequence features.
In conclusion, compared with the standard multi-layer BLSTM, the multi-input BLSTM enhances the ability of the standard multi-layer BLSTM with multi-input functionality. Compared with traditional HRRP target recognition methods, the multi-input BLSTM can deeply learn rich features contained in the HRRP data. Here, we use a multi-input BLSTM network consisting of three layers: input, hidden, and output layer. The input layer is three inputs mentioned above, the hidden layer includes two BLSTM networks, and the output layer is a fully connected layer with sequence features. The multi-input BLSTM is trained in the same way as the shallow CNN. Finally, the multi-domain HRRP sequences go into the trained multi-input BLSTM to produce the fused feature sequences, which are sent to a HMM classifier to learn and classify the multi-aspect features.

3.4. Multi-Aspect Features Learning and Classification with HMM

Multi-domain HRPP sequences are fused by the multi-input BLSTM to obtain a sequence feature. Each sequence feature corresponds to an aspect angle of the target. Before entering the HMM classifier, these sequence features need to be processed as follows.
In the HMM training stage, these sequence features are connected in series to form a matrix. As shown in Figure 5, the horizontal direction represents the aspect angle and the vertical direction represents the sequence features. Here, we use this feature matrix to construct the HMM model, i.e., to establish the state of HMM and the mode of transition of these states, and then calculate the probability π and transition probability matrix A of the initial state of the corresponding HMM model. Since the details of the HMM-based multi-aspect feature classification algorithm used here have been described in several papers [16,17,18], we briefly summarize the basic concepts in this section.
(1) Establishing the state of the HMM model and the mode of these state transitions. Using the average aspect angle, the omni-directional angular space of the radar observation target, namely 360 degrees, is divided into L states, and the angular extent of each state is 360 / L degrees. In addition, to reduce the parameters of the HMM model, the initial transition mode of the HMM model is from left to right, as shown in Figure 5.
(2) Calculating the probability π and transition probability matrix A of the initial state of the HMM model. When the radar observation position is constant, let δ θ represent the range of the target aspect change when the radar continuously measures the target. Let θ i represent the range of the state i (angle change). Regarding the state transition of the HMM, the state i is subjected to δ θ < θ i . Assuming that A = { a i j } , a i j represents the probability of state i to the next state j, then the initial value of A is as follows:
a i , i 1 = a i , i + 1 = δ θ / 2 θ i , a i , i = ( θ i δ θ ) / θ i
Moreover, the initial orientation of the target is generally considered to be uniformly distributed, so the probability of the initial state can be calculated:
π i = θ i / i = 1 N θ i
As discussed above, Equations (5) and (6) constitute the initial estimates of A and π . These parameters can be better estimated by the Baum-Welch method. Through the above methods, the HMM model can be established and trained. Each type of target corresponds to an HMM model, i.e., the number of HMM models depend on the type of targets. In addition, it should be noted that this paper adopts the continuous HMM model. This is because when HRRP is vectorized to obtain discrete signals, HRRP will be distorted, which reduces the recognition performance of the HMM.
In the HMM test phase, the sequence features obtained by the multi-input BLSTM are directly go into the trained HMM model, and the likelihood estimates in the output of each HMM model are obtained. If the i-th HMM yields the largest likelihood, then we declare the sequence features are associated with the i-th target type. For example, if we obtain the following aspect sequence o = { o 1 , o 2 , o 3 , , o M } from the unknown target type T, then the probability of the observation sequence o is given by summing the joint probability P ( o | q , T ) P ( q | T ) over all possible state paths q .
P ( o | T ) = a l l q P ( o | q , T ) P ( q | T )
If the target type T i gives the maximum likelihood for the observation sequence o , i.e.,
P ( o | T i ) P ( o | T k ) , T k
Then we declare the sequence o belongs to the target type T i .
Compared with the traditional HMM-based HRRP target recognition algorithm, this paper builds the aspect sequence of HRRP based on the fusion of multi-domain sequence features, so the HMM can fully learn the features of HRRP to provide the recognition outcome.

4. Benchmark MSTAR Dataset

Due to the lack of open HRRP real data, we use the benchmark real data of moving and stationary targets for acquisition and recognition (MSTAR) SAR images published by the US Department of Defense to test the proposed algorithm. MSTAR is the standard database for evaluating SAR image target recognition algorithms. According to the header information provided with the database, the radar frequency of the collected data is 9.599000 GHz. We know that the radar target’s characteristic dimension is 1 m if we use the high frequency band of 30 MHz–300 MHz, and 0.1 m if using the high frequency band of 300 MHz–3 GHz, and about 0.03 m by using the high frequency band of 9.599000 GHz. Therefore, the HRRP signals generated from the MSTAR database contain sufficient information about vehicle targets such as tanks, artillery, or trucks, which can be used for target recognition research [44,45,46,47,48]. The MSTAR database includes 10 types of military vehicle targets: BMP2, BRT70, T72, BTR60, 2S1, BRDM2, D7, T62, ZIL131, and ZSU234. Among them, the BMP2 contains three different variants: BMP2_9563, BMP2_9566 and BMP2_c21; the T72 also contains three different variants: T72_132, T72_812 and T72_s7. Although these variants have the same design blueprint, they come from different manufacturers and still have some differences in color and shape. In addition to the military vehicle targets, MSTAR also includes a type of man-made target: SLICY. The SLICY is often used as the interference target to test the generalization ability of the proposed recognition algorithm. The optical images of these 11 types of targets are shown in Figure 6.
The MSTAR data is acquired by the X-band spotlight mode SAR. At a certain depression angle, the SAR circles the target several times, and collects images from the target multiple-aspect angles in the range of 360 . The aspect angle change intervals are not uniform. Even the same target is different at different depression angles, which will increase the recognition difficulty of the proposed algorithm. In addition, in the experiments, the observation data at 17 depression angles are generally used for training, while 15 are used for testing. Table 1 shows the acquisition of SAR images in the MSTAR database.

4.1. Inversing HRRP from the SAR Image

Figure 7 shows the basic process of inversing HRRP from the SAR image, which comprises three steps:
Firstly, we remove the zero padding and the Taylor window. The MSTAR images are formed by taking a 2-D inverse fast Fourier transform (IFFT) of the Taylor-windowed and zero-padded phase data on a rectangular grid. The 2-D FFT is undertaken at first. Then the transformed signal is shifted so that the low frequencies occur in the center. Figure 7 shows the resulting 2-D signal of an example of the MSTAR SAR images. A noticeable band of near-zero values appears at the border of the 2-D signal, as shown in Figure 7. This band is the zero-padding result, which must be removed from the border of the signal. Next, we remove the Taylor window based on its parameters (35 dB sidelobe suppression level).
Secondly, we remove the clutter. Once zero padding and the Taylor window have been removed, a 2-D inverse FFT is applied to produce a de-convolved and Nyquist-sampled SAR image. Then the target segmentation procedure is undertaken to achieve the clutter removal.
Finally, we extract the HRRP. Before extracting the HRRP, the segmented SAR image requires zero padding to restore the original image size. Then, the zero-padding SAR image is transformed using the FFT, and the HRRP is extracted. Table 1 shows the number of HRRP that can be extracted from each SAR image for each type of target.
In Table 1, the number of HRRPs extracted from different sized SAR images is different. However, since all SAR images have the same down-range and cross-range resolutions, i.e., Δ r d = Δ r c = 0.305 m , the radar aperture angle represented by HRRP generated from each SAR image is equal, i.e.,
Δ θ = c 2 f Δ r c 3.0
where f = 9.6 GHz, the center frequency of the radar waveform, and c is the speed of light.

4.2. Generating a Multi-Domain HRRP Sequence

According to the description in Section 3.1, the construction of the MSTAR multi-domain HRRP sequence goes through two steps.
First, the HRRP sequence is extracted. The HRRPs obtained from a slice of SAR image are taken as a set of HRRPs collected continuously by the radar over a period of time. Figure 8 shows the basic process of extracting a HRRP sequence from the BMP2 target. 101 HRRPs are divided into N groups. For ease of calculation, if 101 is not divisible by N, the remaining HRRP will be deleted. If N = 4 , an HRRP sequence containing four elements is generated. In the training set, each element is the average result of 25 HRRPs, which contains 0.75 aperture information. In the test set, if not averaged, then each element is selected from one of the original 25 HRRPs, which contains 0.03 aperture information. Alternatively, an appropriate averaging number can be selected. For example, we select the amount of information for each element to be 0.3 aperture, i.e., 10 of the original 25 HRRPs for each element. This ensures that each element of the test set contains a certain amount of information to improve the recognition accuracy. Each element of the test set in this paper takes 0.3 aperture.
Secondly, we generate multi-domain HRPP data. The data domains in this paper include: time-, frequency-, and time-frequency (T-F)-domains. The amplitude in the time domain and the power spectrum in the time-frequency domain are used as target features for identification. For the T-F domain, we adopt AGR to extract T-F features of HRRP. In Figure 8, if N = 4, the 4-element multi-domain HRPP sequence in these three domains will be generated respectively, as shown in Figure 9. For each subfigure, the top-left part is the power spectrum in the frequency domain, the top-right part is the amplitude in the time domain, the middle part is the 3-D T-F feature generated from the power spectrum and the amplitude, and the middle-lower part shows 2-D T-F features.

5. Experiments

To test the validity and generalization capability of the proposed algorithm, we designed a series of experiments under extended operating conditions (EOC) [49,50]. The so-called EOC is to test the algorithm according to various conditions in reality. These test conditions are different from those of the training algorithms. Specific EOC are set as follows:
(1) The test set is very different from the training set. Generally, the data of the 17 depression angle is selected as the training set, and the 15 depression angle is used as the test set. More stringently, one of the variants of BMP2 and T72 is used as the training set, while the other two variants are used as the test set.
(2) The aspect information of the target in the test set is assumed to be unknown and covers 360 aspect during the test. This hypothesis is meaningful in reality. First, it reduces dependence of the radar on other resources. For example, an auxiliary device such as a moving target indicator, a tracker, or the like is not required to acquire the azimuth information of the target. The azimuth coverage of the target is generally required to be 360 . However, this will increase the difficulty of the recognition algorithm, and will represent a significant challenge to our algorithm.
(3) The target in the test set is considered to be a non-cooperative target. This means that the target type that appears in the test set may not be present in the training set. This situation is inevitable in reality. Because of the incompleteness of the database, a new test sample may not find its similarity in the database. Therefore, the recognition algorithm needs evaluation criteria to determine whether the test sample belongs to the library (in-targets) or not (out-targets).
In this paper, three groups of experiments are set up under the above EOC: baseline experiments, validity verification experiments, and robustness evaluation experiments. These are described next.

5.1. Baseline Experiments

To verify the validity of the proposed multi-domain HRRP sequence features, we designed two sets of comparative experiments:
(1) B1 experiment: Comparing the validity of the HRRP random data and the sequence data in a single data domain;
(2) B2 experiment: Comparing the validity of the HRRP sequence data in the same data domain with different data domains.

5.1.1. B1 Experiment

To compare the validity of the HRRP random data and sequence data in a single data domain, we constructed two similar test models, namely single-domain random data model (SRDM) and single-domain sequential data model (SSDM). The design of these two models is shown in Figure 10. Figure 10a represents the SRDM model and Figure 10b shows the SSDM model.
In Figure 10a,b, the first column shows the number of layers of the model; the second column shows the name of the model and its structure; the third and fourth columns show which data domains are entered and configuration of the model in the current data domain. Overall, the two models have similar structures and configurations. They comprise five layers; the first four layers are for data preprocessing and the fifth layer is the classifier. C o n v , B L S T M , and F-C represent a convolutional layer, a bidirectional LSTM layer, and a fully connected layer, respectively. Since HRRP is one-dimensional data in the time and frequency domain, but two-dimensional data in time-frequency domain, C o n v has two configurations: 32 @ 3 represents one-dimensional convolution of convolution kernel 3, which results in 32 features; 32 @ 3 3 represents a two-dimensional convolution of the convolution kernel 3 3 , and also results in 32 features, after this convolution operation has been implemented. In the B L S T M configuration, 32 @ ( 1 , N , 64 ) means that 32 features are obtained after the processing of this layer, −1 indicates that the number of the input sequence samples is not limited, N denotes that each sequence sample contains N elements, 64 implies each element is a 64-dimensional vector. In these two models, the configuration of the first three layers is identical, while the configuration of the fourth layer is different. In the fourth layer, the SRDM model uses a convolution layer, while the SSDM model uses the B L S T M layer. This is because, unlike the SRDM model, the SSDM model needs to extract sequence features. After the first four layers are set up for preprocessing, the two models eventually use the same classifier for recognition. The SRDM and SSDM models are designed in this way to ensure their experimental results are comparable. In the test results of the SRDM model, the validity of HRRP in time domain, frequency domain, and time-frequency domain can be compared under random conditions. In the test results of the SSDM model, the validity of HRRP in these domains can be compared under ordered conditions. In the test results of these two models, the effectiveness of HRRP in the same data domain under random and sequential conditions can be compared.
From Table 1, the training and test sets containing the “out-targets” are selected to test the SRDM and SSDM models. In the training set, we choose 9 types of targets; in the test set, because of the non-cooperative characteristics of the targets, we choose 2 types as the “out-targets” and the remaining as “in-targets”. These target types are shown in Table 2. In Table 2, the targets except “Others” in the first line, are the targets of the training set, i.e., 9 types of “in-targets”. In addition, the targets in the first column are the targets of the test set, in total 10 types, including 2 types of “out-targets”. In addition, to better demonstrate the validity of the data, the HRRP data of these targets are respectively generated into four sets of data to train and test the two models, according to the description in Section 4. In these four sets of data, each sequence of the training set contains 5, 4, 2, and 1 elements, which correspond to the aperture angles of 0.6 , 0.75 , 1.5 and 3 respectively. The aperture angle of each element in the test set is 0.3 . The order of the HRRP sequence is ignored in the SRDM model. Under such conditions, the experimental results of the SRDM and SSDM models are shown in Figure 11a,b, respectively. The horizontal axis represents the aperture angle of the training set corresponding to different experiments, and the vertical axis represents the recognition rate of the model. Electronics 08 00535 i001, Electronics 08 00535 i002 and Electronics 08 00535 i003 represent the recognition rate curves in time, frequency, and time-frequency domain, respectively.
In terms of the data domain, Figure 11a,b show that the feature recognition results in the time-frequency domain are the best, followed by the time-domain features, and finally the frequency-domain features. The recognition rate of the sequence features shown in Figure 11b is higher than that of the corresponding random features shown in Figure 11a. This shows that the HRRP sequence proposed in this paper is effective, which can improve the recognition rate of the targets. Figure 11a shows that the target recognition rate increases with the increase of aperture angles. When the aperture angle is 3.0 , i.e., all the HRRPs extracted from each SAR image in the training set are processed by the multi-look scheme at the sample layer to obtain one HRRP, where the target recognition rate is the highest. This is consistent with other conclusions reported in the literature [51]. Unlike Figure 11a, the target recognition rates shown in Figure 11b do not increase with the increase in aperture angle. This shows that the improved target recognition rate is not caused by the increase in aperture angles, and is, in fact, due to the HRRP sequence proposed here.
In the above comparative experiments, the SSDM model trained with the HRRP sequence in the time-frequency domain with 0.75 aperture produces the highest recognition rate. Table 2 shows the recognition results in the form of a confusion matrix. Since there are the “out-targets" in the test set, we need to use reasonable evaluation criteria.
The introduction of the “Others” type in the first line of Table 2 is the strategy adopted in this paper. Given a test sample, we first determine whether or not it belongs to the “in-targets”. If so, we continue to determine which types it belongs to; if not, it is considered to belong to the “out-targets” and we classify it as the “Others” type. Ideally, the “in-targets” in the test set are correctly identified and the “out-targets” are correctly classified as the “Others”. However, in the actual test, the “in-targets” may be incorrectly recognized as the “Others” type, and the “out-targets” may be incorrectly recognized as the “in-targets”. Therefore, we need three evaluation criteria to evaluate the experimental results. The three criteria are defined as:
(1)
Correct recognition rate of “in-targets”:
P c i d = N c c N t N o
where N t is the total number of “in-targets” in the test set, N o is the total number of “in-targets” identified as “Others” type in the test set, and N c c is the total number of “in-targets” correctly identified in the test set.
(2)
Detection rate of “in-targets”:
P d = N t N o N t
(3)
False alarm rate of “out-targets”:
F A R = N int N n t
where N i n t is the total number of “out-targets” identified as “in-targets” in the test set, and N n t is the total number of “out-targets” in the test set.
In Table 2, we set P d to 0.9, which means that 10% of “in-targets” in the test set are judged as “out-targets” when they are less than a certain threshold. In this way, on the one hand, the reliability of identification of “in-targets” in the test set can be improved; and on the other hand, a reference threshold can be found to judge the “in-targets” as “out-targets” for subsequent experiments. Under such conditions, P c i d = 0.70 , F A R = 0.62 are obtained in Table 2, which is still a long way from the ideal results ( P c i d = 1 , F A R = 0 ). Therefore, in the next set of experiments, we explore the fusion of multi-domain HRRP sequence features to further improve the target recognition rate.

5.1.2. B2 Experiment

In Section 5.1.1, both the SRDM and SSDM models have only one input, so these two models belong to the single-look processing type. The experimental results shown in Section 5.1.1 show that the recognition rates of these two models are not satisfactory. Compared to the single-look processing model, the multi-look processing model uses multiple inputs to extract and fuse multi-view features of data, which can improve recognition performance. Normally, the input data of the multi-look processing model are from the same data domain. Here, we will build a multi-input BLSTM model, and then, on the one hand, we use the same data domain HRRP sequence data for testing; and on the other hand, we use different data domains’ HRRP sequence data for testing. The purpose is to compare the validity of sequence features in the same data domain and in different data domains.
Figure 12 shows the multi-input BLSTM model built in this section. The middle column represent the structure of this model, and the left and right sides are the corresponding configurations under different input data. The model comprises six layers, where the first five layers represent data preprocessing, and the sixth layer is the classifier, with use of s o f t m a x . Therefore, we use MIBL-softmax to represent the model. The MIBL-softmax model has three input interfaces: one main input interface and two branch input interfaces. The main input interface is at Layer 1, the first branch input interface is at Layer 4, and the second branch input interface is at Layer 5. When the input data is from the same data domain HRRP sequence, these data do not need to distinguish the importance level, and directly feed in the input interface of the MIBL-softmax model. When the input data is from different data domains, we send them to the primary and secondary input interfaces of the MIBL-softmax model, from large to small, according to the importance level.
Looking at the experimental results shown in Section 5.1.1, when the HRRP sequence is at 0.75 aperture, the time-frequency domain features are the best, followed by the time domain features, and finally the frequency domain features. Therefore, the HRRP sequence of 0.75 aperture in the time-frequency domain is used and sent to three input interfaces of the MIBL-softmax model. On the other hand, we use the time-frequency domain, time domain, and frequency domain HRRP sequence of 0.75 aperture as the data of the different data domains, and use the main input interface, the first branch input interface and the second branch input interface of the MIBL-softmax model respectively. The test results of these two sets of data in the MIBL-softmax model are represented by the confusion matrix as shown in Table 3 and Table 4.
Table 3 shows test results of the same data domain HRRP sequence in the MIBL-softmax model. The correct recognition rate P c i d of the MIBL-softmax model is 0.8205, which is significantly higher than P c i d = 0.70 shown in Table 2. Under the same threshold as shown in Section 5.1.1, the detection rate P d of the MIBL-softmax model reaches 0.9078, which is higher than P d = 0.900 shown in Table 2. Regarding the “out-targets”, the F A R of the MIBL-softmax model is reduced to 0.5857, which is significantly lower than F A R = 0.62 shown in Table 2. This shows that, compared to the single-look processing model, the multi-look processing model has a better confidence level and the recognition rate for “in-targets”, and a better distinction between “out-targets” and “in-targets”, for the traditional case of using the same data domain HRRP sequence as input. This is because the multi-look processing model extracts the features of the HRRP sequence several times and continuously fuses them to obtain the best discriminative features. This in turn, improves the recognition performance of the targets to some extent. However, due to the limitation of the HRRP sequence in the same data domain, the recognition performance of the MIBL-softmax model is difficult to be further improved.
Unlike Table 3, Table 4 shows the test results of different data domain HRRP sequence in the MIBL-softmax model. Compared to Table 3, the three evaluation criteria shown in Table 4 have been significantly improved. The correct recognition rate P c i d of the “in-targets” increases from 0.8205 to 0.8844 ; the detection rate P d of the “in-targets” increases from 0.9078 to 0.9126 ; and the false alarm rate F A R of the “out-targets” decreases from 0.5857 to 0.4397 . This indicates that in the same multi-look processing model, the different data domain HRRP sequence is more favorable for target recognition than the same data domain HRRP sequence. Specifically, the data with the highest importance level, i.e., the main input data, can be better extracted through the whole network; secondly, the data with the highest importance level are seen to be the non-linear transformation of input data from each branch, i.e., the time-frequency domain data is the non-linear transformation of the time-domain data and the frequency-domain data. When the multi-look processing model fuses the input features, then, firstly, the data of each branch can provide similar features to the system, which can enhance these features in the fusion process, and thus improve the target recognition rate. Secondly, they can also add the features of different data domains in the process of non-linear transformation, which makes the multi-look processing model exhibit better generalization ability and lower false alarm rate. Therefore, in the process of multi-look sequence recognition, the HRRP sequence features in different data domains are more effective than those in the same data domain.

5.2. Validity Verification Experiments

In the baseline experiment of Section 5.1, we validated the validity of the multi-domain HRRP sequence features. In this section, we will verify the validity of the proposed MIBL-HMM algorithm for multi-domain HRRP sequence feature learning. Here, we continue to use the 0.75 aperture multi-domain HRPP sequence shown in Table 4 as the training data, and set up two sets of comparative experiments:
(1) V1 experiment: the comparative experiment between the MIBL-HMM algorithm and the MIBL-softmax algorithm. For the case of the multi-input BLSTM algorithm, the MIBL-HMM algorithm uses the HMM as the classifier, while the MIBL-softmax algorithm uses the s o f t m a x as the classifier. To verify the effectiveness of the MIBL-HMM in further learning multi-aspect features contained in the multi-domain HRRP sequences, the MIBL-HMM and the MIBL-softmax are compared under the same design and configuration, except for the different classifiers.
(2) V2 experiment: comparisons between MIBL-HMM and MIL-HMM algorithms. The MIL-HMM algorithm is an algorithm in which the multi-input LSTM algorithm uses the HMM as a classifier. To validate the effectiveness of the MIBL-HMM in learning multi-domain HRPP sequence features, the MIBL-HMM and the MIL-HMM are compared under the same design and configuration, except for the different numbers of the directions of LSTM.
Before starting the comparison, we need to train the HMM models. The algorithms reviewed in Section 3.4 are first used to construct the HMM models. These are then trained using the forward and reverse HRRP aspect sequences. The forward sequence is the aspect angle of the HRRP sequence sorted from small to large, i.e., 0 1 ... 359 360 , while the reverse sequence is arranged in the opposite direction. These sequences contain all HRRP aspect information and state statistics, and are able train HMM models quickly. The training set used in Section 5.1 contains 9 types of targets. Therefore, 9 HMM models need to be trained, with each target type corresponding to one HMM model. In the training process, the Baum-Welch algorithm is used to estimate the parameter A and the state density function of the HMM, so that the HMM can better reflect the scattering characteristics of the targets.

5.2.1. V1 Experiment

In the test phase of the MIBL-HMM algorithm, the multi-domain fusion sequence features of the unknown targets are fed into the 9 trained HMM models, and the corresponding likelihood values are generated, respectively. If the maximum of these likelihood values is less than the threshold set shown in Section 5.1.1, the unknown target is identified as an “out-targets”. Otherwise, it is identified as “in-targets”, and its specific type is the target type represented by the HMM model which obtains the maximum likelihood value. Table 5 shows the recognition results of the MIBL-HMM algorithm. The correct recognition rate of the “in-targets” is seen to be P c i d = 0.9132 , the detection rate of the “in-targets” is P d = 0.9211 , and the false alarm rate of the “out-targets” is F A R = 0.3467 . For the same test set, compared with the test results of the MIBL-softmax algorithm ( P c i d = 0.8844 , P d = 0.9126 , F A R = 0.4397 , Table 4), the MIBL-HMM algorithm test results can be seen to have improved. The results show that the proposed method can further learn the multi-aspect features contained in the HRRP multi-domain sequence features to improve the recognition performance.
It should be noted that the HMM model is sensitive to the number of hidden states and the weight of the Gaussian probability density function. Changing these configurations will lead to additional complexity in the HMM model for sequential data processing. Figure 13 shows the recognition rate of the HMM model using the MIBL-HMM algorithm under different hidden state numbers S and weights w of the Gaussian probability density function. When S = 60 and w = 2, the MIBL-HMM algorithm produces the best recognition outcome. The HMM model in the above experiments is set up in the same way. Therefore, in the following experiments, the parameters of the HMM model are set to S = 60, w = 2.

5.2.2. V2 Experiment

Similar to the testing of the MIBL-HMM algorithm, Table 6 shows the recognition results of the MIL-HMM algorithm. The correct recognition rate of the “in-targets” is seen to be P c i d = 0.8501, and the detection rate of the “in-targets” is P d = 0.9132. The false alarm rate of the “out-targets” is F A R = 0.4525. Under the same test set, compared with the test results of the MIBL-HMM algorithm ( P c i d = 0.9132, P d = 0.9211, F A R = 0.3467, Table 5), the MIBL-HMM algorithm test results are significantly higher than the MIL-HMM algorithm. The results show that the proposed method can further learn the context features contained in the HRRP multi-domain sequence features to improve the recognition performance.
In addition, we compare the proposed algorithm with several state-of-the-art methods. In general, the recognition performance of the template matching method is used as the baseline for HRRP target recognition [52]. As can be seen from Table 5, the correct recognition rate of our algorithm is 91.32%, which is better than 90% of the template matching method presented in [52]. Moreover, unlike the template matching method, our algorithm does not require aspect information of the target.
The algorithm proposed in this paper is also compared with the multi-look processing methods mentioned in Section 1. [20] used a multi-look processing method to identify the same 9 types of targets as those used in this paper, and obtained 97% recognition accuracy under the 3 aperture test set. In the HRRP sparse feature-based target recognition, [16] also used the 3 aperture test set, and obtained 92% recognition accuracy. In time-frequency domain feature recognition, Zhang et al. [18] used the 6 aperture test set, and obtained 95.62% recognition accuracy. Although the recognition accuracy of these algorithms is relatively high, they are all obtained at the expense of computational time of the algorithms. To compare with these algorithms, our proposed algorithm is also tested on a 3 aperture test set, and the recognition accuracy is found to be 96.2%, as shown in Table 7. This recognition accuracy is higher than the results reported in literature, which demonstrates the superiority of this algorithm.

5.3. Robustness Evaluation Experiments

To verify the generalization ability of the proposed algorithm in multi-domain HRRP sequence feature recognition, we designed the following two sets of comparative experiments, with rigorous experimental conditions for robustness evaluation.
(1) R1 experiment: One of the variant types of BMP2 and T72 targets is used as the training set, while the other three variants are used as test sets. Under such conditions, the robustness of the recognition algorithms is evaluated and compared.
(2) R2 experiment: The training set is the same as the training set of the R1 experiment, while the targets of the test set are set to “out-targets”. Under such conditions, the robustness of the recognition algorithms is evaluated and compared.
In the experiments reported in Section 5.2, although our MIBL-HMM algorithm outperforms the MIBL-softmax algorithm on three evaluation criteria, the comparative recognition accuracy has little difference. We will further compare the two algorithms in robustness evaluation experiments described next.

5.3.1. R1 Experiment

The 0.75 aperture HRRP multi-domain sequences are used to train the MIBL-softmax and MIBL-HMM algorithms respectively. Comparative recognition results are shown in Table 8 and Table 9, respectively.
Under the same threshold conditions, as can be seen from Table 8 and Table 9, the correct recognition rate P c i d of MIBL-softmax and MIBL-HMM algorithms for the “in-targets” has reached 98%. This shows that both the MIBL-softmax and the MIBL-HMM algorithms can learn the common features of targets from 0.75 aperture multi-domain HRRP sequence to accurately identify the targets. It can also be seen by comparing the “in-targets” detection rate P d , the MIBL-HMM algorithm obtains an accuracy of 0.9597, while the MIBL-softmax algorithm achieves an accuracy of 0.9274. The detection rate of the MIBL-HMM algorithm for “in-targets” is thus significantly higher than that of the MIBL-softmax algorithm. This shows that the MIBL-HMM algorithm is more robust than the MIBL-softmax algorithm based on the same multi-domain HRRP sequence. This is attributed to the ability of the MIBL-HMM algorithm to further learn the multi-aspect features contained in the multi-domain HRRP sequence, to enhance the generalization ability of target recognition.

5.3.2. R2 Experiment

The training set of the R2 experiment is the same as that of the R1 experiment, while the targets of the test set are the seven other types of targets shown in Table 1. The latter are also used as “out-targets" in the R2 experiment. Further, the MIBL-softmax and the MIBL-HMM algorithms are still trained using the 0.75 aperture multi-domain HRRP sequence. Comparative recognition results are shown in Table 10 and Table 11, respectively.
Table 10 and Table 11 show both, the total false alarm rate of “out-targets” in the whole test set, and the false alarm rate of each type of target. Overall, the total false alarm rate of the MIBL-softmax algorithm is 0.288, while that of the MIBL-HMM algorithm is approximately 10% less, at 0.191. This shows that the MIBL-HMM algorithm performs better in rejecting the “out-targets” test samples. In addition, it can also be seen that the MIBL-HMM algorithm can enhance the generalization capability of target recognition by more deeply learning the multi-aspect features contained in the multi-domain HRRP sequence. The MIBL-HMM algorithm more effectively rejects the “out-targets” compared to the MIBL-softmax algorithm. In Table 11, the MIBL-HMM algorithm achieves an ideal false alarm rate ( F A R < 0.100 ) on the exclusion of three types of “out-targets”: D7, T62, and ZIL131, where D7 is a bulldozer, T62 is a main battle tank, and ZIL131 is a freight truck. In addition, it can be seen in Table 11, that the MIBL-HMM algorithm does not demonstrate much improvement in detecting false alarm rate of the BTR60 target, which can be attributed to the test samples of BTR60 comprising abnormal samples [53]. Therefore, it is reasonable to expect that the MIBL-HMM algorithm will not reduce the false alarm rate detection of BTR60.
In summary, we extract the sequence of HRRP and on this basis, generate multi-domain sequences which are effective for target recognition. At the same time, the results show our proposed algorithm can deeply learn and more effectively fuse features of HRRP sequences in multiple data domains.

6. Conclusions

The construction of an optimal classifier based on limited HRRP signals to correctly extract, learn, and fuse HRRP features, in order to improve the target recognition ability is an open research problem. In this paper, we propose to learn features contained in the HRRP data, by exploiting the acquisition process of HRRP to extract a HRRP sequence, and generate multi-domain sequence features in time-, frequency- and time-frequency domains. A novel algorithm for HRRP target recognition based on BLSTM and HMM is proposed, which, on the one hand, extracts and fuses the multi-domain HRRP sequence features effectively, using a multi-input approach with primary and secondary branches. On the other hand, our proposed algorithm improves recognition performance by combining multi-aspect sequence features contained in multi-domain HRRP sequences, with the standard HMM model.
In benchmark recognition tasks using the MSTAR database with 10 non-cooperative targets, our proposed algorithm achieves 91% correct recognition rate. Compared with other state-of-the-art methods, our approach exhibits enhanced recognition performance, lower false alarm rate, and higher confidence. In addition, our proposed algorithm is simple to implement and, in the future, can be explored for implementing real-time human–machine interaction designs.

Author Contributions

All the authors make contribution to this work. F.G. and T.H. proposed the idea and wrote the paper; J.W. conceived and designed the experiments; J.S. performed the experiments; H.Z. and A.H. co-wrote the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (61771027; 61071139; 61471019; 61171122; 61501011; 61671035), the Guangxi Science and Technology Project (Guike AB16380273), the Scientific Research Foundation of Guangxi Education Department (KY2015LX444), and the Scientific Research and Technology Development Project of Wuzhou, Guangxi, China (201402205). A. Hussain was supported by the UK Engineering and Physical Sciences Research Council (EPSRC: grant no. EP/M026981/1), and a Visiting Professorship at Taibah Valley, Taibah University (Madinah, Saudi Arabia). H. Zhou was supported by UK EPSRC under Grant EP/N011074/1, Royal Society-Newton Advanced Fellowship under Grant NA160342, and European Union’s Horizon 2020 research and innovation program under the Marie-Sklodowska-Curie grant agreement No 720325.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AGRadaptive Gaussian representation
B1 experimentthe 1st baseline experiment
B2 experimentthe 2nd baseline experiment
BLSTMbidirectional long short-term memory
CNNconvolutional neural network
Convconvolutional layer
F domainfrequency domain
F-Cfully connected layer
FARFalse alarm rate
HMMhidden Markov model
HRRPhigh-resolution range profile
LSTMlong short-term memory
MIBL-HMMthe algorithm based on the multi-input BLSTM and HMM
MIBL-softmaxthe algorithm based on the multi-input BLSTM and softmax
MIL-HMMthe algorithm based on the multi-input LSTM and HMM
MSTARmoving and stationary target acquisition and recognition database
RBMsrestricted Boltzmann machines
R1 experimentthe 1st robustness evaluation experiment
R2 experimentthe 2nd robustness evaluation experiment
SAEstacked autoencoders
SARsynthetic aperture radar
SRDMsingle-domain random data model
SSDMsingle-domain sequential data model
STFTshort-time Fourier transform
T domaintime domain
T-F domaintime-frequency domain
V1 experimentthe 1st validity verification experiment
V2 experimentthe 2nd validity verification experiment
WVDWigner-Ville distribution

References

  1. Golovachev, Y.; Etinger, A.; Pinhasi, G.; Pinhasi, Y. Millimeter Wave High Resolution Radar Accuracy in Fog Conditions—Theory and Experimental Verification. Sensors 2018, 18, 2148. [Google Scholar] [CrossRef]
  2. Baker, C.J.; Smith, G.E.; Balleri, A.; Holderied, M.; Griffiths, H.D. Biomimetic Echolocation With Application to Radar and Sonar Sensing. Proc. IEEE 2014, 102, 447–458. [Google Scholar] [CrossRef] [Green Version]
  3. Gruber, A.; Gadringer, M.; Schreiber, H.; Amschl, D.; Bosch, W.; Metzner, S.; Pflugl, H. Highly scalable radar target simulator for autonomous driving test beds. In Proceedings of the Radar Conference, Oklahoma City, OK, USA, 23–27 April 2018; pp. 147–150. [Google Scholar]
  4. Tan, Q.J.O.; Romero, R.A. Ground vehicle target signature identification with cognitive automotive radar using 24–25 and 76–77 GHz bands. IET Radar Sonar Navig. 2018, 12, 1448–1465. [Google Scholar] [CrossRef]
  5. Gurbuz, S.Z.; Soraghan, J.; Balleri, A.; Clemente, C. Micro-Doppler Based In-Home Aided and Unaided Walking Recognition with Multiple Radar and Sonar Systems. IET Radar Sonar Navig. 2017, 11, 107–115. [Google Scholar] [CrossRef]
  6. Saho, K.; Sakamoto, T.; Sato, T.; Inoue, K.; Fukuda, T. Pedestrian classification based on radial velocity features of UWB Doppler radar images. In Proceedings of the International Symposium on Antennas and Propagation, Nagoya, Japan, 29 October–2 November 2012; pp. 90–93. [Google Scholar]
  7. Yue, Z.; Gao, F.; Xiong, Q.; Wang, J.; Huang, T.; Yang, E.; Zhou, H. A Novel Semi-Supervised Convolutional Neural Network Method for Synthetic Aperture Radar Image Recognition. Cogn. Comput. 2019. [Google Scholar] [CrossRef]
  8. Gao, F.; Ma, F.; Wang, J.; Sun, J.; Yang, E.; Zhou, H. Visual Saliency Modeling for River Detection in High-Resolution SAR Imagery. IEEE Access 2018, 6, 1000–1014. [Google Scholar] [CrossRef] [Green Version]
  9. Fan, Z.; Yao, X.; Tang, H.; Qiang, Y.; Hu, Y.; Lei, B. Multiple Mode SAR Raw Data Simulation and Parallel Acceleration for Gaofen-3 Mission. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2018, 11, 2115–2126. [Google Scholar]
  10. Hudson, S.; Psaltis, D. Correlation filters for aircraft identification from radar range profiles. IEEE Trans. Aerosp. Electron. Syst. 2002, 29, 741–748. [Google Scholar] [CrossRef]
  11. Liao, X.; Bao, Z.; Xing, M. On the aspect sensitivity of high resolution range profiles and its reduction methods. In Proceedings of the Record of the IEEE 2000 International Radar Conference [Cat. No. 00CH37037], Alexandria, VA, USA, 12 May 2000; pp. 310–315. [Google Scholar]
  12. Beckman, D.; Frame, S. Comparison of features from SAR and GMTI imagery of ground targets. In Proceedings of the International Society for Optics and Photonics Algorithms for Synthetic Aperture Radar Imagery X, Orlando, FL, USA, 21 April 2003; Volume 5095, pp. 224–233. [Google Scholar]
  13. Jacobs, S.P.; O’Sullivan, J.A. Automatic target recognition using sequences of high resolution radar range-profiles. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 364–381. [Google Scholar] [CrossRef]
  14. Copsey, K.; Webb, A. Bayesian gamma mixture model approach to radar target recognition. IEEE Trans. Aerosp. Electron. Syst. 2003, 39, 1201–1217. [Google Scholar] [CrossRef]
  15. Lan, D.; Liu, H.; Zheng, B.; Zhang, J. A two-distribution compounded statistical model for Radar HRRP target recognition. IEEE Trans. Signal Process. 2006, 54, 2226–2238. [Google Scholar] [CrossRef]
  16. Liao, X.; Runkle, P.; Carin, L. Identification of ground targets from sequential high-range-resolution radar signatures. IEEE Trans. Aerosp. Electron. Syst. 2002, 38, 1230–1242. [Google Scholar] [CrossRef] [Green Version]
  17. Albrecht, T.W.; Gustafson, S.C. Hidden Markov models for classifying SAR target images. Proc. SPIE 2004, 5427, 302–308. [Google Scholar]
  18. Zhang, X.; Liu, Z.; Liu, S.; Li, G. Time-Frequency feature extraction of HRRP using AGR and NMF for SAR ATR. J. Electr. Comput. Eng. 2015, 2015, 36. [Google Scholar] [CrossRef]
  19. Williams, R.; Westerkamp, J.; Gross, D.; Palomino, A. Automatic target recognition of time critical moving targets using 1D high range resolution (HRR) radar. IEEE Aerosp. Electron. Syst. Mag. 2000, 15, 37–43. [Google Scholar] [CrossRef]
  20. Wong, S. Multi-look fusion identification: A paradigm shift from quality to quantity in data samples. Proc. Spie Int. Soc. Opt. Eng. 2009, 7335. [Google Scholar]
  21. Gross, D. A neural network ATR for high range resolution radar signature recognition of moving ground targets. In Proceedings of the IEEE Conference Record of the Thirty-Third Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 24–27 October 1999; Volume 2, pp. 1235–1239. [Google Scholar]
  22. Kahler, B.; Blasch, E. Robust multi-look HRR ATR investigation through decision-level fusion evaluation. In Proceedings of the 2008 IEEE 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; pp. 1–8. [Google Scholar]
  23. Kittler, J.; Hatef, M.; Duin, R.P.; Matas, J. On combining classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 226–239. [Google Scholar] [CrossRef] [Green Version]
  24. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef]
  25. Gao, F.; Huang, T.; Wang, J.; Sun, J.; Hussain, A.; Yang, E. Dual-Branch Deep Convolution Neural Network for Polarimetric SAR Image Classification. Appl. Sci. 2017, 7, 447. [Google Scholar] [CrossRef]
  26. Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.A. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
  27. Hinton, G.E. A Practical Guide to Training Restricted Boltzmann Machines. Momentum 2012, 9, 599–619. [Google Scholar]
  28. Tai, K.S.; Socher, R.; Manning, C.D. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks. Comput. Sci. 2015, 5, 36. [Google Scholar]
  29. Huang, Y.L.; Xu, B.B.; Ren, S.Y. Analysis and pinning control for passivity of coupled reaction-diffusion neural networks with nonlinear coupling. Neucom. 2018, 272, 334–342. [Google Scholar] [CrossRef]
  30. Zhao, F.; Liu, Y.; Huo, K.; Zhang, S.; Zhang, Z. Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine. Sensors 2018, 18, 173. [Google Scholar] [CrossRef] [PubMed]
  31. Peng, X.; Gao, X.; Zhang, Y.; Li, X. An Adaptive Feature Learning Model for Sequential Radar High Resolution Range Profile Recognition. Sensors 2017, 17, 1675. [Google Scholar] [CrossRef]
  32. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735. [Google Scholar] [CrossRef]
  33. Soltau, H.; Liao, H.; Sak, H. Neural speech recognizer: Acoustic-to-word LSTM model for large vocabulary speech recognition. arXiv 2016, arXiv:1610.09975. [Google Scholar]
  34. Wen, T.H.; Gasic, M.; Mrksic, N.; Su, P.H.; Vandyke, D.; Young, S. Semantically Conditioned LSTM-based Natural Language Generation for Spoken Dialogue Systems. Comput. Sci. 2015, 1711–1721. [Google Scholar] [CrossRef]
  35. Jithesh, V.; Sagayaraj, M.J.; Srinivasa, K.G. LSTM recurrent neural networks for high resolution range profile based radar target classification. In Proceedings of the International Conference on Computational Intelligence & Communication Technology, Ghaziabad, India, 9–10 February 2017; pp. 1–6. [Google Scholar]
  36. Bin, X.U.; Bo, C.; Liu, H.; Lin, J. Attention-based Recurrent Neural Network Model for Radar High-resolution Range Profile Target Recognition. J. Electron. Inf. Technol. 2016. [Google Scholar] [CrossRef]
  37. Xu, Z.; Li, S.; Deng, W. Learning temporal features using LSTM-CNN architecture for face anti-spoofing. In Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 3–6 November 2015; pp. 141–145. [Google Scholar]
  38. Ma, X.; Hovy, E. End-to-end sequence labeling via bi-directional lstm-cnns-crf. arXiv 2016, arXiv:1603.01354. [Google Scholar]
  39. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. Off. J. Int. Neural Netw. Soc. 2005, 18, 602–610. [Google Scholar] [CrossRef] [Green Version]
  40. Schmidhuber, J.; Gers, F.A. Recurrent Nets that Time and Count. In Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium, Como, Italy, 27 July 2007; Volume 03, p. 3189. [Google Scholar]
  41. Weninger, F.; Bergmann, J.; Schuller, B. Introducing CURRENNT: The Munich Open-Source CUDA Recurrent Neural Network Toolkit. J. Mach. Learn. Res. 2015, 16, 547–551. [Google Scholar]
  42. Liu, Q.; Zhou, F.; Hang, R.; Yuan, X. Bidirectional-Convolutional LSTM Based Spectral-Spatial Feature Learning for Hyperspectral Image Classification. Remote Sens. 2017, 9, 1330. [Google Scholar] [CrossRef]
  43. Zhang, F.; Hu, C.; Yin, Q.; Li, W.; Li, H.; Hong, W. SAR Target Recognition Using the Multi-aspect-aware Bidirectional LSTM Recurrent Neural Networks. arXiv 2017, arXiv:1707.09875. [Google Scholar]
  44. Chen, V.C.; Ling, H. Time-Frequency Transforms for Radar Imaging and Signal Analysis; Artech House: Norwood, MA, USA, 2002. [Google Scholar]
  45. Lui, H.S.; Persson, M.; Shuley, N.V. Joint time-frequency analysis of transient electromagnetic scattering from a subsurface target. IEEE Antennas Propag. Mag. 2012, 54, 109–130. [Google Scholar]
  46. Sullivan, R. Radar Foundations for Imaging and Advanced Concepts; Scitech Publishing: Boston, MA, USA, 2004. [Google Scholar]
  47. Lui, H.S.; Shuley, N. Joint time-frequency analysis on UWB radar signals. In Proceedings of the International Conference on Signal Processing and Communication Systems, Gold Coast, Australia, 17–19 December 2007. [Google Scholar]
  48. Lui, H.S.; Shuley, N.V. Evolutions of partial and global resonances in transient electromagnetic scattering. IEEE Antennas Wirel. Propag. Lett. 2008, 7, 435–439. [Google Scholar]
  49. Ross, T.D.; Velten, V.J.; Mossing, J.C. Standard SAR ATR evaluation experiments using the MSTAR public release data set. Proc. SPIE Int. Soc. Opt. Eng. 1998, 3370, 566–573. [Google Scholar]
  50. Mossing, J.C.; Ross, T.D. Evaluation of SAR ATR algorithm performance sensitivity to MSTAR extended operating conditions. Proc. SPIE Int. Soc. Opt. Eng. 1998, 3370, 13. [Google Scholar]
  51. Xing, M.; Bao, Z.; Pei, B. Properties of high-resolution range profiles. Opt. Eng. 2002, 41, 493–504. [Google Scholar] [CrossRef]
  52. Dudgeon, D.E.; Lacoss, R.T. An overview of automatic target recognition. Linc. Lab. J. 1993, 6, 3–9. [Google Scholar]
  53. Gao, F.; Huang, T.; Sun, J.; Wang, J.; Hussain, A.; Yang, E. A New Algorithm of SAR Image Target Recognition Based on Improved Deep Convolutional Neural Network. Cogn. Comput. 2018, 1–16. [Google Scholar] [CrossRef]
Figure 1. LSTM and BLSTM model.
Figure 1. LSTM and BLSTM model.
Electronics 08 00535 g001
Figure 2. A flowchart of the proposed HRRP ATR framework based on BLSTM and HMM.
Figure 2. A flowchart of the proposed HRRP ATR framework based on BLSTM and HMM.
Electronics 08 00535 g002
Figure 3. Extraction of the HRRP sequence.
Figure 3. Extraction of the HRRP sequence.
Electronics 08 00535 g003
Figure 4. The proposed multi-input BLSTM.
Figure 4. The proposed multi-input BLSTM.
Electronics 08 00535 g004
Figure 5. The basic process of HMM model construction.
Figure 5. The basic process of HMM model construction.
Electronics 08 00535 g005
Figure 6. Images of the MSTAR targets.
Figure 6. Images of the MSTAR targets.
Electronics 08 00535 g006
Figure 7. The procedure from a SAR image to HRRPs.
Figure 7. The procedure from a SAR image to HRRPs.
Electronics 08 00535 g007
Figure 8. Extracting the HRRP sequence.
Figure 8. Extracting the HRRP sequence.
Electronics 08 00535 g008
Figure 9. Multi-domain HRPP sequence.
Figure 9. Multi-domain HRPP sequence.
Electronics 08 00535 g009
Figure 10. Two similar single-domain HRRP test models.
Figure 10. Two similar single-domain HRRP test models.
Electronics 08 00535 g010
Figure 11. Test results of four sets of single-domain HRRP data in SRDM and SSDM models respectively.
Figure 11. Test results of four sets of single-domain HRRP data in SRDM and SSDM models respectively.
Electronics 08 00535 g011
Figure 12. MIBL-softmax model.
Figure 12. MIBL-softmax model.
Electronics 08 00535 g012
Figure 13. Different configuration of HMM model in MIBL-HMM algorithm obtains different recognition rates.
Figure 13. Different configuration of HMM model in MIBL-HMM algorithm obtains different recognition rates.
Electronics 08 00535 g013
Table 1. Target Types in MSTAR Dataset.
Table 1. Target Types in MSTAR Dataset.
Target TypesNo. of SAR ImagesPixel Size of
SAR Image
No. of HRRPs
Extracted per SAR Image
15 17
BMP2_9563195233128*128101
BMP2_9566196233128*128101
BMP2_c21196233128*128101
T72_132196232128*128101
T72_812195232128*128101
T72_s7191232128*128101
BTR70196233128*128101
BTR60195256128*128101
BRDM2274298128*129101
ZSU234274299158*158120
T62273299172*173136
ZIL131274299192*193152
2S1274299158*158120
D7274299177*178136
SLICY27429854*5448
Table 2. Test results of the SSDM model trained in time-frequency domain with 0.75 aperture.
Table 2. Test results of the SSDM model trained in time-frequency domain with 0.75 aperture.
Target TypesBMP2_c21T72_132BTR70SLICYBTR60BRDM2ZSU234T62ZIL131Others
BMP2_c2113853104927216
T72_13201338313105825
BTR701171151012683222
BTR6079211122582524
BRDM21471412161501217725
ZSU2343498124188111223
T62810561011141741223
ZIL13178691511101217422
The number of the “in-targets" HRRP sequences in the test set = 1878.
out-targets
2S127328891324312795
D7122010191232211718113
The number of the “out-targets” HRRP sequences in the test set = 548.
Correct recognition rate of the “ in-targets”, P c i d = 1194/1698 = 0.70;
Detection rate of the “in-targets”, P d = 1698/1878 = 0.90;
False alarm rate of the “out-targets”, F A R = 340/548 = 0.62.
Table 3. Test results of the same data domain HRRP sequence in the MIBL-softmax model.
Table 3. Test results of the same data domain HRRP sequence in the MIBL-softmax model.
Target TypesBMP2_c21T72_132BTR70SLICYBTR60BRDM2ZSU234T62ZIL131Others
BMP2_c211553132432221
T72_1324148282343220
BTR705015574421018
BTR602436150350616
BRDM2787610182512631
ZSU2347464562104622
T625637836211420
ZIL131056979101518825
The number of “in-targets” HRRP sequences in the test set = 1878.
out-targets
2S1161520431514151114111
D7131121481713141110116
The number of the “out-targets” HRRP sequences in the test set = 548.
Correct recognition rate of the “ in-targets”, P c i d = 1399/1705 = 0.8205;
Detection rate of the “in-targets”, P d = 1705/1878 = 0.9078;
False alarm rate of the “out-targets”, F A R = 321/548 = 0.5857.
Table 4. Test results of different data domain HRRP sequence in the MIBL-softmax model.
Table 4. Test results of different data domain HRRP sequence in the MIBL-softmax model.
Target TypesBMP2_c21T72_132BTR70SLICYBTR60BRDM2ZSU234T62ZIL131Others
BMP2_c211621041332119
T72_1326152231252518
BTR701214836257121
BTR6021010160121216
BRDM23158322027322
ZSU2349025752132724
T620202114234524
ZIL1310073207822720
The number of “in-targets” HRRP sequences in the test set = 1878.
out-targets
2S1711143451111107164
D712101127122071715143
The number of the “out-targets” HRRP sequences in the test set = 548.
Correct recognition rate of the “in-targets”, P c i d = 1516/1714 = 0.8844;
Detection rate of the “in-targets”, P d = 1714/1878 = 0.9126;
False alarm rate of the “out-targets”, F A R = 241/548 = 0.4397.
Table 5. Test results of the MIBL-HMM algorithm trained with 0.75 aperture multi-domain HRRP sequence.
Table 5. Test results of the MIBL-HMM algorithm trained with 0.75 aperture multi-domain HRRP sequence.
Target TypesBMP2_c21T72_132BTR70SLICYBTR60BRDM2ZSU234T62ZIL131Others
BMP2_c211554330256315
T72_1321161324212020
BTR700217012001119
BTR601200172210017
BRDM20326323235317
ZSU2342423512305319
T6212435282131322
ZIL1310000000824719
The number of “in-targets” HRRP sequences in the test set = 1878.
out-targets
2S1581155681011205
D7510647213815204
The number of “out-targets” HRRP sequences in the test set = 548.
Correct recognition rate of the “in-targets”, P c i d = 1580/1730 = 0.9132;
Detection rate of the “in-targets”, P d = 1730/1878 = 0.9211;
False alarm rate of the “out-targets”, F A R = 190/548 = 0.3467.
Table 6. Test results of the MIL-HMM algorithm trained with 0.75 aperture multi-domain HRRP sequence.
Table 6. Test results of the MIL-HMM algorithm trained with 0.75 aperture multi-domain HRRP sequence.
Target TypesBMP2_c21T72_132BTR70SLICYBTR60BRDM2ZSU234T62ZIL131Others
BMP2_c21115821011956822
T72_1324121016351051121
BTR700017210330116
BTR600120167012022
BRDM256716519157230
ZSU234073456193141032
T622013010252212
ZIL131223640202478
The number of “in-targets” HRRP sequences in the test set = 1878.
out-targets
2S1693711111510129154
D7151240591211420146
The number of the “out-targets” HRRP sequences in the test set = 548.
Correct recognition rate of the “ in-targets”, P c i d = 1458/1715 = 0.8501;
Detection rate of the “in-targets”, P d = 1715/1878 = 0.9132;
False alarm rate of the “out-targets", F A R = 248/548 = 0.4525.
Table 7. Test results of MIBL-HMM testing with 3 aperture multi-domain HRRP sequence.
Table 7. Test results of MIBL-HMM testing with 3 aperture multi-domain HRRP sequence.
Target TypesBMP2_c21T72_132BTR70SLICYBTR60BRDM2ZSU234T62ZIL131Others
BMP2_c211650110250319
T72_1320174320212012
BTR700217512001114
BTR601200172212015
BRDM20020325201313
ZSU2342023012500313
T621201021247316
ZIL1310000000025915
Total number of test input sequences = 1878
confuser-targets(confusers)
2S1581155681011205
D7510647213815204
Total number of test input sequences = 548
Overall correct identification rate, P c i d = 1694/1761 = 0.962;
Overall “in-target" detection rate, P d = 1761/1878 = 0.939;
Overall false alarm rate, F A R = 139/548 = 0.2536
Table 8. Test results of the MIBL-softmax algorithm under the condition that test set targets are “in-targets”.
Table 8. Test results of the MIBL-softmax algorithm under the condition that test set targets are “in-targets”.
Target TypesBMP2_c21BTR70T72 _132Others
BMP2_95631752315
BMP2_95661822012
BMP2_c211850011
BTR700180016
T72_1320018610
T72_8122517018
T72_s71017317
The number of “in-targets” HRRP sequences in the test set = 1365.
Correct recognition rate of the “in-targets”, P c i d = 1251/1266 = 0.9881;
Detection rate of the “in-targets”, P d = 1266/1365 = 0.9274.
Table 9. Test results of the MIBL-HMM algorithm under the condition that the test set targets are “in-targets”.
Table 9. Test results of the MIBL-HMM algorithm under the condition that the test set targets are “in-targets”.
Target TypesBMP2_c21BTR70T72_132Others
BMP2_9563185127
BMP2_9566186019
BMP2_c21190006
BTR700185011
T72_132221848
T72_812421809
T72_s7101855
The number of the “in-targets” HRRP sequences in the test set = 1365.
Correct recognition rate of the “in-targets”, P c i d = 1295/1310 = 0.9885;
Detection rate of the “in-targets”, P d = 1310/1365 = 0.9597.
Table 10. Test results of the MIBL-softmax algorithm under the condition that test set targets are “out-targets”.
Table 10. Test results of the MIBL-softmax algorithm under the condition that test set targets are “out-targets”.
Target TypesBMP2_c21BTR70T72_132OthersFAR
BRDM2149010160114/274 = 0.416
BTR6010150530165/195 = 0.846
D713191123143/274 = 0.157
T6214151822647/273 = 0.172
2S121122022153/274 = 0.193
ZIL13116172821361/274 = 0.222
ZSU23413181522846/274 = 0.168
The number of “out-targets” HRRP sequences in the test set = 1838.
Overall false alarm rate of the “out-targets”, FAR = 529/1838 = 0.288.
Table 11. Test results of the MIBL-HMM algorithm under the condition that test set targets are “out-targets”.
Table 11. Test results of the MIBL-HMM algorithm under the condition that test set targets are “out-targets”.
Target TypesBMP2_c21BTR70T72_132OthersFAR
BRDM21250820470/274 = 0.255
BTR607145538157/195 = 0.805
D7751125123/274 = 0.084
T6284625518/273 = 0.066
2S111121024133/274 = 0.120
ZIL13147825519/274 = 0.069
ZSU234781524430/274 = 0.109
The number of the “out-targets” HRRP sequences in the test set = 1838.
Overall false alarm rate of the “out-targets”, FAR = 350/1838 = 0.191.

Share and Cite

MDPI and ACS Style

Gao, F.; Huang, T.; Wang, J.; Sun, J.; Hussain, A.; Zhou, H. A Novel Multi-Input Bidirectional LSTM and HMM Based Approach for Target Recognition from Multi-Domain Radar Range Profiles. Electronics 2019, 8, 535. https://doi.org/10.3390/electronics8050535

AMA Style

Gao F, Huang T, Wang J, Sun J, Hussain A, Zhou H. A Novel Multi-Input Bidirectional LSTM and HMM Based Approach for Target Recognition from Multi-Domain Radar Range Profiles. Electronics. 2019; 8(5):535. https://doi.org/10.3390/electronics8050535

Chicago/Turabian Style

Gao, Fei, Teng Huang, Jun Wang, Jinping Sun, Amir Hussain, and Huiyu Zhou. 2019. "A Novel Multi-Input Bidirectional LSTM and HMM Based Approach for Target Recognition from Multi-Domain Radar Range Profiles" Electronics 8, no. 5: 535. https://doi.org/10.3390/electronics8050535

APA Style

Gao, F., Huang, T., Wang, J., Sun, J., Hussain, A., & Zhou, H. (2019). A Novel Multi-Input Bidirectional LSTM and HMM Based Approach for Target Recognition from Multi-Domain Radar Range Profiles. Electronics, 8(5), 535. https://doi.org/10.3390/electronics8050535

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop