Next Article in Journal
Estimating Scattering Coefficient in a Large and Complex Terrain through Multifactor Association
Next Article in Special Issue
Radar-Jamming Decision-Making Based on Improved Q-Learning and FPGA Hardware Implementation
Previous Article in Journal
Current Status of the Community Sensor Model Standard for the Generation of Planetary Digital Terrain Models
Previous Article in Special Issue
Multi-Mission Oriented Joint Optimization of Task Assignment and Flight Path Planning for Heterogeneous UAV Cluster
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radar High-Resolution Range Profile Rejection Based on Deep Multi-Modal Support Vector Data Description

1
National Key Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, China
2
Hangzhou Institute of Technology, Xidian University, Hangzhou 311200, China
3
Institute of Information Sensing, Xidian University, Xi’an 710071, China
4
Shanghai Aerospace Electronic Technology Institute, Shanghai 201109, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(4), 649; https://doi.org/10.3390/rs16040649
Submission received: 10 December 2023 / Revised: 2 February 2024 / Accepted: 7 February 2024 / Published: 9 February 2024
(This article belongs to the Special Issue Advances in Remote Sensing, Radar Techniques, and Their Applications)

Abstract

:
Radar Automatic Target Recognition (RATR) based on high-resolution range profile (HRRP) has received intensive attention in recent years. In practice, RATR usually needs not only to recognize in-library samples but also to reject out-of-library samples. However, most rejection methods lack a specific and accurate description of the underlying distribution of HRRP, which limits the effectiveness of the rejection task. Therefore, this paper proposes a novel rejection method for HRRP, named Deep Multi-modal Support Vector Data Description (DMMSVDD). On the one hand, it forms a more compact rejection boundary with the Gaussian mixture model in consideration of the high-dimensional and multi-modal structure of HRRP. On the other hand, it captures the global temporal information and channel-dependent information with a dual attention module to gain more discriminative structured features, which are optimized jointly with the rejection boundary. In addition, a semi-supervised extension is proposed to refine the boundary with available out-of-library samples. Experimental results based on measured data show that the proposed methods demonstrate significant improvement in the HRRP rejection performance.

Graphical Abstract

1. Introduction

Nowadays, radar realizes all-weather observation without relying on ambient radiation and is widely applied in remote sensing, such as ship detection and instance segmentation [1,2] and target tracking [3,4,5]. Radar Automatic Target Recognition (RATR) came into being to satisfy the growing demands of military surveillance, civil monitoring and environmental assessments. RATR is usually implemented based on high-resolution range profiles (HRRP) [6], synthetic aperture radar images [7,8,9], and inverse synthetic aperture radar images [10]. HRRP is more accessible and reflects physical structure information, such as the size of the target and the distribution of scattering points, which receives extensive attention in RATR [6]. The construction of a database containing complete target classes is the basis for RATR. However, it is incapable of guaranteeing the completeness of the HRRP database due to the difficulty of obtaining HRRPs of non-cooperative or even hostile targets in advance, like aircraft that mistakenly intrude and invade airspace. In such a situation, a false alarm or missed alarm will emerge when the target to be identified comes from an unknown class. Therefore, rejecting these targets before recognition, called out-of-library target rejection, is one of the most important yet difficult issues in the practical application of the RATR system.
The rejection task has the following unique problem complexities: (1) missing out-of-library samples are associated with unknown spatial distribution; and (2) the data structure of various in-library samples is complicated. The aforementioned problems give rise to several challenges in designing reasonable rejection models.
Conventional rejection models, such as One-Class Support Vector Machine (OC-SVM, [11]), Support Vector Data Description (SVDD, [12]), and Isolated Forest (IF, [13]), have been applied in the field of RATR. These models have relatively simple shallow structures and demonstrate stable rejection performance. However, the conventional rejection model can no longer accomplish the rejection task well due to the high-dimension and multi-modality characteristics of HRRP.
Deep learning has shown tremendous capabilities in learning expressive representations of such complexly distributed datasets, so deep rejection methods have been advancing to further satisfy practical requirements. Initially, well-trained neural networks are only used to extract features from high-dimensional samples, and then these features are fed into conventional rejection models [14,15]. Such hybrid methods exploit the powerful nonlinear mapping capability of deep networks and reduce the information loss in feature extraction. Nevertheless, they mostly ignore the unique and multi-modal distribution of HRRP. The optimization criteria for feature extraction and rejection boundary learning are usually inconsistent in two-phase procedures, resulting in extracting sub-optimal features for rejection [16]. Hence, many studies explore methods that can unify feature extraction with rejection models. The mainstream is divided into three categories, i.e., deep rejection methods based on reconstruction, self-supervised learning, and one-class classification. The first category has been proposed in the literature [17,18,19,20], which is inspired by the Auto-encoder (AE) and the Generative Adversarial Network. Such methods assume that in-library samples can be reconstructed well, so samples with larger reconstruction errors are rejected as out-of-library samples. Therefore, they usually regard reconstruction error as the rejection criterion and train a well-behaved generator of in-library samples. However, these methods are often irrespective of common features between in-library samples and out-of-library samples that benefit the reconstruction of out-of-library samples. Thus, they may be sub-optimal [16]. And they are weak in the utilization of multi-modal information of HRRP. Recently, self-supervised learning has been proven to be effective in feature extraction, and deep methods based on self-supervised learning have emerged endlessly [21,22,23]. Nevertheless, similar to the first category, most of them leave the latent overlap and multi-modality out of consideration. In addition, they usually have requirements for data types and are more suitable for image data [16]. Deep rejection methods based on one-class classification achieve an integrated design of feature extraction and rejection, such as Deep Support Vector Data Description (DSVDD, [24]). The core idea is similar to the traditional SVDD, with the difference that neural networks instead of kernel functions are leveraged to map in-library samples into a hyper-sphere in the feature space. And it outperforms conventional rejection methods and deep hybrid rejection methods. However, it assumes that all of the in-library samples follow the same distribution, which is unsuitable for describing the multi-modal structure of HRRP [25]. Therefore, Ghafoori et al. [25] embedded in-library samples into multiple hyper-spheres, but this still cannot effectively portray the anisotropic structure of HRRP and has limited performance on rejection.
To address the above problems, this paper proposes a novel rejection method for HRRP, named Deep Multi-modal Support Vector Data Description (DMMSVDD). It mainly contains a data preprocessor, a feature extractor with a dual attention module, and a rejector with a more compact rejection boundary. The data preprocessor aims to address the sensitivity problems specific to HRRP. The feature extractor with the dual attention module obtains distinguishing features with global temporal information and channel-dependent information. Based on the high-dimensional and multi-modal structure of HRRP, the rejector is competent to adaptively fit the complicated underlying distribution of in-library samples with an adjustable Gaussian mixture model (GMM). In the geometric sense, multiple closed hyper-ellipsoids are substituted for hyper-spheres in the existing SVDD-like methods design, which is conducive to constructing a tighter and more explicit rejection boundary. In the proposed method, the feature extraction and rejection boundary learning are jointly optimized under the unified distance-based rejection criterion. It can achieve a high matching between features and the rejection boundary and provide the foundation for improving rejection performance. Moreover, to deal with the situation in which limited out-of-library samples are available, a semi-supervised method is extended to tighten the rejection boundary with those out-of-library samples.
The contributions of this paper are summarized as follows:
  • An efficient rejection method for HRRP is proposed, which jointly optimizes feature extraction and rejection boundary learning under a unified distance-based criterion;
  • The dual attention module in the feature extractor is capable of capturing the global and local structure information of HRRP, which further strengthens the feature discrimination between the in-library and out-of-library samples.
  • Considering the high-dimensional and multi-modal structure of HRRP, a more compact and explicit rejection boundary is formed with an adjustable GMM;
  • A semi-supervised method is extended to take advantage of available out-of-library samples to assist rejection boundary learning;
  • Experiments demonstrate that the proposed methods can significantly promote the rejection performance on the measured HRRP dataset.
The rest of this paper is organized as follows: Section 2 describes the major related rejection methods; Section 3 provides the framework of the proposed methods; the experimental results are presented and analyzed in Section 4; and finally, Section 5 summarizes this paper.

2. Related Work

For the sake of notational uniformity,  x n X ( n = 1 , 2 , , N )  with  X R d  denotes an in-library sample, where  N  denotes the total number of in-library samples. This section provides a brief introduction to several typical rejection methods.

2.1. Support Vector Data Description

Inspired by the support vector classifier, Tax and Duin propose SVDD [12], in which in-library samples are mapped into a hyper-sphere by a kernel function in the feature space, whereas the projected out-of-library samples are located outside the hyper-sphere. The compactness of the hyper-sphere is guaranteed by minimizing the volume of the hyper-sphere, and the cost function of SVDD is as follows:
min R , c , ξ R 2 + 1 v N n = 1 N ξ n s . t . φ x n ; ο c 2 R 2 + ξ n ξ n 0 , n = 1 , 2 , N
where  R  and  c  are the radius and center of the hyper-sphere in the feature space, respectively,  v  is a penalty factor to balance the sample classification error with the complexity of the algorithm,  ξ n  is a relaxation factor, and  φ · ; ο  is a kernel function with a parameter  ο .

2.2. Deep Support Vector Data Description

DSVDD, proposed by Ruff et al. [24], employs neural networks to extract features instead of kernel functions in SVDD. The powerful expressiveness of neural networks breaks through the limitations of traditional kernel mapping and improves the differentiation between in-library and out-of-library samples in low-dimensional space. Similar to SVDD, it minimizes the volume of the hyper-sphere in the feature space so that the features of the in-library samples mapped by the neural network are contained inside the hyper-sphere as much as possible, while the features of the out-of-library samples are as far away from the hyper-sphere as possible, and the loss function of Deep SVDD is as follows.
min R , W R 2 + 1 v N n = 1 N max 0 , ϕ x n ; W c 2 R 2 + λ 2 l = 1 L W l F 2
where  ϕ · ; W  is a neural network with parameters  W λ  is a regularization parameter,  L  is the number of network layers, and  W l  is l-th layer network parameters.
The training samples for the rejection task are mostly in-library samples, so this problem is ordinarily regarded as a one-class classification. Consequently, Ruff et al. [24] further simplify the above model by directly minimizing the mean distance between the features of training samples and the center, which is equivalent to minimizing the volume of the hyper-sphere.
min W 1 N n = 1 N ϕ x n ; W c 2 + λ 2 l = 1 L W l F 2

2.3. Deep Multi-Sphere Support Vector Data Description

Deep Multi-sphere Support Vector Data Description (DMSVDD) proposed by Ghafoori et al. [25] further improves the rejection performance by mapping in-library samples from different distributions to multiple hyper-spheres in the feature space.
min W , R 1 K k = 1 K R k 2 + 1 v N n = 1 N max 0 , ϕ x n ; W c i 2 R i 2 + λ 2 l = 1 L W l F 2
where  K  denotes the number of hyper-spheres in the feature space, and  R k  and  c k  denote the radius and center of the k-th hyper-sphere, respectively.  i = argmin k ϕ x n ; W c k 2 .

3. The Proposed Method

Considering the complex distribution of radar HRRP with high-dimension and multi-modality structure characteristics, this paper proposes a specific HRRP rejection method. As shown in Figure 1, the framework of the proposed method is mainly divided into a data preprocessor, a feature extractor, and a rejector. Firstly, data preprocessing alleviates the sensitivity problems of HRRP. Secondly, the convolution module and the dual attention module jointly realize the feature extraction of HRRP. The convolution module aims to extract the local spatial features of HRRP envelopes, and the dual attention module further obtains the global temporal and channel-dependent information ignored by the convolution module. Finally, the rejector fits the complex distribution of various in-library samples with the GMM. In-library samples are projected into multiple hyper-ellipsoids in the feature space, and out-of-library samples fall outside hyper-ellipsoids; therefore, a more delicate and closed rejection boundary is formed.

3.1. Data Preprocessor

Taking two HRRP samples as an example, the pre-processing procedure is shown in Figure 2. In particular, it mainly includes normalization and alignment, which alleviate the amplitude-scale and time-shift sensitivity of HRRP, respectively.
In practice, the received HRRP data contain amplitude information and phase information of targets, but the phase is greatly affected by the range variation. Hence, the phase of the target echo is usually ignored, and only the amplitude is preserved. Unfortunately, the amplitude is influenced by the radar antenna gain, target distance, target size, and other factors [26]. In order to alleviate the amplitude-scale sensitivity of HRRP, the L-2 normalization is usually used to erase the amplitude difference among samples and only retain their shape information. For an HRRP  x r , r = 1 , 2 , , R , where  x r  denotes the echo signal amplitude in the r-th range cell and  R  denotes the number of range cells, its normalized result can be expressed as:
x r = x r u = 1 R x u , r = 1 , 2 , , R
An HRRP is usually obtained by sliding the range window and intercepting part of the echo containing the target signal with a certain margin. As a result, the support area of the moving target may produce different degrees of shifting within the window. The center of gravity alignment is utilized to impair the impact of shifting support area on rejection [26]. Specifically, circular translation is operated to move the center of HRRPs to the center of gravity, i.e.,
G = r = 1 R r x r 2 / r = 1 R x r 2

3.2. Feature Extractor

The feature extraction network is divided into two convolution modules and a dual attention module. The convolution module can well reflect the local spatial information of targets [27]. A convolution module consists of several blocks, and one block includes a convolution layer, a batch-norm layer, a Leaky ReLU activation layer, and a pooling layer.
However, Ristea N C et al. [23] showed that stacked convolution layers aggregate low-level local features into high-level semantic features without comprehending the global arrangement. Specifically, given features from different channels, the convolution operation ignores their inter-dependencies and directly concatenates them. Given features from one channel, the convolution operation cannot acquire long-range dependency among all range cells of HRRP within the limitation of the local receptive field size. Attention modules have been widely applied in radar to solve the above-mentioned problems [28,29,30,31]. In this paper, the dual attention module [32] is added between convolution modules to further grasp the long-range and channel dependencies, respectively.
As illustrated in Figure 3, a dual attention module consists of a position attention module and a channel attention module.
The channel attention module adaptively recalibrates channel-wise features by a self-attention mechanism. Specifically, given a local feature  A R C × L , where  C  denotes convolution channels and  L  denotes the feature dimensions, inter-dependencies among channels, called the channel attention map  S , are calculated by applying a softmax layer to the matrix multiplication result of the transpose of  A  and itself, i.e.,
s i j = exp a i · a j c = 1 C exp a c · a j
where  s i j  denotes the impact of the  i -th channel on the  j -th channel, and  a i  denotes the feature of the  i -th channel in the local feature  A . Then, the feature  A  and the channel attention map  S  are matrix multiplied with a scale factor  α , and finally, the result is element-wise added with the local feature  A  to obtain the feature map  E 1 , i.e.,
E j 1 = α · i = 1 C s i j a i + a j
where  E j 1  denotes the feature map of the j-th channel obtained by the channel attention module.
Differently, the position attention module first feeds  A  into three convolutional layers to get three new feature maps  B R C × L C R C × L , and  D R C × L . Then, the position attention map  P R L × L  is computed by the feature map  B  and  C , i.e.,
p i j = exp b i · c j l = 1 L exp b l · c j
where  p i j  denotes the impact of the i-th dimensional feature on the j-th dimensional feature,  b i  denotes the i-th dimensional feature of the feature map  B , and  c j  denotes the j-th dimensional feature of the feature map  C . The feature map  E 2  is calculated as
E j 2 = β · i = 1 L p i j d i + a j
where  E j 2  denotes the feature map of the j-th dimension obtained by the position attention module, and  d i  denotes the i-th dimensional feature of the feature map  D .
The final feature map  E  is a fusion of the  E 1  from the channel attention module and  E 2  from the position attention module. The dual attention module enables the support area of HRRP with plentiful target information to gain more attention and strengthens the feature discrimination. Moreover, the dynamic capture induces the affection of the shifting support area within HRRP on the rejection task.

3.3. Rejector

The key to rejecting out-of-library samples is to form an accurate distribution of in-library samples. The study [33] manifests that each range cell of HRRP is non-Gaussian and statistically correlated with others in high-dimensional space, which results in anisotropy and multi-modality characteristics of various in-library samples. Hence, it is unreasonable to curtly push them to obey one or several standardized Gaussian distributions. It is known that the GMM is competent at depicting the underlying distribution with the appropriate number of Gaussian components selected [34]. As shown in Figure 4, the left figure briefly depicts the probability distribution of HRRP, and the GMM with three components describes the distribution of in-library samples more accurately. In the geometric sense, the GMM fitting the in-library sample distribution means multiple closed hyper-ellipsoids wrapping in-library samples inside while excluding out-of-library samples outside. Compared with the other SVDD-like methods, the proposed method can better adapt to the non-uniformity and discontinuity of the in-library sample distribution. Note that a sample is considered to affect only one Gaussian component for efficiency. As for details of the structure selection and parameter estimation of the GMM, please refer to Section 3.5.

3.4. Objective Function

As mentioned above, we hope that the GMM can fit the distribution of in-library samples as closely as possible, which is interpreted in geometric terms to push the in-library samples into their nearest hyper-ellipsoids in the feature space as tightly as possible, so that out-of-library samples that deviate from this distribution are excluded from hyper-ellipsoids. Consequently, the goal of modeling the distributions turns to minimize the volume of every hyper-ellipsoid. Thus, the objective function is defined as:
min θ , μ k , Σ k , K 1 N n = 1 N min k = 1 , 2 , , K ϕ x n ; θ μ k T Σ k   1 ϕ x n ; θ μ k + λ 2 p = 1 P θ p F 2
where  ϕ (   ; θ )  denotes the feature extractor with parameters  θ K  denotes the number of Gaussian components,  μ k  and  Σ k  denote the mean and covariance matrix of the k-th Gaussian component, respectively,  λ  denotes a non-negative hyperparameter that controls the trade-off between compact loss and regularization,  P  denotes the number of layers of the feature extractor, and  θ p  denotes the p-th layer network parameters. The first term in (11) minimizes the volume of hyper-ellipsoids, and the second term ensures the generalization performance and lessens the risk of over-fitting.
The curse of rejection is that there is no prior knowledge about out-of-library samples. The rejection boundary only relies on limited in-library samples and could be inexplicit, which leads to insufficient rejection performance. To tackle the issue, it is urgent to exploit available out-of-library samples as much as possible. In recent years, semi-supervised rejection methods with outlier exposure [35,36,37] come into being. These methods attempt to regard out-of-library samples as negative samples to adjust the rejection boundary and obtain performance gains.
Based on DMMSVDD, we designed a semi-supervised method called Deep Multi-modal Semi-supervised Anomaly Detection (DMMSAD) for the HRRP rejection task. It makes use of the out-of-library samples to aid in learning a more explicit rejection boundary. The new objective function is defined as:
min θ , μ k , Σ k , K 1 N + M n = 1 N min k = 1 , , K ϕ x n ; θ μ k T Σ k   1 ϕ x n ; θ μ k + η N + M m = 1 M min s = 1 , , K ϕ x ˜ m ; θ μ s T Σ s   1 ϕ x ˜ m ; θ μ s 1 + λ 2 p = 1 P θ p F 2
where  x ˜ m  denotes an out-of-library sample,  M  denotes the total number of out-of-library samples, and  η  denotes a non-negative hyperparameter that balances the effects of in-library and out-of-library samples on the boundary optimization. Compared with (11), the added term in (12) constrains out-of-library samples to be excluded from their nearest hyper-ellipsoids and further guarantees the compactness and clarity of the rejection boundary, as shown in Figure 5.

3.5. Training

The training procedure is explained in this section, including initializing and updating parameters.

3.5.1. Initializing

The initialization of the feature extractor parameters and the GMM parameters are described below.
  • Initialization of feature extractor
The proposed method uses in-library samples for initializing, and the optimization may get stuck in local optima due to the lack of constraints on the distribution of out-of-library samples [38]. To improve the convergence, the network parameters are initialized with the AE, for which the structure is shown in Figure 6.
Assuming that  f e ( · ) : X Z  is the encoder and  f d ( · ) : Z X  is the decoder, both form an AE, where the encoder structure is consistent with the feature extractor structure. Given a sample  x n X , its reconstructed sample  x ^ n  is obtained by the AE, i.e.,
x ^ n = f d f e x n ; θ e ; θ d
where  θ e  and  θ d  denote the parameters of the encoder and decoder, respectively. During the training, the above AE is optimized using the mean square error as the loss function, i.e.,
min θ e , θ d 1 N n = 1 N x n x ^ n 2 2
The training ends when the loss of AE converges. Then, the encoder parameters initialize the feature extractor parameters.
  • Initialization of GMM
The parameters of GMM contain the number of Gaussian components  K , the corresponding mean  μ k , and the covariance matrix  Σ k  of each Gaussian component. The number of Gaussian components is selected empirically, and the mean and covariance matrix are initialized by K-means clustering [39].

3.5.2. Updating

Considering the different scales of the network parameters  θ , mean  μ k  and covariance matrix  Σ k , it is difficult to update them directly using SGD [25], so we alternatively optimize Equation (11).
Step 1: Given the network parameters, the parameters of the GMM are optimized. Firstly, the number of Gaussian components of the GMM is selected. When the component number is too large, the redundant components could describe superfluous noise information and decrease the generalization performance. While the component number is too small, the GMM is incapable of delicately portraying the multi-modal distribution. Therefore, in order to select a fitted model structure, we adopt the ISODATA algorithm [40] to adaptively adjust the number of Gaussian components according to Equation (15).
K = k = 1 K 1 { n k > = ν · max { n 1 , n 2 , , n K } }
where  n k  is the sample number corresponding to the k-th Gaussian components, and  ν  controls a fraction of hyper-ellipsoids to be abandoned. Secondly, the mean  μ k  and covariance matrix  Σ k  are estimated using Gaussian mixture clustering [41].
Step 2: Parameters of the GMM are fixed, Equation (11) is optimized using the Adam optimizer [42], and the backpropagation algorithm updates the network parameters.

3.6. Theoretical Analysis

Essentially, parameter optimization is meant to maximize the log-likelihood of all samples in the Expectation-Maximization framework [43], regarding the GMM as latent variables. To be specific, the objective function is to maximize the log-likelihood of all samples, i.e.,
max n = 1 N log p x n ; θ
In the GMM, the objective function can be further expressed as
max n = 1 N k = 1 K p μ k , Σ k ; x n , θ · log p x n , μ k , Σ k ; θ
E-step: Estimate the posterior probability  p ( μ k , Σ k ; x n , θ ) . Given the parameters of the GMM, we apply a hard shrinkage operation on the posterior probability, i.e.,
p ( μ k , Σ k ; x n , θ ) = 1 k = i
where  i = argmin k { ( ϕ ( x n ; θ ) μ k ) T Σ k   1 ( ϕ ( x n ; θ ) μ k ) } .
M-step: Maximize the log-likelihood. Under the assumption of a uniform prior over Gaussian components, we have
p x n , μ k , Σ k ; θ = p x n ; μ k , Σ k , θ · p μ k , Σ k ; θ = 1 K · 1 2 π d 2 Σ k 1 2 · exp 1 2 ϕ x n ; θ μ k T Σ k 1 ϕ x n ; θ μ k
Substituting Equations (18) and (19) into (17) and ignoring the constant term, the objective function can be expressed as
min n = 1 N k = 1 K 1 k = i ϕ x n ; θ μ k T Σ k   1 ϕ x n ; θ μ k
which is consistent with (11).

3.7. Rejection Criterion

In theory, in-library sample distribution fits closely, and the samples that deviate from the distribution are the out-of-library samples. Geometrically, the in-library samples are compactly wrapped inside hyper-ellipsoids in the feature space, while out-of-library samples fall outside. Therefore, the location of the test sample in the feature space can determine its category according to (21).
C x = in - library   samples ,   dist x < dist threshold out - of - library   samples ,   otherwise
where  dist x = min k = 1 , 2 , , K ϕ x ; θ μ k T Σ k   1 ϕ x ; θ μ k , and  dist threshold  denotes the distance threshold, which can be set according to the false alarm probability designed by the recognition system.

4. Results

In this section, we verify the rejection performance of proposed methods on the measured HRRP dataset. First, we briefly describe the details of the dataset. Then, implementation details are presented to ensure the reproduction of experiments. Next, we provide the definition and meaning of evaluation metrics. Finally, four experiments are conducted to exhibit the superior rejection performance of the proposed methods.

4.1. Dataset

The experimental radar transmits linear frequency modulation signals with a signal bandwidth of 400 MHz. The collected dataset consists of measured HRRPs of 10 types of aircraft targets. The training set contains 6000 HRRPs for every type, and the test set also contains 6000 HRRPs for every type. In order to verify the generalization performance of the method, the pitch angles of targets in the test set are slightly different from those in the training set.

4.2. Implementation Details

In this section, the performance of the proposed method is compared with conventional rejection methods and newly developed deep rejection methods. Traditional methods, including one-class SVM (OCSVM), isolated forest (IF), and kernel density estimation (KDE), and deep methods such as AE [44], VAE [45], MPN [46], HRN [47], NeuTraLAD [48], DSVDD, and DMSVDD, all use pre-processed HRRPs as inputs. The main structure and parameter settings of the feature extractor and decoder in the AE are shown in Table 1, and the encoder in the AE is consistent with the feature extractor. The initial number of Gaussian components is set to 10, and the hyperparameters  η λ , and  v  are set to 0.1, 10−5, and 0.1, respectively. The learning rate of the Adam optimizer is set to 10−5. To be fair, we use the same convolution modules and deconvolution modules in other convolutional deep rejection methods.

4.3. Evaluation Metrics

In order to compare the rejection performance of different methods, we use the Receiver Operating Characteristics (ROC) curve and Area Under the Curve (AUC) for evaluation [24]. The ROC curve is obtained by plotting the true positive rate (TPR) versus the false positive rate (FPR), and they are defined as follows:
TPR = TP TP + FN FPR = FP FP + TN
where  TP  denotes the total number of in-library samples being identified correctly in the test set,  FN  denotes the total number of in-library samples being misjudged,  FP  denotes the total number of out-of-library samples being misjudged, and  TN  denotes the total number of out-of-library samples being identified correctly. Obviously, the closer the ROC curve is to the upper left region, the larger the AUC and the better the rejection performance. In addition, following the previous works [48], we also use the F1-score as the evaluation indicator, which is calculated according to Equation (23), and the larger the F1-score, the better the rejection performance.
Precision = T P T P + F P Recall = T P T P + F N F 1 = 2 · Precision · Recall Precision + Recall

4.4. Experiment with All Training Samples

In this section, we study how the methods deal with the HRRP rejection task. We randomly select 6 classes from 10 classes as in-library, and the remaining classes are treated as out-of-library. The above operations are repeated 4 times, corresponding to four experimental settings, and each experiment is repeated 10 times. Figure 7 shows the ROC curves of different methods, and Table 2 gives the corresponding AUC and F1-score.
We can draw the following conclusions from the experiment results in Figure 7 and Table 2. Firstly, the rejection performance under the four settings is different. This means that the performance is associated with the classes of out-of-library samples, and some classes are difficult to reject. Secondly, the rejection performance of deep rejection methods is generally better than that of traditional methods. Limited by the shallow structure, features extracted by traditional methods, such as OCSVM, IF, and KDE, cannot effectively distinguish multi-class in-library samples and out-of-library samples. In contrast, deep rejection methods use neural networks to nonlinearly project high-dimensional samples to the feature space, where in-library samples are well separated from out-library samples. Finally, the results also illustrate the superiority of the proposed methods in most situations. As shown in Figure 7, the ROC curves of the proposed methods are closer to the upper left region, corresponding to larger AUC and F1-score in Table 2. In the HRRP rejection task, in-library samples are usually multi-class, which makes the multi-modal characteristic more prominent. MPN and NeuTraLAD aim to learn latent features of the in-library samples by data reconstruction or self-supervised learning, and HRN addresses the output bias to in-library samples with a gradient penalization. However, it weakens performance by neglecting the influence of multi-modal information on features. The reason for the decreasing performance of DSVDD and DMSVDD is that they restrict the in-library samples to obey the Gaussian distribution with zero mean and unit variance, which is obviously inconsistent with the multi-modality characteristic of in-library samples. The GMM in our methods fits closely to the complex distribution of multi-class in-library samples and tightens the rejection boundary. The dual attention module in the feature extractor optimizes the integration of convolutional features so that the support area of HRRP attracts more attention, which reflects the abundant physical structure information of the targets. Meanwhile, the dual attention module automatically focuses on the shifting target support area within HRRP and weakens the influence of the time-shift sensitivity on the performance to some extent. Furthermore, available out-of-library samples assist the optimization to a more explicit boundary, and the rejection of unknown out-of-library samples is more effective.

4.5. Experiment with Different Training Sample Sizes

Considering the difficulty of acquiring HRRPs in practice, the rejection performance of DMMSVDD is evaluated on several small sample sets without out-of-library samples. Four training sets are constituted from the aforementioned training set by uniform sampling, and the number of samples of each class target is 6000, 3000, 1500, and 750. For each training set, five classes are randomly selected as in-library targets, and the remaining classes are treated as out-of-library targets. The test set is kept constant. Each experiment is repeated 10 times, and Table 3 shows the AUC and F1-score of each method with different training sample sizes.
Traditional rejection methods work stably yet poorly, and the AUC of OCSVM, IF, and KDE remains at 79%, 76%, and 75%, respectively. Although their shallow structure does not require too many training samples to maintain performance, it limits the feature extraction ability of high-dimensional HRRP. As the training sample size decreases, the performance of most deep rejection methods weakens, but the proposed method still achieves remarkable performance. When the number of training samples drops from 30,000 to 3750, the AUC of AE, MPN, and DMSVDD declines to 73%, 68%, and 78%, respectively, whereas the AUC of DMMSVDD remains above 83%, which is the highest among all rejection methods. This is because the dual attention module grasps more comprehensive information and enhances feature discrimination by dynamically focusing on the HRRP support area. Moreover, the GMM characterizes the anisotropic and multi-modal structure of HRRP, which is beneficial to a more compact and delicate rejection boundary. In summary, the proposed method can cope with the HRRP rejection task in the case of a limited number of samples.

5. Discussion

In this section, we analyze the effectiveness of the proposed method by ablation study and visualization.

5.1. Ablation Study

The effectiveness of different components in the proposed method is evaluated in this section. The experimental setting is the same as in Section 4.4, and the results are shown in Table 4. The performance comparison between DMMSVDD and DMMSVDD-GMM verifies the significance of the rejection boundary formed with the GMM. DMSVDD-GMM eases restrictions of DMSVDD on the mean and covariance of the spatial distribution, which means the in-library samples are more compactly wrapped with multiple hyper-ellipsoids, then the rejection boundary tightens up. It significantly promotes the rejection performance, and the AUC increases from 66.07% to 73.77%. To quantify the effectiveness of the dual attention module, we add the channel attention module and the position attention module stage by stage to DMSVDD-GMM. Compared with DMSVDD-GMM, DMSVDD-GMMCA adopts the channel attention module, taking the dependency of different channel features into account. DMMSVDD further employs the position attention module, which obtains and utilizes the global temporal information among range cells within HRRP. Both attention modules enlarge the feature distinction between in-library and out-of-library samples and improve the performance. Further, the results of DMMSVDD and DMMSAD illustrate the rational leverage of out-of-library samples is indeed a simple but effective way for performance enhancement. The rejection boundary is more explicit with some out-of-library samples; therefore, the AUC of DMMSAD increases to 75.41%.

5.2. Visualization

5.2.1. Visualization of Separability

To demonstrate the feature extraction ability explicitly, Figure 8 visualizes the features of different methods. As these high-dimensional features are difficult to display directly, we visualize them by using the t-SNE [49] technique to reduce the dimension. The experimental setting is the same as in Section 4.4. Compared with other methods, features extracted by the proposed method show more regularity. The proposed method can obtain more compact features of multi-class in-library samples, while the features of other methods are relatively scattered. At the same time, there is less overlap between the features of in-library samples and out-of-library samples, so the proposed method can distinguish in-library samples well from out-of-library samples. This explains why the proposed methods outperform others.

5.2.2. Visualization of Position Attention Maps

To illustrate the effectiveness of the position attention module explicitly, Figure 9 visualizes measured HRRPs and the corresponding position attention maps of three types of targets with different sizes. The horizontal and vertical coordinates in measured HRRPs are range cell and magnitude, and the horizontal and vertical coordinates in position maps are both feature dimensions. Measured HRRPs in Figure 9 demonstrate that the wider target support area within an HRRP represents the larger actual size of the target. Obviously, the HRRP of target-A has the narrowest target support area, indicating that its actual size is the smallest, and the actual size of target-B and target-C is increasing. Figure 9 also shows the discrepancy among the position attention maps of three targets, with the relevant area in the position attention map of target-B being longer than that of target-A and shorter than that of target-C. In conclusion, the position attention map relates to the target support area within its HRRP, similarly reflecting the actual size of the target. The position attention module is capable of adaptively identifying the region where the target support area is located and assigning more weight to features in this region. It is conducive to strengthening the influence of the support area on the rejection performance and thus learning more discriminative features.

6. Conclusions

In this paper, a novel method named DMMSVDD is proposed to solve the HRRP-based rejection problem. The core ideas are to obtain more discriminative features with an attention module and form a more compact and explicit rejection boundary with the multi-modality of HRRP. Firstly, the data preprocessor is designed to alleviate the sensitivity problems of HRRP. Then, the feature extractor with a dual attention module is responsible for refining convolutional features with the global-dependency and channel-dependency information, which is effective in capturing abundant information in the support area of HRRP. Next, the rejector takes the anisotropy and multi-modality characteristics of HRRP into account and forms a more compact rejection boundary of multiple closed hyper-ellipsoids. The rejector is jointly optimized with the feature extractor under a unified rejection criterion. Moreover, its semi-supervised version, named DMMSAD, is extended to take advantage of out-of-library samples to obtain a more explicit boundary. Experiments demonstrate the promising performance of the proposed methods on the HRRP rejection task.
In applications, the HRRP database can be continuously expanded by merging accumulated out-of-library samples. And the users need to update the rejection model in a timely manner to improve the rejection performance. However, retraining a new rejection model every time is costly. Hence, in the future, we will introduce continual learning methods to update the existing model online based on newly acquired data, which is more efficient and cost-effective.

Author Contributions

Methodology, Y.D. and P.W.; writing—original draft preparation, Y.D.; writing—review and editing, P.W., H.L., M.F., Y.G., L.C. and J.Y.; supervision, P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (61701379 and 62192714), the Stabilization Support of National Radar Signal Processing Laboratory (KGJ202204), the Fundamental Research Funds for the Central Universities (QTZX22160), the Industry-University-Research Cooperation of the 8th Research Institute of China Aerospace Science and Technology Corporation (SAST2021-011), the Open Fund Shaanxi Key Laboratory of Antenna and Control Technology, and the 111 Project.

Data Availability Statement

The data are not publicly available due to privacy.

Acknowledgments

The authors thank the editors and reviewers for their constructive comments and professional suggestions to improve the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, T.; Zhang, X.; Liu, C.; Shi, J.; Wei, S.; Ahmad, I.; Zhan, X.; Zhou, Y.; Pan, D.; Li, J. Balance learning for ship detection from synthetic aperture radar remote sensing imagery. ISPRS J. Photogramm. 2021, 182, 190–207. [Google Scholar] [CrossRef]
  2. Zhang, T.; Zhang, X. A mask attention interaction and scale enhancement network for SAR ship instance segmentation. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4511005. [Google Scholar] [CrossRef]
  3. Yan, J.; Jiao, H.; Pu, W.; Shi, C.; Dai, J.; Liu, H. Radar sensor network resource allocation for fused target tracking: A brief review. Inform. Fusion 2022, 86, 104–115. [Google Scholar] [CrossRef]
  4. Yan, J.; Pu, W.; Zhou, S.; Liu, H.; Bao, Z. Collaborative detection and power allocation framework for target tracking in multiple radar system. Inform. Fusion 2020, 55, 173–183. [Google Scholar] [CrossRef]
  5. Yang, Y.; Zheng, J.; Liu, H.; Ho, K.; Chen, Y.; Yang, Z. Optimal sensor placement for source tracking under synchronization offsets and sensor location errors with distance-dependent noises. Signal Process. 2022, 193, 108399. [Google Scholar] [CrossRef]
  6. Liu, X.; Wang, L.; Bai, X. End-to-end radar HRRP target recognition based on integrated denoising and recognition network. Remote Sens. 2022, 14, 5254. [Google Scholar] [CrossRef]
  7. Zhang, T.; Zhang, X. A polarization fusion network with geometric feature embedding for SAR ship classification. Pattern Recogn. 2022, 123, 108365. [Google Scholar] [CrossRef]
  8. Zhang, T.; Zhang, X. Squeeze-and-excitation Laplacian pyramid network with dual-polarization feature fusion for ship classification in SAR images. IEEE Geosci. Remote Sens. Lett. 2021, 19, 4019905. [Google Scholar] [CrossRef]
  9. Zhang, T.; Zhang, X. Injection of traditional hand-crafted features into modern CNN-based models for SAR ship classification: What, why, where, and how. Remote Sens. 2021, 13, 2091. [Google Scholar] [CrossRef]
  10. Li, X.; Ran, J.; Wen, Y.; Wei, S.; Yang, W. MVFRnet: A novel high-accuracy network for ISAR air-target recognition via multi-view fusion. Remote Sens. 2023, 15, 3052. [Google Scholar] [CrossRef]
  11. Schölkopf, B.; Platt, J.C.; Shawe-Taylor, J.; Smola, A.J.; Williamson, R.C. Estimating the support of a high-dimensional distribution. Neural Comput. 2001, 13, 1443–1471. [Google Scholar] [CrossRef]
  12. Tax, D.M.; Duin, R.P. Support vector data description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
  13. Liu, F.T.; Ting, K.M.; Zhou, Z.-H. Isolation forest. In Proceedings of the Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008. [Google Scholar]
  14. Erfani, S.M.; Rajasegarar, S.; Karunasekera, S.; Leckie, C. High-dimensional and large-scale anomaly detection using a linear one-class SVM with deep learning. Pattern Recogn. 2016, 58, 121–134. [Google Scholar] [CrossRef]
  15. Andrews, J.; Tanay, T.; Morton, E.J.; Griffin, L.D. Transfer representation-learning for anomaly detection. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016. [Google Scholar]
  16. Pang, G.; Shen, C.; Cao, L.; Hengel, A.V.D. Deep learning for anomaly detection: A review. ACM Comput. Surv. 2021, 54, 1–38. [Google Scholar] [CrossRef]
  17. Perera, P.; Nallapati, R.; Xiang, B. OCGAN: One-class novelty detection using GANs with constrained latent representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  18. Akcay, S.; Atapour-Abarghouei, A.; Breckon, T.P. GANomaly: Semi-supervised anomaly detection via adversarial training. In Proceedings of the 14th Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2019. [Google Scholar]
  19. Zaheer, M.Z.; Lee, J.-h.; Astrid, M.; Lee, S.-I. Old is gold: Redefining the adversarially learned one-class classifier training paradigm. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  20. Sabokrou, M.; Fathy, M.; Zhao, G.; Adeli, E. Deep end-to-end one-class classifier. IEEE Trans. Neur. Net. Lear. 2020, 32, 675–684. [Google Scholar] [CrossRef]
  21. Golan, I.; El-Yaniv, R. Deep anomaly detection using geometric transformations. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  22. Tack, J.; Mo, S.; Jeong, J.; Shin, J. CSI: Novelty detection via contrastive learning on distributionally shifted instances. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020. [Google Scholar]
  23. Ristea, N.-C.; Madan, N.; Ionescu, R.T.; Nasrollahi, K.; Khan, F.S.; Moeslund, T.B.; Shah, M. Self-supervised predictive convolutional attentive block for anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  24. Ruff, L.; Vandermeulen, R.; Goernitz, N.; Deecke, L.; Siddiqui, S.A.; Binder, A.; Müller, E.; Kloft, M. Deep one-class classification. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  25. Ghafoori, Z.; Leckie, C. Deep multi-sphere support vector data description. In Proceedings of the 2020 SIAM International Conference on Data Mining, Cincinnati, OH, USA, 7–9 May 2020. [Google Scholar]
  26. Xia, Z.; Wang, P.; Dong, G.; Liu, H.J. Radar HRRP Open Set Recognition Based on Extreme Value Distribution. IEEE Trans. Geosci. Remote Sens. 2023, 61, 3257879. [Google Scholar] [CrossRef]
  27. Chen, J.; Du, L.; Guo, G.; Yin, L.; Wei, D. Target-attentional CNN for radar automatic target recognition with HRRP. Signal Process 2022, 196, 108497. [Google Scholar] [CrossRef]
  28. Zhang, T.; Zhang, X.; Shi, J.; Wei, S. HyperLi-Net: A hyper-light deep learning network for high-accurate and high-speed ship detection from synthetic aperture radar imagery. ISPRS J. Photogramm. 2020, 167, 123–153. [Google Scholar] [CrossRef]
  29. Zhang, T.; Zhang, X.; Ke, X. Quad-FPN: A novel quad feature pyramid network for SAR ship detection. Remote Sens. 2021, 13, 2771. [Google Scholar] [CrossRef]
  30. Zhang, T.; Zhang, X. A full-level context squeeze-and-excitation ROI extractor for SAR ship instance segmentation. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4506705. [Google Scholar] [CrossRef]
  31. Zhang, T.; Zhang, X.; Ke, X.; Liu, C.; Xu, X.; Zhan, X.; Wang, C.; Ahmad, I.; Zhou, Y.; Pan, D. HOG-ShipCLSNet: A novel deep learning network with hog feature fusion for SAR ship classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–22. [Google Scholar] [CrossRef]
  32. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  33. Wang, P.; Shi, L.; Du, L.; Liu, H.; Xu, L.; Bao, Z. Radar HRRP statistical recognition with temporal factor analysis by automatic Bayesian Ying-Yang harmony learning. Front. Electr. Electron. Eng. China 2011, 6, 300–317. [Google Scholar] [CrossRef]
  34. Liu, H.; Du, L.; Wang, P.; Pan, M.; Bao, Z. Radar HRRP automatic target recognition: Algorithms and applications. In Proceedings of the 2011 IEEE CIE International Conference on Radar, Chengdu, China, 24–27 October 2011. [Google Scholar]
  35. Hendrycks, D.; Mazeika, M.; Dietterich, T. Deep anomaly detection with outlier exposure. arXiv 2018, arXiv:1812.04606. [Google Scholar]
  36. Ruff, L.; Vandermeulen, R.A.; Görnitz, N.; Binder, A.; Müller, E.; Müller, K.-R.; Kloft, M. Deep semi-supervised anomaly detection. arXiv 2019, arXiv:1906.02694. [Google Scholar]
  37. Yao, X.; Li, R.; Zhang, J.; Sun, J.; Zhang, C. Explicit Boundary Guided Semi-Push-Pull Contrastive Learning for Supervised Anomaly Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 24490–24499. [Google Scholar]
  38. Zaheer, M.Z.; Mahmood, A.; Khan, M.H.; Segu, M.; Yu, F.; Lee, S.-I. Generative cooperative learning for unsupervised video anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  39. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Caliafornia, CA, USA, 21 June–18 July 1965 and 27 December 1965–7 January 1966. [Google Scholar]
  40. Ball, G.H.; Hall, D.J. A clustering technique for summarizing multivariate data. Behav. Sci. 1967, 12, 153–155. [Google Scholar] [CrossRef] [PubMed]
  41. Yang, M.-S.; Lai, C.-Y.; Lin, C.-Y. A robust EM clustering algorithm for Gaussian mixture models. Pattern Recogn. 2012, 45, 3950–3961. [Google Scholar] [CrossRef]
  42. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Pernkopf, F.; Bouchaffra, D. Genetic-based EM algorithm for learning Gaussian mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1344–1348. [Google Scholar] [CrossRef] [PubMed]
  44. Masci, J.; Meier, U.; Cireşan, D.; Schmidhuber, J. Stacked convolutional auto-encoders for hierarchical feature extraction. In Proceedings of the 21st International Conference on Artificial Neural Networks, Espoo, Finland, 14–17 June 2011. [Google Scholar]
  45. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  46. Lv, H.; Chen, C.; Cui, Z.; Xu, C.; Li, Y.; Yang, J. Learning normal dynamics in videos with meta prototype network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  47. Hu, W.; Wang, M.; Qin, Q.; Ma, J.; Liu, B. HRN: A holistic approach to one class learning. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020. [Google Scholar]
  48. Qiu, C.; Pfrommer, T.; Kloft, M.; Mandt, S.; Rudolph, M. Neural transformation learning for deep anomaly detection beyond images. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  49. Van der Maaten, L.; Hinton, G. Visualizing non-metric similarities in multiple maps. Mach. Learn. 2012, 87, 33–55. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed method.
Figure 1. The framework of the proposed method.
Remotesensing 16 00649 g001
Figure 2. HRRP pre-processing procedure.
Figure 2. HRRP pre-processing procedure.
Remotesensing 16 00649 g002
Figure 3. The details of the dual attention module.
Figure 3. The details of the dual attention module.
Remotesensing 16 00649 g003
Figure 4. The design principle of the rejector and its geometric explanation.
Figure 4. The design principle of the rejector and its geometric explanation.
Remotesensing 16 00649 g004
Figure 5. The geometric explanation of DMMSAD.
Figure 5. The geometric explanation of DMMSAD.
Remotesensing 16 00649 g005
Figure 6. The structure of AE.
Figure 6. The structure of AE.
Remotesensing 16 00649 g006
Figure 7. ROC curves of different methods under different experiment settings. (a) Experimental setting I; (b) experimental setting II; (c) experimental setting III; (d) experimental setting IV.
Figure 7. ROC curves of different methods under different experiment settings. (a) Experimental setting I; (b) experimental setting II; (c) experimental setting III; (d) experimental setting IV.
Remotesensing 16 00649 g007
Figure 8. The t-SNE visualization of different methods. The blue dots indicate the in-library sample features, and the yellow dots indicate the out-of-library sample features. (a) AE; (b) NeuTraLAD; (c) DSVDD; (d) DMSVDD; (e) DMMSVDD (ours); (f) DMMSAD (ours).
Figure 8. The t-SNE visualization of different methods. The blue dots indicate the in-library sample features, and the yellow dots indicate the out-of-library sample features. (a) AE; (b) NeuTraLAD; (c) DSVDD; (d) DMSVDD; (e) DMMSVDD (ours); (f) DMMSAD (ours).
Remotesensing 16 00649 g008
Figure 9. Measured HRRPs and position attention maps of three types of aircraft targets with different sizes. (a) Target-A with small size; (b) target-B with medium size; (c) target-C with large size.
Figure 9. Measured HRRPs and position attention maps of three types of aircraft targets with different sizes. (a) Target-A with small size; (b) target-B with medium size; (c) target-C with large size.
Remotesensing 16 00649 g009
Table 1. Main structure and parameter settings of our method.
Table 1. Main structure and parameter settings of our method.
ModuleLayerOutput SizeNormalization/Activation
Feature
Extractor
Convolution
Module 1
Conv1D8 × 256BN/Leaky ReLU
Max Pooling8 × 128-
Conv1D16 × 128BN/Leaky ReLU
Max Pooling16 × 64-
Dual Attention
Module
Channel Attention16 × 64-
Position Attention16 × 64-
Convolution
Module 2
Conv1D8 × 64BN/Leaky ReLU
Max Pooling8 × 32-
Flattening1 × 256-
Conv1D1 × 32-
DecoderDeconvolution
Module 1
Reshape2 × 16-
Upsample2 × 32-
Deconv1D8 × 32BN/Leaky ReLU
Upsample8 × 64-
Deconv1D16 × 64BN/Leaky ReLU
Upsample16 × 128-
Dual Attention
Module
Channel Attention16 × 128-
Position Attention16 × 128-
Deconvolution
Module 2
Deconv1D8 × 128BN/Leaky ReLU
Upsample8 × 256-
Deconv1D1 × 256BN/Leaky ReLU
Sigmoid1 × 256-
Table 2. AUC and F1-score of different methods under different experiment settings, with the best results bolded and the second-best underlined.
Table 2. AUC and F1-score of different methods under different experiment settings, with the best results bolded and the second-best underlined.
MethodAUC/F1-Score
Setting ISetting IISetting IIISetting IVAverage
OCSVM50.80/41.0783.98/73.5259.03/46.3861.88/50.4663.92/52.86
IF52.62/43.3180.58/63.5954.91/45.2251.82/42.6659.98/48.70
KDE47.06/38.1081.88/69.6653.91/41.4055.67/45.3859.63/48.64
AE52.06/42.2564.55/53.1954.56/43.9368.02/54.9359.80/48.58
VAE48.30/39.0967.62/56.7048.28/38.9548.40/37.7653.15/43.13
MPN64.15/53.2378.64/67.0159.58/47.6470.16/57.0568.13/56.23
HRN56.55/46.0968.87/57.6454.60/44.6459.43/49.1359.86/49.38
NeuTraLAD62.96/51.1881.55/68.2856.25/45.4565.60/53.3966.59/54.58
DSVDD62.82/51.0882.07/69.4667.32/55.1066.69/54.7669.73/57.60
DMSVDD66.07/53.9886.77/74.8167.79/55.6367.01/55.2371.91/59.91
DMMSVDD (ours)74.77/62.4690.64/79.5780.79/68.2569.04/56.1978.81/66.61
DMMSAD (ours)75.41/62.3991.16/81.0981.01/68.5076.03/64.0080.90/68.99
Table 3. AUC and F1-score of different methods with different training sample sizes, with the best results bolded and the second-best underlined.
Table 3. AUC and F1-score of different methods with different training sample sizes, with the best results bolded and the second-best underlined.
MethodTraining Sample Sizes (AUC/F1-Score)
3750750015,00030,000
OCSVM79.54/73.9379.59/73.7679.63/74.0179.67/77.46
IF75.91/71.3276.07/71.4176.15/71.4576.18/71.54
KDE75.50/71.7275.79/71.9875.80/72.0175.83/73.76
AE73.93/67.9677.46/70.4478.19/70.7979.01/70.54
VAE64.83/61.5464.91/61.5764.95/61.6265.13/61.85
MPN68.75/63.8473.26/66.8378.40/71.4978.77/71.89
HRN68.12/62.8871.09/65.4171.44/66.3272.19/66.47
NeuTraLAD78.26/71.1478.49/71.5879.01/71.7679.26/71.93
DSVDD78.00/71.5478.19/71.6878.85/72.0981.25/74.16
DMSVDD78.85/71.8479.86/72.7980.27/73.0983.87/73.77
DMMSVDD (ours)83.06/76.0684.97/78.0886.49/79.9987.70/79.88
Table 4. Impact of different modules on AUC, with the best results bolded and the second-best underlined.
Table 4. Impact of different modules on AUC, with the best results bolded and the second-best underlined.
MethodGMMChannel
Attention
Position
Attention
Out-of-Library
Samples
AUC
DMSVDD 66.07
DMSVDD-GMM 73.77
DMSVDD-GMMCA 74.22
DMMSVDD (ours) 74.77
DMMSAD (ours)75.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, Y.; Wang, P.; Fang, M.; Guo, Y.; Cao, L.; Yan, J.; Liu, H. Radar High-Resolution Range Profile Rejection Based on Deep Multi-Modal Support Vector Data Description. Remote Sens. 2024, 16, 649. https://doi.org/10.3390/rs16040649

AMA Style

Dong Y, Wang P, Fang M, Guo Y, Cao L, Yan J, Liu H. Radar High-Resolution Range Profile Rejection Based on Deep Multi-Modal Support Vector Data Description. Remote Sensing. 2024; 16(4):649. https://doi.org/10.3390/rs16040649

Chicago/Turabian Style

Dong, Yue, Penghui Wang, Ming Fang, Yifan Guo, Lili Cao, Junkun Yan, and Hongwei Liu. 2024. "Radar High-Resolution Range Profile Rejection Based on Deep Multi-Modal Support Vector Data Description" Remote Sensing 16, no. 4: 649. https://doi.org/10.3390/rs16040649

APA Style

Dong, Y., Wang, P., Fang, M., Guo, Y., Cao, L., Yan, J., & Liu, H. (2024). Radar High-Resolution Range Profile Rejection Based on Deep Multi-Modal Support Vector Data Description. Remote Sensing, 16(4), 649. https://doi.org/10.3390/rs16040649

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop