Next Article in Journal
Soil Moisture Mapping with Moisture-Related Indices, OPTRAM, and an Integrated Random Forest-OPTRAM Algorithm from Landsat 8 Images
Next Article in Special Issue
Airborne Passive Bistatic Radar Clutter Suppression Algorithm Based on Root Off-Grid Sparse Bayesian Learning
Previous Article in Journal / Special Issue
Influence of Plasma Sheath’s Velocity Field on ISAR Imaging of Hypersonic Target
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Multiple-Measurement Sparsity-Aware STAP with Bayesian Variational Autoencoder

1
National Lab of Radar Signal Processing, Xidian University, Xi’an 710071, China
2
The 38th Research Institute of China Electronics Technology Corporation, Hefei 230088, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(15), 3800; https://doi.org/10.3390/rs14153800
Submission received: 13 June 2022 / Revised: 18 July 2022 / Accepted: 20 July 2022 / Published: 6 August 2022
(This article belongs to the Special Issue Radar High-Speed Target Detection, Tracking, Imaging and Recognition)

Abstract

:
Due to the shortage of independent and identically distributed (i.i.d.) training samples, space−time adaptive processing (STAP) often suffers remarkable performance degradation in the heterogeneous clutter environment. Sparse recovery (SR) techniques have been introduced into STAP for the benefit of the drastically reduced training requirement, but they are incompletely robust for involving the tricky selection of hyper−parameters or the undesirable point estimation for parameters. Given this issue, we incorporate the Multiple−measurement Complex−valued Variational relevance vector machines (MCV) to model the space−time echoes and provide a Gibbs−sampling−based method to estimate posterior distributions of parameters accurately. However, the Gibbs sampler require quantities of iterations, as unattractive as traditional Bayesian type SR−STAP algorithms when the real−time processing is desired. To address this problem, we further develop the Bayesian Autoencoding MCV for STAP (BAMCV−STAP), which builds the generative model according to MCV and approximates posterior distributions of parameters with an inference network pre−trained off−line, to realize fast reconstruction of measurements. Experimental results on simulated and measured data demonstrate that BAMCV−STAP can achieve suboptimal clutter suppression in terms of the output signal to interference plus noise ratio (SINR) loss, as well as the attractive real−time processing property in terms of the convergence rate and computational loads.

1. Introduction

Airborne phased−array radar is widely used to detect moving targets because of its mobility and freedom from the radius of the earth’s curvature, but it is always plagued by the existence of the ‘nuisance’ clutter coupling in the space−time domain. To ensure the detection of weak targets, space-time adaptive processing (STAP) methods [1,2,3,4] are developed to mitigate the clutter and achieve better performance compared with traditional non−adaptive methods. For classical STAP methods, the performance is substantially determined by the accuracy of the clutter plus noise covariance matrix (CCM). Typically, the CCM of the cell under test (CUT) is estimated by training samples adjacent to the CUT, which are supposed to be independent and identically distributed (i.i.d.) [5]. According to the Reed−Mallett−Brennan (RMB) rule [6], the required training samples should be at least twice the system degree of freedom (DOF) to achieve steady performance. Nevertheless, since the clutter background is usually non−homogeneous and the DOF is usually high in practice, obtaining adequate qualified training samples is challenging, leading to dramatically degraded performance of CCM estimation and clutter suppression. A large number of methods to improve the performance of STAP with finite training samples have been proposed in the past few years, such as the reduced−dimension (RD) methods [7,8,9,10], the reduced−rank methods (RR) [11,12,13], the direct data domain (D3) methods [14] and the knowledge−aided STAP (KA−STAP) methods [15,16,17,18,19]. However, the number of required samples is still large in the RD and RR methods, the DOF of D3 is significantly reduced resulting in severe performance loss, and the KA−STAP methods depend heavily on prior knowledge. By contrast, the sparse recovery STAP (SR−STAP) methods [20,21,22,23], which are not only more tolerant with the number of training samples but also more independent from the prior knowledge, have received significant attention. In SR−STAP, every atom of the dictionary is a space−time steering vector denoting a grid of the space−time plane, which is uniformly over−sampled in both the angle and Doppler. The space−time profile, recovered by solving the formulated optimization problem, will be leveraged to reconstruct the CCM for effective clutter suppression. Since the clutter suppression performance mainly relies on the accuracy of recovery results, many efforts have been put into the problem formulation of sparse recovery. The pioneering work, such as greedy algorithms [24,25,26,27], considers the problem of SR−STAP as an l 0 −norm minimization problem which is known as NP−hard [28,29]. To tackle the NP−hard problem, appealing approaches transform it into a convex problem [20,30,31,32], but the strong coherence among the atoms of sparse dictionary often leads to the severe performance degradation of sparse reconstruction. Besides, the recovery results obtained by convex optimization are particularly sensitive to the selection of regularization parameters.
Due to the superior ability to deal with the highly coherent dictionary situation and the tricky selection of regularization parameters, Bayesian type sparse recovery (SR) algorithms [22,33,34,35], represented by the sparse Bayesian learning (SBL) methods [36] and its extension for the multiple measurement vector (MMV) case, i.e., M−SBL [37], have received much attention in the field of STAP. However, the potential need for a large number of iterations per testing in these methods is a fatal drawback for STAP where real−time processing is desired. Some methods [33,34,35,38] are devoted to modifying the time−consuming process and accelerating the convergence, but they are at the price of either significant signal loss or expensive computational loads, which also hinder the real−time processing. In addition, all these Bayesian type SR methods are incompletely robust as a result of only providing point estimates for some parameters, such as the parameter controlling the variance of the space−time profile and the parameter denoting the noise power. In summary, the completely robust Bayesian type SR−STAP algorithms that ensure both the stable performance of sparse reconstruction and the feasibility of real−time processing are rarely studied.
To model the space−time echoes with a completely robust Bayesian framework, this paper introduces variational relevance vector machines (VRVM) [39] into STAP due to its attractive advantage of providing distributional estimates for all parameters without any user parameters (known as the parameter−free property) and sufficient ability to model the sparsity. However, the original VRVM is proposed in the real domain for the single measurement vector (SMV) case, which is not suitable for STAP where the complex−valued multiple measurement vector (MMV) case exists. As a result, we first develop the multiple−measurement complex−valued variational relevance vector machines (MCV), which can address the condition where complex−valued multiple measurements occur. To achieve accurate inference, we employ the Gibbs sampler as a general technique to estimate all parameters for MCV. Then, we present the novel STAP algorithm via MCV, named MCV−STAP, to design the adaptive filters. However, a large number of iterations, as well as expensive computational loads, are required in the Gibb sampler, which limits the efficiency of the model for out−of−sample prediction. To render the real−time processing of the MCV−STAP, inspired by the structure of the variational autoencoder (VAE) [40,41,42,43], the powerful models for unsupervised learning in the field of deep learning, we introduce a VAE into our model to provide the variational inference and expand it to a VAE−structured model named Bayesian Autoencoding MCV−STAP (BAMCV−STAP), which is scalable in the training phase and fast in the testing phase. Then, our model can be optimized by minimizing the evidence lower bound (ELBO) via a gradient ascent scheme. We note that due to the gradient ascent scheme without involving inverse operations when learning the parameters, BAMCV−STAP can approximate posterior densities of space−time profiles with low computational loads. Moreover, since the inference network can be pre−trained off−line, only a few iterations are needed before testing. Finally, the space−time profiles recovered via BAMCV−STAP are adopted to estimate the CCM for STAP. To the best of our knowledge, BAMCV−STAP is the first work that attempts a VRVM−with−VAE strategy in the context of STAP, and our method appears to outperform several other sparsity−enhanced methodologies in terms of clutter suppression performance and real−time processing, as will be demonstrated in the experimental results.
The main contributions of this paper are summarized as follows:
  • Generalized from the original VRVM to the multiple measurements case existing in the complex domain, a parameter−free probabilistic model called MCV is derived to recover space−time profiles via the Gibbs sampling method for STAP.
  • Since all parameters are estimated based on their posterior distributions in MCV, the robustness to the number of training samples and noise power estimation is significantly improved compared with other SR−STAP methods for the MMV case.
  • Incorporating a suitable VAE into MCV, a novel method called BAMCV is developed to accelerate the convergence of iterative procedures for estimating parameters. As the inference network is pre−trained off−line, BAMCV−STAP can realize the sparse reconstruction with lower computational loads and much fewer iterations compared with conventional SR−STAP methods.
  • As demonstrated on both simulated and measured data, the final proposed method BAMCV−STAP can process space−time echoes in real−time without degrading clutter suppression performance.
The rest of this paper is organized as follows. Section 2 introduces the signal model for airborne radar. Section 3 details the proposed algorithm. Numerical experiments with both simulated and measured data are shown in Section 4 before discussion in Section 5. Section 6 presents the conclusion and future work of findings.

2. System Model and Derivation of Optimal Filters

For tractability purpose, a side−looking airborne phased−array radar with a uniformly linear antenna (ULA) [34,44] consisting of M half wavelength spaced elements on the airborne radar platform is considered. As shown in Figure 1, the platform is at altitude h p and moving with constant velocity v p . φ and θ are the azimuth angle and the elevation angle, respectively. The radar transmits N pulses at a constant pulse repetition frequency (PRF) in a coherent processing interval (CPI). The model for the space−time clutter plus noise snapshot x C M N × 1 from the CUT is [45]
x = x c + n = i = 1 N c α i v f d , i , f s , i + n ,
where x c C M N × 1 is the space−time clutter vector and n C M N × 1 is the Gaussian white thermal noise whose variance is σ 2 ; N c denotes the number of independent clutter patches in the CUT; α i , f d , i , f s , i , v f d , i , f s , i C M N × 1 are the complex amplitude, the Doppler frequency, the spatial frequency and the space−time steering vector of the ith clutter patch, respectively. Considering the criterion of linearly constrained minimum variance (LCMV), the optimal STAP weight vector is [45]
S w = κ R 1 v t ,
where κ = v t H R 1 v t 1 is the normalization constant, v t is the space−time steering vector of the target, and R denotes the CCM, which is normally unknown and estimated from i.i.d training samples. ( · ) H denotes the conjugate operation. In conventional STAP algorithms, the CCM can be calculated by [45]
R = 1 L l = 1 L x l x l H ,
where L is the number of training samples and x l C M N × 1 ( l = 1 , 2 , , L ) denotes the space−time clutter plus noise vector of lth training sample. In order to achieve stable performance, L is required to be at least twice of M N . However, it is difficult to obtain sufficient i.i.d. training samples in practice. As a result, SR−STAP algorithms, which exhibit better performance using far fewer training samples, stand out from the crowd. They start by dividing the whole space−time plane uniformly into K = N s N t bins, where N s = ρ s M ρ s > 1 and N t = ρ t N ρ t > 1 are the number of spatial frequency grid points and Doppler frequency grid points, respectively; ρ s and ρ d determine the resolution of the spatial frequency and Doppler frequency axes. As each grid point can be represented as a space−time steering vector v k ( k = 1 , 2 , 3 , , K ) , STAP dictionary Φ C M N × K can be denoted as a collection of all these space−time steering vectors, i.e.,
Φ = v 1 , v 2 , , v K .
For the SMV case, the signal model in Equation (1) can be re−expressed as
x = Φ w + n ,
where w C K × 1 denotes the space−time profile with non−zero elements representing the presence of the clutter. The formulation of the CCM for the SMV case can be expressed as
R = Φ w w H Φ H + σ 2 I .
For the MMV case, the signal model of L range cells X = x 1 , x 2 , , x L C M N × L can be expressed as
X = Φ W + N ,
where W = w 1 , w 2 , , w L C K × L is an unknown solution matrix, N = n 1 , n 2 , , n L C M N × L is a Gaussian white thermal noise matrix. Inconsistent with the SMV case, row sparsity is encouraged simultaneously in the MMV case, i.e., each training sample in X has the same indices of direction of arrival (DOA). The calculation of the CCM for the MMV case can be expressed as
R = 1 L Φ W W H Φ H + σ 2 I .
The optimal solution to W is discussed in the next section.

3. Proposed MCV−STAP Algorithm and BAMCV−STAP Algorithm

In this section, we first introduce the MCV approach by generalizing the complex−valued VRVM from the SMV case to the MMV case. Then we propose an MCV−STAP algorithm for designing the adaptive filters to suppress clutter and develop a Gibbs sampler to estimate parameters. To remedy the unattractive property requiring quantities of iterations in the Gibbs sampler of the MCV−STAP algorithm, we provide an efficient autoencoding variational inference network for the generative model built on the MCV approach and develop a VAE−structured method named BAMCV−STAP.

3.1. Derivation of MCV

We now derive the MCV approach by generalizing the complex−valued VRVM from the SMV case to the MMV case. Firstly, for the SMV case, assuming the noise is white complex Gaussian distribution with unknown power σ 2 , we can obtain the Gaussian likelihood function of the signal model in Equation (5) as
p x w , σ 2 CN Φ w , σ 2 I .
Drawing on the concept of automatic relevance determination (ARD) [46], we assign the Gaussian prior on the kth row of w R K as
p w k γ k CN 0 , γ k 1
where γ k is an unknown variance parameter. Combining these row priors, a full prior of variable w can be expressed as
p w γ k = 1 K p w k γ k
where γ = γ 1 , γ 2 , , γ K T R + K is a vector of hyper−parameters. In this way, every weight has an individual hyper−parameter to moderate the strength of the prior. To ensure the parameter−free property of our method and avoid point estimation of γ and σ 2 [36], the Gamma prior [47] over γ k as well as over the noise variance σ 2 in the signal model are defined, i.e.,
p γ = k = 1 K G a m m a γ k a k , b k ,
p β = G a m m a β c , d ,
where β = σ 2 and
G a m m a ( γ k | a k , b k ) = Γ a k 1 b k a k γ k a k 1 exp ( b k γ k ) .
Γ ( a ) is the gamma function expressed as
Γ a = 0 t a 1 exp ( t ) d t .
By providing posterior distribution estimates for γ and β , our method increases the robustness of the noise power estimates while requiring fewer training samples.
Extending the above model for the SMV case to the MMV case to obtain the proposed model MCV, the Gaussian likelihood function of the signal model can be rewritten as
p X W , β CN Φ W , β 1 I .
The prior of the kth row of W in Equation (7) denoted by w k · is
p w k · γ k CN 0 , γ k 1 I .
So the prior of the lth column of W denoted by w · l is
p w · l γ CN ( 0 , A 1 )
where A = d i a g ( γ ) and d i a g ( · ) denotes the diagnonal operation. Hence the full prior of W  is
p ( W | γ ) = Δ l = 1 L p ( w · l | γ ) .
The priors of γ and β are the same as the case of SMV.

3.2. Proposed MCV−STAP Algorithm

The Gibbs sampler to estimate all parameters for MCV are derived from the SMV case in this subsection. Firstly, for the SMV case, the conditional posterior distribution over w can be written as
p w x , γ , β = p x | w , β p ( w γ ) p x γ , β = Σ 1 π K e x p w μ H Σ 1 w μ
where the posterior covariance and mean are respectively
Σ = β i = 1 M N ϕ i H ϕ i + A 1 ,
μ = β Σ Φ H x ,
with ϕ i denoting the i-th row vector of Φ .
Similarly, the conditional posterior distribution over γ can be expressed as
p γ | w , a 0 , b 0 = p w γ p γ a 0 , b 0 p w a 0 , b 0 = k = 1 K G a m m a γ k a k , b k
where the shape parameter and the scale parameter can be expressed as
a k = a k , 0 + 1 , b k = b k , 0 + w k 2 .
a 0 = [ a 1 , 0 , a 2 , 0 , . . . , a K , 0 ] T denotes the vector of initial values for shape parameters of γ , and b 0 = [ b 1 , 0 , b 2 , 0 , . . . , b K , 0 ] T denotes the vector of initial values for scale parameters of γ . The conditional posterior distribution over β is
p β x , w , c 0 , d 0 = p x w , β p β c 0 , d 0 p x w , c 0 , d 0 = G a m m a β c , d ,
where the shape parameter and the scale parameter respectively become
c = c 0 + M N , d = d 0 + x H x 2 w H Φ H x + i = 1 M N ϕ i ( w w H ) ϕ i H .
c 0 denotes the initial value for shape parameters of β and d 0 denotes the initial values for scale parameters of β .
Extending the above process to the MMV case, we develop the Gibbs sampler for MCV here. The conditional posterior distribution over the latent variable W is thus given by
p W X , γ , β = p X | W , β p ( W γ ) p X γ , β = Σ M 1 π K e x p { t r W μ M H Σ M 1 W μ M }
where the covariance and mean are respectively
Σ M = fi β i = 1 MN ϕ i H ϕ i + A 1 ,
μ M = β Σ M Φ H X ,
and t r ( · ) denotes the trace of a square matrix. The conditional posterior distribution over γ can be expressed as
p γ | W , a 0 , b 0 = p W γ p γ a 0 , b 0 p W a 0 , b 0 = k = 1 K G a m m a γ k a k , b k
where the shape parameter and the scale parameter can be expressed as
a k = a k , 0 + L , b k = b k , 0 + w k . 2 2 .
The conditional posterior distribution over β is
p β X , W , c 0 , d 0 = p X W , β p β c 0 , d 0 p X W , c 0 , d 0 = G a m m a β c , d ,
where the shape parameter and the scale parameter respectively become
c = c 0 + M N L , d = d 0 + l = 1 L x l H x l 2 l = 1 L W H Φ H x l + i = 1 M N ϕ i ( W W H ) ϕ i H .
As conditional distributions over all parameters are specified clearly in the above, the space−time profiles estimated accurately by the Gibbs sampler of MCV will be adopted to design filters to suppress the clutter. For clarity, the pseudo−codes of the abovementioned MCV−STAP algorithm are provided in Algorithm 1.
Algorithm 1 MCV−STAP algorithm.
Step1: Give initial values a k , 0 = b k , 0 = c 0 = d 0 = 10 6 , k = 1 , 2 , 3 , . . . , K .
Step2: Sample γ ( 0 ) from Equation (12) and sample β ( 0 ) from Equation (13).
Step3: For t = 1 , 2 , . . . do
     Calculate μ M ( t ) and Σ M ( t ) via Equations (28) and (29);
     Sample W ( t ) from Equation (20);
     Calculate a ( t ) and b ( t ) via Equation (31);
     Sample γ ( t ) from Equation (30);
     Calculate c ( t ) and d ( t ) via Equation (33);
     Sample β ( t ) from Equation (32);
     Check for convergence.
    End For
    Denote the last iter number as T.
Step4: Assume W = μ M ( T ) , calculate the CCM via Equation (8) and the space−time adaptive optimal
    weight vector via Equation (1).
Step5: Denote the CUT as x v t , the output of MCV−STAP algorithm solved by Gibbs sampling
    is y = S v t H x v t .
While having closed−form updating equations and providing accurate estimation for parameters, the Gibbs sampler is still not attractive since it takes a large number of iterations to infer the sparse representation during the testing phase, which hinders real−time processing of the incoming CUT and motivates us to construct a VAE with fast testing, as described below.

3.3. BAMCV−STAP Algorithm

In this section, to ensure the real−time processing of MCV−STAP, we extend it into a VAE−structured method named BAMCV−STAP to map the observation x directly to the latent representation for out−of−sample prediction. Specifically, the generative model (decoder) of BAMCV−STAP is built based on MCV as shown in Figure 2a, and the inference network (encoder) is accomplished by the neural network with the structure shown in Figure 2b. Instead of needing expensive iterative inference schemes with Gibbs samplers introduced above, the variational inference network of BAMCV allows us to efficiently realize sparse reconstructions by fitting an approximate inference model to the unknown posterior using standard stochastic gradient methods. According to the theory of mean−field variational Bayes [48,49,50], the ELBO of BAMCV can be expressed as
L = D K L [ q ( γ ) | | p ( γ ) ] D K L [ q ( w | x ) | | p ( w | γ ) ] D K L [ q β | | p ( β ) ] + E q ( w , γ , β | x ) p ( x | w , γ , β ) .
where the approximate conditional posterior distributions q ( γ ) , q w | x and q β can be written as
q ( γ ) = G a m m a a h , b h , q w | x = CN μ h , Σ h , q β = G a m m a c h , d h .
Noticeably, D K L [ q ( · ) | | p ( · ) ] denotes the Kullback–Leibler (KL) divergence between q ( · ) and  p ( · ) .
Different from traditional Gaussian distribution based VAEs, the gamma distribution, modeling the sparsity and satisfying non−negative constraint, is defined over γ and β here. However, it is hard to compute the gradient of the ELBO to γ and β due to the difficulty of reparameterizing gamma distributed random variables [51,52]. To reparameterize γ and β easily without deviating from gamma distributions, the Weibull distribution is considered an ideal choice to replace the gamma distribution, as demonstrated in [53]. Consequently, gamma distributed conditional posteriors, q ( γ ) and q ( β ) , are approximated by Weibull distributions expressed as
q ( γ ) = i = 1 K q ( γ i ) = i = 1 K W e i b u l l k γ i , λ γ i , q β = W e i b u l l k β , λ β .
where k γ i and λ γ i are the parameters of γ i , and λ β and k β are the parameters of γ i . Latent variables γ i and β can be easily reparameterized as [53,54]
γ i = λ γ i ( ln ( 1 ε γ i ) ) 1 k γ i , ε γ i U n i f o r m ( 0 , 1 ) , β = λ β ( ln ( 1 ε β ) ) 1 k β , ε β U n i f o r m ( 0 , 1 ) .
In addition, the Gaussian distribution based variable w can be simply reparameterized as [40]
w r = μ h r + Σ h z r , z r N ( 0 , I ) , w i = μ h i + Σ h z i , z i N ( 0 , I )
where the superscript r of all parameters represents their real part and the superscript i of all parameters represents their imaginary part. ( ) denotes matrix multiplication. According to the illustration in [53], the KL divergence from the gamma distribution to the Weibull distribution, which is the first term of the ELBO in Equation (34), can be expressed analytically as
D K L [ q ( γ ) | | p ( γ ) ] = D K L [ W e i b u l l ( k γ , λ γ ) | | G a m m a ( a 0 , b 0 ) ] = i = 1 K γ e a i , 0 k γ i a i , 0 ln λ γ i + ln k γ i + b i , 0 λ γ i Γ ( 1 + 1 k γ i ) γ e 1 a i , 0 ln ( b i , 0 ) + ln Γ ( a i , 0 ) .
where γ e is the Euler–Mascheroni constant. The second term of the ELBO can be expressed analytically as
D K L [ q ( w | x ) | | p ( w | γ ) ] = D K L [ CN ( μ h , Σ h ) | | CN ( 0 , A 1 ) ] = i = 1 K ( 1 + l o g ε k k | μ h , i | 2 ε k k ) .
ε k k means the k-th diagonal element of Σ h , and μ h , i denotes the i-th element of μ h . | · | means the modulus operation. The third term and the last term of the ELBO can be expressed analytically as
D K L [ q ( β ) | | p ( β ) ] = D K L [ W e i b u l l ( k β , λ β ) | | G a m m a ( c 0 , d 0 ) ] = γ e c 0 k β c 0 l n λ β + ln k β + d 0 λ β Γ ( 1 + 1 k β ) γ e 1 c 0 ln ( d 0 ) + ln Γ ( c 0 ) .
E q ( w , γ , β | x ) p ( x | w , γ , β ) = x Φ w 2 2
As illustrated in Figure 2a, all the parameters relative to conditional distributions are transformed from the observation x i n p u t with the neural network. Specifically,
μ h r = R e L U C μ r h 1 + g μ r , μ h i = R e L U C μ i h 1 + g μ i , μ h = μ h r + j μ h i ,
where μ h is the mean vector of w ,
e h = R e L U C e h 1 + g e ,
where e h = d i a g ( Σ h ) , and
h 1 = R e L U C 1 h x i n p u t + g 1 h
where R e L U ( · ) denotes the non−linear activation function. Note that x i n p u t denotes the concatenation of the real part and the imaginary part of training data, as x and latent representation w are complex−valued. Similarly,
k γ = S o f t p l u s C a h 2 + g a , λ γ = S o f t p l u s C b h 2 + g b k β = S o f t p l u s C c h β + g c , λ β = S o f t p l u s C d h β + g d
h 2 = R e L U C 2 h h 1 + g 2 h h β = R e L U C β x i n p u t + g β
where k γ = [ k γ 1 , k γ 2 , . . . , k γ K ] R K × 1 , λ γ = [ λ γ 1 , λ γ 2 , . . . , λ γ K ] R K × 1 and S o f t p l u s ( · ) applies l o g [ 1 + e x p ( · ) ] non-linearity to each element to ensure positive Weibull shape and scale parameters.
Obviously, C μ r , C μ i , C e , C 1 h , g μ r , g μ i , g e , g 1 h , C a , C b , C c , C d , C 2 h , C β , g a , g b , g c , g d , g 2 h and g β are parameters in the neural network. The superscript r of all parameters represents the real part, and the superscript i of all parameters represents the imaginary part. For clarity, the dimensions of parameters in the inference network are summarized in Table 1.
Above all, due to the simple reparameterization tricks, our proposed VAE−structured MCV−STAP methods can obtain the latent representations directly with the aid of an autoencoding inference network.
Moreover, to make the proposed method further attractive in real−time processing, we pre−train the inference network off−line with simulated data preparing for sequential fine−tuning facing with realistic clutter data to be processed [55]. The main steps of the proposed BAMCV for STAP (BAMCV−STAP) algorithm are summarized in Algorithm 2. For clarity, the flow chart of BAMCV is plotted in Figure 3.
Algorithm 2 BAMCV−STAP algorithm.
Step1: Simulate data for pre−training using radar system parameters of realistic clutter data under
   test. Denote the simulated dataset as D p r e and the realistic dataset as D t e s t .
Step2: Pre−train the inference network with dataset D p r e off−line until the ELBO converges.
Step3: Select CUT in D t e s t and choose L training samples around CUT.
Step4: Fine−tune the inference network with the selected L training samples until the ELBO
   converges again. Assume W = μ h .
Step6: Estimate the CCM via Equation (8) and the space−time adaptive optimal weight vector via Equation (2).
   Denote the weight vector as S v t .
Step7: Denote data in CUT as x c u t , the output of BAMCV−STAP solved by the inference network is
    y = S v t H x c u t .
Compared with MCV, BAMCV has two explicit advantages which can be summarized as follows:
  • Decreasing the computational loads. Inspired by the gradient ascent scheme for parameters optimization, the parameters of BAMCV are updated by the backpropagation of the gradient, which only involves some linear operations, exponential operations and logarithmic operations, instead of involving complex inverse operations and multiple samplings appeared in MCV.
  • Improving the convergence rate for testing. As the inference network is pre−trained off−line, only a few iterations are taken in BAMCV to obtain the recovery results of observations rather than a large number of time−consuming iterations per testing in MCV.
Experimental results shown in the following section will verify these advantages.

4. Numerical Results

In this section, we evaluate the proposed algorithms, MCV−STAP and BAMCV−STAP, using both simulated data and measured data in terms of clutter suppression performance, time efficiency and computational complexity. Significantly, the metric to evaluate clutter suppression performance for the simulated data is the signal to interference plus noise ratio (SINR) loss [45], which is defined as follows
L SINR = σ n 2 M N S w H v t 2 S w H R S w .
where S w denotes the STAP weight vector, R denotes the true CCM, v t denotes the target steering vector and σ n 2 denotes the actual noise power. In addition, as the true CCM is unable to be obtained for the measured data, we evaluate the target detection performance for targets with normalized Doppler frequency of 0.1 by the probability of detection (PD) versus signal to noise ratio (SNR) curves, which are achieved by utilizing the adaptive matched filter (AMF) detector [56] and averaged over 10 5 Monte Carlo trials. The probability of the false alarm rate (PFA) is set as 10 6 , the PD is estimated based on 10 6 randomly sampled CUTs and the PD to SNR curves are averaged over 10 5 Monte Carlo trials. SNR is defined as the ratio of signal power (for a single pulse per antenna element) to the rough estimate of noise power (the average energy of echoes within the support region) here. In addition, to evaluate the time efficiency, we evaluate the time spent on the convergence rate per testing. The computational complexity is also evaluated by the comparison of computational loads per testing. For clarity, the performance metrics used in this paper are summarized in Table 2, in which the check mark means the metric is used in the data and the blank means not.
Moreover, the proposed algorithm is compared with some other classical methods especially for the MMV case, such as loading sample matrix inversion (LSMI) for STAP [57,58], multiple orthogonal matching pursuit (M−OMP) for STAP [59], M−SBL for STAP and multiple fast converging SBL (M−FCSBL) for STAP [34]. The LSMI for STAP [57,58], which is the representative of classical STAP algorithm, is an adaptive algorithm imposing diagonal loading (DL) to improve the robustness of conventional Capon beamformers but fails to achieve satisfactory performance unless the number of training samples is more than twice the clutter rank. The M−OMP for STAP, orthogonal matching pursuit (OMP) [25] for the MMV case, obtains sparse representations through a greedy algorithm but is particularly unfriendly to the highly coherent dictionary. As the classical technique to solve sparse recovery issues utilizing the greedy strategy, M−OMP is chosen as one of comparison methods in this paper. The M−SBL for STAP, suppressing the clutter imposing M−SBL algorithm based on Bayesian theory as well as ours, receives much attention in the field of STAP for the excellent performance with a highly coherent dictionary but potentially requires a large number of iterations per testing leading to unsuccessful real−time processing. The M−FCSBL for STAP, a fast converging M−SBL algorithm, improves the convergence by incorporating a simple approximation term and achieves a favorable recovery result with only a few iterations. However, the experimental results reveal that this approximation term leads to sub−optimal recovery results, as shown in its convergence curve.
Furthermore, the elements of all weight matrices in the neural network employed in this paper are initialized with Gaussian distributions with standard deviations set to 0.1, and all bias terms are set to zero. During the pre−training stage, we set the mini−batch size to 32. For optimization, the Adam optimizer [60] with a learning rate of 10 4 is used. On a Pentium PC with a 3.7 GHz CPU and 64 GB RAM, the inference network is pre−trained and fine−tuning in non−optimized Python software before being moved to MATLAB for testing. The amount of samples in the pre−training data set is ten times the number of parameters in the neural network to ensure the generalization of the pre−trained neural network. The settings of each layer in the neural network have been stated in Table 1.

4.1. Simulated Data

In this paper, a side−looking airborne phased array radar with a uniformly linear antenna (ULA), whose inter−element spacing is half wavelength, is considered. The parameters of radar system are shown in Table 3. In the simulations, we set N s = 4 M , N t = 4 N , σ n 2 = 1 and 180 clutter patches uniformly distributed from π / 2 to π / 2 . According to Algorithm 2, the data set D p r e and D t e s t for BAMCV−STAP are simulated with the same radar system parameters, respectively. Obviously, MCV−STAP is executed on D t e s t directly without need of D p r e . To make the inference network easily generalized to the training samples of CUT, We simulated as many training samples as ten times the number of parameters with random clutter to signal ratio (CNR) sampled from 30 dB to 70 dB in D p r e . As for D t e s t , to investigate the performance of all methods under the heterogeneous clutter scenario conveniently, we assume the 150−th range cell as CUT which is located at the interested area with the average CNR = 45 dB and around by only 10 training samples satisfying the i.i.d constraint, as shown in Figure 4. Compared with the clairvoyant spectrum calculated by known CCM in Figure 5a, the clutter spectrum estimated by 128 training samples, as many as twice the DOF, exhibits too large value along the clutter ridge due to the heterogeneity of samples, as depicted in Figure 5b. Noticeably, the clutter suppression performance with the different number of i.i.d training samples in the interested area will also be investigated later.

4.1.1. Analysis of Clutter Suppression Performance

The clutter suppression performance of the proposed method is evaluated here. The clutter spectrum of clutter data under test utilizing various approaches are depicted in Figure 6. The number of training samples is fixed to 10 since the number of i.i.d training samples is set as 10 according to Figure 4. As is customary in LSMI, the load factor is set at 1. The clutter spectrum computed by MCV in Figure 6f and BAMCV in Figure 6h is the closest to the ideal one in Figure 6a in terms of both position and value. As the posterior distributions are strictly derived via Bayesian theory, the CCM computed using Gibbs sampling is as close to the clairvoyant one as BAMCV−STAP. Because of limited i.i.d. training samples, the approximated spectrum by M−OMP in Figure 6c and LSMI in Figure 6b expose their powerlessness in clutter subspace estimation. Instead, the clutter spectrum obtained by M−SBL, as shown in Figure 6d, is extremely close to the ideal one, with a small spreading and a few noise peaks due to imprecise noise power assessment. In Figure 6e, M−FCSBL [34] achieves an accurate clutter subspace and performs no noise peaks. However, upon further investigation of these SR−STAP approaches, as shown in Figure 7, we discover that values along the clutter ridge in the clutter spectrum calculated by M−FCSBL are not as well as expected. The clutter ridge formed by our method, on the other hand, is closer to the ideal result.
Moreover, to demonstrate the importance of the pre−training to the inference network, we utilize the inference network after pre−training to recover the space−time profile of clutter data under test, as shown in Figure 6g. Although the significant values along the clutter ridge are not accurate enough, the sparsity of the clutter spectrum and the shape of the clutter ridge are initialized successfully as the preparation for sequential fine−tuning. Apart from the sparse subspace, the noise power estimation is also an important issue for the CCM, according to Equation (8). Consequently, we enumerate the final results estimated by all comparison methods in Table 4. Clearly, our methods outperform M−SBL and M−FCSBL, proving the robustness to the estimation for noise power.
When the number of training samples is 10, the clutter suppression performance measured by SINR loss is evaluated in Figure 8. The optimal SINR loss curve is completed using the clairvoyant CCM. In comparison to previous Bayesian SR−STAP approaches, LSMI and M−OMP display severe performance loss in the sidelobe region, as depicted in Figure 8a. Specifically, the output SINR loss of MCV−STAP is the closest to the optimal one among various Bayesian SR−STAP methods, as shown in Figure 8b (which is an enlargement of Figure 8a). BAMCV−STAP achieves the lower SINR loss than MCV−STAP as shown in Figure 8 but still surpasses other comparison methods.
Besides, we compare the effect of reducing the i.i.d training samples on the SINR loss of each method, and the results are shown in Figure 9. The average SINR loss is defined in this section as the mean of the SINR loss values across the entire normalized Doppler frequency range. According to Figure 9, on average sense, when L is a large value, all the methods have almost equivalent performance. But with an decrease in L, the Bayesian type SR−STAP methods, including M−SBL−STAP, MCV−STAP and BAMCV−STAP, have the larger SINR loss. Notably, our methods, both MCV−STAP and BAMCV−STAP, maintain stable performance as the number of training samples decreases and achieve a 3 dB higher SINR loss than M−SBL−STAP with only two training samples, demonstrating the robustness of the proposed methods to the number of training samples.

4.1.2. Analysis of Computational Loads and Convergence Rate

To study the real−time processing of several SR−STAP algorithms with similarly superior performance i.e., M−SBL, M−FCSBL and our methods, we detail the analysis of the convergence rate and computational loads in Figure 10.
For the on−line phase, convergence curves, utilizing the negative log−likelihood as the unified loss function [34] for fair comparison, are shown in Figure 10a, i.e.,
L cost = ln M + t r ( M 1 R cost )
where
M = Φ H A 1 Φ + β 1 I
R cost = 1 L X X H .
Different from MCV converging as slowly as M−SBL, BAMCV, which is pre−trained with lots of simulated data off−line, exhibits the dominant superiority at the beginning of the iteration. Then, after fine−tuning with several snapshots around CUT over a few iterations, BAMCV converges to a stable and optimal state. The time spent on each iteration of each method is also listed in Table 5. Indeed, since it uses a gradient ascent scheme without any inverse operation or sampling, BAMCV takes much less time per iteration than the other three approaches. However, M−SBL and MCV require not only a large number of iterations but also a long time for each iteration. Furthermore, even though M−FCSBL quickly converges to its steady−state value, the final solution is not optimal, which is exactly coincident with the estimated clutter spectrum in Figure 6e and Figure 7.
In addition, the computational loads of M−SBL, M−FCSBL, MCV and BAMCV are detailed in Table 6. Specifically, N s is the iteration number of the M−SBL algorithm, N f is the iteration number of the M−FCSBL algorithm and N g is the iteration number of MCV for our method. Due to the intractable inverse operations and the large number of iterations, the computational loads of M−SBL and M−FCSBL, as well as MCV are about three times of the DOF. By contrast, since the time−consuming training process of neural networks is finished before testing, BAMCV achieves fast super−resolution by directly mapping out sparse representation vectors of training samples with only twice of the DOF. To provide an explicit illustration, we also display the curve of computational loads versus DOF in Figure 10b with setting K = 16 M N , L = 10 and N s = N f = N g = 1 . Obviously, the computational complexity of BAMCV is much lower than that of other methods, and the gap between BAMCV and other methods diverges as DOF increasing.
In summary, the modified method BAMCV−STAP not only requires less online running time and lower computational loads than other comparison methods while achieving comparable clutter suppression performance, but also remains robust to the number of training samples and noise power. Consequently, BAMCV−STAP is suitable for airborne early warning (AEW) radar applications.

4.2. Measured Data

In this subsection, MCV−STAP and BAMCV−STAP are applied to the public available data set, the Multi−Channel Airborne Radar Measurements (MCARM) data set [61], to verify their attractiveness. Some indispensable parameters of the radar system are listed in Table 7. The clutter spectrum estimated using 10 training samples around the selected target range cell by different methods is exhibited in Figure 11. Consistent with simulated results, limited by the number of training samples, the clutter spectrum estimated by the LSMI method in Figure 11a is as rough as the one by the M−OMP method in Figure 11b. Due to the approximation term, the clutter spectrum estimated by M−FCSBL in Figure 11d method is too sparse to reconstruct the information in the sidelobe region. M−SBL exhibits a better capability of recovering the clutter spectrum using only 10 training samples but with a slight broadening of the clutter ridge as shown in Figure 11c.
As it is difficult to judge whether our methods or M−SBL achieve the best performance according to the clutter spectrum estimated shown in Figure 11c,e,f, we explore the detection performance averaged over all range cells for the targets with the normalized Doppler frequency of 0.1 (corresponding velocity is 10 m/s) by the PD versus SNR curves as depicted in Figure 12a. The probability of the false alarm rate (PFA) is set as 10 6 , the PD to SNR for each CUT is averaged over 10 5 Monte Carlo trials. To further explore the detection performance of targets with different velocity values, we exhibit the PD versus velocity curves (SNR = 20 dB), which are averaged over 10 5 Monte Carlo trials, in Figure 12b.
Moreover, the detection performance to explore the robustness to the number of training samples is provided in Figure 12c. As shown in Figure 12, similar to the LSMI method, M−FCSBL suffers a severe performance degradation as the information in the sidelobe region is ultimately given up due to the approximation term, leading to the too sparse reconstructions to obtain the accurate CCM. M−SBL method achieves sub−optimal performance among comparison methods but with very slow convergence, as shown in Figure 13 and Table 8. On the contrary, BAMCV−STAP achieves the sub−optimal target detection performance. Besides that, according to the computational loads analyzed in Figure 10b and Table 6, the computational loads of BAMCV is about O ( M N ) but the other methods are about O ( ( M N ) 3 ) . Consequently, BAMCV−STAP can achieve favorable detection of targets and keep robust to the number of training samples, with the most efficient convergence and the lowest computational loads among comparison methods.
To sum up, among these comparison approaches, our method exhibits its superiority in recovering the clutter spectrum in terms of both sparsity and accuracy on measured data, which is consistent with the results on simulated data.

5. Discussion

Along with the impressive performance achieved by our methods, the above−mentioned experimental results indicate that it is substantially more challenging for traditional SR−STAP algorithms, such as the proposed MCV−STAP, to ensure both the stable performance of sparse reconstruction and real−time processing. However, deep learning techniques, such as the proposed BAMCV−STAP, can efficiently realize parameter estimates for real−time on−line processing at the expense of nonsignificant off−line pre−training.

6. Conclusions

Traditional SR−STAP algorithms are unattractive as they only provide point estimates for several parameters and rely on a large number of iterations per testing at the testing stage. To avoid the iterative procedure, we first propose MCV−STAP, a parameter−free SR−STAP method that incorporates a modified VRVM into the field of STAP, to model the space−time echoes with a completely robust Bayesian framework. Then, to realize the real−time processing of the incoming space−time echoes, we introduce BAMCV−STAP, a VAE−structured method that builds its encoder from MCV but with an efficient inference network instead of the time−consuming Gibbs sampler in MCV−STAP. Compared with the conventional Bayesian type SR−STAP algorithms, the inference network of BAMCV−STAP can be pre−trained off−line using a gradient ascent scheme, enabling more efficient sparse reconstruction. The experimental results show that the BAMCV−STAP algorithm can achieve comparable clutter suppression with lower computational cost.
This paper focuses on the suppression of non−homogeneous clutter environments with side−looking airborne phased−array radar by means of deep learning methods. Inspired by their attractive characteristics, approaches that can solve the other tricky problems of radar signal processing, such as the suppression of the non−stationary clutter and the detection of slow−moving targets, are the promising directions that we plan to explore in the future.

Author Contributions

Conceptualization, P.W. and B.C.; methodology, C.Z.; software, H.Z. and C.Z.; validation, H.Z. and C.Z.; formal analysis, H.Z. and C.Z.; investigation, C.Z.; resources, P.W.; data curation, C.J.; writing—original draft preparation, C.Z.; writing—review and editing, C.Z. and W.C.; visualization, C.Z. and W.C.; supervision, C.Z. and W.C.; project administration, B.C. and H.L.; funding acquisition, B.C. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation Of China, Grant No. 61771361.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, X.; Cheng, Y.; Wu, H.; Wang, H. Heterogeneous clutter suppression for airborne radar STAP based on matrix manifolds. Remote Sens. 2021, 13, 3195. [Google Scholar] [CrossRef]
  2. Li, H.; Liao, G.; Xu, J.; Zeng, C. Sub-CPI STAP based clutter suppression and target refocusing with airborne radar system. Digit. Signal Process. 2022, 123, 103418. [Google Scholar] [CrossRef]
  3. Liu, K.; Wang, T.; Wu, J.; Chen, J. A Two-Stage STAP Method Based on Fine Doppler Localization and Sparse Bayesian Learning in the Presence of Arbitrary Array Errors. Sensors 2022, 22, 77. [Google Scholar] [CrossRef] [PubMed]
  4. Xiao, H.; Wang, T.; Zhang, S.; Wen, C. A robust refined training sample reweighting space–time adaptive processing method for airborne radar in heterogeneous environment. IET Radar Sonar Nav. 2021, 15, 310–322. [Google Scholar] [CrossRef]
  5. Yifeng, W.; Tong, W.; Jianxin, W.; Jia, D. Robust training samples selection algorithm based on spectral similarity for space–time adaptive processing in heterogeneous interference environments. IET Radar Sonar Nav. 2015, 9, 778–782. [Google Scholar] [CrossRef]
  6. Liu, J.; Liu, W.; Liu, H. A simpler proof of rapid convergence rate in adaptive arrays. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 135–136. [Google Scholar] [CrossRef]
  7. Zhang, W.; He, Z.; Li, J.; Liu, H.; Sun, Y. A method for finding best channels in beam-space post-Doppler reduced-dimension STAP. IEEE Trans. Aerosp. Electron. Syst. 2014, 50, 254–264. [Google Scholar] [CrossRef]
  8. Zhang, W.; He, Z.; Li, J.; Li, C. Beamspace reduced-dimension space–time adaptive processing for multiple-input multiple-output radar based on maximum cross-correlation energy. IET Radar Sonar Nav. 2015, 9, 772–777. [Google Scholar] [CrossRef]
  9. Wang, X.; Yang, Z.; Huang, J.; de Lamare, R.C. Robust two-stage reduced-dimension sparsity-aware STAP for airborne radar With coprime arrays. IEEE Trans. Signal Process. 2020, 68, 81–96. [Google Scholar] [CrossRef]
  10. Zhang, W.; An, R.; He, N.; He, Z.; Li, H. Reduced dimension STAP based on sparse recovery in heterogeneous clutter environments. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 785–795. [Google Scholar] [CrossRef]
  11. Guerci, J.R.; Goldstein, J.S.; Reed, I.S. Optimal and adaptive reduced-rank STAP. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 647–663. [Google Scholar] [CrossRef]
  12. Fa, R.; de Lamare, R.C. Reduced-Rank STAP Algorithms using Joint Iterative Optimization of Filters. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 1668–1684. [Google Scholar] [CrossRef]
  13. Yang, Z.; Wang, X. Reduced-rank space-time adaptive processing algorithm based on multistage selections of angle-Doppler filters. IET Radar Sonar Nav. 2022, 16, 327–345. [Google Scholar] [CrossRef]
  14. Sarkar, T.K.; Wang, H.; Park, S.; Adve, R.; Koh, J.; Kim, K.; Zhang, Y.; Wicks, M.C.; Brown, R.D. A deterministic least-squares approach to space-time adaptive processing (STAP). IEEE Trans. Aerosp. Electron. Syst. 2001, 49, 91–103. [Google Scholar] [CrossRef] [Green Version]
  15. Guerci, J.R.; Baranoski, E.J. Knowledge-aided adaptive radar at DARPA: An overview. IEEE Signal Process. Mag. 2006, 23, 41–50. [Google Scholar] [CrossRef]
  16. Capraro, C.T.; Capraro, G.T.; Bradaric, I.; Weiner, D.D.; Wicks, M.C.; Baldygo, W.J. Implementing digital terrain data in knowledge-aided space-time adaptive processing. IEEE Trans. Aerosp. Electron. Syst. 2006, 42, 1080–1099. [Google Scholar] [CrossRef]
  17. Cui, N.; Xing, K.; Duan, K.; Yu, Z. Knowledge-aided block sparse Bayesian learning STAP for phased-array MIMO airborne radar. IET Radar Sonar Navig. 2021, 15, 1628–1642. [Google Scholar] [CrossRef]
  18. Liu, M.; Zou, L.; Yu, X.; Zhou, Y.; Wang, X.; Tang, B. Knowledge Aided Covariance Matrix Estimation via Gaussian Kernel Function for Airborne SR-STAP. IEEE Access 2020, 8, 5970–5978. [Google Scholar] [CrossRef]
  19. Zhu, X.; Li, J.; Stoica, P. Knowledge-aided space-time adaptive processing. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 1325–1336. [Google Scholar] [CrossRef]
  20. Yang, Z.; de Lamare, R.C.; Li, X. L1 regularized STAP algorithms With a generalized sidelobe canceler architecture for airborne radar. IEEE Trans. Signal Process. 2012, 60, 674–686. [Google Scholar] [CrossRef]
  21. Sun, K.; Zhang, H.; Li, G.; Meng, H.; Wang, X. A Novel STAP Algorithm using Sparse Recovery Technique. In Proceedings of the IEEE International Geoscience & Remote Sensing Symposium (IGARSS 2009), Cape Town, South Africa, 12–17 July 2009; pp. 336–339. [Google Scholar]
  22. Wu, Q.; Zhang, Y.D.; Amin, M.G.; Himed, B. Space-Time Adaptive Processing and Motion Parameter Estimation in Multistatic Passive Radar Using Sparse Bayesian Learning. IEEE Trans. Geosci. Remote Sens. 2016, 54, 944–957. [Google Scholar] [CrossRef]
  23. Yang, X.; Sun, Y.; Zeng, T.; Long, T.; Sarkar, T.K. Fast STAP Method Based on PAST with Sparse Constraint for Airborne Phased Array Radar. IEEE Trans. Signal Process. 2016, 64, 4550–4561. [Google Scholar] [CrossRef]
  24. Mallat, S.; Zhang, Z. Matching pursuits with time-frequency dictionaries. IEEE Trans. Signal Process. 1993, 41, 3397–3415. [Google Scholar] [CrossRef] [Green Version]
  25. Pati, Y.; Rezaiifar, R.; Krishnaprasad, P. Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition. In Proceedings of the 27th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 1–3 November 1993; pp. 40–44. [Google Scholar]
  26. Cotter, S.F.; Rao, B.D.; Engan, K.; Kreutz-Delgado, K. Sparse solutions to linear inverse problems with multiple measurement vectors. IEEE Trans. Signal Process. 2005, 53, 2477–2488. [Google Scholar] [CrossRef]
  27. Blumensath, T.; Davies, M.E. Gradient Pursuits. IEEE Trans. Signal Process. 2008, 56, 2370–2382. [Google Scholar] [CrossRef]
  28. Bruckstein, A.M.; Donoho, D.L.; Elad, M. From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images. SIAM Rev. 2009, 51, 34–81. [Google Scholar] [CrossRef] [Green Version]
  29. Tibshirani, R. Regression shrinkage and selection via the lasso: A retrospective. J. R. Stat. Soc. Series B Stat. Methodol. 2011, 73, 273–282. [Google Scholar] [CrossRef]
  30. Zibulevsky, M.; Elad, M. L1-L2 Optimization in Signal and Image Processing. IEEE Signal Process. Mag. 2010, 27, 76–88. [Google Scholar] [CrossRef]
  31. Chen, S.S.; Donoho, D.L.; Saunders, M.A. Atomic Decomposition by Basis Pursuit. SIAM Rev. 2001, 43, 129–159. [Google Scholar] [CrossRef] [Green Version]
  32. Wei, S.; Zhang, L.; Ma, H.; Liu, H. Sparse Frequency Waveform Optimization for High-Resolution ISAR Imaging. IEEE Trans. Geosci. Remote Sens. 2020, 58, 546–566. [Google Scholar] [CrossRef]
  33. Wu, Q.; Zhang, Y.D.; Amin, M.G.; Himed, B. Complex multitask Bayesian compressive sensing. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2014), Florence, Italy, 4–9 May 2014; pp. 3375–3379. [Google Scholar]
  34. Wang, Z.; Xie, W.; Duan, K.; Wang, Y. Clutter suppression algorithm based on fast converging sparse Bayesian learning for airborne radar. Signal Process. 2017, 130, 159–168. [Google Scholar] [CrossRef]
  35. Poli, L.; Oliveri, G.; Viani, F.; Massa, A. MT–BCS-based microwave imaging approach through minimum-norm current expansion. IEEE Trans. Aerosp. Electron. Syst. 2013, 61, 4722–4732. [Google Scholar] [CrossRef]
  36. Tipping, M.E. Sparse Bayesian Learning and the Relevance Vector Machine. J. Mach. Learn. Res. 2001, 1, 211–244. [Google Scholar]
  37. Wipf, D.P.; Rao, B.D. An Empirical Bayesian Strategy for Solving the Simultaneous Sparse Approximation Problem. IEEE Trans. Signal Process. 2007, 55, 3704–3716. [Google Scholar] [CrossRef]
  38. Duan, K.; Chen, H.; Xie, W.; Wang, Y. Deep learning for high-resolution estimation of clutter angle-Doppler spectrum in STAP. IET Radar Sonar Nav. 2022, 16, 193–207. [Google Scholar] [CrossRef]
  39. Bishop, C.M.; Tipping, M.E. Variational Relevance Vector Machines. In Proceedings of the the 16th Conference in Uncertainty in Artificial Intelligence, Stanford University, Stanford, CA, USA, 30 June–3 July 2000; pp. 46–53. [Google Scholar]
  40. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. In Proceedings of the International Conference on Learning Representations (ICLR 2014), Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  41. Notin, P.; Hernández-Lobato, J.M.; Gal, Y. Improving black-box optimization in VAE latent space using decoder uncertainty. In Proceedings of the Neural Information Processing Systems 2021 (NeurIPS 2021), Virtual, 6–14 December 2021; pp. 802–814. [Google Scholar]
  42. Zhou, Y.; Liang, X.; Zhang, W.; Zhang, L.; Song, X. VAE-based Deep SVDD for anomaly detection. Neurocomputing 2021, 453, 131–140. [Google Scholar] [CrossRef]
  43. Ding, M. The road from MLE to EM to VAE: A brief tutorial. AI Open 2022, 3, 29–34. [Google Scholar] [CrossRef]
  44. Hussain, A.; Anjum, U.; Channa, B.A.; Afzal, W.; Hussain, I.; Mir, I. Displaced Phase Center Antenna Processing For Airborne Phased Array Radar. In Proceedings of the 2021 International Bhurban Conference on Applied Sciences and Technologies (IBCAST), Islamabad, Pakistan, 12–16 January 2021; pp. 988–992. [Google Scholar]
  45. Ward, J. Space-time adaptive processing for airborne radar. In Proceedings of the 1995 International Conference on Acoustics, Speech, and Signal Processing, ICASSP, Detroit, MI, USA, 8–12 May 1995; pp. 2809–2812. [Google Scholar]
  46. Wipf, D.; Nagarajan, S. A New View of Automatic Relevance Determination. In Proceedings of the Twentieth Annual Conference On Neural Information Processing Systems (NIPS 2007), Vancouver, BC, Canada, 4–7 December 2006; Platt, J., Koller, D., Singer, Y., Roweis, S., Eds.; Curran Associates, Inc.: West Chester, PA, USA, 2007; Volume 20. [Google Scholar]
  47. Berger, J.O. Statistical Decision Theory and Bayesian Analysis, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 1985. [Google Scholar]
  48. Jordan, M.I.; Ghahramani, Z.; Jaakkola, T.S.; Saul, L.K. An Introduction to Variational Methods for Graphical Models. Mach. Learn. 1999, 37, 183–233. [Google Scholar] [CrossRef]
  49. Worley, B. Scalable Mean-Field Sparse Bayesian Learning. IEEE Trans. Signal Process. 2019, 67, 6314–6326. [Google Scholar] [CrossRef]
  50. Tzikas, D.G.; Likas, A.C.; Galatsanos, N.P. The variational approximation for Bayesian inference. IEEE Signal Process. Mag. 2008, 25, 131–146. [Google Scholar] [CrossRef]
  51. Kingma, D.P.; Welling, M. Stochastic gradient VB and the variational auto-encoder. In Proceedings of the the 2rd International Conference on Learning Representations (ICLR 2014), London, UK, 8–9 July 2014; Volume 19, p. 121. [Google Scholar]
  52. Ruiz, F.J.R.; Titsias, M.K.; Blei, D.M. The Generalized Reparameterization Gradient. In Proceedings of the the 29th Annual Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016; pp. 460–468. [Google Scholar]
  53. Zhang, H.; Chen, B.; Guo, D.; Zhou, M. WHAI: Weibull Hybrid Autoencoding Inference for Deep Topic Modeling. In Proceedings of the the 6th International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  54. Duan, Z.; Wang, D.; Chen, B.; Wang, C.; Chen, W.; Li, Y.; Ren, J.; Zhou, M. Sawtooth Factorial Topic Embeddings Guided Gamma Belief Network. In Proceedings of the the 38th International Conference on Machine Learning (ICML 2021), Virtual Event, 18–24 July 2021; pp. 2903–2913. [Google Scholar]
  55. Hendrycks, D.; Lee, K.; Mazeika, M. Using Pre-Training Can Improve Model Robustness and Uncertainty. In Proceedings of the the 36th International Conference on Machine Learning (ICML 2019), Long Beach, CA, USA, 9–15 June 2019; pp. 2712–2721. [Google Scholar]
  56. Robey, F.; Fuhrmann, D.; Kelly, E.; Nitzberg, R. A CFAR adaptive matched filter detector. IEEE Trans. Aerosp. Electron. Syst. 1992, 28, 208–216. [Google Scholar] [CrossRef] [Green Version]
  57. Carlson, B. Covariance matrix estimation errors and diagonal loading in adaptive arrays. IEEE Trans. Aerosp. Electron. Syst. 1988, 24, 397–401. [Google Scholar] [CrossRef]
  58. Cox, H.; Zeskind, R.M.; Owen, M.M. Robust adaptive beamforming. IEEE Trans. Acoust. Speech Signal Process. 1987, 35, 1365–1376. [Google Scholar] [CrossRef] [Green Version]
  59. Chen, J.; Huo, X. Theoretical Results on Sparse Representations of Multiple-Measurement Vectors. IEEE Trans. Signal Process. 2006, 54, 4634–4643. [Google Scholar] [CrossRef] [Green Version]
  60. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  61. Little, M.; Perry, W. Real-time multichannel airborne radar measurements. In Proceedings of the the 1997 IEEE National Radar Conference, Syracuse, NY, USA, 13–15 May 1997; pp. 138–142. [Google Scholar]
Figure 1. The side−looking airborne phased−array radar with a ULA.
Figure 1. The side−looking airborne phased−array radar with a ULA.
Remotesensing 14 03800 g001
Figure 2. The structure of BAMCV. (a) Generative model of BAMCV. (b) Inference network of BAMCV. The red arrows denote the data generation. The green arrows denote the information propagation in the inference network. The black arrows with imaginary lines represent reparameterization tricks.
Figure 2. The structure of BAMCV. (a) Generative model of BAMCV. (b) Inference network of BAMCV. The red arrows denote the data generation. The green arrows denote the information propagation in the inference network. The black arrows with imaginary lines represent reparameterization tricks.
Remotesensing 14 03800 g002
Figure 3. Flow chart of BAMCV−STAP.
Figure 3. Flow chart of BAMCV−STAP.
Remotesensing 14 03800 g003
Figure 4. CNR versus range cell of D t e s t .
Figure 4. CNR versus range cell of D t e s t .
Remotesensing 14 03800 g004
Figure 5. Estimated angle−Doppler clutter spectrum after pre−training. (a) Clairvoyant spectrum calculated by known CCM. (b) Estimated clutter spectrum using 128 training samples.
Figure 5. Estimated angle−Doppler clutter spectrum after pre−training. (a) Clairvoyant spectrum calculated by known CCM. (b) Estimated clutter spectrum using 128 training samples.
Remotesensing 14 03800 g005
Figure 6. Estimated angle−Doppler clutter spectrum using different methods. (a) Clairvoyant angle−Doppler clutter spectrum. (b) LSMI. (c) M−OMP. (d) M−SBL. (e) M−FCSBL. (f) Clutter spectrum via MCV. (g) BAMCV after pre−training. (h) BAMCV after fine−tuning.
Figure 6. Estimated angle−Doppler clutter spectrum using different methods. (a) Clairvoyant angle−Doppler clutter spectrum. (b) LSMI. (c) M−OMP. (d) M−SBL. (e) M−FCSBL. (f) Clutter spectrum via MCV. (g) BAMCV after pre−training. (h) BAMCV after fine−tuning.
Remotesensing 14 03800 g006
Figure 7. Analysis of clutter ridge estimated by different methods.
Figure 7. Analysis of clutter ridge estimated by different methods.
Remotesensing 14 03800 g007
Figure 8. SINR loss curves. (a) SINR loss curves of all comparison methods. (b) SINR loss curves of Bayesian STAP methods.
Figure 8. SINR loss curves. (a) SINR loss curves of all comparison methods. (b) SINR loss curves of Bayesian STAP methods.
Remotesensing 14 03800 g008
Figure 9. SINR loss versus the number of training samples.
Figure 9. SINR loss versus the number of training samples.
Remotesensing 14 03800 g009
Figure 10. Analysis of convergence and computational loads. (a) Convergence curves of different methods using simulated data. (b) Comparison of computational loads.
Figure 10. Analysis of convergence and computational loads. (a) Convergence curves of different methods using simulated data. (b) Comparison of computational loads.
Remotesensing 14 03800 g010
Figure 11. Estimated clutter spectrum of MCARM data. (a) LSMI. (b) M−OMP. (c) M−SBL. (d) M−FCSBL. (e) MCV. (f) BAMCV.
Figure 11. Estimated clutter spectrum of MCARM data. (a) LSMI. (b) M−OMP. (c) M−SBL. (d) M−FCSBL. (e) MCV. (f) BAMCV.
Remotesensing 14 03800 g011
Figure 12. Average detection results of measured data. (a) PD versus SNR ( f d = 0.1 , L = 10 ). (b) PD versus normalized Doppler frequency (SNR = 20 dB, L = 10 ). (c) PD versus the number of training samples ( f d = 0.1 , SNR = 20 dB). f d denotes the normalized Doppler frequency of targets. L is the number of training samples.
Figure 12. Average detection results of measured data. (a) PD versus SNR ( f d = 0.1 , L = 10 ). (b) PD versus normalized Doppler frequency (SNR = 20 dB, L = 10 ). (c) PD versus the number of training samples ( f d = 0.1 , SNR = 20 dB). f d denotes the normalized Doppler frequency of targets. L is the number of training samples.
Remotesensing 14 03800 g012
Figure 13. Convergence curves of different methods using measured data.
Figure 13. Convergence curves of different methods using measured data.
Remotesensing 14 03800 g013
Table 1. Parameters in the inference network.
Table 1. Parameters in the inference network.
ParametersDimensionsParametersDimensionsParametersDimensions
C μ r K × K g 1 h 2 M N × 1 g a K × 1
C μ i K × K C a K × K g b K × 1
C e K × K C b K × K g c 1 × 1
C 1 h 2 M N × K C c 1 × 2 M N g d 1 × 1
g μ r K × 1 C d 1 × 2 M N g 2 h K × 1
g μ i K × 1 C 2 h K × K g β 2 M N × 1
g e K × 1 C β 2 M N × 2 M N
Table 2. Summary of Performance Metrics.
Table 2. Summary of Performance Metrics.
Performance MetricSimulated DataMeasured Data
Clutter SuppressionSINR loss
PD
Real−time ProcessingConvergence rate
Computational loads
Table 3. Radar System Parameters of Simulated data.
Table 3. Radar System Parameters of Simulated data.
ParametersValueParametersValue
Carrier frequency (Hz)1.25 GPlatform velocity (m/s)125
Bandwidth (Hz)2.5 MPlatform height (m)6000
Mainbeam azimuth (∘)0Pulse number in one CPI8
Mainbeam elevation (∘)0Antenna elements number8
Pulse repetition frequency (Hz)2000Range cell number400
Table 4. Noise power estimation.
Table 4. Noise power estimation.
True Noise PowerM−SBLM−FCSBLMCVBAMCV
0.01 2.49 × 10 10 0.01110.009730.0101
0.1 4.15 × 10 9 0.11970.10610.1103
1 7.92 × 10 8 1.01170.99370.9925
5 2.98 × 10 6 4.76924.97934.9001
10 7.57 × 10 6 11.13829.99929.9856
Table 5. Comparison of running time per iteration using simulated data.
Table 5. Comparison of running time per iteration using simulated data.
ApproachRunning Time Per Iteration (s)
M−SBL 0.125
M−FCSBL 0.055
MCV 0.103
BAMCV 0.011
Table 6. Comparison of computational cost.
Table 6. Comparison of computational cost.
ApproachComputational Loads
M−SBL ( ( K ) 3 + ( M N ) 3 + 3 K 2 M N + [ 2 ( M N ) 2 + 2 M N L + L + 1 ] K + M N L + 1 ) N s
M−FCSBL ( ( M N ) 3 + K 2 M N + 6 K ( M N ) 2 + ( 2 K L + 2 K ) M N + 1 ) N f
MCV ( ( M N ) 3 + 3 K 2 M N + 4 K ( M N ) 2 + ( L 2 + L + 1 ) M N K + M N L 2 + M N L + 1 ) N g
BAMCV 2 K M N L + 5 K 2 L + 4 ( M N ) 2 L + 4 M N L
Table 7. Radar System Parameters of MCARM data.
Table 7. Radar System Parameters of MCARM data.
ParametersValueParametersValue
Pulse repetition frequency (Hz)1984Antenna array spacing of azimuth (m)0.1029
Wavelength (m)0.24Antenna array spacing of elevation (m)0.5629
Pulse number in one CPI128Platform height (m)10,188
Antenna elements number of azimuth11Range cell number400
Antenna elements number of elevation2
Table 8. Comparison of running time per iteration using MCARM data.
Table 8. Comparison of running time per iteration using MCARM data.
ApproachRunning Time Per Iteration (s)
M−SBL 1.102
M−FCSBL 0.333
MCV 0.951
BAMCV 0.025
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, C.; Zhao, H.; Chen, W.; Chen, B.; Wang, P.; Jia, C.; Liu, H. Robust Multiple-Measurement Sparsity-Aware STAP with Bayesian Variational Autoencoder. Remote Sens. 2022, 14, 3800. https://doi.org/10.3390/rs14153800

AMA Style

Zhang C, Zhao H, Chen W, Chen B, Wang P, Jia C, Liu H. Robust Multiple-Measurement Sparsity-Aware STAP with Bayesian Variational Autoencoder. Remote Sensing. 2022; 14(15):3800. https://doi.org/10.3390/rs14153800

Chicago/Turabian Style

Zhang, Chenxi, Huiliang Zhao, Wenchao Chen, Bo Chen, Penghui Wang, Changrui Jia, and Hongwei Liu. 2022. "Robust Multiple-Measurement Sparsity-Aware STAP with Bayesian Variational Autoencoder" Remote Sensing 14, no. 15: 3800. https://doi.org/10.3390/rs14153800

APA Style

Zhang, C., Zhao, H., Chen, W., Chen, B., Wang, P., Jia, C., & Liu, H. (2022). Robust Multiple-Measurement Sparsity-Aware STAP with Bayesian Variational Autoencoder. Remote Sensing, 14(15), 3800. https://doi.org/10.3390/rs14153800

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop