Next Article in Journal
Green Wave Arterial Cooperative Control Strategy Based on Through-Traffic Priority
Previous Article in Journal
Design Optimization of an Enhanced-Mode GaN HEMT with Hybrid Back Barrier and Breakdown Voltage Prediction Based on Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention Mechanism and Neural Ordinary Differential Equations for the Incomplete Trajectory Information Prediction of Unmanned Aerial Vehicles Using Airborne Radar

1
School of Aeronautics, Northwestern Polytechnical University, Xi’an 710072, China
2
Chengdu Aircraft Design and Research Institute, Chengdu 610041, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(15), 2938; https://doi.org/10.3390/electronics13152938
Submission received: 21 June 2024 / Revised: 22 July 2024 / Accepted: 23 July 2024 / Published: 25 July 2024
(This article belongs to the Topic Radar Signal and Data Processing with Applications)

Abstract

:
Due to the potential for airborne radar to capture incomplete observational information regarding unmanned aerial vehicle (UAV) trajectories, this study introduces a novel approach called Node-former, which integrates neural ordinary differential equations (NODEs) and the Informer framework. The proposed method exhibits high accuracy in trajectory prediction, even in scenarios with prolonged data interruptions. Initially, data outside the acceptable error range are discarded to mitigate the impact of interruptions on prediction accuracy. Subsequently, to address the irregular sampling caused by data elimination, NODEs are utilized to transform computational interpolation into an initial value problem (IPV), thus preserving informative features. Furthermore, this study enhances the Informer’s encoder through the utilization of time-series prior knowledge and introduces an ODE solver as the decoder to mitigate fluctuations in the original decoder’s output. This approach not only accelerates feature extraction for long sequence data, but also ensures smooth and robust output values. Experimental results demonstrate the superior performance of Node-former in trajectory prediction with interrupted data compared to traditional algorithms.

1. Introduction

Unmanned aerial vehicles (UAVs) pose a significant risk to the safety of airplanes [1,2,3] when they intrude into no-fly zones, as ground-based airport facilities struggle to detect such diminutive vehicles. In practical applications, airborne radar systems offer a means of detecting and distinguishing UAV formations from varied perspectives [4], making the prediction of UAV trajectories based on limited historical radar data a valuable pursuit [5].
Due to clutter and jamming signals, a radar’s output often contains discontinuities or errors. Swift and accurate prediction of target trajectories is paramount for preventing target loss [6]. Nevertheless, trajectory prediction based on airborne radar faces two major challenges [7,8].
Firstly, traditional filtering algorithms such as the Kalman filter (KF) and extended Kalman filter (EKF) can indeed predict interrupted data, but their application is limited in the case of prolonged interruptions. As the interruption time interval lengthens, the accuracy of the predicted target trajectory plummets. Additionally, to construct a target trajectory from effective data points, methods such as the nearest neighbor (NN) algorithm, probability data association (PDA) algorithm, and joint probability data association (JPDA) algorithm [9,10,11] are frequently employed. However, the performance of these methods heavily relies on the availability of valid data points. To ensure the validity of these points, a validation gate rule is typically applied, where only plots within certain thresholds for range, velocity, and angle are considered as candidates for trajectory formation after association with previous data points. If the validation gate is set too wide, jamming points may sneak in, resulting in significant errors in the predicted target trajectory. Conversely, a narrow validation gate may lead to target loss when the trajectory point fails to associate with new points for an extended period.
Due to the limited extrapolation capabilities of the aforementioned algorithms, their prediction accuracy suffers when historical data are inaccurate. To address this problem, machine learning algorithms are employed for accurate trajectory prediction. For instance, radar measurement data are used to dynamically fine-tune the modeled aircraft quality, enhancing the precision of aircraft trajectory forecasts. Alternatively, in-flight airspace environmental parameters and spatiotemporal attributes are incorporated into machine learning and hidden Markov models to predict trajectories under uncertain conditions [12]. Furthermore, the target trajectory points captured by airborne radar are analyzed as time-series data. Here, LSTM-based algorithms, which employ gate recurrent units (GRUs), demonstrate excellent performance in predicting target trajectories [13,14,15]. Notably, transformer-based algorithms, leveraging the self-attention mechanism, can discern input data points and assign weights based on their mutual relationships and impact on the output.
Apart from these methods, neural network-based approaches offer flexibility, requiring minimal manual parameter adjustment [16]. In comparison to LSTM and GRUs, neural networks excel in multi-step prediction tasks. One such example is the Informer algorithm, a modified transformer variant, tailored for time-series data processing [17]. This algorithm enhances self-attention with prob-sparse self-attention, significantly reducing time complexity from O L 2 to O L log L  [17]. While LSTM, transformers, and their derivatives are proficient in predicting time-series data, maintaining prediction accuracy in the presence of prolonged data errors or interruptions remains a challenge [18].
Secondly, the interruption of data often results in irregular sampling, rendering traditional data prediction methods ineffective. For instance, recurrent neural network (RNN)-based or transformer-based approaches struggle to handle irregular time-series [18,19]. The conventional RNN approach involves dividing time into equal intervals and utilizing averages for input or aggregation, but this pre-processing can distort data information, particularly for real-time measurements that contain significant hidden details [19,20].
To tackle irregularly sampled data, neural ordinary differential equations (neural ODEs) offer exceptional capabilities. Neural ODEs are a family of continuous-time models that utilize deep neural networks to parameterize the derivatives of hidden states [19,20,21]. Through defining the hidden state as the solution to an ODE initial value problem, the hidden state can be computed at any point in time using an ODE solver. Additionally, the adjoint method is proposed to replace the backpropagation algorithm, as the gradient does not directly pass through the ODE solver during forward propagation. This significantly reduces model memory consumption, as only the hidden state of the last time step needs to be stored, eliminating the need to store intermediate values during forward computation [22,23,24].
In this paper, we introduce the Unite Neural Ordinary Differential and Informer (Node-former) method, which demonstrates robust trajectory prediction accuracy, even in the presence of prolonged and significant data perturbations. The methodology can be succinctly outlined in three stages, as shown in Figure 1. Initially, we selectively retain interference data within an acceptable error range, thereby mitigating the adverse effects of data perturbation. Secondly, following the elimination of disrupted data, the remaining effective data are characterized as irregularly sampled. To address this irregularity, we employ the neural ordinary differential equations (NODEs) approach, transforming the computational interpolation into an initial value problem (IPV). This approach effectively circumvents the loss of informative features that can occur with traditional interpolation methods. Finally, we enhance the Informer algorithm’s encoder with prior knowledge to facilitate the trajectory prediction process. To mitigate fluctuations in the original decoder’s output, we utilize an ODE solver as the decoder. Experimental results demonstrate that, compared to other interpolation and trajectory prediction algorithms, our proposed Node-former method significantly improves trajectory prediction performance under prolonged interrupted conditions.
In this paper, we present the following key contributions:
(1) Integrating Neural Ordinary Differential Equations (NODEs) with Attention Mechanism for UAV Trajectory Prediction: This innovative strategy enables us to gracefully manage irregularly sampled data, leveraging NODEs while also broadening our view to encompass global context through the self-attention mechanism. This synergy allows our model to delicately capture temporal fluctuations and long-distance correlations within UAV trajectory data;
(2) Presenting the Node-former Model for Enhancing Radar Interference Management: We propose the Node-former model, a thoughtful three-stage methodology uniquely designed to tackle radar interference in target trajectory data. By mitigating the disruptive effects of interference, our model strives to produce more refined trajectory forecasts. The Node-former model signifies a valuable advancement in bolstering the dependability and accuracy of trajectory prediction amid radar interference scenarios;
(3) Comprehensive Simulation Studies Illustrating Performance Enhancements: Through meticulous simulation experiments, we have demonstrated that our proposed method consistently outperforms conventional trajectory prediction methods across diverse experimental setups. This performance improvement underscores the effectiveness, resilience, and strength of our approach in the realm of UAV trajectory prediction. Our findings not only reinforce the theoretical underpinnings of our work but also underscore its practical relevance and potential for practical deployment.
This paper is organized as follows. In Section 2, we delve into the problem model and analysis for target trajectory prediction by an airborne radar. This section provides a comprehensive understanding of the challenges and complexities involved in accurate trajectory estimation. Section 3 outlines the workflow of the proposed Node-former method and the target trajectory prediction module. We detail the steps involved in the method, highlighting how it addresses the issues mentioned in Section 2. Section 4 presents the simulation results and experimental outcomes, demonstrating the effectiveness and robustness of our Node-former method for trajectory prediction, especially under the conditions of long-time interrupted data. Finally, Section 5 summarizes our work, highlighting the key contributions and implications of our findings. We discuss potential future directions and applications of the Node-former method.

2. Problem Model and Analysis

In this section, we introduce the problem of target trajectory prediction using an airborne radar, specifically focusing on the challenges posed by interrupted data points. As illustrated in Figure 2, when an airborne radar observes a UAV target, ideal echo data can be collected when the jammer on the UAV is powered off. However, when the jammer is activated, the received data become interrupted, leading to significant errors in tracking and trajectory prediction for the UAV target.
To address this issue, we propose a novel method called the Unite Neural Ordinary Differential and Informer (Node-former). This method aims to achieve high accuracy in trajectory prediction, even in the presence of prolonged and strong jamming interference. Leveraging the combined power of neural ordinary differential equations and the Informer architecture, our Node-former method is designed to effectively handle the challenges posed by interrupted data and ensure reliable performance under adverse conditions.
The key challenge in this research is predicting the target trajectory when the target becomes lost within historical track data. To achieve accurate trajectory prediction, it is crucial to first detect and identify any data that have been interfered with. To tackle this problem, we break it down into two distinct steps: handling jamming data and making predictions using historical data.
When describing the target points, the target trajectory without any interruptions can be mathematically expressed as follows:
x ( t ) = ( x 0 ,   x 1 , ,   x N )
where x ( t ) is the function that represents the normal target trajectory data.
As depicted in Figure 2, when the jammer is active, the radar data being observed experience interruptions. In this study, these interruptions are primarily attributed to ambient noise combined with multipath effects. Ambient noise can be modeled using a Gaussian distribution, whereas the reflected waves resulting from multipath effects adhere to a Rayleigh distribution. To detect these interrupted data points, we use range gate pull-off (RPGO) and velocity gate pull-off (VPGO) techniques [25,26,27]. Due to the presence of interrupted data, the radar’s range and speed gates gradually exhibit significant observation errors. Consequently, the interrupted target trajectory can be mathematically expressed as
x ( t ) = ( x 0 ,   x 1 , ,   x N ) + ( j 0 ,   j 1 , ,   j N )
where j ( t ) represents the jamming trajectory data.
The primary reason for losing track of a target, apart from the misdirection of the transmitted beam caused by jamming that drags the radar gates out of alignment, is the target’s maneuvering, which can cause it to move outside the radar’s tracking range. Accurate prediction of a target’s position during a tracking loss hinges on estimating the position with the highest probability of being the target’s current location by analyzing and learning the target’s maneuvering behavior.
As demonstrated in Equation (2), when the target trajectory is interrupted, our approach can detect the effective, non-interrupted data. It is these effective data samples that we rely on for predicting the target trajectory. Following the detection of jamming data, the remaining effective data samples can be expressed as
x ( t ) = ( x 0 ,   x 1 , ,   x k ,   0   , , 0   ,   x k + L , ,   x N )
where N is the effective data length and L is the interrupted length. Data jamming detection identifies invalid samples. Interrupted samples may miss targets; therefore, predicting their trajectories is crucial.

Problem Analysis

In this research, we initially eliminate from the jamming data the radar measurement traces that deviate from the motion pattern of the intended target, thus establishing a dedicated data-filtering module. When jamming persists for extended durations, it becomes inefficient to discard all jammed data; thus, it suffices to preserve as much data as possible while eliminating those with significant errors.
To execute this filtering function, we adopt a neural network architecture comprising multiple linear layers. The key advantage lies in transforming the target’s performance into a simplified classifier. Through training the neural network to predict whether the target falls within a predetermined error range, we eliminate the need for precise target measurements. However, the filtered data no longer adhere to a regular sampling pattern, necessitating a processing algorithm that is capable of handling irregularly sampled data. Therefore, we employ neural ODEs, which excel in handling such irregular sampling, for data processing.
The historical data encompass multifaceted features, including the target’s time, speed, and distance. The prediction task focuses on identifying the feature that significantly impacts the target’s location and assigning weight parameters accordingly. Inspired by human attention, which filters out unimportant information and focuses on priorities to grasp patterns and predict outcomes, we introduce an attention mechanism. This mechanism applies to the prediction process by observing and learning from historical data, allocating attention to specific features to capture movement patterns and predict the target’s trajectory.
This study introduces an enhanced transformer structure tailored specifically for the task of time-series prediction. While the traditional sequence-to-sequence approach can provide a one-time output, it suffers from a limitation known as the long short-term forecast (LSFT) problem. Specifically, the output values in this approach are generated independently, without considering the previously predicted values, resulting in significant data fluctuations.
To address this issue, we modify the original decoder component of the transformer to incorporate an ODE solver. This modification ensures that each output data value is correlated with the preceding values, resulting in a smoother predicted trajectory. Leveraging the ODE solver, we are able to capture the temporal dependencies within the time-series data and generate predictions that are both accurate and continuous.

3. Proposed Method

To address the challenge of predicting target trajectories in the presence of interference data, this section provides an in-depth examination of the structural framework of the proposed Node-former methodology. We delve into the rationale behind the design choices and elucidate the innovative aspects of each constituent component.

3.1. Algorithm Framework

The proposed algorithm encompasses two integral components. The first is the anti-jamming module, tasked primarily with identifying jamming source data, eliminating it following an error assessment, and subsequently calculating interpolation values to compensate for the eliminated data. The second component, the track prediction module, leverages historical information to generate trajectory forecasts. This historical information is compiled by concatenating radar measurement data with interpolated information processed with the anti-jamming module. Figure 3 visually depicts the algorithm’s structural framework. The subsequent sections elaborate on each algorithmic segment in detail.

3.2. Data Pre-Processing

The data pre-processing is conducted on the input radar’s original measurement data. The radar can identify whether the measurement information is jammed at this time by detecting the channel and then giving a jamming flag. Through the jamming flags, the normal radar measurement data can be separated from the jamming data. As shown in Figure 4, radar jamming data will be sent to the jamming data filter module for processing in the next step.

3.3. Jamming Data Filtering

In addressing the jamming data, this study employs a neural network with multiple linear layers to filter out points with significant errors, retaining measurements with minor errors for use as observations. An error threshold α is set for the selected column vectors, and observations within this threshold are deemed approximations of the radar’s internal anti-jamming strategy’s true values. For irregularly sampled jamming datasets, data exceeding the α threshold are set to null (NA). This is illustrated in Figure 5.
Here, the start and end positions of the jamming data segment are chosen from the radar’s normal measurement state. This approach ensures that, even if most or all jamming data values exceed the error range, the first and last values can serve as endpoints, facilitating the anti-jamming module’s operation and enhancing interpolation accuracy.
Notably, this methodology mitigates the issue of poor historical data quality stemming from prolonged data disturbances, as correcting each individual value in such cases can be challenging.

3.4. Anti-Jamming Module

After the jamming data filtering module, the resulting data exhibit irregular sampling. Given the challenges that the RNN and transformer architectures face with such data, this study proposes using neural ODEs to address this issue. Neural ODEs decode network hidden variables and model dynamics under time continuity, and interpolate missing data points.
This study primarily employs an anti-jamming structure based on neural ODEs (NODEs), which can be interpreted as ResNets with an infinite number of layers [28]. The hidden state h t is formulated as an ODE initial value problem in a continuous time-series model, defined by ordinary differential equations, as shown in Equations (4) and (5) [19,21,22]:
h t + 1 = h t + f ( h t + θ t )
d d t h t = f h t ,   t ,   θ
where f is the function that represents the hidden state dynamics through the neural network parameter θ .
ODE–RNN is a generalized application of RNN to neural differential equations. In Equation (6), the hidden states between observations are defined as
h i = O D E S o l v e f θ ,   h i 1 , t i 1 , t i
RNN is used to update the hidden state of the intermediate process based on the following observations:
h i = R N N C e l l h i ,   x i
The hidden state can be calculated at any expected moment using the ODE solver by iterating Equations (6) and (7), as shown in Equation (8):
h 0 , , h N = O D E S o l v e f θ ,   h 0 , t 0 , , t N
In this study, we enhance the encoder–decoder structure using ODE–RNN, which combines ODE–LSTM as the encoder and an ODE as the decoder. ODE–LSTM updates hidden states between observations, while the ODE handles interpolation between them. To approximate the posterior at t 0 , we propagate the ODE–LSTM encoder backward in time from t N to t 0 , yielding an estimate of the initial state’s posterior q. Solving the ODE, we can obtain the potential state at any point of interest, as illustrated in Figure 6.
According to the time-series model generated using the ODE definition, the initial hidden state determines the following for the entire track:
z 0 p z 0
z 0 , ,   z N = O D E S o l v e f ,   z 0 , t 0 , , t N
x i p x i | z i ,   i = 0 ,   1 , N
where z 0 is represents the initial state randomly drawn from the probability distribution p ( z 0 ) . x i represents the observation conditioned on the latent variable z i .
The mean and standard deviation of the approximate posterior parameters q as a function of the final hidden state concerning the ODE–LSTM are given as follows:
q z 0 | x i , t i i = 0 N = N μ h 0 , σ h 0
μ z 0 , σ z 0 = g O D E _ L S T M x i , t i i = 0 N
where g is the neural network that transforms the mean and variance of the final hidden state in the encoder.
The encoder and decoder are trained by maximizing the evidence lower bound (ELBO) [21,28] using the following equation:
E L B O θ , ϕ = E z 0 q ϕ z 0 | x i , t i i = 0 N log p θ x 0 , , x N KL [ q ϕ z 0 | x i , t i i = 0 N | | p z 0 ]
The ODE solver is a fifth-order solver with adaptive steps. It is found in the torchdiffeq Python package.
The radar measurement data of multiple dimensions are encoded as latent variable z 0 by ODE–RNN, and z 0 is entered into the feed-forward network g to obtain its μ z 0 and σ z 0 . The distribution of z 0 is obtained by N ( μ z 0 , σ z 0 ) , and the latent variable z 0 is obtained next. z i is calculated by the solution O D E S o l v e ( f , z 0 , ( t 0 , , t N ) ) , and the predicted value is found by decoding z i .
The anti-jamming module processes jamming data as shown by the flow in Figure 7.

3.5. Data Contact

The next step is to collocate the anti-jamming processed data in the previous section with the normal radar measurement data. According to the time stamp of each data point, the data corrected by the anti-jamming module are concatenated with the normal measurement data to form the historical data as the input of the prediction module.

3.6. Track Prediction Module

Track prediction is the most important part of this study. For time-series data prediction, this study combines the transformer algorithm architecture with the nodes algorithm connotation. Transformer architecture is more suitable for LSFT problems, and, here, an improved self-attention is applied to effectively reduce the computational complexity of the model [29,30].

3.6.1. Self-Attention

In the self-attention mechanism, the input signals are classified into a query (Q), key (K), and value (V). For a given query, the attention weight (a) corresponding to the k i in the self-attention mechanism is obtained according to the correlation a ( q , k i ) between Q and k i , as follows [16,30,31]:
a ( q , k i ) = q · k i q × k i
On the other hand, the self-attention formula uses the result in Equation (14) to operate on the pair with the corresponding V to obtain the attention output, as shown in Equation (16), which is collated to obtain Equation (15), with d being the dimensions of Q and K:
f q , k 1 , v 1 , k n , v n = j = 1 n S o f t m a x ( a ( q , k i ) ) · v i
A Q , K , V = S o f t m a x Q K T d V
Based on the formula derivation and experiments presented in the relevant literature [17], we can understand that a minority of query and key determines the result of Equation (17), which means that the output in the sequence has a high correlation with only a few inputs. Through randomly sampling the query-wise score from the attention map, a method is defined to select Q. The Q is sorted into active query and lazy query. The probability sparse is found by sampling K to obtain the K-sample, and then finding each q i in Q about the K-sample to find the M ( q i , K ) , as shown in Equation (18) [17,32]:
M q i , K = ln j = 1 L k exp q i k j T d 1 L k j = 1 n q i k j T d
where L K represents the length of the sequence acquired this time. The first half of the formula represents the value of q i on all keys; the latter is the arithmetic mean. More than the arithmetic mean of q i is regarded as an active query, and the rest is regarded as a lazy query.
Setting upper and lower bounds on M ( q i , K ) , Equation (18) can be approximated to derive Equation (19):
M ¯ q i , K = max j q i k j T d 1 L k j = 1 n q i k j T d
where M ¯ q i , K represents the query sparsity measurement, which can be used to find Q ¯ .
For track prediction, because of the correlation between track points, more attention should be assigned to the track points that are closer to the moment of target loss. Therefore, to help improve the allocation of attention, this research introduces a time-series-based residual formula term for the attention scores. Based on the above derivation, a new ProSparse self-attention is acquired. Equation (20) [12,22,33] is as follows:
A Q , K , V = S o f t m a x Q ¯ K T d + ln ( 1 t N t 1 ) f ( t N ) V
where Q ¯ is composed of the top-u q i selected by Equation (18) based on the size of the calculation result and the other unselected q i (the unselected q i are directly given by the mean value) and f is the fitting function obtained from the time residuals. Thus, the space complexity O ( L 2 ) generated by the self-attention mechanism is reduced to O ( L ln L ) .

3.6.2. Knowledge Distillation

A knowledge distillation mechanism is introduced to reduce the computation time by effectively saving memory overhead to halve the sequence length. The layer j to j + 1 distillation operation is shown in Equation (21) [12]:
X j + 1 i = MaxPool ELU Conv 1 d X j i AB
where MaxPool represents the maximum pooling, ELU is the activation function, Conv1d represents the one-dimensional convolution operation, and [ X j i ] A B represents the ProSparse self-attention operation with multi-head attention.
The historical data in Figure 8 are encoded by the encoder and fed into the decoder composed of the ODE solver, which then outputs the predicted data after a fully connected layer, as shown in Figure 9.
The prediction module in this study improves the encoder by introducing a time function variable to adjust the attention weights to accelerate the convergence speed. Additionally, it uses the ODE solver as the decoder so that the output data values can make full use of the previous prediction values to ensure that the trajectory change curve is smooth.

3.7. Implementation Details of the Node-Former Algorithm

To facilitate the understanding of the implementation details of Node-former, we provide a pseudocode outlining the key steps involved in both the training and testing phases of the algorithm. The training pseudocode Algorithm 1 details the process of optimizing the model parameters to fit the training data, while the testing pseudocode Algorithm 2 outlines how the trained model is utilized to make predictions on loss data. These pseudocodes, presented below, offer a concise yet comprehensive view of the Node-former methodology.
Algorithm 1: The training pseudocode of Node-former
Input: Jamming radar training dataset D , jamming filtering network β , anti-jamming network θ , track prediction network ϕ , update epochs for jamming data filtering phase K 1 , update epochs for anti-jamming phase K 2 , update epochs for track prediction phase K 3
  1// Jamming Data Filtering Training Phase
  2for  u   p d a t e   e p o c h   k = 1 K 1  do
  3 Sample a minibatch sample pairs X from D
  4 Divide the sample pairs X into two parts: X J a m and X N o r m , as described in Section 3.2
  5 Train the filtering network β using the divided parts X J a m , as described in Section 3.3
  6end
  7// Anti-jamming Training Phase
  8for  u   p d a t e   e p o c h   k = 1 K 2  do
  9 Sample a minibatch sample pairs X from D and divide into two parts: X J a m and X N o r m
10 Apply the trained filtering network β to X J a m and get the irregular sample pairs X J a m f i l t e r
11 Train the anti-jamming network θ using the irregular pairs X J a m f i l t e r , as described in Section 3.4
12end
13// Track Prediction Training Phase
14for  u   p d a t e   e p o c h   k = 1 K 3  do
15 Sample a minibatch sample pairs X from D and divide into two parts: X J a m  and X N o r m
16 Apply the trained filtering network β and the trained anti-jamming network θ  to X J a m
17 Output the correctd sample pairs X C o r and contact with normal data X N o r m to get X
18 Train the track prediction network ϕ using the new sample pairs X , as described in Section 3.6
19end
Algorithm 2: The testing pseudocode of Node-former
Input: Jamming radar testing dataset D , jamming filtering network β , anti-jamming network θ , track prediction network ϕ , filtering threshold α
  1Sample a test sample X t e s t from D
  2Divide the sample X into two parts: X J a m and X N o r m
  3// Jamming Data Filtering Testing Phase
  4Apply the trained filtering network β to the divided part X J a m
  5Following the filtering threshold α to select X J a m and output the irregular data  X J a m f i l t e r
  6// Anti-jamming Testing Phase
  7Apply the trained anti-jamming network θ to the irregular input X J a m f i l t e r
  8Output the corrected sample X C o r and contact with normal part X N o r m to obtain final input X
  9// Track Prediction Testing Phase
10Apply the trained track prediction network ϕ to the final input X
11Output the subsequent predicted part X p r e d

4. Presentation of Results

This section describes in detail the design of the simulation experiment, the logic of the program operation, and the comparison of the experimental results. Here, the aircraft’s radar was used instead of the airport ground radar, and the jamming data were created by jamming the radar function through the jammer.

4.1. Simulation Design

Firstly, we tested the robustness of the algorithm in terms of its anti-jamming capabilities while tracking the target, as given equation by Equation (22) [34,35,36]:
x t + d t = 1 sin ( ω t Δ t ) ω t 0 cos ( ω t Δ t ) 1 ω t 0 0 cos ( ω t Δ t ) 0 sin ( ω t Δ t ) 0 0 1 cos ( ω t Δ t ) ω t 1 sin ( ω t Δ t ) ω t 0 0 sin ( ω t Δ t ) 0 cos ( ω t Δ t ) 0 0 0 0 0 1 x t + u t
where x t represents the vector that encompasses position, velocity, and rotation rate. u t represents the process noise.
The state vector x t = Δ [ a t , a ˙ t , b t , b ˙ t , ω t ] contains the position ( a t , b t ) , velocity ( a ˙ t , b ˙ t ) , and rotation rate ω t of the tracked target. The initial state of the tracked target was x 0 = Δ [ 10,000 m , 100 m / s , 50,000 m , 300 m / s , 0.0053 rad / s ] . The measurement equation is given by Equation (23):
y t + d t = r t + d t θ t + d t = a t + d t 2 + b t + d t 2 atan ( b t + d t , a t + d t ) + v t + d t
y t represents the vector that encompasses r t + d t and θ t + d t , where r t + d t and θ t + d t , respectively, represent the slant distance and azimuth angle of the target. v d t represents the measurement noise.
The state equation generates a trajectory of motion, and we conducted a comparative analysis of the Node-former algorithm and the Kalman filter algorithm by manipulating the proportion of outliers in the observed data within the measurement equation. This comparison aimed to quantitatively evaluate the performance differences between the two algorithms in handling data with uncertainties and noise.
Next, we subjected the algorithm to a more complex environment and chose a simulated air combat game. The data collected were generated by the DCS (Digital Combat Simulator) world, and the aircraft simulation model consisted of the six-degree-of-freedom motion model and the radar model. We modeled jamming data through using a jammer to interfere with the radar.
In the simulation program, the engagement area and preparation area of both sides were set, and the initial slant distance of both sides of the aircraft in each field was 20 km. The initial altitude was 500 m, and the aircraft was generated at any position in its respective preparation area. The red aircraft’s radar was used as an airport monitoring radar, and the blue aircraft was a drone that was about to break into the no-fly zone.
The red aircraft formation is tasked with driving away the blue aircraft formation. Figure 10 illustrates the approaching of both formations. Figure 11 showcases an operation where the red formation successfully drives away the blue formation. In Figure 11a, the red formation’s radar tracks the blue formation and launches missiles, prompting the blue formation to initiate evasive maneuvers. In Figure 11b, the blue formation activates ECMs (Electronic Countermeasures) to prevent being hit. In Figure 11c, one blue aircraft is shot down, while the other is successfully driven away from the no-fly zone in Figure 11d. Throughout this process, to maximize the number of target tracks, any blue aircraft entering the Field of View (FoV) of any red aircraft’s radar is detected. Figure 12 depicts a radar interface, where, upon encountering interference, the displayed distance may experience fluctuations, and the affected targets are highlighted in yellow as sources of interference.
The specific process shown in Figure 10, Figure 11 and Figure 12 shows the radar interface.
The simulation used the north–east–down (NED) coordinates. The data consisted of the target track measurement output from the radar and the real track of the target according to the NED calculated from the latitude, longitude, and altitude data recorded by the respective INS of the aircraft and target. For the target data R d r R a n g e i , measured by the radar in the current moment, i can be expressed as the equation
R d r m i = R d r R a n g e i , R d r E l i , R d r A z i , R d r V c l o s e i , R d r V r i , J a m _ F l a g
where R d r R a n g e i represents the current moment radar measurement slant distance; R d r A z i represents the current moment’s radar measurement of the target pitch angle under NED coordinates; R d r A z i represents the current moment’s radar measurement of the target azimuth angle under NED coordinates; R d r V c l o s e i represents the current moment’s radar measurement close velocity along the Line of Sight (LOS); R d r R a n g e i represents the current moment’s radar measurement of the target radial velocity along the LOS; and J a m _ F l a g represents the moment that the radar suffers jamming.
The navigation fusion information reported by the Ins can be expressed as I n s i :
I n s i = M y _ R o l l i , M y _ P i t c h i , M y _ V e l _ L o s i
where M y _ R o l l i represents the current moment’s roll angle of the measurement aircraft; M y _ P i t c h i represents the current moment’s pitch angle of the measurement aircraft; and M y _ V e l _ L o s i represents the current moment’s measurement of the aircraft velocity along the LOS.
The target true location information can be expressed as T g t i :
T g t i = R a n g e i , E l i , A z i
where R a n g e i represents the current moment’s true slant distance of both sides; E l i represents the current moment’s true elevation angle; and A z i represents the current moment’s true azimuth angle.

4.2. Mission Flow

In this study, the task was broken into two objectives to be accomplished. The algorithm flow is shown in Figure 13.
The first step was data input and pre-processing. The data in the radar were used as input for the pre-processing and were separated by judging the radar jamming flag.
The second step involved judging when the target state was steadily tracked by the radar. The jammed data were passed through the error judgment evaluation module, and the data output when the radar anti-jamming strategy was judged to be effective was selected, so that the data with smaller error values were sampled in the jamming data.
In the third step, the jamming data sampled with the judgment module were input into the module of error correction to form the corrected radar target data, which were stitched together with the normal, non-jammed radar measurement data to form the target data. Here, the judgment was made whether the target was lost at the current moment. If the target was still in the tracking state, the target data were sent directly to the user interface for display, and, if the target was lost, the next step was taken.
In the fourth step, the target data for some time before the target was lost were sent into the track prediction module to predict the target track for the next period, and the predicted target information was sent to the user interface for display.

4.3. Simulation Experiments

All the experiments were conducted in an environment configured as shown in Table 1.
Initially, the experiment was conducted to generate a motion trajectory using Equation (22). Following this, the trajectory was observed using Equation (23), with a step size of 500. The acquired measurement data were then subjected to a thorough analysis using three distinct algorithms: the Node-former algorithm, the Kalman algorithm, and the Robust-Kalman algorithm. This multi-algorithm approach aimed at providing a comprehensive trajectory analysis. The experimental workflow, incorporating these three algorithms, is visually depicted in Figure 13.
In this experiment, we replaced the perturbed data with outliers. Figure 14 and Figure 15 demonstrates the effectiveness of the three algorithms under different ratios of outliers.
As can be seen from the trajectory and measurement plots in Figure 14a, the performance of the three algorithms was basically the same when there were no outliers and only noise.
As the ratio of outliers increased, as illustrated in Figure 15, we observed a gradual decline in the accuracy of the measured distance and angle. In this context, the Kalman filter algorithm faced challenges in efficiently recognizing and accommodating these outlier points, leading to less-than-optimal tracking performance. Similarly, while the Robust-Kalman algorithm demonstrated an improved ability to handle a higher number of outliers compared to the traditional Kalman algorithm, it too exhibited limitations as the outlier percentage escalated. Ultimately, when the outlier ratio reached 70%, the Robust-Kalman algorithm struggled to maintain effective functionality.
In comparison, our Node-former algorithm showed promising robustness, indicating that its anti-interference capabilities were functioning well. Although we recognize that further improvements are always possible, the Node-former algorithm consistently demonstrated a stronger capacity to process trajectory data amidst high outlier levels, compared to both the Kalman and Robust-Kalman algorithms. This finding underscores the potential of our algorithm in deal with complex and noisy datasets.
Subsequently, we implemented the Node-former algorithm in a more intricate environment and undertook a comparative analysis with several prevalent deep learning algorithms. Through the DCS game, 3000 data sets were generated. This study used 80% of the data as the training set and 20% as the test set.
The simulation set the time interval of each radar measurement target information report to 10 ms. The period of the target data reported by the radar was 0.1 s, the flight altitude was from 100 m to 1500 m, and the maximum speed was 140 km/h.
During training, the batch size was set to 50, the learning rate to 1 × 10 4 , the weight decay to 1 × 10 5 , the input size to 13, and the output size to 3, according to Equations (18)–(20), respectively, for the range, the azimuth angle, and the pitch angle of the target.
The hidden dim of the LSTM algorithm was set to 256, the number of layers was set to 2, and the LSTM dims were set to 256. The embedding dim of the transformer algorithm was set to 128, the number of multi-heads was set to 8, and the number of encoder layers was set to 12.
The screening threshold of the jamming data filter module in the Node-former algorithm was set to 0.5 , the embedding dim of the Node-former was set to 128, the number of multi-heads was set to 8, and the number of encoder layers was set to 12.
After training, the measurement data from the radar in the test set were fed into the algorithm to obtain the results. In this study, the anti-jamming and prediction capabilities of the network were evaluated separately. There were some charts of errors from the test set. Figure 16 and Figure 17 show the errors of the anti-jamming phase and the error curve of the prediction phase in some scenarios from the test set, respectively.
The results of the LSTM output show that, although LSTM could suppress jamming based on the experience obtained from training, the long-time large error data input still slowly affected the output results. The transformer was less sensitive to jamming data than the LSTM, but its output data values varied more per frame, which were less stable compared to the LSTM. The Node-former algorithm had the advantages of both: the accuracy of the output data was more insensitive to the jamming data and the error curve was smooth.
As can be seen from Figure 17, for long sequences of prediction, radar extrapolation could not accurately predict the location of the target. LSTM also began to gradually diverge with the increase in the number of time steps. Case results for the transformer algorithm were better compared to LSTM, but its output error fluctuations were still large. In engineering applications, data fluctuations can be detrimental to the operator. The prediction results of the Node-former algorithm had better convergence, while the output data curves were smooth and conformed to the target motion pattern.
The LSTM, transformer, and Node-former algorithms were tested separately. The experimental results can be found in Table 2.
Table 2 shows that LSTM was better at processing the target distance information and the transformer algorithm was better at processing the target angle information. The Node-former architecture was better at processing both the target range and the azimuth pitch angle. The Node-former range RMSE was better than that for LSTM, and the azimuth and pitch angle RMSEs were better than those for the transformer architecture, proving that the algorithm architecture has advantages in processing such data.
In summary, Node-former improved the azimuth RMSE to 0.555 , pitch RMSE to 0.015 , and distance RMSE to 69.13 m in the anti-jamming phase. In the prediction phase, Node-former improved the azimuth RMSE to 0.140 , pitch RMSE to 0.059 , and distance RMSE to 95.85 m.
Figure 18 shows the comparison of the track generated after inputting the Node-former prediction data into the simulation with the real track of the target. It can be seen from the figure that the track output by the algorithm structure matched the real track and can therefore satisfy the requirement for early removal of a target intending to enter a no-fly zone.

5. Conclusions

Predicting the trajectories of UAVs that may enter no-fly zones can be challenging. This study explored the problem of target track prediction in target loss scenarios when the radar is subjected to jamming. The main conclusions are as follows:
(1) Experiments have proven that Node-former exhibits stronger anti-jamming performance compared to conventional Kalman filter, Robust Kalman filter, LSTM, and Transformer algorithms, regardless of whether it is applied to 2D or 3D trajectory tracking. It employs a novel approach that involves interpolating irregularly sampled data after filtering out the tracking error predictions derived from jamming data through the Nodes algorithm;
(2) The RNN algorithm over-allocates weights to feature dimensions of a large order of magnitude based on the loss function, while the transformer algorithm is more balanced in assigning weights. The RNN output is smooth in the track information, while the transformer output drastically changes the track information. Node-former takes into account the advantages of both, as it inherits the advantages of the transformer algorithm in assigning weights. Additionally, to avoid the track points from jumping during track prediction, the decoder in Node-former uses an ODE solver to make the output data more stable and continuous;
(3) This article provides a three-stage methodology to solve the problem of target trajectory prediction based on interference data scenarios. It can replace each unit module in the form of components while maintaining the architecture in an unchanged manner. This provides room for improvement in adopting cutting-edge algorithms and is easy to implement in engineering applications.
We acknowledge that our proposed approach, while effective in addressing the challenges of anti-jamming and trajectory prediction for UAVs, does have certain limitations. One primary limitation lies in its relatively high computational complexity compared to traditional methods like Kalman filters, which can translate into increased resource consumption. To mitigate this issue and make our approach more efficient, we plan to explore lighter versions of our algorithm in future work. Specifically, we aim to incorporate sparse or frequency-based attention mechanisms, which have shown promise in optimizing performance while reducing computational demands. These mechanisms could contribute to a more resource-conscious implementation, enhancing the practicality of our method. Furthermore, we recognize that our method may encounter difficulties in generalizing to scenarios that it has not been explicitly trained on, indicating a limitation in its generalization ability. To improve this aspect, we plan to incorporate a wider variety of baseline algorithms into our future work. By comparing our approach against these diverse baselines, we can gain a more comprehensive understanding of its strengths and weaknesses and identify potential areas for improvement. Additionally, we intend to test our algorithm in more complex environments, exposing it to a broader range of conditions and challenges. This rigorous testing will help us evaluate the robustness of our method and identify strategies to further enhance its generalization capabilities. By incorporating these enhancements, we aim to make our approach more versatile and reliable, i.e., better suited for real-world applications.

Author Contributions

H.P. contributed to methodology, software, and original draft; W.Y. contributed to review and editing, writing, and supervision; Z.W. contributed to review and editing, writing, and resources; R.C. contributed to review and editing and supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to issues pertaining to the intellectual property rights ownership of the dataset.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Choi, D.; Yim, J.; Baek, M.; Lee, S. Machine learning-based vehicle trajectory prediction using v2v communications and on-board sensors. Electronics 2021, 10, 420. [Google Scholar] [CrossRef]
  2. Wu, Y.; Yu, H.; Du, J.; Liu, B.; Yu, W. An Aircraft Trajectory Prediction Method Based on Trajectory Clustering and a Spatiotemporal Feature Network. Electronics 2022, 11, 3453. [Google Scholar] [CrossRef]
  3. Zhang, X.; Liu, Y.; Zhang, Y.; Guan, X.; Delahaye, D.; Tang, L. Safety assessment and risk estimation for unmanned aerial vehicles operating in national airspace system. J. Adv. Transp. 2018, 2018, 4731585. [Google Scholar] [CrossRef]
  4. Mao, D.; Yang, J.; Zhang, Y.; Huo, W.; Xu, F.; Pei, J.; Zhang, Y.; Huang, Y. Angular superresolution of real aperture radar with high-dimensional data: Normalized projection array model and adaptive reconstruction. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5117216. [Google Scholar] [CrossRef]
  5. Zhang, Y.; Zhang, Q.; Zhang, Y.; Pei, J.; Huang, Y.; Yang, J. Fast split bregman based deconvolution algorithm for airborne radar imaging. Remote Sens. 2020, 12, 1747. [Google Scholar] [CrossRef]
  6. Wang, Y.; Wang, J.; Fan, S.; Wang, Y. Quick intention identification of an enemy aerial target through information classification processing. Aerosp. Sci. Technol. 2023, 132, 108005. [Google Scholar] [CrossRef]
  7. Mao, D.; Zhang, Y.; Pei, J.; Huo, W.; Zhang, Y.; Huang, Y.; Yang, J. Forward-looking geometric configuration optimization design for spaceborne-airborne multistatic synthetic aperture radar. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8033–8047. [Google Scholar] [CrossRef]
  8. Zhang, H.; Chen, H.; Zhang, W.; Zhang, X. Trajectory Planning for Airborne Radar in Extended Target Tracking Based on Deep Reinforcement Learning. Digit. Signal Process. 2024, 153, 104603. [Google Scholar] [CrossRef]
  9. Bar-Shalom, Y.; Daum, F.; Huang, J. The probabilistic data association filter. IEEE Control Syst. Mag. 2009, 29, 82–100. [Google Scholar]
  10. Xiao, C.; Yaan, L.; Yuxing, L.; Li, X. A Novel Probabilistic Data Association for Target Tracking in a Cluttered Environment. Sensors 2016, 16, 2180. [Google Scholar] [CrossRef]
  11. Wang, S.; Bi, D.; Ruan, H.; Du, M. Radar maneuvering target tracking algorithm based on human cognition mechanism. Chin. J. Aeronaut. 2019, 32, 1695–1704. [Google Scholar] [CrossRef]
  12. Kandati, D.R. Introduction to Machine Learning; Duke University: Durham, NC, USA, 2021. [Google Scholar]
  13. Che, Z.; Purushotham, S.; Cho, K.; Sontag, D.; Liu, Y. Recurrent Neural Networks for Multivariate Time Series with Missing Values. Sci. Rep. 2018, 8, 6085. [Google Scholar] [CrossRef] [PubMed]
  14. Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  15. Box, B.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2016. [Google Scholar]
  16. Beltagy, I.; Peters, M.E.; Cohan, A. Longformer: The Long-Document Transformer. arXiv 2020, arXiv:2004.05150. [Google Scholar]
  17. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. Proc. Aaai Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  18. Habiba, M.; Pearlmutter, B.A. Neural Ordinary Differential Equation based Recurrent Neural Network Model. In Proceedings of the 2020 31st Irish Signals and Systems Conference (ISSC), Letterkenny, Ireland, 11–12 June 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  19. Rubanova, Y.; Chen, R.; Duvenaud, D. Latent ODEs for Irregularly-Sampled Time Series. arXiv 2019, arXiv:1907.03907. [Google Scholar]
  20. Kidger, P.; Morrill, J.; Foster, J.; Lyons, T. Neural Controlled Differential Equations for Irregular Time Series. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, BC, Canada, 6–12 December 2020. [Google Scholar]
  21. Kidger, P. On Neural Differential Equations. arXiv 2022, arXiv:2202.02435. [Google Scholar]
  22. Chen, R.; Rubanova, Y.; Bettencourt, J.; Duvenaud, D.K. Neural Ordinary Differential Equations. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  23. Cho, K.; Van Merrienboer, B.; Bahdanau, D.; Bengio, Y. On the Properties of Neural Machine Translation: Encoder-Decoder Approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  24. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to Sequence Learning with Neural Networks. Adv. Neural Inf. Process. Syst. 2014, 2, 3104–3112. [Google Scholar]
  25. Schleher, D.C. Introduction to Electronic Warfare; Artech House: Dedham, MA, USA, 1986. [Google Scholar]
  26. Cardillo, E.; Cananzi, R.; Vita, P.; Caddemi, A. Dual-conversion microwave down converter for nanosatellite electronic warfare systems. Appl. Sci. 2022, 12, 1524. [Google Scholar] [CrossRef]
  27. Greco, M.; Gini, F.; Farina, A.; Ravenni, V. Effect of phase and range gate pull-off delay quantization on jammer signal. IEE Proc. Radar Sonar Navig. 2006, 153, 454–459. [Google Scholar] [CrossRef]
  28. Dmitry, Y. Error bounds for approximations with deep ReLU networks. Neural Netw. Off. J. Int. Neural Netw. Soc. 2017, 94, 103. [Google Scholar]
  29. Guokun, L.; Chang, W.-C.; Yang, Y.; Liu, H. Modeling long-and short-term temporal patterns with deep neural networks. In Proceedings of the The 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 95–104. [Google Scholar]
  30. Li, S.; Jin, X.; Yao, X.; Zhou, X.; Chen, W.; Wang, Y.-X.; Yan, X. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv. Neural Inf. Process. Syst. 2019, 32, 5243–5253. [Google Scholar]
  31. Liu, Y.; Gong, C.; Yang, L.; Chen, Y. DSTP-RNN: A dual-stage two-phase attention-based recurrent neural network for long-term and multivariate time series prediction. Expert Syst. Appl. 2020, 143, 113082. [Google Scholar] [CrossRef]
  32. Neil, D.; Pfeiffer, M.; Liu, S.C. Phased LSTM: Accelerating Recurrent Network Training for Long or Event-based Sequences. Neural Information Processing Systems. Adv. Neural Inf. Process. Syst. 2016, 29, 3889–3897. [Google Scholar]
  33. Chen, R.T.Q.; Behrmann, J.; Duvenaud, D.K.; Jacobsen, J.-H. Residual flows for invertible generative modeling. Adv. Neural Inf. Process. Syst. 2019, 32, 9916–9926. [Google Scholar]
  34. Wang, H.; Li, H.; Fang, J.; Wang, H. Robust Gaussian Kalman filter with outlier detection. IEEE Signal Process. Lett. 2018, 25, 1236–1240. [Google Scholar] [CrossRef]
  35. Wang, H.; Li, H.; Zhang, W.; Zuo, J.; Wang, H.; Fang, J. Outlier-detection-based robust information fusion for networked systems. IEEE Sens. J. 2022, 22, 22291–22301. [Google Scholar] [CrossRef]
  36. Piche, R.; Sarkka, S.; Hartikainen, J. Recursive outlier-robust filtering and smoothing for nonlinear systems using the multivariate Student-t distribution. In Proceedings of the 2012 IEEE International Workshop on Machine Learning for Signal Processing, Santander, Spain, 23–26 September 2012; IEEE: Piscataway, NJ, USA, 2012. [Google Scholar]
Figure 1. Three-stage mission.
Figure 1. Three-stage mission.
Electronics 13 02938 g001
Figure 2. Observation geometric structure.
Figure 2. Observation geometric structure.
Electronics 13 02938 g002
Figure 3. Node-former algorithm structure.
Figure 3. Node-former algorithm structure.
Electronics 13 02938 g003
Figure 4. Data pre-processing.
Figure 4. Data pre-processing.
Electronics 13 02938 g004
Figure 5. Jamming-data filter.
Figure 5. Jamming-data filter.
Electronics 13 02938 g005
Figure 6. ODE–LSTM structure.
Figure 6. ODE–LSTM structure.
Electronics 13 02938 g006
Figure 7. Anti-jamming structure.
Figure 7. Anti-jamming structure.
Electronics 13 02938 g007
Figure 8. Data contact structure.
Figure 8. Data contact structure.
Electronics 13 02938 g008
Figure 9. Track prediction structure.
Figure 9. Track prediction structure.
Electronics 13 02938 g009
Figure 10. Two-aircraft, one-circle fight tactic.
Figure 10. Two-aircraft, one-circle fight tactic.
Electronics 13 02938 g010
Figure 11. Wheel tactics.(a) Red Team and the Blue Team are approaching each other; (b) The Red Team initiates the eviction of the Blue Team, and launches missiles; (c) The Blue Team’s aircraft release jamming signals, the radar of the Red Team’s aircraft to start losing track of the Blue Team’s trajectory information; (d) One of the Blue Team’s aircraft is shot down by the Red Team, while the other Blue Team aircraft leaves the no-fly zone.
Figure 11. Wheel tactics.(a) Red Team and the Blue Team are approaching each other; (b) The Red Team initiates the eviction of the Blue Team, and launches missiles; (c) The Blue Team’s aircraft release jamming signals, the radar of the Red Team’s aircraft to start losing track of the Blue Team’s trajectory information; (d) One of the Blue Team’s aircraft is shot down by the Red Team, while the other Blue Team aircraft leaves the no-fly zone.
Electronics 13 02938 g011
Figure 12. Radar suffers from jamming.
Figure 12. Radar suffers from jamming.
Electronics 13 02938 g012
Figure 13. Algorithmic running logic.
Figure 13. Algorithmic running logic.
Electronics 13 02938 g013
Figure 14. Comparison of algorithm performance under different outlier ratios. (a) 0% outliers; (b) 10% outliers; (c) 20% outliers; (d) 30% outliers.
Figure 14. Comparison of algorithm performance under different outlier ratios. (a) 0% outliers; (b) 10% outliers; (c) 20% outliers; (d) 30% outliers.
Electronics 13 02938 g014
Figure 15. Comparison of algorithm performance under different outlier ratios. (a) 40% outliers; (b) 50% outliers; (c) 60% outliers; (d) 70% outliers.
Figure 15. Comparison of algorithm performance under different outlier ratios. (a) 40% outliers; (b) 50% outliers; (c) 60% outliers; (d) 70% outliers.
Electronics 13 02938 g015
Figure 16. Anti-jamming performance errors. (ac) Radar measurement data included; (df) Radar measurement data not included.
Figure 16. Anti-jamming performance errors. (ac) Radar measurement data included; (df) Radar measurement data not included.
Electronics 13 02938 g016
Figure 17. Track prediction performance errors. (ac) Radar measurement data included; (df) Radar measurement data not included.
Figure 17. Track prediction performance errors. (ac) Radar measurement data included; (df) Radar measurement data not included.
Electronics 13 02938 g017
Figure 18. True track and Node-former output track.
Figure 18. True track and Node-former output track.
Electronics 13 02938 g018
Table 1. Environment configuration.
Table 1. Environment configuration.
ItemParameter
CPUIntel Xeon Gold 6248R @3 GHz
GPUNVIDIA Quadro RTX 4000
RAM512 GB
Table 2. RMSE of algorithms.
Table 2. RMSE of algorithms.
ModelAnti-JammingTrack Prediction
RMSE (Az) RMSE (El) RMSE (Range) RMSE (Az) RMSE (El) RMSE (Range)
LSTM0.188°0.036° 174.56 m0.361°0.105° 145.92 m
Transformer0.180°0.038° 236.46 m0.215°0.070° 218.39 m
Node-former0.055°0.015° 69.13 m0.140°0.059° 95.85 m
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, H.; Yang, W.; Wang, Z.; Chen, R. Attention Mechanism and Neural Ordinary Differential Equations for the Incomplete Trajectory Information Prediction of Unmanned Aerial Vehicles Using Airborne Radar. Electronics 2024, 13, 2938. https://doi.org/10.3390/electronics13152938

AMA Style

Peng H, Yang W, Wang Z, Chen R. Attention Mechanism and Neural Ordinary Differential Equations for the Incomplete Trajectory Information Prediction of Unmanned Aerial Vehicles Using Airborne Radar. Electronics. 2024; 13(15):2938. https://doi.org/10.3390/electronics13152938

Chicago/Turabian Style

Peng, Haojie, Wei Yang, Zheng Wang, and Ruihai Chen. 2024. "Attention Mechanism and Neural Ordinary Differential Equations for the Incomplete Trajectory Information Prediction of Unmanned Aerial Vehicles Using Airborne Radar" Electronics 13, no. 15: 2938. https://doi.org/10.3390/electronics13152938

APA Style

Peng, H., Yang, W., Wang, Z., & Chen, R. (2024). Attention Mechanism and Neural Ordinary Differential Equations for the Incomplete Trajectory Information Prediction of Unmanned Aerial Vehicles Using Airborne Radar. Electronics, 13(15), 2938. https://doi.org/10.3390/electronics13152938

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop