Next Article in Journal
The Measurement Problem Is a Feature, Not a Bug–Schematising the Observer and the Concept of an Open System on an Informational, or (Neo-)Bohrian, Approach
Next Article in Special Issue
Complexity Reduction in Analyzing Independence between Statistical Randomness Tests Using Mutual Information
Previous Article in Journal
Practical Real-Time Phase Drift Compensation Scheme for Quantum Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit

1
School of Aeronautics, Northwestern Polytechnical University, Xi’an 710072, China
2
Chengdu Aircraft Design and Research Institute, Chengdu 610041, China
3
School of Aerospace, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(10), 1409; https://doi.org/10.3390/e25101409
Submission received: 1 August 2023 / Revised: 24 September 2023 / Accepted: 25 September 2023 / Published: 1 October 2023
(This article belongs to the Special Issue Information Theory in Multi-Agent Systems: Methods and Applications)

Abstract

:
This paper proposes an air combat training framework based on hierarchical reinforcement learning to address the problem of non-convergence in training due to the curse of dimensionality caused by the large state space during air combat tactical pursuit. Using hierarchical reinforcement learning, three-dimensional problems can be transformed into two-dimensional problems, improving training performance compared to other baselines. To further improve the overall learning performance, a meta-learning-based algorithm is established, and the corresponding reward function is designed to further improve the performance of the agent in the air combat tactical chase scenario. The results show that the proposed framework can achieve better performance than the baseline approach.

1. Introduction

The application of reinforcement learning (RL) [1,2] in aerial combat has attracted a lot of attention in recent years, and RL has been used to simulate the behavior of pilots and aircraft and to optimize aerial combat strategies [3,4].
Challenges related to these simulations include establishing the interaction model between pilots and aircraft [5,6]; simulating the behavior of pilots maneuvering the aircraft and its impact [7]; introducing enemy aircraft and weapons; simulating the behavior of the enemy aircraft and its impact [8]; and the simulation of multi-aircraft cooperative combat behavior [9,10]. Of these, confrontation behavior in air combat is complex and variable, with various modes [11], and it is difficult for traditional methods such as state machines and differential games to completely characterize the real-time decision-making state of pilots and devise further optimization according to different situations [12,13]. However, by modeling the air combat process as a Markov process [14], reinforcement learning methods can achieve continuous optimization of decision-making algorithms [15,16].
The first application of RL in aerial combat was proposed by Kaelbling et al. [17]. They proposed a model-based RL approach for controlling an unmanned aerial vehicle (UAV) in a simulated air-to-air combat environment. The UAV was equipped with a simulated radar and missile system, and the RL agent was trained to select the optimal action for the UAV to maximize its chances of survival. The results showed that the RL agent was able to outperform the baseline agent in terms of survival rate. More recently, Hu et al. [18] trained long and short-term memory (LSTM) in a deep Q-network (DQN) framework for air combat maneuvering decisions, and this was more forward-looking and efficient in its decision-making than fully connected neural-network- and statistical-principle-based algorithms [19]. In addition, Li proposed a deep reinforcement learning method based on proximal policy optimization (PPO) to learn combat strategies from observation in an end-to-end manner [20,21], and the adversarial results showed that his PPO agent can beat the adversary with a win rate of approximately 97%. Based on the deep deterministic policy gradient algorithm framework, Lu designed and implemented an air warfare decision policy and improved the efficiency of the training process via a preferred experience playback strategy [22]. This method was able to achieve fast convergence while saving training costs.
Because of the sparse nature of the air combat environment, the shaping of the reward function has been a key challenge in the application of reinforcement learning to air combat [23,24]. Piao constructed a high-fidelity air combat simulation environment and proposed a critical air combat event reward-shaping mechanism to reduce episodic win–lose signals [25,26], enabling fast convergence of the training process. The implementation results showed that reinforcement learning can generate a variety of valuable air combat tactical behaviors under beyond-visual-range conditions. Hu et al. [27] designed a reward function based on the original deep reinforcement learning method, and the design dimension of the reward included the real-time gain due to the maneuver as well as the final result gain. For the air combat maneuver decision problem with sparse rewards, Zhan et al. [28,29,30] applied a course-based learning approach to design a decision course of angle, distance, and mixture which improved the speed and stability of training compared to the original method without any course and was able to handle targets from different directions.
In the air combat decision-making process, the combination of various independent states forms a very large situation space which leads to an explosion of state dimensions [31]. Current research focuses on the rationality of the decision logic after the introduction of reinforcement learning training in a specific scenario [32,33], whereas this paper focuses on making the existing decision algorithm rapidly scalable as more and more realistic situations are introduced to quickly adapt to a more realistic air combat countermeasure environment [34,35]. The state space curse of dimensionality problem often leads to insensitivity in the model tracking which eventually fails to converge to a better stable tracking state. Therefore, in this paper, a hierarchical reinforcement learning (HRL)-based air warfare framework is proposed [36], which uses a hierarchical reinforcement learning structure to implement three-dimensional air warfare. Experimental results show that the proposed framework can achieve better performance than existing methods. The main innovations of this study are as follows:
  • We propose a hierarchical reinforcement learning framework in geographic coordination for the training and use of senior and basic policies to solve the MDP in air combat chase scenarios.
  • We propose a meta-learning algorithm applied to the framework proposed in this paper for the complex sub-state and action space learning problem of air warfare. The reward decomposition method proposed in this paper also alleviates the problem of reward sparsity in the training process to some extent.
  • We independently built a three-degrees-of-freedom air combat countermeasure environment and modeled the task as a Markov process problem. Specifically, we defined the key elements of the Markov process, such as state, behavior, and reward functions for this task.
  • We established a quantitative system to evaluate the effectiveness of reinforcement learning methods for training in 3D air combat.
In Section 2, we describe the application of reinforcement learning algorithms to the established air combat environment. In Section 3, we present the algorithm framework, reward function design ideas, algorithm training, and usage process. In Section 4, we establish a standard evaluation method and compare multiple SOTA models. In Section 5, we discuss the experimental results and in Section 6, we summarize the whole paper.

2. Reinforcement Learning for Air Combat

This paper sets out a design for a hierarchical RL algorithm capable of learning effective decision strategies in air combat countermeasure scenarios through interaction with a simulated environment. The core of the algorithm is the use of Markov decision processes (MDPs) to model the decision process of combat aircraft in the presence of uncertainty and dynamic adversaries [37,38]. In this context, the design of MDPs requires careful consideration of factors such as state space representation, action selection, and reward function design. In addition, the construction of realistic and challenging combat environments is critical to evaluate the performance of the HRL algorithms constructed in this paper [39,40].

2.1. Markov Decision Process

Figure 1 describes the feedback loop; each of the subscripts t and t + 1 representing a time step refers to a different state: the state at moment t and the state at moment t + 1. Unlike other forms of learning, such as supervised and unsupervised learning, reinforcement learning can only be thought of as a series of sequential state–action pairs [41].
Intelligence in reinforcement learning requires information from the current state s t + 1 and also from the previous state s t to make the best decision that maximizes its payoff [42]. A state signal is said to have Markovianity if it has the information necessary to define the entire history of past states.
Markov decision processes (MDPs) represent decision-makers who periodically observe systems with Markovianity and make sequential decisions [42,43]. They are the framework used for most problems in reinforcement learning. For each state s and action a , the probability that the next state s may occur is
P s s a = P r s t + 1 = s s t = s , a t = a
where P denotes the transfer probability, meaning the possible change of air combat situation when a certain behavior a is executed in a specific state s . In this paper, the value of P is fixed, and the expectation value of the next reward value R can be determined as
V π ( s ) = a   π ( s , a ) s   P s s a R s s a + γ V π s , s S , a A
Intelligence tries to maximize its payoff over time, and one way to achieve this is to optimize its strategy. A strategy π is optimal when it produces better or equal returns than any other strategy, and π specifies the probability distribution of executing a certain decision action in a given air combat situation. The equation for state values states that at any state, strategy π is better than π ’ if V π ( s ) V π , ∀ s S . The state value function and the state action value function can be optimized according to the following two equations:
V * ( s ) = m a x π   V π ( s ) , s S
Q * ( s , a ) = m a x π   Q π ( s , a ) , s S , a A
The above two equations can calculate the optimal state value V * ( s ) and the optimal action value Q * ( s , a ) when using the strategy π. The Bellman optimal equation for V * ( s ) can be used to calculate the value of states when the reward function R s s a and the transfer probability P s s a are known without reference to the strategy; similarly, the Bellman optimal equation constructed with the state action value function can be used as follows:
V * ( s ) = m a x a   s   P s s a R s s a + W * s Q * ( s , a ) = s   P s s a R s s a + γ m a x a   Q * s , a
The above two equations can calculate the optimal state value V * ( s ) and the optimal action value Q * ( s , a ) when using the strategy π. Additionally, in the case of no reference strategy, when the reward function R s s a and the transfer probability P s s a are known, the Bellman optimal equation of V * ( s ) can be used to calculate the value of states, representing the expected cumulative returns associated with being in a given situation and subsequently following the best decision strategy throughout the air combat. The Bellman optimal equation constructed with the state action value function can also be used.

2.2. Air Combat Environmental Model

The defined air combat adversarial environment for the MDP is implemented as two simulators S i m u i , i { H o r i z o n t a l , V e r t i c a l } , where S n e x t , R i = S i m u i ( S i , A i ) and A i is the action of Agent i in state S i [44]. The simulator S i m u ( i ) receives the action A i and then produces the next state S i and the reward R i , where the state space S i consists of the coordinates (x,y,z), velocity v and acceleration Δ of the red and blue sides under the geographic coordinate system:
S = ( x r , y r , z r , v x r , v y r , v z r , Δ x r , Δ y r , Δ z r , x b , y b , z b , v x b , v x r , v y b , v z b , Δ x b , Δ y b , Δ z b , )
In the next state, the geometric position, the spatial positions of the tracker, and the target are updated after the input actions [45,46]. The action space in H o r i z o n t a l space are discrete, and they are defined as three different actions: forward, left, and right. Again, action space V e r t i c a l is defined as three different actions: up, hold, and down. In addition, we specifically set rules on height for this simulation to match realistic scenarios so that, during training, if the tracker moves beyond the restricted height range, the simulator limits its further descent or ascent and then receives a new movement [47]. We define rewards R i for the corresponding environment, i { H o r i z o n t a l , V e r t i c a l } . The role of the reward function is to encourage the tracker to continuously track the movement of the target. It is defined as follows:
R i = ω 1 S O T ( f t a r g e t , f s t a t e )
where ω 1 is a parameter and is a positive parameter, f t a r g e t represents the real position and velocity of the target, and fstate represents the current position and velocity of the tracker. SOT represents the status of tracking between f s t a t e and f t a r g e t . The DQN [14] algorithm is applied to the learning of each agent in the simulation. It learns an optimal control policy π i : S i , G i A i , i { H o r i z o n t a l , V e r t i c a l } .
The horizontal position between the aircraft and the target is indicated by C = ( φ u , D ) , where D is the azimuth of the aircraft and the distance between the two aircraft, respectively. Figure 2 depicts the position of the tracker relative to the target. The subscripts 𝑢 and 𝑡 indicate the tracker aircraft and the target, respectively, and φ t indicates the azimuth of the tracker relative to the target.
The most important platform capability in air combat countermeasures training systems is flight capability, so this paper presents designs for a set of motion models to model the aircraft platform, which mainly reflect the flight trajectory under the limitation of aircraft flight performance. The six degrees of freedom for aircraft require consideration of the warplane as a rigid body, the complexity of the aircraft structure, and its longitudinal coupling. Here, a three-degrees-of-freedom model is used, ignoring the aircraft as a rigid body, treating it as a mass, and assuming that the flight control system can respond accurately and quickly to form a maneuver trajectory. The core of the maneuvering decision problem is the rapid generation of the dominant maneuvering trajectory, and the aircraft three-degrees-of-freedom model can meet the solution requirements. The aircraft three-degrees-of-freedom model includes a mass point model of the aircraft platform and a dynamics model; the control model is shown in Figure 3. The specific models are
V ˙ = g ( n x sin ( θ ) ) θ ˙ = g ( n z c o s ( γ ) c o s ( θ ) ) V φ ˙ = g ( n z s i n ( γ ) ) V c o s ( θ ) X ˙ = V c o s ( θ ) c o s ( φ ) Y ˙ = V c o s ( θ ) s i n ( φ ) Z ˙ = V s i n ( θ )
where 𝑥, 𝑦, and 𝑧 denote the position of the aircraft in the geographic coordinate system; V is the flight speed; θ is the velocity inclination angle, i.e., the angle between the velocity direction and the horizontal plane, with upward as positive; φ is the heading angle, i.e., the angle between the velocity direction on the horizontal plane and the due north direction, with clockwise as positive; and where it is assumed that the velocity direction is always in line with the direction in which the nose is pointing, i.e., the angle of attack and the sideslip angle are zero.
θ , φ , γ denote the trajectory inclination angle, track deflection angle, and roll angle, respectively; n x and n z denote the tangential overload along the velocity direction and the normal overload in the vertical velocity direction, respectively; and 𝑔 is the gravitational acceleration. In the above equation, the first three terms are the mass kinematics model and the last three terms are the aircraft dynamics model; the state variables include 𝑥, 𝑦, 𝑧, θ , φ , γ , 𝜓, and V ; the control variables include n x , n z , and γ . Because an ideal mass point model is used in this study, the flight performance, control inputs, and control response parameters are restricted to make the trajectory and maneuvers of the target aircraft reasonable. Specifically, the values of n x and n z are limited to within 2 g and 5 g, respectively.

3. Hierarchical Reinforcement Learning Design

In this paper, we propose a hierarchical reinforcement learning training framework that comprises two parts: environment design and framework building. The purpose of the environment design is primarily to define the input and output state data available to the agent and the reward functions that can be obtained, and the framework building is primarily to establish the corresponding hierarchical network structure, realize the reward mapping corresponding to the course learning, and design the optimization algorithm and the corresponding training strategy.

3.1. Geometric Hierarchy in the Aircombat Framework

We formulate the intelligent body motion decision for a 3D air combat 1V1 confrontation as a Markov decision process (MDP), supplemented by a goal state G that we want the two agents to learn. We define this MDP as a tuple (S, G, A, T, μ ), where S is the set of states, G is the goal, A is the set of actions, and T is the transition probability function. In this paper, a hierarchical reinforcement learning-based approach called a hierarchical reinforcement learning framework in geographic coordination for air combat, referred to as HRL-GCA, is used to build a shared multilevel structure. The method uses a technique called meta-learning, which learns from a set of tasks and applies this knowledge to new tasks. The algorithm can effectively build a shared multilevel structure, thus improving learning efficiency.
As shown in Figure 4, the global state S is a geometric representation of the tracker and target aircraft in a 3D simulated air combat scenario, including the positions S = (x, y, z) and velocities v = (vx, vy, vz) of both aircraft. At the beginning of each episode of each state s i in the MDP, for a given initial state s 0 and target g i , the solution to the sub-policy ω is a control policy π i : S i , G i A i that maximizes the following value function:
v π i s i , g i E π i [ t = 0 μ i t R t i | s 0 i = s i , g i = G i ]
An agent consists of an algorithm that updates a parameter vector ( θ , ω ) defining a stochastic policy π θ , ω ( s | a ) , where the ω parameter is shared among all sub-policies, whereas the θ parameter is learned for each senior policy starting from zero, encoding the state of the learning process on that task. In the considered setup, an MDP is first sampled from the P S , the agent is represented by the shared parameter ω and the randomly initialized θ parameter, and the agent iteratively updates its θ parameter during the T-step interaction with the sampled MDP. The objective of the HRL-GCA is to optimize the value contributed by the sub-policy over the sampled tasks:
V = maximize θ E S ~ P S , k = 0 , , T 1 [ v π s k , g k ]
where π consists of a set of sub-policies π 1 , π 2 , …, π N , and each sub-policy π i is defined by a subvector ω n . The network constructed by the parameter θ works as a selector. That is, the senior policy parameterized by θ selects the most appropriate behavior from index n ϵ { 1 , 2 , , N } to maximize the global value function V .

3.2. Reward Shaping

The senior action reward is used to train senior behaviors, which guide the sub-action to make further behavioral decisions. We take inspiration from the Meta-Learning Shared Hierarchies architecture to train the sub-policy independently, solidify its parameters, and then train senior action adaptively. Our approach is similar to Alpha-Dogfight [48], but we differ in that we implement further layering in the behavioral layer and map global rewards to local rewards by transformations under geographic coordination, and experimental results demonstrate that performance in the behavioral layer is further enhanced.

3.3. Senior Policy Reward

The senior policy performs discrete actions at a frequency five times lower than the sub-policy, which is 1 Hz and is trained using the same DQN as the sub-policy. The state space of the senior policy differs from that of the sub-policy, which is described in detail later in this paper. The reward for senior policy is given by
r t o t a l = α r a n g l e + β r d i s
where α and β are positive parameters and α + β = 1 .
Firstly, the angle reward r a n g l e can help the model learn how to control the angle of the aircraft toward the target, and φ u is related to the limits of the detection angle of the airborne radar and the off-axis angle of the missile. Specifically, the attack advantage increases the closer φ u is to the desired angle, and r a n g l e reaches its maximum when φ u = 0°, i.e., when the velocity is aligned with the target:
r a n g l e = e ( a b s ( f ( φ u ) f ( φ t ) ) / 180 )
Secondly, the distance redirection r d i s is designed based on the distance between the aircraft and the target, which can help the model learn how to control the position of the aircraft to achieve a reasonable position about the target. Specifically, the smaller the distance D between the aircraft and the target, the higher the r d i s value:
r d i s = e ( a b s ( f ( D ) ) / 100 )
We used the above rewards for the initial training, and then in subsequent experiments, for comparison with other models, we adjusted the design of the reward to achieve the same state as the baseline. A description of how the three model rewards are adjusted in this paper can be found in Appendix A.

3.4. Sub-Policy Reward

However, the objective of this paper requires the mapping of rewards to the two subtask spaces, and we redistribute rewards for Agent 1 and Agent 2 via transformation in the geometric space. Because Agent 1 and Agent 2 are mainly implemented in two planes of control, as shown in Figure 3, r t o t a l is achieved by mapping φ u and D to the x-y and x-z planes using the function δ to reconstruct the G i .
r t o t a l h δ φ u h , δ φ t h , δ D h = α h r a n g l e h δ φ u h , δ φ t h + β h r d i s h δ φ u h , δ D h r t o t a l v δ φ u v , δ φ t v , δ D v = α v r a n g l e v δ φ u v , δ φ t v + β v r d i s v δ φ u v , δ D v
Here,
r a n g l e h δ φ u h , δ φ t h = e ( a b s ( f ( φ u h ) f ( φ t h ) ) / 180 ) ,
r d i s h δ φ u h , δ D h = e ( a b s ( D h f ( φ u h ) / 100 ) ,
r a n g l e v δ φ u v , δ φ t v = e ( a b s ( f ( φ u v ) f ( φ t v ) ) / 180 ) ,
r d i s v δ φ u v , δ D v = e ( a b s ( D v f ( φ u v ) ) / 100 ) .
The redistribution of rewards is achieved by the function δ . The δ function is a spatial projection operator that maps reward elements φ u , φ t , and D to the x-y and x-z planes, respectively. This ensures that the reward functions r t o t a l h and r t o t a l v , which are used for training in the x-y and x-z planes, have the same expression. However, their auto-covariates are the result of the projection through the δ posterior: φ u h , φ t h , D h , and φ u v , φ t v , D v , respectively, as detailed in Appendix B. Of these, reward r t o t a l h allows the tracker to follow the target better on the x-y surface, and reward r t o t a l v is used to suppress the altitude difference and, as much as possible, encourage the aircraft to be at the same altitude level as the target at high altitude. In addition, in this paper, the treatment in Equation (7) is also applied to its rewards in the comparisons with other baselines.

3.5. Hierarchical Training Algorithm

In this paper, a course learning approach is used for hierarchical training; the definition of the algorithm is detailed in Appendix C, and the policy network is trained to interact with the environment at a frequency of 10 Hz. The same observation space is used for both policies.
We then explore cooperative learning between Agent 1 and Agent 2 in the training of horizontal control and height control policies. In each iteration of the learning, firstly, Agent 1 moves the tracker on the x-y surface of the 3D geographic coordination scenes; secondly, the next state s t + 1 1 and the intermediate state s t 2 update after action a t 1 ; and thirdly, Agent 2 moves the tracker on the x-z surface. The next state s t + 1 2 updates after a t 2 .
Initial conditions: These initial conditions are divided into tracking targets that start moving from different positions and take different forms of motion in the height and horizontal planes. Concerning stochastic multistep payoffs, for time–distance learning, multistep payoffs tend to lead to faster learning when appropriately tuned for the number of steps to be used in the future. Instead of tuning a fixed value, we define the maximum number of steps in the future and uniformly sample the maximum value. A common expression for future value is
Q S , A Q S , A + R + τ max a Q ( S , a ) Q S , A
The tactical objective of the horizontal plane tracking subtask is to enable the tracker to continuously track the target aircraft in the x-y plane. Formally, motion in the x-y plane is achieved by outputting horizontal motion, successive horizontal left turns, and successive horizontal right turns at each simulation step with a constant steering speed of 18°/s. The initial and termination conditions for the x-y subtasks are designed as shown in Figure 2. The tactical objective of the altitude tracking subtask is to enable the tracker to follow the target aircraft consistently at altitude. The mission can start in any state. Formally, motion in the x-z plane is achieved by outputting horizontal motion, continuous climb, and continuous descent in each simulation step, with a constant climb and descent rate of 20 m/s.
This in turn contains one output, namely, the value of Q ( s , x i ) . The activation function is the logsoftmax function:
Q ( s , x i ) = x i x m l o g ( j = 0 n e x j x m )
and Equation (20) directly outputs the value of each action using the logsoftmax nonlinear function, where x m is the largest element of X = ( x 1 , x 2 , x n ) .

3.6. Hierarchical Runtime Algorithm

In the hierarchical runtime algorithm, we explore the cooperation of Agent 1 and Agent 2 in a 3D simulated air combat situation. The algorithm is defined in detail in Appendix D. In each iteration of learning, firstly, Agent 1 moves in the x-y plane of the 3D air combat scenario; secondly, the next state s t 1 + 1 and intermediate state s t 2 are updated after action; and thirdly, Agent 2 moves up or down in the x-z plane. The next state s t 2 + 1 is updated after a t 2 .
For each action 𝑚𝑖, a minimum period 𝑡 = 1500 milliseconds and a maximum period 𝑢𝑖 = 4 milliseconds are set. When the reinforcement learning intelligence outputs the action 𝑚𝑖 (including the stop action) at moment T, it starts to execute 𝑚𝑖 if no action is executed at the previous moment T − 1. If moment T – 1 performs action 𝑚𝑗 and the execution time is greater than or equal to 𝑡, then at moment T, the agent will be allowed to execute 𝑚𝑖 to replace the action 𝑚𝑗, otherwise not. If moment T – 1 performs action 𝑚𝑗 and the execution time is less than 𝑡𝑗, then the output behavior 𝑚𝑖 at moment T is ignored. When the reinforcement learning intelligence outputs no behavior (which is not the same as the stopping behavior) at moment T, if the previous moment T – 1 performed the behavior 𝑚𝑘 and the execution time is greater than or equal to 𝑢𝑖, then the execution of the no-behavior starts; otherwise, the execution of the behavior 𝑚𝑘 continues. The setting of the min-max period can to some extent prevent incorrect behavior of the flight unit.

4. Results

4.1. Experimental Environment Setup

The experiments in this paper use a hierarchical reinforcement learning framework to solve the problem in an air combat simulation environment. The hardware environment used in the experiments is an Intel Core i7-8700K CPU, 16 GB RAM, and an NVIDIA GeForce GTX 4090 Ti graphics card. The size of the 3D space in the experiment is 100 km × 100 km × 10 km; there are 20,000 × 480 s training episodes for each model; and the actual data sampling frequency is 10 HZ. The experimental results show that the performance of the algorithm improves significantly after the training of 20,000 episodes.

4.2. Performance Metrics during Training and Validation

To select the best-performing agent, we create an evaluation metric to compare the training results of various methods. The qualitative and quantitative results demonstrate the usefulness of our proposed model. The tracking performance of the tracker is evaluated when the target is moving at 0–180° relative north in an air combat environment. For comparison, we trained 2400 episodes for each angle type, for a total of 11.5 × 106 simulation steps, and tested 500 samples for the corresponding angle types.
The meanings of each indicator are as follows: miss distance represents the average distance between the tracker and the target during the entire tracking process; miss angle represents the average track angle φ u between the tracker and the target during the entire tracking process; approach time represents the time taken to approach the target for the first time to a certain distance; hold distance time is the length of time that the tracker stays within a certain distance of the target; hold angle time is the time that the tracker stays within a certain angle of the target; and cost time refers to the time spent by each strategy model when outputting the current action command.
P M i s s   D i s t a n c e = a c c ( t o t a l   d i s ) t e p o c h   t i m e
P M i s s   A n g l e = a c c ( t r a c k   a n g l e   φ u ) t e p o c h   t i m e
P A p p r o a c h   t i m e = τ T O A τ 0
P H o l d   d i s t a n c e   t i m e = τ ( d i s σ ) τ ( e p o c h   t i m e )
P H o l d   a n g l e   t i m e = τ ( a n g l e ) τ ( e p o c h   t i m e )
P C o s t   T i m e = τ ( θ )

4.3. Validation and Evolution of the Hierarchical Agents

In this experiment, we reproduce the models and algorithms in three papers [9,15,49], and apply the hierarchical reinforcement learning framework established in this paper to learn and train them, respectively, while mapping the reward functions shaped in the three papers in the corresponding sub-state spaces; then, in the air combat environment established in this paper, different models are compared in the same test scenarios, and the performance of the three original models is compared with that of the models after applying HRL. We use the benchmark performance comparison method proposed in Section 4.2 to compare the models proposed in the paper, as shown in Table 1. Models 1, 2, and 3 denote the performance of the three models. The experimental results show that the HRL-GCA proposed in this paper can achieve higher scores in all three dimensions under the six test metrics compared with the other three models: the miss distance, miss angle, and approach time decreased by an average of 5492 m, 6.93 degrees, and 34.637 s, respectively, and the average improvement of angle maintenance and distance maintenance time is 8.13% and 16.52%, respectively. Of the other models, Model 2 has the highest hold distance and hold angle time with percentages of 41.12 and 15.44, respectively. In addition, the HRL-GCA model can converge faster and achieve higher accuracy in the training process. Therefore, we conclude that HRL-GCA demonstrates better performance in this experiment.
As shown in Table 1, the implementation of HRL models results in a 40–50% increase in runtime compared to the baseline models. This can be attributed to the fact that HRL models involve more complex computations and require more processing time. This is mostly because HRL incorporates several learning layers. Consequently, the HRL will execute over two extra neural networks in addition to the base models.
Notwithstanding, we consider the time cost to be acceptable based on the comparative results presented in Table 1. For instance, Model 2 benefited from HRL improvement, requiring only a minimum of 87.56 s for Approach Time and making approximately 65 decisions for the approach to the target. In contrast, the corresponding model without HRL improvements required 137.06 s for Approach Time, making about 145 decisions for the approach to the goal. The HRL-improved model achieves its goal with only 65 decisions compared to the original model’s 145, resulting in a 55% improvement in decision-making efficiency. This increase in efficiency of the HRL-improved model (55%) offsets the additional time overhead required to execute the model (43.40%).
Furthermore, as an example, Model 2 shows improved Hold Distance Time and Holding Angle Time by 16.33% and 8.24%, respectively, after implementing HRL. Furthermore, compared to the model without HRL improvement, the distance and angle tracking stability are enhanced by 65% and 114%, respectively. In summary, although the computation time spent increased by 43.40%, the HRL improvement resulted in a 55% increase in decision efficiency within the same timeframe. The distance and angle tracking stability also increased by 65% and 114%, respectively. Therefore, this improvement is deemed reasonable.

5. Discussion

5.1. Trajectory of Air Combat Process

As shown in Figure 5, Figure 6, Figure 7 and Figure 8, we deploy the algorithm of this paper in a typical air combat scenario and compare its tracking of the target aircraft with a model that does not use this algorithm. During air combat, continuous tracking of the target aircraft in a given scenario is necessary to shoot it down. In the test cases, the target aircraft maneuvers continuously at altitude and moves away from the tracker by turning away from it, as seen in the 3D and 2D tracking trajectories, but the tracker ensures continuous alignment with the target in both altitude and direction. In contrast, the use of the other model fails to achieve continuous tracking of the target in either direction or altitude. Furthermore, the red dashed line in Figure 6 and Figure 8 shows the desired tracking trajectory for the target.
In our experiments, we use a hierarchical reinforcement learning framework to optimize and enhance the vehicle tracking trajectories. The trajectories in Figure 9 show the tracking states of the modified model 2 based on HRL and the model set out in this paper in the XY plane, XZ plane, and XYZ 3D space, respectively. Of these, in Figure 10, the red line is the tracking flight, the blue line is the tracked flight, and the number represents the flight trajectory sequence of both flights. The experimental results show that the use of the hierarchical reinforcement learning framework can effectively improve the accuracy and stability of aircraft tracking trajectories and can effectively reduce their deviation. It is found that Model 3 is more sensitive to the weighting parameters α , β in Equation (11) and has the best test results when the two reward ratios in the original paper are set to 0.5, 0.5. Irrespective of the rewards in Models 1, 2, and 3 or the rewards used in this paper, in Figure 11 and Figure 12, the tracking performance of a single network simultaneously controlling the motion of the horizontal plane and the motion of the height layer is inferior to that of multiple networks controlling them separately. In addition, the experimental results show that the use of the reinforcement learning method can effectively improve the accuracy of aircraft tracking trajectories, thus improving the timeliness of target tracking.

5.2. Training Process

The analysis of the experimental results in this paper shows that we can compare the changes in reward and loss of Models 1, 2, and 3 with the HRL-GCA model during the training process. From the experimental results, the reward and loss of HRL-GCA converge as the episodes increase and reach their optimal state after stabilization. In Figure 13, from the change in reward, our research model reward reaches its maximum at episode 21, whereas the rewards of the three standard models still show large fluctuations at episode 21, indicating that the reward of our research model has better convergence performance. Figure 14 illustrates the loss parameters during training after normalization. From the change in loss, the loss of our research model reaches its minimum at episode 592, whereas the losses of the three standard models still show a large fluctuation at episode 592, indicating that our research model has a better convergence performance of loss.
In summary, Figure 13 and Figure 14 show that our research model has the best convergence performance during the training process, as well as an optimal state after stabilization. Therefore, when introducing a new sub-policy, the framework in this paper can achieve fast adaptation in training and learning for the corresponding task. Furthermore, in Figure 13, Model 3 and the proposed model decrease significantly after the first peak around episode 21; such behavior makes these models inferior to Model 2. This is mainly due to overfitting. From Equations (9) and (10) in Section 1, it can be deduced that each sub-policy π i can quickly learn its corresponding subvector ω n after the initial learning phase, but ω n learns only a tiny portion of the state space, and it needs to further learn the θ corresponding to the selector to maximize the global value function V.
In addition, the θ corresponding to the selector can be learned from the ω n subvector, but due to the large amount of training data required to train θ , as π θ , ω ( s | a ) performs stochastic exploration, the variance is large, resulting in a decrease in v π s k , g k until S ~ P S accumulates sufficiently to train a valid θ parameter. Furthermore, model 2 can maintain a more stable exploration-utilization capability throughout the training process, but the model proposed in this paper has a higher final reward value compared to the other models due to a better memory of what has been learned. Finally, we use black dashed lines in Figure 13 and Figure 14 to show the variation of reward and loss with the training process in the ideal case.

6. Conclusions

The hierarchical reinforcement learning framework in geographic coordination for air combat proposed in this paper trains two types of neural networks using distance reward, angle reward, and a combination of both to control the vehicle in multiple dimensions. The model has achieved good results in tracking targets in multiple dimensions. In thousands of tests, it achieved an average improvement of 8.13% in angle tracking and 16.52% in distance tracking over the baseline model, demonstrating its effectiveness. However, the model has limitations, especially in complex environments or when the goal is to perform complex maneuvers, and it is not yet able to achieve optimal control. Future research should focus on improving the tracking performance in such scenarios, along with exploring additional reward functions to improve stability and accuracy. Furthermore, numerous challenges remain, such as addressing two-agent game combat and extending to 2v2 and multi-agent combat scenarios in air combat pair control, which warrant further exploration. In addition, the application potential of this model in other real-world scenarios should be investigated.

Author Contributions

Conceptualization, R.C. and Q.Z.; methodology, R.C. and H.L.; software, R.C.; validation, R.C. and H.P.; formal analysis, R.C.; investigation, G.Y.; resources, H.L.; data curation, H.P.; writing—original draft preparation, R.C.; writing—review and editing, Q.Z.; visualization, R.C.; supervision, G.Y.; project administration, H.L.; funding acquisition, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by a Project funded by China Postdoctoral Science Foundation (Grant No. 61932012), and in part by the Natural Science Basic Research Program of Shaanxi (Program No. 2022JQ-061).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Design of Reward Functions

We adjusted the reward functions in the three models appropriately according to the characteristics and goals of the tasks. Reward functions play a crucial role in reinforcement learning in that they guide model learning and decision-making. However, in complex real-world tasks, it is very challenging to design ideal reward functions that apply to all situations. To realize the model in this paper to adapt the reward functions of the three models [9,15,49], we mainly adjust the f ( φ u ) , f ( φ t ) , and f ( D ) in Equations (12) and (13).
In model 1:
f ( φ u ) = 1 φ u 150 0 φ u 30 0 30 φ u 180
f ( φ t ) = ( 180 φ t ) π 180 , 180 φ t 180
f ( D ) = 1 D < 10 D 10 30 10 D 40 0 D > 40
In model 2:
f ( φ u ) = 0
f ( φ t ) = 360 φ u φ t 360 , 0 φ u 180 , 0 φ t 180
f ( D ) = 0 D 80 D 60 40 60 < D 80 D 40 20 40 < D 60 1 20 < D 40 0 D 20
In model 3:
f ( φ u ) = ( φ u 30 ) 2 , 0 φ u 180
f ( φ t ) = ( φ t 60 ) 2 , 0 φ t 180
f ( D ) = ( R / 80 ) 2

Appendix B. The Spatial Projection

The main role of the spatial mapping operator δ is to map the base auto-morphisms of the reward function to the x-y and x-z planes, such that the x-y and x-z planes share the same reward function form but have different auto-morphisms.
Where V u and V t represent the velocity vectors of the tracker and target in the geographic coordinate system, respectively, with scalar forms ( v u , x , v u , y , v u , z ) and ( v t , x , v t , y , v t , z ) . D represents the vector expression of the line that connects the center of gravity of the tracker and the center of gravity of the target in the geographic coordinate system, and its scalar form is ( d x , d y , d z ) . As shown in Figure 2, φ u is the angle between vector V u and D , while φ t is the angle between vector V t and D .
In x-y plane:
φ u h = δ φ u x , y , z x , y h = δ a r c c o s V u D V u D = a r c c o s ( v u , x d y + v u , y d x ) φ t h = δ φ t x , y , z x , y h = δ a r c c o s V t D V t D = a r c c o s ( v t , x d y + v t , y d x ) D h = δ D x , y , z x , y h = d x d x + d y d y
In x-z plane:
φ u v = δ φ u x , y , z x , y v = δ a r c c o s V u D V u D = a r c c o s ( v u , x d z + v u , z d x ) φ t v = δ φ t x , y , z x , y v = δ a r c c o s V t D V t D = a r c c o s ( v t , x d z + v t , z d x ) D v = δ D x , y , z x , y v = d x d x + d z d z

Appendix C. Algorithm A1

The training algorithm for hierarchical reinforcement learning proposed in this paper involves the steps of hierarchical policy optimization, subtask policy optimization, hierarchical reward design, and network parameter training. Among them, the reward function of each layer can be adjusted and optimized according to the objectives and characteristics of the task. By reasonably designing the reward function, the hierarchical reinforcement learning network can be guided to learn decision-making and behavioral strategies suitable for the task. Through iteration and optimization of these steps, we can obtain a hierarchical reinforcement learning model adapted to the complex task.
Algorithm A1 The hierarchical training algorithm
Initialize a one-on-one air combat simulation environment
Initialize replay buffer R1,R2 to capacity N
Initialize the action-value function Q with random weights
Initialize Agent DQN1with (Q,R1), DQN2 with (Q,R2)
for episode = 1, MAX do
Initialize   state   s t 1 = env.reset()
 for t = 1, T do
   Agent   1   samples   action   a t 1 = a t , θ 1 s t 1 + ε t using DQN1
   Execute   a t 1 in an air combat simulation environment
   The   aircraft   observes   s t + 1 1 ,   reward   r t 1
  Agent 2 samples action a t 2   = a t , θ 2 s t 2 + ε t using DQN2
   Execute   a t 2 in an air combat simulation environment
   The   aircraft   observes   s t + 1 2 ,   reward   r t 2
 end for
Store   ( s t 1 , a t 1 , r t 1 , s t + 1 1 )   into   R 1 ,   Store   ( s t 2 , a t 2 , r t 2 , s t + 1 2 ) into R2
if the Update condition is reached, then
 Sample random mini-batch of m from the replay buffer
Calculate   the   target   Q   value   for   each   both   DQN   in   each   transition   Q ^ m = r m + λ m a x a θ Q m ( s m t + 1 , a θ s m t + 1 ; θ )
Compute   gradient   estimation   θ 1   and   θ 2
Update   the   parameters   of   DQN 1   and   DQN 2   based   on   the   optimizer   using   θ 1   and   θ 2
 end for
end for

Appendix D. Algorithm A2

In practice, the hierarchical reinforcement learning algorithm proposed in this paper gradually improves the performance and decision-making ability of the whole system through the steps of hierarchical structure, policy execution, reward signaling, and parameter updating. Through decision-making and learning at different levels, the hierarchical reinforcement learning algorithm can be more flexible in solving complex, high-dimensional tasks and achieve improved adaptability and generalization ability.
Algorithm A2 The hierarchical running algorithm
Load   trained   neural   networks   Q 1 ( s , a , θ 1 ) , Q 2 ( s , a , θ 2 )
Initialize   the   state   of   the   tracker   and   target   ( s T = 0 u , s T = 0 t )
Initialize   target   maneuver   strategy   π ( a | s ) t
for step = 1 to max step do
for   i   in   ( θ 1 , θ 2 ) do
   Calculate   ( φ u h , φ t h , D )
   Execute   a t ~ π θ 1 ( a t | s T u )
   | X T h , Y T h , Z T h | is obtained according to (2)
   Get   the   next   state   s T + 1 u
   Calculate   ( φ u v , φ t v , D )
   Execute   a t ~ π θ 2 ( a t | s T + 1 u )
   | X T v , Y T v , Z T v | is obtained according to (2)
   Get   the   next   state   s T + 2 u
s T u = s T + 2 u
 end for
The   target   moves   to   the   next   state   s T + 1 t   according   to   the   strategy   π ( a | s ) t
s T t = s T + 1 t
end for

References

  1. Sutton, R.S.; Barto, A. Reinforcement Learning: An Introduction, Nachdruck; Adaptive Computation and Machine Learning; The MIT Press: Cambridge, MA, USA, 2014; ISBN 978-0-262-19398-6. [Google Scholar]
  2. Silver, D.; Schrittwieser, J.; Simonyan, K.; Antonoglou, I.; Huang, A.; Guez, A.; Hubert, T.; Baker, L.; Lai, M.; Bolton, A.; et al. Mastering the game of Go without human knowledge. Nature 2017, 550, 354–359. [Google Scholar] [CrossRef] [PubMed]
  3. Hu, D.; Yang, R.; Zuo, J.; Zhang, Z.; Wu, J.; Wang, Y. Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat. IEEE Access 2021, 9, 32282–32297. [Google Scholar] [CrossRef]
  4. Jiang, Y.; Yu, J.; Li, Q. A novel decision-making algorithm for beyond visual range air combat based on deep reinforcement learning. In Proceedings of the 2022 37th Youth Academic Annual Conference of Chinese Association of Automation (YAC), Beijing, China, 19–20 November 2022; IEEE: New York, NY, USA, 2022; pp. 516–521. [Google Scholar] [CrossRef]
  5. Shi, W.; Song, S.; Wu, C.; Chen, C.L.P. Multi Pseudo Q-learning Based Deterministic Policy Gradient for Tracking Control of Autonomous Underwater Vehicles. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3534–3546. Available online: http://arxiv.org/abs/1909.03204 (accessed on 3 April 2023). [CrossRef] [PubMed]
  6. Byrnes, C.M.W. Nightfall: Machine Autonomy in Air-to-Air Combat. Air Space Power J. 2014, 28, 48–75. [Google Scholar]
  7. Kim, C.-S.; Ji, C.-H.; Kim, B.S. Development of a control law to improve the handling qualities for short-range air-to-air combat maneuvers. Adv. Mech. Eng. 2020, 12, 168781402093679. [Google Scholar] [CrossRef]
  8. Xu, J.; Zhang, J.; Yang, L.; Liu, C. Autonomous decision-making for dogfights based on a tactical pursuit point approach. Aerosp. Sci. Technol. 2022, 129, 107857. [Google Scholar] [CrossRef]
  9. Li, W.; Shi, J.; Wu, Y.; Wang, Y.; Lyu, Y. A Multi-UCAV cooperative occupation method based on weapon engagement zones for beyond-visual-range air combat. Def. Technol. 2022, 18, 1006–1022. [Google Scholar] [CrossRef]
  10. Kong, W.; Zhou, D.; Du, Y.; Zhou, Y.; Zhao, Y. Hierarchical multi-agent reinforcement learning for multi-aircraft close-range air combat. IET Control Theory Appl 2022, 17, cth2.12413. [Google Scholar] [CrossRef]
  11. Ernest, N.; Carroll, D. Genetic Fuzzy based Artificial Intelligence for Unmanned Combat Aerial Vehicle Control in Simulated Air Combat Missions. J. Def. Manag. 2016, 06, 2167-0374. [Google Scholar] [CrossRef]
  12. Li, Q.; Jiang, W.; Liu, C.; He, J. The Constructing Method of Hierarchical Decision-Making Model in Air Combat. In Proceedings of the 2020 12th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Hangzhou, China, 22–23 August 2020; IEEE: New York, NY, USA, 2020; pp. 122–125. [Google Scholar] [CrossRef]
  13. Mulgund, S.; Harper, K.; Krishnakumar, K.; Zacharias, G. Air combat tactics optimization using stochastic genetic algorithms. In SMC’98 Conference Proceedings, Proceedings of the 1998 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No. 98CH36218), San Diego, CA, USA, 14 October 1998; IEEE: New York, NY, USA, 1998; Volume 4, pp. 3136–3141. [Google Scholar] [CrossRef]
  14. Lee, G.T.; Kim, C.O. Autonomous Control of Combat Unmanned Aerial Vehicles to Evade Surface-to-Air Missiles Using Deep Reinforcement Learning. IEEE Access 2020, 8, 226724–226736. [Google Scholar] [CrossRef]
  15. Li, Y.; Lyu, Y.; Shi, J.; Li, W. Autonomous Maneuver Decision of Air Combat Based on Simulated Operation Command and FRV-DDPG Algorithm. Aerospace 2022, 9, 658. [Google Scholar] [CrossRef]
  16. Cao, Y.; Kou, Y.-X.; Li, Z.-W.; Xu, A. Autonomous Maneuver Decision of UCAV Air Combat Based on Double Deep Q Network Algorithm and Stochastic Game Theory. Int. J. Aerosp. Eng. 2023, 2023, 3657814. [Google Scholar] [CrossRef]
  17. Kaelbling, L.P.; Littman, M.L.; Moore, A.W. Reinforcement learning: A survey. J. Artif. Intell. Res. 1996, 4, 237–285. [Google Scholar] [CrossRef]
  18. Wang, Y.; Ren, T.; Fan, Z. Autonomous Maneuver Decision of UAV Based on Deep Reinforcement Learning: Comparison of DQN and DDPG. In Proceedings of the 2022 34th Chinese Control and Decision Conference (CCDC), Hefei, China, 21–23 May 2022; IEEE: New York, NY, USA, 2022; pp. 4857–4860. [Google Scholar] [CrossRef]
  19. Chen, Y.; Zhang, J.; Yang, Q.; Zhou, Y.; Shi, G.; Wu, Y. Design and Verification of UAV Maneuver Decision Simulation System Based on Deep Q-learning Network. In Proceedings of the 2020 16th International Conference on Control, Automation, Robotics and Vision (ICARCV), Shenzhen, China, 13 December 2020; IEEE: New York, NY, USA, 2020; pp. 817–823. [Google Scholar] [CrossRef]
  20. Li, L.; Zhou, Z.; Chai, J.; Liu, Z.; Zhu, Y.; Yi, J. Learning Continuous 3-DoF Air-to-Air Close-in Combat Strategy using Proximal Policy Optimization. In Proceedings of the 2022 IEEE Conference on Games (CoG), Beijing, China, 21–24 August 2022; IEEE: New York, NY, USA, 2022; pp. 616–619. [Google Scholar] [CrossRef]
  21. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal Policy Optimization Algorithms. arXiv 2017, arXiv:1707.063472017. Available online: http://arxiv.org/abs/1707.06347 (accessed on 7 April 2023).
  22. Lu, J.; Zhao, Y.-B.; Kang, Y.; Wang, Y.; Deng, Y. Strategy Generation Based on DDPG with Prioritized Experience Replay for UCAV. In Proceedings of the 2022 International Conference on Advanced Robotics and Mechatronics (ICARM), Guilin, China, 9–11 July 2022; IEEE: New York, NY, USA, 2022; pp. 157–162. [Google Scholar] [CrossRef]
  23. Wei, Y.-J.; Zhang, H.-P.; Huang, C.-Q. Maneuver Decision-Making For Autonomous Air Combat Through Curriculum Learning And Reinforcement Learning With Sparse Rewards. arXiv 2023, arXiv:2302.05838. Available online: http://arxiv.org/abs/2302.05838 (accessed on 7 March 2023).
  24. Hu, Y.; Wang, W.; Jia, H.; Wang, Y.; Chen, Y.; Hao, J.; Wu, F.; Fan, C. Learning to Utilize Shaping Rewards: A New Approach of Reward Shaping. Adv. Neural Inf. Process. Syst. 2020, 33, 15931–15941. [Google Scholar]
  25. Piao, H.; Sun, Z.; Meng, G.; Chen, H.; Qu, B.; Lang, K.; Sun, Y.; Yang, S.; Peng, X. Beyond-Visual-Range Air Combat Tactics Auto-Generation by Reinforcement Learning. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; IEEE: New York, NY, USA, 2020; pp. 1–8. [Google Scholar] [CrossRef]
  26. Wang, A.; Zhao, S.; Shi, Z.; Wang, J. Over-the-Horizon Air Combat Environment Modeling and Deep Reinforcement Learning Application. In Proceedings of the 2022 4th International Conference on Data-driven Optimization of Complex Systems (DOCS), Chengdu, China, 28–30 October 2022; IEEE: New York, NY, USA, 2022; pp. 1–6. [Google Scholar] [CrossRef]
  27. Hu, J.; Wang, L.; Hu, T.; Guo, C.; Wang, Y. Autonomous Maneuver Decision Making of Dual-UAV Cooperative Air Combat Based on Deep Reinforcement Learning. Electronics 2022, 11, 467. [Google Scholar] [CrossRef]
  28. Zhan, G.; Zhang, X.; Li, Z.; Xu, L.; Zhou, D.; Yang, Z. Multiple-UAV Reinforcement Learning Algorithm Based on Improved PPO in Ray Framework. Drones 2022, 6, 166. [Google Scholar] [CrossRef]
  29. Narvekar, S.; Sinapov, J.; Stone, P. Autonomous Task Sequencing for Customized Curriculum Design in Reinforcement Learning. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2536–2542. [Google Scholar] [CrossRef]
  30. Schmidhuber, J. Learning to generate subgoals for action sequences. In Proceedings of the IJCNN-91-Seattle International Joint Conference on Neural Networks, Seattle, WA, USA, 8–12 July 1991; IEEE: New York, NY, USA, 1991; Volume II, p. 453. [Google Scholar] [CrossRef]
  31. Rane, S. Learning with Curricula for Sparse-Reward Tasks in Deep Reinforcement Learning; Massachusetts Institute of Technology: Cambridge, MA, USA, 2020. [Google Scholar]
  32. Tobin, J.; Fong, R.; Ray, A.; Schneider, J.; Zaremba, W.; Abbeel, P. Domain randomization for transferring deep neural networks from simulation to the real world. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, USA, 24 September 2017; IEEE: New York, NY, USA, 2017; pp. 23–30. [Google Scholar] [CrossRef]
  33. Comanici, G.; Precup, D. Optimal Policy Switching Algorithms for Reinforcement Learning. In Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems, Montreal, QC, Canada, 10–14 May 2010; Volume 1, pp. 709–714. [Google Scholar]
  34. Frans, K.; Ho, J.; Chen, X.; Abbeel, P.; Schulman, J. Meta Learning Shared Hierarchies. arXiv 2017, arXiv:1710.09767. Available online: http://arxiv.org/abs/1710.09767 (accessed on 20 February 2023).
  35. Zhao, H.; Stretcu, O.; Smola, A.J.; Gordon, G.J. Efficient Multitask Feature and Relationship Learning. PMLR 2020, 115, 777–787. [Google Scholar]
  36. Barto, A.G.; Mahadevan, S. Recent Advances in Hierarchical Reinforcement Learning. Discret. Event Dyn. Syst. 2003, 13, 41–77. [Google Scholar] [CrossRef]
  37. Kaelbling, L.P.; Littman, M.L.; Cassandra, A.R. Planning and acting in partially observable stochastic domains. Artif. Intell. 1998, 101, 99–134. [Google Scholar] [CrossRef]
  38. Sutton, R.S.; Precup, D.; Singh, S. Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artif. Intell. 1999, 112, 181–211. [Google Scholar] [CrossRef]
  39. Eppe, M.; Gumbsch, C.; Kerzel, M.; Nguyen, P.D.H.; Butz, M.V.; Wermter, S. Intelligent problem-solving as integrated hierarchical reinforcement learning. Nat Mach Intell 2022, 4, 11–20. [Google Scholar] [CrossRef]
  40. Wen, Z.; Precup, D.; Ibrahimi, M.; Barreto, A. On Efficiency in Hierarchical Reinforcement Learning. Adv. Neural Inf. Process. Syst. 2020, 33, 6708–6718. [Google Scholar]
  41. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing Atari with Deep Reinforcement Learning. arXiv 2013, arXiv:1312.5602. Available online: http://arxiv.org/abs/1312.5602 (accessed on 7 April 2023).
  42. Littman, M.L. A tutorial on partially observable Markov decision processes. J. Math. Psychol. 2009, 53, 119–125. [Google Scholar] [CrossRef]
  43. White, D.J. A Survey of Applications of Markov Decision Processes. J. Oper. Res. Soc. 1993, 44, 1073–1096. [Google Scholar] [CrossRef]
  44. Wang, L.; Wei, H. Research on Autonomous Decision-Making of UCAV Based on Deep Reinforcement Learning. In Proceedings of the 2022 3rd Information Communication Technologies Conference (ICTC), Nanjing, China, 6–8 May 2022; IEEE: New York, NY, USA, 2022; pp. 122–126. [Google Scholar] [CrossRef]
  45. Duan, Y.; Chen, X.; Houthooft, R.; Schulman, J.; Abbeel, P. Benchmarking Deep Reinforcement Learning for Continuous Control. PMLR 2016, 48, 1329–1338. [Google Scholar]
  46. Vogeltanz, T. A Survey of Free Software for the Design, Analysis, Modelling, and Simulation of an Unmanned Aerial Vehicle. Arch. Comput. Methods Eng. 2016, 23, 449–514. [Google Scholar] [CrossRef]
  47. Chandak, Y.; Theocharous, G.; Kostas, J.E.; Jordan, S.M.; Thomas, P.S. Learning Action Representations for Reinforcement Learning. PMLR 2019, 97, 941–950. [Google Scholar]
  48. Pope, A.P.; Ide, J.S.; Micovic, D.; Diaz, H.; Twedt, J.C.; Alcedo, K.; Walker, T.T.; Rosenbluth, D.; Ritholtz, L.; Javorsek, D. Hierarchical Reinforcement Learning for Air Combat At DARPA’s AlphaDogfight Trials. IEEE Trans. Artif. Intell. 2022. Early Access. [Google Scholar] [CrossRef]
  49. Chen, W.; Gao, C.; Jing, W. Proximal policy optimization guidance algorithm for intercepting near-space maneuvering targets. Aerosp. Sci. Technol. 2023, 132, 108031. [Google Scholar] [CrossRef]
Figure 1. Markov process model.
Figure 1. Markov process model.
Entropy 25 01409 g001
Figure 2. Vehicle situational model.
Figure 2. Vehicle situational model.
Entropy 25 01409 g002
Figure 3. Vehicle control model.
Figure 3. Vehicle control model.
Entropy 25 01409 g003
Figure 4. Model structure and training framework.
Figure 4. Model structure and training framework.
Entropy 25 01409 g004
Figure 5. Angle tracking performance: comparison of models ((af) represents the horizontal tracking trajectory of model 1 with HRL framework, the horizontal tracking trajectory of model 2 with HRL framework, the horizontal tracking trajectory of model 3 with HRL framework, the horizontal tracking trajectory of model 1 without HRL framework, the horizontal tracking trajectory of model 2 without HRL framework, and the horizontal tracking trajectory of model 3 without HRL framework, respectively).
Figure 5. Angle tracking performance: comparison of models ((af) represents the horizontal tracking trajectory of model 1 with HRL framework, the horizontal tracking trajectory of model 2 with HRL framework, the horizontal tracking trajectory of model 3 with HRL framework, the horizontal tracking trajectory of model 1 without HRL framework, the horizontal tracking trajectory of model 2 without HRL framework, and the horizontal tracking trajectory of model 3 without HRL framework, respectively).
Entropy 25 01409 g005
Figure 6. Comparison of angle tracking states of different models in the same scene.
Figure 6. Comparison of angle tracking states of different models in the same scene.
Entropy 25 01409 g006
Figure 7. Comparison of the height tracking performance of the models ((af) represents the vertical tracking trajectory of model 1 with HRL frame, the vertical tracking trajectory of model 2 with HRL frame, the vertical tracking trajectory of model 3 with HRL frame, the vertical tracking trajectory of model 1 without HRL frame, the vertical tracking trajectory of model 2 without HRL frame, and the vertical tracking trajectory of model 3 without HRL frame, respectively).
Figure 7. Comparison of the height tracking performance of the models ((af) represents the vertical tracking trajectory of model 1 with HRL frame, the vertical tracking trajectory of model 2 with HRL frame, the vertical tracking trajectory of model 3 with HRL frame, the vertical tracking trajectory of model 1 without HRL frame, the vertical tracking trajectory of model 2 without HRL frame, and the vertical tracking trajectory of model 3 without HRL frame, respectively).
Entropy 25 01409 g007
Figure 8. Comparison of different models in the same scene with height-tracked states.
Figure 8. Comparison of different models in the same scene with height-tracked states.
Entropy 25 01409 g008
Figure 9. Tracking the trajectory of the HRL modified model 2 agent.
Figure 9. Tracking the trajectory of the HRL modified model 2 agent.
Entropy 25 01409 g009
Figure 10. Three-dimensional and 2D trajectories when the target and tracker are using the HRL agent.
Figure 10. Three-dimensional and 2D trajectories when the target and tracker are using the HRL agent.
Entropy 25 01409 g010
Figure 11. Tracking the trajectory of the HRL-free model 2 agent.
Figure 11. Tracking the trajectory of the HRL-free model 2 agent.
Entropy 25 01409 g011
Figure 12. Three-dimensional and 2D trajectories when the target and tracker are not using the HRL agent.
Figure 12. Three-dimensional and 2D trajectories when the target and tracker are not using the HRL agent.
Entropy 25 01409 g012
Figure 13. The relationship between the reward and episodes of the models.
Figure 13. The relationship between the reward and episodes of the models.
Entropy 25 01409 g013
Figure 14. The relationship between the loss and episodes of the models.
Figure 14. The relationship between the loss and episodes of the models.
Entropy 25 01409 g014
Table 1. Comparison of experimental results with and without the HRL framework.
Table 1. Comparison of experimental results with and without the HRL framework.
Reward TypeMiss Distance (m)Miss Angle (°)Approach Time (s)
without hrlwith hrlRatio of Decreasewithout hrlwith hrlRatio of Decreasewithout hrlwith hrlRatio of Decrease
Model 144,378.83
(±4020.2)
38,900.41
(±3778.3)
12.34%32.56
(±5.73)
25.71
(±1.46)
21.04%141.99
(±19.89)
101.459
(±20.09)
28.54%
Model 241,696.70
(±3692.7)
35,797.28
(±3494.9)
14.14%31.68
(±2.66)
28.726
(±8.39)
9.32%137.06
(±21.32)
87.56
(±17.83)
36.11%
Model 343,526.82
(±2332.4)
38,427.20
(±2371.3)
11.71%36.69
(±6.93)
25.69
(±6.41)
29.98%102.89
(±25.18)
89.01
(±18.66)
13.49%
Reward TypeHold Distance Time (%)Hold Angle Time (%)Cost Time (ms)
without hrlwith hrlRatio of Increasewithout hrlwith hrlRatio of Increasewithout hrlwith hrlRatio of Increase
Model 117.15%
(±6.39%)
33.14%
(±3.87%)
15.99%5.23%
(±0.22)
11.38%
(±0.99%)
6.15%0.971.42747.11%
Model 224.79%
(±4.99%)
41.12%
(±4.67%)
16.33%7.2%
(±0.63%)
15.44%
(±1.67%)
8.24%0.941.34843.40%
Model 319.30%
(±4.61%)
36.56%
(4.51%)
17.26%4.06%
(±1.07%)
14.07%
(±1.72%)
10.01%0.911.41054.95%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, R.; Li, H.; Yan, G.; Peng, H.; Zhang, Q. Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit. Entropy 2023, 25, 1409. https://doi.org/10.3390/e25101409

AMA Style

Chen R, Li H, Yan G, Peng H, Zhang Q. Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit. Entropy. 2023; 25(10):1409. https://doi.org/10.3390/e25101409

Chicago/Turabian Style

Chen, Ruihai, Hao Li, Guanwei Yan, Haojie Peng, and Qian Zhang. 2023. "Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit" Entropy 25, no. 10: 1409. https://doi.org/10.3390/e25101409

APA Style

Chen, R., Li, H., Yan, G., Peng, H., & Zhang, Q. (2023). Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit. Entropy, 25(10), 1409. https://doi.org/10.3390/e25101409

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop