Next Article in Journal
A Novel Collision Avoidance Strategy with D2D Communications for UAV Systems
Next Article in Special Issue
Minimum-Effort Waypoint-Following Differential Geometric Guidance Law Design for Endo-Atmospheric Flight Vehicles
Previous Article in Journal
Exterminator for the Nests of Vespa velutina nigrithorax Using an Unmanned Aerial Vehicle
Previous Article in Special Issue
A Review of Swarm Robotics in a NutShell
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Dual Aircraft Maneuver Formation Controller for MAV/UAV Based on the Hybrid Intelligent Agent

1
School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China
2
Digital Navigation Center, Beihang University, Beijing 100191, China
3
Science and Technology on Aircraft Control Laboratory, Beihang University, Beijing 100191, China
4
Beijing Aerospace Automatic Control Institute, Beijing 100854, China
*
Author to whom correspondence should be addressed.
Drones 2023, 7(5), 282; https://doi.org/10.3390/drones7050282
Submission received: 28 February 2023 / Revised: 12 April 2023 / Accepted: 18 April 2023 / Published: 22 April 2023
(This article belongs to the Special Issue Intelligent Coordination of UAV Swarm Systems)

Abstract

:
This paper proposes a hybrid intelligent agent controller (HIAC) for manned aerial vehicles (MAV)/unmanned aerial vehicles (UAV) formation under the leader–follower control strategy. Based on the high-fidelity three-degrees-of-freedom (DOF) dynamic model of UAV, this method decoupled multiple-input-multiple-output (MIMO) systems into multiple single-input-single-output (SISO) systems. Then, it innovatively combined the deep deterministic policy gradient (DDPG) and the double deep Q network (DDQN) to construct a hybrid reinforcement learning-agent model, which was used to generate onboard desired state commands. Finally, we adopted the dynamic inversion control law and the first-order lag filter to improve the actual flight-control process. Under the working conditions of a continuous S-shaped large overload maneuver for the MAV, the simulations verified that the UAV can achieve accurate tracking for the complex trajectory of the MAV. Compared with the traditional linear quadratic regulator (LQR) and DDPG, the HIAC has better control efficiency and precision.

1. Introduction

Aiming at increasingly fast-paced and high-intensity air combat, the use of MAVs as combat operations leaders with a certain number of UAVs as wingers to form a hybrid formation of UAV/MAV has become the development trend for future air confrontations. Among them, the two-aircraft formation consisting of an MAV and a UAV is one of the most typical combat styles. In MAV/UAV formations, the unmanned system must be able to share information and carry out cooperative operations with the manned systems across systematic boundaries [1]. The Fast Lightweight Autonomy (FLA) Program by the Defense Advanced Research Projects Agency (DARPA) has developed an advanced algorithm that enables an MAV or a UAV to operate autonomously without a human operator, the Global Positioning System (GPS), or any data resources. DARPA’s Lifelong Learning Machines (L2M) Project also aims to develop new machine learning methods that enable unmanned systems to continuously adapt to new environments and remember what they have learned [2]. Meanwhile, the U.S. Air Force’s Loyal Wingman Program aims to enhance the autonomy of UAVs and improve their combat capabilities in complex war environments [3]. Moreover, the recently proposed Skyborg program is working on the combination of manned and unmanned combat aerial vehicles. Therefore, improving the capability of autonomous flight control has become an important direction for the development of future UAV technology.
One of the research hotspots of UAV autonomous control capability is the formation flight-control problem [4]. In terms of the traditional design of the formation controller, Ref. [5] proposed a sliding mode controller for MAV/UAV formation flights based on a layered architecture. However, it makes extensive simplifications on the strong nonlinear dynamic model of MAV/UAVs, and was only validated by simulations for flat trajectories. Ref. [6] considered sensor noise and developed a leader–follower formation PID controller for multi-robots, which can achieve better performance in limiting position deviations. Furthermore, Ref. [7] proposed a parallel approach control law for fixed-wing UAV formations under the leader–follower strategy. Ref. [8] referred to an idea of multi-channel decoupling that split the MIMO system into multiple SISO systems and used sliding mode control to track the reference trajectory, which can be further applied to formation-control problems. Refs. [9,10] proposed a consensus-based multiple aircraft cooperative formation control method, but the consensus theory analysis was highly dependent on the linearized dynamic model, which limited its further application in a complex nonlinear dynamic system. Refs. [11,12] developed a formation controller where the commands were generated independently of the dynamic model, decreasing the control precision in extreme working conditions. Refs. [13,14] considered the confrontation situation and adopted pre-defined maneuver strategy collections, taking typical maneuvers as the basic units and building a collection of maneuver strategies with free combinations of various basic units. However, due to the model uncertainty and non-cooperative environment, this method hardly dealt with complex working conditions. Therefore, the intelligent agent method has become a novel research trend because of its weak model dependence and strong ability in terms of strategy exploration. Refs. [15,16] adopted deep neural networks to learn aircraft-maneuvering strategies and made progress in enhancing the autonomous maneuvering capability of UAVs. However, UAV formation control is a high-dimensional dynamical control problem with tightly coupled variables. When traditional neural networks learn such complex behaviors, they cause problems such as low training efficiency and difficulty in stable convergence [17]. Among the novel neural networks, the double deep Q network (DDQN) algorithm has shown good performance in control problems with discrete action sets by fitting the value functions of state actions through neural networks [18,19,20], but it cannot be applied to control problems with continuous variables. Based on the deterministic policy gradient (DPG) algorithm, DeepMind proposed the deep deterministic policy gradient (DDPG) algorithm which is proven to perform well on many kinds of continuous control problems [21,22,23]. However, in the field of aircraft control, the large variation in the angle of attack commands will increase the load on the attitude control loop [24]. Meanwhile, when it comes to complex tasks with multiple continuous control variables problems, DDPG has problems with unstable networks and low exploration efficiency [25,26,27,28]. For the above dilemma, some scholars have turned to hybrid reinforcement learning methods in recent years. By adding discrete “meta-actions” to continuous control problems, Ref. [29] partially solved the reinforcement learning traps and improved exploration efficiency. The experiments verified its superiority to the traditional continuous strategy algorithm in some cases. [30] proposed the parametrized deep Q-network for the hybrid action space without approximation or relaxation, which provides a reference for solving the hybrid control problem.
Based on the above analysis, it is obvious that the formation controller must be able to better adapt to complex flight conditions in future confrontation situations, e.g., continuous large overload maneuvers for the MAV, etc. Therefore, inspired by [29], we propose a hybrid reinforcement intelligent agent controller based on the decoupling of multi-channels, which can effectively solve the problem of formation-tracking under continuous maneuvering conditions. It should be emphasized that when designing controllers based on artificial intelligent methods, especially when the reinforcement learning controller is directly applied to the generation of underlying flight-control commands, the lack of flight dynamic constraints can easily bring about problems. Due to the lack of dynamic constraints, the attitude control system cannot quickly track the commands, leading to flight instability. Therefore, this paper introduced the dynamic inversion controller and the first-order lag filter to the hybrid reinforcement learning agent to enhance the smoothness and executability of control commands.
In summary, the main contributions of this paper are as follows:
(1) A hybrid intelligent agent was designed based on the novel concept of “meta-action” to further enhance formation control performance. The hybrid intelligent agent combined DDPG and DDQN according to the specific formation control targets;
(2) The framework of the HIAC was developed that combined the dynamic inversion controller and the first-order lag filter with the hybrid intelligent agent to effectively overcome the common drawbacks of reinforcement learning;
(3) The superiority of the HIAC method was validated with experiments of nominal conditions. Monte Carlo simulations with different initial conditions were then conducted to verify the adaptability of the HIAC.
The organization of this paper is as follows: Section 2 establishes the UAV dynamic model and formation-control targets. Section 3 designs the novel formation controller HIAC based on the DDPG/DDQN hybrid intelligent agent. The dynamic inversion controller and first-order lag filter are introduced to the framework of the HIAC as well. Section 4 conducts the experiments of nominal conditions and 100 Monte Carlo simulations with varying initial conditions. Finally, we summarize the research conclusion of this paper in Section 5.

2. Mathematical Modeling

2.1. UAV Dynamic Model

The main concern in dual aircraft formation flights is the real-time position, velocity, and attitude of the two aircraft, so it is necessary to establish a dynamic model of the UAV according to the forces on the mass as shown in Figure 1. To simplify the problem, the constraints flight envelope is ignored.
In the ground inertial coordinate system o x y z , V is the UAV flight velocity. γ and ψ are the flight path angle and flight azimuth angle, respectively. The flight adopts the Bank-To-Turn (BTT), which is considered to have no sideslip. α is the attack angle, and σ is the bank angle. The engine thrust and drag of the aircraft are denoted by T and D , respectively. n is the normal overload of the UAV in the velocity coordinate system o x V y V z V . Ignoring the wind disturbance in the flight, the three-degrees-of-freedom of the dynamic model for the UAV is established as follows [31,32,33]:
H = x ˙ = V cos γ sin ψ y ˙ = V cos γ cos ψ z ˙ = V sin γ V ˙ = T D / m g sin γ γ ˙ = g n cos σ cos γ / V ψ ˙ = g n sin σ / V cos γ ,
where m is the weight of the aircraft, which is considered constant in this paper, and g is the local gravity.
The engine thrust T can be denoted by
T = η T max ,
where η is the throttle manipulator, and its range is defined as 0 ,   1 . T max is the maximum thrust that the engine can achieve.
The air drag D consists of the parasite drag and the induced drag, which can be expressed as follows [31]:
D = C D P ρ V 2 S / 2 + 2 C D I n 2 m 2 g 2 / ρ V 2 S ,
where S is the reference area of the UAV. C D P is the parasite drag coefficient. C D I is the induced drag coefficient. ρ is the atmospheric density, which varies with the altitude of the aircraft in the stratosphere. It is calculated by [34]
ρ = ρ 0 e z / z 0 ,
where ρ 0 = 1.225 kg/m3 and z 0 = 6700 m.

2.2. Formation Control Targets

In this paper, the formation control target of the UAV was determined based on the leader–follower formation strategy. Taking a typical dual aircraft formation flight as an example, the formation configuration of the MAV/UAV was designed as shown in Figure 2. Since the reference trajectory of the MAV as the leader aircraft is known, the flight velocity, attitude, and position can be obtained from the sensors mounted within the MAV. The winger aircraft can receive real-time flight data from the MAV through the onboard data chain and complete the trajectory tracking and formation control autonomously. During the flight, it is required that the UAV and MAV keep a specific formation throughout the whole flight, as shown in Figure 2.

2.2.1. Flight Velocity Control Targets

The MAV and UAV keep the same formation flight velocity. The reference velocity of the MAV is V L , and the UAV velocity is V W , then the velocity deviation Δ V is
Δ V = V L V W .
The MAV and UAV keep the same flight path angle in formation flight. The MAV flight path angle is γ L , and the UAV flight path angle is γ W , then the flight path angle deviation Δ γ is
Δ γ = γ L γ W .
The MAV and UAV keep the same flight azimuth angle in formation flight. The flight azimuth angle of the MAV is ψ L , and the flight azimuth angle of the UAV is ψ W , then the deviation of the flight azimuth angle Δ ψ is
Δ ψ = ψ L ψ W .
The flight velocity and attitudes of the UAV should be consistent with the MAV within an allowable error
Δ V V Δ max , Δ γ γ Δ max , Δ ψ ψ Δ max ,
where V Δ max , γ Δ max , ψ Δ max represent the error thresholds of the velocity, flight path angle, and flight azimuth angle of the UAV, respectively.

2.2.2. Flight Distance Control Targets

The UAV is located around the MAV and maintains the specified formation distance. Δ D denote the distance between the MAV and the UAV in the ground inertial coordinate system. Δ D x , Δ D y and Δ D z denote the spatial distance of Δ D as follows:
Δ D = Δ D x 2 + Δ D y 2 + Δ D z 2 .
Summarily, the UAV should keep a distance larger than the safe flight distance from the MAV, which is as follows:
D Δ min Δ D D Δ max ,
where D Δ min and D Δ max represent the thresholds of the safe distance.

3. Design of the HIAC

The HIAC first adopted a DDPG/DDQN hybrid reinforcement learning method to train the agent model to generate the tracking commands. Then, we further designed a dynamic inversion controller and a first-order lag filter to construct an improved formation flight controller. Overall, the HIAC consists of three parts, i.e., desired state command solver, dynamic inversion controller, and first-order lag filter. The framework of the HIAC is shown in Figure 3.
In order to track the MAV, the HIAC adopted the current deviation between the states of the UAV and the states of the MAV, i.e., Δ V , Δ γ and Δ ψ as inputs, and outputs the control commands of the thrust, normal overload, and bank angle, i.e., η , n and σ . The main difference of the HIAC from other traditional controllers is that to further enhance the control accuracy, the HIAC adopted a hybrid intelligent agent as the desired command solver to generate the desired commands, V c , γ c and ψ c . Then, these commands were sent to the dynamic inversion controller to generate the control commands, η c , n c and σ c . Finally, the first-order lag filter further smoothed η c , n c and σ c to improve the executability of these commands. The three parts will be introduced in detail as the order of the information flow.

3.1. Desired Command Solver

Learning from the idea of “meta-action”, we partially discretized the control variables in the continuous control problems and developed a continuous–discrete mixed action space according to the characteristics of these control variables. Based on this process, we constructed a hybrid intelligent agent based on DDPG and DDQN to control V , γ , ψ and D of the UAV.

3.1.1. Framework of Hybrid Intelligent Agent Based on DDPG/DDQN

Based on the traditional Q-Learning algorithm, DDQN uses the neural network to fit the value function. It adopts discrete action sets to define the strategy and evaluates the Q value of the generated strategy through the Critic network. Compared with the traditional DQN algorithm [18,19,20], DDQN decouples the action selection strategy of the Q value and the calculation of the Q value and solves the problem of overestimation of the Q value compared with the traditional methods.
DDPG adopts the Actor–Critic network based on DQN and uses continuous action sets to define the control strategy. The model consists of the Actor–Critic network, where the Critic evaluates the actions generated by the Actor, and the Actor feeds back the evaluation results to the Critic for policy optimization [23]. More proofs and conclusions of the DDQN and DDPG can be found in [18,23], respectively.
However, the DDQN and DDPG suffer from different drawbacks when applied in practical engineering. Although the DDQN is easier to converge when compared with DDPG, it can only deal with discrete and low-dimensional action spaces. However, most of the practical targets, especially physical control targets, have continuous and high-dimensional action spaces. Moreover, even though the continuous space can be transferred into the discrete space, DDQN will generate high high-dimensional action space in this process and finally cause quite low computational efficiency. Meanwhile, although DDPG can solve the problem of continuous and high-dimensional action spaces, it is more likely to diverge than DDQN. Therefore, learning from “meta-action”, we proposed a hybrid intelligent agent combining the DDQN and DDPG according to their complementary characteristics. Considering the value range and the control precision of V , γ and ψ , we adopted the idea of multi-channel decoupling to perform partial discretization of the action space. For the velocity control agent V c , the DDPG was used to generate the set of continuous state commands. Because the value range of V c is larger than γ c and ψ c , discretizing the continuous action space with high precision will lead to dimension explosion. Meanwhile, for the angle control agents γ c and ψ c , the DDQN is used to generate the set of discretized state commands. Combining the DDQN and DDPG can improve the capability of convergence when these two agents are trained together.
The framework of the desired commands solver was designed as shown in Figure 4. It includes three agents which process the variation of the state commands V c , γ c and ψ c , respectively. Based on the decoupling principles between different agents, each agent calculates the action A V , A γ and A ψ , and updates the desired state commands respectively. The outputs are executed by the flight-control system of the UAV and fed back the rewards of each agent. The total reward function R is expressed by
R = R D , V + R γ + R ψ ,
where R D , V , R γ and R ψ are components of R in each agent.
To construct an intelligent agent based on the DDPG/DDQN hybrid reinforcement learning network, it was necessary to transform the trajectory tracking problem into a Markov decision process, which mainly includes three parts, i.e., the state space, the action space, and the reward function.

3.1.2. State Space S

According to the targets of formation flight control, the state space S is designed as follows:
S = [ Δ V , Δ D , Δ V ˙ , Δ γ , Δ γ d t ,   Δ γ ˙ , Δ ψ , Δ ψ d t   , Δ ψ ˙ ] ,
where Δ D equals
Δ D = Δ D 0 + Δ V d t ,
where Δ D 0 is the flight distance deviation between the MAV and the UAV at the initial epoch. The integral items Δ V , Δ γ and Δ ψ are the cumulative deviation from the initial epoch till the current epoch. Δ V ˙ , Δ γ ˙ and Δ ψ ˙ is the deviation rate of the velocity, flight azimuth angle, and flight path angle.

3.1.3. Action Space A

The action space A is defined as follows:
A = A V , A γ , A ψ ,
where the action A V denotes the correction value of the UAV velocity commands Δ V c , the action A γ denotes the correction value of the UAV flight path angle commands Δ γ c , and the action A ψ denotes the correction value of the UAV flight azimuth angle commands Δ ψ c , i.e.,
Δ V c = A V , Δ γ c = A γ , Δ ψ c = A ψ Δ V c λ V c max c , Δ γ c λ γ c max , Δ ψ c λ ψ c max ,
where λ V c max , λ γ c max and λ ψ c max are the maximum of corrections, respectively. A V is used to generate the set of continuous velocity commands. A γ , A ψ is used to generate the set of discretized angle commands. Specifically, the discretization can be further expressed as follows:
Ω γ c = 2 λ γ c max / γ c + 1 Ω ψ c = 2 λ ψ c max / ψ c + 1
Ω γ c = A γ 0 , ± γ c , ± 2 γ c , , ± Ω γ c 1 γ c / 2 , ± λ γ c max Ω ψ c = A ψ 0 , ± ψ c , ± 2 ψ c , , ± Ω ψ c 1 ψ c / 2 , ± λ ψ c max .
Then, the update of desired state commands is
V c V c + Δ V c γ c γ c + Δ γ c ψ c ψ c + Δ ψ c .

3.1.4. Reward Function R

According to the formation control targets of UAVs, the reward function R was designed as follows:
R = R P + R N + R C
where R P is the reward sub-function, which gives a positive response when the flight state of the UAV meets the control targets. R N is the penalty sub-function, which gives a negative response when the flight states exceed the allowable error of the control target. R C is the command limiting function, which can limit the values of the control commands η c , n c , σ c . More specifically, R C can smooth the variation of the control commands to finally reduce energy consumption.
R P is calculated by
R P = 10 × ε D 2 + ε V 2 + ε γ 2 + ε ψ 2
where ε D , ε V , ε γ , and ε ψ are reward coefficients, which are defined as follows
ε D = 1 , D Δ min Δ D D Δ max 0 , Δ D < D Δ min   or   Δ D > D Δ max , ε V = 1 Δ V / V Δ max , Δ V V Δ max 0 , Δ V > V Δ max , ε γ = 1 Δ γ / γ Δ max , Δ γ < γ Δ max 0 , Δ γ γ Δ max , ε ψ = 1 Δ ψ / ψ Δ max , Δ ψ < ψ Δ max 0 , Δ ψ ψ Δ max .
R N is calculated by
R N = 100 × e D 2 + e V 2 + e γ 2 + e ψ 2
where e D , e V , e γ , and e ψ are penalty coefficients, which are defined as follows:
e D = 1 , Δ D < D Δ min   or   Δ D > D Δ max 0 , D Δ min Δ D D Δ max , e V = 1 , Δ V > 2 V Δ max Δ V / V Δ max 1 , V Δ max Δ V 2 V Δ max 0 , Δ V < V Δ max , e γ = 1 , Δ γ > 2 γ Δ max Δ γ / γ Δ max 1 , γ Δ max Δ γ 2 γ Δ max 0 , Δ γ < γ Δ max , e ψ = 1 , Δ ψ > 2 ψ Δ max Δ ψ / ψ Δ max 1 , ψ Δ max Δ ψ 2 ψ Δ max 0 , Δ ψ < ψ Δ max .
R C is calculated by
R C = 0.2 η c / η max + n c / n max + σ c / σ max .

3.2. Dynamic Inversion Controller

To realize tracking of the commands of a given flight trajectory, the dynamic inversion control law was designed as follows [35]
V ˙ c = ϖ V V c V γ ˙ c = ϖ γ γ c γ ψ ˙ c = ϖ ψ ψ c ψ
where ϖ V , ϖ γ , and ϖ ψ denote the bandwidth of the controller, respectively. V c , γ c , ψ c denote the desired state commands of the flight velocity, the flight path angle, and the flight azimuth angle, respectively.
Since the UAV commands follow the dynamic constraints by Equation (1), considering Equations (1), (2), and (25) yields
T c = η c T max = D + m ϖ V V c V + m g sin γ , N γ = ϖ γ V γ c γ / g + cos γ , N ψ = ϖ ψ V ψ c ψ cos γ / g ,
where N γ and N ψ denote the normal overload and lateral overload, respectively. The throttle δ c , normal overload n c and bank angle σ c are selected as the control commands. Then, the UAV control command was designed as follows:
F = η c = D + m ϖ V V c V + m g sin γ / T max n c = N γ 2 + N ψ 2 σ c = arctan N ψ / N γ .
Moreover, the control command must satisfy the constraints:
η min η c η max , 0 n c n max , σ c σ max
where η min and η max is the minimum and maximum values of the throttle commands, respectively. n max is the maximum value of the normal overload, and σ max is the maximum value of the bank angle.

3.3. First-Order Lag Filter

Considering the fact that the UAV cannot instantly complete the change of the engine thrust, normal overload, and bank angle, a first-order lag filter model was constructed to simulate the delayed variation processes of these three variables:
G = η ˙ = η c η / τ δ n ˙ = n c n / τ n σ ˙ = σ c σ / τ σ ,
where η c , n c , σ c represent the control commands of the throttle, normal overload, and bank angle, respectively. τ δ , τ n , and τ σ represent the response time of the UAV control system accordingly.
Summarily, considering Equations (1), (27), and (29), the UAV flight process can be presented by the control equations as follow:
F V c , γ c , ψ c T = η c , n c , σ c T G η c , n c , σ c T = η ˙ , n ˙ , σ ˙ T H η , n , σ T = V , γ , ψ T .
Equation (30) reveals the calculation process from the desired control commands to the actual control commands. It is clear that the premise to realize the formation flight is to acquire the desired control commands of the UAV V c , γ c , ψ c under the specific formation strategy. Then, the ultimate flight trajectory can be obtained by the Runge–Kutta method.

4. Simulation Validation

4.1. Simulation Design

Based on the 3-DOF dynamic model in this paper, the MAV was designed to make a complex maneuver and provide the reference trajectory and control commands, accordingly. Under the leader–follower formation strategy, the UAV adopts the HIAC, DDPG, and LQR to track the MAV and keep the dual aircraft formation, respectively. LQR is a commonly used guidance method for tracking multi-state trajectories in aerospace engineering and it has been validated by extensive flight tests [36,37]. Therefore, we compared the proposed method with LQR and DDPG to verify its superiority in the following Section 4.2 and Section 4.3. The design of DDPG is described in Section 3.1.
First, the experiment of nominal conditions was conducted to analyze the superiority of the proposed method in detail. Meanwhile, the initial values greatly affect the performance of the reinforcement learning models. Therefore, the generalization ability of the model was required to be fully verified. Then, 100 Monte Carlo experiments were conducted to verify the adaptability of this method to different initial conditions.
The simulations were conducted by Matlab2021a and the 3-DOF dynamic model was built by Simulink. The total simulation time was T , the simulation interval was Δ T , and the specific experimental parameters are shown in Table 1.
The training methods of DDPG and DDQN refer to [18,23], respectively. Learning rate, max episode, discount factor, and experience buffer length were set as the same for both DDPG and DDQN. In addition, the batch size of DDPG was set to 256, and the batch size of DDQN was set to 64. The specific parameters are shown in Table 2.

4.2. Basic Principles of LQR

The implementation of LQR mainly includes three parts: linearization of the motion model, design of the tracking controller for the reference trajectory, and solution of the feedback gain matrix.
By linearizing the dynamic model of the UAV in Equation (1) with small deviations, the linear system can be obtained as follows:
X ˙ = A X + B u .
Equation (31) can be expressed by
δ x ˙ δ y ˙ δ z ˙ δ V ˙ δ γ ˙ δ ψ ˙ = A 11 A 12 A 13 A 14 A 15 A 16 A 21 A 22 A 23 A 24 A 25 A 26 A 31 A 32 A 33 A 34 A 35 A 36 A 41 A 42 A 43 A 44 A 45 A 46 A 51 A 52 A 53 A 54 A 55 A 56 A 61 A 62 A 63 A 64 A 65 A 66 δ x δ y δ z δ V δ γ δ ψ + B 11 B 12 B 13 B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43 B 51 B 52 B 53 B 61 B 62 B 63 δ η δ n δ σ .
Set the given MAV trajectory as the reference, the state space is defined as follows:
δ x = x W x L , δ y = y W y L , δ z = z W z L , δ V = V W V L , δ γ = γ W γ L , δ ψ = ψ W ψ L .
The control commands are defined as follows:
δ η = η η L , δ n = n n L , δ σ = σ σ L ,
where A and B are the partial derivative coefficient matrix calculated according to the motion differential equation and the feature points of the reference trajectory. The calculation results are as follows:
A 11 = A 12 = A 13 = 0 , A 14 = cos γ sin ψ , A 15 = V sin γ sin ψ , A 16 = V cos γ cos ψ , A 21 = A 22 = A 23 = 0 , A 24 = cos γ cos ψ , A 25 = V sin γ cos ψ , A 26 = V cos γ sin ψ , A 31 = A 32 = A 33 = A 36 = 0 , A 34 = sin γ , A 35 = V cos γ , A 41 = A 42 = A 46 = 0 , A 43 = D z / m , A 44 = D V / m , A 45 = g cos γ , A 51 = A 52 = A 53 = A 56 = 0 , A 54 = g n cos σ cos γ / V 2 , A 55 = g sin γ / V , A 61 = A 62 = A 63 = A 66 = 0 , A 64 = g sin σ n / V 2 cos γ , A 65 = g n sin σ sin γ / V cos 2 γ , B 11 = B 12 = B 13 = B 21 = B 22 = B 23 = B 31 = B 32 = B 33 = 0 , B 41 = T max / m , B 42 = B 43 = 0 , B 51 = D n / m , B 52 = g cos σ / V , B 53 = g n sin σ , B 61 = 0 , B 62 = g sin σ / V cos γ , B 63 = g n cos σ / V cos γ .
where D z , D V and D n are the partial derivatives of the drag D on the feature point of the reference trajectory to the flight height z , velocity V and normal overload n respectively. Define the optimal control performance index from t 0 to t f as follows:
J = 0.5 t 0 t f X T t Q X t + u T t R u t   dt ,
where Q and R are the weight matrices of state and control respectively. Q is positive semi-definite and R is positive-definite. Then, there exists an optimal control law u * = K * X to minimize the above performance index, and the feedback gain matrix K * is
K * = K η 1 K η 2 K η 3 K η 4 K η 5 K η 6 K n 1 K n 2 K n 3 K n 4 K n 5 K n 6 K σ 1 K σ 2 K σ 3 K σ 4 K σ 5 K σ 6 ,
K * = R 1 B T P
where P is the solution of the Riccati equation. It is calculated by
P A A T P + P B R 1 B T P Q = 0 .
Define Q and R as follows:
Q = diag Q 1 , Q 2 , Q 3 , Q 4 , Q 5 , Q 6 , R = diag R 1 , R 2 , R 3 .
To reflect the impact of the flight relative distance in the dual aircraft formation flight. Set Q 1 = Q 2 = Q 3 and define δ D 2 = Δ D 2 = δ x 2 + δ y 2 + δ z 2 , then
J = 0.5 t 0 t f Q 1 δ D 2 + Q 4 δ V 2 + Q 5 δ γ 2 + Q 6 δ ψ 2 + R 1 δ η 2 + R 2 δ n 2 + R 3 δ σ 2 dt .
According to Bryson Law [38], Q and R are set as follows:
Q 1 D 2 Δ max = Q 4 V 2 Δ max = Q 5 γ 2 Δ max = Q 6 ψ 2 Δ max = R 1 η 2 max = R 2 n 2 max = R 3 σ 2 max .
Set Q 1 = 1 , then other parameters can be obtained. According to u * , the control commands can be obtained:
η = η L K η 1 δ x + K η 2 δ y + K η 3 δ z + K η 4 δ V + K η 5 δ γ + K η 6 δ ψ , n = n L K n 1 δ x + K n 2 δ y + K n 3 δ z + K n 4 δ V + K n 5 δ γ + K n 6 δ ψ , σ = σ L K σ 1 δ x + K σ 2 δ y + K σ 3 δ z + K σ 4 δ V + K σ 5 δ γ + K σ 6 δ ψ .
Since the feedback gains obtained at different feature points of the reference trajectory are different, the monotonic flights can be selected as an independent variable, and the feedback gain coefficient of the offline design can be interpolated to obtain the corresponding control commands.

4.3. Experiment of Nominal Conditions

In the experiment of nominal conditions, the initial position of the MAV was x L 0 = 0 m, y L 0 = 0 m, z L 0 = 10,000 m, V L 0 = 400 m/s, γ L 0 = π / 6 , ψ L 0 = 0 . The initial position of the UAV was x W 0 = 100 m, y W 0 = 100 m, z W 0 = 10,000 m, V W 0 = 400 m/s, γ W 0 = π / 6 , ψ W 0 = 0 .
The formation flight trajectories of MAV and UAV of the three methods are shown in Figure 5. The MAV is designed to make continuous S-shaped large maneuver with a maximum overload of about 4 g at 1 s, 11 s, 29 s and 41 s, respectively. Figure 5 indicates that the LQR, DDPG, and HIAC can realize the stable tracking of the given trajectory of the MAV under large, overloaded maneuvers and reach the target of the designed formation.
Figure 6a–c shows the control commands of the UAV, i.e., the thrust, the normal overload, and the bank angle, generated by the LQR, DDPG and HIAC, respectively, with reference commands of the MAV. Figure 7a–c shows the errors between the control commands of the LQR, DDPG, and HIAC and the reference commands of the MAV. Figure 6 illustrates that there are four peaks in the curves of the control commands due to the four large, overloaded maneuvers. Moreover, compared with the LQR and DDPG, the trend of the control commands of the HIAC can be better consistent with the MAV in thrust, normal overload, and bank angle. Especially in the control of the normal overload, the HIAC has mitigated the sharp change of the commands generated by the reinforcement learning controller to a certain extent. It can provide more smooth and executable control commands under large maneuvers. However, during the large maneuver of the MAV, in order to track the reference commands, it inevitably generates a certain amount of extra adjustment for the thrust, overload, and bank angle for the three methods.
Figure 8 shows the change of the three controlled states of the UAV, i.e., the velocity, the flight path angle, and the flight azimuth angle, generated by LQR, DDPG and HIAC. Figure 9 shows the deviation of the three controlled states and the relative distance. It can be seen from Figure 8 that the change trend of the controlled state of the HIAC is basically the same as that of the MAV, and the formation maintenance performance is obviously better than that of the LQR and DDPG. Especially, the HIAC can keep up with most of the fluctuations of the MAV in the flight velocity and the flight azimuth angle. Moreover, Figure 9 shows that compared with the LQR and DDPG, the control precision of the HIAC has been significantly improved, and the control deviation can rapidly decrease to nearly 0 under the large maneuver. Figure 9d indicates that the HIAC successfully limits the formation distance within the safe distance between 100 m and 600 m while LQR and DDPG fail. The LQR continuously accumulates distance deviation due to the velocity deviation during the flight, and ultimately, the formation distance reveals a divergent trend. Meanwhile, although the relative distance of the DDPG gradually converges, it still extends beyond the safe distance at the end of the flight.
Table 3 presents the root mean square (RMS) errors and maximum errors of the four controlled states of the LQR, DDPG and HIAC. It is clear that both the RMS error and maximum error of the HIAC are smaller than those of the LQR and DDPG. Moreover, the HIAC has a reduction of 5.81%, 70.44%, and 64.95%, respectively, in the RMS error of the velocity, flight path angle and flight azimuth angle compared with the LQR, and has a reduction of 60.35%, 55.32% and 69.47% in the maximum error of velocity, flight path angle and flight azimuth angle, respectively, compared with the LQR. The HIAC has a reduction of 36.10%, 35.85% and 51.61%, respectively, in the RMS error of velocity, flight path angle and flight azimuth angle compared with the DDPG, and has a reduction of 54.43%, 31.57% and 55.01% in the maximum error of velocity, flight path angle and flight azimuth angle, respectively, compared with the DDPG.
In summary, the proposed HIAC significantly improves the state control performance and guarantees that the flight distance stays within a safe distance as well.

4.4. Monte Carlo Experiments

In order to further test how the HIAC adapts to various initial conditions, 100 Monte Carlo simulations were carried out by adding random deviations to the nominal conditions.
The initial position of the MAV is x L 0 = 0 m, y L 0 = 0 m, z L 0 = 10,000 m, V L 0 = 400 m/s, γ L 0 = π / 6 , ψ L 0 = 0 . The baseline of initial values of the UAV is x W 0 = 100 m, y W 0 = 100 m, z W 0 = 10,000 m, V W 0 = 400 m/s, γ W 0 = π / 6 , ψ W 0 = 0 . Then, random deviations which follow the uniform distributions were added to these six baselines, respectively. The specific values of the deviations are presented in Table 4.
Figure 10 is the scatterplot of the Monte Carlo simulation results of the velocity errors, flight path angle error, flight azimuth angle, and relative distance for the LQR, DDPG and HIAC. For each evaluation index, the horizontal axis is the RMS error, and the vertical axis is the maximum error. It can be seen that the HIAC can fulfill the control target in the magnitude of velocity. Meanwhile, because the training threshold is set quite strictly in order to achieve better control performance, the maximum error and RMS error of the flight path angle and flight azimuth angle may extend out of the threshold when the extreme deviations are added to the initial values. However, the HIAC can still present a satisfactory control accuracy of the angle compared with the DDPG and LQR. Moreover, in terms of the safety distance, the HIAC can stay within a safe distance of 100 m to 600 m from the MAV, which reaches the distance control target. However, the DDPG and LQR gradually extend out of the safe distance as the initial values vary. Statistically, compared with LQR and DDPG, the HIAC has smaller values in both the RMS error and maximum error of these four evaluation indices. In summary, the performance of the HIAC in formation control is better than that of the other two methods, which is consistent with the simulation results under nominal conditions. It is believed that the HIAC has significant adaptability to the varying initial conditions.

5. Conclusions

In this study, a novel HIAC method was proposed, which is able to enhance the smoothness and executability of control commands and improve the control performance of the MAV/UAV flight formation. First, based on the idea of “meta-action” in hybrid reinforcement learning, the formation control was modeled as a continuous–discrete space control problem. Then, we proposed the framework of the HIAC, and the hybrid intelligent agent model based on the DDPG/DDQN was designed through multi-channel decoupling. Finally, we carried out simulations of nominal conditions and 100 Monte Carlo simulations in varying initial conditions. The simulation results showed that, compared with the traditional LQR and DDPG, the HIAC has better performance of high control precision and rapid convergence. Meanwhile, the adaptability of HIAC to the varying initial conditions was verified as well.
For further practical applications, HIAC can gradually support practical scenarios such as formation military operations and terrain surveys. In particular, two aspects should be considered when applying HIAC. The first is the reliability of the method. HIAC should be preliminarily trained with a large number of ground tests before the real flights, to ensure that intelligent control gradually takes authority over traditional flight-control methods. The second is the portability of the method. At present, the method supports the deployment of reinforcement learning on hardware such as DSP, and FPGA, and can realize airborne portability and the online training of agent models.

Author Contributions

Conceptualization, methodology, software, validation, formal analysis, resources, data curation, writing—original draft preparation, L.Z. (Luodi Zhao); writing—review and editing, Y.L. and Q.P.; visualization, investigation, Y.L.; supervision, project administration, funding acquisition, L.Z (Long Zhao). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation of China, grant number 42274037 and 41874034, the National key research and development program of China, grant number 2020YFB0505804 and the Beijing Natural Science Foundation, grant number 4202041.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, L.; Wang, T.; Jiang, Q. Key Technology Develop Trends of Unmanned Systems Viewed from Unmanned Systems Integrated Roadmap 2017—2042. Unmanned Syst. Technol. 2018, 1, 79–84. [Google Scholar]
  2. Mishory, J. DARPA Solicits Information for New Lifelong Machine Learning Program. Inside Pentagon 2017, 33, 10. [Google Scholar]
  3. Pittaway, N. Loyal Wingman. Air Int. 2019, 96, 12–13. [Google Scholar]
  4. Oh, K.; Park, M.; Ahn, H. A survey of multi-agent formation control. Automatica 2015, 53, 424–440. [Google Scholar] [CrossRef]
  5. Wang, H.; Liu, S.; Lv, M.; Zhang, B. Two-Level Hierarchical-Interaction-Based Group Formation Control for MAV/UAVs. Aerospace 2022, 9, 510. [Google Scholar] [CrossRef]
  6. Choi, I.S.; Choi, J.S. Leader-Follower formation control using PID controller. In Proceedings of the International Conference on Intelligent Robotics & Applications, Montreal, QC, Canada, 3–5 October 2012. [Google Scholar]
  7. Gong, Z.; Zhou, Z.; Wang, Z.; Lv, Q.; Xu, Q.; Jiang, Y. Coordinated Formation Guidance Law for Fixed-Wing UAVs Based on Missile Parallel Approach Metho. Aerospace 2022, 9, 272. [Google Scholar] [CrossRef]
  8. Liang, Z.; Ren, Z.; Shao, X. Decoupling trajectory tracking for gliding reentry vehicles. IEEE/CAA J. Autom. Sin. 2015, 2, 115–120. [Google Scholar]
  9. Kuriki, Y.; Namerikawa, T. Formation Control of UAVs with a Fourth-Order Flight Dynamics. J. Control. Meas. Syst. Integr. 2014, 7, 74–81. [Google Scholar] [CrossRef]
  10. Kuriki, Y.; Namerikawa, T. Consensus-based cooperative formation control with collision avoidance for a multi-UAV system. In Proceedings of the American Control Conference, Portland, OR, USA, 4–6 June 2014. [Google Scholar]
  11. Atn, G.M.; Stipanovi, D.M.; Voulgaris, P.G. Collision-free trajectory tracking while preserving connectivity in unicycle multi-agent systems. In Proceedings of the American Control Conference, Washington, DC, USA, 17–19 June 2013. [Google Scholar]
  12. Tsankova, D.D.; Isapov, N. Potential field-based formation control in trajectory tracking and obstacle avoidance tasks. In Proceedings of the Intelligent Systems, Sofia, Bulgaria, 6–8 September 2012. [Google Scholar]
  13. Hu, J.; Wang, L.; Hu, T. Autonomous Maneuver Decision Making of Dual-UAV Cooperative Air Combat Based on Deep Reinforcement Learning. Electronics 2022, 11, 467. [Google Scholar] [CrossRef]
  14. Luo, Y.; Meng, G. Research on UAV Maneuver Decision-making Method Based on Markov Network. J. Syst. Simul. 2017, 29, 106–112. [Google Scholar]
  15. Yang, Q.; Zhang, J.; Shi, G. Maneuver Decision of UAV in Short-Range Air Combat Based on Deep Reinforcement Learning. IEEE Access 2020, 8, 363–378. [Google Scholar] [CrossRef]
  16. Li, Y.; Han, W.; Wang, Y. Deep Reinforcement Learning with Application to Air Confrontation Intelligent Decision-Making of Manned/Unmanned Aerial Vehicle Cooperative System. IEEE Access 2020, 99, 67887–67898. [Google Scholar] [CrossRef]
  17. Wang, X.; Gu, Y.; Cheng, Y. Approximate Policy-Based Accelerated Deep Reinforcement Learning. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 1820–1830. [Google Scholar] [CrossRef] [PubMed]
  18. Hasselt, H.V.; Guez, A.; Silver, D. Deep Reinforcement Learning with Double Q-Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Canberra, Australia, 30 November–5 December2015. [Google Scholar]
  19. Mnih, V.; Kavukcuoglu, K.; Silver, D. Playing Atari with deep reinforcement learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  20. Silver, D.; Huang, A.; Maddison, C.J. Mastering the game of go with deep neural networks and the tree search. Nature 2016, 529, 484. [Google Scholar] [CrossRef]
  21. Silver, D.; Lever, G.; Heess, N. Deterministic policy gradient algorithms. In Proceedings of the 31st International Conference on Machine Learning, Beijing, China, 21–26 June 2014. [Google Scholar]
  22. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  23. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  24. Wada, D.; Araujo-Estrada, S.A.; Windsor, S. Unmanned Aerial Vehicle Pitch Control Using Deep Reinforcement Learning with Discrete Actions in Wind Tunnel Test. Aerospace 2021, 8, 18. [Google Scholar] [CrossRef]
  25. Haarnoja, T.; Zhou, A.; Abbeel, P. Soft Actor-Critic Algorithms and Applications. arXiv 2018, arXiv:1812.05905. [Google Scholar]
  26. Heess, N.; Silver, D.; Teh, Y.W. Actor-critic reinforcement learning with energy-based policies. In Proceedings of the Tenth European Workshop on Reinforcement Learning, Edinburgh, UK, 30 June–1 July 2012. [Google Scholar]
  27. Schaul, T.; Quan, J.; Antonoglou, I. Prioritized experience replay. arXiv 2015, arXiv:1511.05952. [Google Scholar]
  28. Hu, Z.; Wan, K.; Gao, X.; Zhai, Y.; Wang, Q. Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments. Sensors 2020, 20, 1890. [Google Scholar] [CrossRef] [PubMed]
  29. Neunert, M.; Abdolmaleki, A.; Wulfmeier, M. Continuous-Discrete Reinforcement Learning for Hybrid Control in Robotics. In Proceedings of the Conference on Robot Learning, Virtual Event, 30 October–1 November 2020. [Google Scholar]
  30. Xiong, J.; Wang, Q.; Yang, Z. Parametrized Deep Q-Networks Learning: Reinforcement Learning with Discrete-Continuous Hybrid Action Space. arXiv 2018, arXiv:1810.06394. [Google Scholar]
  31. Anderson, M.R.; Robbins, A.C. Formation flight as a cooperative game. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Boston, MA, USA, 10–12 August 1998. [Google Scholar]
  32. Kelley, H.J. Reduced-order modeling in aircraft mission analysis. AIAA J. 2015, 9, 349–350. [Google Scholar] [CrossRef]
  33. Williams, P. Real-time computation of optimal three-dimensional aircraft trajectories including terrain-following. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Keystone, CO, USA, 24–26 August 2006. [Google Scholar]
  34. Wang, X.; Guo, J.; Tang, S. Entry trajectory planning with terminal full states constraints and multiple geographic constraints. Aerosp. Sci. Technol. 2019, 84, 620–631. [Google Scholar] [CrossRef]
  35. Snell, S.A.; Enns, D.F.; Garrard, W.L. Nonlinear inversion flight control for a supermaneuverable aircraft. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Portland, OR, USA, 20–22 August 1990. [Google Scholar]
  36. Dukeman, G. Profile-Following Entry Guidance Using Linear Quadratic Regulator Theory. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Monterey, CA, USA, 5–8 August 2002. [Google Scholar]
  37. Wen, Z.; Shu, T.; Hong, C. A simple reentry trajectory generation and tracking scheme for common aero vehicle. In Proceedings of the AIAA Guidance, Navigation, and Control Conference, Minneapolis, MN, USA, 13–16 August 2012. [Google Scholar]
  38. Bryson, A.E.; Ho, Y. Applied Optimal Control. Technometrics 1979, 21, 3. [Google Scholar]
Figure 1. The forces on the center of gravity of the aircraft.
Figure 1. The forces on the center of gravity of the aircraft.
Drones 07 00282 g001
Figure 2. Dual aircraft formation for MAV/UAV.
Figure 2. Dual aircraft formation for MAV/UAV.
Drones 07 00282 g002
Figure 3. The framework of the HIAC.
Figure 3. The framework of the HIAC.
Drones 07 00282 g003
Figure 4. The framework of the desired commands solver.
Figure 4. The framework of the desired commands solver.
Drones 07 00282 g004
Figure 5. The formation flight trajectory of MAV and UAV of three methods.
Figure 5. The formation flight trajectory of MAV and UAV of three methods.
Drones 07 00282 g005
Figure 6. The control commands of the MAV generated by the LQR, DDPG, and HIAC, respectively, with reference commands of the MAV. The results of the thrust, the thrust, the normal overload, and the bank angle are presented in (ac), respectively.
Figure 6. The control commands of the MAV generated by the LQR, DDPG, and HIAC, respectively, with reference commands of the MAV. The results of the thrust, the thrust, the normal overload, and the bank angle are presented in (ac), respectively.
Drones 07 00282 g006
Figure 7. The errors between the control commands of the LQR, DDPG, HIAC and the reference commands of the MAV. The errors of the thrust, the thrust, the normal overload, and the bank angle are presented in (ac), respectively.
Figure 7. The errors between the control commands of the LQR, DDPG, HIAC and the reference commands of the MAV. The errors of the thrust, the thrust, the normal overload, and the bank angle are presented in (ac), respectively.
Drones 07 00282 g007
Figure 8. The change of the three controlled state of the UAV generated by the LQR, DDPG and HIAC. The results of the velocity, the flight path angle, and the flight azimuth angle are presented in (ac), respectively.
Figure 8. The change of the three controlled state of the UAV generated by the LQR, DDPG and HIAC. The results of the velocity, the flight path angle, and the flight azimuth angle are presented in (ac), respectively.
Drones 07 00282 g008
Figure 9. The deviation of the velocity, the flight path angle, the flight azimuth angle and the relative distance are presented in (ad), respectively.
Figure 9. The deviation of the velocity, the flight path angle, the flight azimuth angle and the relative distance are presented in (ad), respectively.
Drones 07 00282 g009
Figure 10. Monte Carlo simulation results of LQR, DDPG and HIAC.
Figure 10. Monte Carlo simulation results of LQR, DDPG and HIAC.
Drones 07 00282 g010
Table 1. The experimental parameter settings.
Table 1. The experimental parameter settings.
ParametersSettingsParametersSettings
T (s)50 η min 0
Δ T (s)0.1 η max 1
T max (lb)25,600 n max 6
m (kg)14,470 σ max (rad) π / 2
g (m/s2)9.81 D Δ max (m)600
S (ft2)400 D Δ min (m)100
C D P 0.02 V Δ max (m/s)50
C D I 0.1 ψ Δ max (rad)0.2
τ δ (s)0.6 γ Δ max (rad)0.2
τ n (s)0.5 λ V c max (m/s)50
τ σ (s)0.5 λ γ c max (rad) π / 2
ϖ V (s)0.3 λ ψ c max (rad) π / 2
ϖ γ (s)0.2 γ c (rad) π / 180
ϖ ψ (s)0.2 ψ c (rad) π / 180
Table 2. The training parameters of DDPG/DDQN.
Table 2. The training parameters of DDPG/DDQN.
ParametersSettings
Learning Rate0.0001
Max Episode25,000
Batch Size (DDPG)256
Batch Size (DDQN)64
Discount Factor0.99
Experience Buffer Length1 × 106
Table 3. RMS and maximum errors of the four states of the LQR, DDPG, and HIAC in nominal conditions.
Table 3. RMS and maximum errors of the four states of the LQR, DDPG, and HIAC in nominal conditions.
ControllerVelocity (m/s)Flight Path Angle (rad)Flight Azimuth Angle (rad)Relative Distance (m)
Safe Distance [100, 600]
LQRRMS5.69570.47370.6202516.7072
Max.15.33070.80270.7833710.2799
DDPGRMS8.39530.21830.4493444.1190
Max.13.33790.52410.5315610.6078
HIACRMS5.36470.14010.2174460.0709
Max.6.07800.35860.2391552.1845
Table 4. Uniform distribution of deviations for the six initial values.
Table 4. Uniform distribution of deviations for the six initial values.
Numbers of Monte Carlo SimulationsX (m)Y (m)Z (m)Velocity (m/s)Flight Path Angle (rad)Flight Azimuth Angle (rad)
100[−50, 550][−50, 550][−1000, 1000][−100, 100][− π /18, π /18][− π /18, π /18]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, L.; Liu, Y.; Peng, Q.; Zhao, L. A Dual Aircraft Maneuver Formation Controller for MAV/UAV Based on the Hybrid Intelligent Agent. Drones 2023, 7, 282. https://doi.org/10.3390/drones7050282

AMA Style

Zhao L, Liu Y, Peng Q, Zhao L. A Dual Aircraft Maneuver Formation Controller for MAV/UAV Based on the Hybrid Intelligent Agent. Drones. 2023; 7(5):282. https://doi.org/10.3390/drones7050282

Chicago/Turabian Style

Zhao, Luodi, Yemo Liu, Qiangqiang Peng, and Long Zhao. 2023. "A Dual Aircraft Maneuver Formation Controller for MAV/UAV Based on the Hybrid Intelligent Agent" Drones 7, no. 5: 282. https://doi.org/10.3390/drones7050282

APA Style

Zhao, L., Liu, Y., Peng, Q., & Zhao, L. (2023). A Dual Aircraft Maneuver Formation Controller for MAV/UAV Based on the Hybrid Intelligent Agent. Drones, 7(5), 282. https://doi.org/10.3390/drones7050282

Article Metrics

Back to TopTop