Next Article in Journal
YOLO-DroneMS: Multi-Scale Object Detection Network for Unmanned Aerial Vehicle (UAV) Images
Previous Article in Journal
CMDN: Pre-Trained Visual Representations Boost Adversarial Robustness for UAV Tracking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Insensitive Mechanism-Based Nonlinear Model Predictive Guidance for UAVs Intercepting Maneuvering Targets with Input Constraints

College of Intelligence Science and Technology, National University of Defence Technology, Changsha 410005, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(11), 608; https://doi.org/10.3390/drones8110608
Submission received: 13 August 2024 / Revised: 17 October 2024 / Accepted: 22 October 2024 / Published: 24 October 2024

Abstract

:
This paper proposed an innovative guidance strategy, denoted as NMPC-IM, which integrates the Insensitive Mechanism (IM) with Nonlinear Model Predictive Control (NMPC) for Unmanned Aerial Vehicle (UAV) pursuit-evasion scenarios, with the aim of effectively intercepting maneuvering targets with consideration of input constraints while minimizing average energy expenditure. Firstly, the basic principle of IM is proposed, and it is transformed into an additional cost function in NMPC. Secondly, in order to estimate the states of maneuvering target, a fixed-time sliding mode disturbance observer is developed. Thirdly, the UAV’s interception task is formulated into a comprehensive Quadratic Programming (QP) problem, and the NMPC-IM guidance strategy is presented, which is then improved by the adjustment of parameters and determination of maximum input. Finally, numerical simulations are carried out to validate the effectiveness of the proposed method, and the simulation results show that the NMPC-IM guidance strategy can decrease average energy expenditure by mitigating the impact of the target’s maneuverability, optimizing the UAV’s trajectory during the interception process.

1. Introduction

With the rapid advancement of Unmanned Aerial Vehicle (UAVs) technology, its role has evolved from singular tracking and reconnaissance to multiple types of mobility tasks [1,2,3]. This includes tasks such as airborne interception and multi-UAV cooperative tracking. This transformation signifies that UAVs are transitioning from traditional reconnaissance platforms to multi-tasking platforms, with the enhancement of their intelligence and autonomy levels becoming a critical factor in the future [4].
The UAV interception task primarily refers to the UAV employing tactical maneuvers to bring the target within its fire range. However, there are constraints of control input, and differences in maneuverability between the UAV and the target. When the UAV is confronted with a more maneuverable target, how it can accomplish the interception task under input constraints while minimizing average energy expenditure is a question that merits investigation. In this paper, it is simplified to a point-to-point guidance problem under conditional constraints.
A number of methods have been proposed to solve the UAV pursuit-evasion game. Ref. [5], considering obstacle avoidance and input constraint, proposed a collaborative strategy combining the Apollonius circle algorithm and the geometric algorithm. Ref. [6] introduced a Reinforcement Learning (RL)-based control strategy for the UAV pursuit-evasion game, ensuring equal formation and stability against disturbances, with a novel Euler-based proof and Hamilton–Jacobi–Isaacs (HJI) equation solution for the Nash equilibrium. Ref. [7] presents a robust swarm strategy for UAVs in urban pursuit-evasion scenarios, integrating predictive decision-making and demonstrating enhanced performance under normal and damaged conditions. Ref. [8] established a mathematical model of UAV pursuit and evasion countermeasures based on the DDPG (deep deterministic policy gradient) algorithm, and designed countermaneuver strategies for evading UAVs. Ref. [9] introduced an autonomous exploration and mapping system utilizing heterogeneous UAVs and UGVs in GPS-denied environments, featuring a two-layered exploration strategy and an optimized next-view planning framework for efficient 3D environment perception and navigation. Ref. [10] proposed a robust adaptive event-triggered control strategy for USV-UAV cooperative formation systems to enhance maritime parallel search efficiency. Ref. [11] proposed a varying cells strategy for UAV collision avoidance that integrates aerodynamic constraints, offering more flexible maneuvers and overcoming the limitations of fixed cell methods.
Model Predictive Control (MPC) utilizes a model of the system to predict its future behavior over a certain horizon, optimizing the control inputs to achieve desired objectives while satisfying constraints. This predictive capability makes MPC particularly attractive for the design of guidance laws, which require precise trajectory planning and control under various dynamic conditions and constraints. This solves the problem of violating constraint of lateral acceleration when using conventional proportional navigation [12].
There have been some applications of MPC to the guidance problem. Ref. [13] proposed a guidance law using Robust MPC optimized by a neural network. Ref. [14] proposed an MPC-based cooperative guidance approach, which launches a salvo attack for multi-missile engagement against a stationary target. Ref. [15] introduced an LPMPC-based suboptimal guidance law for 3D interception of ballistic missiles, and achieved high accuracy in impact angle and miss distance. Similarly, ref. [16] presented an LPMPC cooperative guidance law for multi-missile simultaneous attacks with impact angle constraints. It converts a nonlinear control problem into a solvable Quadratic Programming task. Ref. [17] proposed a hybrid missile guidance algorithm for planar engagement scenarios, which combined Nonlinear MPC with a collision cone approach and ensured that the impact angle lies in a predefined range. More recently, ref. [18] presented an MPC-based guidance system for drone interception with visual tracking and impact angle constraints. It optimizes reduced maneuvering at interception, tested across scenarios with both maneuvering and non-maneuvering targets.
In order to solve the problem of input constraints for the UAV interception task, we incorporate IM into the traditional NMPC guidance model and integrate it into an algorithm for the interception task with some additional operations. The primary contributions of this paper are as follows:
1.
We propose the basic concept of the Insensitive Mechanism (IM), which is expressed by concrete formulations through a UAV interception scenario. This concept is then transformed into an additional cost function within the NMPC framework.
2.
To estimate the target states while reducing the required information acquisition, we design a fixed-time sliding mode disturbance observer to obtain information on the target’s maneuverability.
3.
The UAV interception task is then formulated into a comprehensive Quadratic Programming (QP) problem with the consideration of soft constraints. Consequently, a guidance algorithm that integrates NMPC with IM is presented.
4.
To refine the algorithm, we adjust the parameters and qualitatively analyze the reasons. For the reason that NMPC considers more about the predicted position, we design a determination of maximum input to correct the trajectory in a timely manner.
The structure of the paper is as follows: Section 2 presents the engagement kinematics of UAVs and the basic theories of NMPC in guidance. Section 3 introduces the concept and description of IM, designs a fixed-time sliding mode disturbance observer, and proposes the enhanced NMPC algorithm with IM. In Section 4, the effectiveness and robustness of the proposed NMPC-IM algorithm are validated through simulation, and conclusions are drawn in Section 5.

2. Preliminaries and Problem Statement

To concretely represent the UAV interception, we assume that a UAV intercepts a target. The target is more maneuverable due to its smaller size. The task of the UAV is to minimize the average energy expenditure within the constraints to catch up with the target and destroy it once it reaches the attack range.
As shown in Figure 1, the target may be inside the cone of the UAV at the next moment or outside the cone, the latter meaning that the UAV is constrained by control input and cannot be pointed at the target in time.
In this section, we mainly introduce engagement kinematics of UAVs and NMPC in guidance.

2.1. Engagement Kinematics of UAVs

Assuming that the motion between the UAV and the target is in the same plane, a motion schematic is shown in Figure 2, where Oxy is the inertial frame of reference. Since we consider the UAV and target as particles during the interception and ignore their dynamics, we only need to consider the relative motion of the two points in the local inertial coordinate system.
In this figure, V M and V T represent the speed of the UAV and the target, respectively, which remain constant during the engagement. α M and α T represent their heading angles. The quantities of a M and a T represent the accelerations of M and T, respectively. r and θ represent the range and line-of-sight (LOS) angle to the T, respectively. The relationships between the above parameters are represented by the following differential equation:
r ˙ = V r = V T cos α T θ V M cos α M θ r θ ˙ = V θ = V T sin α T θ V M sin α M θ
In order to establish the state equations describing the motion between M and T, the state vector and control input are defined as x   = r V r θ V θ α M T , u = a M = V M α ˙ M , where V r and V θ represent the relative velocity components along and perpendicular to the LOS. The kinematics of the engagement can be expressed in the following:
r ˙ V ˙ r θ ˙ V ˙ θ α ˙ M = V r V θ 2 / r V θ / r V r V θ / r 0 + 0 sin α M θ 0 cos α M θ 1 / V M u + 0 a T sin θ α T 0 a T cos θ α T 0
And the compact form is:
x ˙ = f x + g x u + d
where u R , x R 5 , f : R 5 R 5 , g : R 5 R 5 , d R 5 . The maneuvrability of the target is represented by d , which is treated as an unknown bounded disturbance in many articles [17,19]. In this paper, d is considered to be measurable or estimatable, and is used to make improvements to the results.

2.2. Nonlinear Model Predictive Control

MPC typically requires transforming differential equations into discrete form. According to Equation (3), and without the disturbances, the discrete-time form using Euler discretization is expressed in the following:
x k + 1 = f d x k + g d x k u k
We use p and c to represent the prediction and control horizons, respectively. The increment in the control input is given by Δ u k = u k u k 1 . Then, the “referenced predictive form” [20] is
x k + j k = f d x k + j 1 k + g d x k + j 1 k u k + j 2 k + Δ u k + j 1 k
where j = 1 , , p , and k + j k means that the current time step is k and the distance from the current time step is j. With that, the standard prediction form of Equation (5), required for the NMPC formulations, is given by
X k = F k + g k u k 1 k 1 + G k Δ U k
where u k 1 k 1 is the control input of the last time step, X k R 5 p × 1 , F k R 5 p × 1 , g k R 5 p × 1 , G k R 5 p × c , and Δ U k R c × 1 , Δ U k = Δ u k k Δ u k + c 1 k T . Since the computation of x k + j k j = 1 , , p 1 requires Δ u k + i k i = 0 , , c 1 , which needs to be solved, the predicted states x k + j k cannot be obtained at the current step k, nor can F k , g k , G k . It is a good method to use the predicted states of the previous time step to calculate the above matrics [13]. Therefore, Equation (6) can be reformulated as
X k = F k 1 + g k 1 u k 1 k 1 + G k 1 Δ U k
where F k 1 = f d x k 1 k 1 f d x k k 1 f d x k + c 2 k 1 f d x k + p 2 k 1 , g k 1 = g d x k 1 k 1 g d x k k 1 g d x k + c 2 k 1 g d x k + p 2 k 1 ,
G k 1 = g d x k 1 k 1 0 5 × 1 0 5 × 1 g d x k k 1 g d x k k 1 0 5 × 1 g d x k + c 2 k 1 g d x k + c 2 k 1 g d x k + c 2 k 1 g d x k + p 2 k 1 g d x k + p 2 k 1 g d x k + p 2 k 1 .
In order to design the cost function, it is necessary to define a more appropriate output such as V θ . V θ = r θ ˙ , which contains both the position information between the M and T as well as the LOS rate. Therefore, the output is expressed as y = Cx , where C = 0 0 0 1 0 . Similarly, y d denotes the desired output, and usually takes a value like y d = c y r ,   c y 0 [21]. Additionally, the magnitude of the control increments needs to be minimized. Based on the above conditions, the cost function, for time step k, is designed as follows:
J k = Y k Y d k T Q Y k Y d k + Δ U k T R Δ U k
where Y k = y k + 1 k y k + p k T , Y d k = y d k + 1 k y d k + p k T , Q = q I p , R = r I c , Y k = C ¯ X k = I p C X k . “⊗” denotes the Kronecker product of matrices.
With q > 0 , r > 0 , which represent the weighting factor of output error and control increment, respectively. Therefore, the cost function J k in Equation (8) can be expressed as
J k = C ¯ X k Y d k T Q C ¯ X k Y d k + Δ U k T R Δ U k = J c k + Δ U k T H k Δ U k + f k T Δ U k
where H k = G k 1 T C ¯ T Q C ¯ G k 1 + R , f k T = 2 C ¯ F k 1 + g k 1 Y d k T Q C ¯ G k 1 , J c k   =   C ¯ F k 1 + g k 1 Y d k T Q C ¯ F k 1 + g k 1 Y d k . Note that the later optimization problem is to find Δ U k to minimize J k , while the change of Δ U k does not affect J c k , which can be disregarded.
Like most MPC problems, some constraints on the amount of control and control increments are required. They are constrained as
U min U k U max
U k = U k 1 + I 1 t Δ U k
Δ U min Δ U k Δ U max
where U k 1 = u k 1 k 1 1 c , I 1 t = 1 0 0 0 1 1 0 1 1 1 c × c .
Additionally, V r is required to remain negative for a successful interception, for the reason that V r represents the relative velocity components along the LOS. Only when V r < 0 does the distance between the UAV and the target gradually decrease so that interception can be achieved [17]. The terminal state in the prediction is constrained to be negative, for instance, less than a negative number v r d . We choose v r d = s i g n V r 0   ×   0.1 V r 0 here, where V r 0 represents the initial value of V r . The terminal constraint is given by
EX k V r d
where E = 0 1 × 5 0 1 × 5 0 1 × 5 e p × 5 p , V r d = 0 0 v r d p × 1 , e = 0 1 0 0 0 . Substituting X k from Equation (7) in Equation (13), we obtain
EG k 1 Δ U k V r d E F k 1 + g k 1
Considering the above constraints, a compact form can be expressed as
A k Δ U k b k
where A k = I c I c I 1 t I 1 t EG k 1 , b k = Δ U max Δ U min U max U k 1 U min + U k 1 V r d E F k 1 + g k 1 . With the above formulations of the cost function and constraints, the NMPC in guidance can be described as a QP problem:
min Δ U k Δ U k T H k Δ U k + f k T Δ U k s . t . A k Δ U k b k
Assuming that the solution to the QP in Equation (16) at time step k is given by Δ U k , then U k can be calculated by Equation (11). The first element of U k is u k k , which is used as the control input at time step k. Using U k and the measurement of state at the current time step x k k , we can compute X s k = x T k k x T k + 1 k x T k + p 1 k T by Equation (5). X s k is then used at the next time step to calculate the matrices F k 1 , g k 1 , G k 1 . Further, U k is reformulated as U k = u k k 1 c . After the above algorithm, a closed loop control is accomplished.
Remark 1.
At the begining step of the simulation, u k 1 k 1 = 0 , U k 1 = 0 c × 1 and X s k 1 = 1 p x 0 , where x 0 is the initial state.

2.3. Problem Statement

In order to guide the UAV under input constraints, we designed the basic guidance law by transforming the guidance conditions into a cost function and the constraints into inequality constraints using NMPC. Furthermore, it is our intention to attenuate the oscillatory variation of the control inputs throughout the guidance in order to reduce the impact of the target maneuvering, so that the average energy expenditure is as small as possible. In order to quantify this, we set two parameters as follows:
S u = 0 t f u d t
s u = S u / t f
where t f represents the impact time and also the time the simulation stops.
S u in Equation (17) represents accumulated control losses, s u in Equation (18) represents average energy expenditure during guidance.

3. Insensitive Mechanism-Based Nonlinear Model Predictive Guidance

In this section, we first introduce the Insensitive Mechanism (IM) in NMPC, then design a fixed-time sliding mode disturbance observer, and finally elaborate a combination algorithm of NMPC and IM used for guidance. In order to make the algorithm more complete, the parameters are adjusted and a determination condition is added as to whether the maximum input is required or not. The general structure of what is designed in this section is shown in Figure 3.
M and T provide the initial position information, and the information related to LOS is calculated based on Equation (1). The maneuvering states of target are obtained after the disturbance observer. The control inputs to the UAV are calculated by the NMPC-IM algorithm. Meanwhile, Equation (41) is utilized for Maximum Input Determination, and if the conditions are satisfied, the input calculated by NMPC-IM will be substituted.

3.1. Design of Insensitive Mechanism

When using NMPC, as well as other methods of guidance on maneuvering targets, if T is bending forward, M may behave sensitively to the movement of T and thus make unnecessary steering control. Especially when M is far away from T, the steering control may produce wasteful lateral overload. Due to this, the guidance method may need to be less sensitive.
One simple idea is that if we can measure or estimate the maneuvering state of T, we might be able to make a prediction of its position some time in the future, choosing to strike directly toward the predicted position without or with little regard for the current position of T [22,23].
A schema of the basic idea of the IM is shown in Figure 4. M k and T k represent the position of M and T at time step k, respectively. After a p-step prediction, the position of T is T k + p . The blue dotted line represents the predicted trajectory of T, the red solid line is M’s guidance to the current position of T, and the green solid line is M’s guidance to the predicted position of T. In the IM, we would rather choose the green one than choose the red one.
We first assuming that the maneuvering state of T is known to be a T , the position and the heading angle of the target can be measured. In order to represent the insensitivity of M to the maneuverability of T, it is assumed that M remains stationary while the position of T is predicted. According to Equation (1), the kinematic relationship between M and T is simplified as
r ˙ = V T cos α T θ r θ ˙ = V T sin α T θ α ˙ T = a T / V T
In fact, Equation (19) is not a description of the objective world motion, but a hypothesis required at the moment of calculating the control inputs to the UAV. In this instant, the UAV assumes that it remains stationary and predicts the position of the target in the future, which allows the UAV not to be limited to the target’s current true position, but to reduce unnecessary inputs by anticipating the future. Equation (1), on the other hand, is an objective description of the relative motion of objects that exists in the objective world and follows objective laws. Therefore, it can be considered that Equation (19) is a subjective hypothesis required for prediction and is where the basic idea of IM lies, and it does not conflict with Equation (1).
At the time step k, the predicted position of T after p steps is calculated by Equation (19). The step-by-step calculation of the Insensitive Mechanism is shown in Figure 5.
The red arrows represent the positional shift of the target at each step during the prediction time period, while the missile remains stationary. For j = 1 , , p , r s k + j k , and θ s k + j k represent the range and LOS angle from M k to the T k + j , respectively. Similarly, the heading angle of T can be expressed as α T s k + j k , which is given by linear prediction with known a T . Using Equation (19) and the measurement of M and T’s state at time step k, we can compute all the above parameters in the prediction time domain step by step. Once the position of T k + p is obtained, all the M needs to do is to keep the heading angle θ s k + p k , which is abbreviated as θ s d .
However, in actual motion, M does not remain stationary. Here, we design an ideal trajectory based on (1) and θ s d . Before this, a new output is expressed as y s = C s x , where C s = 1 0 0 0 0 ; 0 0 k θ 0 0 . In this case, y s = r k θ θ T . Similarly, y s d denotes the desired output, where y s d k + j k = r s d k + j k k θ θ s d k + j k T . Generally speaking, the range of the variables can be used as a reference for the value of k θ . For instance, r ranges from 0 to 200, and θ ranges from 0 to 6.28. In order to balance the range of variation of the two variables, k θ is supposed to be 32. However, it is unlikely that the two variables will change in the same way, i.e., error values between expected and actual may not even be of the same order of magnitude in each step. Since we design the algorithm for θ s d in IM, we expect its corresponding error value to be smaller. In order for this error term in the QP problem to have a larger impact on the results, a larger k θ is required. After testing the number in different orders of magnitude, we choose k θ = 5000 . The calculations of r s d k + j k and θ s d k + j k are displayed as follows
f o r j = 1 r s d k + 1 k = r k + Δ t V T cos α T k θ k Δ t V M cos θ s d θ k θ s d k + 1 k = θ k + Δ t / r k V T sin α T k θ k Δ t / r k V M sin θ s d θ k f o r j = 2 , , p r s d k + j k = r s d k + j 1 k + Δ t V T cos α T s k + j 1 k θ s d k + j 1 k Δ t V M cos θ s d θ s d k + j 1 k θ s d k + j k = θ s d k + j 1 k + Δ t / r s d k + j 1 k V T sin α T s k + j 1 k θ s d k + j 1 k Δ t / r s d k + j 1 k V M sin θ s d θ s d k + j 1 k
Similar to Equation (9), the cost function, designed for IM, at time step k, is as follows:
J s k = Y s k Y s d T S Y s k Y s d = J s c k + Δ U k T H s k Δ U k + f s k T Δ U k
where Y s k = y s T k + 1 k y s T k + p k T , Y s d = y s d T k + 1 k y s d T k + p k T , S = s I 2 p , Y s k = C s ¯ X k = I p C s X k .
H s k = G k 1 T C s ¯ T S C s ¯ G k 1 , f s k T = 2 C s ¯ F k 1 + g k 1 Y s d T S C s ¯ G k 1 ,
J s c k = C s ¯ F k 1 + g k 1 Y s d T S C s ¯ F k 1 + g k 1 Y s d
with s > 0 , which represents the weighting factor of the output error in IM.
It is noteworthy that Equations (9) and (21) have the same structure, which allows the addition of IM to NMPC to produce certain effects
Remark 2.
Although the target maneuvering state assumed here is known, it can still be set to a condition that cannot be measured, which could be estimated by using the disturbance observer later. Further, a time series prediction of a T can be used, for example, a neural network. The more accurate the prediction of the target’s movement, the more effective IM is.
Moreover, the concept of IM is not limited to this paper. Scenarios such as a robotic arm gripping a flexible moving object under constraints, or autonomous driving with more soothing control inputs for the comfort of the passengers, are examples of scenarios where IM has a promising application.

3.2. NMPC-IM Guidance Law with Fixed-Time Sliding Mode Observer

In the previous sections, we derived the prediction equation, cost function, and control constraints in NMPC and then transformed the optimization problem into a QP problem. In addition, the conditions of IM are transformed into another cost function to be added into NMPC. We combine the two into a QP problem and add some additional conditions to the composition of the algorithm. To address the case where the maneuvering state of the target is unknown, we introduce a fixed-time sliding mode observer to estimate the state of the target, and eventually give a complete proof of feasibility.

3.2.1. Full Description of the QP Problem

In many cases, when the MPC problem is transformed into a QP problem, it is usually necessary to use a slack variable to solve the unsolvable problem due to hard constraints. To soften the terminal constraint for V r in Equation (14), we relax the inequality with
EG k 1 Δ U k + 0 1 × p 1 γ k T V r d E F k 1 + g k 1
where γ k 0 is the slack variable. With the addition of γ k , and the cost function in IM, the QP in Equation (16) is modified as
min ζ k ζ k T H ˜ k ζ k + f ˜ k T ζ k s . t . A ˜ k ζ k b ˜ k
where ζ k = Δ U k T γ k T R c + 1 , H ˜ k = diag H k + H s k , r 1 , f ˜ k = f k + f s k r 2 , A ˜ k = A k 0 ˜ 0 1 × c 1 4 c + p + 1 × c + 1 , b ˜ k = b k 0 4 c + p + 1 × 1 .
with r 1 , r 1 > 0 , 0 ˜ = 0 c × 1 0 c × 1 0 c × 1 0 c × 1 0 ¯ 4 c + p × 1 , 0 ¯ = 0 p 1 × 1 1 .
The combination of NMPC and IM belongs to a linear superposition; however, with a fixed linear relationship, it is certainly more difficult to fulfill the demand. For example, when the maneuverability of T is relatively large, we would like to consider IM more.
We choose a ˙ T to represent the maneuverability of T; the corresponding weighting factors in the cost function of NMPC and IM are adjusted, respectively.
q = 0.1 × ( 1 2 π arc tan a ˙ T ) r = 1 s = 10 × 5 + 10 a ˙ T
According to Equation (24), it can be seen that q and s vary monotonically decreasing and monotonically increasing with respect to a ˙ T , respectively. Additionally, the effect of target maneuvering on the result of prediction is considered. In the IM, the predicted target is considered to have constant a ˙ T , which means the result of prediction could be better. In contrast, when a ˙ T changes greatly, the result could be worse, and we hope that p could be smaller in order to avoid a larger error. Therefore, p should be inversely proportional to a ˙ T . In addition, the NMPC-based guidance law may focus on the predicted position of T, resulting in an excessive amount of eventual off-targeting. Thus, once r < r 0 , a threshold, we expect p to decrease as r decreases, which can expressed as follows:
p = 1 2 π arc tan a ˙ T p max p min + p min , r > r 0 r r 0 1 2 π arc tan a ˙ T p max p min + p min , r r 0
where “ [ ] ” represents the Gauss mark; p m a x > p m i n > 0 are positive integers.
Remark 3.
Because p is changeable at each step, X s k 1 R 5 p × 1 changes too. While calculating F k 1 , g k 1 , G k 1 using X s k 1 , we need to change the scale of X s k 1 first. Actually, X s k 1 R 5 p k 1 × 1 , where p k 1 represents p the previous time step. After computing p k at this time step by Equation (25), change X s k 1 as
X s k 1 = x k 1 k 1 x k + p k 1 2 k 1 x k + p k 1 2 k 1 5 p k × 1 , p k 1 < p k x k 1 k 1 x k k 1 x k + p k 2 k 1 5 p k × 1 , p k 1 p k
After adjusting the weighting factors and the prediction time domain, the IM in NMPC can work better.
This is a complete description of the QP problem in NMPC-IM.

3.2.2. Fixed-Time Sliding Mode Observer for Acceleration Estimation of Maneuvering Target

Sliding Mode Observer (SMO) is a type of nonlinear observer commonly used for estimating the state of a system. It operates by constructing an observation error system and designing a switching control law achieving an accurate estimation of the system’s state. There are also successful examples of combining observer and MPC [24,25].
To ensure that the state of the system converges to the origin in a finite time T and the convergence time is minimally affected by the initial state, ref. [26] proposed a control law by improving the super-twisting algorithm.
Assuming that the maneuvering states of T are unknown and bounded. We let a T r = a T sin ( θ α T ) and a T θ = a T cos ( θ α T ) represent the maneuvering states of T, and a T r ,   a T θ L , a ˙ T r , a ˙ T θ H . According to Equation (2), there exist two first-order systems: V ˙ r = u ¯ 1 + a T r , where u ¯ 1 = V θ 2 / r + sin α M θ u and V ˙ θ = u ¯ 2 + a T θ , where u ¯ 2 = V r V θ / r cos α M θ u .
With the addition of the designed fixed-time sliding mode disturbance observer, a theorem about the feasibility of the QP in Equation (23) is given as follows:
Theorem 1.
Observer (27) can observe the actual value of the maneuvering states a T r and a T θ , respectively. The errors converge to zero in a fixed time T f , which is independent of the initial states. The QP in Equation (23) is a Strictly Convex QP (SCQP) for all k, and the errors will not effect the feasibility of the solution.
x ˙ = u + d σ 1 = z 1 x σ 2 = z 2 d z ˙ 1 = k 1 σ 1 p s i g n σ 1 k 2 σ 1 q s i g n σ 1 + z 2 + u z ˙ 2 = k 3 s i g n σ 1
T f M 1 k 1 p 1 ε p 1 + ε 1 q k 2 1 q 1 m 1 M h k 2 / k 2
where k 1 , k 2 > 0 , k 3 > H , p > 1 , 0 < q < 1 , ε > 0 , h k 2 = 1 / k 2 + 2 e / m k 2 1 / 3 , M = k 3 + H , m = k 3 H .
Proof of Theorem 1.
We use the estimation of a T r as an example. For the purpose of the proof, the initial time of the system is considered to be 0 and the value of the sliding mode surface at the initial moment is σ 1 0 , and σ 2 0 = 0 [27]. The time derivatives of σ 1 and σ 2 are
σ ˙ 1 = z ˙ 1 V ˙ r = k 1 σ 1 p s i g n σ 1 k 2 σ 1 q s i g n σ 1 + z 2 a T r σ ˙ 2 = z ˙ 2 a ˙ T r = k 3 s i g n σ 1 a ˙ T r
Given ε > 0 , when σ 1 0 > ε , according to Equation (29), we have
d σ 1 t d t k 1 σ 1 t p
It is easy to analyze that when t > 0 , the signs of σ 1 and σ 2 are opposite, i.e., the signs of σ 1 remain unchanged, so Equation (30) can be expressed as
σ 1 t p d σ 1 t k 1 d t
Integrating both sides of the inequality yields:
σ 1 t 1 p 1 p σ 1 0 1 p 1 p k 1 t 0
σ 1 t 1 p 1 p k 1 t 0 + σ 1 0 1 p 1 p k 1 t
Multiply both sides simultaneously by ( p 1 ) :
σ 1 t 1 p k 1 p 1 t
σ 1 t p 1 1 k 1 p 1 t
After analysis, it can be obtained that σ 1 t gradually decreases to ε in 0 , T 1 , where T 1 1 / k 1 p 1 ε p 1 . The signs of σ 2 and σ 1 are opposite and k 3 > H , σ 2 t increases in 0 , T 1 .
In the time period t > T 1 , σ 1 t decreases from ε to 0. According to Equation (29), we have
d σ 1 t d t k 2 σ 1 t q
Solving this inequality gives
σ 1 t 1 q k 2 1 q t T 1 + ε 1 q
In this case, σ 1 t gradually decreases to 0 in T 1 , T 2 , where T 2 ε 1 q / k 2 1 q . σ 2 t increases in T 1 , T 2 , and σ 2 T 2 < M T 2 = M 1 k 1 p 1 ε p 1 + ε 1 q k 2 1 q .
In the time period t > T 2 , the trajectory of the system state starts from 0 , σ 2 T 2 and gradually converges to the origin. Since the parameters of the observer satisfy the inequality M h k 2 / k 2 < 1 , it can be obtained that the convergence time satisfies the inequality T f σ 2 T 2 / m 1 M h k 2 / k 2 according to the proof in ref. [26].
In summary, the sliding mode surfaces of the observer σ 1 t and σ 2 t can converge at a fixed time T f , where T f satisfies
T f M 1 k 1 p 1 ε p 1 + ε 1 q k 2 1 q 1 m 1 M h k 2 / k 2
We use a ^ T r and a ^ T θ to represent the estimated maneuvering states, and they will converge to the real value in T f from the above proof.
The feasible set for the optimization problem in Equation (23) is a polytope for all k. Hence, the feasible set is convex for all k [28]. The Hessian matrix H k + H s k can be rewritten as H k + H s k = G k 1 T C ¯ T Q C ¯ G k 1 + R + G k 1 T C s ¯ T S C s ¯ G k 1 . The maneuvering states have been abandoned in Equation (4), so G k 1 will not contain a ^ T r or a ^ T θ according to Equation (7). Obviously, C ¯ and C s ¯ are also not related to a ^ T r , a ^ T θ . As for the parameters q, r, s in Equation (24), a T and a ˙ T are given by
a T 2 = a ^ T r 2 + a ^ T θ 2
a ˙ T = a ^ T r a ^ ˙ T r + a ^ T θ a ^ ˙ T θ a ^ T r 2 + a ^ T θ 2
Clearly, q, r, s remain larger than zero, the corresponding Q , R , S are positive definite. Therefore, G k 1 T C ¯ T Q C ¯ G k 1 and G k 1 T C s ¯ T S C s ¯ G k 1 are positive semidefinite (at least). Thus, H k + H s k is positive definite and the cost function J k + J s k is strictly convex for all k. In this way, due to the positive definiteness of H k + H s k and optional r 1 > 0 , the Hessian H ˜ k is positive definite for all k. Thus, the cost function ζ k T H ˜ k ζ k + f ˜ k T ζ k is strictly convex for all k. With that, the corresponding QP in Equation (23) is an SCQP for all k.
Due to the strict convexity property, the optimal solution for the QP in Equation (23) can be efficiently obtained using existing convex optimization tools and solvers [17], which means the QP is feasible even with the estimated states of T. □
It is worth stating that although the error in the estimation does not affect the feasibility of the QP problem, it affects the prediction of the target’s direction and determines whether the target can be intercepted quickly and accurately. Since the estimated states will converge in a fixed time, which can usually be limited to a few seconds, errors have little effect on the overall guidance process. Unlike general control processes, the guidance process does not need to deliberately follow a particular trajectory, but only needs to ensure that the relative distance to the target decreases gradually. In Equation (13), V r is limited to less than 0, which means that r is decreasing, satisfying the basic requirement that the UAV approaches the target gradually.

3.2.3. Maximum Input Determination

In contemplating an exceedingly extreme scenario, α M , the heading angle of M has deviated significantly from the LOS angle θ to the extent that even with the application of maximum control input, it is not feasible to nullify the deviation angle for the remainder of the trajectory. Despite this limitation, it remains imperative to employ the maximum control input in an attempt to effect a timely correction to the trajectory. Considering constraints on control increments, the determination of maximum input is designed as
l e t α M θ π , π , t g o = k t r V M i f s i g n α M θ α M θ u max t g o V M s i g n α M θ > 0 u = u p r e s i g n α M θ Δ u max
where u p r e represents the last control input, and k t > 0 is an influencing factor that determines how early the effect will take place.

4. Simulation Results

In this section, we present simulation results for the NMPC-IM guidance algorithm. Firstly, we discuss the selection of the desired output y d in NMPC-IM. The NMPC-IM guidance algorithm without SMO and Maximum Input Determination can be called pure NMPC-IM, which is compared with the PNG to verify the effectiveness of the algorithm. We then analyzed the parameter selection for Maximum Input Determination. The NMPC-IM algorithm is compared with PNG and NMPC and conclusions are drawn. Finally, assuming that the maneuvering states of the target are unknown, SMO is added to estimate them and obtain a satisfactory guidance effect.
The experiments were conducted on a Dell computer equipped with an Intel(R) Core(TM) i7-8565U CPU, operating at a base frequency of 1.80 GHz. The system is furnished with 8 GB of RAM and runs on a 64-bit operating system. For the simulation and computational tasks, Python 3.7 was utilized as the primary programming language. Moreover, we use the Euler method for numerical solutions.

4.1. Experimental Settings and Datasets

We consider the maneuvering target with time-varying acceleration. The basic simulation conditions are shown in Table 1. Additionally, PNG is set as a comparison test, and the proportional coefficient is K.

4.2. Selection of the Desired Output y d

Because there is a determination of attack range, r will not converge to 0. With the protection of determination conditions, and after experiments, it was found that it is more appropriate to use y d = 1 / r rather than y d = r for the algorithm. To verify this statement, we set the initial range and LOS angle to T as r 0 = 200 m and θ 0 = 45 , respectively. T moves with acceleration as a T = 10 sin ( 2 t ) g, and the initial heading angle is 180 . The initial heading angle of M is set to four sets of experiments α M 0 = 0 , 90 , 180 , 270 . The results are shown in Figure 6.
According to the results, we can find that when choosing y d = 1 / r , the UAV intercepts the target faster than choosing y d = r . Especially in the cases of 90 and 270 , the average energy expenditure is much smaller when choosing y d = 1 / r . Thus, it seems that the previous choice of y d is relatively correct.

4.3. Compare pure NMPC-IM with PNG

In order to better validate the performance of pure NMPC-IM, we set up two different scenarios with the target acceleration set to a T = 20 sin ( t ) g and a T = 15 sin ( 1.5 t ) g, respectively, based on the conditions r 0 = 200 m, θ 0 = 45 , and the initial heading angle of T is 0 . The initial heading angles of M are set to four sets of experiments α M 0 = 0 , 90 , 180 , 270 , and α M 0 = 0 , 30 , 60 , 90 , respectively. The results are shown in Figure 7 and Table 2, and Figure 8 and Table 3, respectively.
From the results, it can be found that the impact time of PNG is shorter than that of NMPC-IM in most cases, but the average energy expenditure is larger. It is worth noting that in the 90 counterpart case in Figure 7a, the guidance trajectory of the NMPC-IM re-intercepts the target after making a large circle, resulting in a much longer impact time than normal. This is due to the fact that the UAV’s miss distance on its first approach to the target is too large to intercept the target, resulting in a missed intercept. As for the reason, it could be that the control inputs are constrained, and it is difficult to increase the inputs without limit to make adjustments when faced with such a situation, as in the case of the PNG. The Maximum Input Determination proposed before is intended to solve this problem.

4.4. Effectiveness Validation of Maximum Input Determination

The effectiveness of Maximum Input Determination is validated first. We set the initial range and LOS angle to T as r 0 = 200 m and θ 0 = 120 , respectively. The initial heading angle of M is α M 0 = 45 . T moves with acceleration as a T = 10 sin ( t ) g, and the initial heading angle is 0 . After approximate testing, a range of more pronounced results due to changes in k t was identified. Within that range, a series of representative points were selected so that the results were distinguishable from each other. k t is chosen sequentially as 0.2 , 0.5 , 0.9 , 1.5 , 2.5 , and the results are shown in Figure 9.
We use colored spots to mark the portion of Maximum Input Determination, and it is obvious that when k t is relatively small, the maximum input is used at the very beginning of the guidance in order to head to T, but the trajectory is correspondingly more curved. As k t gradually increases, the Maximum Input Determination effects occur later. In general, at the end stage of guidance, since both M and T are dynamic, it is easier to produce angular deviation and the Maximum Input Determination effects occur more easily.

4.5. Comparison of NMPC-IM with PNG and NMPC

To compare NMPC-IM with PNG and NMPC, we designed two sets of experiments with different maneuvering states of T for a more complete validation of the algorithm.
We first set T moving with acceleration as a T = 10 sin ( 1.5 t ) g, and the initial heading angle as 0 . r 0 = 150 m and θ 0 = 45 , respectively. The initial heading angle of M is α M 0 = 90 . k t is chosen as 0.9 , for a more balanced effect as shown in Figure 9, i.e., not determined too early or too late. Additionally, we use the computation time to represent the real time it takes for the computer to finish running. The results are shown in Figure 10 and Table 4.
Let q = 0 , which means completely disregarding the cost function in the original NMPC, to show the IM individually.
As shown in Figure 11, it can be seen that the IM is slow with respect to the target’s maneuverability, and will not respond very quickly. Comparing with Figure 10, we can find that IM works mainly by attenuating the implication effect of frequent oscillations of control inputs due to frequent maneuvers of the target (e.g., sometimes left, sometimes right), so that unnecessary energy loss can be reduced by certain prediction.
We then set T moving with acceleration as a T = 20 exp ( 0.1 t ) g, and the initial heading angle as 0 . r 0 = 180 m and θ 0 = 45 , respectively. The initial heading angle of M is α M 0 = 90 . k t is chosen as 0.9 , for a more balanced effect as shown in Figure 9, i.e., not determined too early or too late. The results are shown in Figure 12 and Table 5.
It can be seen from the above results that the addition of IM does improve the trajectory guided by NMPC. Although the trajectory of NMPC looks straighter than that of NMPC-IM in Figure 10a, the trajectory is still more curved from the details due to the fact that its control inputs have been oscillating between poles. More importantly, NMPC-IM reduces average energy expenditure during the UAV interception. In general, for s u , NMPC-IM < NMPC < PNG, because unnecessary bends as well as control waste can be reduced with prediction. In terms of computational performance as shown in Table 3 and Table 4, NMPC and NMPC-IM are considerably slower than PNG, but NMPC-IM has some improvement compared to NMPC. We believe that it may be the inclusion of IM that makes the QP problem easier to solve, as well as the Maximum Input Determination to correct the trajectory in time and optimize the subsequent computation. All in all, the addition of IM is an improvement to the NMPC guidance method in terms of trajectory optimization and expenditure of average energy.

4.6. Further Testing of The NMPC-IM Algorithm

In order to check the universality and robustness of the algorithm, we consider the guidance effect in the case of different initial heading angles of M and with the addition of disturbance observer, respectively.

4.6.1. Different Initial Heading Angles of M

The acceleration of T is set as a T = 20 exp ( 0.2 t ) g, with initial heading angle 0 . r 0 = 250 m and θ 0 = 45 , respectively. The initial heading angle of M is α M 0 = 90 . k t is chosen as 0.9 , and the results are shown in Figure 13.
With different initial heading angles, M always intercepts T, though t f differs.

4.6.2. Add Disturbance Observer

Previously, we assumed that the maneuvering states of T are known, including a T and α T . In this simulation, the maneuvering states of T are unknown; the fixed-time sliding mode observer in Equation (27) is used to estimate them. Take the same conditions as Figure 10 except Δ t = 0.02 s, and set k 1 = 5 , k 2 = 5 , k 3 = 15 , p = 1.5 , q = 0.5 ; the results are shown in Figure 14.
Based on the comparison of the estimated and true values, it can be observed that the disturbance observer has a relatively good estimate of the maneuvering state of T, converging to the vicinity of the true value in a very short time, despite the inevitable presence of buffeting. By obtaining the approximate maneuvering information of T, under the guidance of the NMPC-IM algorithm, M is indeed able to accomplish the task of pursuing T, and the average energy expenditure remains smaller than NMPC. It is worth noting that, due to the buffeting, the estimation of the state of T may sometimes be diametrically opposed to the true value, but this does not immediately have an irreversible impact on the control inputs, further illustrating the robustness of NMPC-IM.

5. Conclusions

In this paper, a guidance method based on NMPC is developed, on the basis of which IM is proposed. IM is formulated as an additional cost function for incorporation into the NMPC, which is finally transformed into a complete QP problem. Meanwhile, the parameters are adjusted and optimized to make IM play a better role in NMPC. We add a determination condition, which can be regulated, for the timely correction of the guidance of NMPC-IM. Moreover, a fixed-time sliding mode disturbance observer is designed to reduce the information required and verify the robustness of this algorithm. It has been shown by simulation experiments that the inclusion of IM reduces the average energy expenditure by minimizing the impact of the target’s maneuverability on guidance. This holds positive significance for the UAV’s trajectory optimization during the interception of a more maneuverable target under constraints. The parameters used in this paper require further qualitative analysis and optimization through alternative methods to achieve superior results. Additionally, predicting the target’s maneuverability remains a significant challenge, and the current predictions are quite limited. Accurately predicting the target’s state in the future would represent a significant optimization.

Author Contributions

Conceptualization, J.M., D.H. and T.Z.; methodology, J.M., D.H. and M.Z.; software, D.H. and M.Z.; validation, D.H. and T.Z.; formal analysis, D.H. and T.Z.; writing—original draft preparation, D.H.; writing—review and editing, J.M. and M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The dataset is available on request from the authors.

Acknowledgments

The authors are thankful to the anonymous reviewers for their useful suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ramezani, M.; Amiri Atashgah, M.A. Energy-Aware Hierarchical Reinforcement Learning Based on the Predictive Energy Consumption Algorithm for Search and Rescue Aerial Robots in Unknown Environments. Drones 2024, 8, 283. [Google Scholar] [CrossRef]
  2. Yan, T.; Liu, C.; Gao, M.; Jiang, Z.; Li, T. A Deep Reinforcement Learning-Based Intelligent Maneuvering Strategy for the High-Speed UAV Pursuit-Evasion Game. Drones 2024, 8, 309. [Google Scholar] [CrossRef]
  3. Shima, T.; Shinar, J. Time-Varying Linear Pursuit-Evasion Game Models with Bounded Controls. J. Guid. Control. Dyn. 2002, 25, 425–432. [Google Scholar] [CrossRef]
  4. Shao, J.; Xu, Y.; Luo, D. Cooperative combat decision-making research for multi UAVs. Inf. Control. 2018, 47, 347–354. [Google Scholar] [CrossRef]
  5. Liang, X.; Zhou, B.; Jiang, L.; Meng, G.; Xiu, Y. Collaborative pursuit-evasion game of multi-UAVs based on Apollonius circle in the environment with obstacle. Connect. Sci. 2023, 35, 2168253. [Google Scholar] [CrossRef]
  6. Xiong, H.; Zhang, Y. Reinforcement learning-based formation-surrounding control for multiple quadrotor UAVs pursuit-evasion games. ISA Trans. 2024, 145, 205–224. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, R.; Zong, Q.; Zhang, X.; Dou, L.; Tian, B. Game of Drones: Multi-UAV Pursuit-Evasion Game With Online Motion Planning by Deep Reinforcement Learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 7900–7909. [Google Scholar] [CrossRef] [PubMed]
  8. Fu, X.; Xu, Z.; Wang, H. Generalization strategy design of UAVs pursuit evasion game based on DDPG. J. Northwestern Polytech. Univ. 2022, 40, 47–55. [Google Scholar] [CrossRef]
  9. Qin, H.L.; Meng, Z.H. Autonomous Exploration and Mapping System Using Heterogeneous UAVs and UGVs in GPS-Denied Environments. IEEE Trans. Veh. Technol. 2019, 68, 1339–1350. [Google Scholar] [CrossRef]
  10. Li, J.Q.; Zhang, G.Q. Integrating Dynamic Event-Triggered and Sensor-Tolerant Control: Application to USV-UAVs Cooperative Formation System for Maritime Parallel Search. IEEE Trans. Intell. Transp. Syst. 2023, 25, 3986–3998. [Google Scholar] [CrossRef]
  11. Zhou, X.B.; Yu, X. UAV Collision Avoidance Based on Varying Cells Strategy. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 1743–1755. [Google Scholar] [CrossRef]
  12. Ghawghawe, S.N.; Ghose, D. Pure proportional navigation against time-varying target manoeuvres. IEEE Trans. Aerosp. Electron. Syst. 1996, 32, 1336–1347. [Google Scholar] [CrossRef]
  13. Li, Z.; Xia, Y.; Su, C.; Deng, J.; Fu, J.; He, W. Missile Guidance Law Based on Robust Model Predictive Control Using Neural-Network Optimization. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 1803–1809. [Google Scholar] [CrossRef] [PubMed]
  14. Kang, S.; Wang, J.; Li, G.; Shan, J. MPC-consensus based optimal cooperative guidance law design. In Proceedings of the 2016 IEEE Chinese Guidance, Navigation and Control Conference (CGNCC), Nanjing, China, 12 August 2016; pp. 336–341. Available online: https://kns.cnki.net/KCMS/detail/detail.aspx?filename=ZGHU201608003062&dbname=IPFDTEMP (accessed on 23 January 2017).
  15. He, X.; Chen, W.; Yang, L. Suboptimal Impact-Angle-Constrained Guidance Law Using Linear Pseudospectral Model Predictive Spread Control. IEEE Access 2020, 8, 102040–102050. [Google Scholar] [CrossRef]
  16. Zhao, S.; Chen, W.; Yang, L. Cooperative Optimal Guidance Law with Simultaneous Attack and Impact Angle Constraint Using Linear Pseudospectral Model Predictive Control. In Proceedings of the 2021 12th International Conference on Mechanical and Aerospace Engineering (ICMAE), Athens, Greece, 16–19 July 2021; pp. 511–517. [Google Scholar] [CrossRef]
  17. Bhattacharjee, D.; Chakravarthy, A.; Subbarao, K. Nonlinear Model Predictive Control and Collision-Cone-Based Missile Guidance Algorithm. J. Guid. Control. Dyn. 2021, 44, 1481–1497. [Google Scholar] [CrossRef]
  18. Çetin, A.T.; Koyuncu, E. Model Predictive Control-Based Guidance with Impact Angle Constraints for Visual Quadrotor Interception. In Proceedings of the 2023 9th International Conference on Control, Decision and Information Technologies (CoDIT), Rome, Italy, 3–6 July 2023; pp. 1–6. [Google Scholar] [CrossRef]
  19. Bhattacharjee, D.; Chakravarthy, A.; Subbarao, K. Nonlinear Model Predictive Control Based Missile Guidance for Target Interception. In Proceedings of the AIAA Scitech 2020 Forum, Orlando, FL, USA, 6–10 January 2020. AIAA Paper 2020–0865. [Google Scholar] [CrossRef]
  20. Kwon, W.H.; Han, S.H. Receding Horizon Control: Model Predictive Control for State Models; Springer: London, UK, 2006; pp. 8–11, 84–85. [Google Scholar]
  21. Lianos, D.; Shtessel, Y.; Shkolnikov, I. Integrated guidance-control system of a homing interceptor—Sliding mode approach. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Montreal, QC, Canada, 6–9 August 2001. AIAA 2001–4218. [Google Scholar] [CrossRef]
  22. Paul, N.; Ghose, D. Longitudinal-Acceleration-Based Guidance Law for Maneuvering Targets Inspired by Hawk’s Attack Strategy. J. Guid. Control. Dyn. 2023, 46, 1437–1447. [Google Scholar] [CrossRef]
  23. Moore, T.Y.; Biewener, A.A. Outrun or Outmaneuver: Predator–Prey Interactions as a Model System for Integrating Biomechanical Studies in a Broader Ecological and Evolutionary Context. Integr. Comp. Biol. 2015, 55, 1188–1197. [Google Scholar] [CrossRef] [PubMed]
  24. Chen, J.; Xu, Y.; Zheng, Z. Neural Network and Extended State Observer-Based Model Predictive Control for Smooth Braking at Preset Points in Autonomous Vehicles. Drones 2024, 8, 273. [Google Scholar] [CrossRef]
  25. Wang, D.; Gao, Y.; Wei, W. Sliding mode observer-based model predictive tracking control for Mecanum-wheeled mobile robot. ISA Trans. 2024, 151, 51–61. [Google Scholar] [CrossRef] [PubMed]
  26. Basin, M.; Bharath Panathula, C.; Shtessel, Y. Multivariable continuous fixed-time second-order sliding mode control: Design and convergence time estimation. IET Control. Theory Appl. 2017, 11, 1104–1111. [Google Scholar] [CrossRef]
  27. Zhang, M.J. Research of Multi-interceptor Cooperative Guidance Law with Line-of-sight Angle Constraint. Master’s Thesis, National University of Defense Technology, Changsha, China, October 2019. [Google Scholar] [CrossRef]
  28. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004; pp. 31–32. [Google Scholar]
Figure 1. The UAV pursuit-evasion scenario. The blue and orange conical areas are the variable range of orientation angles of the UAV and the target, respectively, and represent their maneuvering capabilities, and also the constraints of control input. Larger top angles of the cones mean more maneuverability.
Figure 1. The UAV pursuit-evasion scenario. The blue and orange conical areas are the variable range of orientation angles of the UAV and the target, respectively, and represent their maneuvering capabilities, and also the constraints of control input. Larger top angles of the cones mean more maneuverability.
Drones 08 00608 g001
Figure 2. Kinematics between UAV (M) and target (T).
Figure 2. Kinematics between UAV (M) and target (T).
Drones 08 00608 g002
Figure 3. The structure of NMPC-IM in guidance.
Figure 3. The structure of NMPC-IM in guidance.
Drones 08 00608 g003
Figure 4. The basic idea of IM.
Figure 4. The basic idea of IM.
Drones 08 00608 g004
Figure 5. Step-by-step schematic of Insensitive Mechanism.
Figure 5. Step-by-step schematic of Insensitive Mechanism.
Drones 08 00608 g005
Figure 6. Different y d in NMPC-IM: (a) Trajectories. (b) Control input. Green represents y d = r and blue represents y d = 1 / r .
Figure 6. Different y d in NMPC-IM: (a) Trajectories. (b) Control input. Green represents y d = r and blue represents y d = 1 / r .
Drones 08 00608 g006
Figure 7. Comparisons of PNG and pure NMPC-IM ( a T = 20 sin ( t ) g): (a) Trajectories. (b) Control input. Green represents PNG and purple represents pure NMPC-IM.
Figure 7. Comparisons of PNG and pure NMPC-IM ( a T = 20 sin ( t ) g): (a) Trajectories. (b) Control input. Green represents PNG and purple represents pure NMPC-IM.
Drones 08 00608 g007
Figure 8. Comparisons of PNG and pure NMPC-IM ( a T = 15 sin ( 1.5 t ) g): (a) Trajectories. (b) Control input. Green represents PNG and purple represents pure NMPC-IM.
Figure 8. Comparisons of PNG and pure NMPC-IM ( a T = 15 sin ( 1.5 t ) g): (a) Trajectories. (b) Control input. Green represents PNG and purple represents pure NMPC-IM.
Drones 08 00608 g008
Figure 9. Different k t in Maximum Input Determination: (a) Trajectories. (b) Control input.
Figure 9. Different k t in Maximum Input Determination: (a) Trajectories. (b) Control input.
Drones 08 00608 g009
Figure 10. Comparisons of PNG, NMPC, NMPC-IM (sin): (a) Trajectories. (b) Control input. (c) Prediction horizon of NMPC-IM.
Figure 10. Comparisons of PNG, NMPC, NMPC-IM (sin): (a) Trajectories. (b) Control input. (c) Prediction horizon of NMPC-IM.
Drones 08 00608 g010
Figure 11. Individually considerate of IM: (a) Trajectories. (b) Control input. (c) Prediction horizon.
Figure 11. Individually considerate of IM: (a) Trajectories. (b) Control input. (c) Prediction horizon.
Drones 08 00608 g011
Figure 12. Comparisons of PNG, NMPC, NMPC-IM (exp): (a) Trajectories. (b) Control input. (c) Prediction horizon of NMPC-IM.
Figure 12. Comparisons of PNG, NMPC, NMPC-IM (exp): (a) Trajectories. (b) Control input. (c) Prediction horizon of NMPC-IM.
Drones 08 00608 g012
Figure 13. Different initial heading angles of M: (a) Trajectories. (b) Control input.
Figure 13. Different initial heading angles of M: (a) Trajectories. (b) Control input.
Drones 08 00608 g013
Figure 14. Addition of disturbance observer: (a) Trajectories. (b) Control input. ( t f , s u ) = (36.64 s, 1.4346 g). (c) Prediction horizon of NMPC-IM. (d) Estimated and true values of a T r . (e) Estimated and true values of a T θ .
Figure 14. Addition of disturbance observer: (a) Trajectories. (b) Control input. ( t f , s u ) = (36.64 s, 1.4346 g). (c) Prediction horizon of NMPC-IM. (d) Estimated and true values of a T r . (e) Estimated and true values of a T θ .
Drones 08 00608 g014
Table 1. Simulation conditions.
Table 1. Simulation conditions.
VariablesValueVariablesValue
  Δ t  10.1 s r b 210 m
V M 14 m/s V T 10 m/s
  u max 2 g u min −2 g
  Δ u max 1 g Δ u min −1 g
  p ( N M P C ) 50 c ( N M P C ) 25
  q ( N M P C ) 0.1 r ( N M P C ) 1
  r 1  3100 r 2 200
  p ( N M P C I M ) Equation (25) c ( N M P C I M ) [p/2]
  p m a x 50 p m i n 3
q , r , s ( N M P C I M ) Equation (24) K ( P N G ) 3
a M 0  40 g r 0  550 m
  a ˙ T  6 a T k a T k 1 / Δ t a ˙ T 0 0
1 Sample time. 2 The simulation will stop once r < r b . 3 Slack variable. 4 The initial control input, which can also be the acceleration of M. 5 A threshold used in Equation (25). 6 Euler discretization.
Table 2. Impact time and average energy expenditure in different guidance laws ( a T = 20 sin ( t ) g).
Table 2. Impact time and average energy expenditure in different guidance laws ( a T = 20 sin ( t ) g).
Guidance Laws 0 90 180 270
P N G t f = 16.0  s t f = 15.4  s t f = 30.6  s t f = 32.8  s
  s u  1 = 3.1628 g  s u = 2.7524  g  s u = 2.5487  g  s u = 3.3881  g
N M P C I M t f = 16.8  s t f = 68.0  s t f = 32.4  s t f = 41.3  s
s u = 1.6464  g s u = 1.4575  g s u = 1.5242  g s u = 1.6239  g
1 Refer to Equation (18).
Table 3. Impact time and average energy expenditure in different guidance laws ( a T = 15 sin ( 1.5 t ) g).
Table 3. Impact time and average energy expenditure in different guidance laws ( a T = 15 sin ( 1.5 t ) g).
Guidance Laws 0 30 60 90
P N G t f = 33.2  s t f = 31.1  s t f = 30.8  s t f = 31.5  s
  s u = 3.2696  g s u = 3.3183  g s u = 3.2152  g s u = 2.7818  g
  N M P C I M t f = 36.2  s t f = 33.3  s t f = 30.0  s t f = 33.5  s
  s u = 1.2263  g s u = 0.9766  g s u = 0.4707  g s u = 1.2486  g
Table 4. Impact time and average energy expenditure in different guidance laws (sin).
Table 4. Impact time and average energy expenditure in different guidance laws (sin).
Guidance Laws t f  (s) su  (g)Computation Time 1 (s)
P N G 30.32.68290.2784
N M P C 30.71.54503.3304
NMPC-IM31.81.28761.5531
1 After five repetitions of the experiment and taking the average.
Table 5. Impact time and average energy expenditure in different guidance laws (exp).
Table 5. Impact time and average energy expenditure in different guidance laws (exp).
Guidance Laws t f  (s) su  (g)Computation Time (s)
P N G 15.03.03510.2098
N M P C 14.71.52931.1526
NMPC-IM14.61.43040.9277
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, D.; Zhang, M.; Zhan, T.; Ma, J. Insensitive Mechanism-Based Nonlinear Model Predictive Guidance for UAVs Intercepting Maneuvering Targets with Input Constraints. Drones 2024, 8, 608. https://doi.org/10.3390/drones8110608

AMA Style

Huang D, Zhang M, Zhan T, Ma J. Insensitive Mechanism-Based Nonlinear Model Predictive Guidance for UAVs Intercepting Maneuvering Targets with Input Constraints. Drones. 2024; 8(11):608. https://doi.org/10.3390/drones8110608

Chicago/Turabian Style

Huang, Danpeng, Mingjie Zhang, Taideng Zhan, and Jianjun Ma. 2024. "Insensitive Mechanism-Based Nonlinear Model Predictive Guidance for UAVs Intercepting Maneuvering Targets with Input Constraints" Drones 8, no. 11: 608. https://doi.org/10.3390/drones8110608

APA Style

Huang, D., Zhang, M., Zhan, T., & Ma, J. (2024). Insensitive Mechanism-Based Nonlinear Model Predictive Guidance for UAVs Intercepting Maneuvering Targets with Input Constraints. Drones, 8(11), 608. https://doi.org/10.3390/drones8110608

Article Metrics

Back to TopTop