Next Article in Journal
Mann-Type Inertial Subgradient Extragradient Rules for Variational Inequalities and Common Fixed Points of Nonexpansive and Quasi-Nonexpansive Mappings
Next Article in Special Issue
Analyzing Uncertain Dynamical Systems after State-Space Transformations into Cooperative Form: Verification of Control and Fault Diagnosis
Previous Article in Journal
Fractional Coupled Hybrid Sturm–Liouville Differential Equation with Multi-Point Boundary Coupled Hybrid Condition
Previous Article in Special Issue
Nonlinear Control System Design of an Underactuated Robot Based on Operator Theory and Isomorphism Scheme
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Analytic and Numerical Investigation of a Differential Game

1
Department of Mathematics, ORT Braude College, Karmiel 2161002, Israel
2
The Center for Mathematics and Scientific Computation, U. Haifa, Mt. Carmel, Haifa 3498838, Israel
3
Department of Mathematics, The Technion—Israel Institute of Technology, Haifa 3200003, Israel
*
Author to whom correspondence should be addressed.
Axioms 2021, 10(2), 66; https://doi.org/10.3390/axioms10020066
Submission received: 2 March 2021 / Revised: 12 April 2021 / Accepted: 15 April 2021 / Published: 17 April 2021
(This article belongs to the Special Issue Advances in Analysis and Control of Systems with Uncertainties)

Abstract

:
In this paper we present an appropriate singular, zero-sum, linear-quadratic differential game. One of the main features of this game is that the weight matrix of the minimizer’s control cost in the cost functional is singular. Due to this singularity, the game cannot be solved either by applying the Isaacs MinMax principle, or the Bellman–Isaacs equation approach. As an application, we introduced an interception differential game with an appropriate regularized cost functional and developed an appropriate dual representation. By developing the variational derivatives of this regularized cost functional, we apply Popov’s approximation method and show how the numerical results coincide with the dual representation.

1. Introduction

In this paper, we present a zero-sum differential game with linear dynamics and a quadratic cost functional. Such games appear in many areas of control theory, for example, robust controllability [1], pursuit-evasion [2,3,4,5], robust tracking [6] and robust investment [7], to name but a few.
The singularity of such a game is caused by the weight matrix of the minimizer’s control cost, meaning that problem of minimization its variational Hamiltonian with respect to the minimizer control has either infinitely many solutions or no solutions. This makes the game challenging, since it can not be solved by some well-known approaches, such as the Isaacs min-max principle [8] and the Bellman–Isaacs equation [8,9,10,11].
Known techniques in the literature involve high order optimality conditions, but their practical usage is quite limited and not general enough [2,5]. Regularization approaches with additional assumption have been studied by several researchers, such as [12]. Other related results include [13,14,15].
In [16] the differential game is tackled by introducing a cost functional containing the minimizer’s control cost. A regularization approach is then considered, yielding an auxiliary differential game with partial cheap control of the minimizer. While differential games with total cheap control of at least one of the players has been studied enough already—see, e.g., [1,6,17,18,19,20]—partial cheap control of at least one of the players has been studied by only a few [16].
In the recent work of the authors [21], a saddle-point reformulation of the zero-sum singular differential game was studied, and two gradient methods were presented and analyzed. This work considering a slightly more general game than [16] and a pursuit–evasion game illustrates the applicability of the numerical methods.
Following the above and as a continuing work of [21], the objectives of this work are as follows:
  • Introduce an appropriate cost functional that includes in addition to relative lateral velocity, lateral relative separation.
  • For the above functional, develop a dual representation and present its variational derivatives.
  • Present numerical calculations of the dual representation that gives the interception.
  • Validate the above via Popov’s approximation method.
The paper is organized as follows. We first recall some basic definitions and results in Section 2. In Section 3, an interception game is considered. In Section 4, the dual representation of the game’s cost functional is derived, which follows by numerical validation of the double projection methods for finding saddle points in Section 5. Final conclusions are given in Section 6.

2. Preliminaries

A standard linear zero-sum differential game consists of constraint
d z ( t ) d t = A z ( t ) + B u ( t ) + C v ( t ) , z ( 0 ) = z 0 , t [ 0 , t f ] ,
and a cost functional
J ( u , v ) z T ( t f ) F z ( t f ) + 0 t f z T ( t ) D z ( t ) + u T ( t ) G u u ( t ) v T ( t ) G v v ( t ) d t ,
to be minimized by u and maximized by v. The involved parameters in (1) and (2) are: t f , a given final time moment; T, the transpose; E n , an n dimensional Euclidean space; z ( t ) E n , a state vector; u ( t ) E r ; ( r n ); v ( t ) E s , the players’ controls and quadratic integrable; A, B and C are the given matrices, where B is fully ranked. Moreover, z 0 E n is a given initial vector; F, D and G u are given positive semi-definite symmetric matrices; and G v is a positive, definite symmetric matrix.
Next we recall several definitions.
Definition 1.
The differential control game (1) and (2) is called singular if all or some of the coordinates of the minimizer’s control are singular, that is, G u = 0 or
G u = diag g u 1 , , g u q , 0 , , 0 r q , g u j > 0 , j = 1 , , q ,
Definition 2.
The control game (1) and (2) are called regular if the cost functional (2) for a small enough ε > 0 has one of the following structures:
J ε ( u , v ) = z ( t f ) T F z ( t f ) + 0 t f [ z T ( t ) D z ( t ) + ε 2 u T ( t ) u ( t ) v T ( t ) G v v ( t ) ] d t .
Or
J ε ( u , v ) = z T ( t f ) F z ( t f ) + 0 t f z T ( t ) D z ( t ) + u T ( t ) G u + E u ( t ) v T ( t ) G v v ( t ) d t ,
where
G u + E = diag g u 1 , , g u q , ε 2 , , ε 2 r q .
Such regular cheap/partial cheap control games were analyzed in [12,16,22,23] using the study of the Riccati matrix differential equation for a finite-horizon game, and the Riccati matrix algebraic equation fora n infinite-horizon game.
In this work we study an interception game, which is a special regular differential game, in order to minimize with respect to u and maximize with respect to v the associated regularized cost functional. This then suggested exploring the equivalent saddle-point reformulation.
Consider the Hilbert space L 2 [ 0 , t f ] and let Q L 2 [ 0 , t f ] and R L 2 [ 0 , t f ] be two closed, convex and bounded sets of admissible controls. Then, solving (1)–(4) is equivalent to solving the following min-max problem.
min u Q max v R J ε ( u , v )
where J ε is continuous, convex–concave (convex in u and concave in v) and differentiable. See [24] for further details.
A saddle-point reformulation of the min-max problem (7) is formulated as finding a point u * , v * Q × R such that
J ε ( u * , v ) J ε ( u * , v * ) J ε ( u , v * )
for all v R and u Q .
Another known relationship with the above assumptions on J ε is the following variational inequality reformulation. The saddle-point problem (8) is equivalent to finding a point u * , v * Q × R such that
δ J ε δ u ( u * , v * ) δ J ε δ v ( u * , v * ) , u v u * v * 0 for all ( u , v ) Q × R
where · , · is an appropriate inner product and δ J ε / ( δ u ) and δ J ε / ( δ v ) are the variational derivatives of the functional J ε , as will be explained later (can be thought of as partial derivatives in the case of real functions).
Saddle-point problems (as well as variational inequalities) stand at the core of many real-world applications in convex programming, game theory and many more instances; see, e.g., Rockafellar [25]. In [21] we considered two gradient methods for solving saddle-point problems, the Arrow–Hurwicz–Uzawa algorithm [26] and Korpelevich’s extragradient method [27]; see also [28,29,30,31,32] and the many references therein.
These methods use gradients in each of their update rules. Thus we recall the variational/functional derivative definition next (see, for example, [33]).
Definition 3.
Consider an integral functional J ( x ( t ) ) of an argument x ( t ) . The variational/functional derivative of J ( x ( t ) ) with respect to x ( t ) , δ J δ x : [ 0 , T ] R n is defined as
δ J ( x ( t ) ) h ( t ) = 0 t f δ J δ x ( t ) h ( t ) d t .
Recall that for a function f : R n R , the gradient f is defined by
d d ϵ f ( x + ϵ h ) ϵ = 0 = f ( x ) · h
Thus the variational derivative works similarly as such; that is,
d d ϵ J ( x + ϵ h ) ϵ = 0 = δ J ( x ) · h ,
and with (9), we get
d d ϵ J ( x + ϵ h ) ϵ = 0 = 0 t f δ J δ x h ( t ) d t .
Remark 1.
In a more general setting, where the zero-sum differential cost functional is not differentiable, subdifferentials/subgradients are needed; see [24,25,34] and references therein.
Now we recall the Arrow–Hurwicz–Uzawa method [26] and Korpelevich’s extragradient method [27], which we applied in our previous paper [21]. For simplicity we present the algorithms for solving variational inequalities, and clearly the translation to min-max and saddle-point problems can be easily derived.
Let H 1 , H 2 be two real Hilbert spaces and a bifunction F : H 1 × H 2 R with partial derivatives u F and v F . Choose an arbitrary starting point ( u 0 , v 0 ) Q × R and step-size α > 0 . Given the current iterate ( u k , v k ) , the Arrow–Hurwicz–Uzawa update rule is formulated as follows:
u k + 1 = P Q u k α u F ( u k , v k ) v k + 1 = P R v k + α v F ( u k , v k )
where P Q and P R are the orthogonal projection operators for the sets Q and R, respectively.
A related modification with weaker convergence assumptions is Korpelevich’s method, in which additional mid-points computations are done, corresponding to gradients, and thus its known name is the extragradient method:
u k ¯ = P Q u k α u F ( u k , v k ) v k ¯ = P R v k + α v F ( u k , v k ) u k + 1 = P Q u k α u F ( u k ¯ , v k ¯ ) v k + 1 = P R v k + α v F ( u k ¯ , v k ¯ )
Although convergence of the extragradient method is guaranteed under weaker assumptions than the Arrow–Hurwicz–Uzawa method, there is still the need to calculate two evaluations of F = ( u F , v F ) and two projections onto Q and R. One step in the direction of simplifying the extragradient method with respect to the double projections is Censor et al.’s [30,31] subgradient extragradient method. In this method, the second orthogonal projection onto R is replaced by an easy computed projection onto some constructible set T k .
u k ¯ = P Q u k α u F ( u k , v k ) v k ¯ = P R v k + α v F ( u k , v k ) u k + 1 = P T k u k α u F ( u k ¯ , v k ¯ ) v k + 1 = P T k v k + α v F ( u k ¯ , v k ¯ )
For avoiding the extra evaluations of F per each iteration, Popov [35] proposed the following modification introducing the so-called “leading” point:
u k ¯ = P Q u k α u F ( u k , v k ) v k ¯ = P R v k + α v F ( u k , v k ) u k + 1 = P Q u k ¯ α u F ( u k , v k ) v k + 1 = P R v k ¯ + α v F ( u k , v k )
A standard assumption, which we also use, for the convergence of the above methods, is the so-called roundedness of the derivatives, which means that the functional derivatives of J ε ( u k , v k ) and J ε ( u k , v k ) with respect to u and v, respectively, are uniformly bounded; i.e., there is a constant M > 0 such that
δ J ε ( u k , v k ) δ u t M , δ J ε ( u k , v k ) δ v t M ,
for all k 0 .
Since the introduction of the above methods, many modifications and extensions have been offered using various techniques—inertial, hybrid, viscosity and more; see, e.g., [36,37] and the references therein.

3. Interception Game

In this section we consider a particular singular problem (1)–(4), namely, n = 2 , r = 1 , s = 1 . The matrices of coefficients in (1) and (2) are
A = 0 1 0 0 , D = 0 0 0 0 , F = f 1 0 0 f 2
B T = ( 0 , 1 ) , C T = ( 0 , 1 ) , G v = g
with the scalars g , f 1 , f 2 > 0 .
The initial position z 0 is
z 0 T ( 0 ) = ( 0 , 1 ) .
The system (1) subject to the data (15), (16) has the following form:
d z 1 ( t ) d t = z 2 ( t ) d z 2 ( t ) d t = u ( t ) + v ( t )
The solution of (18) with initial position (17) has the following integral form:
z ( t ) = z 1 ( t ) T z 2 ( t ) T = M t M 0 1 z 0 ( 0 ) + 0 t M t M s 1 f s d s ,
where
M t = 1 t 0 1
is a fundamental matrix solution of the corresponding homogeneous system
d z 1 ( t ) d t = z 2 ( t ) d z 2 ( t ) d t = 0
and
f s = 0 u ( s ) + v ( s ) .
Thus, the analytical solution (after some technical calculations in (19) with (20) and (22)) can be written as
z 1 ( t ) = t + 0 t t s ( u ( s ) + v ( s ) ) d s
z 2 ( t ) = 1 + 0 t ( u ( s ) + v ( s ) ) d s
The system (18), with (17) is a linearized kinematic model of a planar engagement between two vehicles—an interceptor (pursuer) and a target (evader) where both vehicles are directly controlled by their lateral accelerations u ( t ) = a p ( t ) and v ( t ) = a e ( t ) , respectively. The coordinates of the state vector z ( t ) = ( z 1 ( t ) , z 2 ( t ) ) are the relative lateral separation and the relative lateral velocity of the vehicles. The basic schematic view of the planar engagement geometry is shown in Figure 1, where:
  • The x-axis of the coordinate system is aligned with the initial line of sight;
  • The points ( x p , y p ) , ( x e , y e ) are the current coordinates;
  • The origin is collocated with the initial pursuer position;
  • V p and V e are the velocities;
  • a p and a e are the lateral accelerations;
  • φ p and φ e are the respective aspect angles between velocities vectors and reference line of sight;
  • y = y e y p is the relative lateral separation normal to the initial sight of line;
  • r is the current range between the vehicles;
  • The line-of-sight angle λ is the angle between the current and initial lines of sight.
More details of such an engagement can be found, for instance, in [38,39].
The behavior of each player in this singular game is evaluated by the following regularized cost functional.
J ε ( u , v ) = f 1 z 1 2 ( t f ) + f 2 z 2 2 ( t f ) + 0 t f ε 2 u 2 t g v 2 ( t ) d t .
The cost functional (25) has to be minimized by the pursuer u ( t ) and maximized by the evader v ( t ) .
The game, consisting of the dynamics (1), with the data (15) and (16), initial condition (17) and cost functional (25) is called the interception differential game.
By combining the above data and definitions and substituting in the functional (25) we obtain:
J ε ( u , v ) = f 1 t f + 0 t f t f t ( u ( t ) + v ( t ) ) d t + f 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t 2 + 0 t f ε 2 u 2 t g v 2 ( t ) d t .
For the numerical validation of the saddle-point approximation of this functional, we next develop the appropriate variational derivatives.
Theorem 1.
The variational derivatives of (26) are given by:
δ J ε δ u = 2 f 1 ( t f t ) t f + 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t + 2 f 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t + 2 ε 2 u ( t ) ,
and
δ J ε δ v = 2 f 1 ( t f t ) t f + 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t + 2 f 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t 2 g v ( t ) .
Proof. 
Our functional J ε has two dependent variables of t (u and v), and we have to find the variational derivative with respect to u and v. In the view of above arguments, we have
0 t f δ J ε δ u h 1 ( t ) d t = δ J ε ( u ) · h 1 = ϵ 1 J ε ( u + ϵ 1 h 1 , v + ϵ 2 h 2 ) ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) ,
and
0 t f δ J ε δ v h 2 ( t ) d t = δ J ε ( v ) · h 2 = ϵ 2 J ε ( u + ϵ 1 h 1 , v + ϵ 2 h 2 ) ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) .
Equations (29) and (30) will render the functional derivatives δ J ε δ u and δ J ε δ v , respectively. Therefore, we start with the (29):
ϵ 1 J ε ( u + ϵ 1 h 1 , v + ϵ 2 h 2 ) ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) = ϵ 1 f 1 t f + 0 t f ( t f t ) ( u ( t ) + ϵ 1 h 1 ( t ) + v ( t ) + ϵ 2 h 2 ( t ) ) d t 2 ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) + ϵ 1 f 2 1 + 0 t f ( u ( t ) + ϵ 1 h 1 ( t ) + v ( t ) + ϵ 2 h 2 ( t ) ) d t 2 ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) + ϵ 1 0 t f ( ε 2 ( u ( t ) + ϵ 1 h 1 ( t ) ) 2 g ( v ( t ) + ϵ 2 h 2 ( t ) ) 2 ) d t ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) = 2 f 1 t f + 0 t f ( t f t ) ( u ( t ) + ϵ 1 h 1 ( t ) + v ( t ) + ϵ 2 h 2 ( t ) ) d t 0 t f ( t f t ) h 1 ( t ) d t ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) + 2 f 2 1 + 0 t f ( u ( t ) + ϵ 1 h 1 ( t ) + v ( t ) + ϵ 2 h 2 ( t ) ) d t 0 t f h 1 ( t ) d t ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) + 0 t f 2 ε 2 ( u ( t ) + ϵ 1 h 1 ( t ) ) h 1 ( t ) ) d t ( ϵ 1 , ϵ 2 ) = ( 0 , 0 ) = 2 f 1 t f + 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t 0 t f ( t f t ) h 1 ( t ) d t + 2 f 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t 0 t f h 1 ( t ) d t + 0 t f 2 ε 2 u ( t ) h 1 ( t ) d t = 2 f 1 t f 0 t f ( t f t ) h 1 ( t ) d t + 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t 0 t f ( t f t ) h 1 ( t ) d t + 2 f 2 0 t f 1 + 0 t f ( u ( t ) + v ( t ) ) d t h 1 ( t ) d t + 0 t f 2 ε 2 u ( t ) h 1 ( t ) d t = 2 f 1 t f 0 t f ( t f t ) h 1 ( t ) d t + 0 t f ( t f t ) 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t h 1 ( t ) d t + 2 f 2 0 t f 1 + 0 t f ( u ( t ) + v ( t ) ) d t h 1 ( t ) d t + 0 t f 2 ε 2 u ( t ) h 1 ( t ) d t .
The above expression together with (10) yield
0 t f δ J ε δ u h 1 ( t ) d t = 0 t f [ 2 f 1 t f ( t f t ) + 2 f 1 ( T t ) 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t
+ 2 f 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t + 2 ε 2 u ( t ) ] h 1 ( t ) d t .
It follows that the variational derivative of J ε with respect to u is given by
δ J ε δ u = 2 f 1 t f ( t f t ) + 2 f 1 ( t f t ) 0 t f ( t f t ) ( u ( t ) + v ( t ) ) d t + 2 f 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t + 2 ε 2 u ( t ) .
with some simplifications, we obtain (27), and in a similar way we derive (28); and the proof is complete. □

4. Duality Representation

We consider the functional (26) with the constants f 1 , f 2 , g > 0 and a small parameter ε > 0 , and present its variational derivatives.
Theorem 2.
The dual representation of the functional (26) is given by:
u * ( t ) = 1 2 ε 2 h T ( t ) * ,
v * ( t ) = 1 2 g h T ( t ) * ,
where
* = arg max E 2 T d 1 4 T G .
or equivalently in scalar form
* = arg max E 2 1 t f + 2 1 4 1 2 f 1 + 2 2 f 2 1 4 t f 1 ε 2 1 g 1 2 t f 3 3 + 1 2 t f + 2 2
and all the coefficients will be presented in the proof itself.
Proof. 
In the lines of [40], let us calculate the program maxi-min:
ρ = max v ( · ) L 2 [ 0 , t f ] min u ( · ) L 2 [ 0 , t f ] J ε
Using [41]
z T R z = max E n T z 1 4 T R 1 ,
where R is symmetric positive definite matrix. Note that
f 1 z 1 2 ( t f ) + f 2 z 2 2 ( t f ) = z T ( t f ) F z ( t f ) ,
in (25) where z = ( z 1 , z 2 ) T ,
F = f 1 0 0 f 2 .
Then,
J ε = max E 2 φ ( , u ( · ) , v ( · ) ) ,
where
φ ( , u ( · ) , v ( · ) ) = T z ( t f ) 1 4 T F 1 + 0 t f ε 2 u 2 t g v 2 ( t ) d t ,
F 1 = 1 f 1 0 0 1 f 2 .
Consequently,
ρ = max v ( · ) min u ( · ) max E 2 φ ( , u ( · ) , v ( · ) ) .
Then (40) becomes:
φ ( , u ( · ) , v ( · ) ) = 1 t f + 0 t f t f t ( u ( t ) + v ( t ) ) d t + 2 1 + 0 t f ( u ( t ) + v ( t ) ) d t
1 4 1 f 1 1 2 + 1 f 2 2 2 + 0 t f ε 2 u 2 t g v 2 ( t ) d t =
1 T + 2 1 4 1 f 1 1 2 + 1 f 2 2 2 + φ u ( , u ( · ) ) + φ v ( , v ( · ) ) ,
where
φ u ( , u ( · ) ) 0 t f 1 ( t f t ) + 2 u ( t ) + ε 2 u 2 ( t ) d t ,
φ v ( , v ( · ) ) 0 t f 1 ( t f t ) + 2 v ( t ) g v 2 ( t ) d t ,
In (42), the operations of maximum over E 2 and minimum over u ( · ) L 2 [ 0 , t f ] commute. Therefore,
ρ = max E 2 max v ( · ) min u ( · ) φ ( , u ( · ) , v ( · ) ) =
max 1 , 2 E 1 t f + 2 1 4 1 f 1 1 2 + 1 f 2 2 2 + min u ( · ) φ u ( , u ( · ) ) + max v ( · ) φ v ( , v ( · ) ) .
The inner minimizer and maximizer in (46) are
u * ( t , ) = 1 ( t f t ) + 2 2 ε 2 ,
v * ( t , ) = 1 ( t f t ) + 2 2 g .
Substituting (47) and (48) into (43),
χ ( ) = φ ( , u * ( · , ) , v * ( · , ) ) =
1 T + 2 1 4 1 f 1 1 2 + 1 f 2 2 2 + φ u ( , u * ( · , ) + φ v ( , v * ( · , ) ) =
T d 1 4 T G ,
where
d = ( T , 1 ) T ,
G = F 1 + μ 0 t f h ( t ) h T ( t ) d t ,
μ = 1 ε 2 1 g ,
h ( t ) = ( t f t , 1 ) T .
Thus,
G = 1 f 1 + μ t f 3 3 μ t f 2 2 μ t f 2 2 1 f 2 + μ t f
Observe that the above is associated with the theory of symmetrical matrices [42]. Moreover, note that the studied interception differential game is solvable if the matrix G given by (54) is positive definite. Hence, the desired result has been obtained. □

5. Numerical Validation

In this section we examine the numerical behavior of the double projection methods described in Section 2 (Arrow–Hurwicz–Uzawa [26], Korpelevich’s extragradient [27] and Popov [35]). We show that under boundedness of the derivatives assumption (14) the results such as ε 0 + coincide with the dual development of the previous section. We choose g = 4 , f 1 , f 2 = 0.5 , t f = 4 and present the results for ε = 0.1 , 0.01 , 0.001 . Observe that the Table 1, Table 2 and Table 3 and Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6 show that for ε 0 + , the values z 1 ( 4 ) and z 2 ( 4 ) go to zero, meaning that the relative lateral velocity z 2 and relative lateral separation z 1 of the vehicles tend to zero at the end of the game, and the objective of the interception control is achieved. Observe also that the graphs of the inner minimizer and maximizer in Figure 2 coincide with the results in Table 2. Moreover, the values of the cost functional (26) decrease to 0.
Since we barely noticed any major differences between the numerical methods, we decided to present only the results of Popov [35].

6. Conclusions

In this work we presented a singular, zero-sum, linear-quadratic differential game in which the weight matrix of the minimizer’s control cost in the cost functional is singular. As an application we focused on an interception differential game and introduced a regularized cost functional; we examined its dual representation and validated it via numerical schemes for finding saddle points.

Author Contributions

All authors contributed equally to the following: conceptualization, methodology, formal analysis, writing—original draft preparation. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The study does not report any data.

Acknowledgments

The authors would like to thank the referees for their comments on the manuscript which helped in improving earlier versions of this paper. Moreover, we acknowledge Vladimir Turetsky useful help regarding the duality representation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Turetsky, T.; Glizer, V.Y. Robust state-feedback controllability of linear systems to a hyperplane in a class of bounded controls. Optim. Theory Appl. 2004, 123, 639–667. [Google Scholar] [CrossRef]
  2. Forouhar, K. Singular Differential Game Numerical Techniques and Closed Loop Strategies; University of California Press: Los Angeles, CA, USA, 1980. [Google Scholar]
  3. Glizer, V.Y.; Shinar, J. On the structure of a class of time-optimal trajectories. Optim. Control Appl. Methods 1993, 14, 271–279. [Google Scholar] [CrossRef]
  4. Shinar, J.; Glizer, V.Y. Application of receding horizon control strategy to pursuit-evasion problems. Optim Control Appl Methods 1995, 16, 127–141. [Google Scholar] [CrossRef]
  5. Simakova, E.N. Differential pursuit game. Autom. Remote Control 1967, 28, 173–181. [Google Scholar]
  6. Turetsky, T.; Glizer, V.Y.; Shinar, J. Robust trajectory tracking: Differential game/cheap control approach. Int. J. Syst. Sci. 2014, 45, 2260–2274. [Google Scholar] [CrossRef]
  7. Hu, Y.; Oksendal, B.; Sulem, A. Singular mean-field control games with applications to optimal harvesting and investment problems. arXiv 2014, arXiv:1406.1863v1. [Google Scholar]
  8. Isaacs, R. Differential Games; John Wiley and Sons: New York, NY, USA, 1967. [Google Scholar]
  9. Basar, T.; Olsder, G.J. Dynamic Noncooperative Game Theory; Academic Press: London, UK, 1992. [Google Scholar]
  10. Bryson, A.E., Jr.; Ho, Y.-C. Applied Optimal Control; Taylor & Francis Group: New York, NY, USA, 1975. [Google Scholar]
  11. Krasovskii, N.N.; Subbotin, A.I. Game-Theoretical Control Problems; Springer: New York, NY, USA, 1988. [Google Scholar]
  12. Shinar, J.; Glizer, V.Y.; Turetsky, V. Solution of a singular zero-sum linear-quadratic differential game by regularization. Int. Game Theory Rev. 2014, 16, 14–32. [Google Scholar] [CrossRef]
  13. Amato, F.; Pironti, A. A Note on singular zero-sum linear quadratic differential games. In Proceedings of the 33rd Conference on Decision and Control, Lake Buena Vista, FL, USA, 14–16 December 1994; IEEE Publishing: New York, NY, USA, 1994; pp. 1533–1535. [Google Scholar]
  14. Stoorvogel, A.A. The singular zero-sum differential game with stability using H control theory. Math. Control Signals Syst. 1991, 4, 121–138. [Google Scholar] [CrossRef]
  15. Stoorvogel, A.A. The H Control Problem: A State Space Approach; University of Michigan Press: Ann Arbor, MI, USA, 2000. [Google Scholar]
  16. Glizer, V.Y.; Kelis, O. Solution of a zero-sum linear quadratic differential game with singular control cost of minimiser. Control Decis. 2015, 2, 155–184. [Google Scholar] [CrossRef]
  17. Glizer, V.Y. Asymptotic solution of zero-sum linear-quadratic differential game with cheap control for the minimizer. Nonlinear Differ. Equ. Appl. 2000, 7, 231–258. [Google Scholar] [CrossRef]
  18. Petersen, I.R. Linear-quadratic differential games with cheap control. Syst. Control Lett. 1986, 8, 181–188. [Google Scholar] [CrossRef]
  19. Starr, A.W.; Ho, Y.-C. Nonzero-sum differential games. Optim. Theory Appl. 1969, 3, 184–206. [Google Scholar] [CrossRef]
  20. Turetsky, T.; Glizer, V.Y. Robust solution of a time-variable interception problem: A cheap control approach. Int. Game Theory Rev. 2007, 9, 637–655. [Google Scholar] [CrossRef]
  21. Gibali, A.; Kelis, O. Gradient methods for solving zero-sum linear-quadratic differential games. Appl. Anal. Optim. 2018, 2, 237–252. [Google Scholar]
  22. Glizer, V.Y.; Kelis, O. Solution of a singular infinite horizon zero-sum linear-quadratic differential game: A regularization approach. In Proceedings of the IEEE 23rd Mediterranean Conference on Control and Automation (MED2015), Torremolinos, Spain, 16–19 June 2015; pp. 390–397. [Google Scholar]
  23. Glizer, V.Y.; Kelis, O. Singular infinite horizon zero-sum linear-quadratic differential game: Saddle-point equilibrium sequence. Numer. Algebra Control Optim. 2017, 7, 1–20. [Google Scholar] [CrossRef] [Green Version]
  24. Goebel, R. Convexity in zero-sum differential games. Control Optim. 2002, 40, 1491–1504. [Google Scholar] [CrossRef]
  25. Rockafellar, R.T. Convex Analysis; Princeton University Press: Princeton, NJ, USA, 1970. [Google Scholar]
  26. Arrow, K.J.; Hurwicz, L.; Uzawa, H. Studies in Linear and Non-Linear Programming; Stanford University Press: Stanford, CA, USA, 1958. [Google Scholar]
  27. Korpelevich, G.M. The extragradient method for finding saddle points and other problems. Ekon. Mat. Metod. 1976, 12, 747–756. [Google Scholar]
  28. Antipin, A.S. On a method for convex programs using a symmetrical modification of the Lagrange function. Ekon. Mater. Metod. 1976, 12, 1164–1173. [Google Scholar]
  29. Censor, Y.; Gibali, A.; Reich, S. Strong convergence of subgradient extragradient methods for the variational inequality problem in Hilbert space. Optim. Methods Softw. 2011, 26, 827–845. [Google Scholar] [CrossRef]
  30. Censor, Y.; Gibali, A.; Reich, S. The subgradient extragradient method for solving variational inequalities in Hilbert space. Optim. Theory Appl. 2011, 148, 318–335. [Google Scholar] [CrossRef] [Green Version]
  31. Censor, Y.; Gibali, A.; Reich, S. Extensions of Korpelevich’s extragradient method for solving the variational inequality problem in Euclidean space. Optimization 2012, 61, 1119–1132. [Google Scholar] [CrossRef]
  32. Zaslavski, A.J. Numerical Optimization with Computational Errors; Springer: Cham, Switzerland, 2016. [Google Scholar]
  33. Gelfand, I.M.; Fomin, S.V. Calculus of Variations; Dover Publications, Inc.: Mineola, NY, USA, 2000. [Google Scholar]
  34. Nedić, A.; Ozdaglar, A. Subgradient methods for saddle-point problems. Optim. Theory Appl. 2009, 142, 205–228. [Google Scholar] [CrossRef]
  35. Popov, L.D. A modification of the Arrow-Hurwicz method for finding saddle points. Math. Notes 1980, 28, 845–848. [Google Scholar] [CrossRef]
  36. Malitsky, Y.V.; Semenov, V.V. An extragradient algorithm for monotone variational inequalities. Cybern. Syst. Anal. 2014, 50, 271–277. [Google Scholar] [CrossRef]
  37. Thong, D.V.; Li, X.H.; Dong, Q.L. An inertial Popov’s method for solving pseudomonotone variational inequalities. Optim. Lett. 2021, 15, 757–777. [Google Scholar] [CrossRef]
  38. Gutman, S.; Leitmann, G. Optimal strategies in a neighborhood of a collision course. AIAA J. 1976, 14, 1210–1212. [Google Scholar] [CrossRef]
  39. Shinar, J. Solution techniques for realistic pursuit-evasion games. In Control and Dynamic Systems; Academic Press: New York, NY, USA, 1981; Volume 17, pp. 63–124. [Google Scholar]
  40. Shinar, J.; Glizer, V.Y.; Turetsky, V.; Ianovsky, E. Solvability of linearquadratic differential games associated with pursuit-evasion problems. Int. Game Theory Rev. 2008, 10, 481–515. [Google Scholar] [CrossRef]
  41. Bellman, R. Introduction to Matrix Analysis; McGraw-Hill Book Co., Inc.: New York, NY, USA; Toronto, ON, Canada; London, UK, 1960. [Google Scholar]
  42. Crasmareanu, M. The determinant inner product and the Heisenberg product of Sym(2). Int. Electron. J. Geom. 2021, 14, 145–156. [Google Scholar]
Figure 1. Geometry of the interception game.
Figure 1. Geometry of the interception game.
Axioms 10 00066 g001
Figure 2. u * ( t ) and v * ( t ) .
Figure 2. u * ( t ) and v * ( t ) .
Axioms 10 00066 g002
Figure 3. The relative lateral separation z 1 ( t ) for varying ε .
Figure 3. The relative lateral separation z 1 ( t ) for varying ε .
Axioms 10 00066 g003
Figure 4. A zoom of z 1 ( 4 ) for varying ε .
Figure 4. A zoom of z 1 ( 4 ) for varying ε .
Axioms 10 00066 g004
Figure 5. The relative lateral velocity z 2 ( t ) for varying ε .
Figure 5. The relative lateral velocity z 2 ( t ) for varying ε .
Axioms 10 00066 g005
Figure 6. A zoom of z 2 ( 4 ) for varying ε .
Figure 6. A zoom of z 2 ( 4 ) for varying ε .
Axioms 10 00066 g006
Table 1. Duality.
Table 1. Duality.
ε 1 2 *
0.1 0.007417428195 0.00977333591 0.009948188435
0.01 0.00007499156434 0.00009997687968 0.00009999468885
0.001 7.499991563 × 10 7 9.999976875 × 10 7 9.999994680 × 10 7
Table 2. The inner minimizer and maximizer for varying ε .
Table 2. The inner minimizer and maximizer for varying ε .
ε u * ( t ) v * ( t )
0.1 0.9948188435 + 0.3708714098 · t 0.002487047109 0.0009271785245 · t
0.01 0.9999468885 + 0.3749578217 · t 0.00002499867221 0.000009373945540 · t
0.001 0.9999994690 + 0.3749995782 · t 2.499998672 × 10 7 9.374989455 × 10 8 · t
Table 3. Numerical calculations for varying ε .
Table 3. Numerical calculations for varying ε .
ε z 1 ( 4 ) z 2 ( 4 ) the Cost Functional (26)
0.1 0.007417429 0.009773335 0.009948188431
0.01 0.000074991 0.000099977 0.00009999468883
0.001 7.49 × 10 7 0.000001 9.999994685 × 10 7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gibali, A.; Kelis, O. An Analytic and Numerical Investigation of a Differential Game. Axioms 2021, 10, 66. https://doi.org/10.3390/axioms10020066

AMA Style

Gibali A, Kelis O. An Analytic and Numerical Investigation of a Differential Game. Axioms. 2021; 10(2):66. https://doi.org/10.3390/axioms10020066

Chicago/Turabian Style

Gibali, Aviv, and Oleg Kelis. 2021. "An Analytic and Numerical Investigation of a Differential Game" Axioms 10, no. 2: 66. https://doi.org/10.3390/axioms10020066

APA Style

Gibali, A., & Kelis, O. (2021). An Analytic and Numerical Investigation of a Differential Game. Axioms, 10(2), 66. https://doi.org/10.3390/axioms10020066

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop