Next Article in Journal
Illustration of Convective Boundary Conditions on the Darcy–Forchheimer Flow of Nanofluid with the Impact of Chemical Reaction
Previous Article in Journal
Fractional Integrals Associated with the One-Dimensional Dunkl Operator in Generalized Lizorkin Space
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Equivalence Conditions of Optimal Feedback Control-Strategy Operators for Zero-Sum Linear Quadratic Stochastic Differential Game with Random Coefficients

1
College of Computer Science, Chengdu University, Chengdu 610106, China
2
Faculty of Science, Yibin University, Yibin 644000, China
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(9), 1726; https://doi.org/10.3390/sym15091726
Submission received: 17 July 2023 / Revised: 1 September 2023 / Accepted: 4 September 2023 / Published: 8 September 2023

Abstract

:
From the previous work, when solving the LQ optimal control problem with random coefficients (SLQ, for short), it is remarkably shown that the solution of the backward stochastic Riccati equations is not regular enough to guarantee the robustness of the feedback control. As a generalization of SLQ, interesting questions are, “how about the situation in the differential game?”, “will the same phenomenon appear in SLQ?”. This paper will provide the answers. In this paper, we consider a closed-loop two-person zero-sum LQ stochastic differential game with random coefficients (SDG, for short) and generalize the results of Lü–Wang–Zhang into the stochastic differential game case. Under some regularity assumptions, we establish the equivalence between the existence of the robust optimal feedback control strategy operators and the solvability of the corresponding backward stochastic Riccati equations, which leads to the existence of the closed-loop saddle points. On the other hand, the problem is not closed-loop solvable if the solution of the corresponding backward stochastic Riccati equations does not have the needed regularity.

1. Introduction

In this paper, we consider the following controlled linear stochastic differential equation (SDE, for short) on a finite horizon [ s , T ] :
d x ( t ) = [ A ( t ) x ( t ) + B 1 ( t ) u 1 ( t ) + B 2 ( t ) u 2 ( t ) ] d t + i = 1 d [ C i ( t ) x ( t ) + D 1 i ( t ) u 1 ( t ) + D 2 i ( t ) u 2 ( t ) ] d W i ( t ) , t [ s , T ] x ( s ) = θ ,
with the following quadratic objective functional:
J θ ( u 1 , u 2 ) = 1 2 E { G x ( T ) , x ( T ) R n + s T [ R ( t ) x ( t ) , x ( t ) R n + N 1 ( t ) u 1 ( t ) , u 1 ( t ) R m 1 + N 2 ( t ) u 2 ( t ) , u 2 ( t ) R m 2 ] d t } ;
for simplicity of notion, sometimes we denote m = m 1 + m 2 and:
B = B 1 B 2 , D i = D 1 i D 2 i , i = 1 , 2 , , d , u = u 1 u 2 , N = N 1 O O N 2 .

1.1. Literature Review

Thanks to the key role of economics and finance in the real world, much attention has been paid by many researchers to mathematical financial problems, including the differential game theory, not only because of its useful basic theory, but also because of its practical applications to financial markets. The differential game (DG, for short), which involves the decision-making of multiple players being affected by each other, could be traced back to 1965 and the work carried out by Isaacs [1], which inspired later research in this field. For research in the early years, we refer to the literature [2,3,4,5,6,7] and the references therein. Among them, Ho-Bryson-Baron [7] may be the first paper about deterministic two-person zero-sum differential games (DG, for short) where the stochastic perturbation is absent. In 1989, Fleming-Souganidis [6] generalized the DG in the deterministic framework to the stochastic case, which is considered the major work of the research on the problem (SDG). After that, Hamadè-Lepeltier [8] investigated the SDG using the BSDEs approach. Then, Yong and Mou-Yong [9,10] generalized the SDG into a leader–follower case and investigated it in the open-loop form using a Hilbert space method. In 2008, Buckdahn-Li [11] focused on the existence of players’ value through the viscosity solution theory. Yu [12] obtained the existence result of the corresponding Riccati equation by virtue of Hamadè’s linear transform, which enriched the related Riccati equation theory, and studied the general case of SDG (i.e., D 1 i 0 and D 2 i 0 ). Very recently, Moon [13] studied the LQ stochastic leader–follower Stackelberg differential games for jump-diffusion systems with random coefficients. In addition, Wang-Wu [14] considered a kind of time-inconsistent linear-quadratic non-zero sum stochastic differential game problem with random jumps. For more details, we refer to [13,14,15,16,17,18] and the references therein.
To some extent, the problem of SDG can be treated as the generalization of a stochastic LQ optimal control problem with two controls. From this viewpoint, these two problems have a close relationship with each other and benefit from each other. In 2003, Tang [19] obtained the equivalence between the existence of the solution for the Riccati equation and the homomorphism of the stochastic flows derived from the optimally controlled system, which solved the long-standing open problem proposed by Bismut [20] in 1978. After that, Tang [21] improved the well-posedness results in [19] through dynamic programming. In 2017, Lü-Wang-Zhang [22] showed the nonexistence of feedback controls to a solvable SLQ with random coefficients, which is significantly different from its deterministic counterpart. Very recently, Zhang-Dong-Meng [23] investigated the well-posedness of stochastic Riccati equations associated with stochastic linear quadratic optimal control with jumps and random coefficients using an approach similar to that introduced in [21]. For more essential details on SLQ, one can refer to [23,24,25,26,27] and the rich references therein.

1.2. Main Contributions of This Paper

As is well-known, when solving the closed-loop SLQ or SDG, the key point is the solvability of the corresponding backward stochastic Riccati equations in the form of the following:
d P = F ( A , B , C i , D i ; R , N ; P , Q i ) d t + i = 1 d Q i d W i ( t ) , t [ s , T ] P ( T ) = G , Q : = ( Q 1 , Q 2 , , Q d ) ,
which is a highly nonlinear backward stochastic differential equation; the coefficients will be specified later. Fortunately, Tang [19] obtained a unique solution, ( P , Q ) L F t ( 0 , T ; S ( R n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; S ( R n ) ) ) for p [ 1 , ) . After that, Lü-Wang-Zhang [22] remarkably showed that it is not regular enough to establish the feedback control because the second part of the solution Q L F t p ( Ω ; L 2 ( 0 , T ; S ( R n ) ) ) will be involved in the feedback control (for details please see Remark 3). This phenomenon is interesting and important. As a generalization of SLQ, a natural question is,“will the same phenomenon appear in SDG?” Obviously, the answer is affirmative and this paper will show that the same remarkable phenomenon appears in the SDG problem. Then, inspired by the idea from [16,22], we establish the equivalence between the existence of the robust optimal feedback control-strategy operators (please see Section 2 for the definition) and the solvability of the corresponding backward stochastic Riccati equations. Furthermore, the existence of a closed-loop saddle point will be obtained by way of some suitable conditions. Finally, from the example, the SDG problem is not closed-loop solvable without the needed regularity of the solution for the corresponding backward stochastic Riccati equations.
Note that, in the literature mentioned above, no works focus on the above phenomenon in SDG. Especially, when considering recent works [13,14,18,23] ([14,18] only consider the deterministic coefficients case [13,23] and do not pay attention to the robustness of the feedback control), to the best of our knowledge, the robustness of the feedback control-strategy pair in SDG with random coefficients has not yet been studied. This paper can be viewed as a nontrivial extension of [22] in the differential game case. The main generalizations and technical challenges of this paper are compared with [22] as follows:
(a)
The inclusion of two controls appears in Equations (1) and (2).
(b)
The explicit dependence of two controls on both the drift and diffusion terms in Equation (1).
(c)
The backward stochastic Riccati equations above with a high dimension.

1.3. Problem Statement

Throughout the paper, we let R n be the n-dimensional Euclidean space with Euclidean norm | · | R n and Euclidean inner product · , · R n . We also let S ( R n ) be the set of all ( n × n ) symmetric matrices, S + ( R n ) be the set of ( n × n ) positive definite matrices, and S ( R n ) be the set of ( n × n ) negative definite matrices. Furthermore, we suppose that ( Ω , F , ( F t ) t s , P ) is a complete filtered probability space with the natural filtration ( F t ) t s generated by a d-dimensional standard Brownian motion { W t = ( W 1 , W 2 , , W d ) ; s t T } augmented by all P-null sets, where the superscript ( · ) denotes the transpose of vectors or matrices.
In Equations (1) and (2) above, x ( · ) is the state variable in R n , the matrix-valued stochastic processes A ( · ) , B 1 ( · ) , B 2 ( · ) , C i ( · ) , D 1 i ( · ) , D 2 i ( · ) , R ( · ) , N 1 ( · ) , N 2 ( · ) are given F t s -progressively measurable processes, and G is given F T s -measurable random variables satisfying suitable assumptions to be given later such that Equation (1) admits a unique strong solution. In what follows, the time variable t is sometimes suppressed in the form A , B 1 , C i , D 1 i , etc. For any s [ 0 , T ) and i = 1 , 2 , we introduce the following collections of all admissible controls:
U i [ s , T ] : = u i R m i | u i ( · ) is F t s - progressively measurable , E s T | u i ( t ) | 2 d t < .
We also need to introduce the following collections of all admissible non-anticipative strategies for Player i , i = 1 , 2 on [ s , T ] , which is adopted from Buckdahn and Li [11] and Yu [12]:
A i [ s , T ] : = γ i : U j [ s , T ] U i [ s , T ] | γ i ( u j ) γ i ( u j * ) on [ s , τ ] , if u j u j * on [ s , τ ] ,
where j = 3 i and τ : Ω [ s , T ] is any stopping time. In the above collection, for i = 2 , then j = 1 , γ 2 ( u 1 ) means the admissible strategy for Player 2 to deal with the control of Player 1 in the game. Further, Player 1 aims to maximize (2) and Player 2 wants to minimize it. So, J θ ( u 1 , u 2 ) can be regarded as the payoff for Player 1 and the cost for Player 2. Above all, the two-person zero-sum linear quadratic stochastic differential game (SDG, for short) can be stated as follows.
Problem (SDG) For each pair ( s , θ ) , find an admissible control u ¯ 1 U 1 [ s , T ] and an admissible strategy γ ¯ 2 A 2 [ s , T ] such that:
J θ ( u ¯ 1 , γ ¯ 2 ( u ¯ 1 ) ) = inf γ 2 A 2 [ s , T ] sup u 1 U 1 [ s , T ] J θ ( u 1 , γ 2 ( u 1 ) ) ,
which is called Player 1’s value, and ( u ¯ 1 , γ ¯ 2 ) is called an optimal control-strategy pair of Player 1. On the other hand, if we find an admissible control u ¯ 2 U 2 [ s , T ] and an admissible strategy γ ¯ 1 A 1 [ s , T ] such that:
J θ ( γ ¯ 1 ( u ¯ 2 ) , u ¯ 2 ) = sup γ 1 A 1 [ s , T ] inf u 2 U 2 [ s , T ] J θ ( γ 1 ( u 2 ) , u 2 ) ,
which is called Player 2’s value and ( γ ¯ 1 , u ¯ 2 ) is called an optimal control-strategy pair of Player 2.
In the case that the two players’ values are equal, we call this common value the value of the game and ( u ¯ 1 , u ¯ 2 ) is called the saddle point, where u ¯ 1 = γ ¯ 1 ( u ¯ 2 ) and u ¯ 2 = γ ¯ 2 ( u ¯ 1 ) , these two conditions can be replaced by another two matrices (please see Remark 7 in Section 3).
This paper is organized as follows. Some preliminary results and definitions are provided in Section 2. Section 3 is devoted to presenting the main results of this paper. Finally, two examples are provided to illustrate our results.

2. Preliminaries

The following collections will be used in this paper: here, p , p 1 , p 2 [ 1 , ) :
L F T p ( Ω ; R n ) = ξ : Ω R n | ξ is F T - measurable , E | ξ | R n p < , L F t p ( Ω ; C ( s , T ; R n ) ) = { x : Ω × [ s , T ] R n | x ( t ) is F t - adapted , continuous , E sup t [ s , T ] | x ( t ) | R n p 1 p < } , L F t p 1 ( Ω ; L p 2 ( s , T ; R n ) ) = x : Ω × [ s , T ] R n | x ( t ) is F t - adapted , E s T | x ( t ) | R n p 2 d t p 1 p 2 < , L F t p 2 ( s , T ; L p 1 ( Ω ; R n ) ) = x : Ω × [ s , T ] R n | x ( t ) is F t - adapted , s T E | x ( t ) | R n p 1 p 2 p 1 d t < .
All the collections above are Banach spaces with the canonical norms. L F t ( Ω ; L p 2 ( s , T ; R n ) ) , L F t p 1 ( Ω ; L ( s , T ; R n ) ) and L F t ( Ω ; L ( s , T ; R n ) ) can be defined in the same way. For simplicity of notation, we denote L F t p 1 ( s , T ; L p 1 ( Ω ; R n ) ) by L F t p 1 ( s , T ; R n ) . In Problem (SDG), we adopt the following assumptions:
Assumption 1.
(AS1) The coefficients of the state Equation (1) satisfy the following:
A ( · ) L F t ( Ω ; L 1 ( 0 , T ; R n × n ) ) , B 1 ( · ) L F t ( Ω ; L 2 ( 0 , T ; R n × m 1 ) ) , D 1 i ( · ) L F t ( 0 , T ; R n × m 1 ) , C i ( · ) L F t ( Ω ; L 2 ( 0 , T ; R n × n ) ) , B 2 ( · ) L F t ( Ω ; L 2 ( 0 , T ; R n × m 2 ) ) , D 2 i ( · ) L F t ( 0 , T ; R n × m 2 ) .
Assumption 2.
(AS2) The weighting coefficients in Equation (2) satisfy the following:
R ( · ) L F t ( Ω ; L 1 ( 0 , T ; S ( R n ) ) , G L F T ( Ω ; S ( R n ) ) N 1 ( · ) L F t ( 0 , T ; S ( R m 1 ) ) , N 2 ( · ) L F t ( 0 , T ; S ( R m 2 ) ) .
Lemma 1
([16]). Suppose the assumption (AS1) holds. For any ( s , θ ) [ 0 , T ) × L F s 2 ( Ω ; R n ) , and u 1 U 1 L F t 2 ( s , T ; R m 1 ) ) , u 2 U 2 L F t 2 ( s , T ; R m 2 ) ) , the state Equation (1) admits a unique solution x ( · ; s , θ , u 1 ( · ) , u 2 ( · ) ) L F t 2 ( Ω ; C ( s , T ; R n ) ) .
Remark 1.
Further, under the assumption (AS2), Equation (2) is well-defined for any ( s , θ ) [ 0 , T ) × L F s ( Ω ; R n ) and u 1 U 1 [ s , T ] , u 2 U 2 [ s , T ] . So the Problem (SDG) provided before is meaningful.
Next, we present some well-known results, which will be used later.
Lemma 2.
For any ( s , θ ) [ 0 , T ) × L F s 2 ( Ω ; R n ) , the stochastic differential equation:
d x ( t ) = ( A ( t ) x ( t ) + f ( t ) ) d t + i = 1 d ( B i ( t ) x ( t ) + g i ( t ) ) d W i ( t ) , x ( s ) = θ
admits a unique F t -adapted solution x ( · ) L F t 2 ( Ω ; C ( s , T ; R n ) ) , if the coefficients A ( · ) L F t ( Ω ; L 2 ( s , T ; R n × n ) ) , B ( · ) L F t ( Ω ; L 2 ( s , T ; ( R n × n ) d ) ) , f ( · ) L F t 2 ( s , T ; R n × n ) and g ( · ) L F t 2 ( s , T ; ( R n × n ) d ) .
For details of the existence and uniqueness results of Equation (5), we refer to [27,28].
Lemma 3
([29]). Suppose that the coefficient f satisfies that:
(i) 
f ( · , 0 , 0 ) L F t ( Ω ; L 1 ( s , T ; R n ) ) ,
(ii) 
| f ( · , α 1 , β 1 ) f ( · , α 2 , β 2 ) | k 1 ( · ) | α 1 α 2 | + k 2 ( · ) | β 1 β 2 | , α 1 , α 2 , β 1 , β 2 R n , where k 1 ( · ) L F t ( Ω ; L 1 ( s , T ; R ) ) and k 2 ( · ) L F t ( Ω ; L 2 ( s , T ; R ) ) . Then, for any ξ L F T ( Ω ; R n ) , the backward stochastic differential equation:
d y ( t ) = f ( t , y ( t ) , z i ( t ) ) d t + i = 1 d z i ( t ) d W i ( t ) , y ( T ) = ξ
admits a unique solution ( y ( · ) , z ( · ) ) L F t ( Ω ; C ( s , T ; R n ) ) × L F t p ( Ω ; L 2 ( s , T ; R n × d ) ) , which is F t -adapted.
Further, we would like to recall the Pontryagin-type maximum principle for differential games; for further details, we refer to [30] and the references therein.
For any ( ω , t , x , y , z , u 1 , u 2 ) Ω × [ 0 , T ] × R n × R n × R n × d × R m 1 × R m 2 , we define a Hamiltonian function associated with Problem (SDG) by (the variable t is suppressed),
H ( t , x , y , z , u 1 , u 2 ) = A x + B 1 u 1 + B 2 u 2 , y + i = 1 d C i x + D 1 i u 1 + D 2 i u 2 , z i + 1 2 R x , x + N 1 u 1 , u 1 + N 2 u 2 , u 2 .
Lemma 4.
Let ( x ¯ ( · ) , u ¯ 1 ( · ) , u ¯ 2 ( · ) ) L F t 2 ( Ω ; C ( s , T ; R n ) ) × L F t 2 ( s , T ; R m 1 ) × L F t 2 ( s , T ; R m 2 ) be an optimal pair of Problem (SDG). Then there exists a pair ( y ¯ ( · ) , z ¯ ( · ) ) L F t 2 ( Ω ; C ( s , T ; R n ) ) × L F t 2 ( s , T ; R n × d ) satisfying the following BSDE:
d y ¯ ( t ) = A y ¯ ( t ) + i = 1 d ( C i ) z ¯ i ( t ) + R x ¯ ( t ) d t + i = 1 d z ¯ i ( t ) d W i ( t ) , y ¯ ( T ) = G x ¯ ( T ) ,
such that
( B 1 ) y ¯ ( · ) + i = 1 d ( D 1 i ) z ¯ i ( · ) + N 1 u ¯ 1 ( · ) = 0 , ( B 2 ) y ¯ ( · ) + i = 1 d ( D 2 i ) z ¯ i ( · ) + N 2 u ¯ 2 ( · ) = 0 .
As is well-known, stochastic Hamiltonian systems are very useful for deriving the open-loop form control-strategy for Problem (SDG), which is closely linked with the stochastic Riccati equation. Next, in order to investigate the closed-loop form (SDG), the associated backward stochastic Riccati equation takes the following form:
d P = F ( A , B , C i , D i ; R , N ; P , Q i ) d t + i = 1 d Q i d W i ( t ) , t [ s , T ] P ( T ) = G , Q : = ( Q 1 , Q 2 , , Q d ) ,
where the variable t is suppressed, and for any A R n × n , B R n × m , C = ( C 1 , , C d ) ( R n × n ) d , D = ( D 1 , , D d ) ( R n × m ) d ; R R n × n , N R m × m , P R n × n , Q = ( Q 1 , Q 2 , , Q d ) ( R n × n ) d with P and Q i ( i = 1 , , d ) being symmetric, the generator F is defined by:
F ( A , B , C i , D i ; R , N ; P , Q i ) : = A P + P A + R + i = 1 d ( C i ) P C i + i = 1 d [ ( C i ) Q i + Q i C i ] P B + i = 1 d ( C i ) P D i + i = 1 d Q i D i N + i = 1 d ( D i ) P D i 1 × P B + i = 1 d ( C i ) P D i + i = 1 d Q i D i .
It is well-known that Equation (9) is a backward stochastic differential equation with the generator F ( A , B , C i , D i ; R , N ; P , Q i ) being nonlinear in P and Q. For the simplicity of notation, we denote:
N ˜ 11 : = N 1 + i = 1 d ( D 1 i ) P D 1 i , N ˜ 22 : = N 2 + i = 1 d ( D 2 i ) P D 2 i , N ˜ 12 : = i = 1 d ( D 1 i ) P D 2 i , N ˜ : = N + i = 1 d ( D i ) P D i = N ˜ 11 N ˜ 12 ( N ˜ 12 ) N ˜ 22 , B ˜ 1 : = P B 1 + i = 1 d ( C i ) P D 1 i + i = 1 d Q i D 1 i , B ˜ 2 : = P B 2 + i = 1 d ( C i ) P D 2 i + i = 1 d Q i D 2 i , B ˜ : = P B + i = 1 d ( C i ) P D i + i = 1 d Q i D i = B ˜ 1 B ˜ 2 .
Remark 2.
The backward stochastic Riccati equation (BSRE, for short) above becomes the Riccati equation for SLQ when SDG involves only one player (i.e., B 2 , D 2 i , N 2 0 ) in the suitable dimensions, for the formal derivation about the Riccati equation for SLQ, we refer to Bismut [20,31]. Yu [12] solved Equation (9) without a stochastic part by virtue of Hamadè’s linear transform, which needed to restrict the matrices B and D i ( i = 1 , 2 , d ) to some special case,
B j = k 0 K j , D i j = k i K j , i = 1 , 2 , , d ; j = 1 , 2 ,
where K 1 L F t ( 0 , T ; S ( R n × m 1 ) ) and K 2 L F t ( 0 , T ; S ( R n × m 2 ) ) . In this paper, we release this restriction under some sharp regularity on feedback control-strategy operators (the definition will be given later).
With the notation above, Equation (10) can be rewritten as:
F ( A , B , C i , D i ; R , N ; P , Q i ) = A P + P A + R + i = 1 d ( C i ) P C i + i = 1 d [ ( C i ) Q i + Q i C i ] B ˜ N ˜ 1 ( B ˜ ) .
From the above notations, along with Yu’s [12] observation on the following, algebra equations,
N ˜ 11 u 1 ( · ) + N ˜ 12 u 2 ( · ) + ( B ˜ 1 ) x ( · ) = 0 , ( N ˜ 12 ) u 1 ( · ) + N ˜ 22 u 2 ( · ) + ( B ˜ 2 ) x ( · ) = 0 ,
solve them in a useful expression for studying the game problem with strategy, then derive the following two control-strategy forms for Player 1 and Player 2, respectively:
u ¯ 1 ( · ) = N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) x ( · ) , γ ¯ 2 ( u 1 ( · ) ) = N ˜ 22 1 ( B ˜ 2 ) x ( · ) + ( N ˜ 12 ) u 1 ( · )
and
γ ¯ 1 ( u 2 ( · ) ) = N ˜ 11 1 ( B ˜ 1 ) x ( · ) + N ˜ 12 u 2 ( · ) , u ¯ 2 ( · ) = N ˜ 22 ( N ˜ 12 ) N ˜ 11 1 N ˜ 12 1 B ˜ 2 B ˜ 1 N ˜ 11 1 N ˜ 12 x ( · ) .
Taking advantage of these two expressions, Equation (10) can be rewritten in the following two forms:
(i)
From Equation (14), F can be rewritten in the following Form (I):
F ( A , B , C i , D i ; R , N ; P , Q i ) = A P + P A + R + i = 1 d ( C i ) P C i + i = 1 d [ ( C i ) Q i + Q i C i ] B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) .
(ii)
From Equation (15), F can be rewritten in the following Form (II):
F ( A , B , C i , D i ; R , N ; P , Q i ) = A P + P A + R + i = 1 d ( C i ) P C i + i = 1 d [ ( C i ) Q i + Q i C i ] B ˜ 1 N ˜ 11 1 ( B ˜ 1 ) B ˜ 2 B ˜ 1 N ˜ 11 1 N ˜ 12 N ˜ 22 ( N ˜ 12 ) N ˜ 11 1 N ˜ 12 1 B ˜ 2 B ˜ 1 N ˜ 11 1 N ˜ 12 .
These two forms are convenient for completing the square of Players’ control strategies and also investigating the Players’ values.
Remark 3.
In 2003, Tang [19], connected the existence of a solution of the Riccati equation for Problem (SLQ) to the homomorphism of the stochastic flows derived from the optimally controlled system, and obtained a unique solution ( P , Q ) L F t ( 0 , T ; S ( R n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; S ( R n ) ) ) for a given p [ 1 , ) , from which we know that the optimal feedback operator for SLQ (denoted by Λ) only with the regularity Λ L F t p ( Ω ; L 2 ( 0 , T ; S ( R n ) ) ) . Although it showed that if x ¯ is an optimal state, then Λ x ¯ L F t 2 ( 0 , T ; R n ) is the desired optimal control, and such feedback control is not robust, even with regard to some small perturbation, i.e., one cannot conclude that Λ ( x ¯ + ε x ) ( ε x L F t 2 ( Ω ; C ( 0 , T , R n ) ) is a small perturbation) is an admissible control because the second part of solution Q will be involved in the feedback representation, which is not regular enough. Only when the optimal feedback operator has the sharp regularity Λ L F t ( Ω ; L 2 ( 0 , T ; S ( R n ) ) ) , does it become a robust optimal feedback.
Remark 4.
In Problem (SDG), because of the closed relationship with SLQ, it also meets the same difficulty stated in Remark 3, which denies the robustness of the control-strategy pair. Actually, taking advantage of the same direction of Yu’s [12] approach, i.e., Hamadè’s linear transform, we can obtain the solvability of Equation (9) with the existence results of BSRE in Tang’s paper [19]; nevertheless, the second part of the solution Q L F t p ( Ω ; L 2 ( 0 , T ; S ( R n ) ) ) is not regular enough to design a robust feedback control strategy.
Next, let us give the notions of robust optimal feedback control-strategy operators for Problem (SDG).
Definition 1.
The pair of stochastic processes ( ψ 1 ( · ) , Ψ 2 ( · ) ) is called a robust optimal feedback control-strategy operator of Player 1 for Problem (SDG) if
(i) 
ψ 1 ( · ) L F t ( Ω ; L 2 ( 0 , T ; ( R m 1 × n ) ) , Ψ 2 ( ψ 1 ( · ) ) L F t ( Ω ; L 2 ( 0 , T ; ( R m 2 × n ) ) .
(ii) 
For all ( s , θ ) [ 0 , T ) × L F s 2 ( Ω ; R n ) and u 1 ( · ) U 1 [ s , T ] , γ 2 ( · ) , γ ¯ 2 ( · ) A 2 [ s , T ] , it holds that
J θ ( u 1 ( · ) , γ ¯ 2 ( u 1 ( · ) ) ) J θ ( u 1 ( · ) , γ 2 ( u 1 ( · ) ) ) ,
and
J θ ( ψ 1 ( · ) x ¯ ( · ) , Ψ 2 ( ψ 1 ( · ) ) x ¯ ( · ) ) J θ ( u 1 ( · ) , γ ¯ 2 ( u 1 ( · ) ) ) ,
where x ¯ ( · ) is the solution of
d x ( t ) = [ A ( t ) + B 1 ( t ) ψ 1 ( t ) + B 2 ( t ) Ψ 2 ( ψ 1 ( t ) ) ] x ( t ) d t + i = 1 d [ C i ( t ) + D 1 i ( t ) ψ 1 ( t ) + D 2 i ( t ) Ψ 2 ( ψ 1 ( t ) ) ] x ( t ) d W i ( t ) , x ( s ) = θ .
Similarly, we can give the same notion of the robust optimal feedback control-strategy operator of Player 2.
Definition 2.
The pair of stochastic processes ( ψ 2 ( · ) , Ψ 1 ( · ) ) is called a robust optimal feedback control-strategy operator of Player 2 for Problem (SDG) if
(i) 
ψ 2 ( · ) L F t ( Ω ; L 2 ( 0 , T ; ( R m 2 × n ) ) , Ψ 1 ( ψ 2 ( · ) ) L F t ( Ω ; L 2 ( 0 , T ; ( R m 1 × n ) ) .
(ii) 
For all ( s , θ ) [ 0 , T ) × L F s 2 ( Ω ; R n ) and u 2 ( · ) U 2 [ s , T ] , γ 1 ( · ) , γ ¯ 1 ( · ) A 1 [ s , T ] , it holds that
J θ ( γ ¯ 1 ( u 2 ( · ) ) , u 2 ( · ) ) J θ ( γ 1 ( u 2 ( · ) ) , u 2 ( · ) ) ,
and
J θ ( Ψ 1 ( ψ 2 ( · ) ) x ¯ ( · ) , ψ 2 ( · ) x ¯ ( · ) ) J θ ( γ ¯ 1 ( u 2 ( · ) ) , u 2 ( · ) ) ,
where x ¯ ( · ) is the solution of
d x ( t ) = [ A ( t ) + B 1 ( t ) Ψ 1 ( ψ 2 ( t ) ) + B 2 ( t ) ψ 2 ( t ) ] x ( t ) d t + i = 1 d [ C i ( t ) + D 1 i ( t ) Ψ 1 ( ψ 2 ( t ) ) + D 2 i ( t ) ψ 2 ( t ) ] x ( t ) d W i ( t ) , x ( s ) = θ .
Remark 5.
In the Definitions above, γ ¯ 1 ( u 2 ( · ) ) and γ ¯ 2 ( u 1 ( · ) ) are the feedback strategies for Player 1 and Player 2, respectively, which means Player 1 (Player 2) in the game knows how to deal with the control for Player 2 (Player 1) in the following feedback control-strategy form:
γ ¯ 1 ( u 2 ( · ) ) = N ˜ 11 1 ( B ˜ 1 ) x ¯ ( · ) + N ˜ 12 u 2 ( · ) , γ ¯ 2 ( u 1 ( · ) ) = N ˜ 22 1 ( B ˜ 2 ) x ¯ ( · ) + ( N ˜ 12 ) u 1 ( · ) .
Remark 6.
Taking Definition 1, for example, the optimal feedback control-strategy operator ( ψ 1 ( · ) , Ψ 2 ( · ) ) is required to be independent of the initial state θ L F s 2 ( Ω ; R n ) . For a fixed ( s , θ ) [ 0 , T ) × L F s 2 ( Ω ; R n ) , Equations (18) and (19) imply that the control-strategy pair
u ¯ 1 ( · ) = ψ 1 ( · ) x ¯ ( · ) = N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) x ¯ ( · ) γ ¯ 2 ( u ¯ 1 ( · ) ) = Ψ 2 ( ψ 1 ( · ) ) x ¯ ( · ) = N ˜ 22 1 ( B ˜ 2 ) ( N ˜ 12 ) ψ 1 x ¯ ( · )
is optimal for Player 1 for Problem (SDG). Therefore, the existence of a robust optimal feedback operator implies the existence of an optimal control-strategy pair ( u ¯ 1 ( · ) , γ ¯ 2 ( u ¯ 1 ( · ) ) ) for Player 1. Player 2’s situation can be stated similarly.

3. Main Results

In this section, we state the main results of this paper, inspired by the idea from [12,16,22].
Theorem 1.
Suppose that the assumptions (AS1)–(AS2) hold. Then, Problem (SDG) admits a robust optimal feedback control-strategy operator ( ψ 1 ( · ) , Ψ 2 ( · ) ) for Player 1 if and only if:
(i) 
The Equation (9) admits a F t -adapted solution, for p [ 1 , ) ,
( P ( · ) , Q ( · ) ) L F t ( 0 , T ; S ( R n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) ;
(ii) 
The matrices N ˜ 22 are positive, and N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) are negative—we denote this condition Condition (I);
(iii) 
The feedback operator of Player 1 can be written as:
ψ 1 = N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) L F t ( Ω ; L 2 ( 0 , T ; ( R m 1 × n ) ) ) , Ψ 2 ( ψ 1 ) = N ˜ 22 1 ( B ˜ 2 ) ( N ˜ 12 ) ψ 1 L F t ( Ω ; L 2 ( 0 , T ; ( R m 2 × n ) ) ) ;
(iv) 
The value of Player 1 is:
inf γ 2 A 2 [ s , T ] sup u 1 U 1 [ s , T ] J θ ( u 1 , γ 2 ( u 1 ) ) = 1 2 E P ( s ) θ , θ .
Proof. 
“⇐” Suppose the Equation (9) admits a unique solution,
( P ( · ) , Q ( · ) ) L F t ( 0 , T ; S ( R n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) , p [ 1 , ) ,
such that Conditions (ii) and (iii) hold. Then, the function pair ( ψ 1 ( · ) , Ψ 2 ( · ) ) is a robust feedback operator. We only need to show it is an optimal one.
Let u 1 ( · ) U 1 [ s , T ] , u 2 ( · ) = γ 2 ( u 1 ( · ) ) A 2 [ s , T ] be any given admissible control for Player 1 and strategy for Player 2, respectively, and also suppose x ( · ) L F t 2 ( Ω ; C ( s , T ; R n ) ) is the corresponding state trajectory of state Equation (1). By using state Equation (1) and Equation (16), and applying the Itô formula to P x ( t ) , we obtain:
d ( P x ( t ) ) = { A P x ( t ) R x ( t ) i = 1 d [ ( C i ) P C i + Q i C i ] x ( t ) + P B 1 u 1 ( t ) + P B 2 u 2 ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) + i = 1 d [ Q i D 1 i u 1 ( t ) + Q i D 2 i u 2 ( t ) ] + B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) x ( t ) } d t + i = 1 d [ P ( C i x ( t ) + D 1 i u 1 ( t ) + D 2 i u 2 ( t ) ) + Q i x ( t ) ] d W i ( t ) .
Further, applying the Itô formula to P x ( t ) , x ( t ) , and noticing the Equation (11) for simple computing, we obtain:
d P x ( t ) , x ( t ) = { R x ( t ) , x ( t ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + i = 1 d ( D 1 i ) P D 1 i u 1 ( t ) , u 1 ( t ) + i = 1 d ( D 2 i ) P D 2 i u 2 ( t ) , u 2 ( t ) + N ˜ 12 u 2 ( t ) , u 1 ( t ) + B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) x ( t ) , x ( t ) } d t + i = 1 d [ 2 P ( C i x ( t ) + D 1 i u 1 ( t ) + D 2 i u 2 ( t ) ) , x ( t ) + Q i x ( t ) , x ( t ) ] d W i ( t ) .
From Equation (24), we have:
ψ 1 ( · ) = N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) ,
together with the symmetry of [ N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ] , which implies that:
B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) = ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) .
Since ψ 1 ( · ) L F t ( Ω ; L 2 ( 0 , T ; ( R m 1 × n ) ) and [ N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ] are bounded, we have [ B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) ] L F t ( Ω ; L 2 ( 0 , T ; ( R m 1 × n ) ) . Moreover, from Equation (24), we derive that:
B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) = ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) = ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 .
Therefore, we rewrite Equation (26) as:
d P x ( t ) , x ( t ) = { R x ( t ) , x ( t ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + i = 1 d ( D 1 i ) P D 1 i u 1 ( t ) , u 1 ( t ) + i = 1 d ( D 2 i ) P D 2 i u 2 ( t ) , u 2 ( t ) + N ˜ 12 u 2 ( t ) , u 1 ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t + i = 1 d [ 2 P ( C i x ( t ) + D 1 i u 1 ( t ) + D 2 i u 2 ( t ) ) , x ( t ) + Q i x ( t ) , x ( t ) ] d W i ( t ) .
In order to ensure that the stochastic integral above makes sense, we introduce the following stopping times τ k as, for any s [ 0 , T ) ,
τ k : = inf r s | i = 1 d s r | Q i ( t ) | 2 d t k T .
Obviously, when k , then τ k T a . s . Integrating from s to T and having an expectation on both sides of Equation (27), we have:
E P ( τ k ) x ( τ k ) , x ( τ k ) P ( s ) θ , θ = E s T X [ s , τ k ] { R x ( t ) , x ( t ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + i = 1 d ( D 1 i ) P D 1 i u 1 ( t ) , u 1 ( t ) + i = 1 d ( D 2 i ) P D 2 i u 2 ( t ) , u 2 ( t ) + N ˜ 12 u 2 ( t ) , u 1 ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t .
Clearly,
| P ( τ k ) x ( τ k ) , x ( τ k ) | | P | L F t ( 0 , T ; S ( R n ) ) | x | L F t 2 ( Ω ; C ( 0 , T ; R n ) ) 2
by virtue of the Dominated Convergence Theorem, we have:
lim k P ( τ k ) x ( τ k ) , x ( τ k ) = P ( T ) x ( T ) , x ( T ) = G x ( T ) , x ( T ) .
Similarly, since all the terms in the integral can be dominated by some bounded functions, we can prove the right hand side of Equation (28):
lim k E s T X [ s , τ k ] { R x ( t ) , x ( t ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + i = 1 d ( D 1 i ) P D 1 i u 1 ( t ) , u 1 ( t ) + i = 1 d ( D 2 i ) P D 2 i u 2 ( t ) , u 2 ( t ) + N ˜ 12 u 2 ( t ) , u 1 ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t = E s T { R x ( t ) , x ( t ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + i = 1 d ( D 1 i ) P D 1 i u 1 ( t ) , u 1 ( t ) + i = 1 d ( D 2 i ) P D 2 i u 2 ( t ) , u 2 ( t ) + N ˜ 12 u 2 ( t ) , u 1 ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t .
Then, letting k , and adding the same term:
E s T [ R x ( t ) , x ( t ) + N 1 u 1 ( t ) , u 1 ( t ) + N 2 u 2 ( t ) , u 2 ( t ) ] d t
on both sides of Equation (28), and by the Equation (11), it follows that:
2 J θ ( u 1 , u 2 ) P ( s ) θ , θ = E s T { R x ( t ) , x ( t ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + i = 1 d ( D 1 i ) P D 1 i u 1 ( t ) , u 1 ( t ) + i = 1 d ( D 2 i ) P D 2 i u 2 ( t ) , u 2 ( t ) + 2 N ˜ 12 u 2 ( t ) , u 1 ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t + E s T [ R x ( t ) , x ( t ) + N 1 u 1 ( t ) , u 1 ( t ) + N 2 u 2 ( t ) , u 2 ( t ) ] d t = E s T { 2 B ˜ 1 u 1 ( t ) , x ( t ) + 2 B ˜ 2 u 2 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + N ˜ 11 u 1 ( t ) , u 1 ( t ) + N ˜ 22 u 2 ( t ) , u 2 ( t ) + 2 N ˜ 12 u 2 ( t ) , u 1 ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t ,
from the notation γ ¯ 2 ( u 1 ) in Remark 5, using the completion of a square for u 2 , we get:
2 J θ ( u 1 , u 2 ) P ( s ) θ , θ = E s T { N ˜ 22 u 2 ( t ) , u 2 ( t ) 2 N ˜ 22 u 2 ( t ) , γ ¯ 2 ( u 1 ( t ) ) + N ˜ 22 γ ¯ 2 ( u 1 ( t ) ) , γ ¯ 2 ( u 1 ( t ) ) + 2 B ˜ 1 u 1 ( t ) , x ( t ) + B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) x ( t ) , x ( t ) + N ˜ 11 u 1 ( t ) , u 1 ( t ) N ˜ 22 γ ¯ 2 ( u 1 ( t ) ) , γ ¯ 2 ( u 1 ( t ) ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t = E s T { N ˜ 22 ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) , ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) + N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) u 1 ( t ) , u 1 ( t ) + 2 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) u 1 ( t ) , x ( t ) + ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 x ( t ) , x ( t ) } d t ,
noticing that:
B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) = ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ,
and using the completion square for u 1 , we have:
2 J θ ( u 1 , γ 2 ( u 1 ) ) = 2 J θ ( ψ 1 x ¯ , Ψ 2 ( ψ 1 ) x ¯ ) + E s T { N ˜ 22 ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) , ( u 2 γ ¯ 2 ( u 1 ( t ) ) ) + N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ( u 1 ( t ) ψ 1 x ( t ) ) , ( u 1 ( t ) ψ 1 x ( t ) ) } d t .
From Condition (ii), we obtain
J θ ( u 1 , γ ¯ 2 ( u 1 ) ) J θ ( u 1 , γ 2 ( u 1 ) ) , u 1 U 1 [ s , T ] , u 2 = γ 2 ( u 1 ) A 2 [ s , T ] ,
and
J θ ( ψ 1 x ¯ , Ψ 2 ( ψ 1 ) x ¯ ) J θ ( u 1 , γ ¯ 2 ( u 1 ) ) , u 1 U 1 [ s , T ] ,
which means that ( ψ 1 ( · ) , Ψ 2 ( · ) ) is an optimal operator for Player 1 by Definition 1.
“⇒” In this part, we prove the necessity in Theorem 1 and we divide the proof into three steps.
Step 1. We provide some existence and uniqueness results of SDEs. Suppose that ( ψ 1 ( · ) , Ψ 2 ( · ) ) is a robust optimal feedback control-strategy operator for Player 1. Then, from Lemmas 2–4, for any μ R n , there is the following forward–backward SDE:
d x ( t ) = [ A + B 1 ψ 1 + B 2 Ψ 2 ( ψ 1 ) ] x ( t ) d t + i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] x ( t ) d W i ( t ) , d y ( t ) = A y ( t ) + i = 1 d ( C i ) z i ( t ) + R x ( t ) d t + i = 1 d z i ( t ) d W i ( t ) , t [ 0 , T ] , x ( 0 ) = μ , y ( T ) = G x ( T ) ,
which admits a unique solution:
( x ( t ) , y ( t ) , z ( t ) ) L F t 2 ( Ω ; C ( 0 , T ; R n ) ) × L F t 2 ( Ω ; C ( 0 , T ; R n ) ) × L F t 2 ( 0 , T ; R n × d )
such that:
( B 1 ) y ( t ) + i = 1 d ( D 1 i ) z i ( t ) + N 1 ψ 1 x ( t ) = 0 , ( B 2 ) y ( t ) + i = 1 d ( D 2 i ) z i ( t ) + N 2 Ψ 2 ( ψ 1 ) x ( t ) = 0 .
Additionally, we introduce another SDE:
d x ˜ ( t ) = [ A B 1 ψ 1 B 2 Ψ 2 ( ψ 1 ) + i = 1 d ( C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ) 2 ] x ˜ ( t ) d t i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] x ˜ ( t ) d W i ( t ) , t [ 0 , T ] , x ( 0 ) = μ .
From Lemma 3, Equation (33) admits a unique solution x ˜ ( t ) L F t 2 ( Ω ; C ( 0 , T ; R n ) ) .
Moreover, consider the following R n × n -valued SDEs:
d X ( t ) = [ A + B 1 ψ 1 + B 2 Ψ 2 ( ψ 1 ) ] X ( t ) d t + i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] X ( t ) d W i ( t ) , d Y ( t ) = A Y ( t ) + i = 1 d ( C i ) Z i ( t ) + R X ( t ) d t + i = 1 d Z i ( t ) d W i ( t ) , t [ 0 , T ] , X ( 0 ) = I n , Y ( T ) = G X ( T ) ,
and,
d X ˜ ( t ) = [ A B 1 ψ 1 B 2 Ψ 2 ( ψ 1 ) + i = 1 d ( C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ) 2 ] X ˜ ( t ) d t i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] X ˜ ( t ) d W i ( t ) , t [ 0 , T ] , X ˜ ( 0 ) = μ .
Equations (34) and (35) above also admit unique solutions X ˜ ( t ) L F t 2 ( Ω ; C ( 0 , T ; R n × n ) ) and ( X ( t ) , Y ( t ) , Z ( t ) ) L F t 2 ( Ω ; C ( 0 , T ; R n × n ) ) × L F t 2 ( Ω ; C ( 0 , T ; R n × n ) ) × L F t 2 ( 0 , T ; ( R n × n ) d ) . Obviously, from Equations (32) and (37), we obtain:
( B 1 ) Y ( t ) + i = 1 d ( D 1 i ) Z i ( t ) + N 1 ψ 1 X ( t ) = 0 , ( B 2 ) Y ( t ) + i = 1 d ( D 2 i ) Z i ( t ) + N 2 Ψ 2 ( ψ 1 ) X ( t ) = 0 .
Also from Equations (31) to (35), for any μ R n , it easily follows that:
x ( t ; μ ) = X ( t ) μ , y ( t ; μ ) = Y ( t ) μ , z ( t ; μ ) = Z ( t ) μ , x ˜ ( t ; μ ) = X ˜ ( t ) μ , a . e . t [ 0 , T ] ,
where Z ( · ) = ( Z 1 ( · ) , , Z d ( · ) ) and z ( · ) = ( z 1 ( · ) , , z d ( · ) ) . Further, for any μ , ν R n and t [ 0 , T ] , applying the Itô formula to x ( t ; μ ) , x ˜ ( t ; ν ) , we have:
x ( t ; μ ) , x ˜ ( t ; ν ) μ , ν = 0 t [ A + B 1 ψ 1 + B 2 Ψ 2 ( ψ 1 ) ] x ( r ; μ ) , x ˜ ( r ; ν ) d r + 0 t i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] x ( r ; μ ) , x ˜ ( r ; ν ) d W i ( r ) + 0 t x ( r ; μ ) , [ A B 1 ψ 1 B 2 Ψ 2 ( ψ 1 ) + i = 1 d ( C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ) 2 ] x ˜ ( r ; ν ) d r 0 t x ( r ; μ ) , i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] x ˜ ( r ; ν ) d W i ( r ) 0 t i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] 2 x ( r ; μ ) , x ˜ ( r ; ν ) d r = 0 .
Therefore, from Equation (37), it follows that:
x ( t ; μ ) , x ˜ ( t ; ν ) = X ( t ) μ , X ˜ ( t ) ν = μ , ν , a . s . ,
which implies that X ( t ) [ X ˜ ( t ) ] = I n , i.e., [ X ˜ ( t ) ] = X ( t ) 1 .
Step 2. We try to prove that Equation (9) admits a solution through the idea of making the connection between the solvable SDG and the solvability of the Riccati equations. We assume that:
P ( t ) : = Y ( t ) [ X ˜ ( t ) ] , Υ i ( t ) = Z i ( t ) [ X ˜ ( t ) ] .
Applying the Itô formula to P ( t ) , we have:
d P = { A Y + i = 1 d ( C i ) Z i + R X X 1 i = 1 d Z i X 1 [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] + Y X 1 A B 1 ψ 1 B 2 Ψ 2 ( ψ 1 ) + i = 1 d ( C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ) 2 } d t + i = 1 d Z i X 1 i = 1 d Y X 1 [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] d W i ( t ) = { A P i = 1 d ( C i ) Υ i R i = 1 d Υ i [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] + P A B 1 ψ 1 B 2 Ψ 2 ( ψ 1 ) + i = 1 d ( C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ) 2 } d t + i = 1 d Υ i P [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] d W i ( t ) .
Put Q i : = Υ i P [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] , then ( P ( · ) , Q ( · ) ) solves the following SDE:
d P = { A P + P A + i = 1 d ( C i ) P C i + i = 1 d [ ( C i ) Q i + Q i C i ] + B ˜ 1 ψ 1 + B ˜ 2 Ψ 2 ( ψ 1 ) + R } d t + i = 1 d Q i d W i P ( T ) = G . t [ 0 , T ] .
From Lemma 3, we deduce that ( P , Q ) L F t ( Ω ; C ( 0 , T ; R n × n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; ( R n × n ) d ) ) .
Next, we show that ( P , Q ) are symmetric. We introduce the following forward–backward SDEs, for any t [ 0 , T ) and α L F t 2 ( Ω ; R n ) ,
d x t ( r ) = [ A + B 1 ψ 1 + B 2 Ψ 2 ( ψ 1 ) ] x t ( r ) d r + i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] x t ( r ) d W i ( r ) , d y t ( r ) = A y t ( r ) + i = 1 d ( C i ) z i t ( r ) + R x t ( r ) d r + i = 1 d z i t ( r ) d W i ( r ) , r [ t , T ] , x t ( t ) = α , y t ( T ) = G x t ( T ) ,
and,
d X t ( r ) = [ A + B 1 ψ 1 + B 2 Ψ 2 ( ψ 1 ) ] X t d r + i = 1 d [ C i + D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] X t ( r ) d W i ( r ) , d Y t ( r ) = A Y t ( r ) + i = 1 d ( C i ) Z i t ( r ) + R X t ( r ) d r + i = 1 d Z i t ( r ) d W i ( r ) , r [ t , T ] , X t ( t ) = I n , Y t ( T ) = G X t ( T ) .
Clearly, the above two SDEs admit unique solutions:
( x t , y t , z t ) L F t 2 ( Ω ; C ( t , T ; R n ) ) × L F t 2 ( Ω ; C ( t , T ; R n ) ) × L F t 2 ( t , T ; R n × d )
and,
( X t , Y t , Z t ) L F t 2 ( Ω ; C ( t , T ; R n × n ) ) × L F t 2 ( Ω ; C ( t , T ; R n × n ) ) × L F t 2 ( t , T ; ( R n × n ) d ) .
Obviously, from Equations (40) and (41), we have:
x t ( r ; α ) = X t ( r ) α , y t ( r ; α ) = Y t ( r ) α , z t ( r ; α ) = Z t ( r ) α , a . e . r [ t , T ] .
Due to the uniqueness of the solution to Equation (31), for any μ R n and t [ 0 , T ] , it follows that:
x t ( r ; X ( t ) μ ) = X t ( r ) X ( t ) μ = x ( r ; μ ) , a . s .
therefore,
y t ( t ; X ( t ) μ ) = Y t ( t ) X ( t ) μ = Y ( t ) μ , a . s .
which implies that:
Y t ( t ) = Y ( t ) [ X ˜ ( t ) ] = P ( t ) . a . s .
Further, let α , β L F t 2 ( Ω ; R n ) and apply the Itô formula to x t ( r ; α ) , y t ( r ; β ) , noticing x t ( r ; α ) = X t ( r ) α , y t ( r ; β ) = Y t ( r ) β ; then, we have:
E α , P ( t ) β = E t T { X t ( r ) α , R X t ( r ) β [ B 1 ψ 1 + B 2 Ψ 2 ( ψ 1 ) ] X t ( r ) α , Y t ( r ) β i = 1 d [ D 1 i ψ 1 + D 2 i Ψ 2 ( ψ 1 ) ] X t ( r ) α , Z i t ( r ) β } d r + E X t ( T ) α , G X t ( T ) β .
Noticing that Equation (36), we obtain:
E P ( t ) β , α = E t T { R X t ( r ) β , X t ( r ) α + N 1 ψ 1 X t ( r ) β , ψ 1 X t ( r ) α + N 2 Ψ 2 ( ψ 1 ) X t ( r ) β , Ψ 2 ( ψ 1 ) X t ( r ) α } d r + E G X t ( T ) β , X t ( T ) α = E [ X t ( r ) ] G X t ( T ) β + t T { [ X t ( r ) ] R X t ( r ) β + [ X t ( r ) ] [ ψ 1 ] N 1 ψ 1 X t ( r ) β + [ X t ( r ) ] [ Ψ 2 ( ψ 1 ) ] N 2 Ψ 2 ( ψ 1 ) X t ( r ) β } d r , α ,
which means:
P ( t ) = E { [ X t ( r ) ] G X t ( T ) + t T { [ X t ( r ) ] R X t ( r ) + [ X t ( r ) ] [ ψ 1 ] N 1 ψ 1 X t ( r ) + [ X t ( r ) ] [ Ψ 2 ( ψ 1 ) ] N 2 Ψ 2 ( ψ 1 ) X t ( r ) } d r | F t } ;
therefore, P ( t ) is symmetric, since G , R ( · ) , N 1 ( · ) , N 2 ( · ) are symmetric in (AS2). On the other hand, ( P , Q ) satisfies that:
d P = { A P + P A + i = 1 d ( C i ) P C i + i = 1 d [ ( C i ) ( Q i ) + ( Q i ) C i ] + ( ψ 1 ) ( B ˜ 1 ) + [ Ψ 2 ( ψ 1 ) ] ( B ˜ 2 ) + R } d t + i = 1 d ( Q i ) d W i ( t ) , [ P ( T ) ] = G . t [ 0 , T ]
Let Equation (39) minus Equation (43) and integrate from 0 to t; noticing that P ( t ) is symmetric, we obtain the diffusion term:
i = 1 d 0 t [ Q i ( Q i ) ] d W i ( r ) = 0
and then we conclude that Q ( t ) is symmetric due to the uniqueness of the decomposition of semi-martingale.
Next we show that ( P , Q ) is the solution of Equation (9) in which F is in the Equation (16) and the matrices N ˜ 22 and N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) are invertible. Actually, multiply X 1 on both sides of Equation (36); it follows that:
( B 1 ) P + i = 1 d ( D 1 i ) Υ i + N 1 ψ 1 = 0 , ( B 2 ) P + i = 1 d ( D 2 i ) Υ i + N 2 Ψ 2 ( ψ 1 ) = 0 .
By the definition of Q and by Equations (11) and (44), we obtain:
( B ˜ 1 ) + N ˜ 12 Ψ 2 ( ψ 1 ) + N ˜ 11 ψ 1 = 0 , ( B ˜ 2 ) + ( N ˜ 12 ) ψ 1 + N ˜ 22 Ψ 2 ( ψ 1 ) = 0 ,
from which we have:
ψ 1 = N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) Ψ 2 ( ψ 1 ) = N ˜ 22 1 ( B ˜ 2 ) ( N ˜ 12 ) ψ 1 .
Putting these two equalities into Equation (39), for simplicity of computing, we have:
B ˜ 1 ψ 1 + B ˜ 2 Ψ 2 ( ψ 1 ) = B ˜ 2 N ˜ 22 1 ( B ˜ 2 ) B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) ,
which implies that ( P , Q ) L F t ( Ω ; C ( 0 , T ; S ( R n ) ) ) × L F t p ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) solves the Equation (9) and ends the proof of Condition (i).
Step 3. In this part, we show that Conditions (ii)–(iv) hold. Obviously, from Equation (46) and ( ψ 1 ( · ) , Ψ 2 ( · ) ) being robust, it easily follows that Condition (iii) holds. We only need to prove Conditions (ii) and (iv). Actually, from Equation (46), it follows that:
B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) 1 B ˜ 1 B ˜ 2 N ˜ 22 1 ( N ˜ 12 ) = ( ψ 1 ) N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ψ 1 ,
and repeating the procedures to derive (30), we have:
2 J θ ( u 1 , γ 2 ( u 1 ) ) = P ( s ) θ , θ + E s T { N ˜ 22 ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) , ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) + N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ( u 1 ( t ) ψ 1 x ( t ) ) , ( u 1 ( t ) ψ 1 x ( t ) ) } d t . = 2 J θ ( ψ 1 x ¯ , Ψ 2 ( ψ 1 ) x ¯ ) + E s T { N ˜ 22 ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) , ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) + N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ( u 1 ( t ) ψ 1 x ( t ) ) , ( u 1 ( t ) ψ 1 x ( t ) ) } d t .
Hence, from the optimality of feedback control-strategy operator ( ψ 1 ( · ) , Ψ 2 ( · ) ) , we obtain that Condition (iv) holds, and from Definition 1 we have:
E s T N ˜ 22 ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) , ( u 2 ( t ) γ ¯ 2 ( u 1 ( t ) ) ) d t 0 , u 2 ( t ) = γ 2 ( u 1 ( t ) ) A 2 [ s , T ] , E s T N ˜ 11 N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) ( u 1 ( t ) ψ 1 x ( t ) ) , ( u 1 ( t ) ψ 1 x ( t ) ) d t 0 , u 1 ( t ) U 1 [ s , T ] ,
from which we obtain that Condition (ii) holds and we finish the proof of the necessity. □
A similar statement of Player 2 is made in the following sense.
Theorem 2.
Suppose that the Assumptions (AS1)–(AS2) hold. Then, Problem (SDG) admits a robust optimal feedback control-strategy operator ( ψ 2 ( · ) , Ψ 1 ( · ) ) for Player 2 if and only if:
(i) 
The Equation (9) admits a F t -adapted solution for p [ 1 , ) ,
( P ( · ) , Q ( · ) ) L F t ( 0 , T ; S ( R n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) ;
(ii) 
The matrices N ˜ 11 are negative, N ˜ 22 ( N ˜ 12 ) N ˜ 11 1 N ˜ 12 is positive, we denote this condition by Condition (II).
(iii) 
The feedback operator of Player 2 can be written as:
Ψ 1 ( ψ 2 ) = N ˜ 11 1 ( B ˜ 1 ) N ˜ 12 ψ 2 L F t ( Ω ; L 2 ( 0 , T ; ( R m 1 × n ) ) ) . ψ 2 = N ˜ 22 ( N ˜ 12 ) N ˜ 11 1 N ˜ 12 1 B ˜ 2 B ˜ 1 N ˜ 11 1 N ˜ 12 L F t ( Ω ; L 2 ( 0 , T ; ( R m 2 × n ) ) ) .
(iv) 
The value of Player 2 is:
sup γ 1 A 1 [ s , T ] inf u 2 U 2 [ s , T ] J θ ( γ 1 ( u 2 ) , u 2 ) = 1 2 E P ( s ) θ , θ .
Proof. 
This theorem can be proved with the same approach in Theorem 1, so we omit it here. □
Corollary 1.
Suppose that the Assumptions (AS1)–(AS2) hold. Then, Problem (SDG) admits robust optimal feedback control-strategy operators ( ψ 1 ( · ) , Ψ 2 ( · ) ) and ( ψ 2 ( · ) , Ψ 1 ( · ) ) with:
ψ 1 ( · ) = Ψ 1 ( ψ 2 ( · ) ) , ψ 2 ( · ) = Ψ 2 ( ψ 1 ( · ) ) .
if and only if:
(i) 
The Equation (9) admits a unique F t -adapted solution ( P ( · ) , Q ( · ) ) L F t ( 0 , T ; S ( R n ) ) × L F t p ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) (where p [ 1 , ) );
(ii) 
The matrices N ˜ 11 is negative, N ˜ 22 is positive, we denote it by Condition (III);
(iii) 
ψ 1 ( · ) , Ψ 2 ( ψ 1 ( · ) ) , Ψ 1 ( ψ 2 ( · ) ) , ψ 2 ( · ) satisfy the condition given by Equations (24) and (47);
(iv) 
Further, the value of Problem (SDG) is:
inf γ 2 A 2 [ s , T ] sup u 1 U 1 [ s , T ] J θ ( u 1 , γ 2 ( u 1 ) ) = sup γ 1 A 1 [ s , T ] inf u 2 U 2 [ s , T ] J θ ( γ 1 ( u 2 ) , u 2 ) = 1 2 E P ( s ) θ , θ .
Proof. 
“⇒” If (i)–(iv) hold, we can obtain that robust optimal feedback control-strategy operators ( ψ 1 ( · ) , Ψ 2 ( · ) ) and ( ψ 2 ( · ) , Ψ 1 ( · ) ) exist from Theorems 1 and 2, because Condition (III) in this Corollary holds if and only if Condition (I) and Condition (II) hold, which are the conditions (ii) in Theorems 1 and 2, respectively. Furthermore, from Condition (III) in this Corollary,
( u ¯ 1 , γ ¯ 2 ( u ¯ 1 ) ) = ( ψ 1 x ¯ , Ψ 2 ( ψ 1 ) x ¯ ) , ( γ ¯ 1 ( u ¯ 2 ) , u ¯ 2 ) = ( Ψ 1 ( ψ 2 ) x ¯ , ψ 2 x ¯ )
are the unique solution of the algebra Equation (13), which leads to the Equation (49).
“⇐” If Problem (SDG) admits robust optimal feedback control-strategy operators ( ψ 1 ( · ) , Ψ 2 ( · ) ) and ( ψ 2 ( · ) , Ψ 1 ( · ) ) with:
ψ 1 ( · ) = Ψ 1 ( ψ 2 ( · ) ) , ψ 2 ( · ) = Ψ 2 ( ψ 1 ( · ) ) .
We can easily prove the results (i)–(iv) from Theorems 1 and 2, which ends the proof. □
Remark 7.
Equation (49) can be replaced by the following matrices form if the matrices N ˜ 11 and N ˜ 22 are invertible:
( N ˜ 12 ) 1 N ˜ 22 N ˜ 12 1 ( B ˜ 1 ) = N ˜ 11 1 I N ˜ 12 N ˜ 22 1 ( N ˜ 12 ) N ˜ 11 1 ( B ˜ 1 ) , N ˜ 12 1 N ˜ 11 ( N ˜ 12 ) 1 ( B ˜ 2 ) = N ˜ 22 1 I ( N ˜ 12 ) N ˜ 11 1 N ˜ 12 N ˜ 22 1 ( B ˜ 2 ) .
These two conditions can be obtained by using Equations (24), (47) and (49) with the observation that N ˜ 11 and N ˜ 22 are symmetric matrices.
Remark 8.
Actually, the solution ( P , Q ) of the Equation (9) is also unique, because the Lemma 3 ensures the uniqueness of the solution for the Equation (39), which is another form of Equation (9).
Remark 9.
In the Corollary 1, if the Equation (9) admits a solution ( P , Q ) with the feedback operator having the sharp regularity, i.e., Equation (24) and (47) hold, then Problem (SDG) admits robust optimal feedback control-strategy operators ( ψ 1 ( · ) , Ψ 2 ( · ) ) and ( ψ 2 ( · ) , Ψ 1 ( · ) ) , and we can obtain that Problem (SDG) is closed-loop solvable and the saddle point exists in the following sense:
( u ¯ 1 , u ¯ 2 ) = ( Ψ 1 ( ψ 2 ) x ¯ , Ψ 2 ( ψ 1 ) x ¯ )
and,
J θ ( u ¯ 1 , u ¯ 2 ) = inf γ 2 A 2 [ s , T ] sup u 1 U 1 [ s , T ] J θ ( u 1 , γ 2 ( u 1 ) ) = sup γ 1 A 1 [ s , T ] inf u 2 U 2 [ s , T ] J θ ( γ 1 ( u 2 ) , u 2 ) = 1 2 E P ( s ) θ , θ = J θ ( Ψ 1 ( ψ 2 ) x ¯ , Ψ 2 ( ψ 1 ) x ¯ ) .
However, if the solution ( P , Q ) does not have the needed regularity, Problem (SDG) may fail to obtain a robust optimal feedback control-strategy; this phenomenon is different from the deterministic case.

4. Examples

In this section, we provide two examples to illustrate our theoretical results. Example 1 shows that the solvability of the backward stochastic Riccati equation is not sufficient enough to guarantee that Problem (SDG) admits a robust optimal feedback control-strategy. Example 2 shows that there does exist a feedback control-strategy pair under some suitable conditions. These two examples are inspired by [16,22,26,32].
Example 1.
Firstly, we introduce the following stochastic processes Γ ( · ) , σ ( · ) and stopping time τ 1 :
Γ ( t ) : = 0 t 1 T r d W ( r ) , t [ 0 , T ) , σ ( t ) : = π 2 2 T t χ [ 0 , τ 1 ] ( t ) , t [ 0 , T ) , τ 1 : = inf { t [ 0 , T ) , | Γ ( t ) | > 1 } T .
From [22,32], we have:
| 0 T σ ( r ) d W ( r ) | = π 2 2 | 0 τ 1 1 T r d W ( r ) | = π 2 2 | Γ ( τ 1 ) | π 2 2 ,
E exp 0 T | σ ( r ) | 2 d r = .
In this example, we consider Problem (SDG) with the following form:
d x ( r ) = [ u 1 ( r ) + u 2 ( r ) ] d W ( r ) , r [ 0 , T ] x ( 0 ) = θ ,
and the objective functional:
J θ ( u 1 , u 2 ) = 1 2 κ 1 E x ( T ) 2 + 1 2 E 0 T u 2 ( r ) 2 d r ,
for any θ L F 0 2 ( Ω ; R ) , where m = n = d = 1 . Obviously, we have the following data:
A = 0 , B 1 = 0 , B 2 = 0 , C = 0 , D 1 = 1 , D 2 = 1 , R = 0 , N 1 = 0 , N 2 = 1 , G = κ 1 ,
where κ = φ ( T ) and ( φ ( t ) , ϕ ( t ) ) satisfy the following backward stochastic differential equation:
φ ( t ) = 0 T σ ( r ) d W ( r ) + π 2 2 + 1 t T ϕ ( r ) d W ( r ) , t [ 0 , T ] .
It admits a unique solution ( φ ( t ) , ϕ ( t ) ) in the following sense:
φ ( t ) = 0 t σ ( r ) d W ( r ) + π 2 2 + 1 , ϕ ( t ) = σ ( t ) , t [ 0 , T ] .
By Equations (53), (54), and (55) above, it has:
1 φ ( t ) π 2 + 1 , ϕ ( t ) L F t ( Ω ; L 2 ( 0 , T ; R ) ) .
Then, from Equations (11) and Data (56) above, for simple computing, it follows that:
N ˜ 11 = P , N ˜ 22 = P + 1 , N ˜ 12 = P , B ˜ 1 = Q , B ˜ 2 = Q ,
and the Equation (9), in which F is in the Form (I) in (16), is given by:
d P ( r ) = P ( r ) 1 Q ( r ) 2 d r + Q ( r ) d W ( r ) , r [ 0 , T ] P ( T ) = κ 1 .
Applying the Itô formula to φ ( · ) 1 , we deduce that ( P ( · ) , Q ( · ) ) = ( φ ( · ) 1 , φ ( · ) 2 ϕ ( · ) ) is the unique solution of Equation (58). Furthermore, from Theorem 1 and Equation (57), we obtain:
ψ 1 = φ ( · ) 1 ϕ ( · ) L F t ( Ω ; L 2 ( 0 , T ; R ) ) , Ψ 2 ( ψ 1 ) = 2 φ ( · ) 2 ϕ ( · ) φ ( · ) 1 + 1 L F t ( Ω ; L 2 ( 0 , T ; R ) ) .
Therefore, there is no robust feedback control-strategy operator for Player 1, that is, Problem (SDG) is not closed-loop solvable.
Example 2.
In this example, we shall apply the Itô formula to cos W ( t ) , which shows that:
cos W ( T ) cos W ( t ) = t T sin W ( s ) d W ( s ) 1 2 t T cos W ( s ) d s .
From the above equality, we suppose that:
δ = 3 2 + T 2 + cos W ( T ) + 1 2 0 T cos W ( s ) d s , γ ( t ) = 3 2 + T 2 + cos W ( t ) + 1 2 0 t cos W ( s ) d s . Υ ( t ) = sin W ( t ) , t [ 0 , T ] .
From Equation (61), we have:
1 2 γ ( · ) 5 2 + T , 1 2 δ 5 2 + T .
And ( γ ( · ) , Υ ( · ) ) satisfies:
γ ( t ) = δ t T Υ ( s ) d W ( s ) , t [ 0 , T ] .
In this example, we suppose the following data:
m = n = d = 1 , A = 0 , B 1 = 0 , B 2 = 0 , C = 0 , D 1 = 1 , D 2 = 1 , R = 0 , N 1 = 0 , N 2 = 1 , G = δ 1 .
Then, from Equations (11) and (63) above, for simple computing, it follows that:
N ˜ 11 = P , N ˜ 22 = P + 1 , N ˜ 12 = P , B ˜ 1 = Q , B ˜ 2 = Q ,
and the Equation (9), in which F is in the Form (I) in (16), is given by:
d P ( r ) = P ( r ) 1 Q ( r ) 2 d r + Q ( r ) d W ( r ) , r [ 0 , T ] P ( T ) = δ 1 .
Applying the Itô formula to γ ( · ) 1 , we deduce that ( P ( · ) , Q ( · ) ) = ( γ ( · ) 1 , γ ( · ) 2 Υ ( · ) ) is the unique solution of Equation (64). Furthermore, from Theorem 1, we obtain that:
ψ 1 = γ ( · ) 1 Υ ( · ) L F t ( Ω ; L 2 ( 0 , T ; R ) ) , Ψ 2 ( ψ 1 ) = 2 γ ( · ) 2 Υ ( · ) γ ( · ) 1 + 1 L F t ( Ω ; L 2 ( 0 , T ; R ) )
is a robust optimal feedback control-strategy operator. For clear presentation, we shall provide the following graphs with T = 10 . Figure 1 shows one controlled sample path of the state process under the optimal control-strategy ( u ¯ 1 , u ¯ 2 ) . Figure 2 shows the optimal control-strategy ( u ¯ 1 , u ¯ 2 ) , which is the solution of the game.
Remark 10.
From the above two examples, although the form of the the Equation (9) is the same, it remarkably shows that, with different endpoints, the robustness of the feedback control-strategy operator will be different. Generally, it is interesting to find some suitable conditions to guarantee that Equation (9) admits a unique solution:
( P ( · ) , Q ( · ) ) L F t ( 0 , T ; S ( R n ) ) × L F t ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) ,
which is unsolved.

5. Concluding Remarks

In this paper, we consider a closed-loop two-person zero-sum LQ stochastic differential game with random coefficients. As we know, the feedback control-strategy contains the robustness when facing small perturbations; however, our results show that the same remarkable phenomenon appears in Problem (SDG) as appeared in the work of Lü-Wang-Zhang, that the regularity of the solutions for the corresponding backward stochastic Riccati equations is not enough to establish the closed-loop control-strategies; this phenomenon has not previously been mentioned in the literature. On the other hand, under some suitable conditions, we have established the equivalence between the existence of the robust optimal feedback control-strategy operators and the solvability of the corresponding backward stochastic Riccati equations in SDG, which shows the importance of the regularity of the solutions for the stochastic Riccati equations when trying to solve the closed-loop SDG. Additionally, the saddle points in the closed-loop case have been obtained by some sharp regularity assumptions on the control-strategy operators. Finally, from the examples, Problem (SDG) is not closed-loop solvable without the sharp regularity of the solution for the corresponding backward stochastic Riccati equations, which illustrate our theoretical results.
As with the generalization, there are several interesting problems that deserve further investigation. One is the solvability of the Equation (9), how to obtain a unique solution ( P ( · ) , Q ( · ) ) L F t ( 0 , T ; S ( R n ) ) × L F t ( Ω ; L 2 ( 0 , T ; ( S ( R n ) ) d ) ) so that the feedback operator has the sharp regularity to guarantee the robustness, which is still open. Another one is to extend the results in this paper to other LQ games, such as the mean-field LQ game, the indefinite LQ game, and so on.

Author Contributions

Conceptualization, C.T.; Writing—original draft, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not Applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Isaacs, R. Differential Games; John Wiley and Sons: New York, NY, USA, 1965. [Google Scholar]
  2. Berkovitz, L.D. Lectures on Differential Gams, Differential Games and Related Topics; Kuhn, H.W., Szego, G.P., Eds.; North-Holland: Amsterdam, The Netherlands, 1971; pp. 3–45. [Google Scholar]
  3. Bernhard, P. Linear-quadratic, two-person, zero-sum differential games: Necessary and sufficient conditions. J. Optim. Theory Appl. 1979, 27, 51–69. [Google Scholar] [CrossRef]
  4. Ellot, R.J.; Kalton, N.J. Existence of value in differential games. Mem. Am. Math. Soc. 1972, 126, 1–23. [Google Scholar] [CrossRef]
  5. Erans, L.C.; Souganidis, P.E. Differential games and representation formulas for solutions of Hamilton-Jacobi-Isaacs equations. Indiana Univ. Math. J. 1984, 33, 773–797. [Google Scholar]
  6. Fleming, W.H.; Souganidis, P.E. On the existence of value functions of two player, zero-sum stochastic differential games. Indiana Univ. Math. J. 1989, 38, 293–314. [Google Scholar] [CrossRef]
  7. Ho, Y.C.; Bryson, A.E.; Baron, S. Differential games and optimal pursuit-evasion strategies. IEEE Trans. AC 1965, 10, 385–389. [Google Scholar] [CrossRef]
  8. Hamadè, S.; Lepeltier, J.P. Zero-sum stochastic differential games and backward equations. Syst. Control Lett. 1995, 24, 259–263. [Google Scholar] [CrossRef]
  9. Mou, L.; Yong, J. Two-person zero-sum linear quadratic stochastic differential games by a Hilbert space method. J. Ind. Manag. Optim. 2006, 2, 95–117. [Google Scholar] [CrossRef]
  10. Yong, J. A leader-follower stochastic linear quadratic differential game. SIAM J. Control Optim. 2002, 41, 1015–1041. [Google Scholar] [CrossRef]
  11. Buckdahn, R.; Li, J. Stochastic differential games and viscosity solutions of Hamilton-Jacobi-Bellman-Isaacs equations. SIAM J. Control Optim. 2008, 47, 444–475. [Google Scholar] [CrossRef]
  12. Yu, Z.Y. An optimal feedback control-strategy pair for zero-sum linear-quadratic stochastic differential game: The Riccati equation approach. SIAM J. Control Optim. 2015, 53, 2141–2167. [Google Scholar] [CrossRef]
  13. Moon, J. Linear quadratic stochastic stackelberg differential games for jump-diffusion systems. SIAM J. Control Optim. 2021, 59, 954–976. [Google Scholar] [CrossRef]
  14. Wang, H.Y.; Wu, Z. Time-inconsistent linear-quadratic non-zero sum stochastic differential games with random jumps. Int. J. Control. 2022, 95, 1864–1874. [Google Scholar] [CrossRef]
  15. Moon, J.; Kim, W. Explicit characterization of feedback Nash equilibria for indefinite, linear-Quadratic, mean-field-type stochastic zero-sum differential games with jump-diffusion models. Mathematics 2020, 8, 1669. [Google Scholar] [CrossRef]
  16. Sun, J.R.; Yong, J. Linear quadratic stochastic differential games: Open-loop and closed-loop saddle points. SIAM J. Control Optim. 2014, 52, 4082–4121. [Google Scholar] [CrossRef]
  17. Sun, J.R.; Yong, J. Linear quadratic stochastic two-person nonzero-sum differential games: Open-loop and closed-loop Nash equilibria. Stoch. Process. Their Appl. 2014, 129, 381–418. [Google Scholar] [CrossRef]
  18. Yu, Z.Y.; Zhang, B.K.; Zhang, F. One kind of linear-quadratic zero-sum stochastic differential game with jumps. Int. J. Control. 2022, 95, 1470–1481. [Google Scholar] [CrossRef]
  19. Tang, S. General linear quadratic optimal stochastic control problems with random coefficients: Linear stochastic Hamilton systems and backward stochastic Riccati equations. SIAM J. Control Optim. 2003, 42, 53–75. [Google Scholar] [CrossRef]
  20. Bismut, J.M. Contrôle des Systèmes Linéaires Quadratiques: Applications de L’IntÉGrale Stochastique; Lecture Notes in Mathematics; Séminaire de Probabilités XII, Université de Strasbourg 1976/77, 180–264; Springer: Berlin/Heidelberg, Germany, 1978; p. 649. [Google Scholar]
  21. Tang, S. Dynamic programming for general linear quadratic optimal stochastic control with random coefficients. SIAM J. Control Optim. 2015, 53, 1082–1106. [Google Scholar] [CrossRef]
  22. Lü, Q.; Wang, T.; Zhang, X. Characterization of optimal feedback for stochastic linear quadratic control problem. Probab. Uncertain. Quant. Risk 2017, 2, 11. [Google Scholar] [CrossRef]
  23. Zhang, F.; Dong, Y.C.; Meng, Q.X. Backward stochastic Riccati equation with jumps associated with stochastic linear quadratic optimal control with jumps and random coefficients. SIAM J. Control Optim. 2020, 58, 393–424. [Google Scholar] [CrossRef]
  24. Ali, I.; KhanA, S.U. Dynamic Competition Analysis of Stochastic Fractional Differential Equation Arising in Finance via Pseudospectral Method. Mathematics 2023, 11, 1328. [Google Scholar] [CrossRef]
  25. Butt, A.I.K.; Ahmad, W.; Rafiq, M.; Ahmad, N.; Imran, M. Computationally efficient optimal control analysis for the mathematical model of Coronavirus pandemic. Expert Syst. Appl. 2023, 234, 121094. [Google Scholar] [CrossRef]
  26. Sun, J.R.; Li, X.; Yong, J. Open-loop and closed-loop solvabilities for stochastic linear quadratic optimal control problem. SIAM J. Control Optim. 2016, 54, 2274–2308. [Google Scholar] [CrossRef]
  27. Yong, J.; Zhou, X.Y. Stochastic controls: Hamiltonian systems and HJB equations. In Applied Mathematics New York; Springer: New York, NY, USA, 1999; p. 43. [Google Scholar]
  28. Protter, P.E. Stochastic Integration and Differential Equations. In Stochastic Modelling and Applied Probability; Springer: Berlin/Heidelberg, Germany, 2005; p. 21. [Google Scholar]
  29. Brian, P.; Delyon, B.; Hu, Y.; Pardoux, E.; Stoica, L. Lp solutions of backward stochastic differential equations. Stochastic Process. Appl. 2003, 108, 109–129. [Google Scholar] [CrossRef]
  30. Wang, G.C.; Yu, Z.Y. A Pontryagin’s Maximum principle for non-Zero sum differential games of BSDEs with applications. IEEE Trans. AC 2010, 55, 1742–1747. [Google Scholar] [CrossRef]
  31. Bismut, J.M. Linear quadratic optimal stochastic control with random coefficients. SIAM J. Control Optim. 1976, 14, 419–444. [Google Scholar] [CrossRef]
  32. Frei, C.; Reis, G.d. A financial market with interacting investors: Does an equilibrium exist? Math. Finan. Econ. 2011, 4, 161–182. [Google Scholar] [CrossRef]
Figure 1. One controlled sample path of the state process under the optimal control-strategy ( u ¯ 1 , u ¯ 2 ) .
Figure 1. One controlled sample path of the state process under the optimal control-strategy ( u ¯ 1 , u ¯ 2 ) .
Symmetry 15 01726 g001
Figure 2. The trajectory of optimal control-strategy ( u ¯ 1 , u ¯ 2 ) .
Figure 2. The trajectory of optimal control-strategy ( u ¯ 1 , u ¯ 2 ) .
Symmetry 15 01726 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, C.; Liu, J. The Equivalence Conditions of Optimal Feedback Control-Strategy Operators for Zero-Sum Linear Quadratic Stochastic Differential Game with Random Coefficients. Symmetry 2023, 15, 1726. https://doi.org/10.3390/sym15091726

AMA Style

Tang C, Liu J. The Equivalence Conditions of Optimal Feedback Control-Strategy Operators for Zero-Sum Linear Quadratic Stochastic Differential Game with Random Coefficients. Symmetry. 2023; 15(9):1726. https://doi.org/10.3390/sym15091726

Chicago/Turabian Style

Tang, Chao, and Jinxing Liu. 2023. "The Equivalence Conditions of Optimal Feedback Control-Strategy Operators for Zero-Sum Linear Quadratic Stochastic Differential Game with Random Coefficients" Symmetry 15, no. 9: 1726. https://doi.org/10.3390/sym15091726

APA Style

Tang, C., & Liu, J. (2023). The Equivalence Conditions of Optimal Feedback Control-Strategy Operators for Zero-Sum Linear Quadratic Stochastic Differential Game with Random Coefficients. Symmetry, 15(9), 1726. https://doi.org/10.3390/sym15091726

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop