Next Article in Journal
Extended Divergence on a Foliation by Deformed Probability Simplexes
Next Article in Special Issue
Forward-Backward Sweep Method for the System of HJB-FP Equations in Memory-Limited Partially Observable Stochastic Control
Previous Article in Journal
Musical Emotions Recognition Using Entropy Features and Channel Optimization Based on EEG
Previous Article in Special Issue
Memory-Limited Partially Observable Stochastic Control and Its Mean-Field Control Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

H State-Feedback Control of Multi-Agent Systems with Data Packet Dropout in the Communication Channels: A Markovian Approach

by
Adrian-Mihail Stoica
*,† and
Serena Cristiana Stoicu
Faculty of Aerospace Engineering, University Politehnica of Bucharest, 011061 Bucharest, Romania
*
Author to whom correspondence should be addressed.
Current address: Str. Polizu, No. 1, 011061 Bucharest, Romania.
Entropy 2022, 24(12), 1734; https://doi.org/10.3390/e24121734
Submission received: 19 October 2022 / Revised: 15 November 2022 / Accepted: 25 November 2022 / Published: 28 November 2022
(This article belongs to the Special Issue Information Theory in Control Systems)

Abstract

:
The paper presents an H type control procedure for multi-agent systems taking into account possible data dropout in the communication network. The data dropout is modelled using a standard homogeneous Markov chain leading to an H type control problem for stochastic multi-agent systems with Markovian jumps. The considered H type criterion includes, besides the components corresponding to the attenuation condition of exogenous disturbance inputs, quadratic terms aiming to acquire the consensus between the agents. It is shown that in the case of identical agents, a state-feedback controller with Markov parameters may be determined solving two specific systems of Riccati equations whose dimension does not depend on the number of agents. Iterative procedures to solve such systems are also presented together with an illustrative numerical example.

1. Introduction

Multi-agent systems received a considerable interest in control engineering over the last decades due to their wide area of applications including terrestrial, maritime, aerial and space surveillance and monitoring missions. Some early developments and comprehensive surveys in this field may be found, for instance, in [1,2,3,4]. The design requirements for the control of multi-agent systems may be formulated from different perspectives, a lot of literature treating these topics being available these days. Although the review of the control design methodologies for multi-agent systems is beyond the purpose of this paper, one will mention however some monographs as [5,6,7,8], presenting different multi-agent control problems. An important aspect related the multi-agent systems is their distributed control which is characterised, in contrast with the centralised case, by the absence of a control decision-maker. Such formulations, primarily considered for agents with single and double integrators models (see, e.g., [9]) have been then investigated for more general linear systems, as in [10]. Consensus of nonlinear agents using output feedback has been analysed, for instance, in [11] and an algorithm based on feedback linearisation of nonlinear agents with output measurements may be found in [12]. In [13], a stability analysis from the perspective of a hybrid modelling is proposed using invariant sets and the Lyapunov stability theory. Consensus problems for Bernoulli networks have been considered for instance in [14] and for high-order multi-agent systems under deterministic and Markovian switching network topologies, in [15]. Other consensus problems in stochastic sense have been investigated, for instance, in [16] and in [17] under Markov switching networks for agents with first and second order dynamics. Many formulations of the control problems for multi-agent systems include optimisation criteria. As shown for instance, in [18], the complexity associated with the computation of distributed optimal controller significantly increases. In [19], the authors study the structural properties of optimal control problems with infinite-horizon linear-quadratic criteria by analysing the spatial structure of the solution to corresponding Lyapunov operator and Riccati equations. A problem of synthesising a distributed dynamic output feedback achieving H performance is presented in [20]. In [21], a decentralised Markovian H control problem is considered for first-order dynamics of the agents in presence of time-delay conditions; the optimal control is expressed in terms of the solutions of a system of linear matrix inequalities depending on the dynamics of the multi-agent system. An H state-feedback consensus controller for undirected multi-agent systems is derived in [22] for a more general class of N identical agents; the complexity of the H consensus problem is reduced by representing the problem by N number linear systems. In fact, optimal control problems of multi-agent systems remain a domain of interest due to the design and implementation complexity of the control laws, even in the case of identical agents (see, for instance, [22,23,24,25]). In [26], a linear-quadratic control problem for identical agents is considered, for which it is proved that the optimal solution depends on the number of agents and on the stabilising solutions of two Riccati equations having the same order as the agents dynamics. This conclusion shows that the computational requirements may be significantly reduced, especially for large-scale multi-agent systems. The research has been continued in [27] where the robustness properties of the decentralised linear-quadratic optimal controller are analysed.
More recently, an H optimisation problem for identical stochastic linear models corrupted with multiplicative noises was formulated and solved in [28], aiming to provide disturbance attenuation performance together with robust stability with respect to parametric uncertainties in the agents models. Based on the state-feedback gains of the centralised state-feedback controller, a distributed controller depending on the adjacency matrix associated with the undirected graph of the communication network was obtained using the spectra of Lyapunov operators.
In this paper, the loss of links between agents is modelled by linear stochastic systems with Markovian jumps. The proposed methodology allows to consider different configurations of the network, each of them corresponding to a state of the Markov chain. Markov switching network models may be found in many papers between which one mentions [29,30,31]. The formulation and the developments presented in this paper considers stochastic models both for the agents and for the network. The H cost function for the optimal control design includes, besides the expression for the attenuation of exogenous disturbance inputs, quadratic terms aiming to acquire the state consensus between the agents. The optimal H state-feedback gains are expressed in terms of the stabilising solutions of two systems of coupled game-theoretic Riccati equations having the same order as the dynamics of a single agent. This coupling between the Riccati equations is typical in the optimal control of stochastic Markovian systems, depending on the elements of the stationary transition rate matrix.
The paper is organised as follows: the H control problem for stochastic system with Markovian jumps is formulated and solved in Section 2. The optimal gains of the control law are expressed in terms of the stabilising solution of a system of coupled algebraic Riccati equations with indefinite sign. A convergent iterative algorithm to determine the stabilising solution of this system of coupled Riccati equations is also presented. The Section 3 analyses the case of multi-agent H control. The main result of this section allows to determine the optimal control law for multi-agent systems solving two specific systems of coupled game-theoretic Riccati equations corresponding to the dimension of a single agent. In Section 4, the case of dropout data packages in the communication networks is discussed and illustrated by a numerical example for a large-scale multi-agent system with two states of the Markov chain. The paper ends with some concluding remarks.

2. H Type Control for Stochastic Systems with Markovian Jumps; The Case of a Single Agent

Consider the linear stochastic system
x ˙ ( t ) = A ( η ( t ) ) x ( t ) + B 1 ( η ( t ) ) w ( t ) + B 2 ( η ( t ) ) u ( t ) y 1 ( t ) = C ( η ( t ) ) x ( t ) + D ( η ( t ) ) u ( t ) y 2 ( t ) = x ( t )
where x R n denotes the state vector, w R m 1 is an exogenous input, u R m 2 stands for the control input, y 1 R p 1 is the quality output and y 2 R n denotes the measured output. Throughout the paper η ( t ) , t 0 denotes a continuous Markov chain with the state space D = { 1 , , d } and with the probability transition matrix P ( t ) = p i j ( t ) = e Π t , i , j D , t 0 in which the stationary transition rate matrix of η is Π = π i j with j = 1 d π i j = 0 , i D and π i j 0 if i j .
The triple { Ω , F , P } denotes a given probability space, E [ x ] stands for the expectation of the random variable x, E x | H represents the conditional expectation of x with respect to the σ -algebra H F and E [ x | η ( t ) = i ] is the conditional expectation with respect to the event η ( t ) = i . In the following developments it will be assumed that C ( i ) D ( i ) = 0 and D ( i ) D ( i ) = I m 2 , i D . For invertible D ( i ) D ( i ) , i = 1 , , d , if these assumptions are not accomplished one may perform the following change of the control variable u
u ( η ( t ) ) = D ( η ( t ) ) D ( η ( t ) ) 1 D ( η ( t ) ) C ( η ( t ) ) x ( t ) + D ( η ( t ) ) D ( η ( t ) ) 1 2 u ˜ ( t )
for which one can easily check that with the new control variable u ˜ the orthogonality condition C ( i ) D ( i ) = 0 holds and D ( i ) D ( i ) = I m 2 , i = 1 , , d .
For a multi-agent system with N agents, the indexes k , = 1 , , N will be used to define the connection between the agents k and .
Some known definitions and results used in the following developments will be briefly reminded (more details and proofs may be found, for instance, in [32,33]).
Definition 1. 
The stochastic system with Markov parameters
x ˙ ( t ) = A ( η ( t ) ) x ( t )
is called exponentially stable in mean square (ESMS) if there exists β 1 and α > 0 such that E | Φ ( t ) | 2 | n ( 0 ) = i β e α t , t 0 , i D , where Φ ( t ) denotes the fundamental (random) solution of the differential system (2).
Proposition 1. 
The stochastic system (2) is ESMS if and only if there exist the matrices X ( i ) > 0 , i = 1 , , d verifying the system of Lyapunov-type inequalities
A ( i ) X ( i ) + X ( i ) A ( i ) + j = 1 d π i j X ( j ) < 0 .
Throughout the paper it is assumed that the system
x ˙ ( t ) = A η ( t ) x ( t ) y ( t ) = C η ( t ) x ( t )
is stochastically detectable, namely there exist a set of matrices H ( i ) , i = 1 , . . . , d such that the system x ˙ ( t ) = A η ( t ) + H η ( t ) C η ( t ) x ( t ) is ESMS.
The proof of the next result may be found, for instance, in [33] (Theorem 7 of Chapter 3).
Proposition 2. 
If the system (4) is stochastically detectable and if the system of Lyapunov-type equations
A ( i ) X ( i ) + X ( i ) A ( i ) + C ( i ) C ( i ) + j = 1 d π i j X ( j ) = 0
has a symmetric solution with X ( i ) 0 , i D , then it is ESMS.
Proposition 3. 
If v : R n × D R is a function of C 1 class for every i D then
E v ( x ( t ) , η ( t ) ) | η ( 0 ) = i v ( x 0 , i ) = E 0 t x ( τ ) A ( η ( τ ) ) + j = 1 d v ( x ( τ ) , j ) π η ( τ ) j d τ | η ( 0 ) = i , i D , t 0 ,
where x ( t ) is the solution of the system (2) with the initial condition x 0 .
The main result of this section is the following theorem.
Theorem 1. 
If the system of coupled Riccati equations
A ( i ) X ( i ) + X ( i ) A ( i ) + X ( i ) γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X ( i ) + j = 1 d π i j X ( j ) + C ( i ) C ( i ) = 0
has a stabilizing solution X ( 1 ) , . . . , X ( d ) with X ( i ) 0 , i D for a certain γ > 0 , namely if the stochastic system with Markov jumps
x ˙ ( t ) = A ( η ( t ) ) + γ 2 B 1 ( η ( t ) ) B 1 ( η ( t ) ) B 2 ( η ( t ) ) B 2 ( η ( t ) ) X ( η ( t ) ) ) x ( t )
is ESMS, where
F ( η ( t ) ) : = B 2 ( η ( t ) ) X ( η ( t ) ) ,
then the state-feedback control law u ( t ) = F ( η ( t ) ) x ( t ) stabilises the system (1) and
E 0 | y 1 ( t ) | 2 γ 2 | w ( t ) | 2 d t 0
for all w L η 2 [ 0 , ) , R m 1 , where the quality output y 1 ( t ) is determined with the initial condition x ( 0 ) = 0 of the system (1).
Proof. 
In order to prove that the state feedback gain (6) stabilises (1) one may firstly rewrite the Riccati system (5) as
A ( i ) + B 2 ( i ) F ( i ) X ( i ) + X ( i ) A ( i ) + B 2 ( i ) F ( i ) + γ 2 X ( i ) B 1 ( i ) B 1 ( i ) X ( i ) X ( i ) B 2 ( i ) B 2 ( i ) X ( i ) + j = 1 d π i j X ( j ) + C ( i ) C ( i ) = 0
with X ( i ) 0 , i = 1 , , d . On the other hand, since the system (4) is assumed stochastically detectable, it follows that the system
x ˙ ( t ) = A ( η ( t ) ) + B 2 ( η ( t ) ) F ( η ( t ) ) x ( t ) y ( t ) = γ 1 B 1 ( η ( t ) ) X ( η ( t ) ) B 2 ( η ( t ) ) X ( η ( t ) ) C ( η ( t ) ) x ( t )
is also stochastically detectable and then, based on Proposition 2, one concludes that the above stochastic system with the state matrix A ( η ( t ) ) + B 2 ( η ( t ) ) F ( η ( t ) ) is ESMS. Since the system (1) with the control law u ( t ) = F ( η ( t ) ) x ( t ) has the same state matrix it follows that the state-feedback (6) is stabilising.
In order to prove the last part of the theorem one introduces the function V ( x ( t ) , η ( t ) ) = x ( t ) X ( η ( t ) ) x ( t ) . Using Proposition 3 for a certain initial condition x 0 , it follows that
E x ( t ) X ( η ( t ) ) x ( t ) x 0 X ( η ( 0 ) ) x 0 | η ( 0 ) = i = E 0 2 A ( η ( τ ) ) x ( τ ) + B 1 ( η ( τ ) ) w ( τ ) + B 2 ( η ( τ ) ) u ( τ ) X ( η ( τ ) ) x ( τ ) + j = 1 d π η ( τ ) j x ( τ ) X ( j ) x ( τ ) d τ | η ( 0 ) = i .
Adding J ( i , w , u ) : = E 0 t | y 1 ( τ ) | 2 γ 2 | w ( τ ) | 2 d τ | η ( 0 ) = i , i D and using (5) one obtains
J ( i , w , u ) + E x ( t ) X ( η ( t ) ) x ( t ) x 0 X ( η ( 0 ) ) x 0 | η ( 0 ) = i = E 0 t x ( τ ) C ( η ( τ ) ) C ( η ( τ ) ) x ( τ ) + u ( τ ) u ( τ ) γ 2 w ( τ ) w ( τ ) + x ( τ ) A ( ( η ( τ ) ) X ( η ( τ ) ) + X ( η ( τ ) ) A ( η ( τ ) ) x ( τ ) + w ( τ ) B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) + x ( τ ) X ( η ( τ ) ) B 1 ( η ( τ ) ) w ( τ ) + u ( τ ) B 2 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) + x ( τ ) X ( η ( τ ) ) B 2 ( η ( τ ) ) u ( τ ) + j = 1 d π η ( τ ) j x ( τ ) X ( j ) x ( τ ) d τ | η ( 0 ) = i = E 0 t x ( τ ) A ( η ( τ ) ) X ( η ( τ ) ) + X ( η ( τ ) ) A ( η ( τ ) ) + X ( η ( τ ) ) γ 2 B 1 ( η ( τ ) ) B 1 ( η ( τ ) ) B 2 ( η ( τ ) ) B 2 τ ( η ( τ ) ) X ( η ( τ ) ) + j = 1 d π η ( τ ) j x ( τ ) X ( j ) + C ( η ( τ ) ) C ( η ( τ ) ) x ( τ ) + u ( τ ) + B 2 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) u ( τ ) + B 2 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) γ w ( τ ) γ 1 B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) γ w ( τ ) γ 1 B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) d τ | η ( 0 ) = i .
Taking into account (5), one obtains
J ( i , w , u ) + E x ( t ) X ( η ( t ) ) x ( t ) x 0 X ( η ( 0 ) ) x 0 | η ( 0 ) = i = E 0 t u ( τ ) + B 2 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) u ( τ ) + B 2 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) γ w ( τ ) γ 1 B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) γ w ( τ ) γ 1 B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) d τ | η ( 0 ) = i .
For t and u ( t ) = F ( η ( t ) ) x ( t ) , from the above equation it follows that
E 0 | y 1 ( t ) | 2 γ 2 | w ( t ) | 2 d t = i = 1 d p i ( 0 ) x 0 X ( i ) x 0 i = 1 d p i ( 0 ) E 0 γ w ( τ ) γ 1 B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) × γ w ( τ ) γ 1 B 1 ( η ( τ ) ) X ( η ( τ ) ) x ( τ ) d τ | η ( 0 ) = i
where p i ( 0 ) : = P η ( 0 ) = i . Then, for x 0 = 0 , the inequality (7) directly follows and it becomes equality for w * ( t ) = γ 2 B 1 T X ( η ( t ) ) x ( t ) . □
Remark 1. 
Matrix Riccati equations with indefinite sign as in the system (5) appear in H control ([34]) and in mixed H 2 / H control problems ([35]) in the deterministic framework.
The next result proved in [36] gives a numerical procedure to compute the stabilising solution X ( 1 ) , , X ( d ) of the system of game theoretic Riccati-type Equation (5) assuming that such a solution exists.
Proposition 4. 
Assume that the system (1) is stochastically detectable and that the system of Riccati Equation (5) has a stabilising solution. Then the sequences X k ( i ) k 0 , Z k ( i ) k 0 defined by X 0 ( i ) = 0 and X k + 1 ( i ) = X k ( i ) + Z k ( i ) , i = 1 , , d , where Z 0 ( i ) are the stabilising solutions of the system of Riccati type equations
A ( i ) + 1 2 π i i I n Z 0 ( i ) + Z 0 ( i ) A ( i ) + 1 2 π i i I n Z 0 ( i ) B 2 ( i ) B 2 ( i ) Z 0 ( i ) + C ( i ) C ( i ) + j = 1 , j i d Z 0 ( i ) = 0 ,
Z k ( i ) , k 1 are the stabilising solutions of the un-coupled Riccati equations
M k ( i ) Z k ( i ) + Z k ( i ) M k ( i ) Z k ( i ) B 2 ( i ) B 2 ( i ) Z k ( i ) + R k ( i ) = 0 ,
and where
M k ( i ) = A ( i ) + 1 2 π i i I n + γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X k ( i ) R k ( i ) = γ 2 Z k 1 ( i ) B 1 ( i ) B 1 ( i ) Z k 1 ( i ) + j = 1 , j i d π i j Z k 1 ( j ) ,
are convergent and the limit of X k ( i ) , i = 1 , , d when k is the stabilising solution of the system (5).
An iterative algorithm to solve the system of coupled Riccati equations with definite sign is given for completeness in Appendix A. The proof of the algorithm convergence may be found in [33] (Theorem 21 of Chapter 4).

3. Markovian H Controller Design for Multi-Agent Systems

Consider N > 1 agents with identical dynamics of form
x ˙ k ( t ) = A ( η ( t ) ) x k ( t ) + B 1 ( η ( t ) ) w k ( t ) + B 2 ( η ( t ) ) u k ( t ) y 1 k ( t ) = C ( η ( t ) ) x k ( t ) + D ( η ( t ) ) u k ( t ) y 2 k ( t ) = x k ( t ) , t 0 , k = 1 , , N
with C ( i ) D ( i ) = 0 and D ( i ) D ( i ) = I m 1 , i = 1 , , d , and k = 1 , , N .
Remark 2. 
Although in (9) one considered the same standard homogeneous Markov chain for all agents one may also treat the case when each agent is modelled with its own stochastic process η k ( t ) . Indeed, if each agent dynamics is modelled with a standard homogenous Markov chain η k ( t ) , k = 1 , . . . , N with d states then one may consider for the multi-agent system (9) an extended Markov chain with d N states. Since in the present paper the Markov parameters are used to characterise the availability or the link failure between agents it follows that the maximal number of states of the Markov chain in these applications is 2 N .
The dynamics of the multi-agent system (9) may be written in the following compact form
x ˜ ˙ ( t ) = A ˜ ( η ( t ) ) x ˜ ( t ) + B ˜ 1 ( η ( t ) ) w ˜ ( t ) + B ˜ 2 ( η ( t ) ) u ˜ ( t ) y ˜ 1 ( t ) = C ˜ ( η ( t ) ) x ˜ ( t ) + D ˜ ( η ( t ) ) u ˜ ( t ) y ˜ 2 ( t ) = x ˜ ( t )
in which x ˜ : = x 1 , , x N , w ˜ : = w 1 , , w N , u ˜ : = [ u 1 , , u N ] , y ˜ 1 : = y 11 , , y 1 N and A ˜ ( η ( t ) ) : = I N A ( η ( t ) ) , B ˜ 1 ( η ( t ) ) : = I N B 1 ( η ( t ) ) , B ˜ 2 ( η ( t ) ) : = I N B 2 ( η ( t ) ) , C ˜ ( η ( t ) ) : = I N C ( η ( t ) ) , D ˜ ( η ( t ) ) : = I N D ( η ( t ) ) where ⊗ denotes the Kronecker product.
For γ > 0 define the cost function
J w ˜ , u ˜ = E 0 | y ˜ 1 ( t ) | 2 γ 2 | w ˜ ( t ) | 2 + 1 2 k = 1 N = 1 , k N x k ( t ) x ( t ) Q k ( η ( t ) ) x k ( t ) x ( t ) d t
where Q k ( i ) , k , = 1 , , N and i = 1 , , d are positive semidefinite weighting matrices. Then (11) may be rewritten as
J w ˜ , u ˜ = E 0 y ˜ 1 ( t ) y ˜ 1 ( t ) γ 2 w ˜ ( t ) w ˜ ( t ) + x ˜ ( t ) Q ˜ ( η ( t ) ) x ˜ ( t ) d t
in which Q ˜ ( i ) has the block elements
Q ˜ k k ( i ) = C ( i ) C ( i ) + = 1 , k N Q k ( i ) , Q ˜ k ( i ) = Q k ( i ) , k , = 1 , , N , k ,
i = 1 , , d . Choosing Q k ( i ) = P ( i ) P ( i ) , k , = 1 , , N , k , i = 1 , , d , it follows that the block elements of the matrix Q ˜ ( i ) are
Q ˜ k k ( i ) = C ( i ) C ( i ) + ( N 1 ) P ( i ) P ( i ) , Q ˜ k ( i ) = P ( i ) P ( i ) , k , = 1 , , N , k ,
i = 1 , , d . One can easily check that
y ˜ 1 ( t ) y ˜ 1 ( t ) + x ˜ ( t ) Q ˜ ( η ( t ) ) x ˜ ( t ) = C ˜ ( η ( t ) ) x ˜ ( t ) + D ˜ ( η ( t ) ) u ˜ ( t )
where
C ˜ ( i ) : = P ˜ ( i ) C ( i ) 0 0 0 C ( i ) 0 0 0 C ( i ) , D ˜ ( i ) : = 0 n · N × m 2 · N D ( i ) 0 0 0 D ( i ) 0 0 0 D ( i ) ,
i = 1 , . . . , d with P ˜ ( i ) satisfying the condition
P ˜ ( i ) P ˜ ( i ) = ( N 1 ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) ( N 1 ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) P ( i ) ( N 1 ) P ( i ) P ( i ) ,
i = 1 , , d . Therefore the cost function (11) may be rewritten as
J ( w ˜ , u ˜ ) = E 0 z ˜ ( t ) z ˜ ( t ) γ 2 w ˜ ( t ) w ˜ ( t ) d t
where z ( t ) = C ˜ ( η ( t ) ) x ˜ ( t ) + D ˜ ( η ( t ) ) u ˜ ( t ) . Moreover, since it was assumed that C ( i ) D ( i ) = 0 and D ( i ) D ( i ) = I m 2 , i = 1 , , d it follows that C ˜ ( i ) D ˜ ( i ) = 0 and D ˜ ( i ) D ˜ ( i ) = I m 1 · N , i = 1 , , d and therefore one may apply Theorem 1 for the Markov stochastic system
x ˜ ˙ ( t ) = A ˜ ( η ( t ) ) x ( t ) + B ˜ 1 ( η ( t ) ) w ( t ) + B ˜ 2 ( η ( t ) ) u ( t ) z ˜ ( t ) = C ˜ ( η ( t ) ) x ˜ ( t ) + D ˜ ( η ( t ) ) u ˜ ( t ) y ˜ 2 ( t ) = x ˜ ( t )
with the cost function (15).
The main result of this section is given by the following theorem.
Theorem 2. 
(i) If the system of coupled Riccati equations
A ˜ ( i ) X ˜ ( i ) + X ˜ ( i ) A ˜ ( i ) + X ˜ ( i ) γ 2 B ˜ 1 ( i ) B ˜ 1 ( i ) B ˜ 2 ( i ) B ˜ 2 ( i ) X ˜ ( i ) + j = 1 d π i j X ˜ ( j ) + Q ˜ ( i ) Q ˜ ( i ) = 0 , i = 1 , , d .
has a stabilising solution X ˜ ( 1 ) , , X ˜ ( d ) with X ˜ ( i ) 0 , i = 1 , , d then the stochastic system with Markov parameters
x ˜ ˙ ( t ) = A ˜ ( η ( t ) ) + B ˜ 2 ( η ( t ) ) F ˜ ( η ( t ) ) x ( t ) + B ˜ 1 ( η ( t ) ) w ˜ ( t )
where F ˜ ( i ) = B ˜ 2 ( i ) X ˜ ( i ) , i = 1 , , d , is ESMS and for the initial condition x ˜ ( 0 ) = 0 ,
E 0 | y ˜ 1 ( t ) | 2 γ 2 | w ˜ ( t ) | 2 d t 0
for all w ˜ L η 2 [ 0 , ) , R N · m 1 .
(ii) The solution of (17) has the following structure
X ˜ ( i ) : = X ˜ k k , = 1 , , N w h e r e X ˜ k k ( i ) : = X 1 ( i ) + ( N 1 ) X 2 ( i ) X ˜ k ( i ) : = X 2 ( i ) , k , = 1 , , N , k ,
in which X 1 ( 1 ) , , X 1 ( d ) and X 2 ( 1 ) , , X 2 ( d ) are the solutions of the Riccati type equations
A ( i ) X 1 ( i ) + X 1 ( i ) A ( i ) + X 1 ( i ) γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X 1 ( i ) + j = 1 d π i j X 1 ( j ) + C ( i ) C ( i ) = 0 , i = 1 , , d
and
A ( i ) + γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X 1 ( i ) X 2 ( i ) + X 2 ( i ) ( A ( i ) + γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X 1 ( i ) + N X 2 ( i ) γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X 2 ( i ) + j = 1 d π i j X 2 ( j ) + P ( i ) P ( i ) = 0 , i = 1 , , d ,
respectively.
(iii) If the Riccati type systems (20) and (21) have the stabilising solutions X 1 ( 1 ) , , X 1 ( d ) and X 2 ( 1 ) , , X 2 ( d ) respectively, with X 1 ( i ) 0 and X 2 ( i ) 0 , i = 1 , , d then X ˜ ( 1 ) , . . . , X ˜ ( d ) with X ˜ ( i ) defined in (19), is the stabilising solution of (17) and X ˜ ( i ) 0 , i = 1 , , d .
Proof. 
Part (i) of the statement is a direct consequence of Theorem 1.
The proof of (ii) is inspired by the arguments given in Theorems 1 and 2 of [26] for the multi agent linear quadratic control problem in deterministic framework. Thus, the stabilising solution X ˜ ( 1 ) , , X ˜ ( d ) of (17) has an N × N blocks structure, each of them having the size ( n × n ) . Since all the matrix coefficients in (17) are diagonal it follows that the diagonal and the off-diagonal elements of X ˜ ( i ) , i = 1 , , d are equal, respectively. Then the diagonal block elements will be denoted by X ˜ 1 ( i ) and the off-diagonal ones by X ˜ 2 ( i ) , i = 1 , , d . The blocks ( k , ) , k , { 1 , , N } of (17) have the form
A ( i ) X ˜ k ( i ) + X ˜ k ( i ) A ( i ) + m = 1 N X ˜ k m ( i ) γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X ˜ m ( i ) + Q ˜ k l ( i ) + j = 1 d π i j X ˜ k ( j ) = 0 .
Denoting X 1 ( i ) : = X ˜ 1 ( i ) + ( N 1 ) X ˜ 2 ( i ) , i = 1 , , d and summing up the terms of the k-th row of (17) it follows that X 1 ( i ) , i = 1 , , d verifies (20). Further, for any off-diagonal block ( k , ) with k and k , { 1 , , N } , direct computations give
A ( i ) + γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X ˜ 1 ( i ) X ˜ 2 ( i ) + X ˜ 2 ( i ) A ( i ) + γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X ˜ 1 ( i ) + ( N 2 ) X ˜ 2 ( i ) γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) X ˜ 2 ( i ) + Q ˜ k ( i ) + j = 1 d π i j X ˜ 2 ( j ) = 0 .
Using the fact that X ˜ 1 ( i ) = X 1 ( i ) ( N 1 ) X ˜ 2 ( i ) and that Q ˜ k ( i ) = P ( i ) P ( i ) , i = 1 , , d from the above equation it follows that X 2 ( i ) : = X ˜ 2 ( i ) is a solution of (21).
For part (iii) of the statement, one will prove that the stochastic system
x ˜ ˙ ( t ) = A ˜ ( η ( t ) ) + γ 2 B ˜ 1 ( η ( t ) ) B ˜ 1 ( η ( t ) ) B ˜ 2 ( η ( t ) ) B ˜ 2 ( η ( t ) ) X ˜ ( η ( t ) ) x ˜ ( t )
is ESMS, namely there exist the positive definite matrices S ˜ ( i ) , i = 1 , , d such that
A ˜ ( η ( t ) ) + γ 2 B ˜ 1 ( η ( t ) ) B ˜ 1 ( η ( t ) ) B ˜ 2 ( η ( t ) ) B ˜ 2 ( η ( t ) ) X ˜ ( η ( t ) ) S ˜ ( i ) + S ˜ ( i ) A ˜ ( η ( t ) ) + γ 2 B ˜ 1 ( η ( t ) ) B ˜ 1 ( η ( t ) ) B ˜ 2 ( η ( t ) ) B ˜ 2 ( η ( t ) ) X ˜ ( η ( t ) ) + j = 1 d π i j S ˜ ( j ) < 0 , i = 1 , , d .
Since X 1 ( 1 ) , , X 1 ( d ) is the stabilising solution of the Riccati-type system (20) it follows that the stochastic system
x ˙ ( t ) = A ( η ( t ) ) + M ( η ( t ) ) X 1 ( η ( t ) ) x ( t ) ,
where one denoted M ( i ) : = γ 2 B 1 ( i ) B 1 ( i ) B 2 ( i ) B 2 ( i ) , is ESMS and therefore there exist the positive definite matrices S 1 ( i ) , i = 1 , , d such that
A ( i ) + M ( i ) X 1 ( i ) S 1 ( i ) + S 1 ( i ) A ( i ) + M ( i ) X 1 ( i ) + j = 1 d π i j S 1 ( j ) < 0 .
Similarly, based on the fact that X 2 ( 1 ) , , X 2 ( d ) is the stabilising solution of (21), there exist the positive definite matrices S 2 ( i ) , i = 1 , , d such that
A ( i ) + M ( i ) X 1 ( i ) + N X 2 ( i ) S 2 ( i ) + S 2 ( i ) A ( i ) + M ( i ) X 1 ( i ) + N X 2 ( i ) + j = 1 d π i j S 2 ( j ) < 0 , i = 1 , , d .
Define
S ˜ ( i ) = S 1 ( i ) + ( N 1 ) S 2 ( i ) S 2 ( i ) S 2 ( i ) S 2 ( i ) S 1 ( i ) + ( N 1 ) S 2 ( i ) S 2 ( i ) S 2 ( i ) S 2 ( i ) S 1 ( i ) + ( N 1 ) S 2 ( i )
for which the inequalities (23) become:
P ( i ) : = P 1 ( i ) P 2 ( i ) P 2 ( i ) P 2 ( i ) P 1 ( i ) P 2 ( i ) P 2 ( i ) P 2 ( i ) P 1 ( i ) < 0 , i = 1 , , d
where
P 1 ( i ) = A ( i ) + M ( i ) X 1 ( i ) + ( N 1 ) X 2 ( i ) S 1 ( i ) + ( N 1 ) S 2 ( i ) + S 1 ( i ) + ( N 1 ) S 2 ( i ) A ( i ) + M ( i ) X 1 ( i ) + ( N 1 ) X 2 ( i ) + ( N 1 ) X 2 ( i ) M ( i ) S 2 ( i ) + S 2 ( i ) M ( i ) X 2 ( i ) + j = 1 d π i j S 1 ( j ) + ( N 1 ) S 2 ( j )
and
P 2 ( i ) = A ( i ) + M ( i ) X 1 ( i ) + ( N 1 ) X 2 ( i ) S 2 ( i ) S 2 ( i ) A ( i ) + M ( i ) X 1 ( i ) + ( N 1 ) X 2 ( i ) S 1 ( i ) + S 2 ( i ) M ( i ) X 2 ( i ) X 2 ( i ) M ( i ) S 1 ( i ) + S 2 ( i ) j = 1 d π i j S 2 ( j ) ,
i = 1 , , d . Defining the matrix
T = I n I n I n I n 0 I n 0 0 0 0 0 I n
direct computations give
T P ( i ) T 1 = P 1 ( i ) + ( N 1 ) P 2 ( i ) 0 0 P 1 ( i ) P 2 ( i ) 0 0 P 1 ( i ) P 2 ( i )
in which ∗ denotes irrelevant elements. Due to the triangular structure of the above system it follows that the spectra of P ( i ) , i = 1 , , d is given by the reunion of the spectra of P 1 ( i ) + ( N 1 ) P 2 ( i ) and of P 1 ( i ) P 2 ( i ) , respectively, i = 1 , , d . On the other hand, one may direcly check that P 1 ( i ) + ( N 1 ) P 2 ( i ) coincide with the left hand sides of (24), i = 1 , , d . Further, using the above expressions of P 1 ( i ) and P 2 ( i ) , it follows that
P 1 ( i ) P 2 ( i ) = A ( i ) + M ( i ) X 1 ( i ) + N X 2 ( i ) S 1 ( i ) + N S 2 ( i ) + S 1 ( i ) + N S 2 ( i ) A ( i ) + M ( i ) X 1 ( i ) + N X 2 ( i ) + j = 1 d π i j S 1 ( i ) + N S 2 ( i ) ,
i = 1 , , d . Since (24) remains true if S 1 ( i ) are replaced by ϵ S 1 ( i ) , i = 1 , , d for any ϵ > 0 , from (25) it follows that for a small enough ϵ > 0 , P 1 ( i ) P 2 ( i ) < 0 . Thus one concludes that P ( i ) < 0 , i = 1 , , d and therefore X ˜ ( 1 ) , , X ˜ ( d ) is the stabilising solution of the Riccati system (17). The fact that X ˜ ( i ) , i = 1 , , d given by (19) are positive semidefinite directly follows taking into account that
T X ˜ ( i ) T 1 = X 1 ( i ) 0 0 X 1 ( i ) + N X 2 ( i ) 0 0 X 1 ( i ) + N X 2 ( i )
and using the assumption that X 1 ( i ) and X 2 ( i ) are positive semidefinite. Thus the proof ends. □
Remark 3. 
Based on the expressions of X ˜ ( i ) and F ˜ ( i ) , i = 1 , , d , it follows that F ˜ ( i ) has the following structure
F ˜ ( i ) = F 1 ( i ) F 2 ( i ) F 2 ( i ) F 2 ( i ) F 1 ( i ) F 2 ( i ) F 2 ( i ) F 2 ( i ) F 1 ( i )
where
F 1 ( i ) = B 2 ( i ) X 1 ( i ) + ( N 1 ) X 2 ( i ) F 2 ( i ) = B 2 ( i ) X 2 ( i ) , i = 1 , , d .

4. The Data Packet Dropout Case

From (26) and (29) it follows that the control of each agent is determined as a combination of its own states and the states of all other network agents. The case when the states of all agents are available for every agent represents the nominal case and it will be denoted as the state i = 1 of the set D . In the case when the state of a certain agent is not available, the corresponding terms in the control expression of the other agents will be zero which is equivalent with the fact that F ˜ 2 ( 2 ) = 0 . This case will be associated with the state i = 2 of D . Supposing that the communication network fails at a certain moment of time, it follows that F ˜ ( 2 ) = d i a g F 1 ( 2 ) , , F 1 ( 2 ) and X 2 ( 2 ) = 0 . The condition X 2 ( 2 ) = 0 is accomplished if the weights corresponding to the coupling terms in (11) are taken to be zero, namely if P ( 2 ) = 0 and if the corresponding row in the stationary transition rate matrix Π has null elements. To conclude, in the above considered scenario in which either the communication network properly works or it completely fails leads to a Markovian model with d = 2 states of the set D . If the connection with a single agent, let say k is lost then the gain F ˜ will have all extra diagonal elements of the k-th column equal to zero. Similarly, in the case when the connections with more agents fail, the columns of F ˜ corresponding to these agents will have zero extra diagonal elements; in this situation the Markov chain will have maximum 2 N states.
In order to illustrate these ideas one considers a networked system with N = 100 agents whose planar motions are described by the kinematic equations
x ¨ k ( t ) = u k ( t ) y ¨ k ( t ) = v k ( t ) , k = 1 , , N
where x k and y k denote the Cartesian coordinates and u i and v i their commanded accelerations, respectively. Two states of the Markov chain have been considered in this example. The first one corresponds to the case when the communication properly works and the states of all agents are available for all others; the second state of the Markov chain characterizes the situation in which each agent can only access its own state vector. Assuming that the quality outputs (denoted in Section 3 by y 1 k ) are x k y k u k v k it follows that the matrix coefficients in the representation (1) are identical for both states of the Markov chain, namely
A ( i ) = 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 , B 1 ( i ) = 1 0 0 0 0 1 0 0 , B 2 ( i ) = 0 0 1 0 0 0 0 1 , C ( i ) = 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 , D ( i ) = 0 0 0 0 1 0 0 1 , i = 1 , 2 ,
only the weights Q k ( i ) in the cost function (11) being different for the two states. Thus for i = 1 one chose Q k = 100 · I 4 (namely P ( 1 ) = 10 2 · I 4 ) and for i = 2 , Q k ( 2 ) = 0 4 , for all k . It was assumed that stationary transition rate matrix of η is
Π = 0.5 0.5 0 0 .
The elements of the second row of the stationary transition matrix were taken zero in order to accomplish the condition X 2 ( 2 ) = 0 . Indeed, for P ( 2 ) = 0 and for π 21 = π 22 = 0 , the unique stabilising solution of the Riccati system (21) has the form X 2 ( 1 ) , 0 since X 1 ( 1 ) , X 1 ( 2 ) is the stabilising solution of (20). The elements of the first row of Π have been chosen such that the transition from the nominal first state to the second one corresponding to the network failure takes place in about 5 s as illustrated in Figure 1.
Solving the Riccati systems (20) and (21) for γ = 100 one obtains
X 1 ( 1 ) = 1.7686 1.2373 0 0 1.2373 1.7379 0 0 0 0 1.7686 1.2373 0 0 1.2373 1.7379 , X 2 ( 1 ) = 204.1418 2.0204 0 0 2.0204 1.0006 0 0 0 0 204.1418 2.0204 0 0 2.0204 1.0006 , X 1 ( 2 ) = 1.4147 1.0002 0 0 1.0002 1.4145 0 0 0 0 1.4147 1.0002 0 0 1.0002 1.4145 , X 2 ( 2 ) = 0 4
resulting the following gains
F 1 ( 1 ) = 20.1631 10.0793 0 0 0 0 20.1631 10.0793 , F 2 ( 1 ) = 0.2024 0.1001 0 0 0 0 0.2024 0.1001 , F 1 ( 2 ) = 0.1012 0.1431 0 0 0 0 0.1012 0.1431 , F 2 ( 2 ) = 0 2 × 4 .
Using the aforementioned gains one determined the agents planar trajectories for both cases when the network communication properly works and the case when it fails, respectively. Figure 2 presents two snapshots at t = 0.5 s and at t = 2 s obtained for random initial positions of the agents in both cases. One can see from the two snapshots in the upper half of the figure, that the matrix gain P giving the coupling between agents is important in determining a short settling time. By contrast, numerical simulations show that in the case when the communication network is not available the settling time is about 6.5 s.

5. Concluding Remarks

An optimal H type control method for large-scale multi-agent systems with identical dynamics and independently actuated is presented. It is shown that regardless the number of agents, the optimal solution may be obtained solving two systems of algebraic Riccati equations whose dimension correspond to a single agent. Convergent iterative numerical procedures are presented and used for a case study revealing the benefits of the coupling between agents. The proposed design methodology may be also used in applications in which only some of the links between agents fail. The dimension of the Riccati equations remains the same but their number increases due to the larger number of states of the Markov chain. Future research will be devoted to applications in which a sensitivity analysis with respect to the transition probabilities will be included.

Author Contributions

Conceptualization and methodology, A.-M.S.; software, A.-M.S.; validation, A.-M.S. and S.C.S.; formal analysis, S.C.S.; investigation, S.C.S.; writing—original draft preparation, S.C.S.; writing—review and editing, S.C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The following convergent algorithm is used to solve the system of coupled Riccati equations with definite sign (8), assuming that the stochastic system
x ˙ ( t ) = A ˜ ( η ( t ) x ( t ) + B 2 ( η ( t ) ) u ( t )
where A ˜ ( i ) : = A ( i ) + 1 2 π i i I n , i = 1 , , d , is stabilisable, namely there exist F ( 1 ) , , F ( d ) such that the resulting system obtained for u ( t ) = F ( η ( t ) ) x ( t ) , namely
x ˙ ( t ) = A ˜ ( η ( t ) ) + B 2 ( η ( t ) ) F ( η ( t ) ) x ( t )
is ESMS.
Step 1. Determine Y ( i ) > 0 and F ˜ 0 ( i ) , i = 1 , , d such that
A ˜ ( i ) + B 2 ( i ) F ˜ 0 ( i ) Y ( i ) + Y ( i ) A ˜ ( i ) + B 2 ( i ) F ˜ 0 ( i ) + j = 1 , j i d π j i Y ( j ) < 0 .
The above system of inequalities may be solved using a linear matrix inequalities (LMIs) solver, denoting V ( i ) : = F ˜ 0 ( i ) Y ( i ) and solving the resulting system of LMIs with respect to V ( i ) and Y ( i ) > 0 , i = 1 , , d . Then, F ˜ 0 ( i ) = V ( i ) Y 1 ( i ) , i = 1 , , d .
Step 2. For an arbitrary ε > 0 solve the following system of LMIs
A ˜ ( i ) + B 2 ( i ) F ˜ 0 ( i ) Z ˜ 0 ( i ) + Z ˜ 0 ( i ) A ˜ ( i ) + B 2 ( i ) F ˜ 0 ( i ) + C ( i ) C ( i ) + F ˜ 0 ( i ) F ˜ 0 ( i ) + j = 1 , j i d π i j Z ˜ 0 ( j ) + ε I n < 0
with respect to Z ˜ 0 ( i ) > 0 , i = 1 , , d .
Step 3. With F ˜ 0 ( i ) and Z ˜ 0 ( i ) , i = 1 , , d determined at Steps 1 and 2, respectively and with an arbitrary ε > 0 , solve iteratively for k 1 the system of Lyapunov equations
A ˜ ( i ) + B 2 ( i ) F ˜ k 1 ( i ) Z ˜ k ( i ) + Z ˜ k ( i ) A ˜ ( i ) + B 2 ( i ) F ˜ k 1 ( i ) + F ˜ k 1 ( i ) F ˜ k 1 ( i ) + C ( i ) C ( i ) + ε k + 1 I n + j = 1 , j i d π i j Z ˜ k 1 ( i ) = 0 , F ˜ k ( i ) = B 2 ( i ) Z ˜ k ( i ) , i = 1 , , d .
Since in [33] it is proved that the sequence Z ˜ k ( i ) , i = 1 , , d tends to the stabilising solution of (8) when k , Step 3 will be repeated until Z ˜ k + 1 ( i ) Z ˜ k ( i ) , i = 1 , , d are small enough.

References

  1. Dorri, A.; Kanhere, S.; Jurdak, R. Multi-Agent Systems. A survey. IEEE Access 2018, 6, 28573–28593. [Google Scholar] [CrossRef]
  2. Sandell, N.R.; Varaiya, P.; Athans, M.; Safonov, M. Survey of Descentralised Control Methods for Large Scale Systems. IEEE Trans. Autom. Control 1978, 23, 108–128. [Google Scholar] [CrossRef]
  3. Siljak, D.D. Complex Dynamics Systems: Dimensionality, Structure and Uncertainty. Large Scale Syst. 1983, 4, 279–291. [Google Scholar]
  4. Tsitsiklis, J.N.; Bertsekas, D.P.; Athans, M. Distributed asynchronous deterministic and stochastic gradient optimization algorithms. IEEE Trans. Autom. Control 1986, 31, 803–812. [Google Scholar] [CrossRef]
  5. Ren, W.; Cao, Y. Distributed Coordination of Multi-Agent Networks; Springer: London, UK, 2011. [Google Scholar]
  6. Shamma, J. Cooperative Control of Distributed Multi-Agent Systems; Wiley: Hoboken, NJ, USA, 2007. [Google Scholar]
  7. Wang, Y.; Garcia, E.; Casbeer, D.; Zhang, F. Cooperative Control of Multi-Agent Systems: Theory and Applications; Wiley: Hoboken, NJ, USA, 2017. [Google Scholar]
  8. Wang, J.; Wang, C.; Xin, M.; Ding, Z.; Shan, J. Cooperative Control of Multi-Agent Systems: An Optimal and Robust Perspective; Academic Press: London, UK, 2020. [Google Scholar]
  9. Ren, W.; Beard, R.W.; Atkins, E.M. Information Consensus in Multivehicle Cooperative Control. IEEE Control. Syst. Mag. 2007, 27, 71–81. [Google Scholar]
  10. Seo, J.; Shima, H.; Black, J. Consensus of high-order linear systems using dynamic output feedback compensator: Low gain approach. Automatica 2007, 43, 1049–1057. [Google Scholar] [CrossRef]
  11. Wu, Z.; Wu, Y.; Wu, Z.G. Synchronization of multi-agent systems via static output feedback control. J. Frankl. Inst. 2017, 354, 1049–1057. [Google Scholar] [CrossRef]
  12. Rehak, B.; Lynnyk, V. Consensus of nonlinear multi-agent system with output measurements. IFAC PapersOnLine 2021, 54, 400–405. [Google Scholar] [CrossRef]
  13. Yu, Z.; Li, X.; Xu, L.; Nasr, E.A.; Mahmoud, H. Stability analysis method and application of multi-agent systems from the perspective of hybrid systems. Meas. Control 2021, 54, 1347–1355. [Google Scholar] [CrossRef]
  14. Tahbaz-Salehi, A.; Jadbabaie, A. A necessary and sufficient condition for consensus over random networks. IEEE Trans. Autom. Control 2008, 53, 791–795. [Google Scholar] [CrossRef]
  15. Vengertsev, D.; Kim, H.; Seo, J.H.; Shim, H. Consensus of output-coupled high-order linear multi-agent systems under deterministic and Markovian switching networks. Int. J. Syst. Sci. 2013, 46, 1790–1799. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Tian, Y.P. Consentability and Protocol Design of Multi-Agent Systems with Stable and Unstable Subsystems. An Average Dwell Time Approach. In Proceedings of the American Control Conference, Chicago, IL, USA, 28–30 June 2000; pp. 1055–1061. [Google Scholar]
  17. Miao, G.; Xu, S.; Zou, Y. Necessary and Sufficient Conditions for Mean Square Consensus over Markov Switching Topologies. Int. J. Syst. Sci. 2013, 44, 178–186. [Google Scholar] [CrossRef]
  18. Gupta, V.; Hassibi, B.; Murray, R.M. A sub-optimal algorithm to synthesize control laws for a network of dynamic agents. Int. J. Control. 2005, 78, 1302–1313. [Google Scholar] [CrossRef]
  19. Motee, N.; Jadbabaie, A. Optimal control of spatially distributed systems. IEEE Trans. Autom. Control 2008, 53, 1616–1629. [Google Scholar] [CrossRef]
  20. Langbort, C.; Chandra, R.S.; D-Andrea, R. Distributed control of heterogeneous systems. IEEE Trans. Autom. Control 2008, 49, 1502–1519. [Google Scholar] [CrossRef]
  21. Abdollahi, F.; Khorasani, K. A decentralised Markovian Jump H Control Strategy for Mobile Multi-Agent Networked Systems. IEEE Trans. Control. Syst. Technol. 2011, 19, 269–283. [Google Scholar] [CrossRef]
  22. Wanigasekara, C.; Zhang, L.; Swain, A. H State-Feedback Consensus of Linear Multi-Agent Systems. Proceeding of the 17th International Conference on Control & Automation, Naples, Italy, 27–30 June 2022; pp. 710–715. [Google Scholar]
  23. Bauso, D.; Giarre, L.; Pesenti, R. Consensus for networks with unknown but bounded disturbances. SIAM J. Control Optim. 2009, 48, 1756–1770. [Google Scholar] [CrossRef]
  24. Dal Col, L.; Tarbouriech, S.; Zaccarian, L. H design for synchronisation of identical linear multi-agent systems. Int. J. Control. 2017, 91, 2214–2229. [Google Scholar] [CrossRef]
  25. Zhao, Y.; Chen, G. Distributed H consensus of multi-agent systems: A performance region-based approach. Int. J. Control. 2012, 85, 332–341. [Google Scholar] [CrossRef]
  26. Borrelli, F.; Keviczky, T. Distributed LQR design for identical dynamically decoupled systems. IEEE Trans. Autom. Control 2008, 53, 1901–1912. [Google Scholar] [CrossRef]
  27. Tomic, I.; Haliakis, G.D. Robustness Properties of Distributed Configurations in Multi-Agent Systems; IFAC-PapersOnLine; Elsevier: Amsterdam, The Netherlands, 2016; Volume 49, pp. 86–91. [Google Scholar]
  28. Stoica, A.-M. H Type Control for Multi Agent Systems Subject to Stochastic State Dependent Noise. SICON submitted for publication.
  29. Djouadi, S.M.; Charalambous, C.D.; Repperger, D.W. A convex programming approach to the multiobjective H2/H problem. In Proceedings of the 2002 American Control Conference, Anchorage, AL, USA, 8–10 May 2002; pp. 4315–4320. [Google Scholar]
  30. Jadbabaie, A.; Lin, J.; Morse, A.S. Coordination of groups of mobile autonomous agents using nearest neighbour rules. IEEE Trans. Autom. Control 2003, 48, 988–1001. [Google Scholar] [CrossRef] [Green Version]
  31. Olfati-Saber, R.; Murray, R. Agreement Problems in Networks with Directed Graphs and Switching Topology. In Proceedings of the 42nd IEEE International Conference on Decision and Control, Maui, HI, USA, 9–12 December 2003; pp. 4126–4132. [Google Scholar]
  32. Costa, O.L.V.; Fragoso, M.D.; Todorov, M.G. Continuous-Time Markov Jump Linear Systems; Springer: New York, NY, USA, 2013. [Google Scholar]
  33. Dragan, V.; Morozan, T.; Stoica, A.-M. Mathematical Methods in Robust Control of Linear Stochastic Systems; Springer: New York, NY, USA, 2006. [Google Scholar]
  34. Doyle, J.C.; Glover, K.; Khargonekar, P.P.; Francis, B. State-space solutions to standard H2 and H problems. IEEE Trans. Automat. Control 1989, 34, 831–847. [Google Scholar] [CrossRef]
  35. Khargonekar, P.P.; Rotea, M.A. Mixed H2/H control: A convex optimization approach. IEEE Trans. Automat. Control 1991, 36, 824–837. [Google Scholar] [CrossRef]
  36. Dragan, V.; Freiling, G.; Morozan, T.; Stoica, A.-M. Iterative Algorithms for Stabilizing Solutions of Game Theoretic Riccati Equations of Stochastic Control. In Proceedings of the Mathematical Theory of Networks and Systems (MTNS2008), Virginia Tech, Blacksburg, VA, USA, 28 July–1 August 2008. [Google Scholar]
Figure 1. Transition probabilities: P 11 ( t ) -solid line, P 12 ( t ) -dash-dotted line.
Figure 1. Transition probabilities: P 11 ( t ) -solid line, P 12 ( t ) -dash-dotted line.
Entropy 24 01734 g001
Figure 2. Snapshots for two different coupling weighting matrices of the agents.
Figure 2. Snapshots for two different coupling weighting matrices of the agents.
Entropy 24 01734 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Stoica, A.-M.; Stoicu, S.C. H State-Feedback Control of Multi-Agent Systems with Data Packet Dropout in the Communication Channels: A Markovian Approach. Entropy 2022, 24, 1734. https://doi.org/10.3390/e24121734

AMA Style

Stoica A-M, Stoicu SC. H State-Feedback Control of Multi-Agent Systems with Data Packet Dropout in the Communication Channels: A Markovian Approach. Entropy. 2022; 24(12):1734. https://doi.org/10.3390/e24121734

Chicago/Turabian Style

Stoica, Adrian-Mihail, and Serena Cristiana Stoicu. 2022. "H State-Feedback Control of Multi-Agent Systems with Data Packet Dropout in the Communication Channels: A Markovian Approach" Entropy 24, no. 12: 1734. https://doi.org/10.3390/e24121734

APA Style

Stoica, A. -M., & Stoicu, S. C. (2022). H State-Feedback Control of Multi-Agent Systems with Data Packet Dropout in the Communication Channels: A Markovian Approach. Entropy, 24(12), 1734. https://doi.org/10.3390/e24121734

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop