1. Introduction
In recent years, with the rapid development of communication and computing technology, people pay more and more attention to the control of multi-agent systems [
1,
2,
3]. “Agent” can be broadly defined as a computing entity that is adaptive, autonomous, and has a certain life cycle in a particular environment. Drones, driverless cars, robots, etc., can all be called “Agent”. Apart from these, single-agent has played an important role in aerospace, military, and real life. Furthermore, the development of intelligent control, modern communication technologies, artificial intelligence concepts, etc., as well as people inspired by the phenomenon of animal cluster activity in nature, have pushed agent control research into a new stage of development.
By working together, multi-agent can accomplish some parallel, complex tasks, more efficiently than single-agent. For instance, multiple satellites orbiting the Earth to complete designated tasks and multiple drones can work together to transport goods in disaster relief.
In recent years, many researchers have considered applying iterative learning control (ILC) to multi-agent systems with repetitive work. ILC is a tracking control method for systems that require high precision and repeat the same actions. In a study by Ahn and Chen [
4], ILC was applied to the formation control of a multi-agent system for the first time. Since then, more and more researchers have applied ILC to the practical application of multi-agent system. Ahn, Moore, and Chen [
5] proposed an ILC law to maintain the flight path of satellite formation. Since the satellite rotates the earth periodically, the position related interference can be considered as the time periodic interference. According to Chen and Jia [
6], multiple robots can guarantee formation using iterative learning. Furthermore, in order to ensure the safe running interval of trains, Sun, Hou, and Li [
7] studied the corresponding ILC scheme of multi train cooperation. In studies by Hock and Schoellig [
8] and Schoellig and colleagues [
9], a group of four rotor aircrafts kept a given formation and accurately tracked the desired trajectory with ILC.
However, considering the limited capacity of the communication channel, the actual control design based on wireless communication is difficult. In addition, various industrial production processes and equipment are very complex. For these reasons, communication delay, packet loss, quantization and so on will cause the ILC of multi-agent to not converge [
10,
11]. There are specific researches (see [
12,
13,
14,
15,
16,
17]) on communication delay and quantization, but few on communication packet loss. Zhang and Li [
18] designed an asynchronous event triggering protocol to solve the consistency problem of multi-agent system in the case of external interference, parameter uncertainty, time-delay, and packet loss by ΣΔ quantizer. By using the extended Lyapunov–Krasovskii functional, a consistency criterion in the form of matrix inequality was derived to ensure the consistency of multi-agent system with delay and packet loss under H1 controller. Firstly, Zhang and Li [
19] considered the packet loss problem in the quantitative ILC of a multi-agent system. Considering the nonlinearity of quantization and the randomness of packet loss, a method of reducing learning gain to obtain accurate tracking was proposed. However, the results showed that the convergence rate slows down with the increase of the packet loss rate. All the results from their study [
19] are for homogeneous multi-agent systems and assume that multiple agents have the same packet loss rate. The packet loss compensation for heterogeneous multi-agent systems with different packet loss rates is a more challenging problem.
This paper explores the ILC method with packet loss compensation for a heterogeneous multi-agent system. The question of under what conditions, with the increase of packet loss rate, the convergence speed of iterative learning algorithm will not slow down is solved. Firstly, the ILC model of a heterogeneous multi-agent system is established, and then the frequency domain model is obtained from the time domain model. When different communication packets are lost between agents, the previous iterative data is used to form the current iterative data. ILC in frequency domain is a two-dimensional system in which information propagates on two independent axes: frequency axis and iterative axis. The convergence of ILC is reduced to the convergence of any frequency in the iterative dimension, which is realized by the transformation of frequency and parameter iteration. After the exchange of variables and parameters, Z-transform is performed in the iterative domain. The convergence of the system error is analyzed by using the eigenvalue of the characteristic equation. Since the characteristic equation is in matrix form, it is difficult to get the eigenvalue of the characteristic equation. We use the generalized Nyquist criterion and Gershgorin’s disk theorem to analyze the eigenvalue range of the system and obtain the condition of the system convergence. The simulation results show that the compensation method proposed in this paper can make the system error convergence, and the convergence speed will not slow down.
This paper consists of the following parts. In
Section 2, an ILC model in frequency domain is established for heterogeneous multi-agent systems and the control objective for formation tracking of a multi-agent system is given. A dropout compensation ILC method is proposed in
Section 3. In
Section 4, the convergence conditions are analyzed. The results are verified by numerical simulation in
Section 5.
2. Problem Formulation and Preliminaries
2.1. Preliminaries
The interaction topologies in a multi-agent system are modeled by an undirected graph = (), where is the set of vertices, is the set of the edge, and is the weighted adjacency matrix of the graph . If , , otherwise, . Each edge represents information that can be transmitted between the th agent and the th agent. In addition, we assume that . The neighboring set of agent, , is . The Laplacian matrix of graph is denoted as , where with . A path in the undirected graph is denoted as a finite sequence and .
2.2. Problem Formulation
2.2.1. Control Objective
Consider the heterogeneous linear SISO (Single Input and Single Output) discrete-time multi-agent system of
n agents with topology graph
G. The dynamic of the
agent is modeled by the following dynamic system.
where
is the index of the
agent,
is the discrete-time index, and
is the index of iteration.
is the state of the
th agent in time
and
th iteration. Similarly,
is the control input signal and
is the output signal. Multiple agents have different dynamics which are denoted as
,
and
. Taking the Z-transformation of Equation (1), we get
where
with
and
is frequency, and
is the initial state of
th agent in the
th iteration; let
,
, then (2) can be rewritten as
The desired formation is defined as the relative state values between agents. The relative state values could be position, velocity, and so on. The control objective for the multi-agent systems in this paper is to find an appropriate control input sequence
such that every agent can move in the desired formation tracking trajectory accurately in the sense of expectation as the iterative number
tends to infinity. That is,
.
is the desired output of the
th agent and denoted as
where
,
is the desired reference trajectory of multi-agent systems.
can also be called the desired trajectory of a virtual leader. In practice, not all agents can obtain
, only some of them can. Let
denote the reference-accessibility matrix, which is a diagonal, non-negative, real matrix. If the
th agent can obtain direct information about the
, then
; otherwise,
. In (4),
is the desired output deviation from desired reference trajectory of the
th agent.
represents the desired relative formation between the
th agent and the
th agent. The formation tracking error
is denoted as
and the output error of the
th agent in
th iteration can be denoted as
By applying the Z-transform to Equation (5), we get
If the following condition (7) is met
that is,
, the multi-agent systems realize formation tracking control.
2.2.2. Structure of ILC for Networked Multi-Agent Systems
The following ILC law is used to achieve control objectives (7)
where
is the learning gain for
ith agent. The formulation of control input of the ith agent requires the control input of the last iteration, error, and information of adjacent agents. Due to the limited bandwidth of communication channel, noise interference, and signal fading, when a multi-agent transmits information through the wireless network, random packet loss is introduced into the multi-agent. For the communication graph of multi-agent systems shown in
Figure 1, a block diagram of ILC for multi-agent systems with loss of multiple communication packets is illustrated in
Figure 2. When the
th agent and the
th agent exchange information with each other through the wireless network, the
th agent receives a signal from the
th agent, which is denoted as
where
obeys Bernoulli distribution and
is the formation tracking error of the
th agent. When the
th agent sends a signal to the
th agent and there is no packet loss in the
th iteration, let
. Otherwise,
. Since the
th agent can obtain its own output
and the desired output deviation
without transmission network, there is no packet loss for the
th agent itself. That is,
does not have packet loss for the
th agent. It is worth noting that, unlike a previous study [
18], the scenario discussed in this paper is closer to the engineering practice, assuming that the packet loss rates of communication between agents are different.
3. Dropout Compensation ILC Method for Multi-Agent Systems with Loss of Multiple Communication Packets
In this section, firstly, because the iterative data are changing in the iterative domain, the previous iterative data are used to compensate for the missing data in the iterative process. Secondly, considering that ILC is a two-dimensional system, the convergence of any frequency in the iterative dimension is analyzed by variable iteration and parameter iteration. We make the following assumptions:
Assumption 1. The dynamics for each agent is stable, that is, , . Without loss of generality, the system relative degree is one, that is, .
Remark 1. If an agent’s dynamics is not stable, we can use the local feedback of its output to make it stable.
Assumption 2. The initial reset condition satisfies all agents and the desired input for each iteration, that is, can be abbreviated to and .
Remark 2. Assumption 2 is common in ILC of multi-agent systems (see [13,16,19]). If the initial reset condition is not met, we can regard it as the problem of robustness to the initial shifts. In our future work, the ILC with initial-state learning for formation tracking control of multi-agent systems will be explored. Assumption 3. The graph of multi-agent system (1) is an undirected graph. In addition, its graph is connected, that is, there is a path between any two agents so that they can exchange information.
Lemma 1. [
20]:
If an irreducible matrix is weakly generalized diagonally dominant and at least one of the rows is strictly diagonally dominant, is nonsingular.
Remark 3. Assumption 3 implies that the . In , the magnitude of the diagonal entry is and the sum of the magnitudes of all non-diagonal entries is . Since and at least one of the rows is satisfied , matrix is nonsingular by Lemma.
3.1. Dropout Compensate ILC Method
To achieve the control objective (7), we propose the ILC law to compensate the loss of packets.
When there are packet losses during multiple agents’ communication, we utilize the previous iterative data to compensate this iterative data. In (10),
is data which the
ith agent received from the
jth agent
where
is mentioned above. In practice, we can utilize a register to store each iterative data. If there is no packet loss, the next iterative data replace the old data. Otherwise, the old data are retained.
Remark 4. In order to make the multi-agent system closer to the practical engineering applications, could be different for different agents. That means the multi-agent system is heterogeneous in this paper.
Since
, (6) can be rewritten as
By combining the Z-transform of (10) with (12), we get
where
. According to Equations (3), (11)–(13) and Assumption 2, we have
Let
,
,
,
. Let
,
, then (14) can be rewritten as
where
and
,
The relationship among the (k − 1)th, the kth, and the (k + 1)th system error is shown in (15). Our goal is to converge the system error when the number of iterations tends to infinity. Thus, (15) will be further analyzed in the iterative domain.
3.2. Analysis in Iterative Domain
ILC is a method to control systems operating in a repetitive mode over a finite interval with the requirement that the system accurately follows the desired output trajectory. The same actions are executed again and again. By using the previously executed information, the tracking accuracy is improved from repetition to repetition. It has long been recognized that ILC is a two-dimensional (2D) system in which information travels on two independent axes: the time axis and the iteration axis. Chow and Fang [
21,
22] proposed ILC 2D analysis methods for discrete-time and continue-time systems. After Z-transformation of system dynamics, ILC is considered as a 2D system in frequency domain and iteration domain in this paper. The following analysis is based on the decomposition of 2D dynamics into two separate 1D problems.
is considered as a 2D function
. Exchanging variable frequency and parameter iteration, the convergence of the system is considered as a 1D problem in the iteration domain. Thus, we exchange variable
and parameter
.
can be rewritten as
, that is,
,
can be rewritten as
. Then, (15) can be rewritten as
where
and
,
Take Z-transformation for (17)
where
and
,
(19) leads to
where
,
From Equation (21), we can obtain characteristic equation of the system in iterative domain. In the next section, the convergence of the learning system is discussed based on the characteristic equation.
4. Convergence Analysis
As we know, when we say that an iterative learning algorithm converges, it means that the system error trends to zero as the number of iterations approaches infinity. For the proposed dropout compensation ILC method, we will analyze the convergence of formation tracking for the multi-agent systems in this section.
Theorem 1. Consider the multi-agent system (1), let Assumptions 1–3 hold and the learning law (10) be applied. Given that any one of the following conditions is satisfied for all ,wherethen the formation tracking objective (7) is achieved.
Proof. Let . The characteristic equation is . When , . Thus, . In Remark 3, proves to be nonsingular. Therefore, matrix is nonsingular also and . □
When
, we need to prove that the modules of all roots of (25) are smaller than units.
Let
, based on the general Nyquist stability criterion [
23], the modules of all roots of (25) are smaller than units, if the eigenloci
of
does not enclose the point
for
,
and
. Based on Greshgorin’s disk theorem, we have
for all
,
and
, where
Since
can only be
or
,
is
or
respectively. Taking
as
, we can rewrite (27) as
So the eigenloci
does not enclose the point
for
,
and
when the point
with
is not in the disc
for all
,
,
and
. That is,
for all
,
,
and
, when
. We denote
as
Note that
for all
,
,
and
. So,
as long as
for all
,
,
and
. Further, we can show that
Using the conditions of Theorem 1, we can prove
for all
,
,
and
(see Appendix of [
17]). Then the eigenloci of
for all
,
and
does not enclose the point
. The modules of all roots of (25) are smaller than units. That is, the system achieves a consensus asymptotically.
Multiply
on both sides of Equation (21) and take the limit
According to the final value theorem, we get
As the previous certificate, . Thus, . Based on Sylvester inequality, we get . Thus and the final value of is . Theorem 1 is thus proved.
When considering random communication packet dropout, the ILC algorithm can converge using the compensation method we proposed. When considering the random packet loss, the ILC algorithm can use this proposed compensation method to converge. Theorem 1 gives the convergence condition of learning gain design and proves that the convergence condition is independent of communication packet loss. According to the final value theorem, the error of formation tracking converges to zero as the number of iterations goes to infinity.
5. Simulation
Firstly, consider the multi-agent system in
Section 2.2.1. and let Assumptions 1–3 hold. Next, the learning law (10) is applied and the learning gains satisfy the condition (23). The results of numerical simulations are discussed in this section.
In order to illustrate the effectiveness of the proposed dropout compensation ILC method, we considered a heterogeneous multi-agent system consisting of four agents in the undirected graph as shown in
Figure 3. It can be seen that the graph is a connected graph and satisfies Assumption 3. Only the second agent and the third agent can obtain the desired reference trajectory.
The Laplacian matrix
of graph
and the reference-accessibility matrix
were
Then it was obtained that the information interaction weight between agents can be obtained, as shown in
Table 1.
For every agent in
Figure 3, the system parameters were defined in
Table 2.
Let
,
,
and
. Then, we can obtain the parameters in
Table 3 for Theorem 1.
Therefore, the learning gains satisfied the second condition in Theorem 1. To perform our simulation tests, we considered the desired reference trajectory as
and the desired output deviation of each agent as
Set the initial state of four agents as:
,
,
, and
, which satisfies Assumption 2. When the iteration step is 500 and
, the trajectory of the multi-agent system without packet loss is shown as
Figure 4. It can be seen that the desired formation is well achieved to track the desired reference trajectory by the desired output deviation.
In order to measure the formation accuracy quantitatively, the disagreement among all agents on their output errors was defined as:
. When
, multi-agent system achieved formation tacking.
Figure 5 shows that formation performance of the system at the first 500 iterations. Obviously, when the iteration is around 350 the system error can converge to zero and achieve formation tracking.
It is assumed that there is no compensation when considering communication packet loss, that is
when there is no packet loss, otherwise,
. In
Figure 6, it can be seen that the system error cannot converge to zero in different packet loss rates as iteration increases. When iteration is 500 and packet dropout rate is 10%, the trajectory of the multi-agent system without packet loss compensation is shown as
Figure 7.
When there are different packet loss rates in a multi-agent system (
Figure 8), it can be seen that system errors can converge to zero using (10) to compensate packet loss. The convergence speed hardly gets slower as the packet loss rates increase, as shown in
Figure 8b. In addition, the multi-agent system error still can converge to zero around the 350th iteration. When iteration is 350 and packet loss rate is 10%, the trajectory of the multi-agent system with packet loss compensation is shown as in
Figure 9.
In order to demonstrate the effectiveness of proposed method, the comparison results with the method of Zhang and Li [
19] are shown in
Figure 10. If there is data loss, the information for the current iteration will be discarded without any compensation. Therefore, although the error curve can converge to 0, the convergence speed gets slower as the packet loss rate increases [
19], while the convergence speed hardly gets slower, as shown in
Figure 8b, by the proposed method.