1. Introduction
Stochastic optimization has attracted significant attention in the recent control literature, due to its ability to provide an alternative, often less conservative way to handle uncertainty in dynamical systems. Stochastic optimization takes into account the stochastic characteristics of the uncertainties and thereby the system constraints are treated in a probabilistic sense, i.e., using chance constraints [
1]. Using stochastic optimization in the so-called model predictive control (MPC), one computes an optimal control sequence that minimizes a given objective function subject to the uncertain system dynamics model and chance constraints in a receding horizon fashion [
2]. Chance constraints enable stochastic optimization to offer an alternative approach to achieve a less conservative solution compared to robust optimization [
3], since it directly incorporates the trade-off between constraint feasibility and performance of objective function.
Distributed optimization has been an active research area in recent decades, due to its applicability to handle large-scale problems with constraints. In distributed optimization one replaces large-scale problems stemming from centralized optimization with several smaller-scale problems that can be solved in parallel. These problems make use of information from other subproblems to formulate a local optimization problem. In the presence of uncertainties, however, the main challenge in the formulation of distributed optimization is how one should exchange information through a communication scheme among subproblems (see, e.g., [
4], and references therein). This highlights the necessity of developing distributed strategies to cope with the uncertainties in subproblems while at the same time minimizing information exchange through a communication framework.
To handle uncertainties in distributed setting, some approaches are based on robustification [
5]. Assuming that the uncertainty is bounded, a robust optimization problem is solved, leading to a decision that satisfies the constraints for all admissible values of the uncertainty. The resulting solution using such an approach tends to be conservative in many cases. Tube-based robust optimization, see for example [
6] and the references therein, was considered in a decentralized setup (plug-and-play) in [
7], and it has been recently extended to distributed control systems [
8] for a collection of linear stochastic subsystems with independent dynamics.
1.1. Related Works
Although in [
8] coupled chance constraints were considered separately at each sampling time, in this paper we consider a chance constraint on the feasibility of trajectories of dynamically coupled subsystems. Our approach is motivated by [
7] to reduce the conservativeness of the control design. Other representative approaches for stochastic optimization of a single stochastic system include affine parametrization of the control policy [
9], the randomized (scenario) approach [
10,
11,
12,
13], and the combined randomized and robust approach [
14,
15,
16]. None of these approaches, to the best of our knowledge, have been considered in a distributed control setting.
This paper aims to develop a systematic approach to distributed stochastic optimization using the scenario-based technique. Scenario optimization approximates stochastic problem via the so-called scenario (sample) approach [
17], and if the underlying optimization problem is convex with respect to the decision variables, finite sample guarantees can be provided. Following such an approach, the computation time for a realistic large-scale problem of interest becomes prohibitive, because the number of samples to be extracted tends to be high, and consequently leads to many constraints in the resulting optimization problem.
To overcome the computational burden caused by the large number of constraints, in [
18,
19] a heuristic sample-based approach was used in an iterative distributed fashion via dual decomposition such that all subsystems collaboratively optimize a global performance index. In another interesting work [
20], a multi-agent consensus algorithm was presented to achieve consensus on a common value of the decision vector subject to random constraints such that a probabilistic bound on the tails of the consensus violation was also established. However, in most of the aforementioned references the aim to reduce communication among subsystems, which we refer to as agents, has not been addressed.
1.2. Contributions
Our work in this paper differs from the above references in two important aspects which have not been, to the best of our knowledge, considered in the literature. We develop two different distributed scenario optimization frameworks. In the first version, each agent generates its own scenarios of the local uncertainty sources, and takes into consideration an estimation of the neighboring possible state trajectories, such that using the so-called alternating direction method of multipliers (ADMM), all the agents start to improve their local estimation until they agree on the consistency of exchange scenarios in the proposed distributed setting. In the second version, we develop a new set-based communication setup together with the notion of probabilistically reliable information to reduce the communications between agents required by the proposed first version of distributed scenario optimization framework. To quantify the error introduced by such a new communication scheme, we incorporate the probabilistic reliability notion into existing results and provide new guarantees for the desired level of constraint violations. To summarize the main contributions of this paper are as follows:
(a) By reformulating a centralized scenario optimization problem into a decomposable scenario program, we first present the exactness of decomposition and then, quantify the level of robustness of resulting solutions by providing new a-priori probabilistic guarantees for the desired level of constraint fulfillment under some mild conditions.
(b) To solve the proposed decomposable scenario program, we develop a so-called distributed scenario exchange scheme between subproblems (neighboring agents) using the alternating direction method of multipliers (ADMM). Such a new scheme is used to provide a distributed scenario optimization framework to handle two types of large-scale network problems in control literature, namely dynamically coupled and coupling constraints network of agents.
(c) To reduce the communication in the proposed distributed setting, we develop a novel inter-agent soft communication scheme based on a set parametrization technique together with the notion of probabilistically reliable set to reduce the required communication between the subproblems. We show how to incorporate the probabilistic reliability notion into existing results and provide new guarantees for the desired level of constraint violations.
(d) Using the so-called soft communication scheme, we present a distributed framework such that the neighboring agents just need to communication a set together with the level of reliability of the set. Each agent then solves a local robust-communication scenario program at each sampling time. We also establish a practical plug-and-play (PnP) distributed framework that considers network changes by agents which want to join or leave the network.
The structure of this paper is as follows.
Section 2 describes a problem formulation leading to a large-scale scenario optimization problem. In
Section 3, we first reformulate the large-scale scenario problem into a decomposable scenario problem. In
Section 4, we then provide the results on equivalent relations and analyze robustness of the obtained solutions using the proposed distributed scenario exchange scheme.
Section 5 introduces a novel inter-agent communication scheme between the subproblems, namely
soft communications, and then proceeds to quantify the robustness of the proposed schemes. Using the proposed soft communication scheme, in
Section 6 we establish a practical PnP distributed framework.
Section 7 presents two different simulation studies, whereas in
Section 8, we conclude this paper with some remarks and future work.
2. Problem Formulation
Without loss of generality, this section presents a mathematical problem formulation in the context of advanced control problem, namely stochastic model predictive control (SMPC), leading to a large-scale scenario optimization problem at each time step. Please note that in this paper we focus on the large-scale scenario problem instance from the optimization point of view and derive probabilistic guarantees for constraint fulfillment in a distributed setting. Analyzing the closed-loop asymptotic behavior of SMPC (e.g., stability and recursive feasibility) are not in the scope of this paper.
Consider a discrete-time uncertain linear system with additive disturbance in a compact form as follows:
with a given initial condition
. Here
denotes the time instance,
and
correspond to the state and control input, respectively, and
represents an additive disturbance. The system matrices
and
as well as
are random, since they are known functions of an uncertain variable
that influences the system parameters at each time step
k. It is important to mention that given the initial measured state variable
, the predicted future states are modeled using (
1) at each sampling time
k. For avoiding crowded notation, when there is no ambiguity, in the rest of the paper we will drop the conventional index for the predicted time steps, i.e.,
, and simply use
. We also define
to denote the future predicted time steps.
Throughout this paper, there are three technical assumptions. Assumption 1 is a classical assumption where we define the probability spaces associated with the random variables and highlight the fact that in this paper we only require the availability of a “sufficient number” of samples of such random variables. Using such samples to formulate an optimization problem, Assumption 2 is given as a technical requirement to be able to use a so-called randomization technique and to guarantee a nonempty feasibility domain of such an optimization problem. Assumption 3 is focused on the structure of large-scale system dynamics of interest to be able to explore and exploit its structure for decomposition into an interconnection of smaller subsystems.
Assumption 1. Random variables and are defined on probability spaces and , respectively. w and are two independent random processes, where and are two different probability measures defined over and Δ, respectively, and denotes a Borel σ-algebra. The support sets and Δ of w and , respectively, together with their probability measures and are entirely generic. In fact, , Δ and , do not need to be known explicitly. Instead, the only requirement is availability of a “sufficient number” of samples, which will become concrete in later parts of the paper. Such samples can be for instance obtained by a model learned from available historical data [21]. The system in (
1) is subject to constraints on the system state trajectories and control input. Consider the state and control input constraint sets to be compact convex in the following form
where
,
, and
,
. Keeping the state inside a feasible set
for the entire prediction horizon may be too conservative and result in loss of performance. In particular, this is the case when the best performance is achieved close to the boundary of
, and thus, constraint violations will be unavoidable due to the fact that the system parameters in (
1) are imperfect and uncertain. To tackle such a problem, we will consider chance constraints on the state trajectories to avoid violation of the state variable constraints probabilistically even if the disturbance
w or uncertainty
has unbounded support. Notice that a robust problem formulation [
3] cannot cope with problems having an unbounded disturbance set.
Following the traditional MPC approach to find a stabilizing full-information controller that leads to admissible control inputs
and satisfies the state constraints, one can rely on the standard assumption of the existence of a suitable prestabilizing control law, see, e.g., (Proposition 1, [
7]). Based on [
9], one can also employ a so-called parametrized feedback policy and split the control input to cope with the state prediction under uncertainties:
with
as a free correction input variable to compensate for disturbances and
K is a given pre-stabilizing feedback gain for the nominal (uncertain-free) system (
1).
The control objective is to minimize a cumulative quadratic stage cost of a finite horizon cost
that is defined as follows:
with
, and
. Consider
,
to be detectable and
P to be the solution of the discrete-time Lyapunov equation:
for the closed-loop system with
. Each stage cost term is taken in expectation
, since the argument
is a random variable. Using
, consider now the following stochastic control (optimization) problem:
where
is initialized based on the measured current state, and
is the admissible state constraint violation parameter of the large-scale system (
1). The state trajectory
, has a dependency on the random variables
w and
, and thus, the chance constraint can be interpreted as follows: the probability of violating the state constraint at the future time step
is restricted to
, given that the state of the system in (
1) is measurable at each time step
. Even though
and
are compact convex sets, due to the chance constraint on the state trajectory, the feasible set of the optimization problem in (5) is a non-convex set, in general.
To handle the chance constraint (5c), we recall a scenario-based approximation [
22].
and
at each sampling time
are not necessarily independent and identically distributed (i.i.d.). In particular, they may have time-varying distributions and/or be correlated in time. We assume that a “sufficient number” of i.i.d. samples of the disturbance
and
can be obtained either empirically or by a random number generator. We denote the sets of given finite samples (scenarios) with
and
, respectively. Following the approach in [
23], we approximate the expected value of the objective function empirically by averaging the value of its argument for some number of different scenarios, which plays a tuning parameter role. Using
as the tuning parameter, consider
number of different scenarios of
w and
to build
which has the cardinality
. We then approximate the cost function empirically as follows:
where
represents a compact notation of the objective function
. Notice that
indicates the dependency of the state variables on the random variables.
We are now in a position to formulate an approximated version of the proposed stochastic optimization problem in (5) using the following finite horizon scenario program:
where superscript
s indicates a particular sample realization. The solution of (6) is the optimal control input sequence
. Based on the MPC paradigm, the current input is implemented as
and we proceed in a receding horizon fashion. This means that the problem (6) is solved at each time step
k by using the current measurement of the state
. Please note that new scenarios are needed to be generated at each sampling time
. It is important to note that the proposed large-scale scenario optimization problem (6) should be feasible and its feasibility domain must be a nonempty domain. This is a technical requirement (Assumption 1, [
17]) to be able to use such a randomization technique. In case of infeasible solution, we need to generate new set of scenarios and resolve the problem.
Assumption 2. The proposed large-scale scenario optimization problem (6) is feasible and its feasibility domain has a nonempty domain.
Following Assumption 2 and based on the Weierstrass’ theorem (Proposition A.8, [
24]), the proposed large-scale scenario program (6) admits at least one optimal solution, and therefore the Slater’s constraint qualification [
25] holds for the proposed problem formulations. The key features of the optimization problem (6) are as follows: (1) there is no need to know the probability measures
and
explicitly, only the capability of obtaining random scenarios is enough, (2) formal results to quantify the level of approximations are available. In particular, the results follow the so-called scenario approach [
17], which allows the binding a-priori of the violation probability of the obtained solution via (6).
In the following theorem, we restate the explicit theoretical bound of (Theorem 1, [
17]), which measures the finite scenarios behavior of (6).
Theorem 1. Let and , whereIf the optimizer of problem (6)
is applied to the discrete-time dynamical system (
1)
for a finite horizon of length T, then, with at least confidence , the original constraint (
5c)
is satisfied for all with probability more than . It was shown in [
17] that the above bound is tight. The interpretation of Theorem 1 is as follows: when applying
in a finite horizon control problem, the probability of constraint violation of the state trajectory remains below
with confidence
:
with
where
. It is worth mentioning that while the proposed constraint on the control input (
6d) is also met in a probabilistic sense for all prediction time steps, except at the initial time step, due to the feedback parametrization and the nature of the scenario approach that appears in the proposed optimization problem (6). At the initial time step such a constraint (
6d) is deterministic, and the superscript
s can be dropped, as there is only one measured current state
. This holds for all the following proposed formulations.
3. Decomposed Problem Reformulation
Consider a partitioning of the system dynamics (
1) through a decomposition into
N subsystems and let
be the set of subsystem indices. The state variables
, control input signals
and the additive disturbance
can be considered to be
,
and
, respectively, where
,
,
, and
,
,
. The following assumption is important to be able to partition the system parameters.
Assumption 3. The control input and the additive disturbance of the subsystems are decoupled, e.g., and only affect subsystem for all . Consider the state and control input constraint sets and of large-scale system dynamics (
1)
as defined in (
2)
to be , and such that and for all subsystems can be in the following form:where , , and . Also consider the objective function of each subsystem in the following form:where , such that , and . Please note that and such that and . It is important to note that under Assumption 3 there are no direct coupling constraints between each subsystem
and its neighboring subsystems
. Instead, the subsystem
is dynamically coupled with all its neighboring subsystems
as it is presented in (
8).
We refer to the additive disturbance
as a local uncertainty source of each subsystem
, since it is assumed that it affects only the subsystem
. Motivated by an application to Smart Thermal Grids (STGs) in [
26] where a network of agents is interconnected via an uncertain common resource pool between neighbors, we consider the uncertain variable
as a common uncertainty source between all subsystems
. Such a STG application will also be presented in
Section 7.2 as a second case study. Observe the fact that every common uncertain phenomenon can be considered to be a local uncertain variable, e.g., the outside weather condition for neighboring buildings. Therefore, we also consider having
and refer to both random variables
and
as local uncertainty sources.
We are now able to decompose the large-scale system matrices
∈
,
, and consider
in the following form:
where
,
, and
. Define the set of neighboring subsystems of subsystem
i as follows:
where 0 denotes a matrix of all zeros with proper dimension. Please note that if subsystems are decoupled, they remain decoupled regardless of the uncertainties
for all
. Consider now a large-scale network that consists of
N interconnected subsystems, and each subsystem can be described by an uncertain discrete-time linear time-invariant system with additive disturbance of the form
Following Assumption 3, one can consider a linear feedback gain matrix
for each subsystem
such that
. Using
in each subsystem, we assume that there exists
for each subsystem
such that
preserves the condition in (
4).
Define
such that
, where
, to be auxiliary decision variables, and using
such that
, we are now able to reformulate the optimization problem in (6) into the following decomposable finite horizon scenario program:
where
and
such that
and
. The sets
and
denote sets of given finite samples (scenarios) of disturbance and uncertainties in each subsystem
, such that
and
. It is important to highlight that the proposed optimization problem in (9) is decomposable into
N subproblems with coupling only through the equality constraints (
9d) between neighboring agents. The auxiliary variables
are defined between agent
and its neighboring agent
to represent the local estimation of each scenario information of the neighbor’s state variable
at each sampling time
. By taking into consideration that the interaction dynamics model
by each neighboring agent
is available for agent
, the only information that subsystem
needs is a set of scenarios of the state variable
at each sampling time
from all its neighboring agents
. Please note that the cardinality of
is
, which is assigned by the agent
for all neighboring agents
, and should be sent by the neighboring agents at each sampling time
, and this is known as the
hard communication scheme between the neighboring agents.
In the following proposition, we provide a connection between the proposed optimization problem in (9) and the optimization problem in (6).
Proposition 1. Given Assumption 3 and the block-diagonal structure for the state-feedback controller for the large-scale system dynamics (
1)
, the optimization problem in (9)
is an exact decomposition of the optimization problem in (6)
, such that the optimal objective value of the decomposed problem (9)
is equal to the optimal objective of the problem (6).
An important key feature of the proposed decomposable scenario program in (9) compared to the centralized scenario problem in (6) is as follows. Such a decomposition yields a reduction of computation time complexity of the centralized scenario program compared to the decomposable scenario program (9). This however requires more communication between each subsystem, since at each time each agent should send to and receive from all the neighboring agents a set of their state variable scenarios and , respectively.
Remark 1. The proposed constraint (
9e)
represents an approximation of the following chance constraint on the state of each subsystem :where is the admissible state constraint violation parameter of each subsystem (
8)
. One can also consider as the desired level of state feasibility parameter of each subsystem (
8)
. The following theorem can be considered to be the main result of this section to quantify the robustness of the solutions obtained by (9).
Theorem 2. Let be chosen such that , , and for all subsystem . If , the collection of the optimizers of problem (9)
for all subsystem , is applied to the discrete-time dynamical system (
1)
for a finite horizon of length T, then, with at least confidence , the original constraint (
5c)
is satisfied for all with probability more than . The interpretation of Theorem 2 is as follows. In the proposed decomposable scenario program (9), each subsystem
can have a desired level of constraint violation
and a desired level of confidence level
. To keep the robustness level of the collection of solutions in a probabilistic sense (
5c) for the discrete-time dynamical system (
1), these choices have to follow a certain design rule, e.g.,
and
. This yields a fixed
for the large-scale system (
1) and the individual
for each subsystem
. It is important to mention that to maintain the violation level for the large-scale system with many partitions, i.e.,
the violation level of individual agents needs to decrease, i.e.,
which may lead to conservative results for each subsystem, since the number of samples needs to increase, i.e.,
. Addressing such a limitation is subject of ongoing research work. The following corollary is a direct result of Theorem 2.
Corollary 1. If the optimal solution for each agent obtained via the proposed decomposable problem (9)
is applied to the agent dynamical system (
8)
for a finite horizon of length T, then, with at least confidence , the original constraint (
10)
is satisfied for all with probability more than . 4. Distributed Scenario Optimization
In this section, we continue by developing a distributed framework using ADMM to solve the proposed decomposable formulation in (9). It has been proven that ADMM for this type of problem converges linearly [
27]. We follow a similar approach as in [
28,
29] to solve the decomposable problem (9) in a distributed manner by extending to the scenario program. Define the local feasibility set denoted with
for all agents
:
where all the system parameters together with the local scenario sets,
, are fixed and available locally for each agent
at each sampling time
. We are now able to define the augmented Lagrangian function as follows:
where
is a convex indicator function for the local constraint set
that maps to infinity if one of the constraints is violated, and to zero otherwise. Step size
is a fixed constant, and multipliers
are introduced for every coupling constraint of each scenario. The indicator function makes sure all local constraints are satisfied, and the squared norm penalty term forces constraint (
9d) to be satisfied.
We now describe the steps of the ADMM algorithm as follows:
- (1)
Update primal variables
The multipliers
and the local estimation of neighboring state scenario
for all scenarios
and for all the neighbors
are fixed at their value of the previous iteration. Consider the following minimization problem for all
:
It is important to note that the augmented part of the aforementioned step in each agent
refers to the coupling constraints from the neighboring agent
point of view and
is the estimation of neighboring agent
from the state variable scenario of agent
for all
. Since the minimization is only in
, all terms of
drop out. This results in
separate convex programs.
- (2)
Update and exchange the set of state scenarios
The resulting primarily decision variables
for all
are used to first update and then exchange the set of state scenarios. Please note that each agent only needs to communicate the updated local state scenarios set to all its neighboring agents and receive the updated set of all its neighboring state scenarios. Consider now the following steps which we refer to as the Scenario Updating Steps (SUS) for a generic agent
as follows:
By repeating SUS for all local scenarios,
and
, the agent
should first update the set of its local state scenarios
and send it to all its neighboring agents. Then, the agent
should receive the updated set of all its neighboring state scenarios
. It is important to note that the set of local scenarios,
and
, is fixed for all the SUS steps and iteration steps, i.e.,
.
- (3)
Update and exchange the state estimation variables
Using the updated set of all neighboring state scenarios
and the updated primal variables, consider the following projection problem for all
:
Since the projection is only in
, all terms of
drop out and this results in
separate convex programs. Each agent
should now send the updated local estimation of the neighboring state variable
to its neighbor
.
- (4)
Update and exchange the multiplier variables
The multipliers are updated as follows
, and
:
Notice that no information exchange is needed for the update of the multiplier, since agents receive the updated set of all their neighboring state scenarios
. As the final step, each agent
should receive the updated multiplier variable
from all its neighboring agents
for all the scenarios
.
In Algorithm 1, we summarize the proposed distributed scenario exchange framework using ADMM algorithm to illustrate the calculation and communication steps for each agent . Each agent needs to solve two small-scale scenario programs in Lines 4 and 8 at each iteration that can be considered to be the highest computational cost in the proposed algorithm. In Line 5 each agent first updates using SUS, which works with only simple operations, e.g., addition, subtraction and scaling, and then, it broadcasts and receives the updated local and to and from the neighbors, respectively. Finally, each agent in Lines 9 and 11 broadcasts the updated local estimation from the neighboring agents state variable scenarios and the related multipliers.
Algorithm 1 Distributed Scenario Exchange Algorithm |
- 1:
for all
do - 2:
initialize: , , , , , , and - 3:
while using ( 15) do - 4:
Update using ( 11) - 5:
Update using SUS procedure ( 12) - 6:
Broadcast to all - 7:
Receive from all - 8:
Update using ( 13) for all and for all , - 9:
Broadcast to all - 10:
Update using ( 14) for all and for all - 11:
Broadcast to all and for all - 12:
set - 13:
output:
|
We define the agreement on the local estimation from the state variable scenario of the neighbors
as the convergence criteria of the proposed ADMM algorithm for each agent
as follows:
and the overall convergence criteria is
. If the residual sequence
is sufficiently small, less than or equal to
where
is the desired level of convergence error for each agent
, then all neighboring agents have reached a consistent set of state variable scenarios. In light of Assumption 2, we can now provide convergence of Algorithm 1 based on the results in [
30].
Theorem 3. Assume that Slater’s condition [25] holds for the decomposable scenario optimization problem (9)
, and consider the iterative steps given in Algorithm 1. Then the following statements hold: Proof. The theorem follows from [
30] that studies the convergence of a standard ADMM problem. The details are omitted for brevity. □
Remark 2. The proposed distributed scenario exchange algorithm uses a Gauss–Siedel update on the primal variables and the set of state variable scenarios, after which the multiplier variables are updated. Since either the primal or the set of state variable scenarios are fixed in the Gauss–Siedel steps, the problem can be distributed for all the agents. The main advantage to distribute a large-scale scenario optimization problem (6) is the ability of finding local solutions for each agent based on the information received in the previous iteration. Such calculations can therefore be carried out in parallel. Although an actual parallel implementation is outside the scope of this work, it is important to mention that the proposed algorithm is amenable to such an implementation. ADMM algorithms typically need a large number of iterations to converge to high accuracy, so the local agent problems need to be solved many times before finding a good enough solution. Thus, the ADMM approach without parallelization might not be the quickest method to solve the large-scale scenario program. Such a distributed framework is advantageous especially when the global scenario optimization problem is too large and cannot be solved within polynomial time due to the curse of dimensionality or memory limitations and computational constraints.
We next present the overall steps needed to execute the proposed distributed scenario optimization. Algorithm 2 summarizes all the steps such that after decomposing the largescale dynamical system (
1) and determining the index sets of the neighboring agents, each agent first starts to generate some scenarios of its local uncertainty sources to approximate its cost function and achieve the feasibility of its chance constraint with high confidence level following Theorem 2. Next, all the agents solve their own local problem using the proposed distributed scenario exchange scheme (Algorithm 1) to obtain their local optimal solution. To execute Algorithm 1, it is assumed that the feedback control gain matrices
for all agents
are given (
4), and the coupling terms
are known between each agent
and its neighboring agents
. Using the obtained solution via Algorithm 1, each agent generates a new set of local state variable scenarios to send and receive to and from all its neighboring agents, respectively. Finally, the first optimal control input is applied to the real system and the new state variables are measured, and proceed in the classical receding horizon fashion.
Remark 3. The proposed decomposable scenario program in (9)
is a general deterministic optimization problem for each agent that are coupled via (
9d)
. Therefore, the proposed technique to solve (9)
in Algorithm 1, namely the distributed scenario exchange scheme, can be easily considered to be a general solution method to solve the agents’ problem with the nonempty intersection between their local feasible sets. Such a case where agents are only coupled via coupling constraints has been presented in Section 7.2 as a second case study. Algorithm 2 Distributed Scenario Optimization (DSO) |
- 1:
Decompose the large-scale dynamical system ( 1) into N agents as the proposed form in ( 8) - 2:
Determine the index set of neighboring agents for each agent - 3:
for all agent do - 4:
fix initial state , , and such that - 5:
determine to approximate the objective function, and following Theorem 2 to approximate the chance constraint ( 10) in an equivalent sense - 6:
generate , scenarios of , to determine the sets of and , - 7:
solve the proposed optimization problem in (9) using the proposed distributed scenario exchange algorithm (Algorithm 1) and determine - 8:
generate scenarios of using SUS procedure ( 12) and taking into consideration and , - 9:
send the set to all neighboring agents - 10:
receive the sets from all neighboring agents - 11:
apply the first input of solution into the uncertain subsystem ( 8) - 12:
measure the state and substitute it as the initial state of the next step - 13:
set - 14:
go to Step (4)
|
5. Information Exchange Scheme
When the proposed distributed scenario optimization framework (Algorithm 2) is applied to the large-scale scenario program (6), all neighboring agent of the agent should send a set of scenarios of the state variable to agent i at each sampling time following the proposed distributed scenario exchange scheme in Algorithm 1. Based on ADMM in Algorithm 1, it may require a large number of iterations to achieve an agreement between neighboring agents on the consistency of exchange scenarios, which may also turn out to be too costly in terms of required communication bandwidth. To address this shortcoming, we propose a set-based information exchange scheme which is referred to as a soft communication protocol.
We now describe the soft communication protocol in more detail. The neighboring agent
must first generate
samples of
in order to build the set
. It is important to notice that in the soft communication protocol the number
of samples generated by agent
j may be different than the one needed by agent
i, which is
, as will be remarked later. Let us then introduce
as a bounded set containing all the elements of
. We assume for simplicity that
is an axis-aligned hyper-rectangular set [
14]. This is not a restrictive assumption and any convex set, e.g., ellipsoids and polytopes, could have been chosen instead as described in [
26]. We can define
as an interval, where the vector
defines the hyper-rectangle bounds.
Consider now the following optimization problem that aims to determine the set
with minimal volume:
where
is the number of samples
that neighboring agent
j must take into account in order to determine
. If we denote by
the optimal solution of (
16) computed by the neighbor agent
j, then for implementing the soft communication protocol, agent
j needs to communicate only the vector
along with the level of reliability
to the agent
i.
Definition 1. A set is called reliable ifand we refer to as the level of reliability of the set . We now provide the following theorem to determine as the level of reliability of the set .
Theorem 4. Fix and letwhere is the degree of the root. We then have Theorem 4 implies that given an hypothetical new sample
, agent
has a confidence of at least
that the probability of
is at least
. Therefore, one can
rely on up to probability. The number of samples
in the proposed formulation (
16) is a design parameter chosen by the neighboring agent
. We however remark that one can also set a given
as the desired level of reliability and obtain from (
18) the required number of samples
.
When an agent and its neighbor adopt the soft communication scheme, there is an important effect on the probabilistic feasibility of agent i, following Remark 1. Such a scheme introduces some level of stochasticity on the probabilistic feasibility of agent i, due to the fact that the neighboring information is only probabilistically reliable. This will affect the local probabilistic robustness guarantee of feasibility as discussed in Theorem 2 and consequently in Theorem 1. To accommodate the level of reliability of neighboring information, we need to marginalize a joint cumulative distribution function (cdf) probability of and the generic sample appearing in Theorem 4. We thus have the following theorem, which can be regarded as the main theoretical result of this section.
Theorem 5. Given for all and a fixed , the state trajectory of a generic agent is probabilistically –feasible for all , i.e., where such that . Following the statement of Theorem 5, it is straightforward to observe that if for all neighboring agents
,
then
. This means that if
the level of reliability of the neighboring information is one, i.e.,
, then, the state feasibility of agent
i will have the same probabilistic level of robustness as the hard communication scheme, i.e.,
. Combining this result with the statement of Theorem 2, the proposed soft communication scheme introduces some level of stochasticity on the feasibility of the large-scale system as in (
5c). In particular,
the level of constraint violation in each agent
will increase, since it is proportional with the inverse of
, and therefore,
will also increase. After receiving the parametrization of
and the level of reliability
, agent
should immunize itself against all possible variation of
by taking the worst-case of
, similar to the worst-case reformulation proposed in (Proposition 1, [
16]). It is important to notice that in this way, we decoupled the sample generation of agent
from agent
.
6. Plug-and-Play Operational Framework
Using the soft communication scheme, each agent
should first receive the parametrization of
from all its neighboring agents
together with the level of reliability
. Then, all agents should immunize themselves against all possible variation of
. To this end, we formulate the following
robust-communication scenario program for each agent
:
It is important to highlight that we refer to the aforementioned formulation as the
robust-communication scenario program, since the communicated variable between neighboring agents should be taken into consideration by the worst-case of
. In this way, there is no need for the many iterations of Algorithm 2, and instead, a robustification against all possible variation of
is used. It is also important to mention that another feature of using the soft communication scheme is the relaxation of the condition on the required number of scenarios, e.g., each agent
requests
number of scenarios from all its neighboring agents
, which may give rise to privacy concerns for the neighboring agents. The proposed problem (21), specifically the robust constraint (
21e), can be solved using a robust (worst-case) reformulation similar to (Proposition 1, [
16]) and [
7]. It is important to highlight that based on Assumption 2, the proposed optimization problem (21) should be feasible and its feasibility domain has to be a nonempty domain. In case of infeasible solution, each agent
needs to generate new set of local scenarios,
, and also needs to receive a new set
from all its neighbors
.
We summarize the proposed distributed scenario optimization using soft communication scheme in Algorithm 3. Similar to Algorithm 2, in Algorithm 3 it is also assumed that the feedback control gain matrices
for all agents
are given (
4), and the coupling terms
are known between each agent
and its neighboring agents
. It is important to note that Step 5 of Algorithm 3, initializes
of all neighboring agents
, which is only used for the initial iteration in Step 8, and then, at each iteration all agents
will send and receive
from all its neighboring agents
as in Steps 11 and 12, respectively.
We also summarize the steps that are needed for plug-in and plug-out operations of each agent
in Algorithm 4. Please note that in a plugged-in or plugged-out operation all agents
have to update their
with
to respect the condition in Theorem 2 as in (4) to achieve the desired level of constraint feasibility for the large-scale system (
1). One can also redesign
to potentially improve the local control performance of each agent
.
Algorithm 3 DSO using Soft Communication Scheme |
- 1:
Decompose the large-scale dynamical system ( 1) into N agents as the proposed form in ( 8) - 2:
Determine the index set of neighboring agents for each agent - 3:
for all
do - 4:
fix initial state , , and such that - 5:
initialize for all neighboring agents - 6:
determine to approximate the objective function, and following Theorem 2 to approximate the chance constraint ( 10) in an equivalent sense - 7:
generate , scenarios of , to determine the sets of and , - 8:
solve the optimization problem in (21) by taking into account the worst-case of and determine - 9:
generate scenarios of using the dynamics of agent i in form of ( 8) and together with , - 10:
determine set by solving the optimization problem ( 16) - 11:
send the set to all neighboring agents - 12:
receive the sets from all neighboring agents - 13:
apply the first input of solution into the uncertain subsystem ( 8) - 14:
measure the state and substitute it as the initial state of the next step - 15:
set - 16:
go to Step (7)
|
Algorithm 4 Plug-and-Play Operation |
- 1:
Add the number of new subsystems into the previous number of subsystems, e.g., one additional agent such that - 2:
Update the index set of neighboring agents for each agent - 3:
Go to Step (3) of Algorithm 3
- 4:
Remove the number of departing subsystems from the previous number of subsystems, e.g., one agent leaving leads to such that - 5:
Update the index set of neighboring agents for each agent - 6:
Go to Step (3) of Algorithm 3
|
Comparison: DSO without vs. with Soft Communication
We now provide a comparison between the proposed distributed scenario optimization without vs. with soft communication in Algorithm 2 and Algorithm 3, respectively, in terms of their computational complexities and the conservatism level of obtained solutions as follows:
(1) Computational complexity: It is easy to realize that the proposed distributed scenario optimization (Algorithm 3) using soft communication scheme is computationally advantageous compared to the distributed scenario optimization without soft communication (Algorithm 2). This can be clearly seen by just comparing Step 7 and Step 8 in Algorithms 2 and 3, respectively. In Step 7 of Algorithm 2 each agent
needs to execute the proposed distributed scenario exchange scheme in Algorithm 1 until all its neighboring agents agree on consistency of the set of their state variable scenarios
. However, in Step 8 of Algorithm 3 each agent
needs to solve only once the robust-communication scenario program in (21) and also solve the proposed optimization problem (
16) in Step 10 of Algorithm 3 in order to determine the set
.
(2) Conservatism level: As shown in Theorem 2, the proposed distributed scenario optimization using Algorithm 2 retrieves exactly the same property of the original problem (5) under certain conditions, i.e., , . This is however different in the distributed scenario optimization using the proposed soft communication scheme as presented in Theorem 5. In fact, such a scheme introduces some level of stochasticity on the probabilistic feasibility of the agent , due to the probabilistic reliability of the neighboring information. It is important to mention that guaranteeing the optimality of the obtained solutions in terms of performance objective(s) using the proposed distributed scenario optimization without or with soft communication in Algorithm 2 and Algorithm 3, respectively, is not included in the scope of this paper and it is subject of our ongoing research work.
7. Numerical Study
This section presents two different case studies to illustrate the functionality of our proposed framework to deal with dynamically coupled (
Section 7.1) and coupling constraints (
Section 7.2) neighboring agents with private and common uncertainty sources in networked control problems. The simulation environment for both cases was MATLAB with YALMIP as the interface [
31] and Gurobi as the solver.
We simulate four different problem formulations, namely a centralized SMPC (CSMPC) using (6), a distributed SMPC (DSMPC) via Algorithm 2, and DSMPC with the proposed soft communication scheme with
reliability (DSMPCS
) as described in Definition 1 and DSMPCS
, both following Algorithm 3 in a closed-loop control system framework. For comparison purposes, we also present the results obtained via decoupled SMPC (DeSMPC), where the impact of coupling neighboring dynamics in (
8) are relaxed.
In
Figure 1 and
Figure 2, we evaluate our proposed framework in terms of a-posteriori feasibility validation of the obtained results in both case studies. The “red” line represents the results obtained via DeSMPC, the “blue” line shows the results obtained via CSMPC, the “magenta” presents the results obtained by using DSMPC, the “dark green” and “light green” lines show the results obtained via DSMPCS
and DSMPCS
, respectively. The “black” lines indicate the bounds of the three dynamically coupled systems.
7.1. Three-Room Case Study
We simulate a building climate comfort system with three rooms such that the temperature of rooms is dynamically coupled without any common constraints. The outside weather temperature and the related losses, e.g., through windows, is considered to be the private uncertainty.
Consider now a three-room building system dynamic:
where
,
, and
A as follows:
such that
and
as well as
. The system matrices are a simplified model of a three-room building such that the states
for
, denote the temperature of rooms. The uncertain variable
represents the modeling errors, losses through windows, and
can be realized as the outside weather temperatures.
To generate random scenarios from the additive disturbance, we built a discrete normal process such that one day hourly based forecasted (nominal) outside weather temperature is used which varies within of its nominal scenario at each sampling time. As for the uncertainty , we generate a random variable from a normal distribution with a mean value 0, variance 1 and a maximal magnitude of at each sampling time.
The initial state variables are and the objective is to keep the temperature of rooms within our desired lower and upper bounds at the minimum control input . The control input is also constrained to be within [kWh] and [kWh] for all three rooms, due to actuator saturation. The initialization of the for all neighboring agents as in Step 5 in Algorithm 3 can be done for instance by assuming the initial temperature of the neighboring rooms are known for all rooms.
In
Figure 1, the dynamically coupled state trajectories for all three rooms are shown. One can clearly see in
Figure 1 that the dynamically coupled state trajectories are feasible in a probabilistic sense, since the agent operations are within the lower and upper bounds compared to DeSMPC that violates the constraints completely; the obtained solution via DeSMPC is completely outside of the feasible areas after the first sampling time and thus, we just keep the other results for our discussions.This is a direct result of Theorem 2 such that the obtained solutions via our proposed formulations have to be probabilistically feasible, that can be clearly seen in
Figure 1, since the trajectories are on the lower bounds. Italics, space, bold, superscript, subscript.
7.2. Three-Building (ATES Systems) Case Study
We next simulate the thermal energy demands of three buildings modeled using realistic parameters and the actual registered weather data in the city center of Utrecht, The Netherlands, where these buildings are located and these had been equipped with aquifer thermal energy storage (ATES) systems. A simulation study is carried out for one year based on actual weather conditions from March 2011–March 2012 in a weekly based time steps with three months prediction horizon to meet the coupling constraints between ATES systems. We refer interested readers to [
26] for the detailed explanations on this case study.
To generate random scenarios from the additive disturbance, we built a discrete normal process such that the nominal scenario is of the amplitude of the energy content in a deterministic ATES system model and varies within of its nominal scenario at each sampling time. As for the uncertainty , we generate a random variable from a Gaussian distribution with a mean value 0, variance and a maximal magnitude of at each sampling time.
In
Figure 2 we show a-posteriori feasibility validation of the coupling constraints between each agent
, and neighboring agents, such that (a) shows the results from mid-May to mid-August 2011, (b) presents the results of December 2011 to February 2012, and (c) depicts the results of mid-April to mid-July 2011. As a first desired achievement, one can clearly see in
Figure 2 that the constraints are feasible in a probabilistic sense, since the agent operations are quite close to the upper bounds in the critical time periods compared to DeSMPC that violates the constraints. Strictly speaking, using our proposed framework one can achieve the maximum usage of the aquifer (subsurface) to store thermal energy without affecting the neighboring thermal storage. This is a direct result of Theorem 2 such that the obtained solutions via our proposed formulations have to be probabilistically feasible.
It is worth mentioning that both
Figure 1 and
Figure 2 illustrate our other two main contributions more precisely: (1) the obtained results via CSMPC (blue line) and DSMPC (magenta line) are practically equivalent throughout the simulation studies; this is due to Proposition 1 and Theorem 2. Actually, the solutions via DSMPC are slightly more conservative compared to the results via CSMPC, and this is a direct consequence of Theorem 2. In fact, the level of violation in CSMPC is considered to be
and leading to
for all agents, which is more restrictive. (2) the proposed soft communication scheme yields fewer conservative solutions as explicitly derived in Theorem 5, and can be clearly seen in
Figure 1 and
Figure 2 with the obtained results via DSMPCS
(dark green) and DSMPCS
(light green). Following Theorem 5 the new violation level using DSMPCS
is
, and using DSMPCS
is
. It is important to notice that the violation level of global chance constraint will increase to
and
using DSMPCS
and DSMPCS
, respectively.