Next Article in Journal
Time-Transient Optimization of Electricity and Fresh Water Cogeneration Cycle Using Gas Fuel and Solar Energy
Previous Article in Journal
Viability Selection at Linked Sites
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multilevel Heterogeneous ADMM Algorithm for Elliptic Optimal Control Problems with L1-Control Cost

1
School of Science, Dalian Maritime University, Dalian 116026, China
2
School of Mathematical Sciences, Dalian University of Technology, Dalian 116024, China
3
College of Sciences, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(3), 570; https://doi.org/10.3390/math11030570
Submission received: 31 December 2022 / Revised: 18 January 2023 / Accepted: 19 January 2023 / Published: 21 January 2023
(This article belongs to the Section Computational and Applied Mathematics)

Abstract

:
In this paper, elliptic optimal control problems with L 1 -control cost and box constraints on the control are considered. To numerically solve the optimal control problems, we use the First optimize, then discretize approach. We focus on the inexact alternating direction method of multipliers (iADMM) and employ the standard piecewise linear finite element approach to discretize the subproblems in each iteration. However, in general, solving the subproblems is expensive, especially when the discretization is at a fine level. Motivated by the efficiency of the multigrid method for solving large-scale problems, we combine the multigrid strategy with the iADMM algorithm. Instead of fixing the mesh size before the computation process, we propose the strategy of gradually refining the grid. Moreover, to overcome the difficulty whereby the L 1 -norm does not have a decoupled form, we apply nodal quadrature formulas to approximately discretize the L 1 -norm and L 2 -norm. Based on these strategies, an efficient multilevel heterogeneous ADMM (mhADMM) algorithm is proposed. The total error of the mhADMM consists of two parts: the discretization error resulting from the finite-element discretization and the iteration error resulting from solving the discretized subproblems. Both errors can be regarded as the error of inexactly solving infinite-dimensional subproblems. Thus, the mhADMM can be regarded as the iADMM in function space. Furthermore, theoretical results on the global convergence, as well as the iteration complexity results o ( 1 / k ) for the mhADMM, are given. Numerical results show the efficiency of the mhADMM algorithm.
MSC:
49M41; 49M25; 65K10; 65M32

1. Introduction

Sparse optimal control problems are widespread in many areas, such as the placement of actuators [1], quantum spin systems [2], etc. [3,4]. It is well-known that adding the L 1 -norm penalty can lead to sparse optimal control problems, i.e., the infinite dimensional control variable is localized in its domain of action. In this paper, we consider the elliptic optimal control problem with L 1 -control cost and box constraints on the control:
min ( y , u ) Y × U J ( y , u ) = 1 2 y y d L 2 ( Ω ) 2 + α 2 u L 2 ( Ω ) 2 + β u L 1 ( Ω ) s . t . L y = u + y r in Ω , y = 0 on Ω , u U a d = { v ( x ) | a v ( x ) b , a . e on Ω } U ,
where Y : = H 0 1 ( Ω ) , U : = L 2 ( Ω ) , Ω R n ( n = 2 , 3 ) is a convex, open and bounded domain with C 1 , 1 - or polygonal boundary; the desired state y d H 1 ( Ω ) and the source term y r H 1 ( Ω ) are given; parameters α , β > 0 , < a < 0 < b < + ; L is the uniformly elliptic differential operator given by
L y : = i , j = 1 n x j ( a i j y x i ) + c 0 y ,
where a i j , c 0 L ( Ω ) , c 0 0 , a i j = a j i and there is a constant θ > 0 , such that
i , j = 1 n a i j ( x ) ξ i ξ j θ ξ 2 , a . a . x Ω , ξ R n .
Due to the L 1 -control cost, the objective function J ( y , u ) is non-smooth. This means that the structure of the control significantly differs from that of the smooth one. To solve elliptic optimal control problems with L 1 -control cost, Stadler et al. proposed a semi-smooth Newton method to solve elliptic optimal control problems with L 1 -control cost [1]. Porcelli et al. proposed a semi-smooth Newton method with a robust preconditioner for different formulations of the Newton equation [5]. However, solving Newton equations is expensive, especially when the discretization is at a fine level. Thus, some efficient first-order algorithms have received much attention in recent years. In [6], Schindele and Borzì proposed an inexact accelerated proximal gradient (APG) method in function space to solve elliptic optimal control problems with L 1 -control cost. However, the efficiency of the APG method relies closely on the step length. The backtracking approach is applied to obtain the appropriate step length, but this greatly increases the computational cost. In [7], Song et al. proposed an inexact heterogeneous ADMM (ihADMM) algorithm. Different from the classical ADMM, two different weighted inner products are utilized to define the augmented Lagrangian function for two subproblems. Moreover, the ihADMM was applied to solve PDE-constrained optimization problems with L 2 -control cost [8] and elliptic optimal control problems with pointwise box constraints on the state [9]. For more applications of ADMM-type algorithms to solve PDE-constrained optimization problems, see [10,11,12,13]. Inspired by the efficiency of ADMM-type algorithms on PDE-constrained optimal control problems, we consider using ADMM-type algorithms to solve (1).
In classical finite-element-based algorithms, problems are always discretized and computed at a fixed grid level. As the mesh becomes finer and finer, the scale of the discretized problems will be larger and the computation cost will increase, causing a bottleneck. Thus, it is essential to develop new approaches to solve optimal control problems in an accurate and computationally efficient way. The multigrid method is a modern field of research, which started in the early 1970s. It is well-known that the multigrid method is the optimal solution to many discretized partial differential equations [14]. Different from the classical multigrid method, Deuflhard proposed a cascadic multigrid method for elliptic partial differential equations, which solves problems from the coarse grid to the fine grid [15]. In [16], the multigrid method was applied to tackle infinite dimensional non-linear partial differential equations using Newton methods. Due to the efficiency of the multigrid method, it is a natural idea to construct the multigrid type numerical method for PDE-constrained optimization problems. The classical multigrid method for the optimal control problem is designed to solve the linear algebraic systems formulated on each step of the optimization algorithm; refer to [17] for an overview. In [18], Gong et al. proposed an adaptive multilevel correction method, which solves the optimal control problems from the coarsest mesh to the finest mesh. Chen et al. proposed a strategy of gradually refining the grid, and proposed a multilevel ADMM (mADMM) algorithm for PDE-constrained optimization problems with a L 2 -control cost in [19]. The authors proved the global convergence and the iteration complexity results o ( 1 / k ) for the mADMM.
In this paper, we focus on the first optimize, then discretize [20] approach. This approach provides the freedom to discretize subproblems using different discretization schemes. Motivated by the success of the multigrid method, we combined the multigrid method and the ihADMM algorithm to numerically solve (1). At the early stage of the whole iteration process of the algorithm, using a coarse grid will not make the precision worse, but will reduce the computational cost. Thus, we propose a strategy of gradually refining the grid. Specifically, we first introduced the iterative scheme of the iADMM in function space. Then, nodal quadrature formulas were used to approximately discretize the L 1 -norm and L 2 -norm to ensure that discretized subproblems have decoupled forms. Finally, appropriate methods, such as Krylov-based methods, are used to solve the discretized subproblems.
The main contribution of this paper can be summarized as follows:
  • We propose an efficient multilevel heterogeneous ADMM (mADMM) algorithm for solving sparse optimal control problems with L 1 -control cost. Specifically, we first apply the iADMM in the function space as an outer optimization method. Then, the subproblems in each iteration are discretized by the finite element method. Instead of fixing the grid size before the computation process, we apply the strategy of gradually refining the grid. Moreover, we apply nodal quadrature formulas to approximately L 1 -norm and L 2 -norm to overcome the difficulty that L 1 -norm does not have a decoupled form and ensure z-subproblems have closed form solutions. Finally, we use appropriate methods to inexactly solve the subproblems in each iteration. The global convergence and the iteration complexity results for the mhADMM algorithm are also established.
  • We present numerical experiments to illustrate the effectiveness of the mhADMM algorithm. In addition, we compare the performance of the mhADMM with two benchmark methods: the ihADMM and the classical ADMM. Compared to the two methods, the mhADMM has an evident advantage in terms of computational time. Moreover, the numerical results regarding iterations illustrate the mesh-independent performance of the mhADMM.
The paper is organized as follows. In Section 2, we first briefly review the iteration format of the inexact ADMM algorithm in function space, then introduce the finite element approximation. A strategy of gradually refining the grid is introduced and the mhADMM algorithm is proposed. Numerical computation of the subproblems is also discussed. In Section 3, we present the convergence results of the mhADMM algorithm. Numerical results are given in Section 4, and concluding remarks are drawn in Section 5.

2. An Multilevel Heterogeneous ADMM Algorithm

In this section, we propose an efficient convergent multilevel alternating direction method of multipliers (mhADMM). Moreover, we introduce the numerical computation of the subproblems in the mhADMM algorithm.
To apply ADMM to solve (1), we introduce an artificial variable z; then, we can rewrite (1) in an equivalent form:
min ( y , u , z ) Y × U × U J ( y , u , z ) = 1 2 y y d L 2 ( Ω ) 2 + α 2 u L 2 ( Ω ) 2 + β z L 1 ( Ω ) s . t . L y = u + y r in Ω , y = 0 on Ω , u = z , z U a d = { v ( x ) | a v ( x ) b , a . e on Ω } U .
For the existence and uniqueness of the solution of the elliptic PDE involved in (1), where L is defined by (2), the following proposition holds.
Proposition 1. 
For every u L 2 ( Ω ) and y r H 1 ( Ω ) , the elliptic PDE involved in (1):
L y = u + y r in Ω , y = 0 on Ω ,
has a unique weak solution y = y ( u ) : = S ( u + y r ) , where S : L 2 ( Ω ) H 0 1 ( Ω ) denotes the solution operator. Moreover, S is a well-defined continuous linear injective operator. The adjoint operator S * : H 1 ( Ω ) H 0 1 ( Ω ) is also a continuous linear operator.
Proof. 
The weak formulation of (5) is given by the following:
Find y H 0 1 ( Ω ) : a y , v = u + y r , v L 2 ( Ω ) , v H 0 1 ( Ω ) ,
where a : V × V R is a bilinear form:
a y , v : = Ω i , j = 1 n a i j y x i v x j + c 0 y v d x .
We know from the assumption (3) that a · , · is symmetric, i.e., a y , v = a v , y for all y , v H 0 1 ( Ω ) . Thus, y , v : = a y , v defines a new inner product on H 0 1 ( Ω ) . Then, the existence of a unique solution of (5) directly follows from the Riesz representation theorem. Moreover,
θ y H 0 1 ( Ω ) 2 a y , y = u + y r , y L 2 ( Ω ) u + y r L 2 ( Ω ) y L 2 ( Ω ) u + y r L 2 ( Ω ) y H 0 1 ( Ω ) ,
where θ is a constant, depending only on Ω , and we use the equivalence of norms in H 0 1 ( Ω ) in the first inequality. Then, we have
y H 0 1 ( Ω ) 2 1 θ u + y r L 2 ( Ω ) .
Thus, the solution operator S is well-defined and called the control-to-state mapping, which is a continuous linear injective operator. Since H 0 1 is a Hilbert space, the adjoint operator S * : H 1 ( Ω ) H 0 1 ( Ω ) is also a continuous linear operator.    □
Since (4) is continuous and strongly convex, the solution of (4) exists and is unique. The following Karush–Kuhn–Tucker (KKT) conditions hold at the optimal solution of (4).
Theorem 1. 
(First-order optimality condition) ( y * , u * , z * ) is the optimal solution of (4), if, and only if, adjoint state p * H 0 1 ( Ω ) and Lagrange multiplier λ * L 2 ( Ω ) exists, such that the following conditions hold in the weak sense:
y * = S u * + y r , p * = S * y * y d , α u * p * + λ * = 0 , u * = z * , z * U a d , λ * , z ˜ z * L 2 ( Ω ) + β z ˜ L 1 ( Ω ) z * L 1 ( Ω ) 0 , z ˜ U a d .
Moreover, we have
u * : = Π U a d 1 α soft p * , β ,
where the projection operator Π U a d · and the soft thresholding operator soft · , respectively, are defined as follows:
Π U a d v x : = max { a , min { v ( x ) , b } } , soft ( v ( x ) , β ) : = sgn ( v ( x ) ) max ( | v ( x ) | β , 0 ) .
Using the solution operator S, (4) can be equivalently rewritten as the following reduced form:
min ( u , z ) U × U f ( u ) + g ( z ) s . t . u = z ,
where
f ( u ) : = 1 2 S ( u + y r ) y d L 2 ( Ω ) 2 + α 2 u L 2 ( Ω ) 2 , g ( z ) : = β z L 1 ( Ω ) + δ U a d ( z ) ,
δ U a d denotes the indicator function of U a d , i.e.,
δ U a d ( z ) = 0 , z U a d , , z U a d .
Notice that (6) is a two-block separable convex optimization problem with linear equality constraints; the ADMM algorithm can be used to solve (6). The augmented Lagrangian function of (6) is defined as follows:
L σ ( u , z , λ ; σ ) = f ( u ) + g ( z ) + λ , u z L 2 ( Ω ) + σ 2 u z L 2 ( Ω ) 2 ,
where λ L 2 ( Ω ) denotes the Lagrangian multiplier; σ > 0 is the penalty parameter. Given the initial point ( u 0 , z 0 , λ 0 ) L 2 ( Ω ) × dom ( δ U a d ( · ) ) × L 2 ( Ω ) , parameters σ > 0 , τ 0 , 5 + 1 2 , the iteration format of ADMM in function space is as follows:
Step 1 : u ¯ k + 1 = arg min u f ( u ) + λ ¯ k , u z ¯ k L 2 ( Ω ) + σ 2 u z ¯ k L 2 ( Ω ) 2 , Step 2 : z ¯ k + 1 = arg min z g ( z ) + λ ¯ k , u ¯ k + 1 z L 2 ( Ω ) + σ 2 u ¯ k + 1 z L 2 ( Ω ) 2 , Step 3 : λ ¯ k + 1 = λ ¯ k + τ σ ( u ¯ k + 1 z ¯ k + 1 ) .
However, computing the exact solution of each subproblem is usually expensive and unnecessary. In [7], Song et al. propose an inexact ADMM (iADMM) in function space for (6). Krylov-based methods are used to solve the subproblems, which are equivalent to large-scale linear systems. We show the iterative scheme of the iADMM algorithm in Algorithm 1.
It is easy to see that the z-subproblem has a closed-form solution
z k + 1 = arg min z σ 2 z 1 σ ( σ u k + 1 + λ k ) L 2 ( Ω ) 2 + β z L 1 ( Ω ) + δ U a d ( z ) = Π U a d 1 σ soft σ u k + 1 + λ k , β .
For the global convergence results and the iteration complexity for Algorithm 1, we have the following theorem.
Algorithm 1 Inexact ADMM (iADMM) algorithm for (6)
  • Input: Choose the initial point ( u 0 , z 0 , λ 0 ) L 2 ( Ω ) × dom ( δ U a d ( · ) ) × L 2 ( Ω ) , parameters σ > 0 , τ 0 , 5 + 1 2 . Let { ϵ k + 1 } k = 0 be a sequence satisfying { ϵ k + 1 } k = 0 [ 0 , + ) and k = 0 ϵ k < . Set k = 0 .
  • Output: u k , z k , λ k .
  • Step 1 Compute u k + 1 as an approximation solution of
    min u f ( u ) + λ k , u z k L 2 ( Ω ) + σ 2 u z k L 2 ( Ω ) 2
           such that the residual δ u k + 1 : = f ( u k + 1 ) + λ k + σ ( u k + 1 z k ) satisfies
  •        δ u k + 1 L 2 ( Ω ) ϵ k + 1 .
  • Step 2 Compute z k + 1 as follows:
    z k + 1 = arg min z g ( z ) + λ k , u k + 1 z L 2 ( Ω ) + σ 2 u k + 1 z L 2 ( Ω ) 2 .
  • Step 3 Compute
    λ k + 1 = λ k + τ σ ( u k + 1 z k + 1 ) .
  • Step 4 If a termination criterion is met, stop; otherwise, set k : = k + 1 and go to Step 1.
Theorem 2. 
([7] Theorem 3) Let ( y * , u * , z * , p * , λ * ) be the KKT point of (4), the sequence { ( u k , z k , λ k ) } k = 0 is generated by Algorithm 1 with the associated state { y k } k = 0 and adjoint state { p k } k = 0 ; then, we have
lim k { u k u * L 2 ( Ω ) + z k z * L 2 ( Ω ) + λ k λ * L 2 ( Ω ) } = 0 , lim k { y k y * H 0 1 ( Ω ) + p k p * H 0 1 ( Ω ) } = 0 .
Moreover, a constant C 0 only depends on the initial point ( u 0 , z 0 , λ 0 ) and the optimal solution ( u * , z * , λ * ) , such that, for k 1 ,
min 1 i k R ( u i , z i , λ i ) C 0 k , lim k ( k · min 1 i k R ( u i , z i , λ i ) ) = 0 ,
where the function R : ( u , z , λ ) [ 0 , ) is defined as
R ( u , z , λ ) : = f ( u ) + λ L 2 ( Ω ) 2 + dist 2 ( 0 , λ + g ( z ) ) + u z L 2 ( Ω ) 2 .

2.1. The mhADMM Algorithm

To numerically solve (6), we consider the full discretization, in which both the state y and the control u are discretized by piecewise, linear, globally continuous finite elements. We introduce a family of regular and quasi-uniform triangulations { T h } of Ω ¯ , i.e., Ω ¯ = T T h T ¯ . With each element T T h , we define the diameter of the set T by ρ T : = diam T and let σ T denote the diameter of the largest ball contained in T. The grid size is defined by h : = max T T h ρ T . We suppose the following standard assumption holds (see [20]).
Assumption 1.
(Regular and quasi-uniform triangulations) There are two positive constants κ and τ such that
ρ T σ T κ , h ρ T τ
hold for all T T h and all h > 0 . Moreover, let us define Ω ¯ h = T T h T ¯ and let Ω h Ω and Γ h denote its interior and boundary, respectively. In the case that Ω is a convex polyhedral domain, we have Ω = Ω h . In the case that Ω is a domain with a C 1 , 1 - boundary Γ, we assume that Ω ¯ h is convex and all boundary vertices of Ω ¯ h are contained in Γ, such that
| Ω Ω h | c * h 2 ,
where | · | denotes the measure of the set, and c * > 0 is a constant.
Due to the homogeneous boundary condition of the state equation, we use
Y h : = { y h C ( Ω ¯ ) | y h | T P 1 , T T h , y h = 0 in Ω ¯ Ω h } , U h : = { u h C ( Ω ¯ ) | u h | T P 1 , T T h , u h = 0 in Ω ¯ Ω h } ,
as the discretized state space and the discretized control space, respectively. P 1 denotes the space of polynomials of degree less than or equal to 1. For the given regular and quasi-uniform triangulation T h with nodes { x i } i = 1 N h , let { ϕ i ( x ) } i = 1 N h be a basis of Y h , U h , which satisfies the following properties:
ϕ i ( x ) 0 , ϕ i ( x ) = 1 i = 1 , . . . , N h , i = 1 N h ϕ i ( x ) = 1 .
Then, u h U h , y h Y h can be represented in the following forms, respectively,
y h = i = 1 N h y i ϕ i , u h = i = 1 N h u i ϕ i ,
where y i : = y h ( x i ) , u i : = u h ( x i ) . Moreover, y h can be expressed by y h ( u ) = S h ( u + y r ) , where S h denotes the discretized version of the solution operator S. Let U a d , h denotes the discretized feasible set, which is defined by
U a d , h : = U h U a d = z h = i = 1 N h z i ϕ i | a z i b , i = 1 , , N h U a d .
To overcome the difficulty that the discretized L 1 -norm does not have a decoupled form, we choose the nodal quadrature formulas introduced in [21] to approximately discretized the L 1 -norm:
z h L h 1 Ω h = i = 1 n z i Ω h ϕ i ( x ) d x .
Moreover, in order to obtain a closed form solution for the z-subproblem, a similar quadrature formulae introduced in [8] is also used to discretize the L 2 -norm in the z-subproblem:
z h L h 2 Ω h = i = 1 n ( z i ) 2 Ω h ϕ i ( x ) d x 1 2 .
For the given y r H 1 ( Ω ) and u L 2 ( Ω ) , the unique discretized state y h associated with u is can be expressed by y h ( u ) = S h ( u + y r ) , where S h is the discretized version of the solution operator S. Then, we have the well-known error estimates:
Lemma 1 
([22], Theorem 4.4.6). For a given u L 2 ( Ω ) , let y be the unique weak solution of the state Equation (5) and y h be the unique discretized state. Then, there exists a constant c > 0 independent of h, u and y r , such that
y y h L 2 ( Ω ) + h y y h L 2 ( Ω ) c h 2 ( u L 2 ( Ω ) + y r L 2 ( Ω ) ) .
In particular, this implies S S h L ( L 2 , L 2 ) c h 2 and S S h L ( L 2 , H 1 ) c h .
To project the solution obtained on the coarser grid to the finer grid, we introduce the definition of the node interpolation operator I h .
Definition 1. 
For a given regular and quasi-uniform triangulation T h of Ω with nodes { x i } i = 1 N h , let { ϕ i ( x ) } i = 1 N h denotes the standard set of nodal basis functions. The interpolation operator I h is defined as
I h w ( x ) : = i = 1 N h w ( x i ) ϕ i ( x ) for any w C 0 ( Ω ) H 1 ( Ω ) .
For the interpolation error estimate, we have the following Theorem 3.
Theorem 3 
([19] Theorem 2). For all w C 0 ( Ω ) H 1 ( Ω ) , we have
w I h w L 2 ( Ω ) c I h w H 1 ( Ω ) ,
where c I is a constant independent of h.
At the early stage of the whole process, computing on the coarser grid can reduce the computation cost without making the precision worse. While as the iteration process proceeds, the iteration precision is supposed to increase. In this case, it is necessary to use the finer grid at the late stage. Thus, we apply the strategy of gradually refining the grid. In the initial iteration, we obtained a solution on the coarse grid, then projected the obtained solution to the finer grid. For the convenience of representing subproblems on different grids, we define
f h k + 1 ( u ) : = 1 2 S h k + 1 ( u + I h k + 1 y r ) I h k + 1 y d L 2 ( Ω h k + 1 ) 2 + α 2 u L 2 ( Ω h k + 1 ) 2 , g h k + 1 ( z ) : = β z L h k + 1 1 ( Ω h k + 1 ) + δ U a d , h k + 1 ( z ) .
Moreover, let
I h y r : = i = 1 N h y r i ϕ i , I h y d : = i = 1 N h y d i ϕ i
denotes the L 2 -projection of y r and y d onto Y h , respectively. Then, we show the iterative scheme of the multilevel heterogeneous ADMM alternating direction method of multipliers (mhADMM) in Algorithm 2.
Algorithm 2 Multilevel heterogeneous ADMM (mhADMM) algorithm for (6)
  • Input: Choose the initial point ( u h 1 0 , z h 1 0 , λ h 1 0 ) H 1 ( Ω ) × H 1 ( Ω ) × H 1 ( Ω ) , parameters σ > 0 , τ 0 , 5 + 1 2 . Let { ϵ k + 1 } k = 0 be a sequence satisfying { ϵ k + 1 } k = 0 [ 0 , + ) and k = 0 ϵ k + 1 < , mesh sizes { h k + 1 } k = 0 of each level satisfying k = 0 h k + 1 < . Set k = 0 .
  • Output: u h k k , z h k k , λ h k k .
  • Step 1 Compute u h k + 1 k + 1 as an approximation solution of
    min u f h k + 1 ( u ) + λ h k + 1 k , u z h k + 1 k L 2 ( Ω h k + 1 ) + σ 2 u z h k + 1 k L 2 ( Ω h k + 1 ) 2
           such that the residual δ u , h k + 1 k + 1 : = f h k + 1 ( u h k + 1 k + 1 ) + λ h k + 1 k + σ ( u h k + 1 k + 1 z h k + 1 k ) satisfies δ u , h k + 1 k + 1 L 2 ( Ω h k + 1 ) ϵ k + 1 .
  • Step 2 Compute z h k + 1 k + 1 as follows:
    z h k + 1 k + 1 = arg min z g h k + 1 ( z ) + λ h k + 1 k , u h k + 1 k + 1 z L 2 ( Ω h k + 1 ) + σ 2 u h k + 1 k + 1 z L h k + 1 2 ( Ω h k + 1 ) 2 .
  • Step 3 Compute
    λ h k + 1 k + 1 = λ h k + 1 k + τ σ ( u h k + 1 k + 1 z h k + 1 k + 1 ) .
  • Step 4 If a termination criterion is met, stop; otherwise, set k : = k + 1 and go to Step 1.
It is easy to see that the z-subproblem has a closed-form solution:
z h k + 1 k + 1 = arg min z σ 2 z 1 σ ( σ u h k + 1 k + 1 + λ h k + 1 k ) L 2 ( Ω h k + 1 ) 2 + β z L h k + 1 1 ( Ω h k + 1 ) + δ U a d , h k + 1 ( z ) = Π U a d , h k + 1 1 σ soft σ u h k + 1 k + 1 + λ h k + 1 k , β .

2.2. Numerical Computation of the Subproblems in Algorithm 2

To rewrite the subproblems into matrix-vector forms, we define the following matrices
K h : = ( a ( ϕ i , ϕ j ) ) i , j = 1 N h , M h : = Ω h ϕ i ϕ j d x i , j = 1 N h , W h : = diag Ω h ϕ i ( x ) d x i , j = 1 N h ,
where K h , M h and W h denote the finite element stiffness matrix, mass matrix and lump mass matrix, respectively.
For u h = i = 1 N h u i ϕ i U h , y h = i = 1 N h y i ϕ i Y h , let
u h = ( u 1 , . . . , u N h ) , y h = ( y 1 , . . . , y N h )
be the relative coefficient vectors, respectively. For I h y r = i = 1 N h y r i ϕ i , I h y d = i = 1 N h y d i ϕ i , let
y r , h = ( y r 1 , y r 2 , . . . , y r N h ) , y d , h = ( y d 1 , y d 2 , . . . , y d N h )
be the coefficient vectors, respectively. Let I h denotes the vector version of the interpolation operator. Moreover, we define
f ( u ) : = 1 2 K h 1 M h ( u + y r , h ) y d , h M h 2 + α 2 u M h 2 , g ( z ) : = β W h z 1 2 + δ [ a , b ] N h ( z ) .
Then, the matrix-vector form of Algorithm 2 is given in Algorithm 3.
Algorithm 3 Matrix-vector form of the mhADMM algorithm
  • Input: ( u 0 , z 0 , λ 0 ) R N h × [ a , b ] N h × R N h , parameters σ > 0 , τ 0 , 5 + 1 2 . Let { ϵ k + 1 } k = 0 be a sequence satisfying { ϵ k + 1 } k = 0 [ 0 , + ) and k = 0 ϵ k + 1 M h k + 1 2 < , mesh sizes { h k + 1 } k = 0 of each iteration satisfy k = 0 h k + 1 < . Set k = 0 .
  • Output: u h k k , z h k k , λ h k k .
  • Step 1: Compute u h k + 1 k + 1 as an approximation solution of
    min u f ( u ) + M h k + 1 λ h k + 1 k , u z h k + 1 k + σ 2 u z h k + 1 k M h k + 1 2
           such that the residual δ u , h k + 1 k + 1 : = f ( u h k + 1 k + 1 ) + M h k + 1 λ h k + 1 k + σ M h k + 1 ( u h k + 1 k + 1 z h k + 1 k )
  •        satisfies δ u , h k + 1 k + 1 ϵ k + 1 M h k + 1 2 .
  • Step 2: Compute z h k + 1 k + 1 as follows:
    z h k + 1 k + 1 = arg min z g ( z ) + M h λ h k + 1 k , u h k + 1 k + 1 z + σ 2 u h k + 1 k + 1 z W h 2 .
  • Step 3: Compute
    λ h k + 1 k + 1 = I h k + 1 λ h k k + τ σ ( u h k + 1 k + 1 z h k + 1 k + 1 ) .
  • Step 4: If a termination criterion is met, stop; otherwise, set k : = k + 1 and go to Step 1.
The u-subproblem at the kth iteration is equivalent to the following linear system:
M h k K h k 1 M h k ( K h k 1 M h k ( u h k k + y r , h k ) y d , h k ) + α M h k u h k k + λ h k k + σ ( u h k k z h k k ) = 0 ,
where y h k k : = K h k 1 M h k ( u h k k + y r , h k ) , p h k k : = K h k 1 M h k ( y d , h k y h k k ) denote the discretized state and the discretized adjoint state, respectively. Then (8) can be rewritten as:
M h k 0 K h k 0 ( α + σ ) M h k M h k K h k M h k 0 y h k k u h k k p h k k = M h k y d , h k M h k ( σ I h k z h k 1 k 1 I h k λ h k 1 k 1 ) M h k y r , h k .
We know from (9) that p h k k = ( α + σ ) u h k k σ I h k z h k 1 k 1 + I h k λ h k 1 k 1 . By eliminating the variable p h k k , (9) can be rewritten in the following reduced form without any additional computational cost:
M h k ( α + σ ) K h k K h k M h k y h k k u h k k = M h k y d , h k + K h k ( σ I h k z h k 1 k 1 I h k λ h k 1 k 1 ) M h k y r , h k .
The equation system (10) can be solved by the generalized minimal residual (GMRES) with the preconditioned variant of modified hermitian and skew-hermitian splitting (PMHSS) preconditioner [23,24].
For the z-subproblem, there is a closed-form solution
z h k + 1 k + 1 = Π [ a , b ] N h k + 1 1 σ soft σ u h k + 1 k + 1 + W h 1 M h λ h k + 1 k , β .

3. Convergence Analysis

In this section, we establish the global convergence and the iteration complexity results in non-ergodic sense for the sequence generated by Algorithm 2. Before giving the convergence analysis, we first introduce the exact multi-level heterogeneous ADMM (mhADMM) algorithm. Each subproblem of the exact mhADMM algorithm is solved exactly. Given the initial point ( u 0 , z 0 , λ 0 ) H 1 ( Ω ) × H 1 ( Ω ) × H 1 ( Ω ) , parameters σ > 0 , τ 0 , 5 + 1 2 . The mesh sizes { h k + 1 } k = 0 of each iteration satisfy k = 0 h k + 1 < . Then, each iteration of the exact mhADMM has three main steps:
Step 1 : u ¯ h k + 1 k + 1 = arg min u f h k + 1 ( u ) + λ ¯ h k + 1 k , u z ¯ h k + 1 k L 2 ( Ω h k + 1 ) + σ 2 u z ¯ h k + 1 k L 2 ( Ω h k + 1 ) , Step 2 : z ¯ h k + 1 k + 1 = arg min z g h k + 1 ( z ) + λ ¯ h k + 1 k , u ¯ h k + 1 k + 1 z L 2 ( Ω h k + 1 ) + σ 2 u ¯ h k + 1 k + 1 z L h k + 1 2 ( Ω h k + 1 ) 2 , Step 3 : λ ¯ h k + 1 k + 1 = λ ¯ h k + 1 k + τ σ ( u ¯ h k + 1 k + 1 z ¯ h k + 1 k + 1 ) ,
where λ ¯ h k + 1 k : = I h k + 1 λ ¯ h k k , z ¯ h k + 1 k : = I h k + 1 z ¯ h k k . Then, we use the following lemma to measure the gap between the solution sequence obtained by the ADMM in function space and the exact mhADMM algorithm in finite dimensional space.
Lemma 2. 
Let the initial point be ( z 0 , λ 0 ) H 1 ( Ω ) × H 1 ( Ω ) . Let { ( u ¯ k , z ¯ k , λ ¯ k ) } k = 0 defined in (7) be the sequence generated by the ADMM in function space and { ( u ¯ h k k , z ¯ h k k , λ ¯ h k k ) } k = 0 defined in (11) be the sequence generated by the exact mhADMM algorithm. Then, for all k 1 , we have
u ¯ k u ¯ h k k L 2 ( Ω h k ) C u , k h k , z ¯ k z ¯ h k k L 2 ( Ω h k ) C z , k h k , λ ¯ k 1 λ ¯ h k k 1 L 2 ( Ω h k ) C λ , k h k ,
where C u , k , C z , k , C λ , k are constants independent of h k and there is a constant C such that C u , k C for any k 1 . Thus, we have
k = 1 u ¯ k u ¯ h k k L 2 ( Ω h k ) C k = 1 h k .
Proof. 
We employ the mathematical induction to prove the conclusion. For k = 1 , we know from Theorem 3 that
λ ¯ 0 λ ¯ h 1 0 L 2 ( Ω h 1 ) = λ 0 I h 1 λ 0 L 2 ( Ω h 1 ) c I h 1 λ 0 H 1 ( Ω h 1 ) C λ , 1 h 1 ,
where C λ , 1 : = c I λ ¯ 0 H 1 ( Ω h 1 ) is a constant independent of h.
For u-subproblems, u ¯ 1 and u ¯ h 1 1 satisfy the following optimality conditions, respectively,
S * [ S ( u ¯ 1 + y r ) y d ] + α u ¯ 1 + λ 0 + σ ( u ¯ 1 z 0 ) = 0 , S h 1 * [ S h 1 ( u ¯ h 1 1 + I h 1 y r ) I h 1 y d ] + α u ¯ h 1 1 + I h 1 λ 0 + σ ( u ¯ h 1 1 I h 1 z 0 ) = 0 .
By subtracting the two equalities above, we have
( α + σ ) I S h 1 * S h 1 ( u ¯ 1 u ¯ h 1 1 ) = S * S u ¯ 1 S h 1 * S h 1 u ¯ 1 + S * S y r S h 1 * S h 1 I h 1 y r S * y d + S h 1 * I h 1 y d + λ 0 I h 1 λ 0 + σ I h 1 z 0 σ z 0 .
Then, we know from Lemma 2 in [19] that u ¯ 1 u ¯ h 1 1 L 2 ( Ω h 1 ) C u , 1 h 1 , where C u , 1 is a constant independent of h 1 .
For z-subproblems, z ¯ 1 and z ¯ h 1 1 satisfy
z ¯ 1 = Π U a d 1 σ soft σ u ¯ 1 + λ 0 , β , z ¯ h 1 1 = Π U a d , h 1 1 σ soft σ u h 1 1 + λ h 1 0 , β ,
respectively. Then, we know from the projection operator Π and the soft thresholding operator soft · are nonexpansive, such that
z ¯ 1 z ¯ h 1 1 L 2 ( Ω h 1 ) = Π U a d 1 σ soft σ u ¯ 1 + λ 0 , β Π U a d , h 1 1 σ soft σ u ¯ h 1 1 + λ h 1 0 , β L 2 ( Ω h 1 ) 1 σ σ u ¯ 1 σ u ¯ h 1 1 + λ 0 λ h 1 0 L 2 ( Ω h 1 ) u ¯ 1 u ¯ h 1 1 L 2 ( Ω h 1 ) + 1 σ λ 0 λ h 1 0 L 2 ( Ω h 1 ) C u , 1 h 1 + 1 σ C λ , 1 h 1 = C z , 1 h 1 ,
where C z , 1 : = C u , 1 + 1 σ C λ , 1 . Hence, the statement is true for k = 1 .
For k > 1 , we assume the statement is true for j k . Then, for j = k + 1 , we have
λ ¯ k λ ¯ h k + 1 k L 2 ( Ω h k + 1 ) = λ ¯ k λ ¯ h k k + λ ¯ h k k I h k + 1 λ ¯ h k k L 2 ( Ω h k + 1 ) λ ¯ k λ ¯ h k k L 2 ( Ω h k ) + λ ¯ h k k I h k + 1 λ ¯ h k k L 2 ( Ω h k + 1 ) λ ¯ k 1 λ ¯ h k k 1 + τ σ ( u ¯ k u ¯ h k k ) τ σ ( z ¯ k z ¯ h k k ) L 2 ( Ω h k ) + λ ¯ h k k I h k + 1 λ ¯ h k k L 2 ( Ω h k + 1 ) λ ¯ k 1 λ ¯ h k k 1 L 2 ( Ω h k ) + τ σ u ¯ k u ¯ h k k L 2 ( Ω h k ) + z ¯ k z ¯ h k k L 2 ( Ω h k ) + λ ¯ h k k I h k + 1 λ ¯ h k k L 2 ( Ω h k + 1 ) ( C λ , k + τ σ C u , k + τ σ C z , k ) h k + c I h k + 1 λ ¯ h k k H 1 ( Ω h k + 1 ) C λ , k + 1 h k + 1 ,
where C λ , k + 1 : = C k + 1 ( C λ , k h k + τ σ C u , k + τ σ C z , k ) + c I λ ¯ h k k H 1 ( Ω h k + 1 ) is a constant independent of h k + 1 . In the last equality, we use the property k = 0 h k < ; thus, there exists a constant C k + 1 such that h k < C k + 1 h k + 1 .
For u-subproblems, u ¯ k + 1 and u ¯ h k + 1 k + 1 satisfy the following optimality conditions respectively,
S * [ S ( u ¯ k + y r ) y d ] + α u ¯ k + 1 + λ ¯ k + σ ( u ¯ k + 1 z ¯ k ) = 0 , S h k + 1 * [ S h k + 1 ( u ¯ h k + 1 k + 1 + I h k + 1 y r ) I h k + 1 y d ] + α u ¯ h k + 1 k + 1 + λ ¯ h k + 1 k + σ ( u ¯ h k + 1 k + 1 z ¯ h k + 1 k ) = 0 .
By subtracting the two equalities above, we have
( α + σ ) I + S h k + 1 * S h k + 1 ( u ¯ k + 1 u ¯ h k + 1 k + 1 ) = S * S u ¯ k + 1 S h k + 1 * S h k + 1 u ¯ k + 1 + S * S y r S h k + 1 * S h k + 1 I h k + 1 y r S * y d + S h k + 1 * I h k + 1 y d + λ ¯ k λ ¯ h k + 1 k σ ( z ¯ k z ¯ h k + 1 k ) .
Then, we know from Lemma 2 in [19] that
u ¯ k + 1 u ¯ h k + 1 k + 1 L 2 ( Ω h k + 1 ) C u , k + 1 h k + 1 ,
where C u , k + 1 is a constant independent of h k + 1 .
For z-subproblems, z ¯ k + 1 and z ¯ h k + 1 k + 1 satisfy
z ¯ k + 1 = Π U a d 1 σ soft σ u ¯ k + 1 + λ k , β , z ¯ h k + 1 k + 1 = Π U a d , h 1 1 σ soft σ u ¯ h k + 1 k + 1 + λ h k + 1 k , β ,
respectively. We know the projection operator Π and the soft thresholding operator soft · are nonexpansive, such that
z ¯ k + 1 z ¯ h k + 1 k + 1 L 2 ( Ω h k + 1 ) 1 σ σ u ¯ k + 1 σ u ¯ h k + 1 k + 1 + λ k λ h k + 1 k L 2 ( Ω h k + 1 ) u ¯ k + 1 u ¯ h k + 1 k + 1 L 2 ( Ω h k + 1 ) + 1 σ λ k λ h k + 1 k L 2 ( Ω h k + 1 ) C u , k + 1 h k + 1 + 1 σ C λ , k + 1 h k + 1 = C z , k + 1 h k + 1 ,
where C z , k + 1 : = C u , k + 1 + 1 σ C λ , k + 1 . Hence, the statement is true for j = k + 1 . We complete the whole proof of Lemma 2.    □
Similarly, we have the following lemma. Lemma 3 shows the gap between the sequence ( u k , z k , λ k ) generated by Algorithm 1 and the sequence ( u h k k , z h k k , λ h k k ) generated by Algorithm 2.
Lemma 3. 
Let the initial point be ( u 0 , z 0 , λ 0 ) H 1 ( Ω ) × H 1 ( Ω ) × H 1 ( Ω ) . Let { ( u k , z k , λ k ) } k = 0 be the sequence generated by Algorithm 1 and { ( u h k k , z h k k , λ h k k ) } k = 0 be the sequence generated by Algorithm 2. Then, for all k 1 , we have
u k u h k k L 2 ( Ω h k ) C ^ u , k ( h k + δ u , h k k L 2 ( Ω h k ) ) , z k z h k k L 2 ( Ω h k ) C ^ z , k ( h k + δ u , h k k L 2 ( Ω h k ) ) , λ k 1 λ h k k 1 L 2 ( Ω h k ) C ^ λ , k ( h k + δ u , h k k L 2 ( Ω h k ) ) ,
where C ^ u , k , C ^ z , k , C ^ λ , k are constants independent of h k , and there exists a constant C ^ such that C ^ u , k C ^ for any k 1 . Thus we have
k = 1 u k u h k k L 2 ( Ω ) C ^ k = 1 h k + δ u , h k k L 2 ( Ω h k ) .
Proof. 
We employ the mathematical induction to prove the conclusion. The proof is similar to Lemma 2. We do not discuss this in detail here.    □
The total error of utilizing numerical methods to solve PDE-constrained optimal control problem consists of two parts: the discretization error and the iteration error. These two kinds of errors can be regarded as the error of inexactly solving infinite-dimensional subproblems. Thus, the mhADMM algorithm can be regarded as the iADMM algorithm in function space. Inspired by the results of Theorem 2, we have the following convergence results.
Theorem 4. 
Let ( y * , u * , z * , p * , λ * ) be the KKT point of (1), ( u h k k , z h k k , λ h k k ) k = 0 be the sequence generated by Algorithm 2 with the associated state { y h k k } k = 0 and the adjoint state { p h k k } k = 0 . Then we have
lim k { u h k k u * L 2 ( Ω h k ) + z h k k z * L 2 ( Ω h k ) + λ h k k λ * L 2 ( Ω h k ) } = 0 , lim k { y h k k y * H 0 1 ( Ω h k ) + p h k k p * H 0 1 ( Ω h k ) } = 0 .
Moreover, there exists a constant C ¯ that only depends on the initial point ( u 0 , z 0 , λ 0 ) and the optimal solution ( u * , z * , λ * ) such that, for k 1 ,
min 1 i k R h i ( u h i i , z h i i , λ h i i ) C ¯ k , lim k ( k · min 1 i k R h i ( u h i i , z h i i , λ h i i ) ) = 0 ,
where R h i : ( u h i i , z h i i , λ h i i ) [ 0 , ) is defined as
R h ( u h i i , z h i i , λ h i i ) : = f h ( u h i i ) + λ h i i 1 L 2 ( Ω h i ) 2 + dist 2 ( 0 , λ h i i 1 + g h ( z h i i ) ) + u h i i z h i i L 2 ( Ω h i ) 2 .
Proof. 
Note that ( u h k k , z h k k , λ h k k ) can be regarded as the inexact solution obtained by Algorithm 1. The total error δ u k consists of two parts, the discretization error from gradually refining the grid and the iteration error from the inexactly solving the subproblems. Then, we know from the optimality conditions of the u-subproblem in Algorithm 1 that
S * [ S ( u h k k + y r ) y d ] + α u h k k + λ h k 1 k 1 + σ ( u h k k z h k 1 k 1 ) = δ u k .
Moreover, we know from the optimality condition of the u-subproblem in ADMM in function space, the optimality conditions of the u-subproblem in Algorithm 2 and the exact multi-level ADMM that
S * [ S ( u ¯ k + y r ) y d ] + α u ¯ k + λ ¯ k 1 + σ ( u ¯ k z ¯ k 1 ) = 0 ,
S h k * [ S h k ( u h k k + I h k y r ) I h k y d ] + α u h k k + λ h k k 1 + σ ( u h k k z h k k 1 ) = δ u , h k k ,
S h k * [ S h k ( u ¯ h k k + I h k y r ) I h k y d ] + α u ¯ h k k + λ ¯ h k k 1 + σ ( u ¯ h k k z ¯ h k k 1 ) = 0 .
Then, we know from (12)–(15) that
δ u k = δ u k δ u , h k k + δ u , h k k = δ u , h k k + S * S ( u h k k u ¯ k ) + ( α + σ ) ( u h k k u ¯ k ) + ( λ k 1 λ ¯ k 1 ) σ ( z k 1 z ¯ k 1 ) + S h k * S h k ( u ¯ h k k u h k k ) + ( α + σ ) ( u ¯ h k k u h k k ) + ( λ ¯ h k k 1 λ h k k 1 ) σ ( z ¯ h k k 1 z h k k 1 ) = δ u , h k k + [ ( α + σ ) I + S * S ] ( u ¯ h k k u ¯ k ) I 1 + ( λ k 1 λ h k k 1 ) I 2 σ ( z k 1 z h k k 1 ) I 3 + ( λ ¯ h k k 1 λ ¯ k 1 ) I 4 σ ( z ¯ h k k 1 z ¯ k 1 ) I 5 + ( S * S S h k * S h k ) ( u h k k u ¯ h k k ) I 6 .
For the term I 1 and I 4 , we know from Lemma 2 that
I 1 L 2 ( Ω h k ) ( α + σ + S * S ) C u , k h k ,
I 4 L 2 ( Ω h k ) C λ , k h k .
For the term I 5 , we know from Theorem 3 and Lemma 2 that
I 5 L 2 ( Ω h k ) = σ z ¯ h k k 1 z ¯ h k 1 k 1 + z ¯ h k 1 k 1 z ¯ k 1 L 2 ( Ω h k ) σ z ¯ h k k 1 z ¯ h k 1 k 1 L 2 ( Ω h k ) + σ z ¯ h k 1 k 1 z ¯ k 1 L 2 ( Ω h k ) σ c I h k z ¯ h k 1 k 1 H 1 ( Ω h k ) + σ C z , k 1 h k 1 σ c I h k z ¯ h k 1 k 1 H 1 ( Ω h k ) + σ C z , k 1 C k h k c 5 h k ,
where c 5 : = σ c I z ¯ h k 1 k 1 H 1 ( Ω h k ) + σ C z , k 1 C k is a constant, h k 1 C k h k .
For the term I 2 , we know from Lemma 3 that
I 2 L 2 ( Ω h k ) C ^ λ , k ( h k + δ u , h k k L 2 ( Ω h k ) ) .
For the term I 3 , we know from Theorem 3 and Lemma 2 that
I 3 L 2 ( Ω h k ) c 3 ( h k + δ u , h k k L 2 ( Ω h k ) ) ,
where c 3 is a constant.
Finally, for the term I 6 , we make use of the decomposition
u h k k u ¯ h k k L 2 ( Ω h k ) = u h k k u k + u k u * + u * u ¯ k + u ¯ k u ¯ h k k L 2 ( Ω h k ) u h k k u k L 2 ( Ω h k ) + u k u * L 2 ( Ω h k ) + u * u ¯ k L 2 ( Ω h k ) + u ¯ k u ¯ h k k L 2 ( Ω h k ) C ^ u , k ( h k + δ u , h k k L 2 ( Ω h k ) ) + C * ,
where C ^ u , k , C * are constants in dependent of h k . In the last equality, we used Lemma 2, Lemma 3, the convergence property of ADMM in function space and the inexact ADMM in function space. Then, we know from Proposition 1 and Lemma 1 that
I 6 L 2 ( Ω h k ) = ( S * S S * S h k + S * S h k S h k * S h k ) ( u h k k u ¯ h k k ) L 2 ( Ω h k ) ( S * S S * S h k ) ( u h k k u ¯ h k k ) L 2 ( Ω h k ) + ( S * S h k S h k * S h k ) ( u h k k u ¯ h k k ) L 2 ( Ω h k ) S * S S h k u h k k u ¯ h k k L 2 ( Ω h k ) + S * S h k * S h k u h k k u ¯ h k k L 2 ( Ω h k ) c 6 h k ,
where c 6 is a constant.
Then, we know from (16)–(23) that there are constants C 1 * and C 2 * such that
k = 1 δ u k L 2 ( Ω ) C 1 * k = 1 δ u , h k k L 2 ( Ω h k ) + C 2 * k = 1 h k .
Moreover, the mesh sizes { h k + 1 } k = 0 of each mhADMM iteration satisfy k = 0 h k + 1 < , the residuals of each mhADMM iteration satisfy k = 0 δ u , h k + 1 k + 1 L 2 ( Ω h k + 1 ) k = 0 ϵ k + 1 < , thus we have
k = 1 δ u k L 2 ( Ω ) < .
Then, we know from Theorem 2 that the global convergence and the iteration complexity results o ( 1 / k ) for Algorithm 2 are guaranteed.    □

4. Numerical Experiments

In this section, we illustrate the numerical performance of the mhADMM algorithm for the elliptic PDE-constrained optimization problems with L 1 -control cost. For our numerical experiment, we used MATLAB R2021b with the FEM package iFEM [25] on a Thinkpad laptop with 2.8 GHz Intel Core i7 processor with 16GB of RAM.
In the mhADMM algorithm, the accuracy of a numerical solution is measured by the following residual. Let ϵ be a given accuracy tolerance, we terminate the algorithm when η < ϵ , where η : = max { η 1 , η 2 , η 3 , η 4 , η 5 } , where
η 1 : = K h y M h u M h y r 1 + M h y r , η 2 : = M h ( u z ) 1 + u , η 3 : = M h ( y y d ) + K h p 1 + M h y d , η 4 : = α M h u M h p + M h λ 1 + u , η 5 : = z Π [ a , b ] N h soft W h 1 M h λ , β 1 + z .
To present the finite element error estimates’ results, we introduce the experimental order of convergence, a brief EOC defined by
EOC : = log E h 1 log E h 2 log h 1 log h 2 ,
where h 1 , h 2 > 0 , h 1 h 2 denotes different grid sizes, E denotes the positive error functional
E ( h ) : = u u h L 2 ( Ω ) .
We note that if E ( h ) = O h γ holds, then EOC γ .
As shown in Section 2.1, instead of using the standard piecewise linear and continuous finite elements, nodal quadrature formulas are used to approximately discretize the L 1 -norm and L 2 -norm in Algorithm 2. In both examples, the mhADMM algorithm, the ihADMM algorithm and the classical ADMM algorithm are employed to obtain numerical solutions of different grid sizes. For both numerical examples and all algorithms, we chose ( u 0 , z 0 , λ 0 ) = ( 0 , 0 , 0 ) as the initial values. The penalty parameter σ was chosen as σ = α . For the step length τ , we chose τ = 1.618 . We terminate the algorithms when the residual η < 10 6 with the maximum number of iterations set to 500.
In numerical experiments, we show the numerical results for different final mesh sizes. In Table 1 and Table 2, h denotes the final mesh size, ‘#dofs’ denotes the dimension of the control variable on each grid level, and ‘iter’ represents the times of iteration. To guarantee the sequence { ϵ k + 1 } k = 0 [ 0 , + ) satisfies k = 0 ϵ k + 1 M h k + 1 2 2 < , and the mesh sizes { h k } k = 0 [ 0 , + ) of each iteration satisfy k = 0 h k + 1 , we choose ϵ k + 1 = C ( k + 1 ) 2 , where C is a constant and h k = 2 2 k + 3 , k Z , k 1 in both examples. Moreover, we would like to point out that, in the iteration of mhADMM algorithm, once the mesh size h k reaches the final mesh size h, we continue the iteration in the final mesh until the stopping criterion above is satisfied.
Before providing examples, we first introduce the following algorithm, which can help us formulate sparse optimal control problems.
According to the first-order optimality condition given in Theorem 1, it is easy to see that Algorithm 4 provides a construction strategy for problems with known optimal solutions ( y * , u * ) .
Algorithm 4 Construct the optimal control problem
  • Step 1 Choose y * H 0 1 ( Ω ) and p * H 0 1 ( Ω ) arbitrarily.
  • Step 2
    u * : = Π U a d 1 α soft p * , β = min p * β α , b ,           on x Ω : p * ( x ) > β , max p * + β α , a ,           on x Ω : p * ( x ) < β 0 ,                                                             elsewhere .
  • Step 3 Set y r = S 1 y * u * and y d = y * ( S * ) 1 p .
Example 1. 
Consider
min ( y , u ) H 0 1 ( Ω ) × L 2 ( Ω ) J ( y , u ) = 1 2 y y d L 2 ( Ω ) 2 + α 2 u L 2 ( Ω ) 2 + β u L 1 ( Ω ) s . t . Δ y = u + y r in Ω , y = 0 on Ω , u U a d = { v ( x ) | a v ( x ) b , a . e on Ω } ,
where Ω = ( 0 , 1 ) 2 , the parameters α = 0.5 , β = 0.5 , a = 0.5 , b = 0.5 . As this is a constructed problem, we set y * = sin π x 1 sin π x 2 and p * = 2 β sin 2 π x 1 exp 0.5 x 1 sin 4 π x 2 . Then, through Algorithm 4, we can obtain the optimal control solution u * = Π U a d 1 α soft p * , β , the source term y r = S 1 y * u * and the desired state y d = y * ( S * ) 1 p * . Thus, we construct the example for which we know the exact solution.
We then test the mhADMM, the ihADMM and the classical ADMM for Example 1. The exact optimal control u and an example for the numerical optimal control obtained by mhADMM on the grid with h = 2 2 7 are shown in Figure 1.
In Table 1, we report the dimension of the control variable on each grid level, the error E of the control u, the EOC, the residual η , the computational time and the number of iterations obtained by the mhADMM, the ihADMM and the classical ADMM. As can be seen from the fourth and fifth column of Table 1, a high-subfigures are correct.precision solution does not improve the accuracy of the discretization error and the EOC. The computational time on the seventh, eighth and ninth columns show that the mhADMM is much faster than the ihADMM and the classical ADMM, especially when the discretization is at a fine level. The mhADMM algorithm can significantly reduce the computational cost and make the algorithm faster. This is mainly because the mhADMM adopts the strategy of gradually refining the grid, while the ihADMM and the classical ADMM compute the problem on a fixed grid size, which is their computational bottleneck. Moreover, the seventh column illustrates the mesh-independent performance of mhADMM; that is, the number of iteration of the mhADMM is independent of the discretization level. Above all, we can see that the mhADMM is much more efficient than the ihADMM and the classical ADMM.
Example 2. 
Consider
min ( y , u ) H 0 1 ( Ω ) × L 2 ( Ω ) J ( y , u ) = 1 2 y y d L 2 ( Ω ) 2 + α 2 u L 2 ( Ω ) 2 + β u L 1 ( Ω ) s . t . Δ y = u in Ω , y = 0 on Ω , u U a d = { v ( x ) | a v ( x ) b , a . e on Ω } ,
where Ω = ( 0 , 1 ) 2 , the parameters α = 10 4 , β = 10 3 , a = 10 , b = 10 . The exact sparse solution of this problem is not known in advance. Instead, we use the numerical solutions computed on the grid with the grid size h = 2 2 9 as reference solutions.
As an example, Figure 2 presents the numerical optimal control for Example 2 on the grid with h = 2 2 7 .
Table 2 shows the dimension of the control variable at each grid level, the error E of the control u, the EOC, the residual η , the computational time and the number of iterations for three methods. Similar to Example 1, the mhADMM still outperforms the ihADMM and the classical ADMM in terms of the computational time. The fourth and fifth column of Table 2 clearly show that a high-precision solution does not improve the accuracy of the discretization error and the EOC. Furthermore, the numerical results in the seventh column also illustrate the mesh-independent performance of our mhADMM algorithm. These results demonstrate that the mhADMM is more efficient than the ihADMM and the classical ADMM.

5. Conclusions

In this paper, we propose a new, efficient, multilevel, heterogeneous ADMM (mhADMM) algorithm for solving sparse elliptic PDE-constrained optimal control problems with L 1 -control cost and box constraints on the control. Specifically, the inexact ADMM is first applied in the function space. Then, we propose the strategy of gradually refining the grid and employ the standard piecewise linear finite element to discretize the related subproblems appearing in each iteration of the inexact ADMM algorithm. Moreover, nodal quadrature formulas are utilized to approximately discretize the L 1 -norm and L 2 -norm to overcome the difficulty that the L 1 -norm does not have a decoupled form. Finally, subproblems are solved by appropriate numerical methods. Theoretical results regarding the global convergence and iteration complexity are presented. In our numerical experiments, we show that the proposed mhADMM is superior to the ihADMM and the classical ADMM in terms of the efficiency.

Author Contributions

Conceptualization, X.C. and Z.C.; methodology, X.S.; software, X.C.; validation, X.C. and L.X.; formal analysis, X.C. and X.S.; writing—original draft preparation, X.C.; writing—review and editing, L.X. and Z.C.; visualization, X.C.; supervision, X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 11971092), the National Natural Science Foundation of China (No. 42274166), the China Postdoctoral Science Foundation of China (No. 2020M670717), the Fundamental Research Funds for the Central Universities (No. N2105019), the Fundamental Research Funds for the Central Universities (No. 3132022200) and the Fundamental Research Funds for the Central Universities.

Data Availability Statement

The data presented in this study are available on request from the first author and corresponding author.

Acknowledgments

We will thank the reviewers for taking time off their busy schedule to review this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Stadler, G. Elliptic optimal control problems with L1-control cost and applications for the placement of control devices. Comp. Optim. Appls. 2009, 44, 159–181. [Google Scholar] [CrossRef]
  2. Ciaramella, G.; Borzì, A. A LONE code for the sparse control of quantum systems. Comput. Phys. Commun. 2016, 200, 312–323. [Google Scholar] [CrossRef]
  3. Garcke, H.; Lam, K.F.; Signori, A. Sparse optimal control of a phase field tumor model with mechanical effects. SIAM J. Control Optim. 2021, 59, 1555–1580. [Google Scholar] [CrossRef]
  4. Casas, E.; Tröltzsch, F. Sparse optimal control for a semilinear heat equation with mixed control-state constraints-regularity of Lagrange multipliers. ESAIM Control Optim. Calc. Var. 2021, 27, 2. [Google Scholar] [CrossRef]
  5. Porcelli, M.; Simoncini, V.; Stoll, M. Preconditioning PDE-constrained optimization with L1- sparsity and control constraints. Comput. Math. Appl. 2017, 74, 1059–1075. [Google Scholar] [CrossRef]
  6. Schindele, A.; Borzì, A. Proximal methods for elliptic optimal control problems with sparsity cost functional. Appl. Math. 2016, 7, 967–992. [Google Scholar] [CrossRef] [Green Version]
  7. Song, X.; Yu, B.; Wang, Y.; Zhang, X. An FE-inexact heterogeneous ADMM for elliptic optimal control problems with L1-control cost. J. Syst. Sci. Complex. 2018, 31, 1659–1697. [Google Scholar] [CrossRef] [Green Version]
  8. Song, X.; Yu, B. A two-phase strategy for control constrained elliptic optimal control problem. Numer. Linear Algebra Appl. 2018, 25, e2138. [Google Scholar] [CrossRef] [Green Version]
  9. Chen, Z.; Song, X.; Zhang, X.; Yu, B. A FE-ADMM algorithm for Lavrentiev-regularized state-constrained elliptic control problem. ESAIM Control Optim. Calc. Var. 2019, 25, 5. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, K.; Li, J.; Song, Y.; Wang, X. An alternating direction method of multipliers for elliptic equation constrained optimization problem. Sci. Chin. Math. 2017, 60, 361–378. [Google Scholar] [CrossRef]
  11. Li, J.; Wang, X.; Zhang, K. An efficient alternating direction method of multipliers for optimal control problems constrained by random Helmholtz equations. Numer. Algorithms 2018, 78, 161–191. [Google Scholar] [CrossRef]
  12. Glowinski, R.; Song, Y.; Yuan, X. An ADMM numerical approach to linear parabolic state constrained optimal control problems. Numer. Math. 2020, 144, 931–966. [Google Scholar] [CrossRef]
  13. Glowinski, R.; Song, Y.; Yuan, X.; Yue, H. Application of the alternating direction method of multipliers to control constrained parabolic optimal control problems and beyond. Ann. Appl. Math. 2022, 38, 115–158. [Google Scholar] [CrossRef]
  14. Shaidurov, V.V. Multigrid Methods for Finite Elements; Kluwer Academic Publics: Dordrecht, The Netherlands, 1995. [Google Scholar]
  15. Bornemann, F.A.; Deuflhard, P. The cascadic multigrid method for elliptic problems. Numer. Math. 1996, 75, 135–152. [Google Scholar] [CrossRef]
  16. Deuflhard, P. Newton Methods for Nonlinear Problems: Affine Invariance and Adaptive Algorithms; Springer: Berlin/Heildeberg, Germany, 2011. [Google Scholar]
  17. Borzì, A.; Schulz, V. Multigrid methods for PDE optimization. SIAM Rev. 2009, 51, 361–395. [Google Scholar] [CrossRef]
  18. Gong, W.; Xie, H.; Yan, N. Adaptive multilevel correction method for finite element approximations of elliptic optimal control problems. J. Sci. Comput. 2017, 72, 820–841. [Google Scholar] [CrossRef]
  19. Chen, X.; Song, X.; Chen, Z.; Yu, B. A multilevel ADMM algorithm for elliptic PDE-constrained optimization problems. Comp. Appl. Math. 2020, 39, 331. [Google Scholar] [CrossRef]
  20. Hinze, M.; Pinnau, R.; Ulbrich, M.; Ulbrich, S. Optimization with PDE Constraints; Springer: Berlin/Heildeberg, Germany, 2009. [Google Scholar]
  21. Wachsmuth, G.; Wachsmuth, D. Convergence and regularization results for optimal control problems with sparsity functional. ESAIM Control Optim. Calc. Var. 2011, 17, 858–886. [Google Scholar] [CrossRef] [Green Version]
  22. Ciarlet, P.G. The Finite Element Method for Elliptic Problems; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2002. [Google Scholar]
  23. Bai, Z.; Benzi, M.; Chen, F.; Wang, Z. Preconditioned MHSS iteration methods for a class of block two-by-two linear systems with applications to distributed control problems. IMA J. Numer. Anal. 2013, 33, 343–369. [Google Scholar] [CrossRef]
  24. Cao, S.; Wang, Z. PMHSS iteration method and preconditioners for Stokes control PDE-constrained optimization problems. Numer. Algorithms 2021, 87, 365–380. [Google Scholar] [CrossRef]
  25. Chen, L. iFEM: An Integrated Finite Element Methods Package in MATLAB; Technical Report; University of California at Irvine: Irvine, CA, USA, 2008. [Google Scholar]
Figure 1. (a) The exact optimal control on the grid with h = 2 2 7 for Example 1. (b) The numerical optimal control obtained by the mhADMM on the grid with h = 2 2 7 for Example 1.
Figure 1. (a) The exact optimal control on the grid with h = 2 2 7 for Example 1. (b) The numerical optimal control obtained by the mhADMM on the grid with h = 2 2 7 for Example 1.
Mathematics 11 00570 g001
Figure 2. The numerical optimal control for Example 2 on the grid with h = 2 2 7 .
Figure 2. The numerical optimal control for Example 2 on the grid with h = 2 2 7 .
Mathematics 11 00570 g002
Table 1. The convergence behavior of the mhADMM, the ihADMM and the classical ADMM for Example 1.
Table 1. The convergence behavior of the mhADMM, the ihADMM and the classical ADMM for Example 1.
Caseh#dofsEEOCIndexmhADMMihADMMADMM
1 2 2 4 2259.66 × 10 2 0.99residual η 8.93 × 10 7 6.69 × 10 7 6.46 × 10 7
time (s)0.120.130.13
#iter201618
2 2 2 5 9614.46 × 10 2 1.05residual η 9.44 × 10 7 6.60 × 10 7 8.63 × 10 7
time (s)0.400.420.53
#iter201824
3 2 2 6 39691.49 × 10 2 1.23residual η 3.30 × 10 7 7.37 × 10 7 9.40 × 10 7
time (s)2.162.678.51
#iter222149
4 2 2 7 16,1294.92 × 10 3 1.32residual η 5.57 × 10 7 5.54 × 10 7 8.87 × 10 7
time (s)14.8618.35262.56
#iter2123120
5 2 2 8 65,0251.65 × 10 3 1.37residual η 7.02 × 10 7 4.61 × 10 7 9.97 × 10 7
time (s)114.33200.437576.14
#iter2025257
6 2 2 9 261,1215.83 × 10 4 1.39residual η 6.91 × 10 7 2.82 × 10 7 1.17 × 10 5
time (s)2457.433850.03279,881.28
#iter2027500
Table 2. The convergence behavior of the mhADMM, the ihADMM and the classical ADMM for Example 2.
Table 2. The convergence behavior of the mhADMM, the ihADMM and the classical ADMM for Example 2.
Caseh#dofsEEOCIndexmhADMMihADMMADMM
1 2 2 4 2259.69 × 10 1 1.19residual η 7.12 × 10 7 7.12 × 10 7 8.04 × 10 7
time (s)0.250.300.33
#iter222235
2 2 2 5 9615.72 × 10 1 0.97residual η 7.71 × 10 7 7.77 × 10 7 9.37 × 10 7
time (s)0.951.101.66
#iter191988
3 2 2 6 39691.64 × 10 1 1.25residual η 6.03 × 10 7 6.02 × 10 7 7.63 × 10 7
time (s)5.035.5735.73
#iter2020198
4 2 2 7 16,1294.44 × 10 2 1.41residual η 8.48 × 10 7 8.25 × 10 7 8.10 × 10 7
time (s)28.1337.70851.48
#iter2020454
5 2 2 8 65,0251.72 × 10 2 1.40residual η 7.15 × 10 7 7.74 × 10 7 1.66 × 10 5
time (s)134.46196.078659.08
#iter2121500
6 2 2 9 261,121--residual η 9.09 × 10 7 9.01 × 10 7 6.17 × 10 5
time (s)1052.591972.46151,623.99
#iter2121500
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, X.; Song, X.; Chen, Z.; Xu, L. A Multilevel Heterogeneous ADMM Algorithm for Elliptic Optimal Control Problems with L1-Control Cost. Mathematics 2023, 11, 570. https://doi.org/10.3390/math11030570

AMA Style

Chen X, Song X, Chen Z, Xu L. A Multilevel Heterogeneous ADMM Algorithm for Elliptic Optimal Control Problems with L1-Control Cost. Mathematics. 2023; 11(3):570. https://doi.org/10.3390/math11030570

Chicago/Turabian Style

Chen, Xiaotong, Xiaoliang Song, Zixuan Chen, and Lijun Xu. 2023. "A Multilevel Heterogeneous ADMM Algorithm for Elliptic Optimal Control Problems with L1-Control Cost" Mathematics 11, no. 3: 570. https://doi.org/10.3390/math11030570

APA Style

Chen, X., Song, X., Chen, Z., & Xu, L. (2023). A Multilevel Heterogeneous ADMM Algorithm for Elliptic Optimal Control Problems with L1-Control Cost. Mathematics, 11(3), 570. https://doi.org/10.3390/math11030570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop