Next Article in Journal
Existence and Uniqueness Results of Fractional Differential Inclusions and Equations in Sobolev Fractional Spaces
Previous Article in Journal
Algebrodynamics: Shear-Free Null Congruences and New Types of Electromagnetic Fields
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Weighted, Relaxed Gradient-Based Iterative Algorithm for the Generalized Coupled Conjugate and Transpose Sylvester Matrix Equations

College of Mathematics and Physics, Center for Applied Mathematics of Guangxi, Guangxi Minzu University, Nanning 530006, China
*
Author to whom correspondence should be addressed.
Axioms 2023, 12(11), 1062; https://doi.org/10.3390/axioms12111062
Submission received: 8 October 2023 / Revised: 10 November 2023 / Accepted: 17 November 2023 / Published: 20 November 2023
(This article belongs to the Section Mathematical Analysis)

Abstract

:
By applying the weighted relaxation technique to the gradient-based iterative (GI) algorithm and taking proper weighted combinations of the solutions, this paper proposes the weighted, relaxed gradient-based iterative (WRGI) algorithm to solve the generalized coupled conjugate and transpose Sylvester matrix equations. With the real representation of a complex matrix as a tool, the necessary and sufficient conditions for the convergence of the WRGI algorithm are determined. Also, some sufficient convergence conditions of the WRGI algorithm are presented. Moreover, the optimal step size and the corresponding optimal convergence factor of the WRGI algorithm are given. Lastly, some numerical examples are provided to demonstrate the effectiveness, feasibility and superiority of the proposed algorithm.

1. Introduction

Matrix equations are often used in mathematics and engineering applications, such as control theory, signal processing and computational mathematics [1,2,3,4]. For example, the forward and backward periodic Sylvester matrix equations (PSMEs) with the following forms
A i X i B i + C i X i + 1 D i = F i , i = 1 , 2 , , ω
and
A i X i + 1 B i + C i X i D i = F i , i = 1 , 2 , , ω ,
are an indispensable part of pole assignment and the design of state observers for linear discrete periodic systems [5]. Thus, studying the computational methods of matrix equations has become an important subject in the field of computational mathematics and control. For matrix equations, computing their exact solutions is very meaningful for many practical problems. However, in many applications, such as stability analysis of control systems, it is usually not necessary to calculate the exact solution, as the approximate solution is sufficient. Therefore, the research of iterative solutions to matrix equations has attracted many researchers [6,7,8,9,10,11,12].
One of the important ways to study the approximate solutions of matrix equations is to establish iterative methods. In recent years, many researchers have proposed a great deal of iterative methods to solve different kinds of Sylvester matrix equations. For the generalized coupled Sylvester matrix equations
j = 1 q A i j X j B i j = F i , i = 1 , 2 , , p ,
where A i j R r i × m j , B i j R n j × s i , F i R r i × s i and X j R m j × n j , i = 1 , 2 , , p ,   j = 1 , 2 , , q , Ding and Chen [13] applied the hierarchical identification principle and introduced the block matrix inner product to propose the GI algorithm. Based on the idea of the GI algorithm, some researchers established many improved versions of the GI algorithm and investigated their convergence properties [14,15]. To improve the convergence rate of the GI algorithm, Zhang [16] proposed the residual norm steepest descent (RNSD), conjugate gradient normal equation (CGNE) and biconjugate gradient stabilized (Bi-CGSTAB) algorithms to solve the matrix in Equation (1). Subsequently, Zhang [17] constructed the full-column rank, full-row rank and reduced-rank gradient-based algorithms, with the main idea of them being to construct an objective function and use the gradient search. Afterward, Zhang and Yin [18] developed the conjugate gradient least squares (CGLS) algorithm for Equation (1), which can be convergent within the finite iteration steps in the absence of round off errors.
The generalized coupled Sylvester-conjugate matrix equations have the following form:
j = 1 q A i j X j B i j + C i j X j ¯ D i j = F i , i = 1 , 2 , , p ,
where A i j , C i j C m i × r j , B i j , D i j C t j × n i , F i C m i × n i and X j C r j × t j , i = 1 , 2 , , p ,   j = 1 , 2 , , q . The matrix in Equation (2) can be regarded as the generalization of Equation (1) in the complex field. For the matrix in Equation (2), Wu et al. [19] extended the GI algorithm to solve it and derived the sufficient condition for the convergence of the GI algorithm. Due to the fact that the sufficient condition in [19] is somewhat conservative, Huang and Ma [20] established the sufficient and necessary conditions for convergence of the GI algorithm based on the properties of the real representation of a complex matrix and the vec operator. Also, they made use of different definitions of the real representation to derive another sufficient and necessary condition for the convergence of the GI algorithm. In [21], Huang and Ma introduced l relaxation factors into the GI algorithm and proposed two relaxed gradient-based iterative (RGI) algorithms. They proved that the RGI algorithms are convergent under suitable restrictions in light of the real representation of a complex matrix and the vec operator. Quite recently, Wang et al. [22] developed a cyclic gradient-based iterative (CGI) algorithm by introducing a modular operator, which is different from previous iterative methods. The most remarkable advantage of the CGI algorithm is that less information is used in each iteration update, which helps save memory and improve efficiency.
In addition, the generalized, coupled Sylvester-transpose matrix equations
j = 1 l A i j X j B i j + C i j X j T D i j = F i , i = 1 , 2 , , s ,
where A i j R m i × r j , C i j R m i × t j , B i j R t j × n i , D i j R r j × n i , F i R m i × n i and X j R r j × t j ,   i = 1 , 2 , , s , j = 1 , 2 , , l are related to fault detection, observer design and so forth. Due to the important role of the generalized, coupled Sylvester-transpose matrix equations in several applied problems, numerous methods have been developed to solve them. For example, Song et al. [23] constructed the GI algorithm for the matrix in Equation (3) by using the principle of hierarchical identification. According to the rank of the related matrices of the matrix in Equation (3), Huang and Ma [24] developed three relaxed gradient-based iterative (RGI) algorithms recently by minimizing the objective functions.
In [25], Beik et al. considered the following matrix equations:
T v X = i = 1 p μ = 1 s 1 A v i μ X i B v i μ + μ = 1 s 2 C v i μ X i T D v i μ + μ = 1 s 3 M v i μ X i ¯ N v i μ + μ = 1 s 4 H v i μ X i H G v i μ = F v ,
where A v i μ , B v i μ , C v i μ , D v i μ , M v i μ , N v i μ , H v i μ , G v i μ and F v , v = 1 , 2 , , N are the known matrices with suitable dimensions in a complex number field, X = X 1 , X 2 , , X p is a group of unknown matrices and X j H represents the conjugate transpose of the matrix X j . The matrix in Equation (4) is quite general and includes several kinds of Sylvester matrix equations, and it can be viewed as a general form of the aforementioned matrices in Equations (1)–(3). By using the hierarchical identification principle, the authors in [25] put forward the GI algorithm over a group of reflexive (anti-reflexive) matrices.
Inspired by the above work, this paper focuses on solving the iterative solution of the generalized coupled conjugate and transpose Sylvester matrix equations:
j = 1 l A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j = M i , i = 1 , 2 , , s ,
where A i j , C i j C m i × r j , B i j , D i j C s j × n i , E i j , G i j C m i × s j , F i j , H i j C r j × n i and M i C m i × n i ,   i = 1 , 2 , , s , j = 1 , 2 , , l are the known matrices and X j C r j × s j for j = 1 , 2 , , l are the unknown matrices that need to be determined. When l = 1 , Wang et al. [26] presented the relaxed gradient iterative (RGI) algorithm for Equation (5), which has four system parameters. Note that the matrices in Equations (1)–(3) are special cases of the matrix in Equation (5), and thus the results obtained in this work contain the existing ones in [20,23,24,27]. Owing to the fact that the convergence speed of the GI algorithm is slow in many cases, it is quite meaningful to further improve the numerical performance of the GI algorithm for the matrix in Equation (5). According to [21,24,26,28,29], it can be seen that the relaxation technique can ameliorate the numerical behaviors of the existing GI-like algorithms. Then, this motivated us to apply the weighted relaxation technique to the GI algorithm. By using different step size factors and the weighted technique, we construct the weighted, relaxed gradient-based iterative (WRGI) algorithm for solving the matrix in Equation (5). The WRGI algorithm contains s relaxation factors. When all relaxation factors are equal, the WRGI algorithm will reduce to the GI one proposed in [25]. In [25], the optimal convergence factor of the GI algorithm was not derived. Compared with the GI algorithm in [25], the WRGI algorithm proposed in this paper can own a higher computational efficiency, and its convergence properties are analyzed in detail, including the convergence conditions, optimal parameters and corresponding optimal convergence factor. The proposed WRGI algorithm has a faster convergence rate than the GI one by adjusting the values of the relaxation factors so it is more conducive to solving matrix equations in control theory, signal processing and computational mathematics, among other applications. The main contributions of this paper are given below:
  • By using a series of step size factors and the weighted relaxation technique, we establish the weighted, relaxed gradient-based iterative (WRGI) algorithm for solving the matrix in Equation (5), which generalizes and improves the existing GI one in [25]. Aside from that, the proposed WRGI algorithm contains the RGI ones in [21,24].
  • We analytically provide the necessary and sufficient condition for the convergence of the proposed WRGI algorithm. Also, the expressions of the optimal step size and the corresponding optimal convergence factor of the WRGI algorithm are derived.
The rest of this paper is organized as follows. In Section 2, some definitions and previous results are given, and the GI algorithm that has been proposed before is reviewed. In Section 3, we first design a new algorithm referred to as the WRGI algorithm to solve the matrix in Equation (5). Then, we prove that the WRGI algorithm is convergent for any started matrices under proper conditions, and we explicitly give the optimal step factor such that the convergence rate of the WRGI algorithm is maximized. Section 4 gives two numerical examples to demonstrate the effectiveness and advantages of the proposed WRGI algorithm. Finally, we give some concluding remarks to end this paper in Section 5.

2. Preliminaries

In this section, as a matter of convenience to discuss the main results of this paper, we describe the following notations which will be used throughout this paper. Let C n × n be the set of all n × n complex matrices. For a given matrix A C n × n , some related notations are the following:
  • A ¯ denotes the conjugate of the matrix A;
  • A T stands for the transpose of the matrix A;
  • A H represents the conjugate transpose of the matrix A;
  • A 1 denotes the inverse of the matrix A;
  • A denotes the Frobenius norm of the matrix A;
  • A 2 denotes the spectrum norm of the matrix A;
  • σ max A indicates the maximum singular value of the matrix A;
  • σ min A indicates the minimum singular value of the matrix A;
  • ρ ( A ) stands for the spectral radius of the matrix A;
  • λ ( A ) stands for the spectrum of the matrix A;
  • rank ( A ) stands for the rank of the matrix A.
Moreover, some useful definitions and lemmas are given below:
Definition 1
([30]). For two matrices E C m × n and F C k × l , the Kronecker product of the matrices E and F is defined as follows:
E F = e 11 F e 12 F e 1 n F e 21 F e 22 F e 2 n F e m 1 F e m 2 F e m n F = [ e i j F ] m × n C m k × n l .
Definition 2
([27]). Let e i n be the n-dimensional column vector whose ith element of e i n is one, and other elements are zero. Then, the vec permutation matrix P ( m , n ) can be defined as follows:
P ( m , n ) = I m e 1 n T I m e 2 n T I m e n n T .
Definition 3
([30]). Let A = [ a 1 , a 2 , , a n ] C m × n , with a i being the ith column of A. The vector-stretching function of A is defined as follows:
vec ( A ) = [ a 1 T , a 2 T , , a n T ] T C m n .
Lemma 1
([20]). Let C Y D = G be a matrix equation, with the matrices C and D being full-column rank and full-row rank, respectively. Then, the iterative solution Y ( l ) produced by the GI iteration process
Y ( l + 1 ) = Y ( l ) + τ C H ( G C Y ( l ) D ) D H
converges to the exact solution Y * of C Y D = G for any initial matrix Y ( 0 ) if and only if
0 < τ < 2 σ max ( C ) 2 σ max ( D ) 2 .
In addition, the optimal step size τ 0 is   
τ 0 = 2 λ max C H C λ max D D H + λ min C H C λ min D D H .
Lemma 2
([30]). Let A C m × n , B C s × t and X C n × s . Then, we have the following:
(i) 
vec ( A X B ) = ( B T A ) vec ( X ) ;
(ii) 
vec ( X T ) = P ( n , s ) vec ( X ) .
Lemma 3
([19]). Let A , B C n × n . If tr ( A ) + tr ( B ) R , then
tr A + tr B = tr A + tr B ¯ = tr A ¯ + tr B ¯ .
Next, we introduce two definitions of the real representation of a complex matrix. For A C m × n , A can be uniquely expressed as A = A 1 + i A 2 C m × n , where A 1 , A 2 R m × n . We define the operators · and · as follows:
A = A 1 A 2 A 2 A 1 , A = A 2 A 1 A 1 A 2 .
It can be seen from Equation (11) that the sizes of A and A are two times that of A. Then, by combining Equation (11) with the definition of the Frobenius norm, we can obtain
A 2 = 2 A 2 .
Aside from that, for the identity matrix I n with the matrix order n, we define the following matrices:
Q n = 0 I n I n 0 , P n = 2 2 i I n I n I n i I n .
The properties of the real representation of several complex matrices are given by the following lemma:
Lemma 4
([21]). Let A C m × n and B C n × r . Then, the following statements hold:
(1) 
( A B ) = A B , ( A T ) = Q n ( A ) T Q m ,
( A H ) = ( A ) T , ( A ¯ ) = Q m A Q n , A = Q m A .
(2) 
If A C n × n is nonsingular, then ( A 1 ) = ( A ) 1 .
(3) 
B 2 = B 2 = B 2 .
(4) 
If n = r , then it holds that ρ ( B ) = ρ ( B ) and ρ ( B ) = ρ B B ¯ .
The lemma below gives the norm relationship between the block matrices and its submatrices:
Lemma 5
([31]). Let B be a block-partitioned matrix with
B = B 11 B 12 B 1 n B 21 B 22 B 2 n B m 1 B m 2 B m n
and B i j be a matrix with a proper size. Then, it holds that
B 2 = σ max B B = B 11 B 12 B 1 n B 21 B 22 B 2 n B m 1 B m 2 B m n = i = 1 m j = 1 n B i j 2 .
If rank ( B ) = 1 , or if B is a vector, then we have
B 2 = B = σ max B .
The following result gives the norm inequality for the p norms of the block matrix and its submatrices, which is essential for analyzing the convergence of the WRGI algorithm:
Lemma 6
([21]). Let B = [ B i j ] m × n be a partitioned matrix with the form in Equation (13), and let the orders of the matrices B i j   ( i = 1 , , m , j = 1 , , n ) be compatible. Then, for any deduced q norm, we have
B q B 11 q B 12 q B 1 n q B 21 q B 22 q B 2 n q B m 1 q B m 2 q B m n q q .

3. The Weighted, Relaxed Gradient-Based Iterative (WRGI) Algorithm and Its Convergence Analysis

In this section, we first propose the weighted, relaxed gradient-based iterative (WRGI) algorithm to solve the matrix in Equation (5) based on the hierarchical identification principle. Then, we discuss the convergence properties of the WRGI algorithm, which include the convergence conditions, optimal step size and the corresponding optimal convergence factor of the WRGI algorithm.
First, we define the following intermediate matrices for i = 1 , , s , p = 1 , , l by applying the hierarchical identification principle:
Π i p = M i j = 1 l A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j + A i p X p B i p ,
γ i p = M i j = 1 l A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j + C i p X p ¯ D i p ¯ ,
Φ i p = M i j = 1 l A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j + E i p X p T F i p T ,
Ω i p = M i j = 1 l A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j + G i p X p H H i p H .
For the sake of the following discussions, we define
Γ i j = A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j ,
Γ i j k = A i j X j k B i j + C i j X j k ¯ D i j + E i j X j T k F i j + G i j X j H k H i j .
Then, Equations (14)–(17) can be written as
Π i p = M i j = 1 l Γ i j + A i p X p B i p ,
γ i p = M i j = 1 l Γ i j + C i p X p ¯ D i p ¯ ,
Φ i p = M i j = 1 l Γ i j + E i p X p T F i p T ,
Ω i p = M i j = 1 l Γ i j + G i p X p H H i p H .
Therefore, the matrix in Equation (5) can be decomposed into the following matrix equations
Π i p = A i p X p B i p ,
γ i p = C i p ¯ X p D i p ¯ ,
Φ i p = F i p T X p E i p T ,
Ω i p = H i p H X p G i p H ,
for i = 1 , , s , p = 1 , , l . By applying Lemma 1 to Equations (22)–(25), we can construct the recursive forms as follows
X p 1 , i k + 1 = X p 1 , i k + μ A i p H Π i p A i p X p 1 , i k B i p B i p H ,
X p 2 , i k + 1 = X p 2 , i k + μ C i p T γ i p C i p ¯ X p 2 , i k D i p ¯ D i p T ,
X p 3 , i k + 1 = X p 3 , i k + μ F i p ¯ Φ i p F i p T X p 3 , i k E i p T E i p ¯ ,
X p 4 , i k + 1 = X p 4 , i k + μ H i p Ω i p H i p H X p 4 , i k G i p H G i p ,
for i = 1 , , s , p = 1 , , l , where μ is a step size factor.
For convenience, we define the following notations:
Γ i j r , i k = A i j X j r , i k B i j + C i j X j r , i k ¯ D i j + E i j X j r , i k T F i j + G i j X j r , i k H H i j , r = 1 , 2 , 3 , 4 .
Substituting Equations (18)–(21) into Equations (26)–(29), respectively, and then using X p 1 , i k , X p 2 , i k , X p 3 , i k and X p 4 , i k to replace X p for i = 1 , s , p = 1 , , l gives
X p 1 , i k + 1 = X p 1 , i k + μ A i p H M i j = 1 l Γ i j 1 , i k B i p H ,
X p 2 , i k + 1 = X p 2 , i k + μ C i p T M i j = 1 l Γ i j 2 , i k ¯ D i p T ,
X p 3 , i k + 1 = X p 3 , i k + μ F i p ¯ M i j = 1 l Γ i j 3 , i k T E i p ¯ ,
X p 4 , i k + 1 = X p 4 , i k + μ H i p M i j = 1 l Γ i j 4 , i k H G i p .
By taking the average of X p 1 , i k + 1 , X p 2 , i k + 1 , X p 3 , i k + 1 and X p 4 , i k + 1 in the above equations, one can obtain the following iterative algorithm:
X p 1 k + 1 = X p 1 k + μ 4 A 1 p H M 1 j = 1 l Γ 1 j k B 1 p H + C 1 p T M 1 j = 1 l Γ 1 j k ¯ D 1 p T + F 1 p ¯ M 1 j = 1 l Γ 1 j k T E 1 p ¯ + H 1 p M 1 j = 1 l Γ 1 j k H G 1 p ,
X p s k + 1 = X p s k + μ 4 A s p H M s j = 1 l Γ s j k B s p H + C s p T M s j = 1 l Γ s j k ¯ D s p T + F s p ¯ M s j = 1 l Γ s j k T E s p ¯ + H s p M s j = 1 l Γ s j k H G s p .
Then, we construct the calculation forms of X p k + 1   ( p = 1 , , l ) by introducing the suitable relaxation parameters and using the balanced strategies
X p k + 1 = α 1 X p 1 k + 1 + α 2 X p 2 k + 1 + + α s X p s k + 1 ,
with α i > 0   ( i = 1 , , s ) being the weighted relaxation factors.
Remark 1.
We apply Lemma 1 to Equations (22)–(25) and then use the same step size factor μ to establish the iterative sequences in Equations (26)–(29), which is conducive to deducing the convergence conditions of the proposed algorithm. It is noteworthy that we also can use different step size factors in Equations (26)–(29) and design a new algorithm, but it may be difficult to derive the convergence conditions of the new algorithm this way. Using different step size factors to construct iterative sequences and proposing a new algorithm will be the direction and focus in our future work.
Based on the above discussions, we obtain the following weighted, relaxed gradient-based iterative (WRGI) algorithm (Algorithm 1) to solve the generalized coupled conjugate and transpose Sylvester matrix equations.
Algorithm 1: The weighted, relaxed gradient-based iterative (WRGI) algorithm
Step 1: Given the matrices A i j , C i j C m i × r j , B i j , D i j C s j × n i , E i j , G i j C m i × s j , F i j ,   H i j C r j × n i and M i C m i × n i , i = 1 , 2 , , s , j = 1 , 2 , , l , two constants ε > 0 and μ > 0 and the relaxation parameters α i > 0 , i = 1 , 2 , , s , choose the initial matrices X p i   0   ( i = 1 , 2 , , s , p = 1 , 2 , , l ) , and set k = 0 ;
Step 2: If δ k = i = 1 s M i j = 1 l Γ i j k 2 i = 1 s M i 2 < ε , then stop; otherwise, go to Step 3;
Step 3: Compute X p ( k + 1 ) using
X p 1 k + 1 = X p 1 k + μ 4 A 1 p H M 1 j = 1 l Γ 1 j k B 1 p H + C 1 p T M 1 j = 1 l Γ 1 j k ¯ D 1 p T + F 1 p ¯ M 1 j = 1 l Γ 1 j k T E 1 p ¯ + H 1 p M 1 j = 1 l Γ 1 j k H G 1 p
X p s k + 1 = X p s k + μ 4 A s p H M s j = 1 l Γ s j k B s p H + C s p T M s j = 1 l Γ s j k ¯ D s p T + F s p ¯ M s j = 1 l Γ s j k T E s p ¯ + H s p M s j = 1 l Γ s j k H G s p
X p k + 1 = α 1 X p 1 k + 1 + α 2 X p 2 k + 1 + + α s X p s k + 1 , p = 1 , , l .
Step 4: Set k = k + 1 and return to Step 2.
In the following, we will discuss the convergence properties of the WRGI algorithm by applying the properties of the real representation of a complex matrix and the Kronecker product of the matrices. For convenience, we introduce the following notations. Let
V = V 1 + V 2
and
R = R 1 + R 2 ,
where
V 1 = B 11 A 11 T + Q s 1 D 11 Q r 1 C 11 T B s 1 A s 1 T + Q s 1 D s 1 Q r 1 C s 1 T B 12 A 12 T + Q s 2 D 12 Q r 2 C 12 T B s 2 A s 2 T + Q s 2 D s 2 Q r 2 C s 2 T B 1 l A 1 l T + Q s l D 1 l Q r l C 1 l T B s l A s l T + Q s l D s l Q r l C s l T ,
R 1 = B 11 T A 11 + D 11 T Q s 1 C 11 Q r 1 B 1 l T A 1 l + D 1 l T Q s l C 1 l Q r l B 21 T A 21 + D 21 T Q s 1 C 21 Q r 1 B 2 l T A 2 l + D 2 l T Q s l C 2 l Q r l B s 1 T A s 1 + D s 1 T Q s 1 C s 1 Q r 1 B s l T A s l + D s l T Q s l C s l Q r l ,
V 2 = Q s 1 E 11 T Q r 1 F 11 + G 11 T H 11 P 2 m 1 , 2 n 1 Q s 1 E s 1 T Q r 1 F s 1 + G s 1 T H s 1 P 2 m s , 2 n s Q s 2 E 12 T Q r 2 F 12 + G 12 T H 12 P 2 m 1 , 2 n 1 Q s 2 E s 2 T Q r 2 F s 2 + G s 2 T H s 2 P 2 m s , 2 n s Q s l E 1 l T Q r l F 1 l + G 1 l T H 1 l P 2 m 1 , 2 n 1 Q s l E s l T Q r l F s l + G s l T H s l P 2 m s , 2 n s ,
R 2 = F 11 T Q r 1 E 11 Q s 1 + H 11 T G 11 P 2 r 1 , 2 s 1 F 1 l T Q r l E 1 l Q s l + H 1 l T G 1 l P 2 r l , 2 s l F 21 T Q r 1 E 21 Q s 1 + H 21 T G 21 P 2 r 1 , 2 s 1 F 2 l T Q r l E 2 l Q s l + H 2 l T G 2 l P 2 r l , 2 s l F s 1 T Q r 1 E s 1 Q s 1 + H s 1 T G s 1 P 2 r 1 , 2 s 1 F s l T Q r l E s l Q s l + H s l T G s l P 2 r l , 2 s l .
Theorem 1.
The matrix in Equation (5) has a unique solution if and only if rank ( R ) = p = 1 l 4 s p r p (that is, R is of a full-column rank) and rank ( R ) = rank ( ( R , f ) ) . The unique solution is given by
x = R T R 1 R T f ,
and the corresponding homogeneous matrix equations
j = 1 l A i j X j B i j + C i j X j ¯ D i j + E i j X j T F i j + G i j X j H H i j = 0 , i = 1 , 2 , , s
have a unique solution X * = X 1 * , X 2 * , , X l * = 0 , 0 , , 0 , where
x = vec X 1 vec X 2 vec X l , f = vec M 1 vec M 2 vec M s
and R is defined by Equation (31).
Proof. 
Applying the real representation of the complex matrix to Equation (5) leads to
j = 1 l A i j X j B i j + C i j Q r j X j Q s j D i j + E i j Q s j X j T Q r j F i j + G i j X j T H i j = M i , i = 1 , 2 , , s .
By using the Kronecker products of the matrices and the vector-stretching operator in the above equations, we have
j = 1 l B i j T A i j + D i j T Q s j C i j Q r j + F i j T Q r j E i j Q s j + H i j T G i j P 2 r j , 2 s j vec X j = vec M i ,
where i = 1 , 2 , , s , which can be equivalently transformed into the following linear system:
R x = f .
Therefore, Equation (33) has a unique solution if and only if R has a full-column rank and rank ( R ) = rank ( ( R , f ) ) . In this case, the unique solution in Equation (32) for the matrix in Equation (5) can be obtained. The conclusions follow immediately. □
Based on the properties of the matrix norms and Algorithm 1, we establish the sufficient condition for the convergence of the proposed WRGI algorithm in the following theorem:
Theorem 2.
Suppose that the matrix in Equation (5) has a unique solution X * = X 1 * , X 2 * , , X l * . Then, the iterative sequences { X p ( k ) }   ( p = 1 , , l ) generated by Algorithm 1 converge to X * for any initial matrices X p 0   ( p = 1 , , l ) if μ satisfies
0 < μ < min 1 i s 8 α i V 2 2 .
Proof. 
Define the error matrices
X ˜ p k = X p k X p * , X ˜ p i k = X p i k X p * , i = 1 , , s , p = 1 , , l ,
and
Γ ˜ i j k = A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j , Z i k = j = 1 l A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j , Z i k = Z i k , i = 1 , 2 , , s .
It follows from Algorithm 1 that
X ˜ p i k + 1 = X p i k + 1 X p * = X p i k X p * + μ 4 A i p H M i j = 1 l Γ i j k B i p H + C i p T M i j = 1 l Γ i j k ¯ D i p T + F i p ¯ M i j = 1 l Γ i j k T E i p ¯ + H i p M i j = 1 l Γ i j k H G i p = X ˜ p i k μ 4 A i p H j = 1 l Γ ˜ i j k B i p H + C i p T j = 1 l Γ ˜ i j k ¯ D i p T + F i p ¯ j = 1 l Γ ˜ i j k T E i p ¯ + H i p j = 1 l Γ ˜ i j k H G i p = X ˜ p i k μ 4 A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p ,
and
X ˜ p k + 1 = X p k + 1 X p * = α 1 X p 1 k + 1 + α 2 X p 2 k + 1 + + α s X p s k + 1 X p * = i = 1 s α i X ˜ p i k + 1 = i = 1 s α i X ˜ p i k μ 4 i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p = X ˜ p k μ 4 i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p .
Taking the Frobenius norm in Equation (36) and using the properties of the norm leads to
X ˜ p k + 1 2 = tr X ˜ p H k + 1 X ˜ p k + 1 = X ˜ p k 2 μ 4 tr i = 1 s α i X ˜ p H k A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p μ 4 tr i = 1 s α i B i p Z i H k A i p + D i p ¯ Z i T k C i p ¯ + E i p T Z i k ¯ F i p T + G i p H Z i k H i p H X ˜ p k + μ 2 16 i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 = X ˜ p k 2 μ 4 tr i = 1 s α i B i p H X ˜ p H k A i p H Z i k μ 4 tr i = 1 s α i D i p T X ˜ p H k C i p T Z i k ¯ μ 4 tr i = 1 s α i E i p ¯ X ˜ p H k F i p ¯ Z i T k μ 4 tr i = 1 s α i G i p X ˜ p H k H i p Z i H k μ 4 tr i = 1 s α i Z i H k A i p X ˜ p k B i p μ 4 tr i = 1 s α i Z i T k C i p ¯ X ˜ p k D i p ¯ μ 4 tr i = 1 s α i Z i k ¯ F i p T X ˜ p k E i p T μ 4 tr i = 1 s α i Z i k H i p H X ˜ p k G i p H + μ 2 16 i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 .
It is not difficult to verify that
tr i = 1 s α i D i p T X ˜ p H k C i p T Z i k ¯ + tr i = 1 s α i Z i T k C i p ¯ X ˜ p k D i p ¯
and
tr i = 1 s α i E i p ¯ X ˜ p H k F i p ¯ Z i T k + tr i = 1 s α i Z i k ¯ F i p T X ˜ p k E i p T
are real. Then, under Lemma 3, it holds that
tr i = 1 s α i D i p T X ˜ p H k C i p T Z i k ¯ + tr i = 1 s α i Z i T k C i p ¯ X ˜ p k D i p ¯ = tr i = 1 s α i D i p H X ˜ p T k C i p H Z i k + tr i = 1 s α i Z i H k C i p X ˜ p k ¯ D i p ,
and
tr i = 1 s α i E i p ¯ X ˜ p H k F i p ¯ Z i T k + tr i = 1 s α i Z i k ¯ F i p T X ˜ p k E i p T = tr i = 1 s α i E i p X ˜ p T k F i p Z i H k + tr i = 1 s α i Z i k F i p H X ˜ p k ¯ E i p H .
With the relations in Equations (38) and (39), it follows from Equation (37) that
X ˜ p k + 1 2 = X ˜ p k 2 μ 4 tr i = 1 s α i B i p H X ˜ p H k A i p H + D i p H X ˜ p T k C i p H + F i p H X ˜ p k ¯ E i p H + H i p H X ˜ p k G i p H Z i k μ 4 tr i = 1 s α i Z i H k A i p X ˜ p k B i p + C i p X ˜ p k ¯ D i p + E i p X ˜ p T k F i p + G i p X ˜ p H k H i p + μ 2 16 i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 .
By making use of the relation in Equation (12), as well as Lemmas 2 and 4, we have
p = 1 l i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 = 1 2 p = 1 l i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 = 1 2 p = 1 l i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 = 1 2 p = 1 l i = 1 s α i A i p T Z i k B i p T + Q r p C i p T Q m i Q m i Z i k Q n i Q n i D i p T Q s p + Q r p F i p Q n i Q n i Z i k T Q m i Q m i E i p Q s p + H i p Z i k T G i p 2 = 1 2 p = 1 l i = 1 s α i A i p T Z i k B i p T + Q r p C i p T Z i k D i p T Q s p + Q r p F i p Z i k T E i p Q s p + H i p Z i k T G i p 2 = 1 2 p = 1 l i = 1 s B i p A i p T + Q s p D i p Q r p C i p T + Q s p E i p T Q r p F i p + G i p T H i p P 2 m i , 2 n i vec α i Z i k 2 2 = 1 2 i = 1 s B i 1 A i 1 T + Q s 1 D i 1 Q r 1 C i 1 T + Q s 1 E i 1 T Q r 1 F i 1 + G i 1 T H i 1 P 2 m i , 2 n i vec α i Z i k i = 1 s B i 2 A i 2 T + Q s 2 D i 2 Q r 2 C i 2 T + Q s 2 E i 2 T Q r 2 F i 2 + G i 2 T H i 2 P 2 m i , 2 n i vec α i Z i k i = 1 s B i l A i l T + Q s l D i l Q r l C i l T + Q s l E i l T Q r l F i l + G i l T H i l P 2 m i , 2 n i vec α i Z i k 2 2 = 1 2 V b 2 2 ,
where V is defined by Equation (30) and
b = vec α 1 Z 1 k T , vec α 2 Z 2 k T , , vec α s Z s k T T .
By applying the properties of the 2-norms of the matrices and vectors, we obtain that
p = 1 l i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 = 1 2 V b 2 2 1 2 V 2 2 b 2 2 .
That aside, in light of the relation in Equation (12), we deduce that
b 2 2 = i = 1 s α i Z i k 2 = 2 i = 1 s α i 2 Z i k 2 .
Substituting Equation (43) into Equation (42) leads to
p = 1 l i = 1 s α i A i p H Z i k B i p H + C i p T Z i k ¯ D i p T + F i p ¯ Z i T k E i p ¯ + H i p Z i H k G i p 2 V 2 2 i = 1 s α i 2 Z i k 2 .
By combining Equations (40) and (44) and changing the order of addition, we derive
p = 1 l X ˜ p k + 1 2 p = 1 l X ˜ p k 2 μ 4 tr i = 1 s α i p = 1 l B i p H X ˜ p H k A i p H + D i p H X ˜ p T k C i p H + F i p H X ˜ p k ¯ E i p H + H i p H X ˜ p k G i p H Z i k μ 4 tr Z i H k i = 1 s α i p = 1 l A i p X ˜ p k B i p + C i p X ˜ p k ¯ D i p + E i p X ˜ p T k F i p + G i p X ˜ p H k H i p + μ 2 16 V 2 2 i = 1 s α i 2 Z i k 2 = p = 1 l X ˜ p k 2 μ 4 tr i = 1 s α i Z i H k Z i k μ 4 tr i = 1 s α i Z i H k Z i k + μ 2 16 V 2 2 i = 1 s α i 2 Z i k 2 = p = 1 l X ˜ p k 2 μ 4 i = 1 s α i tr Z i H k Z i k μ 4 i = 1 s α i tr Z i H k Z i k + μ 2 16 V 2 2 i = 1 s α i 2 Z i k 2 = p = 1 l X ˜ p k 2 μ 2 i = 1 s α i Z i k 2 + μ 2 16 V 2 2 i = 1 s α i 2 Z i k 2 = p = 1 l X ˜ p k 2 μ 2 i = 1 s α i 1 μ 8 α i V 2 2 Z i k 2 p = 1 l X ˜ p k 1 2 μ 2 i = 1 s α i 1 μ 8 α i V 2 2 Z i k 2 + Z i k 1 2 p = 1 l X ˜ p 0 2 μ 2 i = 1 s α i 1 μ 8 α i V 2 2 ω = 0 k Z i ω 2 .
If the parameter μ is chosen to satisfy
0 < μ < min 1 i s 8 α i V 2 2 ,
then
0 < μ 2 i = 1 s α i 1 μ 8 α i V 2 2 ω = 0 Z i ω 2 p = 1 l X ˜ p 0 2 ,
and therefore
ω = 0 Z i ω 2 < + , i = 1 , 2 , , s ,
According to the convergence theorem of series, we have lim ω + Z i ω = 0   ( i = 1 , 2 , , s ) . Having in mind that the matrix in Equation (5) has a unique solution, then it follows from the definition of Z i k in Equation (34) that
lim k + X j k = X j * , j = 1 , 2 , , l .
The proof is completed. □
In the sequel, by applying the properties of the real representation of a complex matrix and the vector-stretching operator, we study the necessary and sufficient condition for the convergence of the WRGI algorithm. This result can be stated as in Theorem 3:
Theorem 3.
Suppose that the matrix in Equation (5) has a unique solution X * = X 1 * , X 2 * , , X l * . Then, the iterative sequences { X p ( k ) }   ( p = 1 , , l ) generated by Algorithm 1 converge to X * for any initial matrices X p 0   ( p = 1 , , l ) if and only if
0 < μ < 8 T 1 2 R 2 2 .
Proof. 
For convenience, we define
X ˜ p k = X ˜ p k , p = 1 , 2 , , l .
From Equation (36) and the definition in Equation (34) for Z i k in Theorem 2, it can be obtained that
X ˜ p k + 1 = X ˜ p k μ 4 i = 1 s α i A i p H j = 1 l A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j B i p H + C i p T j = 1 l A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j ¯ D i p T + F i p ¯ j = 1 l A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j T E i p ¯ + H i p j = 1 l A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j H G i p = X ˜ p k μ 4 i = 1 s α i A i p H j = 1 l A i j X ˜ j k B i j + C i j X ˜ j k ¯ D i j + E i j X ˜ j T k F i j + G i j X ˜ j H k H i j B i p H + C i p T j = 1 l A i j ¯ X ˜ j k ¯ B i j ¯ + C i j ¯ X ˜ j k D i j ¯ + E i j ¯ X ˜ j H k F i j ¯ + G i j ¯ X ˜ j T k H i j ¯ D i p T + F i p ¯ j = 1 l B i j T X ˜ j T k A i j T + D i j T X ˜ j H k C i j T + F i j T X ˜ j k E i j T + H i j T X ˜ j k ¯ G i j T E i p ¯ + H i p j = 1 l B i j H X ˜ j H k A i j H + D i j H X ˜ j T k C i j H + F i j H X ˜ j k ¯ E i j H + H i j H X ˜ j k G i j H G i p = X ˜ p k μ 4 i = 1 s j = 1 l α i A i p H A i j X ˜ j k B i j B i p H + A i p H C i j X ˜ j k ¯ D i j B i p H + A i p H E i j X ˜ j T k F i j B i p H + A i p H G i j X ˜ j H k H i j B i p H + C i p T A i j ¯ X ˜ j k ¯ B i j ¯ D i p T + C i p T C i j ¯ X ˜ j k D i j ¯ D i p T + C i p T E i j ¯ X ˜ j H k F i j ¯ D i p T + C i p T G i j ¯ X ˜ j T k H i j ¯ D i p T + F i p ¯ B i j T X ˜ j T k A i j T E i p ¯ + F i p ¯ D i j T X ˜ j H k C i j T E i p ¯ + F i p ¯ F i j T X ˜ j k E i j T E i p ¯ + F i p ¯ H i j T X ˜ j k ¯ G i j T E i p ¯ + H i p B i j H X ˜ j H k A i j H G i p + H i p D i j H X ˜ j T k C i j H G i p + H i p F i j H X ˜ j k ¯ E i j H G i p + H i p H i j H X ˜ j k G i j H G i p .
Then, by combining the real representation with Equation (47) and applying Lemma 4, we obtain
X ˜ p k + 1 = X ˜ p k μ 4 i = 1 s j = 1 l α i A i p T A i j X ˜ j k B i j B i p T + A i p T C i j Q r j X ˜ j k Q s j D i j B i p T + A i p T E i j Q s j X ˜ j k T Q r j F i j B i p T + A i p T G i j X ˜ j k T H i j B i p T + Q r p C i p T Q m i Q m i A i j Q r j Q r j X ˜ j k Q s j Q s j B i j Q n i Q n i D i p T Q s p + Q r p C i p T Q m i Q m i C i j Q r j X ˜ j k Q s j D i j Q n i Q n i D i p T Q s p + Q r p C i p T Q m i Q m i E i j Q s j X ˜ j k T Q r j F i j Q n i Q n i D i p T Q s p + Q r p C i p T Q m i Q m i G i j Q s j Q s j X ˜ j k T Q r j Q r j H i j Q n i Q n i D i p T Q s p + Q r p F i p Q n i Q n i B i j T Q s j Q s j X ˜ j k T Q r j Q r j A i j T Q m i Q m i E i p Q s p + Q r p F i p Q n i Q n i D i j T Q s j X ˜ j k T Q r j C i j T Q m i Q m i E i p Q s p + Q r p F i p Q n i Q n i F i j T Q r j X ˜ j k Q s j E i j T Q m i Q m i E i p Q s p + Q r p F i p Q n i Q n i H i j T Q r j Q r j X ˜ j k Q s j Q s j G i j T Q m i Q m i E i p Q s p + H i p B i j T X ˜ j k T A i j T G i p + H i p D i j T Q s j X ˜ j k T Q r j C i j T G i p + H i p F i j T Q r j X ˜ j k Q s j E i j T G i p + H i p H i j T X ˜ j k G i j T G i p = X ˜ p k μ 4 i = 1 s j = 1 l α i A i p T A i j X ˜ j k B i j B i p T + A i p T C i j Q r j X ˜ j k Q s j D i j B i p T + A i p T E i j Q s j X ˜ j k T Q r j F i j B i p T + A i p T G i j X ˜ j k T H i j B i p T + Q r p C i p T A i j X ˜ j k B i j D i p T Q s p + Q r p C i p T C i j Q r j X ˜ j k Q s j D i j D i p T Q s p + Q r p C i p T E i j Q s j X ˜ j k T Q r j F i j D i p T Q s p + Q r p C i p T G i j X ˜ j k T H i j D i p T Q s p + Q r p F i p B i j T X ˜ j k T A i j T E i p Q s p + Q r p F i p D i j T Q s j X ˜ j k T Q r j C i j T E i p Q r p + Q r p F i p F i j T Q r j X ˜ j k Q s j E i j T E i p Q s p + Q r p F i p H i j T X ˜ j k G i j T E i p Q s p + H i p B i j T X ˜ j k T A i j T G i p + H i p D i j T Q s j X ˜ j k T Q r j C i j T G i p + H i p F i j T Q r j X ˜ j k Q s j E i j T G i p + H i p H i j T X ˜ j k G i j T G i p .
By taking the vector-stretching operator on both sides of Equation (48) and using Lemma 2, we have
vec X ˜ p k + 1 = vec X ˜ p k μ 4 i = 1 s α i B i p A i p T + Q s p D i p Q r p C i p T + P 2 s p , 2 r p Q r p F i p Q s p E i p T + H i p G i p T j = 1 l B i j T A i j + D i j T Q s j C i j Q r j + F i j T Q r j E i j Q s j + H i j T G i j P 2 r j , 2 s j vec X ˜ j k ,
for p = 1 , 2 , , l . This gives the following result:
vec X ˜ k + 1 = vec X ˜ k μ 4 R T T R vec X ˜ k = I μ 4 R T T R vec X ˜ k = I μ 4 R T T 1 2 T 1 2 R vec X ˜ k ,
where
vec X ˜ k = vec X ˜ 1 k T , vec X ˜ 2 k T , , vec X ˜ l k T T ,
T = α 1 I 4 m 1 n 1 0 0 0 α 2 I 4 m 2 n 2 0 0 0 α s I 4 m s n s
and R is defined by Equation (31).
Equation (50) implies that the sufficient and necessary condition for the convergent of Algorithm 1 (the WRGI algorithm) is
ρ I μ 4 R T T 1 2 T 1 2 R < 1 .
Since R T T 1 2 T 1 2 R is a symmetric matrix, it holds that
λ I μ 4 R T T 1 2 T 1 2 R = 1 μ 4 λ q R T T 1 2 T 1 2 R , q = 1 , 2 , , η = 1 μ 4 σ q 2 T 1 2 R , q = 1 , 2 , , η ,
where η = rank R T R = rank R = rank T 1 2 R . Then, ρ I μ 4 R T T 1 2 T 1 2 R < 1 is equivalent to
1 μ 4 σ q 2 T 1 2 R < 1 , q = 1 , 2 , , η .
It follows from Equation (52) that
1 < 1 μ 4 σ q 2 T 1 2 R < 1 , q = 1 , 2 , , η .
After simple computations, we derive
0 < μ < 8 T 1 2 R 2 2 ,
which completes the proof. □
Remark 2.
When the relaxation parameters α i of the WRGI algorithm are α i = 1 s , i = 1 , 2 , , s , then the condition in Equation (46) reduces to the following condition:
0 < μ < 8 s R 2 2 .
Based on Theorem 3, we will study the optimal step size μ and the corresponding optimal convergence factor of the WRGI algorithm in the following theorem:
Theorem 4.
Assume that the conditions of Theorem 2 are valid. Then, it holds that
X k X * ρ k I μ 4 R T T 1 2 T 1 2 R X 0 X * ,
where
X k = X 1 k , X 2 k , , X l k .
In addition, the optimal convergence factor μ o p t is
μ o p t = 8 σ max 2 T 1 2 R + σ min 2 T 1 2 R .
Under this situation, the convergence rate is maximized, and we have
X k X * cond 2 T 1 2 R 1 cond 2 T 1 2 R + 1 k X 0 X * .
Proof. 
Due to the fact that I μ 4 R T T 1 2 T 1 2 R is symmetric, it follows that
I μ 4 R T T 1 2 T 1 2 R 2 = ρ I μ 4 R T T 1 2 T 1 2 R .
In light of Equations (50) and (56), one has
X ˜ k + 1 = vec X ˜ k + 1 = vec X ˜ k + 1 2 = I μ 4 R T T 1 2 T 1 2 R vec X ˜ k 2 I μ 4 R T T 1 2 T 1 2 R 2 vec X ˜ k 2 = ρ I μ 4 R T T 1 2 T 1 2 R X ˜ k ,
where
X ˜ k = X ˜ 1 k , X ˜ 2 k , , X ˜ l k .
Combining the relations in Equations (12) and (57) results in
X ˜ k + 1 ρ I μ 4 R T T 1 2 T 1 2 R X ˜ k ,
where
X ˜ k = X ˜ 1 k , X ˜ 2 k , , X ˜ l k .
Having in mind that X ˜ p k = X p k X p *   ( p = 1 , 2 , , l ) , then it follows from Equation (58) that
X k X * ρ I μ 4 R T T 1 2 T 1 2 R X k 1 X * ρ 2 I μ 4 R T T 1 2 T 1 2 R X k 2 X * ρ k I μ 4 R T T 1 2 T 1 2 R X 0 X * .
It can be seen from Equation (59) that the smaller the ρ I μ 4 R T T 1 2 T 1 2 R , the faster the convergence rate of the WRGI algorithm. Direct calculations give that ρ I μ 4 R T T 1 2 T 1 2 R is minimized if and only if
1 μ 4 σ min 2 T 1 2 R = μ 4 σ max 2 T 1 2 R 1 ,
from which one can deduce that
μ o p t = 8 σ max 2 T 1 2 R + σ min 2 T 1 2 R .
If μ = μ o p t , then we have
ρ I μ 4 R T T 1 2 T 1 2 R = max i 1 2 σ max 2 T 1 2 R + σ min 2 T 1 2 R λ i R T T 1 2 T 1 2 R = max i 1 2 σ max 2 T 1 2 R + σ min 2 T 1 2 R σ i 2 T 1 2 R = 1 2 σ min 2 T 1 2 R σ max 2 T 1 2 R + σ min 2 T 1 2 R = σ max 2 T 1 2 R σ min 2 T 1 2 R σ max 2 T 1 2 R + σ min 2 T 1 2 R = cond 2 T 1 2 R 1 cond 2 T 1 2 R + 1 .
Substituting Equation (61) into Equation (59) yields Equation (55). The proof is completed. □
Although the convergence conditions of the WRGI algorithm are given in Theorems 2 and 3, the Kronecker product and the real representations of the system matrices are involved in computing V 2 and T 1 2 R 2 . This may bring about a high dimensionality problem and lead to high computing consumption. To overcome this drawback, we will give a sufficient condition for the convergence of the proposed WRGI algorithm below.
Corollary 1.
Suppose that the matrix in Equation (5) has a unique solution X * = X 1 * , X 2 * , , X l * . Then, the iterative sequences { X p ( k ) }   ( p = 1 , , l ) generated by Algorithm 1 converge to X * for the arbitrary initial matrices X p 0   ( p = 1 , , l ) if
0 < μ < 2 i = 1 s j = 1 l α i B i j 2 2 A i j 2 2 + D i j 2 2 C i j 2 2 + F i j 2 2 E i j 2 2 + H i j 2 2 G i j 2 2 .
Proof. 
According to Lemmas 4–6 and the fact that E F 2 = E 2 F 2 , we have
T 1 2 R 2 2 i = 1 s j = 1 l α i B i j T A i j + D i j T Q s j C i j Q r j + α i F i j T Q r j E i j Q s j + H i j T G i j P 2 r j , 2 s j 2 2 i = 1 s j = 1 l α i B i j T A i j + D i j T Q s j C i j Q r j 2 + F i j T Q r j E i j Q s j + H i j T G i j P 2 r j , 2 s j 2 2 i = 1 s j = 1 l 2 α i B i j T A i j + D i j T Q s j C i j Q r j 2 2 + F i j T Q r j E i j Q s j + H i j T G i j P 2 r j , 2 s j 2 2 i = 1 s j = 1 l 2 α i B i j T A i j 2 + D i j T Q s j C i j Q r j 2 2 + F i j T Q r j E i j Q s j P 2 r j , 2 s j 2 + H i j T G i j P 2 r j , 2 s j 2 2 i = 1 s j = 1 l 4 α i B i j T A i j 2 2 + D i j T Q s j C i j Q r j 2 2 + F i j T Q r j E i j Q s j P 2 r j , 2 s j 2 2 + H i j T G i j P 2 r j , 2 s j 2 2 = i = 1 s j = 1 l 4 α i B i j T A i j 2 2 + D i j T Q s j C i j Q r j 2 2 + F i j T Q r j E i j Q s j 2 2 + H i j T G i j 2 2 = i = 1 s j = 1 l 4 α i B i j T 2 2 A i j 2 2 + D i j T Q s j 2 2 C i j Q r j 2 2 + F i j T Q r j 2 2 E i j Q s j 2 2 + H i j T G i j 2 2 = i = 1 s j = 1 l 4 α i B i j 2 2 A i j 2 2 + D i j 2 2 C i j 2 2 + F i j 2 2 E i j 2 2 + H i j 2 2 G i j 2 2 .
Substituting Equation (62) into the relation in Equation (46) gives the conclusion of this corollary. □

4. Numerical Experiments

In this section, two numerical examples are provided to verify the effectiveness and advantage of the proposed WRGI algorithm for solving the matrix in Equation (5) in terms of the number of iterations (IT) and the computing times in seconds (CPU). All the computations were performed in MATLAB R2018b on a personal computer with an AMD Ryzen 7 5800H, where the CPU was 3.20 GHz and the memory was 16.0 GB.
Example 1
([32]). Consider the generalized coupled conjugate and transpose Sylvester matrix equations
A 11 X 1 B 11 + E 12 X 2 T F 12 + C 13 X 3 ¯ D 13 + G 14 X 4 H H 14 = M 1 , A 22 X 2 B 22 + E 23 X 3 T F 23 + C 24 X 4 ¯ D 24 + G 21 X 1 H H 21 = M 2 , A 33 X 3 B 33 + E 34 X 4 T F 34 + C 31 X 1 ¯ D 31 + G 32 X 2 H H 32 = M 3 , A 44 X 4 B 44 + E 41 X 1 T F 41 + C 42 X 2 ¯ D 42 + G 43 X 3 H H 43 = M 4 ,
with the following coefficient matrices:
A 11 = 12 7 i 10 11 i 9 + 10 i 2 32 i 27 3 i 1 3 i 10 + 11 i 3 7 i 14 4 i , B 11 = 17 7 i 8 25 i 13 + 1 i 7 + 4 i 2 9 i 0 + 6 i 7 11 i 4 2 i 7 + 6 i ,
A 22 = 11 9 i 8 7 i 18 2 i 33 25 i 3 + 6 i 5 23 i 7 7 i 3 + 11 i 12 15 i , B 22 = 4 + 13 i 7 14 i 10 + 2 i 12 + 5 i 4 + 3 i 8 16 i 1 5 i 19 + 7 i 7 7 i ,
A 33 = 7 + 6 i 5 + 11 i 4 + 8 i 24 1 i 11 3 i 0 + 22 i 0 7 i 2 9 i 0 6 i , B 33 = 4 9 i 20 + 15 i 23 + 20 i 25 + 4 i 4 2 i 11 + 1 i 14 4 i 4 + 8 i 10 1 i ,
A 44 = 12 + 12 i 2 + 5 i 1 9 i 2 2 i 10 6 i 26 + 3 i 19 + 13 i 3 + 15 i 7 17 i , B 44 = 1 5 i 2 + 9 i 3 + 4 i 16 7 i 1 + 4 i 9 5 i 8 6 i 6 + 4 i 1 ,
C 13 = 5 + 6 i 11 + 7 i 12 + 4 i 15 + 2 i 5 + 7 i 0 14 i 11 + 4 i 9 17 i 2 + 21 i , D 13 = 6 + 3 i 22 + 9 i 10 4 i 16 + 17 i 6 + 2 i 0 + 2 i 14 + 3 i 12 2 i 7 2 i ,
C 24 = 16 + 1 i 0 3 i 7 10 i 3 4 + 6 i 9 + 7 i 9 2 i 19 9 i 6 3 i , D 24 = 6 14 i 7 + 20 i 4 0 19 i 6 8 i 5 + 8 i 12 6 i 0 7 17 i ,
C 31 = 13 2 i 15 + 16 i 12 23 i 10 4 i 10 3 i 15 + 12 i 3 9 i 5 + 20 i 5 + 5 i , D 31 = 6 + 6 i 9 + 6 i 20 19 i 9 + 14 i 1 + 21 i 8 12 i 12 12 i 17 4 i 3 + 8 i ,
C 42 = 1 + 5 i 5 + 25 i 2 4 i 19 11 i 2 6 i 2 + 9 i 10 + 2 i 5 9 i 13 + 10 i , D 42 = 5 0 + 1 i 10 + 16 i 2 1 i 24 + 2 i 6 + 1 i 1 12 i 6 + 14 i 4 13 i ,
E 12 = 14 + 2 i 2 + 5 i 4 3 i 9 + 5 i 3 + 10 i 8 + 6 i 8 1 i 5 + 9 i 21 1 i , F 12 = 13 + 1 i 21 + 6 i 11 3 i 2 + 7 i 10 + 4 i 10 + 10 i 3 + 5 i 9 11 i 13 + 13 i ,
E 23 = 8 2 i 8 + 2 i 0 5 i 3 + 6 i 8 2 i 3 1 i 10 3 i 13 + 3 i 14 + 7 i , F 23 = 16 27 i 26 + 2 i 12 10 i 6 8 i 2 + 33 i 1 9 i 5 12 i 12 18 i 29 + 7 i ,
E 34 = 14 + 21 i 9 + 8 i 4 14 i 7 + 7 i 19 2 i 0 + 5 i 7 + 10 i 5 + 8 i 7 18 i , F 34 = 0 + 8 i 16 12 i 8 + 17 i 5 + 18 i 4 + 12 i 10 + 9 i 10 1 i 12 14 i 17 5 i ,
E 41 = 3 + 8 i 18 + 10 i 9 + 14 i 17 + 6 i 3 + 6 i 14 + 7 i 5 5 17 i 6 + 5 i , F 41 = 5 1 i 16 + 3 i 5 13 i 19 + 6 i 12 9 3 i 12 + 4 i 6 12 i 6 + 3 i ,
G 14 = 2 29 i 6 3 i 8 + 8 i 3 + 3 i 6 3 i 1 14 i 1 3 11 i 1 + 19 i , H 14 = 8 8 i 10 2 i 5 + 6 i 1 3 + 6 i 16 + 6 i 9 + 8 i 9 3 2 i ,
G 21 = 15 + 15 i 5 + 19 i 16 4 i 7 14 i 5 + 12 i 11 16 i 1 + 5 i 5 3 i 1 3 i , H 21 = 4 + 15 i 18 + 3 i 6 10 i 6 1 i 4 3 i 9 16 i 10 + 2 i 17 + 12 i 2 8 i ,
G 32 = 14 4 i 25 3 i 5 2 i 4 5 i 2 16 i 3 2 i 7 22 i 2 + 5 i 7 18 i , H 32 = 10 + 5 i 8 + 1 i 12 9 i 12 3 i 6 7 i 3 + 13 i 4 3 i 7 4 i 6 10 i ,
G 43 = 5 + 21 i 1 + 20 i 4 + 12 i 7 8 i 6 6 i 2 + 15 i 13 + 5 i 22 + 4 i 2 + 5 i , H 43 = 13 18 i 3 26 i 3 8 i 2 + 16 i 9 4 i 11 5 i 9 + 7 i 5 + 4 i 2 14 i ,
M 1 = 2418 + 3322 i 10353 966 i 5238 + 1933 i 11927 7210 i 7206 12568 i 4614 + 7638 i 1619 9753 i 16692 + 11938 i 3798 2865 i ,
M 2 = 4750 + 14828 i 10137 3634 i 18315 + 2472 i 11651 + 15269 i 11063 9783 i 16515 + 18210 i 2388 + 17370 i 2934 1222 i 3823 + 2947 i ,
M 3 = 22162 7358 i 22122 18790 i 3570 2827 i 7263 + 4634 i 4142 + 7164 i 6578 + 26838 i 12844 6822 i 11828 13695 i 9789 20700 i ,
M 4 = 5068 9357 i 6306 13376 i 3738 6683 i 9215 + 5146 i 8946 8825 i 14012 7731 i 2927 1660 i 5342 + 4327 i 6279 2721 i .
It can be verified that the unique solution of this matrix equation is
X 1 * = 9 + 8 i 12 5 i 9 + 7 i 6 + 7 i 10 4 i 14 + 2 i 6 10 i 10 + 11 i 1 , X 2 * = 13 + 6 i 19 + 6 i 11 11 i 16 15 i 16 16 i 9 + 11 i 7 10 i 5 + 6 i 5 2 i ,
X 3 * = 13 + 7 i 14 + 3 i 2 + 1 i 3 + 9 i 8 + 3 i 1 + 6 i 15 1 i 6 4 i 5 + 4 i , X 4 * = 2 14 i 9 + 11 i 9 17 i 9 13 i 2 + 13 i 9 + 8 i 10 13 i 2 6 i 7 14 i .
We take the initial iterative matrices
X p i 0 = 10 6 × I 3 , X p 0 = 10 6 × I 3 , i 1 , 4 , p 1 , 4
and set
R E S = M 1 k + M 2 k + M 3 k + M 4 k M 1 0 + M 2 0 + M 3 0 + M 4 0 ξ
to be the termination condition with a constant ξ > 0 , where
M 1 k = M 1 A 11 X 1 k B 11 E 12 X 2 T k F 12 C 13 X 3 k ¯ D 13 G 14 X 4 H k H 14 2 ,
M 2 k = M 2 A 22 X 2 k B 22 E 23 X 3 T k F 23 C 24 X 4 k ¯ D 24 G 21 X 1 H k H 21 2 ,
M 3 k = M 3 A 33 X 3 k B 33 E 34 X 4 T k F 34 C 31 X 1 k ¯ D 31 G 32 X 2 H k H 32 2 ,
M 4 k = M 4 A 44 X 4 k B 44 E 41 X 1 T k F 41 C 42 X 2 k ¯ D 42 G 43 X 3 H k H 43 2 ,
Here, the prescribed maximum iterative number is k m a x = 20,000, and X p k   ( p = 1 , 2 , 3 , 4 ) represents the kth iteration solution.
It follows from Remark 2 that the WRGI algorithm reduces to the GI one [25] as α 1 = α 2 = α 3 = α 4 = 0.25 . In this case, according to Theorem 4, the optimal convergence factor is μ o p t = 4.5603 × 10 6 . However, the optimal convergence factor μ o p t calculated by Theorem 4 may not minimize the IT of the GI algorithm, and the reason for this is that there are errors in the calculation process. Thus, the parameter μ adopted in the GI algorithm is the experimentally found optimal one, which minimizes the IT of the GI algorithm. After experimental debugging, the experimental optimal parameters of the GI algorithm are μ G I 1 = 4.2 × 10 6 , μ G I 2 = 4.53 × 10 6 , μ G I 3 = 4.556 × 10 6 and μ G I 4 = 4.558 × 10 6 when ξ = 0.1 , ξ = 0.01 , ξ = 0.001 and ξ = 0.0001 , respectively. Aside from that, we can obtain that the optimal parameter of the WRGI algorithm is μ = 4.6493 × 10 6 in terms of Theorem 4. Owing to the existence of computation errors, the parameter μ in the WRGI algorithm is adopted to be the experimentally found one as demonstrated above. Through experimental debugging, the experimental optimal parameters of the WRGI algorithm are μ W R G I 1 = 4.2 × 10 6 ,   μ W R G I 2 = 4.61 × 10 6 , μ W R G I 3 = 4.645 × 10 6 and μ W R G I 4 = 4.647 × 10 6 when ξ = 0.1 ,   ξ = 0.01 ,   ξ = 0.001 and ξ = 0.0001 , respectively. Under these circumstances, the relaxation factors are taken to be α 1 = 0.26 , α 2 = 0.26 , α 3 = 0.24 and α 4 = 0.24 .
Table 1 lists the numerical results of the GI and the WRGI algorithms. As observed earlier in the comparisons of the GI and the WRGI algorithms in Table 1, the two tested algorithms were convergent for all cases, and the IT of them increased gradually with the decreases in the value of ξ . Additionally, it can be seen from Table 1 that the proposed WRGI algorithm performed better than the GI one with respect to the IT and CPU times. This implies that by applying the weighted relaxation technique, the WRGI algorithm can own better numerical behaviors than the GI one.
To better validate the advantage of the WRGI algorithm, Figure 1, Figure 2, Figure 3 and Figure 4 describe the convergence curves of the GI and WRGI algorithms for four different values of ξ . These curves reflect that the WRGI algorithm is convergent and its convergence speed is faster than the GI one, except for the case of ξ = 0.1 . Additionally, the advantage of the WRGI algorithm becomes more pronounced as the value of ξ decreases. This is consistent with the results in Table 1.
Example 2.
We consider the generalized coupled conjugate and transpose Sylvester matrix equations
A 11 X 1 B 11 + C 11 X 1 ¯ D 11 + E 12 X 2 T F 12 + G 12 X 2 H H 12 = M 1 , A 21 X 1 B 21 + C 21 X 1 ¯ D 21 + E 22 X 2 T F 22 + G 22 X 2 H H 22 = M 2 ,
with the coefficient matrices
A 11 = 12 5 3 19 8 2 9 3 21 , B 11 = 11 23 30 4 9 3 1 35 20 , C 11 = 8 + 1 i 1 + 21 i 7 1 + 8 i 1 26 1 i 8 41 i , D 11 = 1 1 i 1 1 i 1 1 2 1 i 1 , E 12 = 1 1 i 1 2 i 1 0 0 1 1 , F 12 = 1 1 6 1 0 0 0 1 1 , G 12 = 1 0 1 1 4 1 2 0 2 , H 12 = 1 0 1 0 1 i 0 1 0 1 , A 21 = 1 0 5 i 35 1 45 1 1 1 , B 21 = 1 + 4 i 28 1 1 42 1 1 33 3 , C 21 = 1 i 0 1 1 1 2 0 2 1 , D 21 = 1 1 i 3 1 1 1 1 i 1 i 1 , E 22 = 1 1 i 1 1 i 3 4 2 i 1 2 1 , F 22 = 1 8 2 1 0 0 0 1 3 , G 22 = 1 0 1 1 2 1 2 0 2 H 22 = 1 0 1 1 1 i 1 1 2 1 , M 1 = 136 344 1780 688 i 5777 + 2083 i 1508 670 i 1819 1450 i 8360 + 3079 i 199 1526 i 4844 + 2846 i 4882 + 2714 i , M 2 = 19 76 i 1004 949 i 3 + 166 i 1222 + 1753 i 31239 + 8414 i 2088 + 432 i 28 + 10 i 1133 147 i 69 + 5 i .
The unique solution for these matrix equations is
X 1 * = 12 + 2 i 23 2 i 1 1 8 10 i 2 + 6 i 2 + 2 i 6 , X 2 * = 1 + 2 i 2 3 i 3 + 4 i 2 4 i 6 + 3 i 3 + 2 i 1 + 1 i 2 + 2 i 1 2 i .
In this example, we chose
X p i 0 = 10 6 × I 3 , X p 0 = 10 6 × I 3 , i 1 , 2 , p 1 , 2 ,
as the initial matrices and took
R E S = M 1 k + M 2 k M 1 0 + M 2 0 ξ
as the termination condition with a positive number ξ, where
M 1 k = M 1 A 11 X 1 k B 11 C 11 X 1 k ¯ D 11 E 12 X 2 T k F 12 G 12 X 2 H k H 12 2
and
M 2 k = M 2 A 21 X 1 k B 21 C 21 X 1 k ¯ D 21 E 22 X 2 T k F 22 G 22 X 2 H k H 22 2 ,
and the prescribed maximum iterative number was k m a x = 20,000, while X p k   ( p = 1 , 2 ) denotes the kth iteration solution.
When α 1 = α 2 = 0.5 , the WRGI algorithm degenerates into the GI one, and the optimal convergence factor calculated by Theorem 4 is μ = 1.3151 × 10 6 . When α 1 = 0.6 or α 2 = 0.4 , the optimal parameter of the WRGI algorithm is μ = 1.6269 × 10 6 in light of Theorem 4. Due to the existence of computation errors, as in Example 1, the parameters used in the GI and WRGI algorithms for Example 2 are the experimentally found optimal ones which minimize the IT of the tested algorithms. Through experimental debugging, the experimental optimal parameters of the GI algorithm were μ G I 1 = 1.21 × 10 6 and μ G I 2 = 1.299 × 10 6 for ξ = 0.1 and ξ = 0.01 , respectively. And when α 1 = 0.6 and α 2 = 0.4 , the experimental optimal parameters of the WRGI algorithm were μ W R G I 1 = 1.43 × 10 6 and μ W R G I 2 = 1.625 × 10 6 for ξ = 0.1 and ξ = 0.01 , respectively.
In Table 2, we compare the numerical results of the GI and WRGI algorithms to solve the generalized coupled conjugate and transpose Sylvester matrix equations in Example 2 with respect to two different values of ξ. Since the IT of the algorithms exceeded the maximum number of iterations when ξ = 0.001 and ξ = 0.0001 , we did not list the corresponding numerical results here. From the numerical results listed in Table 2, we can conclude that when the value of ξ decreases, the IT and CPU times of the tested algorithms increase. The changing scope of the IT of the proposed WRGI algorithm was smaller than that for the GI one. In addition, the WRGI algorithm outperformed the GI one from the point of view of computing efficiency, and the advantage of the WRGI algorithm became more pronounced as ξ became small.
Figure 5 describes the convergence curves of the GI and WRGI algorithms for Example 2 with ξ = 0.1 and ξ = 0.01 , respectively. It follows from Figure 5 that the error curves of the WRGI algorithm decreased faster than those for the GI one, which means that the proposed WRGI algorithm is superior to the GI one in terms of IT.

5. Concluding Remarks

In this work, by applying the weighted technique and introducing several relaxation factors into the GI algorithm, the weighted, relaxed gradient-based iterative (WRGI) algorithm was constructed for the generalized coupled conjugate and transpose Sylvester matrix equations. By using the real representation of a complex matrix and the vector-stretching operator, the necessary and sufficient conditions for the convergence of the WRGI algorithm were given, and its optimal convergence factor was settled theoretically. Finally, two numerical examples were offered to show the effectiveness and superiority of the proposed WRGI algorithm.
Note that only one step size factor μ is used in the WRGI algorithm. We will consider to adopt different step size factors in the WRGI algorithm and investigate its algebraic properties. Aside from that, establishing different versions of the WRGI algorithm and their convergent conditions when the system matrix is of a full-row rank, full-column rank or reduced rank deserve to be studied in our future work.

Author Contributions

X.W. wrote the main manuscript text, Z.H. supervised and revised the manuscript, J.C. performed the numerical experiments, and Y.L. examined the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Science Foundations of China (No. 12361078 and 11901123), the Guangxi Natural Science Foundations (No. Guike AD20159056, 2019GXNSFBA185014, 2021GXNSFBA196064 and Guike AD21220129).

Data Availability Statement

Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.

Acknowledgments

We would like to express our sincere thanks to the editor and the anonymous reviewers for their valuable suggestions and constructive comments which greatly improved the presentation of this paper.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Fletcher, L.R.; Kuatslcy, J.; Nichols, N.K. Eigenstructure assignment in descriptor systems. IEEE Trans. Autom. Control. 1986, 31, 1138–1141. [Google Scholar] [CrossRef]
  2. Zhou, B.; Zheng, W.-X.; Duan, G.-R. Stability and stabilization of discrete-time periodic linear systems with actuator saturation. Automatica 2011, 47, 1813–1820. [Google Scholar] [CrossRef]
  3. Zhou, B.; Duan, G.-R. Periodic Lyapunov equation based approaches to the stabilization of continuous-time periodic linear systems. IEEE Trans. Autom. Control 2012, 57, 2139–2146. [Google Scholar] [CrossRef]
  4. Ding, F.; Wang, F.-F.; Xu, L.; Wu, M.-H. Decomposition based least squares iterative identification algorithm for multivariate pseudo-linear ARMA systems using the data filtering. J. Frankl. Inst. 2017, 354, 1321–1339. [Google Scholar] [CrossRef]
  5. Lv, L.-L.; Chen, J.-B.; Zhang, L.; Zhang, F.-R. Gradient-based neural networks for solving periodic Sylvester matrix equations. J. Frankl. Inst. 2022, 359, 10849–10866. [Google Scholar] [CrossRef]
  6. Bai, Z.-Z.; Guo, X.-X.; Yin, J.-F. On two iteration methods for the quadratic matrix equations. Int. J. Numer. Anal. Model. 2005, 2, 114–122. [Google Scholar]
  7. Bai, Z.-Z.; Guo, X.-X.; Xu, S.-F. Alternately linearized implicit iteration methods for the minimal nonnegative solutions of the nonsymmetric algebraic Riccati equations. Numer. Linear Algebra Appl. 2006, 13, 655–674. [Google Scholar] [CrossRef]
  8. Bai, Z.-Z.; Huang, Y.-M.; Ng, M.K. On preconditioned iterative methods for Burgers equations. SIAM J. Sci. Comput. 2007, 29, 415–439. [Google Scholar] [CrossRef]
  9. Dehghan, M.; Hajarian, M. An iterative algorithm for the reflexive solutions of the generalized coupled Sylvester matrix equations and its optimal approximation. Appl. Math. Comput. 2008, 202, 571–588. [Google Scholar]
  10. Deng, Y.-B.; Bai, Z.-Z.; Gao, Y.-H. Iterative orthogonal direction methods for Hermitian minimum norm solutions of two consistent matrix equations. Numer. Linear Algebra Appl. 2006, 13, 801–823. [Google Scholar] [CrossRef]
  11. Guo, K.-H.; Hu, X.-Y.; Zhang, L. A new iteration method for the matrix equation AX = B. Appl. Math. Comput. 2007, 187, 1434–1441. [Google Scholar]
  12. Li, F.-L.; Gong, L.-S.; Hu, X.-Y.; Zhang, L. Successive projection iterative method for solving matrix AX = B. J. Comput. Appl. Math. 2010, 234, 2405–2410. [Google Scholar] [CrossRef]
  13. Ding, F.; Chen, T.-W. On iterative solutions of general coupled matrix equations. Siam J. Control. Optim. 2006, 44, 2269–2284. [Google Scholar] [CrossRef]
  14. Chen, Z.-B.; Chen, X.-S. Modification on the convergence results of the Sylvester matrix equation AX + XB = C. J. Frankl. Inst. 2022, 359, 3126–3147. [Google Scholar] [CrossRef]
  15. Ramadan, M.A.; Bayoumi, A.M.E. A modified gradient-based algorithm for solving extended Sylvester-conjufate matrix equations. Asian J. Control. 2018, 20, 228–235. [Google Scholar] [CrossRef]
  16. Zhang, J.-J. A note on the iterative solutions of general coupled matrix equation. Appl. Math. Comput. 2011, 217, 9380–9386. [Google Scholar] [CrossRef]
  17. Zhang, H.-M. Reduced-rank gradient-based algorithms for generalized coupled Sylvester matrix equations and its applications. Comput. Math. Appl. 2015, 70, 2049–2062. [Google Scholar] [CrossRef]
  18. Zhang, H.-M.; Yin, H.-C. Conjugate gradient least squares algorithm for solving the generalized coupled Sylvester matrix equations. Comput. Math. Appl. 2017, 73, 2529–2547. [Google Scholar] [CrossRef]
  19. Wu, A.-G.; Feng, G.; Duan, G.-R.; Wu, W.-J. Iterative solutions to coupled Sylvester-conjugate matrix equations. Comput. Math. Appl. 2010, 60, 54–66. [Google Scholar] [CrossRef]
  20. Huang, B.-H.; Ma, C.-F. Gradient-based iterative algorithms for generalized coupled Sylvester-conjugate matrix equations. Comput. Math. Appl. 2018, 75, 2295–2310. [Google Scholar] [CrossRef]
  21. Huang, B.-H.; Ma, C.-F. The relaxed gradient-based iterative algorithms for a class of generalized coupled Sylvester-conjugate matrix equations. J. Frankl. Inst. 2018, 355, 3168–3195. [Google Scholar] [CrossRef]
  22. Wang, W.-L.; Qu, G.-R.; Song, C.-Q. Cyclic gradient based iterative algorithm for a class of generalized coupled Sylvester-conjugate matrix equations. J. Frankl. Inst. 2023, 360, 7206–7229. [Google Scholar] [CrossRef]
  23. Song, C.-Q.; Chen, G.-L.; Zhao, L.-L. Iterative solution to coupled Sylvester-transpose matrix equations. Appl. Math. Model. 2011, 35, 4675–4683. [Google Scholar] [CrossRef]
  24. Huang, B.-H.; Ma, C.-F. On the relaxed gradient-based iterative methods for the generalized coupled Sylvester-transpose matrix equations. J. Frankl. Inst. 2022, 359, 10688–10725. [Google Scholar] [CrossRef]
  25. Beik, F.P.A.; Salkuyeh, D.K.; Moghadam, M.M. Gradient-based iterative algorithm for solving the generalized coupled Sylvester-transpose and conjugate matrix equations over reflexive (anti-reflexive) matrices. Trans. Inst. Meas. Control. 2014, 36, 99–110. [Google Scholar] [CrossRef]
  26. Wang, W.-L.; Song, C.-Q.; Ji, S.-P. Iterative solution to a class of complex matrix equations and its application in time-varying linear system. J. Appl. Math. Comput. 2021, 67, 317–341. [Google Scholar] [CrossRef]
  27. Zhang, H.-M.; Yin, H.-C. New proof of the gradient-based iterative algorithm for a complex conjugate and transpose matrix equation. J. Frankl. Inst. 2017, 354, 7585–7603. [Google Scholar] [CrossRef]
  28. Niu, Q.; Wang, X.; Lu, L.-Z. A relaxed gradient based algorithm for solving Sylvester equations. Asian J. Control. 2011, 13, 461–464. [Google Scholar] [CrossRef]
  29. Wang, W.-L.; Song, C.-Q. Iterative algorithms for discrete-time periodic Sylvester matrix equations and its application in antilinear periodic system. Appl. Numer. Math. 2021, 168, 251–273. [Google Scholar] [CrossRef]
  30. Wu, A.-G.; Zhang, Y.; Qian, Y.-Y. Complex Conjugate Matrix Equations; Science Press: Beijing, China, 2017. [Google Scholar]
  31. Khalil, I.S.; Doyle, J.C.; Glover, K. Robust and Optimal Control; Prentice Hall: Hoboken, NJ, USA, 1996. [Google Scholar]
  32. Zhang, H.-M. A finite iterative algorithm for solving the complex generalized coupled Sylvester matrix equations by using the linear operators. J. Frankl. Inst. 2017, 354, 1856–1874. [Google Scholar] [CrossRef]
Figure 1. Convergence curves of the tested algorithms for Example 1 with ξ = 0.1 .
Figure 1. Convergence curves of the tested algorithms for Example 1 with ξ = 0.1 .
Axioms 12 01062 g001
Figure 2. Convergence curves of the tested algorithms for Example 1 with ξ = 0.01 .
Figure 2. Convergence curves of the tested algorithms for Example 1 with ξ = 0.01 .
Axioms 12 01062 g002
Figure 3. Convergence curves of the tested algorithms for Example 1 with ξ = 0.001 .
Figure 3. Convergence curves of the tested algorithms for Example 1 with ξ = 0.001 .
Axioms 12 01062 g003
Figure 4. Convergence curves of the tested algorithms for Example 1 with ξ = 0.0001 .
Figure 4. Convergence curves of the tested algorithms for Example 1 with ξ = 0.0001 .
Axioms 12 01062 g004
Figure 5. Convergence curves of the tested algorithms for Example 2 with ξ = 0.1 (left) and ξ = 0.01 (right).
Figure 5. Convergence curves of the tested algorithms for Example 2 with ξ = 0.1 (left) and ξ = 0.01 (right).
Axioms 12 01062 g005
Table 1. Numerical results of the tested algorithms for Example 1 with different values of ξ .
Table 1. Numerical results of the tested algorithms for Example 1 with different values of ξ .
Algorithm ξ
0.10.010.0010.0001
GIIT17496531212347
CPU0.04500.17841.93723.6098
RES 9.93 × 10 2 1.00 × 10 2 9.9993 × 10 4 9.9969 × 10 5
WRGIIT17492522812128
CPU0.03810.12701.73133.1916
RES 9.9 × 10 2 1.00 × 10 2 9.9998 × 10 4 9.9999 × 10 5
Table 2. Numerical results of the tested algorithms for Example 2 with different values of ξ .
Table 2. Numerical results of the tested algorithms for Example 2 with different values of ξ .
Algorithm ξ
0.10.01
GIIT212495
CPU0.04820.4058
RES 9.82 × 10 2 1.00 × 10 2
WRGIIT152014
CPU0.01790.2831
RES 9.997 × 10 2 1.00 × 10 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, X.; Huang, Z.; Cui, J.; Long, Y. The Weighted, Relaxed Gradient-Based Iterative Algorithm for the Generalized Coupled Conjugate and Transpose Sylvester Matrix Equations. Axioms 2023, 12, 1062. https://doi.org/10.3390/axioms12111062

AMA Style

Wu X, Huang Z, Cui J, Long Y. The Weighted, Relaxed Gradient-Based Iterative Algorithm for the Generalized Coupled Conjugate and Transpose Sylvester Matrix Equations. Axioms. 2023; 12(11):1062. https://doi.org/10.3390/axioms12111062

Chicago/Turabian Style

Wu, Xiaowen, Zhengge Huang, Jingjing Cui, and Yanping Long. 2023. "The Weighted, Relaxed Gradient-Based Iterative Algorithm for the Generalized Coupled Conjugate and Transpose Sylvester Matrix Equations" Axioms 12, no. 11: 1062. https://doi.org/10.3390/axioms12111062

APA Style

Wu, X., Huang, Z., Cui, J., & Long, Y. (2023). The Weighted, Relaxed Gradient-Based Iterative Algorithm for the Generalized Coupled Conjugate and Transpose Sylvester Matrix Equations. Axioms, 12(11), 1062. https://doi.org/10.3390/axioms12111062

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop