Publication Cover
Automatika
Journal for Control, Measurement, Electronics, Computing and Communications
Volume 65, 2024 - Issue 3
143
Views
0
CrossRef citations to date
0
Altmetric
Research Article

The relaxed gradient based iterative algorithm for solving the generalized coupled complex conjugate and transpose Sylvester matrix equations

, , &
Pages 1241-1258 | Received 18 Jul 2023, Accepted 21 May 2024, Published online: 05 Jun 2024

Abstract

Inspired by the idea of Ma et al. (Journal of the Franklin Institute, 2018), we adopt relaxation technique and introduce relaxation factors into the gradient based iterative (GI) algorithm, and the relaxed based iterative (RGI) algorithm is established to solve the generalized coupled complex conjugate and transpose Sylvester matrix equations. By applying the real representation and straighten operation, we contain the sufficient and necessary condition for convergence of the RGI method. In order to effectively utilize this algorithm, we further derive the optimal convergence parameter and some related conclusions. Moreover, to overcome the high dimension calculation problem, a sufficient condition for convergence with less computational complexity is determined. Finally, numerical examples are reported to demonstrate the availability and superiority of the constructed iterative algorithm.

1. Introduction

Solving matrix equations is one of the research focuses of computational mathematics [Citation1–4]. The Sylvester matrix equation is an important type of matrix equations, which has a wide range of applications in control and system theory, pole assignment, model reduction and so further [Citation5–7]. Therefore, finding feasible and effective algorithms for the Sylvester matrix equation has important theoretical significance and practical application value.

In this paper, we aim to find the solution of the generalized coupled complex conjugate and transpose Sylvester matrix equations (1) j=1q(AijYjBij+CijYj¯Dij+EijYjTFij+GijYjHHij)=Mi,iI[1,p],(1) where Aij,CijCmi×rj, Bij,DijCsj×ni, Eij,GijCmi×sj, Fij,HijCrj×ni, MiCmi×ni, iI[1,p], iI[1,q] are the known matrices and YjCrj×sj are unknown matrices to be solved. Equation (Equation1) is involved in both science and engineering. Besides, its form includes many other special matrix equations, such as complex conjugate Sylvester matrix equations, complex transpose Sylvester matrix equations and complex conjugate and transpose Sylvester matrix equations [Citation8–10]. Therefore, it is meaningful to research efficient methods for solving Equation (Equation1).

At present, the methods for solving the Sylvester matrix equation mainly include direct methods and iterative methods. However, when solving high-dimensional matrix equations, the direct methods may lead to lengthy computation time. In order to efficiently solve these matrix equations, we prefer to apply iterative methods. In the past few decades, many scholars are devoted to establishing iterative methods to solve various types of Sylvester matrix equations [Citation7,Citation11–14].

From the previous works [Citation15,Citation16], we know that it is difficult to calculate the exact solution of matrix equations, which consumes large computing costs. In the field of systems and control, calculating approximate solutions is sufficient. So iterative solutions have been widely concerned by researchers, and many researchers paid attention to iterative methods and got excellent results. Ding and Chen developed various iterative algorithms to solve Ax = b, AXB = F and other Sylvester matrix equations [Citation17–19]. Subsequently, many effective iterative methods were proposed. In [Citation10], the least squares based iterative method has been applied to find the solutions of the Sylvester transpose matrix equation AXB+CXTD=F. Xie and Ding constructed the gradient based iterative (GI) methods for the matrix equations AXB + CXD = F [Citation9]. Wu et al. also investigated the GI method for solving the Sylvester conjugate matrix equation AXB+CX¯D=F [Citation20]. Owing to the availability of the GI algorithm, the GI algorithm has been extended to solve general Sylvester matrix equations by researchers. For instance, Wu et al. proposed the GI algorithm to find the solution of coupled Sylvester-conjugate matrix equations [Citation21] (2) i=1l(AijXjBij+CijX¯jDij)=Fi,iI[1,s],(2) where Aij,CijCmi×rj, Bij,DijCtj×ni, FijCmi×ni (iI[1,s], jI[1,l]) are the known matrices. Song et al. applied the GI method to the coupled Sylvester-transpose matrix equations [Citation22] (3) i=1l(AijXjBij+CijXjTDij)=Fi,iI[1,s],(3) where AijRmi×rj, CijRmi×tj, BijCtj×ni, DijCrj×ni, (iI[1,s], jI[1,l]) are the known matrices. The above two matrix equations are important types of matrix equations, which are frequently involved in the fields of systems and control. Not only that, they are also the generalized forms of matrix equations in [Citation10,Citation20], respectively. The convergence properties and the optimal convergence parameters of the GI algorithm for Equations (Equation2) and (Equation3) have been investigated.

Subsequently, Beik et al. proposed the GI algorithm for solving the generalized coupled Sylvester-transpose and conjugate matrix equations [Citation23] (4) Tv(X)=i=1p(μ=1s1AviμXiBviμ+μ=1s2CviμXiTDviμ+μ=1s3MviμXi¯Nviμ+μ=1s4HviμXiHGviμ)=Fv,(4) where Aviμ,Bviμ,Cviμ,Dviμ,Mviμ,Nviμ,Hviμ,Gviμ,Fv (v=1,2,,N) are known matrices with proper dimensions. The form of the above matrix equations is quite general. When p and s1,s2,s3,s4, are taken to be some special values, Equation (Equation4) can be transformed into other matrix equations.

Except for the above classical Sylvester matrix equations, the GI algorithm has also been applied to solve periodic matrix equations [Citation24,Citation25]. Li et al. established the GI method for the forward periodic Sylvester matrix equations and backward forward periodic Sylvester matrix equations [Citation25] (5) AiXiBi+CiXi+1Di=Fi,iI[1,γ],(5) and (6) AiXi+1Bi+CiXiDi=Fi,iI[1,γ],(6) where Ai,Bi,Ci,Di,FiRn×n are the known matrices. In theory, Li et al. proposed the sufficient and necessary conditions for the convergence of the GI algorithm. Numerical experiments have also showed the effectiveness of the GI algorithm.

Although the theory of the GI algorithm has been systematically proposed by researchers, this algorithm still has some drawbacks. In [Citation26], Fan et al. pointed out that the GI algorithm costs large computation time and storage space when encountering ill-posed problems. In order to further optimize the convergence performance of the GI algorithm, the relaxed gradient based iterative (RGI) algorithm has been proposed by introducing the relaxed factor to adjust the weight of the iteration sequences. Niu et al. developed the RGI algorithm to solve the Sylvester matrix equations [Citation27]. Numerical experiments have shown that relaxation techniques can effectively reduce computation time and storage space, and improve the convergence rate of the GI algorithm.

Due to the superiority of the RGI algorithm, many scholars have extended this algorithm to solve more general matrix equations. Recently, in [Citation28,Citation29], Huang et al. applied the RGI algorithm to solve coupled Sylvester-conjugate matrix equation (Equation2) and coupled Sylvester-transpose matrix equation (Equation3). And the experiments results illustrate that the convergence rate of the RGI algorithm is faster than the GI one. Then Wang et al. consider the solution of the complex conjugate and transpose matrix equations (7) A1XB1+A2X¯B2+A3XTB3+A4XHB4=E,(7) where Ai, Bi, ECn×n (iI[1,4]) are the known matrices. By introducing relaxation factors and applying the hierarchical identification principle [Citation30], Wang et al. presented the RGI method to solve Equation (Equation7). However, Wang et al. didn't discuss the generalized form of Equation (Equation7). Based on the ideas of [Citation30], we extend the RGI algorithm to the generalized coupled complex conjugate and transpose Sylvester matrix equations.

Inspired by the idea of [Citation28], we construct the RGI algorithm for solving Equation (Equation1). Its form can be specific written as

(8) {j=1qA1jYjB1j+j=1qC1jYj¯D1j+j=1qE1jYjTF1j+j=1qG1jYjHH1j=M1,j=1qA2jYjB2j+j=1qC2jYj¯D2j+j=1qE2jYjTF2j+j=1qG2jYjHH2j=M2,j=1qApjYjBpj+j=1qCpjYj¯Dpj+j=1qEpjYjTFpj+j=1qGpjYjHHp=Mp.(8) The form of the above matrix equations is quite general, which contains several classic Sylverster matrix equations. Especially, Equations (Equation2)–(Equation3) are special cases of Equation (Equation1). If i = j = p = q = 1, Equation (Equation1) will reduce to Equation (Equation7). Therefore, finding faster algorithms to solve Equation (Equation1) is of great significance.

To accelerate convergence rate of the GI algorithm for Equation (Equation1), we combine relaxation technology with hierarchical identification principle, and we derive the relaxed gradient based iterative (RGI) algorithm to solve Equation (Equation1). This principle regards the unknown matrix as the system parameter matrix to be solved, then it builds a recursive formula to approach the unknown solution [Citation27,Citation28,Citation30,Citation31]. Furthermore, we can effectively control the weight of the iteration sequence by introducing relaxation factors. In theory, we exploit the real representation and the straightening operator to prove the convergence properties of the constructed algorithm. Meanwhile, the sufficient and necessary condition for convergence is presented. Finally, numerical experiments further demonstrate the effectiveness and superiority of the RGI algorithm. The main motivation and contribution of this paper are summarized as follows:

  • In order to accelerate the convergence rate of the GI algorithm [Citation23], we combine the GI algorithm with relaxation technique. By introducing l relaxation factors, we construct the RGI algorithm for Equation (Equation1). Due to that Equation (Equation1) extremely is general, the algorithm constructed in this paper is also more general. It is meaningful to promote the development of the field of solving matrix equations.

  • To optimize convergence theory, we utilize real representation and straighten operation as tool, and present the sufficient and necessary condition for convergence of the RGI method. To overcome high-dimensional computing problems, the sufficient condition for convergence and some related results are proposed. Besides, we use numerical experiments to fully demonstrate the effectiveness and superiority of the RGI algorithm.

The remainder of this paper is structured as follows. In Section 2, we list several useful notations and definitions. Moreover, we construct the relaxed gradient based iterative (RGI) algorithm to find the iterative solution of Equation (Equation1) in Section 3. In Section 4, we deduce the convergence properties of the proposed method, including the sufficient and necessary condition for convergence, the optimal convergence factor and the related corollary. In Section 5, two numerical experiments are reported to validate the superior of convergence for the new algorithm. In the end, Section 6 proposes the some conclusions.

2. Preliminaries

For the sake of convenience, we provide several main notations and lemmas which are used throughout this paper. The set of m×n complex matrix is denoted by Cm×n. For ACm×n, there are some related notations as follows:

  • A¯ indicates the conjugate of the matrix A;

  • AT represents the transpose of the matrix A;

  • AH stands for the conjugate transpose of the matrix A;

  • σmax(A) stands for the maximal singular of the matrix A;

  • σmin(A) stands for the minimal singular of the matrix A;

  • cond(A)=σmax(A)/σmin(A) is defined as the condition number of A;

  • λmax(A) represents the maximal eigenvalue of the matrix A;

  • λmin(A) indicates the minimal eigenvalue of the matrix A;

  • A2 is defined as the the spectral norm of the matrix A.

  • A indicates the Frobenius norm of the matrix A.

  • ρ(A) represents the spectral radius of the matrix A;

Then, some significant definitions and lemmas are listed below.

Definition 2.1

[Citation28]

Let ACm×n, then A can be uniquely expressed as A=A1+iA2 with A1,A2Rm×n. A denotes the real representation of a complex matrix A (9) A=(A1A2A2A1)R2m×2n.(9)

Definition 2.2

[Citation32]

For two matrices A=(aij)Cm×n, B=(bij)Ck×l, the Kronecker product is defined as (10) AB=(a11Ba12Ba1nBa21Ba22Ba2nBam1Bam2BamnB).(10)

Definition 2.3

[Citation28]

Let ein denote an n-dimensional column vector which has 1 in the ith position and 0's elsewhere. The vec-permutation matrix Pmn can be defined as (11) Pmn:=(Ime1nTIme2nTImennT).(11) If X,ACm×n and BCp×q, we have (12) PmnPmn=Imn,PmnT=Pmn1=Pmn,(12) and (13) BA=PmpT(AB)Pnq,(AB)Pnq=Pmp(BA).(13)

Next, we review several lemmas which are used to prove the convergence property.

Lemma 2.1

[Citation33]

If ACm×n, BCs×t, XCn×s, then (14) vec(ABC)=(CTA)vec(B),(14) (15) (AB)(CD)=(AC)(BD).(15)

Lemma 2.2

[Citation29]

For two matrices A and B, it has (16) AB2=A2B2.(16)

Lemma 2.3

[Citation28]

For ACm×r, BCs×n, FCm×n, if the matrix equation AXB = F has unique solution, then the iterative sequences {X(k)} converges to the exact solution X for any initial matrix X(0) by the following algorithm (17) X(k+1)=X(k)+μAH(FAX(k)B)BH,(17) and the algorithm is convergent if and only if (18) 0<μ<2A22B22.(18) Meanwhile, the optimal convergence factor is (19) μ0=2λmax(AHA)λmax(BHB)+λmin(AHA)λmin(BHB).(19)

Proof.

Define error matrix X~(k)=X(k)X.According to the expression (Equation17), it has X~(k+1)=X~(k)μAHAX~(k)BBH.Let Z(k)=AX~(k)B, utilizing the properties of matrix Frobenius norm, Lemmas 2.1 and 2.2, it follows that X~(k+1)2=X~(k)2μtr(BHX~H(k)AHZ(k))μtr(ZH(k)AX~(k)B)+μ2AHZ(k)BH2X~(k)2μtr(ZH(k)Z(k))μtr(ZH(k)Z(k))+μ2(B¯AH)vec(Z(k))2=X~(k)2μ(2μB¯AH22)Z(k)2=X~(k)2μ(2μA22B22)Z(k)2.Repeatedly applying the relationship of the above expression leads to X~(k+1)2X~(k1)2μ(2μA22B22)(Z(k)2+Z(k1)2)X~(0)2μ(2μA22B22)(i=0kZ(i)2).If the convergence parameter μ is selected to satisfy 0<μ<2A22B22,the following inequality holds 0<μ(2μA22B22)i=0Z(i)2X~(0)2.This means that limiZ(i)2=0. Due to that the matrix equation AXB = F has unique solution, then it has limkX~(k)=0. The proof of Equation (Equation18) is completed.

Taking the vec-operator of both sides of the expression (Equation17) and applying Lemma 2.2, it can get vec(X~(k+1))=(IμBBHAHA)vec(X~(k)).The above equation implies that IμBBHAHA is the iterative matrix of the algorithm. Thus, the optimal convergence parameter satisfies the following equation minmax{|1μλ1(BBHAHA)|,,1μλsm(BBHAHA)|}=minmax{|1μλmax(BBHAHA)|,|1μλmin(BBHAHA)|},which means that |1μλmax(BBHAHA)|=|1μλmin(BBHAHA)| has a non-trivial solution. By simple deductions, Expression (Equation19) can be obtained.

Lemma 2.4

[Citation28]

The properties of real representation (.) are as follows:

For two complex matrices ACm×n, BCn×r, then (20) (AB)=AB,(AT)=En(A)TEm,(AH)=(A)T,(A¯)=EmAEn.(20) Here, unitary matrices En is defined as (21) En=(0InIn0).(21) Furthermore, based on the definition of matrix Frobenius norm and real representation, then (22) A2=2A2,(22) (23) A2=A2.(23)

Lemma 2.5

[Citation29]

If mi,iI[1,n] are any given positive number, denote the maximum and minimum values of mi as mmax=max1inmi and mmin=min1inmi, respectively. It has (24) min0<μ<2mmaxmax1in|1μmi|=mmaxmminmmax+mmin,(24) then, the optimal convergence parameter μ is selected as μopt=2mmax+mmin.

Proof.

Build function y=max1in|1μmi|, and then Equation (Equation24) has been obtained by drawing graph in [Citation29]. Besides, |1μmi|<1 if and only if 0<u<2mmax. The optimal convergence factor μ satisfies minmax{|1μm1|,|1μm2|,,|1μmn|}=minmax{|1μm1|,|1μmn|}.The above equation indicates that |1μm|=|1μmn|, that is, 1+μm1=1μmn. By simple calculations, it has μopt=2m1+mn=2mmax+mmin.Thus, the proof is completed.

3. The relaxed gradient-based iterative algorithm

In this section, we mainly propose the relaxed gradient based iterative (RGI) algorithm to solve the generalized coupled complex conjugate and transpose matrix equation. The main idea of this algorithm is to use the hierarchical identification principle to divide Equation (Equation1) into several subsystems. The unknown matrixes Yj are regarded as the identified parameters matrices. Meanwhile, we construct intermediate matrices and adopt an average strategy. Then, the relaxation factors ωl, lI[1,q] are introduced, which are utilized to adjust the weights of matrix schemes. The construction process of the RGI algorithm is as follows.

Firstly, define the following intermediate matrices, iI[1,p], lI[1,q], (25) Πil=Mij=1q(AijYjBij+CijYj¯Dij+EijYjTFij+GijYjHHij)+AilYlBil,(25) (26) Υil=Mij=1q(AijYjBij+CijYj¯Dij+EijYjTFij+GijYjHHij)+CilY¯lDil,¯(26) (27) Φil=(Mij=1q(AijYjBij+CijYj¯Dij+EijYjTFij+GijYjHHij)+EilYlTFil)T,(27) (28) Ψil=(Mij=1q(AijYjBij+CijYj¯Dij+EijYjTFij+GijYjHHij)+GilYlHHil)H.(28) From the expression of Equation (Equation1), some subsystems are given below, iI[1,p], lI[1,q], (29) AilYlBil=Πil,(29) (30) Cil¯YlDil¯=Υil,(30) (31) FilTYlEilT=Φil,(31) (32) HilHYlGilH=Ψil.(32) According to the above fictitious subsystems and Lemma 2.3, we can put forward the iterative schemes as follows, iI[1,p], lI[1,q], (33) Yl1,i(k+1)=Yl1,i(k)+μAilH[ΠilAilYl1,i(k)Bil]BilH,(33) (34) Yl2,i(k+1)=Yl2,i(k)+μCilT[ΥilCil¯Yl2,i(k)Dil¯]DilT,(34) (35) Yl3,i(k+1)=Yl3,i(k)+μFil¯[ΦilFilTYl3,i(k)EilT]Eil¯,(35) (36) Yl4,i(k+1)=Yl4,i(k)+μHil[ΨilHilHYl4,i(k)GilH]Gil.(36) For the sake of convenience, we provide the following notations, sI[1,4], (37) Γijs,i(k)=j=1q(AijYjs,i(k)Bij+CijYjs,i(k)¯Dij+EijYjs,i(k)TFij+GijYjs,i(k)HHij).(37) Combining Equations (Equation25)–(Equation28) with Equations(Equation33)–(Equation36) and utilizing the hierarchical identification principle, the recursive systems are established. Due to that the unknown matrices Yj are included in the expressions, we replace Yj in (Equation25)–(Equation28) with Yj1,i(k),Yj2,i(k),Yj3,i(k),Yj4,i(k), respectively. Therefore, the following expressions are given, iI[1,p], lI[1,q], (38) Xl1,i(k+1)=Xl1,i(k)+μAilH[Mij=1qΓij1,i(k)]BilH,(38) (39) Xl2,i(k+1)=Xl2,i(k)+μCilT[Mij=1qΓij2,i(k)]¯DilT,(39) (40) Xl3,i(k+1)=Xl3,i(k)+μFil¯[Mij=1qΓij3,i(k)]TEil¯,(40) (41) Xl4,i(k+1)=Xl4,i(k)+μHil[Mij=1qΓij4,i(k)]HGil.(41) Then, by taking the average value of Yj1,i(k),Yj2,i(k),Yj3,i(k) and Yj4,i(k), for iI[1,p] we have Yl(k+1)=Yl(k)+μpi=1pAilH×[Mij=1q(AijYj(k)Bij+CijYj(k)¯Dij+EijYj(k)TFij+GijYj(k)HHij)]BilH,Yl′′(k+1)=Yl′′(k)+μpi=1pCilT[Mij=1q(AijYj′′(k)Bij+CijYj′′(k)¯Dij+EijYj′′(k)TFij+GijYj′′(k)HHij)]¯DilT,Ylˇ(k+1)=Ylˇ(k)+μpi=1pFil¯×[Mij=1q(AijYjˇ(k)Bij+CijYjˇ(k)¯Dij+EijYjˇ(k)TFij+GijYjˇ(k)HHij)]TEil¯,Yl´(k+1)=Yl´(k)+μpi=1pHil×[Mij=1q(AijYj´(k)Bij+CijYj´(k)¯Dij+EijYj´(k)TFij+GijYj´(k)HHij)]HGil.Inspire by the idea of the RGI method in [Citation28], we introduce the relaxation factors ωl (lI[1,q]) into the above recursive systems. Based on the previous analysis process, the relaxed gradient-based iterative (RGI) algorithm for Equation (Equation1) is presented as follows.

In the RGI algorithm μ indicates the convergence parameter. The relaxation factors ωl (lI[1,q]) are used to control the weight of iterative sequences, and it can effectively improve the convergence rate of the GI method. In particular, if the relaxed factors are selected as ωl=12 for all l, Algorithm 1 will reduce to the GI algorithm [Citation23]. Besides, Algorithm 1 with ωl=12 and p = q = i = j = 1 will change into the iterative method in [Citation30]. Compared with the RGI algorithm in [Citation28], the new algorithm is more general which includes many kind of iterative formulas.

In what follows, the convergence properties of the RGI method are analysed below. At the same time, we also provide the detailed proof of convergence theory.

4. Convergence analysis of the RGI algorithm

The section presents the sufficient and necessary condition for convergence of the RGI algorithm. Furthermore, to overcome the high-dimensional calculation problem of the iterative matrix, we further discuss the sufficient condition for convergence.

Theorem 4.1

Assume that the generalized coupled complex conjugate and transpose matrix Equation (Equation1) has a unique solution, then the iterative sequences Yl(k) lI[1,q] converge to the exact solution Yl by the RGI algorithm for any initial matrices Yl(0) for all lI[1,q] if and only if the convergence factor μ is selected to satisfy (42) 0<μ<2QM1222,(42) where (43) M=(14ω1(1ω1)I4s1r10014ω2(1ω2)I4s2r2000014ωq(1ωq)I4sqrq),(43) (44) Q=(Q11Q12Q1qQ21Q22Q2qQp1Qp2Qpq),(44) (45) Qij=(Bij)T(Aij)+(Dij)TEsj(Cij)Erj+((Fij)TErj(Eij)Esj)P4rjsj+((Hij)T(Gij))P4rjsj,iI[1,p],lI[1,q].(45)

Proof.

Denote (46) Y~l(1)(k)=Yl(1)(k)Yl,Y~l(2)(k)=Yl(2)(k)Yl,Y~l(3)(k)=Yl(3)(k)Yl,Y~l(4)(k)=Yl(4)(k)Yl,Y~l(k)=Yl(k)Yl,lI[1,q].(46) To facilitate our statement, the expressions of Zi(k) (iI[1,p]) are defined as follows (47) Zi(k)=j=1q(AijY~j(k)Bij+CijY~¯j(k)Dij+EijY~j(k)TFij+GijY~j(k)HHij).(47) From the definition of error matrices and the expression of Yl(1)(k+1) in the RGI algorithm, we derive that for lI[1,q] (48) Y~l(1)(k+1)=Yl(1)(k+1)Yl=Y~l(1)(k)12μωli=1pAilH×[j=1q(AijY~j(k)Bij+CijY~¯j(k)Dij+EijY~j(k)TFij+GijY~j(k)HHijj=1q]BilH=Y~l(1)(k)12μωli=1pAilHZi(k)BilH.(48) It follows from the expression of Yl(2)(k+1) in the RGI method that (49) Y~l(2)(k+1)=Yl(2)(k+1)Yl=Y~l(2)(k)12μωli=1pCilT×[j=1q(AijY~j(k)Bij+CijY~¯j(k)Dij+EijY~j(k)TFij+GijY~j(k)HHij¯]DilT=Y~l(2)(k)12μωli=1pCilTZi(k)¯DilT.(49) By the expression of Yl(3)(k+1) in Algorithm 1, for lI[1,q] one has (50) Y~l(3)(k+1)=Yl(3)(k+1)Yl=Y~l(3)(k)12μ(1ωl)i=1pFil¯×[j=1q(AijY~j(k)Bij+CijY~¯j(k)Dij+EijY~j(k)TFij+GijY~j(k)HHijj=1q]TEil¯=Y~l(3)(k)12μ(1ωl)i=1pFil¯Zi(k)TEil¯.(50) It follows from the expression of Yl(4)(k+1) in Algorithm 1 that for lI[1,q] (51) Y~l(4)(k+1)=Yl(4)(k+1)Yl=Y~l(4)(k)12μ(1ωl)i=1pHil×[j=1q(AijY~j(k)Bij+CijY~¯j(k)Dij+EijY~j(k)TFij+GijY~j(k)HHijj=1q]HGil=Y~l(4)(k)12μ(1ωl)i=1pHilZi(k)HGil.(51) Combing (Equation48)–(Equation51) with Line 5 of the RGI algorithm leads to (52) Y~l(k+1)=Yl(k+1)Yl=12(1ωl)Yl(1)(k+1)+12(1ωl)Yl(2)(k+1)+12ωlYl(3)(k+1)+12ωlYl(4)(k+1)Yl=12(1ωl)Y~l(1)(k+1)+12(1ωl)Y~l(2)(k+1)+12ωlY~l(3)(k+1)+12ωlY~l(4)(k+1)=Y~l(k)14μωl(1ωl)i=1p(AilHZi(k)BilH+CilTZi(k)¯DilT+Fil¯Zi(k)TEil¯+HilZi(k)HGil).(52) According to Lemma 2.4, taking the real representation on both sides of (Equation52) results in (53) (Y~l(k+1))=(Y~l(k))14μωl(1ωl)i=1p(AilHZi(k)BilH+CilTZi(k)¯DilT+Fil¯Zi(k)TEil¯+HilZi(k)HGil)=(Y~l(k))14μωl(1ωl)i=1p×[(AilH)(Zi(k))(BilH)+(CilT)(Zi(k)¯)(DilT)+(Fil¯)(Zi(k)T)(Eil¯)+(Hil)(Zi(k)H)(Gil)]=(Y~l(k))14μωl(1ωl)i=1p×[(Ail)T(Zi(k))(Bil)T+Erl(Cil)TEmiEmi(Zi(k))EniEni(Dil)TEsl+Erl(Fil)EniEni(Zi(k))TEmiEmi(Eil)Esl+(Hil)(Zi(k))T(Gil)]=(Y~l(k))14μωl(1ωl)i=1p×[(Ail)T(Zi(k))(Bil)T+Erl(Cil)T(Zi(k))(Dil)TEsl+Erl(Fil)(Zi(k))T(Eil)Esl+(Hil)(Zi(k))T(Gil)].(53) Using straightening operator in (Equation53) and applying Definition 2.3, for lI[1,q] we have (54) vec[(Y~l(k+1))]=vec[(Y~l(k))]14μωl(1ωl)i=1p×[(Bil)(Ail)T+Esl(Dil)Erl(Cil)T+P4rlslT(Erl(Fil)Esl(Eil)T)+P4rlslT((Hil)(Gil)T)]vec[(Zi(k))].(54) Furthermore, by applying the real representation on two sides of (Equation47), we get (55) (Zi(k))=j=1q(AijY~j(k)Bij+CijY~¯j(k)Dij+EijY~j(k)TFij+GijY~j(k)HHij)=j=1q((Aij)(Y~j(k))(Bij)+(Cij)Erj(Y~j(k))Esj(Dij)+(Eij)Esj((Y~j(k)))TErj(Fij)+(Gij)(Y~j(k))T(Hij)).(55) Then, utilizing the vec-operator in (Equation55) can deduce (56) vec[(Zi(k))]=vec[j=1q((Aij)(Y~j(k))(Bij)+(Cij)Erj(Y~j(k))Esj(Dij)+(Eij)Esj((Y~j(k)))TErj(Fij)+(Gij)(Y~j(k))T(Hij))]=j=1q[(Bij)T(Aij)+(Dij)TEsj(Cij)Erj+((Fij)TErj(Eij)Esj)P4rjsj+((Hij)T(Gij))P4rjsj]vec[Y~j(k)].(56) Finally, substituting (Equation56) into (Equation54) results in (57) vec[(Y~l(k+1))]=vec[(Y~l(k))]14μωl(1ωl)i=1p×[(Bil)(Ail)T+Esl(Dil)Erl(Cil)T+P4rlslT(Erl(Fil)Esl(Eil)T)+P4rlslT((Hil)(Gil)T)]j=1q×[(Bij)T(Aij)+(Dij)TEsj(Cij)Erj+((Fij)TErj(Eij)Esj)P4rjsj+((Hij)T(Gij))P4rjsj]vec[Y~j(k)].(57) Denote (58) vec[(Y~(k))]=[vec((Y~1(k)))T,vec((Y~2(k)))T,,vec((Y~q(k)))T]T.(58) Thus, Equation (Equation57) can be written as the following expression (59) vec[(Y~(k+1))]=vec[(Y~(k))]μMQTQvec[(Y~(k))]=[IμMQTQ]vec[(Y~(k))].(59) The matrices M and Q are given in (Equation43) and (Equation44). It follows from Equation (Equation59) that the matrix [IμMQTQ] is the iterative matrix of Algorithm 1. So the sufficient and necessary condition for convergence of the RGI algorithm is (60) ρ(IμMQTQ)<1.(60) Due to the fact that the iterative matrix MQTQ is similar to M12QTQM12, and M12QTQM12 is symmetric matrix, so one obtains (61) λi(IμMQTQ)=1μλi(M12QTQM12)=1μσi2(QM12),iI[1,p].(61) Since ρ(IμMQTQ)<1, it follows that (62) 1<1μσi2(QM12)<1 or 0<μ<2σi2(QM12),iI[1,p].(62) Finally, the range of convergence parameter μ making the RGI algorithm convergent is (63) 0<μ<2QM1222.(63) Here, we complete the proof of Theorem 4.1.

In order to further effectively utilize the RGI algorithm, we should get the optimal convergence parameter μ of this method. When ρ(IM12QTQM12) reaches minimum value, the convergence behaviour of the RGI method achieves the optimum. According to Lemma 2.5, the necessary and sufficient condition for ρ(IM12QTQM12) is (64) |1μσmin2(QM12)|=|1μσmax2(QM12)|.(64) By simple calculations, the optimal convergence parameter is obtained as (65) μopt=2σmin2(QM12)+σmax2(QM12).(65) Then, we will further discuss the convergence properties of the RGI method with the relaxation parameters ωl=ω for lI[1,q]. Some relevant conclusions are proposed below.

Theorem 4.2

Assume that Y(0)=(Y1(0),Y2(0),,Yq(0)) and Y(k)=(Y1(k),Y2(k),,Yq(k)) represent the initial value and unique solution of the RGI algorithm, respectively. Based on the conditions of Theorem 4.1, if the relaxation factor are selected as ωl=ω for lI[1,q], it holds that (66) Y(k)Yρk(I14μω(1ω)QTQ)Y(0)Y.(66)

And the optimal convergence parameter is (67) μopt=8ω(1ω)(σmax2(Q)+σmin2(Q)).(67) Under this situation, the following inequality holds (68) Y(k)Y(cond2(QM12)1cond2(QM12)+1)kY(0)Y.(68)

Proof.

According to the fact that I14μω(1ω)QTQ is the symmetric matrix, it has (69) I14μω(1ω)QTQ2=ρ(I14μω(1ω)QTQ).(69) Combining Expression (Equation59) with the properties of matrix norms, we derive (70) (Y~(k+1))=vec[(Y~(k+1))]2=[I14μω(1ω)QTQ]vec[(Y~(k))]2[I14μω(1ω)QTQ]2vec[(Y~(k))]2=[I14μω(1ω)QTQ]2vec[(Y~(k))]=ρ(I14μω(1ω)QTQ)(Y~(k)),(70) with (71) (Y~(k))=[(Y~1(k)),(Y~2(k)),,(Y~q(k))].(71) Based on Lemma 2.4 and A2=2A2, it holds (72) Y~(k+1)=12(Y~(k+1))ρ(I14μω(1ω)QTQ)12(Y~(k))=ρ(I14μω(1ω)QTQ)Y~(k).(72) By the definition of the error matrix and Inequality (Equation70), we derive that (73) Y(k)Y=Y~(k)ρk(I14μω(1ω)MQTQ)Y(0)Y.(73) Moreover, when ρ(I14μω(1ω)QTQ) is minimized, the convergence performance of the RGI algorithm can achieve optimal. So we should choose the optimal parameter μopt to minimize ρ(I14μω(1ω)QTQ). The minimum value of ρ(I14μω(1ω)QTQ) is (74) minρ(I14μω(1ω)QTQ)=minmax{|114μω(1ω)σi2(Q)|}=minmax{|114μω(1ω)σmax2(Q)|,|114μω(1ω)σmin2(Q)|},(74) which indicates that |114μω(1ω)σmax2(Q)|=|114μω(1ω)σmin2(Q)| has a non-trivial solution. By simple derivations, the best convergence parameter is (75) μopt=8ω(1ω)(σmax2(Q)+σmin2(Q)).(75) If the convergence parameter μ is selected as the one in (Equation75), it will lead to (76) ρ(I14μω(1ω)QTQ)=max{114μω(1ω)λi(QTQ)}=1148ω(1ω)(σmax2(Q)+σmin2(Q))}×ω(1ω)λmin(QTQ)=12(σmax2(Q)+σmin2(Q))λmin(QTQ)=12σmin2(Q)σmax2(Q)+σmin2(Q)=σmax2(Q)σmin2(Q)σmax2(Q)+σmin2(Q)=cond2(Q)1cond2(Q)+1.(76) Then Equation (Equation68) can be derived by substituting Equation (Equation76) into (Equation73).

Remark 4.1

In Theorem 4.1, the sufficient and necessary condition for convergence of the RGI method is obtained. However, QM1222 involves the calculation of real representation and Kronecker product, which leads to high-dimensional problems. In order to overcome this drawback and develop computational efficiency, we further derive sufficient condition for the convergence with less computational complexity.

Corollary 4.1

Assume that the conditions of Theorem 4.1 are satisfied, then Algorithm 1 is convergent for any initial matrix if the parameters ω and μ are selected to satisfy the following inequality (77) 0<μ2i=1pj=1qωj(1ωj)[Bij22Aij22+Dij22Cij22+Fij22Eij22+Hij22Gij22].(77)

Proof.

By the properties of Frobenius norm of matrix, one has (78) QM1222i=1pj=1q(14ωj(1ωj))(Bij)T(Aij)+(Dij)TEsj(Cij)Erj+((Fij)TErj(Eij)Esj)P4rjsj+((Hij)T(Gij))P4rjsj14ωj(1ωj)22i=1pj=1q14ωj(1ωj)[(Bij)T(Aij)2+(Dij)TEsj(Cij)Erj2+((Fij)TErj(Eij)Esj)P4rjsj2+((Hij)T(Gij))P4rjsj2]2i=1pj=1q12ωj(1ωj)[((Bij)T(Aij)2+(Dij)TEsj(Cij)Erj2)2+(((Fij)TErj(Eij)Esj)P4rjsj2(78) +((Hij)T(Gij))P4rjsj2)2]i=1pj=1qωj(1ωj)[(Bij)T(Aij)22+(Dij)TEsj(Cij)Erj22+((Fij)TErj(Eij)Esj)P4rjsj22+((Hij)T(Gij))P4rjsj22]=i=1pj=1qωj(1ωj)[(Bij)T(Aij)22+(Dij)TEsj(Cij)Erj22+((Fij)TErj(Eij)Esj)22+((Hij)T(Gij))22].Notice the fact that AB2=A2B2,A2=A2, and we have the following inequality (79) QM1222i=1pj=1qωj(1ωj)[(Bij)T22(Aij)22+(Dij)TEsj22(Cij)Erj22+(Fij)TErj22EijEsj22+(Hij)T22(Gij)22]=i=1pj=1qωj(1ωj)[Eni(BijT)Esj22Aij22+Eni(DijT)22Cij22+Eni(FijT)22EijEsj22+Eni(HijT)Eri22Gij22]=i=1pj=1qωj(1ωj)[Bij22Aij22+Dij22Cij22+Fij22Eij22+Hij22Gij22].(79) By combining (Equation79) with (Equation42), the conclusion of Corollary 4.1 is correct.

5. Numerical experimental results

In this section, we present two numerical examples to testify the effectiveness and feasibility of the RGI algorithm proposed in this paper. All experiments are performed on a personal computer with AMD Ryzen 5 5600U with Radeon Graphics 2.30 GHz, 16.0 GB. The programming language is computed in MATLAB R2021b. In our experiment, we compare the convergence behaviour of the RGI algorithm with the GI one in terms of the iterative number (IT), calculation time (CPU) in seconds and the relative error (ERR).

Example 5.1

[Citation34]

We consider the generalized coupled complex conjugate and transpose Sylvester matrix Equation (Equation1) in the case of p = q = 4, and its form is as follows (80) {A11Y1B11+C13Y¯3D13+E12Y2TF12+G14Y4HH14=M1,A22Y2B22+C24Y¯4D24+E23Y3TF23+G21Y1HH21=M2,A33Y3B33+C31Y¯1D31+E34Y4TF34+G32Y2HH32=M3,A44Y4B44+C42Y¯2D42+E41Y1TF41+G43Y3HH43=M4,(80) with A11=[127i1011i9+10i232i273i13i10+11i37i144i],B11=[177i825i13+1i7+4i29i0+6i711i42i7+6i],A22=[119i87i182i3325i3+6i523i77i3+11i1215i],B22=[4+13i714i10+2i12+5i4+3i816i15i19+7i77i],A33=[7+6i5+11i4+8i241i113i0+22i07i29i06i],B33=[49i20+15i23+20i25+4i42i11+1i144i4+8i101i], A44=[12+12i2+5i19i22i106i26+3i19+13i3+15i717i],B44=[15i2+9i3+4i167i1+4i95i86i6+4i1],C13=[5+6i11+7i12+4i15+2i5+7i014i11+4i917i2+21i],D13=[6+3i22+9i104i16+17i6+2i0+2i14+3i122i72i],C24=[16+1i03i710i34+6i9+7i92i199i63i],D24=[614i7+20i4019i68i5+8i126i0717i], C31=[132i15+16i1223i104i103i15+12i39i5+20i5+5i],D31=[6+6i9+6i2019i9+14i1+21i812i1212i174i3+8i],C42=[1+5i5+25i24i1911i26i2+9i10+2i59i13+10i],D42=[50+1i10+16i21i24+2i6+1i112i6+14i413i],E12=[14+2i2+5i43i9+5i3+10i8+6i81i5+9i211i], F12=[13+1i21+6i113i2+7i10+4i10+10i3+5i911i13+13i],E23=[82i8+2i05i3+6i82i31i103i13+3i14+7i],F23=[1627i26+2i1210i68i2+33i19i512i1218i29+7i],E34=[14+21i9+8i414i7+7i192i0+5i7+10i5+8i718i],F34=[0+8i1612i8+17i5+18i4+12i10+9i101i1214i175i], E41=[3+8i18+10i9+14i17+6i3+6i14+7i5517i6+5i],F41=[51i16+3i513i19+6i1293i12+4i612i6+3i],G14=[229i63i8+8i3+3i63i114i1311i1+19i],H14=[88i102i5+6i13+6i16+6i9+8i932i],G21=[15+15i5+19i164i714i5+12i1116i1+5i53i13i],H21=[4+15i18+3i610i61i43i916i10+2i17+12i28i], G32=[144i253i52i45i216i32i722i2+5i718i],H32=[10+5i8+1i129i123i67i3+13i43i74i610i],G43=[5+21i1+20i4+12i78i66i2+15i13+5i22+4i2+5i],H43=[1318i326i38i2+16i94i115i9+7i5+4i214i], M1=[2418+3322i10353966i119277210i720612568i16199753i16692+11938i5238+1933i4614+7638i37982865i],M2=[4750+14828i101373634i11651+15269i110639783i2388+17370i29341222i18315+2472i16515+18210i3823+2947i], M3=[221627358i2212218790i7263+4634i4142+7164i128446822i1182813695i35702827i6578+26838i978920700i],M4=[50689357i630613376i9215+5146i89468825i29271660i5342+4327i37386683i140127731i62792721i].This matrix equation has the following exact solution Y1=[9+8i125i9+7i6+7i104i14+2i610i10+11i1],Y2=[13+6i19+6i1111i1615i1616i9+11i710i5+6i52i],Y3=[13+7i14+3i2+i3+9i8+3i1+6i15i64i5+4i],Y4=[214i9+11i917i913i2+13i9+8i1013i26i714i].

In this example, the initial iterative matrices are taken as Yi(0)=10×I3,iI[1,4], and the relative iterative error is defined as (81) ERR=Y1Y1(k)2+Y2Y2(k)2+Y3Y3(k)2+Y4Y4(k)2Y12+Y22+Y32+Y42,(81) where Yi(k)(iI[1,4]) stand for the kth iteration solution. The iteration is terminated if the relative iterative error is less than δ or the number of the prescribed iteration step kmax=30000 is exceeded. Here, δ is a positive number.

By some calculations, we find that Example 5.1 satisfies the condition of Theorem 4.1. Then the optimal parameter of the RGI algorithm is obtained as μ=5.2559e06 when relaxation factors are chosen as ω1=0.25,ω2=0.52,ω3=0.32,ω4=0.48. However, there are errors in the experiment, and the convergence rate of the RGI method is not the fastest if μ is chosen to be 5.2559e−06. Thus, we try to find the optimal experimental parameter near the value μ=5.2559e06. In Figure , we compare the convergence performance of the RGI algorithm under μ1=5.1499e06, μ2=5.2559e06, μ3=5.0499e06 and μopt=5.2499e06 , respectively. As shown in Figure , if the convergence parameter μ is selected as different values, the convergence curve also has corresponding change. In order to more intuitively observe the performance of the RGI algorithm under different convergence parameters, we list the IT of the RGI algorithm in Table . It is evident that the convergence performance is the best when parameter μ is chosen to be μRGI=5.2499e06.

Figure 1. Comparison of convergence performance of RGI with different parameters μ for Example 5.1.

Figure 1. Comparison of convergence performance of RGI with different parameters μ for Example 5.1.

Table 1. Iterative number of the RGI algorithm with different μ for Example 5.1.

Moreover, the RGI algorithm with ω1=ω2=ω3=ω4=0.5 reduces to the GI algorithm. Similarly, we adopt the method of experiment debugging to find the optimal experimental parameter of the GI algorithm. Finally, the IT of the GI algorithm with μGI=4.5503e06 is the least.

In Figures , we present that the convergence curves of the RGI and GI algorithms with different δ. In this experiment, we compare the convergence curves of two algorithms under the optimal experimental parameters. It follows from Figures that the ERR decreases as the IT increases and gradually approaches 0, which indicates that the tested algorithms are effective and convergent. In addition, Figures clearly show that the IT of the RGI and GI algorithms are decreasing as the increasing of δ. Besides, we can find that the convergence speed of the RGI algorithm is always faster than the GI one in the above four situations.

In order to more specifically verify the advantages of the RGI algorithm, we list detailed numerical results of the RGI and GI algorithms in Table , which includes IT, CPU and ERR. According to Table , the IT and CPU of the tested algorithms are gradually increase with the decreasing of the parameter δ. Moreover, it can be seen that the IT and CPU of the RGI method are always less than the GI one. Therefore, we can conclude that the convergence performance of the RGI algorithm proposed in this paper is better than the GI algorithm [Citation23].

Figure 2. The convergence curves of the tested methods with δ=0.1 (left) and δ=0.01 (right) for Example 5.1.

Figure 2. The convergence curves of the tested methods with δ=0.1 (left) and δ=0.01 (right) for Example 5.1.

Figure 3. The convergence curves of the tested methods with δ=0.001 (left) and δ=0.0001 (right) for Example 5.1.

Figure 3. The convergence curves of the tested methods with δ=0.001 (left) and δ=0.0001 (right) for Example 5.1.

Table 2. Numerical results of the tested methods with different δ for Example 5.1.

Example 5.2

We consider the generalized coupled complex conjugate and transpose Sylvester matrix equation (Equation1) in the special case of p = q = 2, and it has (82) {A11Y1B11+C11Y1¯D11+E12Y2TF12+G12Y2HH12=M1,A21Y1B21+C21Y1¯D21+E22Y2TF22+G22Y2HH22=M2,(82) with the following parametric matrices A11=[3561+7i5+3i12+7i9+6i0],B11=[3+8i29i31+4i3+11i101+8i2+6i3+1i],A21=[53i1143i16i1+11i1+14i795i1+3i],B21=[135131123], C11=[14+2i3+5i41i66i81i5+1i2+3i8+2i19i],D11=[13+1i11+6i112i2+7i10+4i10+1i2+5i1218i19i],C21=[82i82i5i2+7i10+4i10+1i2+5i1218i19i],D21=[12+1i10+6i202+7i10+4i10+1i2+5i1218i19i], E12=[1+8i62i8+17i4+2i10+9i2+9i53i4+10i10+9i],F12=[11+2i28i23i7+7i91i1+5i7+10i5+8i78i],E22=[1i1i11111],F22=[122100013],G12=[111141202],H12=[122100011],G22=[111141202],H12=[1011i1121], M1=[4792+2166i2299+5490i4679+3574i3507629i27202137i4833+2642i33533607i19331090i2164+163i],M2=[562+1926i356+452i1858+1250i1113+610i511+863i6773228i1636+492i2678+3267i3463251i]. It has the exact solution Y1=[2+2i22i1112i22i2+2i1],Y2=[16+16i2028i12+4i294i7+5i4+17i1019i9+11i12i].

The initial iterative matrices are taken to be Yi(0)=106×I3,iI[1,2]. Then, we denote the relative iterative error by (83) ERR=Y1Y1(k)2+Y2Y2(k)2Y12+Y22.(83) In this example, all runs are stopped once ERR is less than ξ or k reaches the maximal iterative steps kmax=50,000. Here, ξ is a positive number.

For Example 5.2, we also compare the convergence performance of the RGI and GI algorithms. The optimal convergence parameters involved in two algorithms are determined by the following method. If relaxation factors are selected as ω1=0.07,ω2=0.18, the optimal convergence factor of the RGI algorithm is adopted as μRGI=1.0821e04 by Theorem 4.1. Moreover, the RGI algorithm with ω1=ω2=0.5 reduce to the GI algorithm. By some calculations, the best convergence parameter of the GI algorithm is μGI=4.5361e05.

In Figures , we plot the graphs of ERR(log10) versus the IT of the RGI and GI algorithms with different ξ. According to the convergence curves, we observe that the two algorithms are both convergent and efficient. It is obvious that the convergence rate of the RGI method (ω1=0.07,ω2=0.18) is always faster than GI one (ω1=ω2=0.5) for the four cases of ξ. In addition, it follows from Figures that the IT and CPU of the tested algorithms are increasing with the decreasing of ξ. In particular, the convergence advantage of the RGI algorithm is more obvious when ξ is smaller. The results illustrate that the RGI algorithm is superior to the GI algorithm if the relaxation parameters are chosen appropriately.

Figure 4. The convergence curves of the tested methods with ξ=0.1 (left) and ξ=0.01 (right) for Example 5.2.

Figure 4. The convergence curves of the tested methods with ξ=0.1 (left) and ξ=0.01 (right) for Example 5.2.

In order to further verify the advantages of the proposed algorithm, we clearly report the numerical results of the RGI and GI methods for Example 5.2 in Table . From Table , it is easy to discover that the IT of the algorithms is increasing with the decreasing of relative error. Furthermore, the IT and CPU of the RGI method are less than those of the GI one. As a whole, the proposed algorithm has better convergence behaviours than the GI method. This means that the relaxation technique can effectively improve the convergence rate of the GI algorithm.

Figure 5. The convergence curves of the tested methods with ξ=0.001 (left) and ξ=0.0001 (right) for of Example 5.2.

Figure 5. The convergence curves of the tested methods with ξ=0.001 (left) and ξ=0.0001 (right) for of Example 5.2.

Table 3. Numerical results of the tested methods with different ξ for Example 5.2.

6. Concluding remarks

In this paper, by adopting the relaxation technique into the GI algorithm, we establish the relaxed gradient-based iterative (RGI) algorithm to solve the generalized coupled complex conjugate and transpose Sylvester matrix equations. The main idea of the algorithm is introducing relaxation parameter to control the weights of iterative sequences. Applying straighten operation and real representation of complex matrices, we derive the necessary and sufficient condition for convergence of the RGI algorithm. Besides, the optimal convergence parameter and some related conclusions are given. To overcome high-dimensional computational problems, we propose sufficient condition for convergence with smaller computational complexity. Finally, numerical experiments verify that the RGI algorithm has more excellent convergence performance than the GI one.

Note that in our experiment, the relaxation factors ωl (lI[1,q]) are obtained through experimental debugging. The selection criteria for the optimal relaxation factors are not provided. The future research direction is to further develop the theory of selecting the optimal relaxation factor. Besides, the value of the convergence parameter μ in the RGI algorithm is fixed. To optimize the convergence performance of the RGI algorithm, we will consider to introduce different step size factors into the RGI algorithm.

Supplemental material

example1.pdf

Download ()

u111.eps

Download ()

u1u2u3.pdf

Download ()

example2.pdf

Download ()

u111.pdf

Download ()

example01.pdf

Download ()

u222.pdf

Download ()

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work is supported by the Guangxi Natural Science Foundations [No. 2021GXNSFBA196064, Guike AD21220129], the National Science Foundation of China [No. 11901123], the Guangxi Natural Science Foundations [2019GXNSFBA185014, Guike AD20159056], the Natural Science Foundation of Guangxi University for Nationalities [No. 2019KJQN001].

References

  • Hajarian M. Computing symmetric solutions of general Sylvester matrix equations via Lanczos version of biconjugate residual algorithm. Comput Math Appl. 2018;76:686–700. doi: 10.1016/j.camwa.2018.05.010
  • Hajarian M. Developing CGNE algorithm for the periodic discrete-time generalized coupled Sylvester matrix equations. Comput Appl Math. 2015;34:755–771. doi: 10.1007/s40314-014-0138-7
  • Zhou Y-H, Zhang X, Ding F. Partially-coupled nonlinear parameter optimization algorithm for a class of multivariate hybrid models. Appl Math Comput. 2022;414:Article ID 126663.
  • Dehghan M, Hajarian M. The generalised Sylvester matrix equations over the generalized bisymmetric and skew-symmetric matrices. Int J Syst Sci. 2012;43:1580–1590. doi: 10.1080/00207721.2010.549584
  • Zhou B, Wei X-Z, Duan G-R. Stability and stabilization of discrete-time periodic linear systems with actuator saturation. Automatica. 2011;47:1813–1820. doi: 10.1016/j.automatica.2011.04.015
  • Zhou B, Duan G-R. Periodic Lyapunov equation based approaches to the stabilization of continuous-time periodic linear systems. IEEE Trans Automat Contr. 2011;57:2139–2146. doi: 10.1109/TAC.2011.2181796
  • Ding F, Wang F. Decomposition based least squares iterative identification algorithm for multivariate pseudo-linear ARMA systems using the data filtering. J Franklin Inst. 2017;354:1321–1339. doi: 10.1016/j.jfranklin.2016.11.030
  • Shen H-L, Peng C, Zhang T. Gradient based iterative solutions for Sylvester conjugate matrix equations. J Math Res Appl. 2017;03:103–118.
  • Li X, Ding J, Ding F. Gradient based iterative solutions for general linear matrix equations. Comput Math Appl. 2009;58:1441–1448. doi: 10.1016/j.camwa.2009.06.047
  • Li X, Liu Y-J, Yang H-Z. Gradient based and least squares based iterative algorithms for matrix equations AXB+CXTD=F. Appl Math Comput. 2010;217:2191–2199.
  • Bai Z-Z, Guo X-X, Yin J-F. On two iteration methods for the quadratic matrix equations. Int J Numer Anal Model. 2005;2:114–122.
  • Chen Z-B, Chen X-S. Modification on the convergence results of the Sylvester matrix equation AX + XB = C. J Franklin Inst. 2022;359:3126–3147. doi: 10.1016/j.jfranklin.2022.02.021
  • Xu L, Ding F, Zhu Q-M. Separable synchronous multi-innovation gradient-based iterative signal modeling from on-line measurements. IEEE Trans Instrum Meas. 2022;71:1–13.
  • Ding F. Least squares parameter estimation and multi-innovation least squares methods for linear fitting problems from noisy data. J Comput Appl Math. 2023;426:Article ID 115107. doi: 10.1016/j.cam.2023.115107
  • Ding J, Liu Y-J, Ding F. Iterative solutions to matrix equations of the form AiXBi=Fi. Comput Math Appl. 2010;59:3500–3507. doi: 10.1016/j.camwa.2010.03.041
  • Ding F, Ding J. Iterative solutions of the generalized Sylvester matrix equations by using the hierarchical identification principle. Appl Math Comput. 2008;197:41–50.
  • Ding F, Chen T-W. Gradient based iterative algorithms for solving a class of matrix equations. IEEE Trans Automat Contr. 2005;50:1216–1221. doi: 10.1109/TAC.2005.852558
  • Ding F, Chen T-W. On iterative solutions of general coupled matrix equations. SIAM J Control Optim. 2006;44:2269–2284. doi: 10.1137/S0363012904441350
  • Ding F, Chen T-W. Iterative least-squares solutions of coupled Sylvester matrix equations. Syst Control Lett. 2005;54:95–107. doi: 10.1016/j.sysconle.2004.06.008
  • Wu A-G, Zeng X-L, Duan G-R, et al. Iterative solutions to the extended Sylvester-conjugate matrix equations. Appl Math Comput. 2010;217:130–142.
  • Wu A-G, Feng G, Duan G-R, et al. Iterative solutions to coupled Sylvester-conjugate matrix equations. Comput Math Appl. 2010;60:54–66. doi: 10.1016/j.camwa.2010.04.029
  • Song C-Q, Chen G-L, Zhao L-L. Iterative solutions to coupled Sylvester-transpose matrix equations. Appl Math Model. 2011;35:4675–4683. doi: 10.1016/j.apm.2011.03.038
  • Beik FPA, Mahmoud MM. Gradient-based iterative algorithm for solving the generalized coupled Sylvester-transpose and conjugate matrix equations over reflexive (anti-reflexive) matrices. Trans Inst Meas Control. 2014;36:99–110. doi: 10.1177/0142331213482485
  • Lv L-L, Chen J-B, Zhang L, et al. Gradient-based neural networks for solving periodic Sylvester matrix equations. J Franklin Inst. 2022;359:10849–10866. doi: 10.1016/j.jfranklin.2022.05.023
  • Li S-H, Ma C-F. Factor gradient iterative algorithm for solving a class of discrete periodic Sylvester matrix equations. J Franklin Inst. 2022;359:9952–9970. doi: 10.1016/j.jfranklin.2022.09.041
  • Fan W, Gu C-Q, Tian Z-L. Jacobi-gradient iterative algorithms for Sylvester matrix equations. In: Linear algebra society topics. Shanghai, China: Shanghai University; 2007. p. 16–20.
  • Niu Q, Wang X, Lu L-Z. A relaxed gradient based algorithm for solving Sylvester equations. Asian J Control. 2011;13:461–464. doi: 10.1002/asjc.v13.3
  • Huang B-H, Ma C-F. The relaxed gradient-based iterative algorithms for a class of generalized coupled Sylvester-conjugate matrix equations. J Franklin Inst. 2018;355:3168–3195. doi: 10.1016/j.jfranklin.2018.02.014
  • Huang B-H, Ma C-F. On the relaxed gradient-based iterative methods for the generalized coupled Sylvester-transpose matrix equations. J Franklin Inst. 2022;359:10688–10725. doi: 10.1016/j.jfranklin.2022.07.051
  • Wang W-L, Song C-Q, Ji S-P. Iterative solution to a class of complex matrix equations and its application in time-varying linear system. J Appl Math Comput. 2021;67:317–341. doi: 10.1007/s12190-020-01486-6
  • Ding F. Hierarchical multi-innovation stochastic gradient algorithm for Hammerstein nonlinear system modeling. Appl Math Model. 2013;37:1694–1704. doi: 10.1016/j.apm.2012.04.039
  • Wu A-G, Zhang Y, Qian Y-Y. Complex conjugate matrix equations. Beijing: Science Press; 2017.
  • Ding F, Chen T-W. Hierarchical gradient-based identification of multivariable discrete-time systems. Automatica. 2005;41:315–325. doi: 10.1016/j.automatica.2004.10.010
  • Zhang H-M. A finite iterative algorithm for solving the complex generalized coupled Sylvester matrix equations by using the linear operators. J Franklin Inst. 2017;354:1856–1874. doi: 10.1016/j.jfranklin.2016.12.011