Publication Cover
Mathematical and Computer Modelling of Dynamical Systems
Methods, Tools and Applications in Engineering and Related Sciences
Volume 22, 2016 - Issue 4: Model Order Reduction
359
Views
2
CrossRef citations to date
0
Altmetric
Articles

STABLE – a stability algorithm for parametric model reduction by matrix interpolation

&
Pages 307-322 | Received 29 Sep 2015, Accepted 02 Jun 2016, Published online: 25 Jun 2016

ABSTRACT

In this article, an algorithm guaranteeing asymptotic stability for parametric model order reduction by matrix interpolation is proposed for the general class of high-dimensional linear time-invariant systems. In the first step, the system matrices of the high-dimensional parameter-dependent system are computed for a set of parameter vectors. The local high-order systems are reduced by a projection-based reduction method and stabilized, if necessary. Secondly, the low-order systems are transformed into a consistent set of generalized coordinates. Thirdly, a new procedure using semidefinite programming is applied to the low-order systems, converting them into strictly dissipative form. Finally, an asymptotically stable reduced order model can be calculated for any new parameter vector of interest by interpolating the system matrices of the local low-order models. We show that this approach works without any limiting conditions concerning the structure of the large-scale model and is suitable for real-time applications. The method is illustrated by two numerical examples.

1. Introduction

Accurate modelling of dynamical systems can lead to large-scale systems of ordinary differential equations. If these systems are used for the purpose of optimization, simulation or control, even powerful computer systems reach their limitations, especially concerning execution time and available memory. As a remedy, methods of model order reduction (MOR) have been developed which approximate the input–output behaviour of the original system by a low-order one [Citation1]. In addition, properties of the high-order system like stability, passivity or the Port-Hamiltonian structure should be preserved, see, e.g., [Citation2,Citation3]. For many engineering applications, the high-order system additionally depends on parameters, such as material parameters. For this case, methods of parametric MOR (pMOR) have been developed [Citation4]. They construct a low-order system that approximates the large-scale one for the entire parameter domain. Hence, one can obtain a reduced system for every parameter value of the domain without the need to repeat the reduction procedure.

Besides other pMOR methods, interpolation-based methods were suggested in the literature. One approach is the interpolation of reduced system matrices [Citation5Citation10], which can handle arbitrary parametric dependencies and has a low computational effort. However, it does not necessarily lead to asymptotically stable interpolated systems. Hence, some stability-preserving approaches were proposed for matrix interpolation, most of them being restricted to systems showing certain conditions concerning the structure of the large-scale model. In [Citation5], stable interpolation on matrix manifolds is proposed for systems of ordinary differential equations in second-order form. Stability-preserving matrix interpolation for dissipative systems is suggested in [Citation11], for Port-Hamiltonian systems in [Citation12] and for passive systems in [Citation13,Citation14]. Concerning the general class of linear time-invariant (LTI) systems, a technique is suggested in [Citation15,Citation16] for passivity-preserving parameterized MOR which exploits the solution of high-order linear matrix inequalities (LMIs). Some stability-preserving methods are proposed in [Citation11,Citation17] for the general class of LTI systems. These approaches have in common that they modify one of the two reduced order bases (ROBs) of the local systems to enforce stability for the interpolated systems. The approach from [Citation11] requires the solution of high-dimensional generalized Lyapunov equations, which is computationally very expensive. The method from [Citation17] determines an optimal stabilizing solution with regard to a quality function that is found by solving high-order optimization problems. For reducing the computational effort, the authors also propose low-order but suboptimal optimization problems.

In the following, a procedure for stability-preserving pMOR for the general class of parameter-dependent, LTI systems is presented, where all steps of the (stability-preserving) matrix interpolation are exclusively performed in low-dimensional vectors and matrices. Preliminary studies for the proposed method were published in [Citation18], and a detailed description can also be found in the dissertation [Citation7]. The method starts with optimization steps similar to [Citation17]. However, it searches for the modified ROBs in the available low-order subspaces, and hence, the optimization steps can be done very efficiently.

A short overview of MOR and pMOR by matrix interpolation is given in Section 2. The new stability-preserving method is presented in Section 3 and is compared to an existing approach in Section 4. Numerical results are presented in Section 5.

2. Preliminaries

2.1. Parameter-dependent LTI systems

In this article,Footnote1 we consider a parameter-dependent, LTI and high-dimensional system of order n. Its representation in the time domain is

(1) G(t,p):{E(p)x˙(t,p)=A(p)x(t,p)+B(p)u(t)y(t,p)=C(p)x(t,p),(1)

where E(p)Rn×n, A(p)Rn×n, B(p)Rn×r and C(p)Rm×n are the system matrices which depend on the parameter vector pD with domain DRd. The vectors u(t)Rr, y(t,p)Rm and x(t,p)Rn denote the inputs, outputs and states of the system at time t, respectively. The initial value of the state vector is x(0,p)=x0(p). Performing a Laplace transformation on system (Equation (1)) and setting x0(p)=0n, we obtain the representation of the system in the frequency domain.

(2) G(s,p)=C(p)[sE(p)A(p)]1B(p),(2)

where sμ is the complex frequency. When we refer to the system independent from its representation, it is denoted by G(p) in the following.

2.2. Asymptotic stability

Stability is an important property of system (Equation (1)). There exist various stability definitions [Citation1,Citation19,Citation20]. In this article, the following concept of stability is examined.

Definition 2.1: The system (Equation (1)) is said to be asymptotically stable if one finds limtx(t,p)=0n for all initial conditions x0(p).

The following theorem gives necessary and sufficient conditions.

Theorem 2.2 ([Citation21]): The system (Equation (1)) with non-singular E(p) is asymptotically stable if and only if

• S1: there exists a Lyapunov function, i.e. a function V(x(t,p)):RnR with V(x(t,p)) > 0 and V˙(x(t,p)) < 0 for all x(t,p)  0n.

• S2: the eigenvalues of the pencil (A(p),E(p)) lie in the open left half of the complex plane.

A special class of systems is of particular interest in this article.

Definition 2.3 ([Citation22]): The system (Equation (1)) is said to be strictly dissipative if it satisfies E(p) > 0n×n and A(p)+A(p)T < 0n×n.

A strictly dissipative system possesses the Lyapunov function V(x(t,p))=x(t,p)TE(p)x(t,p) and hence is asymptotically stable [Citation23]. Strictly dissipative systems have the following interesting property which will be used in Section 3.3 for proposing a stability-preserving interpolation method.

Corollary 2.4 ([Citation11,Citation17]): Given the strictly dissipative systems G(p1),,G(pN) with E(pi) > 0n×n and A(pi)+A(pi)T < 0n×n for i{1,,N}, the system obtained by the superposition of the system matrices A(p)=i=1Nωi(p)A(pi) and E(p)=i=1Nωi(p)E(pi) with weighting functions ωi:DR0+ and at least one ωi:DR+ possesses the Lyapunov function V(x(t,p))=x(t,p)TE(p)x(t,p) and hence is asymptotically stable according to criterion S1 from Theorem 2.2.

2.3. Projection-based MOR

Imagine that we compute the parametric system G(p) for a set of grid points piD with i{1,,N}. We denote E(pi) by Ei and similarly for the other matrices and the resulting non-parametric system is denoted by Gi: =G(pi) with xi(t): =x(t,pi) and yi(t): =y(t,pi). Then, the goal of MOR is to find a low-dimensional system Gˆi of order qn whose output is a good approximation of the output of the original system Gi, i.e., yˆi(t)yi(t) when the same input is fed to the systems. For this, projection matrices Vi: =V(pi)Rn×q and Wi: =W(pi)Rn×q, which are referred to as ROBs, are calculated. The case Wi=Vi is called one-sided reduction and the case Wi  Vi is referred to as two-sided reduction. The ROBs span the right Vi=span(Vi) and left subspace Wi=span(Wi), respectively. For the calculation of the ROBs, we can apply projection-based reduction methods such as Krylov subspace methods, truncated balanced realization (TBR) or proper orthogonal decomposition, see, e.g., [Citation1] and references therein. Using the approximation xi(t)Vixˆi(t) with the reduced state vector xˆi(t)Rq and enforcing the Petrov–Galerkin condition

(3) WiT(EiVix^˙i(t)AiVix^i(t)Biu(t))=0,(3)

leads to the reduced order model

(4) G^i:{E^ix^˙i(t)=A^ix^i(t)+B^iu(t)y^i(t)=C^ix^i(t),(4)

where the reduced system matrices are given by

(5) Eˆi=WiTEiVi,Aˆi=WiTAiVi,Bˆi=WiTBi,Cˆi=CiVi.(5)

2.4. Parametric MOR by matrix interpolation

Methods of pMOR have been developed to approximate the parameter-dependent, high-order system G(p) while preserving the parameter-dependency in the reduced-order model. The method of interest in this article interpolates precomputed reduced system matrices. It was proposed in [Citation5,Citation6,Citation9,Citation10] and a unifying framework was suggested in [Citation7,Citation8]. We sum up four preprocessing steps (offline phase) and the interpolation step (online phase).

  1. The domain μ is sampled for a set of vectors P={p1,,pN}D with N=|P|. Then, the system G(p) is computed for the vectors in the set P which results in a set of high-order systems G={G(p1),,G(pN)}.

  2. Each system in the set μ is reduced individually to order qn, where any projection-based reduction method can be applied according to Section 2.3 resulting in a set of low-order systems Gˆ={Gˆ(p1),,Gˆ(pN)}.

  3. The reduced systems of the set Gˆ have different bases V1,,VN. These right ROBs need to be adjusted to the basis of the reference subspace V0. An approach that captures the most important directions of all subspaces and has a low computational effort uses the economy version of the SVD

    (6) UVSVZVT=svd([V1,,VN],'econ')V0=UV(:,1:q).(6)

    The right ROBs are transformed for i{1,,N} with matrices TiRq×q to V˜i=ViTi so that a quality function is minimized. A possible quality function is the modal assurance criterion (MAC) between the local basis V˜i and the reference basis V0 which can be written according to [Citation7,Citation24]

    (7) JMAC,V,i(Ti)=∥V0TV˜iIqF=∥V0TViTiIqF.(7)

    The solution with the minimum norm value, JMAC,V,i=0, is Ti=(V0TVi)1. A second possible quality function is the distance (DS) between the local basis V˜i and the reference basis V0 measured in the Frobenius norm [Citation5,Citation7,Citation24]

    (8) JDS,V,i(Ti)=∥V˜iV0F=∥ViTiV0F.(8)

    The solution with the minimum value of the cost function is Ti=Vi+V0 using the pseudoinverse Vi+=(ViTVi)1ViT.

  4. In analogy, the bases W1,,WN need to be adjusted to the basis of the reference subspace W0, which can, e.g., be calculated according to Equation (6) by replacing the respective values. The left ROBs are transformed for i{1,,N} with matrices MiRq×q to W˜i=WiMi so that a quality function is minimized. For the MAC approach, we obtain the quality function

    (9) JMAC,W,i(Mi)=∥W0TW˜iIqF=∥W0TWiMiIqF(9)

with the optimal solution Mi=(W0TWi)1. For the DS approach, we obtain

(10) JDS,W,i(Mi)=∥W˜iW0F=∥WiMiW0F(10)

with the optimal solution Mi=Wi+W0.

The steps discussed so far are performed in the offline phase and have to be done only once. They result in a set of reduced-order systems G˜={G˜1,,G˜N} with

(11) G˜i:{MiTE^iTiE˜ix˜˙i(t)=MiTA^iTiA˜ix˜˙i(t)+MiTB^iB˜iu(t)y˜i(t)=C^iTiCix˜i(t).(11)

Note that these systems differ from systems (Equation (4)) only in their state-space representation, but not in their eigenvalues or input–output behaviour.

In the online phase, i.e. when a reduced order model for the new parameter vector pD is supposed to be calculated, this is simply done by the interpolation of the precomputed matrices with weighting functions ω1,,ωN:DR. Then, the new reduced system for the parameter vector p is obtained by

(12) G˜(p):{E˜(p)x˜˙(t,p)=A˜(p)x˜(t,p)+B˜(p)u(t)y˜(t,p)=C˜(p)x˜(t,p)(12)

with

(13) E˜(p)=i=1Nωi(p)E˜i,A˜(p)=i=1Nωi(p)A˜i,B˜(p)=i=1Nωi(p)B˜i,C˜(p)=i=1Nωi(p)C˜i.(13)

The system matrices can also be interpolated element-wise and on matrix manifolds.

2.5. Problem formulation

The matrix interpolation (Equation (13)) does not always lead to an asymptotically stable system (Equation (12)), even if all reduced systems in the set G˜ are asymptotically stable. In other words, the reduction (Equations (12) and (13)) does not guarantee asymptotic stability of the reduced model.

Example: Consider the two asymptotically stable systems x^˙1(t)=A^1x^1(t) and x^˙2(t)=A^2x^2(t) with matrices

(14) Aˆ1=2502,Aˆ2=2052.(14)

For example, the interpolated matrix

(15) Aˆint=0.5Aˆ1+0.5Aˆ2=22.52.52(15)

has the spectrum {–4.5, 0.5} and the resulting system x^˙(t)=A^intx^(t) is therefore unstable according to criterion S2 of Theorem 2.2.

An idea to solve this problem is to minimally change the left ROBs W˜1,,W˜N such that the interpolated reduced system is guaranteed to be asymptotically stable.

3. Stability-preserving interpolation of reduced system matrices

3.1. Calculation of a set of asymptotically stable, reduced systems

The proposed method starts with a set of high-order systems G={G1,,GN}. These systems are reduced individually to the order qn according to Equation (4), and we assume that the reduced systems Gˆ={Gˆ1,,GˆN} are asymptotically stable. In order to ensure this stability, two general approaches are possible: (i) We can use stability-preserving reduction methods such as TBR or extended versions of Krylov subspace methods. (For the latter, the stability-preserving, adaptive rational Krylov (SPARK) from [Citation25] guarantees stability and the iterative rational Krylov algorithm (IRKA) from [Citation26] delivers asymptotically stable systems as long as it converges.) (ii) For MOR methods that do not implicitly guarantee stability, one can easily check if the reduced systems are asymptotically stable by examining their eigenvalues according to criterion S2 from Theorem 2.2; if a system is then found to be unstable, it can be stabilized via post-processing, e.g. by modifying the left or right ROBs [Citation27,Citation28].

3.2. Adjustment of the right ROBs

The reduced systems of the set Gˆ have different bases V1,,VN and it was shown in Section 2.4 that they need to be adjusted to the basis of the reference subspace for accuracy reasons. For this, the right ROBs are transformed for i{1,,N} to V˜i=ViTi with Ti=(V0TVi)1 using the MAC approach or with Ti=Vi+V0 using the DS approach as described in Section 2.4.

3.3. Stabilizing procedure using semidefinite programming

In this section, a stabilization method is described for the interpolation of reduced matrices. The demands that we impose on the method are – besides stability preservation – accuracy and the operation on low-order systems. The main idea is the following: starting from the transformed bases W˜i=WiMi for i{1,,N} with Mi=(W0TWi)1 using the MAC approach or with Mi=Wi+W0 using the DS approach (cf. Section 2.4), we minimally modify the new left ROBs W˜i until stability is guaranteed (again note that the eigenvalues and the input–output behaviour of the local models remain unchanged).

3.3.1. Feasibility problem

To begin with, an abstract optimization problem is formulated in which a convex objective function JW is minimized subject to the constraint that the interpolated system G˜(p) is asymptotically stable for the domain D. According to criterion S1 from Theorem 2.2, this constraint is equivalent to the existence of a Lyapunov function V(x˜(t,p)). Hence, we obtain the optimization problem

(16) minJWs.t.LyapunovfunctionV(x˜(t,p))pD.(16)

According to Corollary 2.4, the existence of a Lyapunov function is guaranteed for the superposition of strictly dissipative systems when non-negative weighting functions are applied. Hence, the optimization problem (Equation (16)) is rewritten as N problems, one for each grid point [Citation17]. We introduce for each i{1,,N} the basis W˜i as optimization variable and we aim to calculate this variable so that the system G˜i is strictly dissipative and the objective function JW,i reaches its minimum value:

(17) minW˜iRn×qJW,iW˜is.t.W˜iTAiV˜i+W˜iTAiV˜iT < 0q×qW˜iTEiV˜i > 0q×q(17)

Unfortunately, the matrices W˜i are of high-order, and hence, the optimization problems are intractable. As a remedy, the new bases W˜i are searched in the available subspaces Wi=span{Wi}. For this, the matrices MiRq×q are introduced as new variables, which provide the new bases W˜i=WiMi and the optimization problems

(18) minMiRq×qJW,i(Mi)(18)
(19) s.t.MiTAˆiTi+(MiTAˆiTi)T < 0q×q(19)
(20) MiTEˆiTi > 0q×q(20)

which only contain low-order matrices. For solving these problems efficiently using interior point methods such as described, e.g., in [Citation29], they are transformed into semidefinite programs. These kinds of optimization problems deal with symmetric positive or negative (semi)definite matrices as variables. To get to this form, the change of variables

(21) Mi=PiEˆiTi(21)

is introduced, where the new variables PiSq are symmetric matrices [Citation21]. Then, Equation (21) is inserted into inequality (Equation (20)), which gives the necessary constraint on the definiteness of the optimization variables

(22) MiTEˆiTi=TiTEˆiTPiEˆiTi=(EˆiTi)TPi(EˆiTi) > 0q×qPi > 0q×q.(22)

For Equation (19) it follows

(23) MiTAˆiTi+(MiTAˆiTi)T=TiTEˆiTPiAˆiTi+TiTAˆiTPiEˆiTi=TiT(EˆiTPiAˆi+AˆiTPiEˆi)Ti < 0q×qEˆiTPiAˆi+AˆiTPiEˆi < 0q×q.(23)

Finally, the new optimization problems are

(24) minPiRqJW,iPis.t. EˆiTPiAˆi+AˆiTPiEˆi< 0q×qPi> 0q×q(24)

For assessing the solvability of problems (Equation (24)), the theory of convex optimization tells us that only the constraints have to be examined and the quality function can be set to 0 which is referred to as feasibility problem [Citation30]. In this case, the constraints are LMIs which are equivalent to the generalized Lyapunov inequality [Citation31]. This inequality and accordingly the optimization problem (Equation (24)) admit a solution because the eigenvalues of the pencil (Aˆi,Eˆi) lie in the open left half of the complex plane as a result of Section 3.1.

3.3.2. Objective functions

In Section 3.3.1, the feasibility problem was introduced which guarantees stable matrix interpolation for all solutions satisfying the constraints. In this section, amongst these solutions, the optimal stabilizing one is determined with regard to an objective function. The objective function, which needs to be minimized, is chosen in such a way that we obtain an accurate matrix interpolation. For this, we can use the objective functions described in Equation (4) which adjust (for each i{1,,N}) the basis Wi to the basis of the reference subspace W0 by transforming it to W˜i=WiMi with MiRq×q. For the MAC approach, we insert the new choice for Mi from Equation (21) into the objective function (Equation (9)) and we obtain

(25) JMAC,W,i(Pi)=∥W0TWiPiEˆiTiIqF.(25)

The matrices Xi=W0TWiRq×q and Yi=EˆiTiRq×q can be precomputed, and hence, the objective function that only comprises low-order matrices is obtained:

(26) JMAC,W,i(Pi)=∥XiPiYiIqF.(26)

This objective function is convex as the composition of the Frobenius norm, which is a convex function, with the affine map PiXiPiYiIq convex [Citation30]. For the DS approach, we insert the new choice for Mi from Equation (21) into the objective function (Equation (10)):

(27) JDS,W,i(Pi)=∥WiPiEˆiTiW0F.(27)

Unfortunately, this objective function is of large order since the matrices Wi,W0 have the size n×q. As a remedy, we take the square of the Frobenius norm, rewrite it using the trace and obtain

(28) JDS,W,i(Pi)2= WiPiEˆiTiW0F2= tr(TiTEˆiTPiWiTWiPiEˆiTi)2tr(W0TWiPiEˆiTi)+tr(W0TW0)= (WiTWi)1/2PiEˆiTiF22tr(W0TWiPiEˆiTi)+tr(W0TW0).(28)

The low-order matrices Xi, Yi and Zi=(WiTWi)1/2Rq×q can again be precomputed. When the third, constant term is left out, the final low-order objective function is obtained:

(29) JDS,W,i(Pi)=∥ZiPiYiF22tr(XiPiYi).(29)

The first term is convex because the composition of the square which is a convex, non-decreasing function for non-negative arguments and the convex Frobenius norm is convex. The entire quality function is convex as the sum of the first convex function and the trace, which is a linear function, is also convex [Citation30].

3.3.3. Algorithm

The proposed method modifies the left ROBs for an accurate interpolation while making the local reduced models’ representations strictly dissipative and thereby guaranteeing stability of any interpolated reduced model. For this, the objective function (Equation (26) or (29)) is minimized for every grid point subject to the constraint (Equations (22) and (23)). As the objective functions are convex and the constraints are linear, a convex optimization problem is obtained which is referred to as Stability Algorithm Based on Linear Matrix Inequalities (STABLE) and which is given in Algorithm 1.

Algorithm 1 STABLE

input: Matrices Eˆi,Aˆi,Ti, Wi for i{1,,N}, reference basis W0

output: Matrices Mi for i{1,,N}

1: for i=1 to N do

2: Compute Xi=W0TWi, Yi=EˆiTi, Zi=(WiTWi)1/2

3: Solve the convex optimization problem (MAC or DS approach):

minPiSqXiPiYiIqF orminPiSq(ZiPiYiF22tr(XiPiYi))
s.t.EˆiTPiAˆi+AˆiTPiEˆi < 0q×qPi > 0q×q

4: Compute Mi=PiEˆiTi

5: end for

As STABLE is a low-order optimization problem, it can easily be implemented and solved, e.g. in MATLAB, using convex optimization solver packages that rely on semidefinite programming such as CVX [Citation32,Citation33]. These packages are based on algorithms which solve an optimization problem iteratively. According to, e.g., [Citation29,Citation31,Citation34Citation36], the complexity is at most O(q6) per iteration and in practice the solvers perform much better. The required number of iterations can be considered in practice almost independent of the problem size, ranging between 5 and 50, as long as the numerics is kept well-conditioned. Hence, the solution of the semidefinite programs is currently restricted to sizes for the reduced order q up to some hundreds.

Then, STABLE delivers for each i{1,,N} the transformation matrix Mi which results in the strictly dissipative reduced system G˜i with E˜i > 0q×q and A˜i+A˜iT < 0q×q. This ensures asymptotic stability of any interpolated reduced model, according to Corollary 2.4. The price paid is a slight increase of the minimum value of the objective function since the left ROBs are distorted as little as possible compared to the method from Section 2.4. For the case that the feasibility problem (Equation (24)) only allows a very limited solution set, one might obtain a more distinct distortion of the left ROBs, which can lead to noticeable interpolation errors.

Remark 3.1 In a dual way, we could adjust the left ROBs for accuracy reasons such as described in Section 2.4 and solve an optimization problem in order to modify the right ROBs for an accurate interpolation while preserving the stability, see dissertation [Citation7] for detailed information.

3.4. Online interpolation process

In the online phase, the reduced order model G˜(p) for a new parameter pD is computed by interpolation of the precomputed matrices as in Equation (13). Following Corollary 2.4, the interpolation process needs to take place with non-negative weighting functions ω1,,ωN:DR0+ and a positive weighting function ωi:DR+ for at least one i{1,,N}. One example for such weighting functions are linear basis functions. Then, the interpolated system G˜(p) possesses the Lyapunov function V(x˜(t,p))=x˜(t,p)T(i=1Nωi(p)E˜i)x˜(t,p) and hence is asymptotically stable for the entire domain D.

4. Comparison to an approach in the literature

A method is proposed in [Citation11] which guarantees stable interpolation of reduced system matrices for the special case of strictly dissipative high-order systems. It applies for each i{1,,N} the one-sided reduction Wi=Vi, adjusts the right ROB with Ti=(V0TVi)1 and chooses for the left ROB the matrices Mi=Ti.

Proposition 4.1: STABLE comprises the method from [Citation11] as a special case for strictly dissipative systems and extends it to the general class of LTI systems.

Proof. For high-order system, Gi with Ei > 0n×n and Ai+AiT < 0n×n perform a one-sided reduction method with Wi=Vi, which leads to a reduced system fulfilling Eˆi > 0q×q and Aˆi+AˆiT < 0q×q [Citation23]. For the adjustment of the right ROBs, just like in [Citation11], take the matrix Ti=(V0TVi)1 which corresponds to the MAC approach (Equation (27)). For adjusting the left ROBs, we consider the objective function (Equation (25))

JMAC,W,i(Pi)=∥V0TViPiEˆiTiIqF=∥Ti1PiEˆiTiIqF

which reaches its minimum value, JMAC,W,i=0, for Pi=Eˆ1. The same holds true for the objective function of the DS approach (Equation (27)). The optimal solution belongs to the feasible set because it fulfils the first constraint (Equation (22)) owing to

Pi=Eˆi1 > 0q×q

and the second constraint (Equation (23)) because of

EˆiTPiAˆi+AˆiTPiEˆi=EˆiTEˆi1Aˆi+AˆiTEˆi1Eˆi=Aˆi+AˆiT < 0q×q.

Hence, STABLE delivers the solution Mi=PiEˆiTi=Eˆi1EˆiTi=Ti which is the result from [Citation11].

5. Numerical examples

5.1. Motivating example

Consider a high-order system E(p)x˙(t,p)=A(p)x(t,p) with x(t,p)Rn. The domain pD with D=[0,1] is sampled for two values p1=0,p2=1. The resulting systems Eix˙i(t)=Aixi(t) with i{1,2} are reduced to order q=2n using projection matrices Vi=V0Rn×2 and Wi=W0Rn×2, where the columns are an orthonormal basis. This is supposed to result according to Equation (4) in the asymptotically stable systems E^ix^˙i(t)=A^ix^i(t) with Eˆi=I2 and Aˆi from Equation (14).

5.1.1. Matrix interpolation without using STABLE

Firstly, the interpolation procedure will be demonstrated without using STABLE. The respective values are labelled with without STABLE (WS). The transformation matrices are found to be TiWS=I2,MiWS=I2 according to Equations (7) or (8) and (9) or (10) with JV,iWS=0,JW,iWS=0. The adjusted ROBs are V˜i=V0,W˜i=W0 and the transformed systems are x˜˙i(t)=A^ix˜i(t) according to Equation (11). The interpolated system is x˜˙(t)=A˜WS(p)x˜(t) with A˜WS(p)=i=12ωi(p)Aˆi, where ωi(p)=1|ppi| are linear weighting functions. The maximum value of the real part of the eigenvalues λmax of A˜WS(p) is positive for p[0.2,0.8] and hence unstable systems are obtained.

5.1.2. Matrix interpolation using STABLE

Secondly, the proposed stability-preserving method is demonstrated. For adjusting the right ROBs, we obtain the transformation matrices Ti=I2 according to objective functions (Equation (7) or (8)), which is equivalent to the approach without using STABLE. For adjusting the left ROBs, we use STABLE which is demonstrated graphically in the following. To begin with, we introduce the variables of the algorithm.

(30) Pi=p11,ip12,ip12,ip22,iR2.(30)

Then, the first constraint (Equation (22)) is described for the two systems by the convex cones Pi={Pi:Pi > 02×2}. The intersection of the convex cone of positive definite matrices with the linear space from the second constraint (Equation (23)) provides for the two systems the convex cones Ci={Pi:Pi > 02×2,PiAˆi+AˆiTPi < 02×2}. The convex cones C1,C2, which are non-empty since the systems E^ix^˙i(t)=A^ix^i(t) are asymptotically stable [Citation31], comprise all stability-preserving solutions Pi.

The convex cones P1,P2,C1,C2 are illustrated in and the axes show the entries of the matrices Pi. Consider the objective function (Equations (26) and (28))

(31) JW,i=∥PiI2F2=tr(PiPi)2tr(Pi)+tr(I2)=(p11,i1)2+(p22,i1)2+2p12,i2.(31)

Figure 1. Convex cones C1 (left), C2 (right) and P1,P2.

Figure 1. Convex cones C1 (left), C2 (right) and P1,P2.

Then, STABLE minimizes the objective function JW,i so that the optimal solution Pi is an element of the respective convex cone Ci:

(32) P1=0.745001.164,P2=1.164000.745.(32)

For solving the convex program STABLE, the package CVX is used [Citation32,Citation33]. The results are obtained using solver SeDuMi [Citation37]. Operating on a 2.20-GHz processor, solving the semidefinite programs lasted tˉSTABLE=0.5s on average for each of the two grid points. A graphical interpretation of the optimization problem is given in . The axes show the entries p11,i,p22,i of the matrices Pi for p12,i=0. The objective functions JW,i from Equation (31) describe circles in the plane p12,i=0 with midpoint (p11,i,p22,i)=(1,1) and radius JW,i. The convex cones C1,C2 are shown as triangles which are the intersection of the convex cones from with the plane p12,i=0. The algorithm STABLE searches for the solutions (p11,i,p22,i) with the minimum value of the objective functions JW,i so that the solutions lie in the convex cones C1,C2. This is the case for the solutions (Equation (32)) which are the points where the circle with radius JW,i=0.303 touches the cones. This results in the transformation matrices Mi=Pi which lead to the reduced systems Pi*x˜˙i(t)=Pi*A^ix˜i(t). Then, the interpolated reduced system is E˜(p)x˜˙(t,p)=A˜(p)x˜(t,p) with E˜(p)=i=12ωi(p)Pi and A˜(p)=i=12ωi(p)PiAˆi with linear weighting functions ωi(p). The interpolated system possesses the Lyapunov function V(x˜(t,p))=x˜(t,p)T(i=12ωi(p)Pi)x˜(t,p) and hence is asymptotically stable for the entire domain D. The price paid is an increase of the minimum value of the objective functions JW,i(Pi) since the left ROBs are distorted as little as possible to W˜i=W0Pi (see ).

Table 1. Comparison of the adjusted ROBs and of the respective minimum values of the objective functions for matrix interpolation with and without STABLE.

Figure 2. Convex cones C1 (top left), C2 (bottom right) in the plane p12,i=0 and optimal (p11,i,p22,i) with minimum cost function JW,i=0.3032.

Figure 2. Convex cones C1 (top left), C2 (bottom right) in the plane p12,i=0 and optimal (p11,i∗,p22,i∗) with minimum cost function JW,i=0.3032.

5.2. Anemometer

The anemometer is a flow sensing device [Citation38,Citation39]. It consists of a heater and two temperature sensors that are located before and after the heater. If there is no flow, the heat dissipates symmetrically into the fluid. This symmetry is disturbed if a flow is applied to the fluid, which leads to convection of the temperature field. Then, the fluid velocity can be determined from the difference between the temperatures measured with the sensors. The governing equation is the convection-diffusion partial differential equation

(33) ρcpTt=(κT)ρcpvT+Q˙,(33)

where ρ denotes the mass density, cp=0.5Jkg1K1 is the specific heat, T is the temperature and Q˙ is the heat flow into the system caused by the heater. The thermal conductivity varies in κ[1,2]Wm1K1 and the fluid velocity in v[0.1,2]ms1. Finite element discretization of the convection-diffusion Equation (33), with zero Dirichlet boundary conditions, leads to the parameter-dependent state-space system of order n=29008

(34) ET˙(t,κ,v)=(A0+κA1+vA2)A(κ,v)T(t,κ,v)+bQ˙(t)y(t,κ,v)=cTT(t,κ,v),(34)

where T(t,κ,v)R29008 is the state vector, containing the temperatures at the discretization nodes. The input is the heat source Q˙(t) and the output is the temperature difference between the sensors. The parameter vector is p=[vκ]T and the parameter domain is normalized to D=[0,1]×[0,1]. For the sake of simplicity, the domain is sampled for the set of points of an equidistant grid P={0,1/5,,1}×{0,1/5,,1}, which result in the set of N=36 high-order systems G={G1,,G36}. The local systems are reduced to order q=10 using IRKA, which result in a set of asymptotically stable reduced systems Gˆ={Gˆ1,,Gˆ36} according to Equation (4). For testing the systems Gˆ, we compute the error in H-norm at the points of the grid P using the formula

(35) eH(p)=||G(p)Gˆ(p)||H||G(p)||Hmax1kKσmax{G(jωk,p)Gˆ(jωk,p)}max1kKσmax{G(jωk,p)},(35)

where σmax is the largest singular value and j=1. For approximating this error, we choose K=50 equally spaced frequencies ω1,ω2,,ωK in the frequency range of interest [101,,106] and we obtain the error eHP:PR0+ with eHP[3.6105,3.0104]. We calculate the reference basis V0 using the non-weighted SVD approach (Equation (6)) and the reference basis W0 is determined in the dual way.

5.2.1. Matrix interpolation without using STABLE

Firstly, the interpolation procedure will be demonstrated without the stability-preserving method, which corresponds to the framework from Section 2.4. The respective values are labelled with WS. The transformation matrices for adjusting the right ROBs are calculated using the MAC or DS approach minimizing the objective function (Equation (7) or (8)) which results in TMAC,iWS=(V0TVi)1 or TDS,iWS=Vi+V0 for i{1,36}. For adjusting the left ROBs, we obtain MMAC,iWS=(W0TWi)1 or MDS,iWS=Wi+W0 using the MAC or DS approach minimizing the objective function (Equation (9) or (10)). The local systems are transformed according to Equation (11) and result in the systems G˜WS={G˜1WS,,G˜36WS}. The parametric reduced system G˜WS(p) is obtained by interpolating the systems of the set G˜WS according to Equations (12) and (13) using bilinear weighting functions. For testing the system, we compute the error in H-norm at the points of a test grid Ptest={0,1/80,,1}×{0,1/80,,1} according to Equation (35). In , the error e:test0+ is plotted for both quality functions. The reader can verify that there are regions in the domain – where the H-norm equals infinity – which exhibit instability of the interpolation procedure.

Figure 3. Error eHPtest without the stability-preserving method and the (a) MAC approach and the (b) DS approach.

Figure 3. Error eH∞Ptest without the stability-preserving method and the (a) MAC approach and the (b) DS approach.

5.2.2. Matrix interpolation using STABLE

Secondly, the proposed stability-preserving method is applied. The transformation matrices for adjusting the right ROBs are Ti=TiWS for i{1,36} for the MAC and DS approaches minimizing the objective function (Equation (7) or (8)). For adjusting the left ROBs, we calculate the transformation matrices M1,,M36 using STABLE with the MAC and DS approach. For specifying and solving the convex program STABLE, the package CVX is used [Citation32,Citation33]. The results are obtained using solver SeDuMi [Citation37] and verified with SDPT3 [Citation40]. Operating on a 2.20GHz processor, solving the semidefinite programs lasted tˉSTABLE=1s on average for each of the N=36 grid points. The local systems of the set Gˆ are transformed according to Equation (11) and result in the systems G˜={G˜1,,G˜36}. The parametric reduced system G˜(p) is obtained by interpolating the systems of the set G˜ according to Equations (12) and (13) using bilinear weighting functions. The error eHPtest:PtestR0+ according to Equation (35) of the reduced system G˜(p) is shown in for the MAC approach and the DS approach. The MAC approach leads to the mean error eˉHPtest,MAC=0.016 and the maximum error eH,maxPtest,MAC=0.070. The DS approach has the mean error eˉHPtest,DS=0.017 and the maximum error eH,maxPtest,DS=0.071. The reader can verify that the proposed method preserves the stability of the interpolated system for the entire domain D. In addition, the method delivers nearly as accurate results as the method without stability preservation – compared for regions where the latter is asymptotically stable – since the left ROBs are distorted as little as possible.

Figure 4. Error eHPtest using STABLE and the (a) MAC approach and (b) DS approach.

Figure 4. Error eH∞Ptest using STABLE and the (a) MAC approach and (b) DS approach.

6. Conclusions

In this article, a method for pMOR by matrix interpolation with stability preservation is presented. This approach makes neither demands on the structure of the original system nor needs post-processing during the online phase to achieve stability and therefore is suitable for real-time applications. The procedure is based on low-order convex optimization problems – so-called semidefinite programs – which can easily be implemented and solved.

Disclosure statement

No potential conflict of interest was reported by the authors.

Notes

1. The matrix 0n×n of size n × n contains only zeros. The matrix In is the identity matrix of size n. Symmetric matrices of size n are denoted by QSn. Symmetric positive definite matrices of size n are denoted by QSn++ or Q > 0n×n. Symmetric negative definite matrices of size n are denoted by Q < 0n×n.

References

  • A.C. Antoulas, Approximation of Large-Scale Dynamical Systems, SIAM, Philadelphia, 2005.
  • S. Grivet-Talocia, Black-box passive macromodeling in electronics: trends and open problems, in ECMI Newsletter, Number 56, European Consortium for Mathematics in Industry, Helsinki, 2014, pp. 40–43.
  • S. Gugercin, R.V. Polyuga, C. Beattie, and A. Van Der Schaft, Structure-preserving tangential interpolation for model reduction of Port-Hamiltonian systems, Automatica 48 (9) (2012), pp. 1963–1974. doi:10.1016/j.automatica.2012.05.052
  • P. Benner, S. Gugercin, and K. Willcox, A survey of projection-based model reduction methods for parametric dynamical systems, SIAM Rev. 57 (4) (2015), pp. 483–531. doi:10.1137/130932715
  • D. Amsallem and C. Farhat, An online method for interpolating linear parametric reduced-order models, SIAM J. Sci. Comput. 33 (5) (2011), pp. 2169–2198. doi:10.1137/100813051
  • J. Degroote, J. Vierendeels, and K. Willcox, Interpolation among reduced-order matrices to obtain parameterized models for design, optimization and probabilistic analysis, Int. J. Numer. Methods Fluids 63 (2) (2010), pp. 207–230.
  • M. Geuss. A Black-box method for parametric model order reduction based on matrix interpolation with application to simulation and control, PhD thesis, submitted, TU München, 2015.
  • M. Geuss, H. Panzer, and B. Lohmann, On parametric model order reduction by matrix interpolation, Proceedings of the European Control Conference, Zurich, Switzerland, 2013, pp. 3433–3438.
  • B. Lohmann and R. Eid, Efficient order reduction of parametric and nonlinear models by superposition of locally reduced models, in Methoden und Anwendungen der Regelungstechnik. Erlangen-Münchener Workshops 2007 und 2008, B. Lohmann and G. Roppenecker, eds., Shaker-Verlag, Aachen, 2009, pp. 1–20.
  • H. Panzer, J. Mohring, R. Eid, and B. Lohmann, Parametric model order reduction by matrix interpolation, at-Automatisierungstechnik 58 (8) (2010), pp. 475–484. doi:10.1524/auto.2010.0863
  • R. Eid, R. Castañé-Selga, H. Panzer, T. Wolf, and B. Lohmann, Stability-preserving parametric model reduction by matrix interpolation, Math. Comput. Model. Dyn. Syst. 17 (4) (2011), pp. 319–335. doi:10.1080/13873954.2011.547671
  • M. Giftthaler, T. Wolf, and H. Panzer, Parametric model order reduction of Port-Hamiltonian systems by matrix interpolation, at-Automatisierungstechnik 62 (9) (2014), pp. 619–628. doi:10.1515/auto-2013-1072
  • O. Farle, S. Burgard, and R. Dyczij-Edlinger, Passivity preserving parametric model-order reduction for non-affine parameters, Math. Comput. Model. Dyn. Syst. 17 (3) (2011), pp. 279–294. doi:10.1080/13873954.2011.562901
  • F. Ferranti, G. Antonini, T. Dhaene, and L. Knockaert, Guaranteed passive parameterized model order reduction of the Partial Element Equivalent Circuit (PEEC) method, IEEE Trans. Electromagn. Compat. 52 (4) (2010), pp. 974–984. doi:10.1109/TEMC.2010.2051949
  • L. Knockaert, T. Dhaene, F. Ferranti, and D. De Zutter, Model order reduction with preservation of passivity, non-expansivity and markov moments, Syst. Control. Lett. 60 (1) (2011), pp. 53–61. doi:10.1016/j.sysconle.2010.10.006
  • E.R. Samuel, F. Ferranti, L. Knockaert, and T. Dhaene, Passivity-preserving parameterized model order reduction using singular values and matrix interpolation, IEEE Trans. Compon. Packag. Manuf. Technol. 3 (6) (2013), pp. 1028–1037. doi:10.1109/TCPMT.2013.2248196
  • B.N. Bond and L. Daniel, Stable reduced models for nonlinear descriptor systems through piecewise-linear approximation and projection, IEEE Trans. Computer-Aided Des. Integr. Circuits Syst. 28 (10) (2009), pp. 1467–1480. doi:10.1109/TCAD.2009.2030596
  • M. Geuss, H. Panzer, T. Wolf, and B. Lohmann, Stability preservation for parametric model order reduction by matrix interpolation, Proceedings of the European Control Conference, Strasbourg, France, 2014, pp. 1098–1103.
  • T. Kailath, Linear Systems, Prentice Hall, Englewood Cliffs, 1980.
  • H.K. Khalil, Nonlinear Systems, 3rd ed., Prentice Hall, Upper Saddle River, NJ, 2002.
  • T. Stykel, Analysis and numerical solution of generalized Lyapunov equations, PhD thesis, TU Berlin, 2002.
  • H. Panzer, B. Kleinherne, and B. Lohmann, Analysis, Interpretation and generalization of a strictly dissipative state space formulation of second order systems, in Methoden und Anwendungen der Regelungstechnik. Erlangen-Münchener Workshops 2011 und 2012, B. Lohmann and G. Roppenecker, eds., Shaker-Verlag, Aachen, 2013, pp. 59–75.
  • R. Castañé-Selga, B. Lohmann, and R. Eid, Stability preservation in projection-based model order reduction of large scale systems, Eur. J. Control 18 (2) (2012), pp. 122–132. doi:10.3166/ejc.18.122-132
  • M. Geuss, H. Panzer, I. Clifford, and B. Lohmann, Parametric model order reduction using pseudoinverses for the matrix interpolation of differently sized reduced models, Proceedings of the 19th IFAC World Congress, Cape Town, South Africa, 2014, pp. 9468–9473.
  • H. Panzer, S. Jaensch, T. Wolf, and B. Lohmann, A greedy rational Krylov method for μ2-pseudooptimal model order reduction with preservation of stability. Proceedings of the American Control Conference, Washington, DC, 2013, pp. 5512–5517.
  • S. Gugercin, A.C. Antoulas, and C. Beattie, μ2 model reduction for large-scale linear dynamical systems, SIAM J. Matrix Anal. Appl. 30 (2) (2008), pp. 609–638. doi:10.1137/060666123
  • D. Amsallem and C. Farhat, Stabilization of projection-based reduced-order models, Int. J. Numer. Methods Eng. 91 (4) (2012), pp. 358–377. doi:10.1002/nme.v91.4
  • B.N. Bond and L. Daniel, Guaranteed stable projection-based model reduction for indefinite and unstable linear systems, Proceedings of the IEEE/ACM International Conference on Computer-Aided Design, San Jose, CA, 2008, pp. 728–735.
  • L. Vandenberghe and S. Boyd, Semidefinite programming, SIAM Rev. 38 (1) (1996), pp. 49–95. doi:10.1137/1038003
  • S. Boyd and L. Vandenberghe, Convex Optimization, Cambridge University Press, New York, NY, 2004.
  • S. Boyd, L. El Ghaoui, E. Feron, and V. Balakrishnan, Linear matrix inequalities in system and control theory, in Studies in Applied Mathematics, SIAM, Philadelphia, PA, 1994.
  • M. Grant and S. Boyd, CVX: MATLAB software for disciplined convex programming, version 2.0, Aug, 2012. Available at http://cvxr.com/cvx.
  • M. Grant and S. Boyd, Graph implementations for nonsmooth convex programs, in Recent Advances in Learning and Control, Lecture Notes in Control and Information Sciences, V. Blondel, S. Boyd, and H. Kimura, eds., Springer, London, 2008, pp. 95–110.
  • Y. Nesterov and A. Nemirovskii, Interior-point polynomial algorithms in convex programming, in Studies in Applied Mathematics, SIAM, Philadelphia, PA, 1994.
  • M.J. Todd, Semidefinite optimization, Acta Numer. 10 (2001), pp. 515–560. doi:10.1017/S0962492901000071
  • L. Vandenberghe and S. Boyd, A polynomial-time algorithm for determining quadratic Lyapunov functions for nonlinear systems. Proceedings of the European Conference Circuit Theory Design, Davos, Switzerland, 1993, pp. 1065–1068.
  • J.F. Sturm, Using SeDuMi 1.02, a MATLAB toolbox for optimization over symmetric cones, Optim. Methods Softw. 11 (1999), pp. 625–653. doi:10.1080/10556789908805766
  • MOR Wiki. Available at http://www.modelreduction.org.
  • C. Moosmann, E.B. Rudnyi, A. Greiner, J.G. Korvink, and M. Hornung, Parameter preserving model order reduction of a flow meter. Proceedings of the NSTI Nanotech, 3, Anaheim, CA, 2005, pp. 684–687.
  • R.H. Tütüncü, K.C. Toh, and M.J. Todd, Solving semidefinite-quadratic-linear programs using SDPT3, Math. Programming Ser. B 95 (2003), pp. 189–217. doi:10.1007/s10107-002-0347-5

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.