277
Views
5
CrossRef citations to date
0
Altmetric
Articles

A double optimal descent algorithm for iteratively solving ill-posed linear inverse problems

Pages 38-66 | Received 04 Sep 2013, Accepted 05 Jan 2014, Published online: 17 Feb 2014

Abstract

In the iterative solution of an ill-posed linear system Bx=b, how to select a fast and easily established descent direction u to reduce the residual r:=Bx-b is an important issue. A mathematical procedure to find a double optimal descent direction u in Bu=r, without inverting B, is developed in an m-dimensional Krylov subspace. The novelty is that we expand u in an affine Krylov subspace with undetermined coefficients, and then two optimization techniques are used to determine these coefficients in closed form, which can greatly accelerate the convergence speed in solving the ill-posed linear problems. The double optimal descent algorithm is proven to be absolutely convergent very fast, accurate and robust against noise, which is confirmed by numerical tests of several linear inverse problems, including the heat source identification problem, the backward heat conduction problem, the inverse Cauchy problem and the external force identification problem.

1 Introduction

In this paper, we develop a double optimal descent algorithm (DODA) in an affine Krylov subspace to iteratively solve the following linear equations system:1 Bx=b,1 where xRn is an unknown vector, to be determined from a given non-singular coefficient matrix BRn×n, and the input bRn. For the existence of the solution x, we suppose that det(B)0, and hence B has a full rank with rank(B)=n.

The linear inverse problems are sometimes obtained via an n-dimensional discretization of a bounded linear operator equation under a noisy input. When B is severely ill conditioned and the data are corrupted by noise, we may encounter the problem that the numerical solution of Equation (Equation1) might deviate from the exact one to a great extent, and in general a generalized solution x=Bb is found, where B is a pseudo-inverse of B in the Penrose sense.

If we only know perturbed input data bδRn with b-bδδ, and if the problem is ill posed, i.e. the range(B) is not closed or equivalently B is unbounded, we have to solve Equation (Equation1) by a regularization method.

A measure of the ill-posedness of Equation (Equation1) can be performed by using the condition number of B [Citation1]:2 cond(B)=BFB-1F,2 where BF denotes the Frobenius norm of B.

For every matrix norm , we have ρ(B)B, where ρ(B) is a radius of the spectrum of B. The Householder theorem states that for every ϵ>0 and every matrix B, there exists a matrix norm B depending on B and ϵ such that Bρ(B)+ϵ. Anyway, the spectral condition number ρ(B)ρ(B-1) can be used as an estimation of the condition number of B by3 cond(B)=maxσ(B)|λ|minσ(B)|λ|,3 where σ(B) is the collection of all the eigenvalues of B. Turning back to the Frobenius norm, we have4 BFnmaxσ(B)|λ|.4 In particular, when B is symmetric, ρ(B)ρ(B-1)=B2B-12. Roughly speaking, the numerical solution of Equation (Equation1) may lose the accuracy of k decimal points when cond(B)=10k.

Due to the ill-posedness of the problem, small errors in the data may lead to computed solution far off the correct one, rendering straightforward approaches to solve Equation (Equation1) useless. To overcome the sensitivity to noise, it is often using a regularization method to solve the ill-posed problem,[Citation2Citation5] where a suitable regularization parameter is used to suppress the approximation error and propagated data error, ensuring stability of the solution x with respect to the perturbed data bδ.

The iterative algorithm for solving linear equations system can be derived from the discretization of a certain ordinary differential equations system.[Citation6] Particularly, some descent methods can be interpreted as the discretizations of gradient flows.[Citation7] For a large-scale linear equations system, the major choice is using an iterative algorithm, where an early stopping criterion is used to prevent the reconstruction of noisy component in the approximate solution, and which has the regularization effect. Recently, some useful and simple methods have been developed by the author for solving Equation (Equation1), like the optimally generalized regularization method,[Citation8] optimally scaled vector regularization method,[Citation9] adaptive Tikhonov regularization method,[Citation10] optimal preconditioner method,[Citation11] the globally optimal iterative method,[Citation12] as well as the optimal and globally optimal tri-vector iterative algorithms (OTVIA).[Citation13, Citation14]

There are a lot of numerical methods that converge significantly faster than the steepest descent method (SDM), and unlike the conjugate gradient method (CGM), they insist their search directions to be the gradient vector at each iteration.[Citation15] The SDM performs poorly, yielding the iteration counts that grow linearly with cond(B),[Citation16] whose unwelcome slowness has to do with the choice of the gradient descent direction R and its original steplength R2/BR2 detected by the SDM. Liu [Citation13] has explored a variant of the SDM by feeding the concept of an optimal descent tri-vector to solve ill-posed linear equations system, which is an optimal combination of the steepest descent vector R:=BTr, the residual vector r and a supplemental vector BR, where not only the direction R but also the steplength are modified from a theoretical foundation of optimization being realized on an invariant manifold. This novel method outperformed better than the generalized minimal residual method (GMRES),[Citation17] the CGM, and other gradient descent variant methods. The concept of optimal vector driving algorithm was explored by Liu and Atluri [Citation18] for solving linear equations. The idea of using two optimizations in the solution of Equation (Equation1) in a Krylov subspace was proposed by Liu [Citation19]. Then, Liu [Citation20] used the scaling invariant property of the merit function for a maximal projection to derive a maximal projection solution of linear problem in an affine Krylov subspace, which is proven better than the least-square solution. As a continuation, we further explore the concept of double optimal iterative algorithm with an m-vector in a Krylov subspace as a descent direction to solve Equation (Equation1).

The remaining parts of this paper are arranged as follows. In Section 2, we review the invariant-manifold method. The Krylov subspace method used to derive an optimal descent direction is demonstrated in Section 3. In Section 4, we solve the optimal parameters appeared in the optimal descent direction by two optimization techniques, of which the doubly optimized solutions of these parameters are given in a closed-form type. Section 5 outlines the numerical procedures of the m-vector DODA in terms of m+1 weighting parameters which are optimized explicitly from two different merit functions, where the convergence of the DODA is proven. The numerical examples of linear inverse problems solved by the iterative algorithm of DODA are given in Section 6, where we compare the numerical performance of DODA with other several algorithms. Finally, the conclusions and discussions are drawn in Section 7.

2 Invariant manifold

For the linear equations system (Equation1), which is expressed to be r=0 in terms of the residual vector:5 r=Bx-b,5 we can introduce a scalar function:6 h(x,t)=Q(t)2r2-12r02=0,6 where Q(t)>0 is a monotonically increasing function of t, x0 is an initial value of x and r0=Bx0-b.

In order to keep the iterative orbit xk evolving on the manifold, Liu and Atluri [Citation18] have derived an iterative algorithm:7 xk+1=xk-ηrk2rkTvkuk,7 where8 v:=Bu,8 9 η=1-γa0,9 10 a0:=r2v2(rTv)21.10 In the above,11 0γ<111 is a relaxation parameter.

Because the convergence speed is closely correlated to a0 by12 Convergence Rate:=r(t)r(t+Δt)=1s>1,12 where13 s=1-1-γ2a0,13 the author has previously developed several optimal iterative algorithms basing on the minimization of a0. On the other hand, the termr·(Bu)Bu2in Equation (Equation7) is a steplength. Liu [Citation21] has proposed a new algorithm to solve Equation (Equation1) by maximizing the steplength in a Krylov subspace.

3 The Krylov subspace method

There are several numerical solution methods of Equation (Equation1), which are originated from the idea of minimization. For the positive definite linear system, solving Equation (Equation1) by the SDM is equivalent to solve the following minimization problem [Citation22, Citation23]:14 minxRnφ(x)=minxRn12xTBx-bTx,14 where B is a positive definite matrix.

Liu [Citation12Citation14] has proposed new methods by minimizing the following merit function:15 mina0=r2Bu2[r·(Bu)]2,15 to obtain a fast descent direction u, in terms of two or three vectors, for the iterative solution of Equation (Equation1). Because the above minimization is quite difficult when u is expressed as a linear combination of multi vectors, Liu [Citation24] has solved, instead of Equation (Equation15), the following minimization problem in a Krylov subspace:16 minr-Bu2=b-Bx2.16 In the solution of linear equations system, the Krylov subspace method is one of the most important classes of numerical methods, and the iterative algorithms that are applied to solve large-scale linear systems are mostly the preconditioned Krylov subspace methods.[Citation24Citation29] In the past few decades, the Krylov subspace methods for solving Equation (Equation1) have been studied in depth since the appearance of pioneering work,[Citation30, Citation31] such as the minimum residual algorithm,[Citation32] the GMRES,[Citation17, Citation27] the quasi-minimal residual method,[Citation28] the biconjugate gradient method,[Citation33] the conjugate gradient squared method [Citation34] and the biconjugate gradient stabilized method.[Citation35] There are more discussions on the Krylov subspace methods in the review papers [Citation26, Citation36] and text books.[Citation37, Citation38] The most are paying more attention to the residual vector and its ‘optimality’ of some residual errors in the Krylov subspace to derive the ‘best approximation’.

It is known that for the iterative algorithm to solve linear equations system (Equation1), the best descent direction u is given by B-1r.[Citation39] Then in order to find the best descent direction u, we require to solve the following linear problem:17 Bu=r.17 Suppose that we have an m-dimensional Krylov subspace generated by the coefficient matrix B from the right-hand side vector r in Equation (Equation17):18 Km:=span{r,Br,,Bm-1r}.18 Let Lm=BKm. The idea of GMRES is using the Galerkin method to search the solution uKm, such that the residual r-Bu=b-Bx is perpendicular to Lm.[Citation17] It can be shown that the solution uKm minimizes the residual [Citation37] in Equation (Equation16). It is known that the GMRES does not always perform well for ill-posed linear systems.[Citation40Citation44] In order to solve ill-posed linear problems based on the GMRES, Calvetti et al. [Citation40] have proposed the range restriction GMRES (RRGMRES), which expands the solution in the following Krylov subspace:19 Lm:=span{Br,B2r,,Bmr}.19 This method restricts the Krylov subspace to generate an approximate solution in the range of the coefficient matrix B. Calvetti et al. [Citation40] confirmed that the RRGMRES performs well for ill-posed linear problems.

Let p(λ) be the characteristic equation of the coefficient matrix B, which can be written as:20 p(λ)=λn+cn-1λn-1++c2λ2+c1λ-c0=0,20 where c0=-det(B)0 because we suppose that B is non-singular. The Cayley-Hamilton theorem [Citation45] asserts that21 p(B)=Bn+cn-1Bn-1++c2B2+c1B-c0In=0.21 From the above equation, we can expand B-1 by22 B-1=c1c0In+c2c0B+c3c0B2++cn-1c0Bn-2+1c0Bn-1,22 and hence, the solution of Equation (Equation17) is given by23 u=B-1r=c1c0In+c2c0B++cn-1c0Bn-2+1c0Bn-1r.23 The above process to find the optimal descent direction of u is quite difficult to be realized in practice, since the coefficients cj,j=0,1,,n-1 are hard to find when n is a quite large positive integer. Moreover, the computation of the higher order powers of B is very expensive. In order to have an effective iterative algorithm, the above series should be truncated properly. Hence, motivated by Equation (Equation23), we can suppose that u is expressed by24 u=βr+k=1mαkuk,24 which is to be determined as an optimal combination of r and the m-vector uk,k=1,,m, when the coefficients αk and β are optimized in Sections 4.2 and 4.3, respectively.

Now we describe how to set up the m-vector uk,k=1,,m by the Krylov subspace method. Suppose that we have an m-dimensional Krylov subspace generated by the coefficient matrix B from the right-hand side vector r in Equation (Equation17):25 Lm:=span{Br,,Bmr}.25 Then, the Arnoldi process is used to normalize and orthogonalize the Krylov vectors Bjr,j=1,,m, such that the resultant vectors ui,i=1,,m satisfy ui·uj=δij,i,j=1,,m, where δij is the Kronecker delta symbol.

Let26 U:=[u1,,um]26 be an n×m Krylov matrix with its jth column being the vector uj. Because u1,,um are linearly independent vectors and m<n, the rank of U is rank(U)=m. Hence, Equation (Equation24) can be written as:27 u=u0+Uα,27 where28 u0=βr,28 and α:=(α1,,αm)T. The superscript T denotes the transpose.

Remark 1

The Krylov subspace method is an iterative method, of which the mth step approximation to the solution of Equation (Equation1) is found in x0+Km. The approximation is of the form xm=x0+pm-1(B)r0, where pm-1 is a polynomial at most with m-1 degree. Usually, the number m is smaller than the degree of the minimal polynomial of B, and indeed in the numerical algorithm we can choose mn.

4 Doubly optimized descent direction of u

The most of the existent numerical methods to solve linear system (Equation1) in the Krylov subspace are paying attention to the minimizations of the residual norm or to the fulfilment of some Galerkin conditions. To the best knowledge of the author, there exists no numerical method to find the numerical solution of Equation (Equation1), simultaneously based on the two minimizations in Equations (Equation15) and (Equation16).

4.1 The first optimization

Let29 v:=Bu,29 and we attempt to establish a merit function, such that its minimization leads to the best fit of v to r, because Bu=r is just the equation we use to find the best descent direction u.

We consider finding the best approximation of v to r. The orthogonal projection of r to v is regarded as the approximation of r by v, whose error vector is written as:30 e:=r-r,vvvv,30 where the brace denotes the inner product. The best approximation can be found with v minimizing the square norm of error vector:31 e2=r2-(r·v)2v20,31 or maximizing the square norm of the orthogonal projection of r to v, i.e.32 maxv(r·v)2v2.32 Let us define the following merit function:33 f:=v2(r·v)21r2,33 of which the inequality follows from the Cauchy-Schwarz inequality:r·vrv.Let J be an n×m matrix:34 J:=[Bu1,,Bum]=BU,34 where U is defined by Equation (Equation26). Then, with the aid of Equations (Equation27) and (Equation29) v can be written as:35 v=v0+Jα,35 where36 v0:=Bu0=βBr.36 Inserting Equation (Equation35) for v into Equation (Equation33), we encounter the following minimization problem:37 minα1,,αmf=v2(r·v)2,37 where38 r·v=r·v0+rTJα,38 39 v1:=α(r·v)=JTr,39 40 v2=v02+2v0TJα+αTJTJα,40 41 v2:=αv2=2JTv0+2JTJα,41 in which α denotes the gradient with respect to α. Equation (Equation41) is obtained from Equation (Equation40) by taking the derivative of v2 with respect to α, which is given below by using the componential form:42 v2αk=αk2v0iJijαj+αiCijαj=2v0iJijδjk+δikCijαj+αiCijδjk=2v0iJik+Ckjαj+αiCik,42 where Cij is the ijth component of43 C:=JTJ,43 which is an m×m positive definite matrix. Because of CT=C, from the last equation in the above we can deduce Equation (Equation41). Here, we have to emphasize that Equation (Equation42) minimization of f is fully equivalent to the minimization of a0 defined by Equation (Equation10), because r is a known vector.

By using44 αv2(r·v)2=0(r·v)2αv2-2r·vv2α(r·v)=0,44 we can derive the following equation to solve α:45 r·vv2-2v2v1=0.45

4.2 A closed-form solution of α

In view of Equation (Equation43), Equations (Equation40) and (Equation41) can be written as:46 v2=v02+2v0TJα+αTCα,46 47 v2=2JTv0+2Cα.47 Because J has a full column rank(J)=m, the positivity of C is guaranteed. From Equation (Equation45), we can observe that v2 is proportional to v1, which is supposed to be48 v2=2v2r·vv1=2λv1,48 where 2λ is a multiplier to be determined.

Then, by Equations (Equation39), (Equation47) and (Equation48), we have49 α=λDJTr-DJTv0,49 where50 D:=C-1=(JTJ)-150 is an m×m positive definite matrix. Inserting Equation (Equation49) into Equations (Equation38) and (Equation46) we have51 r·v=r·v0+λrTEr-rTEv0,51 52 v2=λ2rTEr+v02-v0TEv0,52 where53 E:=JDJT53 is an n×n positive semi-definite matrix.

Now, from Equations (Equation48), (Equation51) and (Equation52), we can derive the following equation:54 λ2rTEr+v02-v0TEv0=λ[r·v0+λrTEr-rTEv0].54 By cancelling the quadratic term λ2rTEr on both sides, we can obtain a linear equation, which renders a closed-form solution of λ:55 λ=v02-v0TEv0r·v0-rTEv0,55 and from Equation (Equation49), we can obtain the closed-form solution of α:56 α=v02-v0TEv0r·v0-rTEv0DJTr-DJTv0.56 Inserting the above α into Equation (Equation27), we can obtain57 u=u0+λQr-Qv0=β[λ0Qr+r-QBr],57 where58 Q:=UDJT,58 59 λ0:=rTBTBr-rTBTEBrrTBr-rTEBr,59 are, respectively, an n×n constant matrix and a constant scalar, both being fully determined by the coefficient matrix B and the right-hand side vector r in Equation (Equation17).

4.3 The second optimization to find β

Upon letting60 v:=λ0Qr+r-QBr,60 u in Equation (Equation57) can be expressed as61 u=βv.61 We can derive the closed-form solution of β by the second optimization of the second merit function:62 Bu-r2=βBv-r2=β2w2-2βw·r+r2,62 where63 w:=Bv=Br+λ0BQr-BQBr.63 Taking the derivative of Equation (Equation62) with respect to β and equating it to zero, we can obtain64 β=w·rw2.64 Inserting it into Equation (Equation61) we arrive at65 u=w·rw2v,65 where v is defined by Equation (Equation60) and w is defined by Equation (Equation63).

4.4 The proof of βλ0=1

Inserting Equation (Equation34) for J into the first J in Equation (Equation53) and comparing the resultant with Equation (Equation58), it immediately follows that66 E=BQ.66 Then, by using Equations (Equation53) and (Equation50), we have67 E2=E,67 such that E is a projection operator. Indeed, E2=E is a well-known result that E=J(JTJ)-1JT is an orthogonal projection operator.

In terms of E, w defined in Equation (Equation63) can be written as:68 w=λ0Er+Br-EBr.68 It follows that69 w2=λ02rTEr+rTBTBr-rTBTEBr,69 70 r·w=λ0rTEr+rTBr-rTEBr,70 where Equation (Equation67) was used in the first equation. With the aid of Equation (Equation59), Equation (Equation70) is further reduced to71 r·w=λ0rTEr+1λ0[rTBTBr-rTBTEBr].71 Now, after inserting Equation (Equation71) into Equation (Equation64), we can obtain72 βλ0w2=λ02rTEr+rTBTBr-rTBTEBr;72 however, in view of Equation (Equation69), the right-hand side is just equal to w2. Hence, we have73 βλ0=1.73 As a consequence, we have a neater form of the doubly optimized descent direction of u, which is given by74 u=β[r-QBr]+Qr,74 where75 β=rTBr-rTEBrrTBTBr-rTBTEBr.75 Equation (Equation74) is better than Equation (Equation65), for saving computations.

Sometimes it is better to use the following normal equation:76 BTBu=R:=BTr76 to find the best descent direction u, because the coefficient matrix BTB is now positive definite. The process to find the best direction u is the same, where we only need to replace B by BTB, and r by R.

5 A double optimal descent algorithm

5.1 The numerical algorithm

The numerical procedure of the DODA is described in this section. Before that we need to compute the inverse matrix D in Equation (Equation50). The accuracy of this matrix is crucial for the DODA, and we can use the following matrix conjugate gradient method (MCGM) developed by Liu et al. [Citation46] to find the the inverse matrix D of C:

(i) Assume an initial D0.

(ii) Calculate R0=Im-CD0 and P1=R0.

(iii) For k=1,2,, we repeat the following iterations:77 αk=Rk-12Pk:(CPk),Dk=Dk-1+αkPk,Rk=Im-CDk,ηk=Rk2Rk-12,Pk+1=Rk+ηkPk.77 If Dk converges according to a given stopping criterion, such that,78 Rk<ε1,78 then stop; otherwise, go to step (iii). In the above, the boldfaced capital letters denote m×m matrices, the norm Rk is the Frobenius norm and the inner product symbol : is used for matrices. Because for the inverse problems to be computed we will use rather small m, the inverse matrix D is easily computed by the above MCGM algorithm, which is convergent very fast.

Thus, we arrive at the following DODA:

(i)

Select m and 0γ<1, and give an initial value of x0.

(ii)

For k=0,1,, we repeat the following computations:79 rk=Bxk-b,Compute Equation(74) to obtainuk,vk=Buk,xk+1=xk-(1-γ)rk·vkvk2uk.79 If xk+1 converges according to a given stopping criterion rk+1<ε, then stop; otherwise, go to step (ii).

Remark 2

Note that uj,j=1,,m are constructed from the Krylov subspace method and following by the Arnoldi process to orthonormalize the Krylov vectors, the resultant optimal algorithm is the DODA with the Krylov subspace method. In the Krylov subspace method, the expansion matrix J is not fixed, and it can adjust its configuration at each iterative step to accelerate the convergence speed, which, however, leads to a consumption of computational time in the calculation of the inverse of C=JTJ to obtain D, E and Q at every iterative step.

5.2 The proof of the convergence of DODA

In this section, we prove that the algorithm DODA is convergent. From the second equality in Equation (Equation48) by cancelling the common term v1 on both sides, we have80 v2=λr·v,80 where λ, by using Equations (Equation55) and (Equation36), can be written as81 λ=βrTBTBr-rTBTEBrrTBr-rTEBr.81 With the help of Equations (Equation59) and (Equation73), we have82 λ=βλ0=1;82 hence, Equation (Equation80) becomes83 v2=r·v.83 In view of Equation (Equation83), the iterative algorithm (Equation79) reduces to84 xk+1=xk-(1-γ)uk.84 By using Equations (Equation8) and (Equation5), it follows that85 Bxk+1=Bxk-(1-γ)Buk,rk+1=rk-(1-γ)vk.85 Taking the square norms of both sides and using Equation (Equation83), we can derive86 rk+12=rk2-(1-γ2)vk2.86 Because of 1-γ2>0 and vk2>0, it immediately leads to87 rk+12<rk2,87 which indicates that the algorithm DODA is absolutely convergent.

Remark 3

The present algorithm DODA can provide a fast reduction of the residual. Indeed for the DODA, from Equations (Equation29), (Equation74), (Equation75) and (Equation66), we have88 vk=β(In-E)Brk+Erk,vk2=rkTErk+[rkT(In-E)Brk]2rkTBT(In-E)Brk,88 where (In-E)E=0 and Equation (Equation67) were used in the derivation of the second equation. In terms of the intersection angle θ between (In-E)rk and (In-E)Brk, we have89 vk2=rkTErk+(In-E)rk2cos2θ.89 If θ=0, vk2=rk2 and the DODA with γ=0 converges with one step, due to rk+12=0 by Equation (Equation86). On the other hand, if we take m=n, then the DODA also converges with one step. We can see that if a suitable value of m is taken then the DODA can converge within n steps. Therefore, we have the following convergence criterion of the DODA. If90 j=0N(1-γ2)vj2r02-ε2,90 then the iterations in Equation (Equation79) terminate, where Nn.

6 Numerical examples

In order to evaluate the performance of the newly developed method of DODA, we test some linear inverse problems. Some numerical results are compared with that computed by the GMRES,[Citation17] RRGMRES,[Citation40] the OMVIA developed by Liu [Citation24], the OTVIA developed by Liu [Citation13] and a recent non-iterative algorithm based on the double optimal solution (DOS).[Citation19] The algorithm in [Citation19] is applied to solve linear problem by selecting a suitable value of m for the dimension of the Krylov subspace.

In the comparison of numerical solution with exact solution, we use two criteria to measure the accuracy of numerical solution. First, the numerical error is defined to be the absolute value of the difference between numerical solution and exact solution. In addition to that, we also compare the root-mean-square error (RMSE), which is defined by91 RMSE:=1nk=1nxkN-xkE2,91 where xkN and xkE are, respectively, the numerical solution and exact solution, and n is the total number of data points to be compared.

6.1 Example 1

Finding an n-order polynomial function p(x)=c0+c1x++cnxn to best match a continuous function f(x) in the interval of x[0,1]:92 mindeg(p)n01[f(x)-p(x)]2dx,92 leads to a problem governed by Equation (Equation1). The coefficient matrix B is the (n+1)×(n+1) Hilbert matrix defined by93 Bij=1i+j-1,93 x is composed of the n+1 coefficients c0,c1,,cn appeared in p(x), and94 b=01f(x)dx[2px]01xf(x)dx01xnf(x)dx94 is uniquely determined by the function f(x).

The Hilbert matrix is a notorious example of highly ill-conditioned matrices. Equation (Equation1) with the matrix B having a large condition number usually displays that an arbitrarily small perturbation of data on the right-hand side may lead to an arbitrarily large perturbation to the solution on the left-hand side.

In this example, we consider a highly ill-conditioned linear system (Equation1) with B given by Equation (Equation93). The ill-posedness of Equation (Equation1) increases fast with n. We consider an exact solution with xj=1,j=1,,n and bi is given by95 bi=j=1n1i+j-1+σR(i),95 where R(i) are random numbers between [-1,1].

It is known that the condition number of Hilbert matrix grows as e3.5n when n is very large. For the case with n=300, the condition number is extremely huge up to the order of 10522. We solve this problem by using the DODA with the Krylov subspace method. Under a noise σ=10-6, and for n=300 and m=5, the DODA is convergent very fast only through three steps with the convergence criterion ε=10-3, where the maximum error is 0.0158. It is very time saving because we only need to calculate the 5×5 matrix C-1 three times, which is using the MCGM developed by Liu et al. [Citation46]. Under a convergence criterion ε1=10-5 in Equation (Equation78), the process to find D is convergent very fast with five or six iterations, although we do not show them at here. The numerical results are shown in Figure . It is interesting that even for a large value of n=300, we do not need a large value of m, of which merely m=5 was used. Also we can observe that the accuracy does not lose although the ill-posedness of the linear Hilbert problem is highly increased for n=300.

Figure 1. For example 1 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) numerical error.

Figure 1. For example 1 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) numerical error.

As mentioned, the algorithm in [Citation19] is applied to solve linear problem by selecting a suitable value of m for the dimension of the Krylov subspace. This algorithm has been named the DOS. Under a large noise with σ=0.05 and for n=300, we take m=2 and apply the GMRES, DODA and DOS to solve this problem, whose numerical errors are compared in Figure . The maximum error of DODA is 0.367, while that for the GMRES is 0.579 and DOS is 0.566. The RMSE obtained by the DODA is 0.154, while that obtained by the GMRES is 0.16, and that obtained by the DOS is 0.214. It can be seen that the DODA is slightly accurate than the GMRES and the DOS.

Figure 2. For example 1 under a large noise with σ=0.05, comparing numerical errors obtained by the DODA, DOS and GMRES.

Figure 2. For example 1 under a large noise with σ=0.05, comparing numerical errors obtained by the DODA, DOS and GMRES.

6.2 Example 2

In this section, we apply the DODA to identify an unknown space-dependent heat source function H(x) for a one-dimensional heat conduction equation:96 ut(x,t)=uxx(x,t)+H(x),0<x<,0<t<tf,96 97 u(0,t)=u0(t),u(,t)=u(t),97 98 u(x,0)=f(x).98 In order to identify H(x), we can impose an extra condition:99 ux(0,t)=q(t).99 We propose a numerical differential method by letting v=ut. Taking the differentials of Equations (Equation96), (Equation97) and (Equation99) with respect to t, and letting v=ut, we can derive100 vt(x,t)=vxx(x,t),0<x<,0<t<tf,100 101 v(0,t)=u˙0(t),101 102 v(,t)=u˙(t),102 103 vx(0,t)=q˙(t).103 This is an inverse heat conduction problem (IHCP) for v(x,t) without using the initial condition.

Therefore, we can first solve the above IHCP for v(x,t) by using the method of fundamental solutions (MFS) to obtain a linear equations system, and then the method introduced in Section 5 is used to solve the resultant linear equations system; hence, we can construct u(x,t) by104 u(x,t)=0tv(x,ξ)dξ+f(x),104 which automatically satisfies the initial condition in Equation (Equation98).

From Equation (Equation104), it follows that105 uxx(x,t)=0tvxx(x,ξ)dξ+f(x),105 which together with ut=v being inserted into Equation (Equation96) leads to106 v(x,t)=0tvxx(x,ξ)dξ+f(x)+H(x).106 Inserting Equation (Equation100) for vxx=vt into the above equation and integrating it, we can derive the following equation to recover H(x):107 H(x)=v(x,0)-f(x).107 For the purpose of comparison, we consider the following exact solutions:108 u(x,t)=x2+2xt+sin(2πx),H(x)=2x-2+4π2sin(2πx).108 In Equation (Equation107), we disregard the ill-posedness of f(x), and suppose that the data f(x) are given exactly. A random noise with an intensity σ=10% is added on the data q˙(t). Under the following parameters m=10 and γ=0.25, we solve this problem by the DODA with 200 steps. In Figure (a) and (b), we plot the residual, a0 and β, and the numerical solution is compared with the exact solution in Figure (c). The numerical error is shown in Figure (b), whose maximum error is 0.00421. It can be seen that the present DODA can provide very accurate numerical result. This result is better than that calculated by Liu [Citation47] using the vector regularization iterative method, whose maximum error is 0.0086. Then under the same parameters as that used in the DODA, we apply the GMRES,[Citation17] the RRGMRES [Citation40] and the OMVIA [Citation24] to solve this problem. The residuals and numerical errors are compared in Figure . When the maximum error of GMRES is 0.053, the maximum error of RRGMRES is 0.02, and the maximum error of OMVIA is 0.0265. The maximum error of DODA as shown in the above is 0.00421, which is the best one among the four numerical methods. The RMSE of GMRES is 0.032, the RMSE of RRGMRES is 0.014, the RMSE of OMVIA is 0.023 and the RMSE of DODA is 0.0024. It can be seen that the DODA is much accurate than other three numerical algorithms.

Figure 3. For example 2 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) numerical error.

Figure 3. For example 2 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) numerical error.

Figure 4. For example 2 solved by the DODA, GMRES, RRGMRES and OMVIA, comparing (a) residuals and (b) numerical errors.

Figure 4. For example 2 solved by the DODA, GMRES, RRGMRES and OMVIA, comparing (a) residuals and (b) numerical errors.

6.3 Example 3

When the backward heat conduction problem (BHCP) is considered in a spatial interval of 0<x< by subjecting to the boundary conditions at two ends of a slab:109 ut(x,t)=κuxx(x,t),0<t<T,0<x<,109 110 u(0,t)=u0(t),u(,t)=u(t),110 we solve u under a final time condition:111 u(x,T)=uT(x).111 The fundamental solution of Equation (Equation109) is given as follows:112 K(x,t)=H(t)2κπtexp-x24κt,112 where H(t) is the Heaviside function.

The MFS has a serious drawback that the resulting linear equations system is always highly ill conditioned, when the number of source points is increased, or when the distances of source points are increased.

In the MFS, the solution of u at the field point z=(x,t) can be expressed as a linear combination of the fundamental solutions U(z,sj):113 u(z)=j=1ncjU(z,sj),sj=(ηj,τj)Ωc,113 where n is the number of source points, cj are unknown coefficients and sj are source points being located in the complement Ωc of Ω=[0,]×[0,T]. For the heat conduction equation, we have the basis functions114 U(z,sj)=K(x-ηj,t-τj).114 It is known that the location of source points in the MFS has a great influence on the accuracy and stability. In a practical application of MFS to solve the BHCP, the source points are uniformly located on two vertical straight lines parallel to the t-axis, not over the final time, which was adopted by Hon and Li [Citation48] and Liu [Citation49], showing a large improvement than the line location of source points below the initial time. After imposing the boundary conditions and the final time condition to Equation (Equation113), we can obtain a linear equations system:115 Bx=b,115 where116 Bij=U(zi,sj),x=(c1,,cn)T,b=(u(ti),i=1,,m1;uT(xj),j=1,,m2;u0(tk),k=m1,,1)T,116 and n=2m1+m2.

Since the BHCP is highly ill posed, the ill condition of the coefficient matrix B in Equation (Equation115) is serious. To overcome the ill-posedness of Equation (Equation115), we can use the DODA to solve this problem. Here, we compare the numerical solution with an exact solution:u(x,t)=cos(πx)exp(-π2t).For the case with T=1, the value of final time data is in the order of 10-4, which is small by comparing with the value of the initial temperature f(x)=u0(x)=cos(πx) to be retrieved, which is O(1). First we impose a relative random noise with an intensity σ=10% being imposed on the final time data. Under the following parameters m1=15, m2=8, m=16, γ=0.005 and ε=10-2 and ε1=10-8, we solve this problem by the DODA. With five steps the DODA is convergent. In Figure (a) and (b), we plot the residual, a0 and β, and the numerical solution is compared with the exact solution in Figure (c), whose maximum error is 9.25×10-3. It can be seen that the present DODA converges very fast and is very robust against noise, and we can provide a very accurate numerical result by using the DODA.

Figure 5. For example 3 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) comparing numerical and exact solutions.

Figure 5. For example 3 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) comparing numerical and exact solutions.

Then under the same parameters as that used in the DODA, we also apply the GMRES,[Citation17] the RRGMRES,[Citation40] the OMVIA [Citation24] and the OTVIA [Citation13] to solve this problem. The residuals and numerical errors are compared in Figure . The OMVIA is convergent with 142 steps and the maximum error is 0.041. The GMRES is convergent with 16 steps and the maximum error is 0.148. The RRGMRES is convergent with four steps and the maximum error is 0.0124. The OTVIA does not converge within 1000 steps and the maximum error is 0.0994. Then we apply the DOS with m=16 to solve this problem, whose numerical error as shown in Figure (b) with the maximum error being 0.0194 is better than that of GMRES, OMVIA and OTVIA, but is worse than that of DODA. It can be seen that the DODA is the best one among the six numerical methods, which is convergent faster and accurate than other five algorithms DOS, GMRES, RRGMRES, OMVIA and OTVIA. The RMSE of GMRES is 0.104, the RMSE of DOS is 0.0136, the RMSE of OMVIA is 0.0282 and the RMSE of OTVIA is 0.0622. The RMSE of DODA is 6.1×10-3, while that for the RRGMRES is 7.9×10-3. For this ill-posed problem, the DODA is better than the RRGMRES.

Figure 6. For example 3 solved by the DODA, DOS, GMRES, RRGMRES, OMVIA and OTVIA, comparing (a) residuals and (b) numerical errors.

Figure 6. For example 3 solved by the DODA, DOS, GMRES, RRGMRES, OMVIA and OTVIA, comparing (a) residuals and (b) numerical errors.

Figure 7. For example 4 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) comparing numerical and exact solutions.

Figure 7. For example 4 solved by the DODA with Krylov subspace, showing (a) residual, (b) a0 and β and (c) comparing numerical and exact solutions.

Figure 8. For example 4 solved by the DODA, DOS, GMRES, RRGMRES, OMVIA and OTVIA, comparing (a) residuals and (b) numerical errors.

Figure 8. For example 4 solved by the DODA, DOS, GMRES, RRGMRES, OMVIA and OTVIA, comparing (a) residuals and (b) numerical errors.

Figure 9. For example 5 solved by the DODA and GMRES, showing (a) residuals, (b) a0 and β, (c) comparing numerical and exact solutions and (d) numerical errors.

Figure 9. For example 5 solved by the DODA and GMRES, showing (a) residuals, (b) a0 and β, (c) comparing numerical and exact solutions and (d) numerical errors.

6.4 Example 4

Let us consider the inverse Cauchy problem for the Laplace equation:117 Δu=urr+1rur+1r2uθθ=0,117 118 u(ρ,θ)=h(θ),0θπ,118 119 un(ρ,θ)=g(θ),0θπ,119 where h(θ) and g(θ) are given function. The inverse Cauchy problem is specified as follows:

To seek an unknown boundary function f(θ) on the part Γ2:={(r,θ)|r=ρ(θ),π<θ<2π} of the boundary under Equations (Equation117)–(Equation119) with the overspecified data being given on Γ1:={(r,θ)|r=ρ(θ),0θπ}.

It is well known that the MFS can be used to solve the Laplace equation when a fundamental solution is known. In the MFS, the solution of u at the field point x=(rcosθ,rsinθ) can be expressed as a linear combination of fundamental solutions U(x,sj):120 u(x)=j=1ncjU(x,sj),sjΩc.120 For the Laplace equation (Equation117), we have the fundamental solutions:121 U(x,sj)=lnrj,rj=x-sj.121 In the practical application of MFS, by imposing the boundary conditions (Equation118) and (Equation119) at N points on Equation (Equation120), we can obtain a linear equations system:122 Bc=b,122 where123 xi=(xi1,xi2)=(ρ(θi)cosθi,ρ(θi)sinθi),sj=(sj1,sj2)=(R(θj)cosθj,R(θj)sinθj),Bij=lnxi-sj,ifiis odd,Bij=η(θi)xi-sj2×ρ(θi)-sj1cosθi-sj2sinθi-ρ(θi)ρ(θi)[sj1sinθi-sj2cosθi],ifiis even,c=(c1,,cn)T,b=(h(θ1),g(θ1),,h(θN),g(θN))T,123 in which n=2N, and124 η(θ)=ρ(θ)ρ2(θ)+[ρ(θ)]2.124 The above R(θ)=ρ(θ)+D with an offset D can be used to locate the source points along a contour with a radius R(θ).

For the purpose of comparison, we consider the following exact solution:125 u(x,y)=cosxcoshy+sinxsinhy,125 defined in a domain with a complex amoeba-like irregular shape as a boundary:126 ρ(θ)=exp(sinθ)sin2(2θ)+exp(cosθ)cos2(2θ).126 After imposing the boundary conditions (Equation118) and (Equation119) at N points on Equation (Equation120), we can obtain a linear equations system. The noise being imposed on the measured data h and g is σ=0.01.

We solve this problem by the DODA with m=5 and γ=0.2. The descent direction u is solved from BTBu=BTr. Through 700 steps the residual and the values of a0 and β are shown in Figure (a) and (b). The numerical solution and exact solution are compared in Figure (c), whose maximum error is smaller than 0.0884. It can be seen that the DODA can accurately recover the unknown boundary condition.

Then under the same parameters as that used in the DODA, we apply the GMRES,[Citation17] the RRGMRES,[Citation40] and the OMVIA [Citation24] to solve this problem. The residuals and numerical errors are compared in Figure . When the maximum error of GMRES is 5.02 (failure) and the maximum error of RRGMRES is 1.16, the maximum error of OMVIA is 0.197. The maximum error of DODA as shown in the above is much better than that calculated by Liu [Citation13] by using the OTVIA, of which the maximum error is 0.34. Then we apply the DOS with m=10 to solve this problem, whose numerical error as shown in Figure (b) with the maximum error being 0.955 is better than that of GMRES, but is worse than that of RRGMRES, OMVIA, OTVIA and DODA. The RMSE of GMRES is 3.03, the RMSE of RRGMRES is 0.553, the RMSE of DOS is 0.602, the RMSE of OMVIA is 0.127 and the RMSE of OTVIA is 0.177. The RMSE of DODA is 0.045, which is better than other five algorithms. Again, it can be seen that the DODA is the best one among the six numerical methods, which is convergent faster and much more accurate than other five algorithms.

6.5 Example 5

Let us consider the following inverse problem to recover the external force F(t) for127 y¨(t)+y˙(t)+y(t)=F(t).127 In a time interval of t[0,tf], the discretized data yi=y(ti) are supposed to be measurable, which are subjected to the random noise with an intensity σ=0.01. Usually, it is very difficult to recover the external force F(ti) from Equation (Equation127) by the direct differentials of the noisy data of displacements, because the differential is an ill-posed linear operator.

To approach this inverse problem by the polynomial interpolation, we begin with128 pm(x)=c0+k=1mckxk.128 Now, the coefficient ck is split into two coefficients ak and bk to absorb more interpolation points; in the meanwhile, cos(kθk) and sin(kθk) are introduced to reduce the condition number of the coefficient matrix. We suppose that129 ck=akcos(kθk)R2kk+bksin(kθk)R2k+1k,129 and130 θk=2kπm,k=1,,m.130 The problem domain is [a,b], and the interpolating points are:131 a=x0<x1<x2<<x2m-1<x2m=b.131 Substituting Equation (Equation129) into Equation (Equation128), we can obtain132 p(x)=a0+k=1makxR2kkcos(kθk)+bkxR2k+1ksin(kθk),132 where we let c0=a0. Here, ak and bk are unknown coefficients. In order to obtain them, we impose the following n interpolated conditions:133 p(xi)=yi,i=0,,n-1.133 Thus, we obtain a linear equations system to determine ak and bk:134 1x0cosθ1R2x0sinθ1R3x0R2mmcosmθmx0R2m+1msinmθm[1ex]1x1cosθ1R2x1sinθ1R3x1R2mmcosmθmx1R2m+1msinmθm[1ex]1x2m-1cosθ1R2x2m-1sinθ1R3x2m-1R2mmcosmθmx2m-1R2m+1msinmθm[1ex]1x2mcosθ1R2x2msinθ1R3x2mR2mmcosmθmx2mR2m+1msinmθma0a1b1ambm=y0y1y2y2m-1y2m.134 We note that the norm of the first column of the above coefficient matrix is 2m+1. According to the concept of equilibrated matrix,[Citation50] we can derive the optimal scales for the current interpolation with a half-order technique as135 R2k=β012m+1j=02mxj2k(coskθk)21/(2k),k=1,2,,m,135 136 R2k+1=β012m+1j=02mxj2k(sinkθk)21/(2k),k=1,2,,m,136 where β0 is a scaling factor.[Citation50] The improved method uses m-order polynomial to interpolate n=2m+1 data nodes, while regular method with a full-order can only interpolate m+1 data points.

Now we fix m0=25 (of the above m) and tf=5 and consider the exact solution to be F(t)=ωcos(ωt)+(1-ω2)sin(ωt), which is obtained by inserting the exact value of y(t)=sin(ωt) into Equation (Equation127). The parameters used are ω=0.5, and β0=2. When we use the DODA with m=4, we let it run 300 steps. The descent direction u is solved from BTBu=BTr. The residual and the values of a0 and β are shown in Figure (a) and (b). The numerical solution and exact solution are compared in Figure (c), whose maximum error is smaller than 0.027. It can be seen that the DODA can accurately recover the unknown external force.

When we apply the GMRES to solve this problem, the parameter β0 is changed to β0=6, and the other parameters are unchanged. In Figure , we plot the results obtained by the GMRES with dashed-dotted lines, of which as shown in Figure (c) and (d) we can see that the GMRES is less accurate with the maximum error being 0.35. The RMSE obtained by the DODA is 1.14×10-2, while that obtained by the GMRES is 0.117. It can be seen that the DODA is much accurate than the GMRES.

7 Conclusions and discussion

In the present paper, we have derived a double optimal algorithm, including an m-vector optimal search direction in an m-dimensional affine Krylov subspace to solve a highly ill-posed linear system. This algorithm is a multi-vector DODA, of which the expansion coefficients in the descent direction are solved in closed-form through double optimizations of two basic merit functions to measure the distance between v=Bu and r. The DODA has a good computational efficiency and accuracy in solving the ill-posed linear equations system. Numerical tests on the linear inverse problems have confirmed the robustness of the DODA against noisy disturbance even with an intensity being large up to 10%. For the test examples, we found that the DODA can converge fast and stable with the iterations account smaller than the dimension of the considered problem. As the usual Krylov subspace method, the value of m cannot be too large; otherwise, many computational time is required to construct the Krylov matrix by the Arnoldi process at each iteration step. When m is large, the Krylov matrix will be highly ill conditioned for the ill-posed problem. The DODA is convergent faster than other numerical methods investigated in this paper, and indeed it can achieve smaller residual errors than the GMRES, the RRGMRES and the OMVIA under the same value of m. In the proof of the convergence of the DODA, we have derived an exact equation to estimate the difference of two consecutive square residual norms:rk+12=rk2-rkTErk+[rkT(In-E)Brk]2rkTBT(In-E)Brk.How to maximize the quantity in the square brackets, i.e.maxrkTErk+[rkT(In-E)Brk]2rkTBT(In-E)Brkmight be an important issue, which will lead to a further study of the structure of the Krylov subspace resulting to the best configuration of the projection operator E. On the other hand, based on Theorem 5 in [Citation20], the square residual obtained by the DODA is smaller than that obtained by the algorithms based on the least square of residual with the following relation:rkDODA2=rkLS2-[rkT(In-E)Brk]2rkTBT(In-E)Brk.This fact revealed that the DODA is more efficient than other algorithms, and all numerical examples investigated in this paper by assessing the convergence speed, the maximum error, the absolute error and the RMSE confirmed the superiority of DODA.

Acknowledgments

The author highly appreciates the constructive comments from anonymous referees, which improve the quality of this paper. Highly appreciated are the project NSC-102-2221-E-002-125-MY3 and the 2011 Outstanding Research Award from the National Science Council of Taiwan, and the 2011 Taiwan Research Front Award from Thomson Reuters. It is also acknowledged that the author has been promoted as being a Lifetime Distinguished Professor of National Taiwan University since 2013.

References

  • Stewart G. Introduction to matrix computations. New York (NY): Academic Press; 1973.
  • Kunisch K, Zou J. Iterative choices of regularization parameters in linear inverse problems. Inverse Probl. 1998;14:1247–1264.
  • Wang Y, Xiao T. Fast realization algorithms for determining regularization parameters in linear inverse problems. Inverse Probl. 2001;17:281–291.
  • Xie J, Zou J. An improved model function method for choosing regularization parameters in linear inverse problems. Inverse Probl. 2002;18:631–643.
  • Resmerita E. Regularization of ill-posed problems in Banach spaces: convergence rates. Inverse Probl. 2005;21:1303–1314.
  • Chehab JP, Laminie J. Differential equations and solution of linear systems. Numer. Algorithms. 2005;40:103–124.
  • Helmke U, Moore JB. Optimization and dynamical systems. Berlin: Springer; 1994.
  • Liu CS. Optimally generalized regularization methods for solving linear inverse problems. CMC: Comput. Mater. Con. 2012;29:103–127.
  • Liu CS. Scaled vector regularization method to solve ill-posed linear problems. Appl. Math. Comput. 2012;218:10602–10616.
  • Liu CS. A dynamical Tikhonov regularization for solving ill-posed linear algebraic systems. Acta Appl. Math. 2012;123:285–307.
  • Liu CS. An optimal preconditioner with an alternate relaxation parameter used to solve ill-posed linear problems. CMES: Comput. Model. Eng. Sci. 2013;92:241–269.
  • Liu CS. A globally optimal iterative algorithm to solve an ill-posed linear system. CMES: Comput. Model. Eng. Sci. 2012;84:383–403.
  • Liu CS. An optimal tri-vector iterative algorithm for solving ill-posed linear inverse problems. Inverse Probl. Sci. Eng. 2013;21:650–681.
  • Liu CS. A globally optimal tri-vector method to solve an ill-posed linear system. J. Comput. Appl. Math. 2014;260:18–35.
  • Barzilai J, Borwein JM. Two point step size gradient methods. IMA J. Numer. Anal. 1988;8:141–148.
  • Akaike H. On a successive transformation of probability distribution and its application to the analysis of the optimum gradient method. Ann. Inst. Stat. Math. Tokyo. 1959;11:1–16.
  • Saad Y, Schultz MH. GMRES: a generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM J. Sci. Stat. Comput. 1986;7:856–869.
  • Liu CS, Atluri SN. An iterative method using an optimal descent vector, for solving an ill-conditioned system Bx = b, better and faster than the conjugate gradient method. CMES: Comput. Model. Eng. Sci. 2011;80:275–298.
  • Liu CS. A doubly optimized solution of linear equations system expressed in an affine Krylov subspace. J. Comput. Appl. Math. 2014;260:375–394.
  • Liu CS. Discussing a more fundamental concept than the minimal residual method for solving linear system in a Krylov subspace. J. Math. Res. 2013;5:58–70.
  • Liu CS. An optimally generalized steepest-descent algorithm for solving ill-posed linear systems. J. Appl. Math. 2013; ID 154358, 15 p.
  • Jacoby SLS, Kowalik JS, Pizzo JT. Iterative methods for nonlinear optimization problems. New Jersey (NJ): Prentice-Hall; 1972.
  • Ostrowski AM. Solution of equations in Euclidean and Banach spaces. 3rd ed. New York (NY): Academic Press; 1973.
  • Liu CS. An optimal multi-vector iterative algorithm in a Krylov subspace for solving the ill-posed linear inverse problems. CMC: Comput. Mater. Con. 2013;33:175–198.
  • Dongarra J, Sullivan F. Guest editors’ introduction to the top 10 algorithms. Comput. Sci. Eng. 2000;2:22–23.
  • Simoncini V, Szyld DB. Recent computational developments in Krylov subspace methods for linear systems. Numer. Linear Algebra Appl. 2007;14:1–59.
  • Saad Y. Krylov subspace methods for solving large unsymmetric linear systems. Math. Comput. 1981;37:105–126.
  • Freund RW, Nachtigal NM. QMR: a quasi-minimal residual method for non-Hermitian linear systems. Numer. Math. 1991;60:315–339.
  • van Den Eshof J, Sleijpen GLG. Inexact Krylov subspace methods for linear systems. SIAM J. Matrix Anal. Appl. 2004;26:125–153.
  • Hestenes MR, Stiefel EL. Methods of conjugate gradients for solving linear systems. J. Res. Nat. Bur. Stand. 1952;49:409–436.
  • Lanczos C. Solution of systems of linear equations by minimized iterations. J. Res. Nat. Bur. Stand. 1952;49:33–53.
  • Paige CC, Saunders MA. Solution of sparse indefinite systems of linear equations. SIAM J. Numer. Anal. 1975;12:617–629.
  • Fletcher R. Conjugate gradient methods for indefinite systems. Lecture notes in Math. Vol. 506. Berlin: Springer-Verlag; 1976. p. 73–89.
  • Sonneveld P. CGS: a fast Lanczos-type solver for nonsymmetric linear systems. SIAM J. Sci. Stat. Comput. 1989;10:36–52.
  • van der Vorst HA. Bi-CGSTAB: a fast and smoothly converging variant of Bi-CG for the solution of nonsymmetric linear systems. SIAM J. Sci. Stat. Comput. 1992;13:631–644.
  • Saad Y, van der Vorst HA. Iterative solution of linear systems in the 20th century. J. Comput. Appl. Math. 2000;123:1–33.
  • Saad Y. Iterative methods for sparse linear systems. 2nd ed. Pennsylvania (PA): SIAM; 2003.
  • van der Vorst HA. Iterative Krylov methods for large linear systems. New York (NY): Cambridge University Press; 2003.
  • Liu CS. The concept of best vector used to solve ill-posed linear inverse problems. CMES: Comput. Model. Eng. Sci. 2012;83:499–525.
  • Calvetti C, Lewis B, Reichel L. GMRES-type methods for inconsistent systems. Linear Algebra Appl. 2000;316:157–169.
  • Kuroiwa N, Nodera T. A note on the GMRES method for linear discrete ill-posed problems. Adv. Appl. Math. Mech. 2009;1:816–829.
  • Matinfar M, Zareamoghaddam H, Eslami M, Saeidy M. GMRES implementations and residual smoothing techniques for solving ill-posed linear systems. Comput. Math. Appl. 2012;63:1–13.
  • Morikuni K, Reichel L, Hayami K. FGMRES for linear discrete ill-posed problems. Appl. Numer. Math. 2014;75:175–187.
  • Yin JF, Hayami K. Preconditioned GMRES methods with incomplete Givens orthogonalization method for large sparse least-squares problems. J. Comput. Appl. Math. 2009;226:177–186.
  • Horn RA, Johnson CR. Matrix analysis. New York (NY): Cambridge University Press; 1985.
  • Liu CS, Hong HK, Atluri SN. Novel algorithms based on the conjugate gradient method for inverting ill-conditioned matrices, and a new regularization method to solve ill-posed linear systems. CMES: Comput. Model. Eng. Sci. 2010;60:279–308.
  • Liu CS. A vector regularization method to solve linear inverse problems. Inverse Probl. Sci. Eng. 2013. dx.doi.org/10.1080/17415977.2013.823415.
  • Hon YC, Li M. A discrepancy principle for the source points location in using the MFS for solving the BHCP. Int. J. Comput. Methods. 2009;6:181–197.
  • Liu CS. The method of fundamental solutions for solving the backward heat conduction problem with conditioning by a new post-conditioner. Numer. Heat Transfer B: Fundam. 2011;60:57–72.
  • Liu CS. A two-side equilibration method to reduce the condition number of an ill-posed linear system. CMES: Comput. Model. Eng. Sci. 2013;91:17–42.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.