268
Views
2
CrossRef citations to date
0
Altmetric
Articles

A vector regularization method to solve linear inverse problems

Pages 765-786 | Received 27 Dec 2012, Accepted 14 Jun 2013, Published online: 01 Aug 2013

Abstract

The linear inverse problem is discretized to be an n-dimensional ill-posed linear equations system Bx=b. In the present paper, an invariant manifold defined in terms of the square norm of a residual vector r:=Bxb is used to derive an iterative algorithm with a fast descent direction Ar, which is close to, but not exactly equal to, the best descent direction B1r. The matrix A is obtained by using a vector regularization method together with a matrix conjugate gradient method to find the right inversion of B: BA=In. The vector regularization iterative algorithm is proven to be Lyapunov stable, and the direct inversion method with solution expressed by x=Ab converges fast. The accuracy and efficiency of them are verified through the numerical tests of linear inverse problems under a large random noise.

Introduction

It is known that an iterative method for solving the system of algebraic equations can be derived from the discretization of a certain ordinary differential equations (ODEs) system.[Citation1, Citation2] In particular, the descent methods can be interpreted as the discretizations of gradient flows.[Citation3] Indeed, it has a long time history that continuous algorithms have been investigated in many literature, for example, Gavurin [Citation4], Alber [Citation5] and Hirsch and Smale [Citation6]. Chu [Citation7] has developed a systematic approach to the continuous realization of several iterative algorithms in numerical linear algebra. The Lyapunov methods used in the analysis of iterative methods have been made by Ortega and Rheinboldt [Citation8], and Bhaya and Kaszkurewicz [Citation1, Citation9, Citation10].

The author and his co-workers have developed several methods to solve ill-posed system of linear equations: the fictitious time integration method as a filter,[Citation11] a modified polynomial expansion method, [Citation12] the Laplacian preconditioners and postconditioners,[Citation13] a vector regularization method,[Citation14] a relaxed steepest descent method,[Citation15, Citation16] an optimal iterative algorithm with an optimal descent vector,[Citation17] an adaptive Tikhonov regularization method,[Citation18] the best vector iterative method,[Citation19] the globally optimal vector iterative method,[Citation20] the optimally scaled vector regularization method,[Citation21] an optimally generalized Tikhonov regularization method,[Citation22] as well as an optimal tri-vector iterative algorithm.[Citation23] As a continuation of these works, in this paper, we propose a more simple and robust method to solve(1) Bx=b,(1) where xRn is an unknown vector determined from a given coefficient matrix BRn×n, which might be unsymmetric, and the input bRn which might be polluted by random noise. Many linear-type inverse problems can be discretized into the above form.

A measure of the ill-posedness of Equation (Equation1) is the condition number of B:(2) Cond(B)=BFB1F,(2) where BF is the Frobenius norm of B.

For every matrix norm , we have ρ(B)B, where ρ(B) is a radius of the spectrum of B. The Householder theorem states that for every ϵ>0 and every matrix B, there exists a matrix norm B depending on B and ϵ such that Bρ(B)+ϵ. So far, the spectral condition number ρ(B)ρ(B1) can be used as an estimation of the condition number of B by(3) Cond(B)=maxσ(B)|λ|minσ(B)|λ|,(3) where σ(B) is the collection of all the eigenvalues of B. If the Frobenius norm in Equation (Equation2) is replaced by the operator norm, then Equations (Equation2) and (Equation3) lead to the same condition number. Turning back to the Frobenius norm, we have(4) BFnmaxσ(B)|λ|.(4) In particular, for the symmetric case ρ(B)ρ(B1)=B2B12. Roughly speaking, the numerical solution of Equation (Equation1) may lose the accuracy of k decimal points when Cond(B)=10k.

Instead of Equation (Equation1), we can solve a normal linear system:(5) Cx=b1,(5) where(6) b1=BTb,C=BTB>0.(6) We consider an iterative method for solving Equation (Equation5) and define for any vector xk the steepest descent vector(7) Rk:=Cxkb1.(7) Ascher et al. [Citation24], and also Liu and Chang [Citation25] have viewed the gradient descent method:(8) xk+1=xkαkRk,(8) as a forward Euler scheme of the following system of ODEs:(9) x·=b1Cx.(9) The absolute stability bound(10) αk2maxσ(C)λ(10) must be obeyed if a uniform stepsize is employed.

Specifically, Equation (Equation8) presents a steepest descent method (SDM), if(11) αk=Rk2RkTCRk.(11) When Rk is small, the calculated Rk may deviate from the real steepest descent direction to a great extent due to a round-off error of computing machine, which usually leads to the numerical instability of SDM. An improvement of SDM is the conjugate gradient method (CGM), which enhances the search direction of the minimum of a quadratic functional by imposing an orthogonality to the residual vector at each iterative step. The algorithm of CGM for solving Equation (Equation5) is summarized as follows.

  1. Give an initial x0, compute R0=Cx0b1 and then set p0=R0.

  2. For k=0,1,2,, we repeat the following iterations:(12) ηk=Rk2pkTCpk,xk+1=xkηkpk,Rk+1=Cxk+1b1,αk+1=Rk+12Rk2,pk+1=αk+1pk+Rk+1.(12)

If xk+1 converges according to a given stopping criterion Rk+1<ε, then stop; otherwise, go to step (ii).

Even if the CGM works very well for most linear systems, the CGM does lose some of its luster for some linear inverse problems. The gradient descent methods enjoy a fictitious time integration method with different stepsizes. This is particularly useful for image deblurring. The concept of optimal descent vector was first developed by Liu and Atluri [Citation26] for solving non-linear algebraic equations. In this paper, we explore the concept of best descent vector [Citation19] and use it as a guidance to develop a new method to solve linear algebraic equations.

The remaining parts of this paper are arranged as follows. The vector regularization method for inverting non-singular matrix is introduced in Section 2. In Section 3, we start from an invariant manifold to derive a non-linear ODEs system for the numerical solution of Equation (Equation1) with the descent direction to be u=Ar, where A is solved from the vector regularization method as a right inversion of the coefficient matrix B given in Equation (Equation1). Then, a Lyapunov stable dynamics on the invariant manifold is constructed in Section 4, resulting in a vector regularization iterative algorithm (VRIA). The linear inverse problems are solved in Section 5 to display some advantages of the VRIA and the direct inversion method (DIM). Finally, the conclusions are drawn in Section 6.

A vector regularization method

The vector regularization method for inverting non-singular matrices was first developed by Liu et al. [Citation14]. Recently, Liu [Citation21] has developed a systematic method to determine the vector regularization parameter.

Let us begin with the following matrix equation:(13) VTUT=Im,i.e.(UV)T=Im,(13) if U is the inversion of a given non-singular m×m matrix V, where Im is the m×m identity matrix. Numerically, we can say that the above U is a left inversion of V due to UV=Im. Then, multiplying the above equation by V from the left side we have(14) DUT=V,D:=VVT,(14) from which we can solve for UT:=C by the following matrix conjugate gradient method (MCGM):

  1. Assume an initial C0.

  2. Calculate R0=VDC0 and P1=R0.

  3. For k=1,2,, we repeat the following iterations:(15) αk=Rk12Pk·(DPk),Ck=Ck1+αkPk,Rk=VDCk,ηk=Rk2Rk12,Pk+1=Rk+ηkPk.(15) If Ck converges according to a given stopping criterion:(16) Rk<ε2,(16) then stop; otherwise, go to step (iii). In the above, the capital boldfaced letters denote m×m matrices, the norm Rk is the Frobenius norm (similar to the Euclidean norm for a vector), and the inner product is for matrices. When C is calculated, the left inversion of V is given by U=CT.

The above MCGM is suitable to find the inversion of a well-conditioned matrix; however, for a given ill-conditioned matrix with a large condition number, we need more study to find its inversion. Let(17) Vx0=y0,(17) through which, given x0, say x0=1=[1,,1]T, we can calculate y0 readily, because V is a given matrix. Hence, multiplying the above equation by U from the left side and using Equation (Equation13) we have(18) y0TUT=x0T,i.e.x0=Uy0.(18) Together, Equations (Equation13) and (Equation22) constitute an over-determined system to calculate UT. This over-determined system can be written as(19) WUT=[Imx0T],(19) where(20) W:=[VTy0T](20) is an n×m matrix with n=m+1. Multiplying Equation (Equation23) by WT, we obtain an m×m matrix equation again:(21) [VVT+y0y0T]UT=V+y0x0T,(21) which, similar to Equation (Equation14), is solved by the MCGM with D=VVT+y0y0T and V being replaced by V+y0x0T. This algorithm for solving the left inversion U of an ill-conditioned matrix V has been labelled as the MCGML method.

The above algorithm is suitable for finding the left inversion of V, i.e. UV=Im; however, we need to solve(22) VU=Im,(22) when we want U to be a right inversion of V. Mathematically, the left inversion is equal to the right inversion. But, numerically, they are hardly being equal, especially for ill-conditioned matrices.

For the right inversion, we can supplement, as in Equation (Equation21), another equation:(23) y0TU=x0T,i.e.y0=VTx0.(23) Then, the combination of Equations (Equation26) and (Equation27) leads to the following over-determined system:(24) [Vy0T]U=[Imx0T].(24) Multiplying the transpose of the leading matrix on both sides, we can obtain an m×m matrix equation:(25) [VTV+y0y0T]U=VT+y0x0T,(25) which is then solved by the following MCGMR:

  1. Let D=VTV+y0y0T, and V1=VT+y0x0T.

  2. Assume an initial value U0.

  3. Calculate R0=V1DU0, and P0=R0.

  4. For k=1,2,, we repeat the following iterations:(26) αk=Rk12Pk·(DPk),Uk=Uk1+αkPk,Rk=V1DUk,ηk=Rk2Rk12,Pk+1=Rk+ηkPk.(26) If Uk converges according to a given stopping criterion,(27) Rk<ε2,(27) then stop; otherwise, go to step (iv).

Equations (Equation25) and (Equation29) are two regularized and non-perturbed algebraic equations to find, respectively, the left and right inversion of a given ill-conditioned matrix V. Due to the appearance of y0, we have a chance to reduce the condition numbers of the coefficient matrices of these two equations; however, we refer the paper [Citation21] for a detailed description of how to find the best value of y0.

Invariant manifold method

The most simple way for solving the linear equations system (Equation1) is(28) x=B1b,(28) where we can apply the MCGML to find B1 from the given matrix B. This inversion method is usually quite dangerous and unstable when there exists a noise on b. The error of noise will be amplified by B1 to cause a large error of x, when the singular values of B are clustered near to zero value. In Section 5.1, we will give a numerical example to reveal this type phenomenon.

Instead of Equation (32), we search an iterative method based on the idea of MCGMR. For the linear equations system (Equation1), which is expressed to be r=0 in terms of the residual vector:(29) r=Bxb,(29) we can introduce a scalar homotopy function:(30) h(x,t)=Q(t)2r(x)212r(x0)2=0.(30) We expect h(x,t)=0 to be an invariant manifold in the space-time domain (x,t) for a dynamical system h(x(t),t)=0 to be specified further. When Q>0, the manifold defined by Equation (Equation34) is continuous and differentiable, and thus the following differential operation makes sense:(31) 12Q·(t)r(x)2+Q(t)r·(Bx·)=0,(31) which is obtained by taking the time differential of Equation (Equation34) with respect to t, considering x=x(t) and using Equation (Equation33).

We suppose that the evolution of x is governed by a vector u:(32) x·=λu,(32) where(33) u=Ar.(33) We hope Ar is near to B1r for a fast convergence (see Section 4.2), i.e.(34) Ar=B1r.(34) Then, we can apply the MCGMR to find the right inversion A of B by(35) BA=In.(35) Inserting Equation (Equation36) into Equation (Equation35) we can derive(36) x·=q(t)r2rTvu,(36) where(37) v:=Bu=BAr,q(t):=Q·(t)2Q(t).(37) Hence, in our algorithm if Q(t) can be guaranteed to be a monotonically increasing function of t, we have an absolutely convergent property in solving the linear equations system (Equation1) by decreasing r2 according to(38) r(x)2=CQ(t),(38) where(39) C=r(x0)2(39) is determined by the initial value x0.

Dynamics on the invariant manifold

In order to keep x on the manifold (Equation43), we can consider the evolution of r along the path x(t) by(40) r·=Bx·=q(t)r2rTvv.(40) Because of q(t)>0 we can introduce a new independent variable τ=0tq(ξ)dξ, such that the above equation can be written as(41) drdτ=r2rTBArBAr.(41) where we have inserted Equation (Equation41) for v. It is an autonomous non-linear dynamical system for r.

First, we note that r=0 is an equilibrium point of Equation (Equation46). It is known that the Lyapunov’s second, or direct, method provides a stronger stability condition of the equilibrium point of a nonlinear dynamical system.[Citation27] To describe the Lyapunov theorem we need to define the positive-definite function. A real scalar function V(r) is defined to be positive-definite in some closed bounded region D of state space if for all r in D,

  1. V(r) is continuously differentiable with respect to r,

  2. V(0)=0, and

  3. V(r)>0, for all r0.

Then V(r) is a Lyapunov function in a neighbourhood D of the origin if
  1. V(r) is positive-definite, and

  2. dV(r)/dτ is negative-semidefinite for all rD.

The Lyapunov theorem of stability says that the origin is stable if there exists a Lyapunov function V(r) throughout D; and that the origin is asymptotically stable if there exists a Lyapunov function V(r) throughout D, such that dV(r)/dτ is negative-definite for all rD/0.

By using the Lyapunov function V=r2/2 and from Equation (Equation45), it is easy to prove that(42) V(0)=0,V(r)=r22>0,r0,V·=r·r·=q(t)r2<0,r0.(42) In Section 4.2, we will prove η=qΔt>0. Then, by using the Lyapunov stability theory, the ODEs in Equation (Equation40) are asymptotically stable to r=0, and the iterative algorithm derived from it with a suitable selection of q>0 is stable.

Discretizing, yet keeping x on the manifold

Now we discretize the foregoing continuous time dynamics (Equation40) into a discrete time dynamics by applying the forward Euler scheme:(43) x(t+Δt)=x(t)ηr2rTvu,(43) where(44) η=q(t)Δt(44) is a steplength.

Similarly, we use the forward Euler scheme to integrate Equation (Equation45):(45) r(t+Δt)=r(t)ηr2rTvv,(45) and taking the square norms of both the sides and using Equation (Equation43), we can obtain(46) CQ(t+Δt)=CQ(t)2ηCQ(t)+η2CQ(t)r2(rTv)2v2.(46) Thus, the following scalar equation is derived:(47) a0η22η+1Q(t)Q(t+Δt)=0,(47) where(48) a0:=r2v2(rTv)21(48) by using the Cauchy–Schwarz inequality:rTvrv.As a result, h(x,t)=0 remains to be an invariant manifold in the space time of (x,t) for the discrete time dynamical system h(x(t),t)=0, t{0,1,2,}, which will be explored further in the next section.

An iterative dynamics

Let(49) s:=Q(t)Q(t+Δt)=r(x(t+Δt))2r(x(t))2,(49) which is an important quantity to assess the convergent property of our numerical algorithm for solving linear equations system (Equation1).

From Equations (Equation53) and (Equation55), it follows that(50) a0η22η+1s=0,(50) of which we can take a preferred solution of η to be(51) η=11(1s)a0a0,if1(1s)a00.(51) Let(52) 1(1s)a0=γ20,s=11γ2a0,(52) and the condition 1(1s)a00 in Equation (Equation57) is automatically satisfied; hence, from Equations (Equation57) and (Equation54), it follows that(53) η=1γa0>0,(53) where(54) 0γ<1(54) is a parameter. Finally, from Equations (Equation49), (Equation54) and (Equation60), we can obtain the following algorithm:(55) x(t+Δt)=x(t)(1γ)rTvv2u.(55) Under conditions (Equation54) and (Equation61), from Equations (Equation55) and (Equation59), we can prove that the new algorithm satisfies(56) r(t+Δt)r(t)=s<1,(56) which means that the residual error is absolutely decreased. In other words, the convergence rate of present iterative algorithm is greater than one:(57) ConvergenceRate:=r(t)r(t+Δt)=1s>1.(57) The property in Equation (Equation64) is very important, since it guarantees that the new algorithm is absolutely convergent to the true solution.

From Equations (Equation48), (Equation50) and (Equation60), we can derive(58) ΔV=1γa0r2.(58) By using the Lyapunov stability, the best choice of a0 is the one that makes ΔV as negative as possible and which can be achieved by rendering a01 as small as possible. In principle, if we can use u=B1r as a search direction, by Equations (Equation41) and (Equation54) we have a0=1 and with ΔV=(1γ)r2 being the minimum of ΔV. This is, however, impossible because the exact inverse B1 is hard to be found, and so instead, we use u=Ar as a search direction and solve A to near B1 by using the MCGMR in Section 2. In doing so, we have a fast convergence iterative algorithm to solve the ill-posed linear problem (Equation1) on one hand, and on the other hand, because A is not exactly equal to B1, we can avoid the ill-posedness to magnify the noisy error by Ar. Indeed, the present iterative algorithm is a trade-off between accuracy and regularization. We need to point out that for an ill-posed linear system (Equation1), the solution given by B1b is unstable. Our algorithms are finding A, not finding B1, which is less ill-conditioned than the finding of B1. The descent direction detected by Ar is not exactly equal to the best descent direction B1r, which is an approximation. Because we have taken the direction B1r into account, the iterative algorithm based on Ar can converge quite fast.

vector regularization iterative algorithm

Since the fictitious time variable is now discrete, t{0,1,2,}, we let xk denote the numerical value of x at the k-th step. Thus, we arrive at a purely iterative algorithm from Equation (Equation62):(59) xk+1=xk(1γ)rkTvkvk2uk.(59) Then, the following VRIA is available:

  1. For a given B, find A by using the MCGMR.

  2. Select γ, and give an initial value of x0 or x0=Ab.

  3. For k=0,1,2,, we repeat the following iterations:(60) rk=Bxkb,uk=Ark,vk=BArk,xk+1=xk(1γ)rk·vkvk2uk.(60) If xk+1 converges according to a given stopping criterion rk+1<ε1, then stop; otherwise, go to step (iii). Here, γ is a relaxation parameter, chosen by the user.

Linear inverse problems

In this section, we will apply the method of fundamental solutions (MFS) to discretize some inverse problems into a set of linear algebraic equations. In recent years, a few different meshless boundary collocation methods, like the singular boundary method, have been proposed and developed which are different but highly related to the MFS examined in this paper.[Citation28, Citation29] The regularization method proposed in this paper can also be an efficient alternative for these methods for solving the ill-posed linear inverse problems.[Citation30, Citation31]

Example 1

When the backward heat conduction problem (BHCP) is considered in a spatial interval of 0<x< by subjecting to the boundary conditions at two ends of a slab:(61) ut(x,t)=αuxx(x,t),0<t<T,0<x<,u(0,t)=u0(t),u(,t)=u(t),(61) we solve u under a final time condition:(62) u(x,T)=uT(x).(62) The fundamental solution of Equation (Equation68) is given as follows:(63) K(x,t)=H(t)2απtexp(x24αt),(63) where H(t) is the Heaviside function.

The method of fundamental solutions (MFSs) has a broad application in engineering computations. However, the MFS has a serious drawback in that the resulting system of linear equations is always highly ill-conditioned. In the MFS, the solution of u at the field point z=(x,t) can be expressed as a linear combination of the fundamental solutions U(z,sj):(64) u(z)=j=1NcjU(z,sj),sj=(ηj,τj)Ωc,(64) where N is the number of source points, cj are unknown coefficients and sj are source points located in the complement Ωc of Ω=[0,]×[0,T]. For the heat conduction equation, we have the basis functions(65) U(z,sj)=K(xηj,tτj).(65) It is known that the location of source points in the MFS has a great influence on the accuracy and stability. In a practical application of MFS to solve the BHCP, the source points are uniformly located on two straight lines parallel to the t-axis and not over t=T, which was adopted by Hon and Li [Citation32] and Liu [Citation33], showing a large improvement than the line location of source points below the initial time. After imposing the boundary conditions and the final time condition on Equation (Equation72), we can obtain a linear equations system:(66) Bx=b,(66) where(67) Bij=U(zi,sj),x=(c1,,cN)T,b=(u(ti),i=1,,m1;uT(xj),j=1,,m2;u0(tk),k=m1,,1)T.(67) The number n=2m1+m2 of collocation points does not necessarily equal to the number N of source points.

Since the BHCP is highly ill-posed,[Citation34] the ill-conditioning of the coefficient matrix B in Equation (Equation74) is serious. To overcome the ill-posedness of Equation (Equation74), we can employ both the DIM and VRIA to solve this problem. Here, we compare the numerical solution with an exact solution:u(x,t)=cos(πx)exp(π2t).For T=1 the value of final data is in the order of 104, which is small in comparison with the value of the initial temperature u0(x)=cos(πx) to be retrieved, which is O(1).

In order to test the stability of the proposed algorithm, we add a random noise on the final time data by(68) bi=u(xi,T)+σR(i),(68) where σ denotes the level of noise and R(i) are random numbers between [1,1].

Fig. 1 For example 1 of a BHCP under an adding noise 0.001, comparing the exact solution with the numerical solutions obtained by the inversions from MCGML and MCGMR.

Fig. 1 For example 1 of a BHCP under an adding noise 0.001, comparing the exact solution with the numerical solutions obtained by the inversions from MCGML and MCGMR.

Fig. 2 For example 1 of a BHCP under a large relative noise 0.1, (a) the residual error, (b) the value of a0, and (c) comparing the numerical errors.

Fig. 2 For example 1 of a BHCP under a large relative noise 0.1, (a) the residual error, (b) the value of a0, and (c) comparing the numerical errors.

We take m1=15, m2=10, and hence n=40, and the noise with an intensity σ=103 is adding on the right-hand side. We first apply the MCGML to find the left inversion of B under a convergence criterion ε2=106, which is convergent with 714 iterations to find the inverse matrix B1, such that the solution of x can be written as(69) x=B1b.(69) Unfortunately, the noisy error is largely amplified in the above solution, which causes an incorrect solution as shown in Figure by the red dashed line with the maximum error being 2.34. Conversely, when we apply the MCGMR to find the right inversion A of B, which is convergent with 914 iterations, and give the solution by(70) x=Ab,(70) which is named a DIM in this paper, the numerical solution is close to the exact solution as shown in Figure by the blue dashed-dotted line, with the maximum error being 0.0478.

The MCGML is a feasible algorithm to find the left inversion of an ill-conditioned matrix. When the MCGML can provide a very accurate solution of B1 in Equation (Equation77), the noise is also being magnified, which pollutes the numerical solution of x. Hence, we do not suggest to directly use Equation (Equation77) to find the solution of x, when the data b are noisy and B is ill-conditioned. The reason is that the noise in b would be enlarged when the elements in B1 are quite large.

We take m1=15, m2=10, and hence n=40, and a relative random noise with intensity σr=10% is added in the final time data:(71) bi=u(xi,T)[1+σrR(i)].(71) We first apply the MCGMR to find the right inversion of B by imposing a convergence criterion ε2=103, which is convergent with 74 iterations to find the right inverse matrix A. Then, we use the DIM as shown in Equation (Equation78) to solve this BHCP, of which the maximum error of initial condition is 0.088. Moreover, when the value of ε2 is reduced to ε2=104, the MCGMR is convergent with 128 iterations to find the inverse matrix A, and the DIM used A in Equation (Equation78) leads to the maximum error being 1.085×103, which is shown in Figure . To the best knowledge of the author, this accuracy is never seen before for this highly ill-posed BHCP under a large noise with σr=0.1.

With γ=0.05, we also apply the VRIA in Section 4.3 to solve this problem, where we take ε1=104 and ε2=104 and with Ab as an initial guess. In Figure , we show the residual error obtained by the VRIA, which is convergent only through 5 iterations. By adding 127 for the number of finding A, the total number of iterations for the VRIA is 132, which is larger than 128 iterations for the DIM. The values of a0 as shown in Figure is very small, which causes a fast convergence of the VRIA. It can be seen that the algorithm of VRIA is convergent very fast only through five iterations that its residual error reduces from 2.5 to 4.448×105. We found that more smaller convergence criterion of ε1 would not lead to more accurate solution. The numerical error is compared with that obtained by the DIM in Figure with the maximum error being 0.014, which is greater than that obtained by the DIM. This result is much better than that obtained by Liu [Citation21, Citation22], and also than that obtained by Liu et al. [Citation35].

Example 2

Let us consider the following inverse problem to recover the external force F(t) for an ODE:(72) y¨(t)+y·(t)+y(t)=F(t).(72) In a time interval of t[0,tf], the discretized data yi=y(ti) are supposed to be measurable, which are subjected to the random noise with an intensity σ=0.01. Usually, it is very difficult to recover the external force F(ti) from Equation (Equation80) by the direct differentials of the noisy data of the displacements, because the differential is an ill-posed linear operator.

To approach this inverse problem by the polynomial interpolation, we begin with(73) pm(x)=c0+k=1mckxk.(73) Now, the coefficient ck is split into two coefficients ak and bk to absorb more interpolation points; in the meanwhile, cos(kθk) and sin(kθk) are introduced to reduce the condition number of the coefficient matrix.[Citation36] We suppose that(74) ck=akcos(kθk)R2kk+bksin(kθk)R2k+1k,(74) and(75) θk=2kπm,k=1,,m.(75) The problem domain is [a,b], and the interpolating points are:(76) a=x0<x1<x2<<x2m1<x2m=b.(76) Substituting Equation (Equation82) into Equation (Equation81), we can obtain(77) p(x)=a0+k=1m[ak(xR2k)kcos(kθk)+bk(xR2k+1)ksin(kθk)],(77) where we let c0=a0. Here, ak and bk are unknown coefficients. In order to obtain them, we impose the following n interpolated conditions:(78) p(xi)=yi,i=0,,n1.(78) Thus, we obtain a linear equations system to determine ak and bk:(79) [1x0cosθ1R2x0sinθ1R3(x0R2m)mcosmθm(x0R2m+1)msinmθm1x1cosθ1R2x1sinθ1R3(x1R2m)mcosmθm(x1R2m+1)msinmθm1x2m1cosθ1R2x2m1sinθ1R3(x2m1R2m)mcosmθm(x2m1R2m+1)msinmθm1x2mcosθ1R2x2msinθ1R3(x2mR2m)mcosmθm(x2mR2m+1)msinmθm][a0a1b1ambm]=[y0y1y2y2m1y2m].(79) We note that the norm of the first column of the above coefficient matrix is 2m+1. According to the concept of equilibrated matrix,[Citation37] we can derive the optimal scales for the current interpolation with a half-order technique as(80) R2k=β0(12m+1j=02mxj2k(coskθk)2)1/(2k),k=1,2,,m,R2k+1=β0(12m+1j=02mxj2k(sinkθk)2)1/(2k),k=1,2,,m,(80) where β0 is a scaling factor.[Citation38] The improved method uses m order polynomial to interpolate n=2m+1 data nodes, while regular method with a full order can only interpolate m+1 data points.

Fig. 3 For example 2 under a large noise 0.01, (a) comparing the numerical and exact solutions, and (b) the numerical errors.

Fig. 3 For example 2 under a large noise 0.01, (a) comparing the numerical and exact solutions, and (b) the numerical errors.

Now, we fix m=25 and tf=5 and consider the exact solution to be F(t)=ωcos(ωt)+(1ω2)sin(ωt), which is obtained by inserting the exact value of y(t)=sin(ωt) into Equation (Equation80). The parameters used are ω=0.5, and β0=5. When we use the MCGMR, it is convergent with nine iterations under ε2=104. We compare the numerical solution obtained by the DIM with the data given by F(t)=ωcos(ωt)+(1ω2)sin(ωt) in Figure , of which the maximum error is found to be 0.0196 as shown in Figure by solid line. Then, we apply the VRIA to solve this problem under γ=0.05 and β0=5, where we first use the MCGMR to find A, which is convergent with five iterations under ε2=103. Then, using the initial guess x0=Ab, we let VRIA run 20 for iterations. We compare the numerical solution obtained by the VRIA with the exact data in Figure by dashed-dotted line, of which the maximum error is 0.0106 as shown in Figure by dashed line.

Example 3

We solve the Cauchy problem of the Laplace equation under the incomplete boundary conditions:(81) Δu=urr+1rur+1r2uθθ=0,r<ρ,0θ2π,u(ρ,θ)=h(θ),0θπ,un(ρ,θ)=g(θ),0θπ,(81) where h(θ) and g(θ) are given functions, and ρ=ρ(θ) is a given contour to describe the boundary shape. The contour in the polar coordinates is specified by Γ={(r,θ)|r=ρ(θ),0θ2π}, which is the boundary of the problem domain Ω, and n denotes the outward normal direction. We need to find the boundary data on the lower half contour for the completeness of boundary data.

In the MFS, the trial solution of u at the field point  z=(rcosθ,rsinθ) can be expressed as a linear combination of the fundamental solutions U(z,sj):(82) u(z)=j=1ncjU(z,sj),sjΩc,(82) where n is the number of source points, cj are the unknown coefficients, sj are the source points and Ωc is the complementary set of Ω. For the Laplace Equation (Equation90), we have the fundamental solutions:(83) U(z,sj)=lnrj,rj=zsj.(83) In the practical application of MFS, usually the source points are distributed uniformly on a circle with a radius R, such that after imposing the boundary conditions (Equation91) and (Equation92) on Equation (Equation93), we can obtain a linear equations system:(84) Bx=b,(84) where(85) zi=(zi1,zi2)=(ρ(θi)cosθi,ρ(θi)sinθi),sj=(sj1,sj2)=(Rcosθj,Rsinθj),Bij=lnzisj,ifiisodd,Bij=η(θi)zisj2(ρ(θi)sj1cosθisj2sinθiρ(θi)ρ(θi)[sj1sinθisj2cosθi]),ifiiseven,x=(c1,,cn)T,b=(h(θ1),g(θ1),,h(θm),g(θm))T,(85) in which n=2m, and(86) η(θ)=ρ(θ)ρ2(θ)+[ρ(θ)]2.(86) This example poses a great challenge to test the efficiency of linear equations solver, because the Cauchy problem is highly ill-posed. A noise with intensity σ=10% is imposed on the given data. We fix n=30 and take a circle with radius R=500 to distribute the source points. Under ε2=104, we can find A through six iterations. Then, we start from the initial guess x0=Ab, and apply the VRIA to solve Equation (Equation95) through 5 iterations as shown in Figure . In Figure , we compare the numerical solution with the exact data given by u=ρ2cos(2θ),πθ<2π, where ρ=106cos(2θ), of which the maximum error is 0.106. The DIM converges with 25 iterations under the convergence criterion ε2=105. As shown in Figure the maximum error of DIM is 0.0957.

Fig. 4 For example 3 of a Cauchy problem under a large noise 0.1, (a) showing residual error, and (b) comparing the numerical and exact solutions.

Fig. 4 For example 3 of a Cauchy problem under a large noise 0.1, (a) showing residual error, and (b) comparing the numerical and exact solutions.

Fig. 5 For example 4 of an inverse heat source problem under a large noise 0.1, (a) the residual error, (b) the value of a0 and (c) comparing the numerical errors.

Fig. 5 For example 4 of an inverse heat source problem under a large noise 0.1, (a) the residual error, (b) the value of a0 and (c) comparing the numerical errors.

Example 4

In this section, we apply the VRIA and DIM to identify an unknown space-dependent heat source function H(x) for a one-dimensional heat conduction equation:(87) ut(x,t)=uxx(x,t)+H(x),0<x<,0<t<tf,u(0,t)=u0(t),u(,t)=u(t),u(x,0)=f(x).(87) In order to identify H(x), we can impose an extra condition:(88) ux(0,t)=q(t).(88) We propose a numerical differential method by letting v=ut. Taking the differentials of Equations (98), (99), (101) with respect to t, and letting v=ut, we can derive(89) vt(x,t)=vxx(x,t),0<x<,0<t<tf,v(0,t)=u·0(t),v(,t)=u·(t),vx(0,t)=q·(t).(89) This is an inverse heat conduction problem (IHCP) for v(x, t) without using the initial condition.

Therefore, as being a numerical method, we can first solve the above IHCP for v(x, t) by using the MFS in Section 5.1 to obtain a linear equations system, and then the method introduced in Section 4 is used to solve the resultant linear equations system; hence, we can construct u(x, t) by(90) u(x,t)=0tv(x,ξ)dξ+f(x),(90) which automatically satisfies the initial condition in Equation (100).

From Equation (106), it follows that(91) uxx(x,t)=0tvxx(x,ξ)dξ+f(x),(91) which together with ut=v being inserted into Equation (98), leads to(92) v(x,t)=0tvxx(x,ξ)dξ+f(x)+H(x).(92) Inserting Equation (102) for vxx=vt into the above equation and integrating it, we can derive the following equation to recover H(x):(93) H(x)=v(x,0)f(x).(93) For the purpose of comparison we consider the following exact solutions:(94) u(x,t)=x2+2xt+sin(2πx),H(x)=2x2+4π2sin(2πx).(94) In Equation (109), we disregard the ill-posedness of f(x), and suppose that the data f(x) are given exactly. We solve this problem by the DIM and VRIA with γ0=0.05 and the convergence criterion is ε2=104. A random noise with an intensity σ=10% is added on the data q·(t). In the solution of A, the DIM is convergent through 57 iterations and its numerical error as shown in Figure by dashed line is small with the maximum error being 0.023. We compute A under ε2=103, which is convergent with 28 iterations. Then, by starting from the initial guess of x by x0=Ab, we let the VRIA run five iterations, whose residual error and a0 are respectively shown in Figure . The numerical error obtained by the VRIA is shown in Figure by solid line with the maximum error being 0.0086. It is amazing that the accuracy of the VRIA is very good when one knows that the maximum value of H(x) is about 38. This example shows again that the proposed DIM and VRIA can solve linear inverse problems with good efficiency and accuracy, and as well is robustness against the large noise with intensity σ=0.1 being imposed on the data in the right-hand side.

Conclusions

For solving an ill-posed linear equations system Bx=b under a large noisy disturbance and with r=Bxb as being a residual vector, we have derived a very simple iterative algorithm including a parameter γ:(95) xk+1=xk(1γ)rk·vkvk2uk,(95) where(96) uk=Ark,vk=Buk.(96) The Lyapunov stability theorem guarantees that the present algorithm is stable and fast convergent because the search direction Ar is close to the best descent direction B1r with A being a right inversion of B. The new algorithm is a VRIA, which has a superior computational efficiency and accuracy in solving ill-posed linear problems. This assertion is further identified by four linear inverse problems, revealing that the proposed DIM and VRIA can solve very well the linear inverse problem with a high robustness against a large noise σ=0.1 being imposed on the data in the right-hand side.

Acknowledgments

First, the author is grateful to the anonymous referees for their comments which substantially improved the quality of this paper. Taiwan’s National Science Council project NSC-99-2221-E-002-074-MY3 granted to the author is highly appreciated. On the other hand, the author would acknowledge the promotion to be a Lifetime Distinguished Professor of National Taiwan University, since 2013.

References

  • Bhaya A, Kaszkurewicz E. Control perspectives on numerical algorithms and matrix problems. Philadelphia, PA: Society for Industrial and Applied Mathematics; 2006.
  • Liu C-S, Atluri SN. A novel time integration method for solving a large system of on-linear algebraic equations. Comput. Model. Eng. Sci. 2008;31:71–83.
  • Helmke U, Moore JB. Optimization and dynamical systems. Berlin: Springer; 1994.
  • Gavurin MK. Nonlinear functional equations and continuous analogs of iterative methods. Izv. Vyssh. Uchebn. Zaved. 1958;5:18–31.
  • Alber YI. Continuous processes of the Newton type. Differ. Equ. 1971;7:1461–1471.
  • Hirsch MW, Smale S. On algorithms for solving f(x) = 0. Commun. Pure Appl. Math. 1979;32:281–312.
  • Chu MT. On the continuous realization of iterative processes. SIAM Rev. 1988;30:375–387.
  • Ortega IM, Rheinboldt WC. Iterative solutions of nonlinear equations in several variables. New York: Academic Press; 1970.
  • Bhaya A, Kaszkurewicz E. Steepest descent with momentum for quadratic functions is a version of the conjugate gradient method. Neural Networks. 2004;17:65–71.
  • Bhaya A, Kaszkurewicz E. A control-theoretic approach to the design of zero finding numerical methods. IEEE Trans. Autom. Control. 2007;52:1014–1026.
  • Liu C-S, Atluri SN. A Fictitious time integration method for the numerical solution of the Fredholm integral equation and for numerical differentiation of noisy data, and its relation to the filter theory. Comput. Model. Eng. Sci. 2009;41:243–261.
  • Liu C-S, Atluri SN. A highly accurate technique for interpolations using very high-order polynomials, and its applications to some ill-posed linear problems. Comput. Model. Eng. Sci. 2009;43:253–276.
  • Liu C-S, Yeih W, Atluri SN. On solving the ill-conditioned system Ax = b: general-purpose conditioners obtained from the boundary-collocation solution of the Laplace equation, using Trefftz expansions with multiple length scales. Comput. Model. Eng. Sci. 2009;44:281–311.
  • Liu C-S, Hong HK, Atluri SN. Novel algorithms based on the conjugate gradient method for inverting ill-conditioned matrices, and a new regularization method to solve ill-posed linear systems. Comput. Model. Eng. Sci. 2010;60:279–308.
  • Liu C-S. A revision of relaxed steepest descent method from the dynamics on an invariant manifold. Comput. Model. Eng. Sci. 2011;80:57–86.
  • Liu C-S. Modifications of steepest descent method and conjugate gradient method against noise for ill-posed linear systems. Commun. Numer. Anal. 2012;Article ID cna-00115:24.
  • Liu C-S, Atluri SN. An iterative method using an optimal descent vector, for solving an ill-conditioned system Bx = b, better and faster than the conjugate gradient method. Comput. Model. Eng. Sci. 2012;80:275–298.
  • Liu C-S. A dynamical Tikhonov regularization for solving ill-posed linear algebraic systems. Acta Appl. Math. 2013;123:285–307.
  • Liu C-S. The concept of best vector used to solve ill-posed linear inverse problems. Comput. Model. Eng. Sci. 2012;83:499–525.
  • Liu C-S. A globally optimal iterative algorithm to solve an ill-posed linear system. Comput. Model. Eng. Sci. 2012;84:383–403.
  • Liu C-S. Optimally scaled vector regularization method to solve ill-posed linear problems. Appl. Math. Comp. 2012;218:10602–10616.
  • Liu C-S. Optimally generalized regularization methods for solving linear inverse problems. Compu. Mater. Contin. 2012;29:103–127.
  • Liu C-S. An optimal tri-vector iterative algorithm for solving ill-posed linear inverse problems. Inv. Prob. Sci. Eng. 2013;21:650–681.
  • Ascher U, van den Doel K, Hunag H, Svaiter B. Gradient descent and fast artificial time integration. M2AN. 2009;43:689–708.
  • Liu C-S, Chang CW. Novel methods for solving severely ill-posed linear equations system. J. Marine Sci. Tech. 2009;17:216–227.
  • Liu C-S, Atluri SN. An iterative algorithm for solving a system of nonlinear algebraic equations, F(x) = b, using the system of ODEs with an optimum α in ẋ = λ[αF + (1 − α)BFF]; Bij = ∂Fi/∂xj. Comput. Model. Eng. Sci. 2011;73:395–431.
  • Mohler RR. Nonlinear systems, Volume 1: dynamics and control. Englewood Cliffs (NJ): Prentice-Hall; 1991.
  • Chen W, Gu Y. An improved formulation of singular boundary method. Adv. Appl. Math. Mech. 2012;4:543–558.
  • Gu Y, Chen W, He X-Q. Singular boundary method for steady-state heat conduction in three dimensional general anisotropic media. Int. J. Heat Mass Transfer. 2012;55:4837–4848.
  • Lin J, Chen W, Wang F. A new investigation into regularization techniques for the method of fundamental solutions. Math. Comput. Simul. 2011;81:1144–1152.
  • Fu Z, Chen W, Zhang C. Boundary particle method for Cauchy inhomogeneous potential problems. Inv. Prob. Sci. Eng. 2012;20:189–207.
  • Hon YC, Li M. A discrepancy principle for the source points location in using the MFS for solving the BHCP. Int. J. Comput. Meth. 2009;6:181–197.
  • Liu C-S. The method of fundamental solutions for solving the backward heat conduction problem with conditioning by a new post-conditioner. Num. Heat Transfer B: Fundam. 2011;60:57–72.
  • Liu C-S. Group preserving scheme for backward heat conduction problems. Int. J. Heat Mass Transfer. 2004;47:2567–2576.
  • Liu C-S, Zhang SY, Atluri SN. The Jordan structure of residual dynamics used to solve linear inverse problems. Comput. Model. Eng. Sci. 2012;88:29–47.
  • Liu C-S. A highly accurate multi-scale full/half-order polynomial interpolation. Comput. Mater. Contin. 2011;25:239–263.
  • Liu C-S. An equilibrated method of fundamental solutions to choose the best source points for the Laplace equation. Eng. Anal. Bound. Elem. 2012;36:1235–1245.
  • Liu C-S. A two-side equilibration method to reduce the condition number of an ill-posed linear system. Comput. Model. Eng. Sci. 2013;91:17–42.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.