585
Views
0
CrossRef citations to date
0
Altmetric
Research Article

An inverse source identification by nonlinear optimization in a two-dimensional hyperbolic problem

, &
Pages 2110-2130 | Received 09 Sep 2020, Accepted 08 Mar 2021, Published online: 11 Apr 2021

Abstract

This study deals with the identification of source function from final time state observation in a two-dimensional hyperbolic problem. The solution to the direct problem is obtained by the weak solution approach and finite element method. In the part of the inverse problem, the trust-region method and Levenberg–Marquardt method, which are nonlinear least-squares optimization methods, are used for the identification of source function. The findings are presented with numerical examples.

Mathematical Subject classifications:

1. Introduction

In abstract and applied mathematics, the hyperbolic problems related to wave equations have attracted the attention of many scientists. The desire to get some knowledge about the internal dynamics of the wave from its traceable behaviour has been the reason for this interest. In hyperbolic equations, solutions of the corresponding direct problems may have discontinuities, singular components of complicated structure. As it is known, the more stable an operator is, the harder it is to work with its inverse. In this respect, inverse problems for hyperbolic equations have always been popular.

Considering the spatial variables in the region xΩR2 and time variable in the interval t(0,T], let us write the hyperbolic equation on the domain QT:=(x,t)Ω×(0,T] in the form (1.1) 2ut2(c(x)u)+a(x,t)u=σ(t)f(x),(x,t)QT(1.1) (1.2) u(x,0)=u0(x),ut(x,0)=v0(x),xΩ(1.2) (1.3) u(x,t)=g(t),xΩ,t(0,T](1.3) where (1.4) c(x)L(Ω)andc(x)>0,a(x,t)L(QT),σ(t)L2(0,T],f(x)L2(Ω)u0(x)H1(Ω),v(x)L2(Ω),g(t)L2(0,T].(1.4) In vibration modelling, the external forces with the form of separation of variables σ(t)f(x) have special meanings. For example, the selection of σ(t)=coswt describes a harmonic spatial force. Generally, the problem (1.1)–(1.3) is admitted as a model for flexible waves corresponding to a point slipping source. This kind of point source can be connected with models in ground-penetrating radar, reflection seismology, oil and gas exploration and many other physical systems [Citation1].

For a given source σ(t)f(x), obtaining the solution u(x,t) of the initial boundary value problem (1.1)–(1.3) is defined as the direct problem. We can conclude from [Citation2] that under the conditions (1.4), the problem (1.1)–(1.3) has a weak solution uC([0,T],H1(Ω)) with the weak derivatives utC([0,T],L2(Ω)) and uttC([0,T],H1(Ω)). Also, this solution satisfies the following estimate: (1.5) uL(0,T;H1(Ω))+utL(0,T;L2(Ω))c(σL2(0,T]fL2(Ω)+u0H1(Ω)+vL2(Ω)+gL2(0,T])(1.5) Besides, the identification of any unknown data from a given observation related to the solution is defined as the inverse problem.

In this study, we consider the following inverse problem: (1.6) find the source functionf(x)fromfinal time state observationw(x):=u(x,T;f(x))(1.6) where u(x,T;f(x)) represents the solution of the direct problem for a given source f(x)L2(Ω).

In the last decades, inverse source problems with the final time observation for hyperbolic problems have attracted great attention due to both their significance in engineering applications and significance in the theory of inverse problems [Citation3–10].

The solution of the (1.6) inverse problem is achieved by minimizing the following least-squares functional (1.7) F(f(x))=u(x,T;f(x))w(x)2(1.7) On the other hand, these types of inverse problems are ill-posed according to Hadamard requirements which are existence, uniqueness and stability of the solution. There are many studies to overcome the ill-posedness by regularization techniques [Citation4,Citation11–14]

The layout of this article is as follows. In Section 2, we have applied the finite element method for weak solution to direct problem. In Section 3, we have implemented a trust-region algorithm and Levenberg–Marquardt method, which are interrelated methods, except for one point, for the identification problem. In the last section, we have tested the proposed methods on two numerical examples.

2. Weak formulation and finite element method for direct problem

Multiplying the equality of (1.1) by a test function v, which is zero on the boundary and integrating over the domain, we get the following integral equality for t(0,T]: Ωu¨vdxΩ(cu)vdx+Ωauvdx=Ωσ(t)f(x)vdxand by Green’s formula, we have (2.1) Ωu¨vdx+Ω(cu)vdxΩ(cu)nvds+Ωauvdx=Ωσ(t)f(x)vdx(2.1) By the spaces Hg1={v:vL2(Ω)+vL2(Ω)<,v|Ω=g} and H01={v:vL2(Ω)+vL2(Ω)<,v|Ω=0}, the variational formulation of the considered problem is:

find uHg1 such that for every fixed t(0,T] and (2.2) Ωu¨vdx+Ω(cu)vdx+Ωauvdx=Ωσ(t)f(x)vdx,vH01.(2.2) Let us apply the finite element method to the direct problem. The domain Ω is approximated by the union of NT elements Ωτhi=1NTτiwhere τi is the ith element. The union of the elements is called τh, and it is the finite element mesh for the domain Ω. The number h indicates the size of the elements.

We seek the approximate solution in the form (2.3) uh(x,y,t)=j=1NpUj(t)ϕj(x,y)(2.3) where Np is the number of the nodes. The basis function ϕj(x,y) is the second-order Lagrangian shape functions ϕj(x,y)=a1+a2x+a3y+a4x2+a5xy+a6y2.The six coefficients aj of this polynomial are determined by requiring ϕj(xk,yk)=δjk,j,k=1,2,,Np.With these six parameters, we consider constructing a quadratic Lagrange polynomial by placing nodes at the vertices and mid-sides of a triangular element [Citation15].

Now, considering the approximate solution uh in (2.2) Ωu¨hvdx+Ω(cuh)vdx+Ωauhvdx=Ωσ(t)f(x)vdxWe get j=1Npu¨j(t)Ωϕjvdx+j=1NpUj(t)Ω(cϕj)vdx+j=1NpUj(t)Ωaϕjvdx=Ωσ(t)f(x)vdxChoosing v to be each of the basis functions ϕi,i=1,,Np, we write j=1Npu¨j(t)Ωϕjϕidx+j=1NpUj(t)Ω(cϕj)ϕidx+j=1NpUj(t)Ωaϕjϕidx=Ωσ(t)f(x)ϕidx,i=1,,Np.So, we have Np unknowns and Np equations above, which will give a unique solution U1,U2,,UNp.

For solution vector U=[U1,U2,,UNp]T, we can write the following time-dependent matrix equations: (2.4) Md2Udt2+KU+A(t)U=F(t)(2.4) where M=Mij=Ωϕjϕidx,i=1,,Np,j=1,,Np, K=Kij=Ω(cϕj)ϕidx,i=1,,Np,j=1,,Np, A(t)=Ai,j=Ωaϕjϕidx,i=1,,Np,j=1,,Np,

and F(t)=Fi=Ωσ(t)f(x)ϕidx,i=1,,Np.The equality of (2.4) is the system of second-order ordinary differential equation and the initial conditions come from (1.2) such as Uj(0)=u0(Pj),dUjdt(0)=v0(Pj),j=1,,Npand by vector form u0=Uj(0)andv0=dUjdt(0).Hence, we have the system (2.5) Md2Udt2+KU+A(t)U=F(t)u0=U(0),v0=dUdt(0)(2.5) This system can be solved on the time interval (0,T] by any of the numerical methods like Euler's method, Runge–Kutta method or an adaptive scheme. Although the requirement on Δt is not so stringent by an implicit method when solving the hyperbolic problems, we must not forget stability issues. In applying (2.5) to approximate a solution to the hyperbolic equation, we realize that the approximate solution is growing implausible in amplitude, and then the time step must be decreased.

The approximate solution uh satisfies the a priori estimate [Citation15] (2.6) u(t)uh(t)Ch2(u¨(t)+0tu¨(.,s)ds)(2.6)

3. Solution of inverse identification problem

In order to get the unknown source function from the given final time observation by (1.7), we will use the nonlinear least-squares optimization for the problem (3.1) f=argminfF(f(x))(3.1) where F(f(x))=u(x,T;f(x))w(x)2=r(f)2=j=1Np[uh(Pj,T;f(Pj))w(Pj)]2and the components of r(f) are defined as rj(f)=uh(Pj,T;f(Pj))w(Pj),j=1,2,,Np.

Considering the compatibility conditions, the structure of identified function f(x) is estimated and appropriate parameters c={c1,c2,,cm} are included to this function. So, it is assumed that we want to get c, which is the minimum of the objective function F(c):RmrRNpR.Namely, the problem (3.1) is reformulated as follows: (3.2) f:=F(c)=argmincuh(Pj,T;f(Pj))w(Pj)2(3.2) Let us write the Taylor approximation by the degree 2 of F around any c: F(c+Δc)L(Δc)=F(c)+(Δc)Tg+12(Δc)T B (Δc)where L(Δc) is the quadratic approximation of F(c) around c, g the gradient of F(c) computed at c, B an approximation of the Hessian matrix H of F(c) at c and H the real hessian matrix of F(c) at c.

Also, we know, from the optimization theory that if c is a local minimizer, then g(c)=0 (Necessary condition for a local minimizer) and (Δc)TB(c)(Δc)>δΔc2 (Positive definiteness – Sufficient condition for a local minimizer), with some number δ>0. So, we must always construct the positive definite B matrix.

In the remaining of the paper, we use the 2-norm, c=c12++cm2.

On the other hand, we know, from the vector analysis that the gradient at c is g(c)=J(c)T r(c) where r(c) is the vector whose components are defined as follows: rj(c)=uh(Pj,T;f(Pj;c))w(Pj),j=1,2,,Np.Hessian at c is H(c)=J(c)TJ(c)+j=1Nprj(c)rj(c) and approximation of Hessian at c is B(c)=J(c)TJ(c). Here, J(c)=[Jij]Np×m=ricj(c) is the partial derivative (Jacobian) of ri with respect to cj.

According to the Gauss–Newton method, the computation of the step Δc involves the solution of the linear system (3.3) (J(c)TJ(c))Δc=J(c)Tr(c).(3.3) One difficulty of using the Gauss–Newton step is that the Jacobi matrix J(c) may be ill-conditioned, which normally leads to a very big step Δc. And a very long step Δc usually causes the algorithm to break down, because of either numerical overflows or failure in line searches. So, when the Gauss–Newton method is used for ill-conditioned problems, it is not efficient about convergence relations. The main difficulty is that the step size Δc is too large and goes in a ‘bad’ direction that gives little reduction in the function.

We can handle this issue via the use of a Lagrange multiplier and thus replace the problem (3.4) (J(c)TJ(c)+λI)Δc=J(c)Tr(c),(3.4) where λ>0 is the Lagrange multiplier for the constraint at the kth iteration.

The parameter λ affects both the direction and the length of the step Δc. Depending on the size of λ, the correction Δc can vary from a Gauss–Newton step for λ=0 to a very short step approximately in the steepest descent direction for large values of λ. As we see from these considerations, this parameter acts similar to the step control for the damped Gauss–Newton method, but it also changes the direction of the correction.

There are two effective methods for solving such nonlinear equations: Levenberg–Marquardt method, first given by Levenberg [Citation16] and re-derived by Marquardt [Citation17], and trust-region method [Citation18,Citation19]. They are both Gauss–Newton-based methods and expose quadratic speed of convergence near c.

The Levenberg–Marquardt step of (3.4) can be interpreted as solving the normal equations used in the Gauss–Newton method, but ‘shifted’ by a scaled identity matrix, so as to convert the problem from having an ill-conditioned (or positive semi-definite) matrix J(c)TJ(c) into a positive definite one. Notice that the positive definiteness implies that the Levenberg–Marquardt direction has always descent and, therefore, the method is well defined. During the process, the size of λ is updated by controlling the gain ratio ρ=actual decreasepredicted decrease=F(c)F(c+Δc)L(0)L(Δc)A large value of ρ implies that L(Δc) is a good approximation to F(c+Δc) and we can decrease λ so that the next Levenberg–Marquardt step is closer to the Gauss–Newton step. If ρ is small or negative then L(Δc) is a poor approximation and we must increase λ to get steepest descent direction and reduce the step length.

The Levenberg–Marquardt method’s algorithm is as follows;

Levenberg–Marquardt method’s algorithm

The choice of ζ should be a small value, e.g. ζ=106. aii is computed by the diagonal components in J(c0)Tr(c0). ε1, ε2 and kmax are chosen by the user. The inequalities J(ck)Tr(ck)ε1 and Δckε2(ck+ε2) are appropriate stopping criteria and k<kmax is a safeguard against an infinite loop.

On the other hand, reduction in the step size Δc is important. It can be proven that, if we apply a proper limitation on the step size Δcδ, we sustain global convergence even if J(c)TJ(c) is an indefinite matrix. The trust-region algorithm is based on this rule (δ is called the region radius).

In a trust-region method, it is assumed that the model is sufficiently accurate inside a ball with radius δ(trust-region radius) centred at c and the step Δc is to solution to a constrained optimization problem (3.5) minimizeL[Δ(c)]=F(c)+(Δc)Tg+12(Δc)T B (Δc)subject toΔcδ(3.5) Indeed, a trust-region algorithm for nonlinear least-squares and Levenberg–Marquardt method are similar, apart from the bound δ is updated from iteration to iteration directly instead of updating parameter λ.

Trust-region method’s algorithm is as follows:

Trust-region method’s algorithm

Modifying δ directly has the advantage of controlling and observing the length of Δc easily. Hence, nowadays, it is regarded that the trust-region approach is better than the original Levenberg–Marquardt method. A detailed information for the trust-region method can be found in [Citation20,Citation21].

4. Numerical illustrations

Now, we aim to test the explained process by giving on two problems. In solving these problems, we have used the MATLAB-2019 software.

Example 1 Consider the material occupying Ω={(x,y):(x,y)(1,4)×(1,5)(2,3)×(3,5)}.

On the time interval t(0,1], we have the problem (4.1) 2ut2(u)=costf(x),(x,t)Q1(4.1) (4.2) u(x,0)=0,ut(x,0)=0,xΩ(4.2) (4.3) u(x,t)=0,xΩ,t(0,1](4.3)

and want to identify the f(x) source function from the final time state information (4.4) w(x)=u(x,1)=sinπxsinπy(cos2πcos1).(4.4) The exact source function is f(x)=(12π2)sinπxsinπy.

Firstly, let us compute the solution of direct problem f(x)u(x,1) by the finite element method.

The domain Ω is approximated by the union of NT elements such as Ωτhi=1NTτi.In the following, there are the figures of the domain and a finite element mesh for the problem.

The numbers of elements and nodes are 1033 and 2188, respectively, for the mesh size h=0.15.

The exact solution u(x,1) and the calculated solution of the finite element method uh(x,1) by (2.5) are presented in Figures .

Figure 1. The domain and finite element mesh for h=0.15.

Figure 1. The domain and finite element mesh for h=0.15.

Figure 2. Exact wave function at T=1.

Figure 2. Exact wave function at T=1.

Figure 3. Calculated wave function at T=1.

Figure 3. Calculated wave function at T=1.

Some error norm values corresponding different h maximum mesh size are given in Table . Due to ill-conditioning of some matrices in (2.5), it can be seen by h=0.15 and h=0.1 in Table  that a smaller h value gives a bigger error norm. Hence, we take the h=0.15 mesh size for this example. Figure  shows the error of the computation for this mesh size.

Figure 4. Error values for h=0.15.

Figure 4. Error values for h=0.15.

Table 1. Some h and corresponding error norm values.

Secondly, we deal with the solution of the inverse identification problem w(x)f(x) by nonlinear least-squares optimization techniques.

From the given final time state function w(x)=sinπxsinπy(cos2πcos1), we can estimate that the source function may have the form (4.5) f(x)=csinπxsinπy.(4.5) On the other hand, the exact source function leading to the given final time state function is fexact(x)=(12π2)sinπxsinπy18.7392sinπxsinπy.The minimization is carried out by the single-parameter functional (4.6) F(c)=argmincu0.15(Pj,1;f(Pj))w(Pj)2.(4.6) If minimization of this functional by the Levenberg–Marquardt method and trust-region method is carried out using MATLAB then the following outcomes are obtained.

For the Levenberg–Marquardt method, we get Table .

Table 2. Levenberg–Marquardt method.

Optimization is completed by the Levenberg–Marquardt method since the relative norm of the current step (3.67e09) is less than the value of the step tolerance (1e06). The found optimal parameter is cLevenbergMarquardt=18.72911864359083. Hence, the identified source function is (4.7) fLevenbergMarquardt(x)=18.72911864359083sinπxsinπy(4.7) The residual norm given by (4.6) for this function is F=1.857800410424216e05.

For the trust-region method, we get Table .

Table 3. Trust-region method.

Optimization is completed by the trust-region method since the size of the gradient (4.08e07), which is the first-order optimality measure, is less than the value of the optimality tolerance (1e06). The found optimal parameter is ctrust - region=18.72911914110111. Hence, the identified source function is (4.8) ftrust - region(x)=18.72911914110111sinπxsinπy(4.8) The residual norm given by (4.6) for this function is F=1.857800460571223e05.

The difference between the optimal parameter values of two methods is cLevenbergMarquardtctrustregion=4.9751e07.As seen the parameters of these methods are too close to each other. If we take the optimal parameter as c18.7291 then we get the following error norm (4.9) fexact(x)f(x)=0.2310.(4.9) As seen in this example, the Levenberg–Marquardt method was able to calculate the result with the same success using less iteration.

Now, we will examine that the problem needs regularization process or not. To do this, we generate random noisy data wnoisy(Pj). The noisy data are simulated as (4.10) wnoisy(Pj)=w(Pj)+ρ(Pj)w(Pj)(4.10) where ρ(Pj) are random variables with the mean zero and standard deviation σ given by σ=γ, where γ represents the percentage of noise. The random variables ρ(Pj)=normrnd(0,σ,length(Pj)) is generated by normrnd MATLAB function.

Firstly, we generate random noisy data with the percentage of 10% in the observation function and measure the difference between the source functions caused by this noise.

The optimal parameters accompanying with noisy source function are ctrustregion(noisy)18.701749615205685 and cLevenbergMarquardt(noisy)=18.701748831381636.

The 10% noise causes the wnoisy(Pj)w(Pj)w(Pj)=0.1587 norm between observation functions and f(x;w(Pj))f(x;wnoisy(Pj))f(x;w(Pj))=0.0018 norm between source functions.

Secondly, we generate random noisy data with the percentage of 2% in the observation function and measure the difference between the source functions caused by this noise (Figures ).

Figure 5. The difference between the exact and 10% noisy observation function.

Figure 5. The difference between the exact and 10% noisy observation function.

Figure 6. The difference between source functions corresponding to the exact and 10% noisy observation function.

Figure 6. The difference between source functions corresponding to the exact and 10% noisy observation function.

Figure 7. The difference between the exact and 2% noisy observation function.

Figure 7. The difference between the exact and 2% noisy observation function.

Figure 8. The difference between source functions corresponding to the exact and 2% noisy observation function.

Figure 8. The difference between source functions corresponding to the exact and 2% noisy observation function.

The optimal parameters accompanying with noisy source function are ctrustregion(noisy)=18.712029543267924 and cLevenbergMarquardt(noisy)=18.712029495721413.

The 2% noise causes the wnoisy(Pj)w(Pj)w(Pj)=0.0159 norm between observation functions and f(x;w(Pj))f(x;wnoisy(Pj))f(x;w(Pj))=0.00091 norm between source functions.

In Table , we present norms of some noisy observations and corresponding source function differences.

Table 4. Some noisy observations and source function differences.

As seen from Table , while noise levels decrease, the corresponding source function differences also decrease. Hence, there is no need for regularization process.

Example 2 Consider the material occupying the domain bounded by the lines y=12x and y=1.

On the time interval t(0,2], we have the problem (4.11) 2ut2((x+y)u)+2u=et2f(x),(x,t)Q2(4.11) (4.12) u(x,0)=x22y2,ut(x,0)=12(x22y2),xΩ(4.12) (4.13) u(x,t)={0on y=12xet2(x22)on y=1,t(0,2].(4.13)

and want to identify the f(x) source function from the final time state information (4.14) w(x)=u(x,2)=e1(x22y2).(4.14) The exact source function is f(x)=0.25x2+5.5y2.

Firstly, let us compute the solution of direct problem f(x)u(x,2) by finite element method.

In the following, there are the figures of domain and a finite element mesh for the problem (Figure ).

Figure 9. The domain and finite element mesh for h=0.15.

Figure 9. The domain and finite element mesh for h=0.15.

The numbers of elements and nodes are 292 and 671, respectively, for the mesh size h=0.15.

The exact solution u(x,2) and the calculated solution of the finite element method uh(x,2) by (2.5) are presented in Figures .

Figure 10. Exact wave function at T=2.

Figure 10. Exact wave function at T=2.

Figure 11. Calculated wave function at T=2.

Figure 11. Calculated wave function at T=2.

Figure 12. Error values for h=0.15.

Figure 12. Error values for h=0.15.

Figure 13. The difference between exact and 10% noisy observation function.

Figure 13. The difference between exact and 10% noisy observation function.

Figure 14. The difference between source functions corresponding to the exact and 10% noisy observation function.

Figure 14. The difference between source functions corresponding to the exact and 10% noisy observation function.

Figure 15. The difference between the exact and 2% noisy observation function.

Figure 15. The difference between the exact and 2% noisy observation function.

Figure 16. The difference between source functions corresponding to the exact and 2% noisy observation function.

Figure 16. The difference between source functions corresponding to the exact and 2% noisy observation function.

Some error norm values corresponding different h maximum mesh size are given in Table . Due to ill-conditioning of some matrices in (2.5), it can be seen by h=0.15 and h=0.1 in Table  that a smaller h value gives a bigger error norm. Hence, we take the h=0.15 mesh size for this example. Figure  shows the error of the computation for this mesh size .

Table 5. Some h and corresponding error norm values.

Secondly, we deal with the solution of inverse identification problem w(x)f(x) by nonlinear least-squares optimization techniques.

From a given final time state function w(x)=e1(x22y2), we can estimate that the source function may have the form (4.15) f(x)=c1x2+c2y2.(4.15) On the other hand, the exact source function leading to a given final time state function is (4.16) fexact(x)=0.25x2+5.5y2(4.16) The minimization is carried out by the single-parameter functional (4.17) F(c1,c2)=argmin(c1,c2)u0.15(Pj,2;f(Pj))w(Pj)2(4.17) If minimization of this functional by the Levenberg–Marquardt method and trust-region method is carried out using MATLAB then the following outcomes are obtained.

For the Levenberg–Marquardt method, we get Table .

Table 6. Levenberg–Marquardt method.

Optimization is completed by the Levenberg–Marquardt method since the final change in the sum of squares relative to its initial value is less than the value of the function tolerance (1e06). The found optimal parameter is cLevenbergMarquardt=[0.246809936178489,5.500964113399660]. Hence, the identified source function is (4.18) fLevenbergMarquardt(x)=0.246809936178489x2+5.500964113399660y2(4.18) The residual norm given by (4.16) for this function is F=3.381226871464835e05.

For the trust-region method, we get Table .

Table 7. Trust-region method.

Optimization is completed by the trust-region method since the final change in the sum of squares relative to its initial value is less than the value of the function tolerance (1e06). The found optimal parameter is ctrust - region=[0.246811657528541,5.500963848795268]. Hence, the identified source function is (4.19) ftrust - region(x)=0.246811657528541x2+5.500963848795268y2(4.19) The residual norm given by (4.6) for this function is F=3.381227400848656e05.

The difference between the optimal parameter values of two methods is cLevenbergMarquardtctrust - region=1.741568e06.As seen, the parameters of these methods are too close to each other. If we take the optimal parameter as c[0.24681,5.50096] then we get the following error norm (4.20) fexact(x)f(x)=0.0398(4.20) As seen in this example, the trust-region method was able to calculate the result with the same success using less iteration.

Now, we generate random noisy data wnoisy(Pj) by (4.10) to decide the regularization process is needed or not (Figures ).

Firstly, we generate random noisy data with the percentage of 10% in the observation function and measure the difference between the source functions caused by this noise.

The optimal parameters accompanying with noisy source function are ctrust - region(noisy)=[1.034134676584133,5.123458167646760] and cLevenbergMarquardt(noisy)=[1.016975683797493,5.127563889242101].

The 10% noise causes the wnoisy(Pj)w(Pj)w(Pj)=0.0930 norm between observation functions and f(x;w(Pj))f(x;wnoisy(Pj))f(x;w(Pj))=0.0986 norm between source functions (for the trust-region method).

Secondly, we generate random noisy data with the percentage of 2% in the observation function and measure the difference between the source functions caused by this noise.

The optimal parameters accompanying with noisy source function are ctrust - region(noisy)=[0.006929996321541,5.548445550509578] and cLevenbergMarquardt(noisy)=[0.006472533875535,5.548249867102941].

The 2% noise causes the wnoisy(Pj)w(Pj)w(Pj)=0.0194 norm between observation functions and f(x;w(Pj))f(x;wnoisy(Pj))f(x;w(Pj))=0.0387 norm between source functions (for the trust-region method).

In Table , we present norms of some noisy observations and corresponding source function differences.

Table 8. Some noisy observations and source function differences

As seen from Table , while noise levels decrease, the corresponding source function differences also decrease. Hence, there is no need for regularization process.

5. Conclusions

The trust-region and Levenberg–Marquardt methods which are nonlinear least-squares optimization methods are powerful tools for identification problems. The methods known to give successful results in one-dimensional problems have also yielded successful results in two-dimensional problems. The fact that the regularization operations are carried out within the methods eliminates the need for adding an extra regularization term to the objective function. This situation can be seen from Tables  and .

Although the difference between two methods is not much in the first example with one identified parameter, in the second example where two identified parameters are included, the trust-region method yielded less iterations than the Levenberg–Marquardt method, in compatible with general predictions.

Acknowledgements

We are very grateful to the reviewers for their comments and suggestions.

Disclosure statement

No potential conflict of interest was reported by the author(s).

References

  • Aki K, Richards PG. Quantitative seismology theory and methods. New York: Freeman; 1980.
  • Evans LC. Partial differential equations, 2nd edn. Graduate studies in mathematics, vol. 19. New York: American Mathematical Society; 2002.
  • Kuliev GF. Problem of optimal control of the coefficients for hyperbolic equations. Izvestiya Vysshikh Uchebnykh Zavedenii. Matematika. 1985;3:39–44.
  • Feng X, Sutton B, Leuhart S, et al. Identification problem for the wave equation with Neumann data input and Dirichlet data abservating. Nonlinear Anal. 2003;52(7):1777–1795.
  • Maciąg A. The usage of wave polynomials in solving direct and inverse problems for two-dimensional wave equation. Int J Numer Method Biomed Eng. 2011;27(7):1107–1125.
  • Tagiyev RK. On optimal control of the hyperbolic equation coefficients. Autom Remote Control. 2012;73(7):1145–1155.
  • Hasanov A, Mukanova B. Relationship between representation formulas for unique regularized solutions of inverse source problems with final overdetermination and singular value decomposition of input–output operators. IMA J Appl Math. 2015;80:676–696.
  • Subaşı M, Kaçar A. A variational technique for optimal boundary control in a hyperbolic problem. Appl Math Comput. 2012;218:6629–6636.
  • Deiveegan A, Prakash P, Nieto JJ. Optimization method for identifying the source term in an inverse Wave equation. Electron J Diff Equat. 2017;2017(200):1–15.
  • Subaşı M, Araz S. Numerical regularization of optimal control for the coefficient function in a wave equation. Iran J Sci Technol Trans A Sci. 2019;43:2325–2333.
  • Eng HW, Scherzer O, Yamamato M. Uniqueness and stable determination of forcing terms in linear partial differential equations with overspecified boundary data. Inverse Probl. 1994;10:1253–1276.
  • Yamamato M. On ill-posedness and a Tikhonov regularization for a multidimensional inverse hyperbolic problem. J Math Kyoto Univ. 1996;36:825–856.
  • Cheng J, Yamamato M. One new strategy for a priori choice of regularization parameters in Tikhonov’s regularization. Inverse Probl. 2000;16:L31–L38.
  • Kabanikhin SI, Satybaev AD, Shishlenin MA. Direct methods of solving multidimensional inverse hyperbolic problems. Utrecht: VSP Science Press; 2005.
  • Larson MG, Bengzon F. The finite element method: theory, implementation, and applications. Berlin: Springer; 2010.
  • Levenberg K. A method for the solution of certain nonlinear problems in least squares. Qart Appl Math. 1944;2:164–166.
  • Marquardt DW. An algorithm for least-squares estimation of nonlinear inequalities. SIAM J Appl Math. 1963;11:431–441.
  • Goldfeld SM, Quandt RE, Trotter HF. Maximization by quadratic hill-climbing. Econometrica. 1966;34(3):541–551.
  • Sorensen DC. Newton's method with a model trust region modification. SIAM J Numer Anal. 1982;19(2):409–426.
  • Conn AR, Gould NIM, Toint PL. Trust-region methods. Philadelphia: Society for Industrial and Applied Mathematics; 2000.
  • Madsen K, Nielsen HB, Tingleff O. Methods for non-linear least squares problems, informatics and mathematical modelling. Kopenhag: Technical University of Denmark; 2004.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.