502
Views
3
CrossRef citations to date
0
Altmetric
Research Article

Nonlinear least squares algorithm for identification of hazards

| (Reviewing Editor)
Article: 1118219 | Received 09 Mar 2015, Accepted 03 Nov 2015, Published online: 15 Dec 2015

Abstract

Given observations of selected concentrations one wishes to determine unknown intensities and locations of the sources for a hazard. The concentration of the hazard is governed by a steady-state nonlinear diffusion–advection partial differential equation and the best fit of the data. The discretized version leads to a coupled nonlinear algebraic system and a nonlinear least squares problem. The coefficient matrix is a nonsingular M-matrix and is not symmetric. Iterative methods are compositions of nonnegative least squares and Picard/Newton methods, and a convergence proof is given. The singular values of the associated least squares matrix are important in the convergence proof, the sensitivity to the parameters of the model, and the location of the observation sites.

AMS subject classifications:

Public Interest Statement

Chemical or biological hazards can move by diffusion or fluid flow. The goal of the research is to make predictions about the location and intensities of the hazards based on the down stream location of observations sites. Mathematical models of the diffusion and flow are used to extrapolate from the observations data sites to unknown upstream locations and intensities of the hazards. The convergence of the iterative process for these extrapolations is studied. The sensitivities of the calculations, because of variations of the model parameters, observation data, and location of the observation sites, are also studied.

1. Introduction

The paper is a continuation of the work in White (Citation2011). The first new result is a convergence proof of the algorithm to approximate a solution to a nonlinear least square problem. A second new result is a sensitivity analysis of the solution as a function of the model parameters, measured data, and observation sites. Here, the singular value decomposition of the least squares matrix plays an important role.

A convergence proof is given for the algorithm to approximate the solution of a coupled nonlinear algebraic system and a nonnegative least squares problem, see lines (9–11) in this paper and in White (Citation2011). These discrete models evolve from the continuous models using the advection–diffusion PDE in Equation (1) (see Pao, Citation1992). In contrast to the continuous models in Andrle and El Badia (Citation2015), Hamdi (Citation2007), and Mahar and Datta (Citation1997), this analysis focuses on the linear and semilinear algebraic systems. In Hamdi (Citation2007), the steady-state continuous linear model requires observations at all of the down stream portions of the boundary. In Andrle and El Badia (Citation2015), the continuous linear velocity free model where the intensities of the source term are time dependent requires observations at the boundary over a given time interval. In both these papers, the Kohn–Vogelius objective function is used for the symmetric models in the identification problems.

Here and in White (Citation2011) we assume the coefficient matrix associated with the steady-state discretized partial differential equation in Equation (1) is an M-matrix and may not be symmetric, which includes coefficient matrices from upwind finite difference schemes for the advection–diffusion PDE. The least squares matrix C( : , ssites), see Equations (6–8), is an k×l matrix where k>l and k and l are the number of observation and source sites, respectively. The rank and singular values (see Meyer, Citation2000, section 5.12), of C( : , ssites) determines whether of not the observation and source sites are located so that the intensities can be accurately computed.

Four examples illustrate these methods. The first two examples in Section 3 are for n=2 and n=3 unknowns, and they show some of the difficulties associated with the nonlinear problem. Sections 4 and 5 contain the convergence proof. Examples three and four in Section 6 are applications of the 1D and 2D versions of the steady-state advection–diffusion PDE. Sensitivity of the computed solutions due to variation of the model parameters, the data and the relative location of the source and observation sites is illustrated. The connection with the sensitivity and the singular values of the least squares matrix is established.

2. Problem formulation

Consider a hazardous substance with sources from a finite number of locations. The sources are typically located at points in a plane and can be modeled by delta functionals. The hazard is described as concentration or density by a function of space and time, u(xyt). This is governed by the advection–diffusion PDE(1) ut-Du+vu+dru=f(u)+S(1)

where S is a source term, D is the diffusion, v is the velocity of the medium, dr is the decay rate, and f(u) is the nonlinear term such as in logistic growth models. The source term will be a finite linear combination of delta functionals(2) S=aiδ(x-x^i)(2)

where the coefficients ai are the intensities and the locations x^i are possibly unknown points in the line or plane. The source coefficients should be nonnegative, which is in contrast with impulsive culling where the coefficient of the delta functionals are negative and have the form -ciu(xi) (see White, Citation2009).

We will assume the source sites and the observation sites do not intersect. The reason for this is the ability to measure concentrations most likely allows one to measure the intensities at the same site.

Identification problem: Given data for u(osites) at a number of given observation sites, osites,  determine the intensities (ai) and locations (x^i) at source sites, ssites, so thatdata-u(osites)is suitably small.

2.1. Linear least squares

Since the governing PDE is to be discretized, we first consider the corresponding linear discrete problemAu=d.

The coefficient matrix A is derived using finite differences with upwind finite differences on the velocity term. Therefore, we assume A is a nonsingular M-matrix (see Berman & Plemmons, Citation1994, chapter 6 or Meyer, Citation2000, section 7.10). The following approach finds the solution in one least squares step. The coefficient matrix is n×n and nodes are partitioned into three ordered sets, whose order represents a reordering of the nodes:ositesobserve sitesssitessource sites andrsitesremaining sites.

The discrete identification problem is given data data at osites, find d(ssites) such thatAu=dsuch thatu(osites)=data.

In general, the reordered matrix has the following 2×2 block structure with other=ssitesrsites(3) A=aefb(3)

where a=A(osites,osites), e=A(osites,other), f=A(other,osites) and b=A(other,other).

Assumptions of this paper:

(1)

A is an M-matrix so that A and a have inverses.

(2)

#(osites)=k, and #(ssites)=l<k so that n=k+l+#(rsites).

(3)

d=0d1TT and the only nonzero terms in d1 are at the nodes in ssites, that is, d0d1(ssites)T0T=0zT0T.

Multiply Equation (3) by a block elementary matrix (as in assumptions 1 and 2)(4) Ik0-fa-1In-kaefbu0u1=Ik0-fa-1In-k0d1(4) (5) ae0b^u0u1=0d1(5)

where b^b-fa-1e. Solve Equation (5) for u1 and then u0=-a-1e(b^)-1d1. Define(6) d1[d1(ssites)T0]TandC-a-1e(b^)-1.(6)

The computed approximations of the observed data are in u0 and the source data are in z=d1(ssites). This gives a least squares problem for z (7) data=C(:,ssites)z.(7)

Since a is k×k, e is k×(n-k), and b^ is (n-k)×(n-k), the matrix C is k×(n-k) and the least squares matrix C( : , ssites) is k×l where k>l. The least squares problem in (7) has a unique solution if the columns of C( : , ssites) are linearly independent (C(:,ssites)z=0 implies z=0).

Impose the nonnegative condition on the components of the least squares solution. That is, consider finding z0 such that(8) r(z)Tr(z)=miny0r(y)Tr(y)wherer(z)data-C(:,ssites)z(8)

If the matrix C( : , ssites) has full column rank, then the nonnegative least squares problem (8) has a unique solution (see Bjorck, Citation1996, section 5.2). The nonnegative least squares solution can be computed using the MATLAB command lsqnonneg.m (see MathWorks Inc., Citation2015).

2.2. Least squares and nonlinear term

Consider a nonlinear variation of Au=d(9) Au=d(z)+d^(u)(9)

where d(z)=0zT0T (as in assumption 3) and d^(u)=[d^i(ui)].

Use the reordering in (3) and repeat the block elementary matrix transformation in (4) and (5) applied to the above. Using the notation d^(u)=[d^0(u)Td^1(u)T]T one getsae0b^u0u1=Ik0-fa-1In-k0d1+d^0(u)d^1(u)=0d1+d^0(u)-fa-1d^0(u)+d^1(u).

Solve for u1 and then for u0u0=a-1d^0(u)-a-1e(b^)-1d1+(-fa-1d^0(u)+d^1(u))=g(u)+Cd1

where C-a-1e(b^)-1 and g(u)a-1d^0(u)+C(-fa-1d^0(u)+d^1(u)).

This leads to a nonlinear least squares problem for nonnegative z (10) data-g(u)=C(:,ssites)z.(10)

The solution of the nonnegative least squares problem is input to the nonzero components in d, that is, d(ssites)=z. The coupled problem from (9) and (10) is to find z0 and u so that(11) data-g(u)=C(:,ssites)zandAu=d(z)+d^(u).(11)

3. Two algorithms for the coupled problem

If the nonnegative least squares problem has a unique solution z=H(u) given a suitable u, then we can write (11) asAu=d(H(u))+d^(u),

or as a fixed pointu=A-1(d(H(u))+d^(u))G(u).

The following algorithm is a composition of nonnegative least squares and a Picard update to form one iteration step. Using additional assumptions we will show this is contractive and illustrate linear convergence. Example 3.2 is a simple implementation with n=3, and in Section 6 Example 6.1 uses the 1D steady-state version of the model in (1) and Example 6.2 uses the 2D steady-state model in (1).

Algorithm 1. NLS-Picard Method for (11)       choose nonnegative u0       for m=0, mmax             solve the nonnegative least squares problem                   data-g(um)=C(:,ssites)zm+1             solve the linear problem                   Aum+1=d(zm+1)+d^(um)             test for convergence       end for-loop

The following algorithm uses a combination of nonnegative least squares and Newton like updates. Define F(zu) and Fu(z,u) asF(z,u)Au-d(z)-d^(u)andFu(z,u)A-0-d^u(u)

where d^u(u)[(d~i)(ui)] is the diagonal matrix with components (d~i)(ui).

The matrix Fu(z,u) is not the full derivative matrix of F(zu) because it ignores the dependence of z on u. Choose λ>0 so that Fu+λI has an inverse.

Algorithm 2. NLS-Newton Method for (11)       choose nonnegative u0and λ>0       for m=0, mmax             solve the nonnegative least squares problem                   data-g(um)=C(:,ssites)zm+1             solve the linear problem                   Aum+1/2=d(zm+1)+d^(um)             compute F(zm+1,um+1/2)Aum+1/2-d(zm+1)-d^(um+1/2)                   =d^(um)-d^(um+1/2)             compute Fu(zm+1,um+1/2)A-0-d^u(um+1/2)             solve the linear problem                   (Fu(zm+1,um+1/2)+λI)Δu=F(zm+1,um+1/2)             um+1=um+1/2-Δu             test for convergence       end for-loop

Variations of the NLS-Newton algorithm include using over relaxation, ω, at the first linear solve and using damping, α, on the second linear solve in the Newton update:um+1/2=(1-ω)um+ωum+1/2andum+1=um+1/2-αΔu.

Appropriate choices of w,α and λ can accelerate convergence.

The calculations recorded in Table 4 and Figure 4 in White (Citation2011), illustrate the NLS-Newton algorithm for the 1D steady-state problem with logistic growth term d^(u)=f(u)=c(100-u)u and variable growth rate c. The larger the c parameter resulted in larger concentrations. The intensities at the three source sites were identified. However, the number of iterations of the NLS-Newton algorithm increased as the c parameter increased. The algorithm did start to fail for c larger than 0.0100.

The lack of convergence is either because the associated fixed point problem does not have a contractive mapping or because the solutions become negative. The following example with n=2, nonzero source term and nonsymmetric A illustrates the difficulty of solving a semilinear algebraic problem where the solution is required to be positive.

Example 3.1

Let n=2 and use fi(ui)=c(100-ui)ui. Equation (9) has the form10-ε1u0u1=0z+f0(u0)f1(u1).

The first equation is u0=c(100-u0)u0 whose solution is either u0=0 or u0=100-1/c>0. Put the nonzero solution into the second equation -ε(100-1/c)+u1=z+c(100-u1)u1 and solve for u1. If ε(100-1/c)+z>0, then the second equation is u1=c(100-u1)u1+ε(100-1/c)+z and will have a positive solution.

The next example with n=3, given data at the first and second nodes and a source term at the third node illustrates the coupled problem in (11). This simple example can be solved both by-hand calculations and the NLS-Picard algorithm, and more details are given in Appendix 1.

Example 3.2

Let the data be given at the first and second nodesdata=2.97115.4971.

Let the matrix A and nonlinear terms be given by system30-103-2-2-24u1u2u3=00z+cu1(10-u1)cu2(10-u2)cu3(10-u3).

The submatrices of A area=3003,e=-1-2,f=-2-2andb=[4].

Then one easily computesb^=b-fa-1e=2andC=-a-1eb^-1=(1/6)12.

The g(u) in the nonlinear least square Equation (11) isg(u)=c(4/9)u1(10-u1)+(1/6)u3(10-u3)+(1/9)u2(10-u2)(5/9)u2(10-u2)+(2/6)u3(10-u3)+(2/9)u1(10-u1).

The calculation with the NLS-Picard algorithm for this example with n=3 in White (Citation2015, testlsnonlc.m) confirms both the exact solution for c=8/100=0.0800 and convergence for c<0.1667, which are given in Appendix 1. The number of iterations needed for convergence increases as c increases, and convergence fails for c=0.3000. Both the exact and NLS-Picard calculations show z>0 for c0.1975 and z=0 for c0.1976. The outputs for five values of c=0.08:0.04:0.24 are given in Figure . The left graphs indicate the five solutions of (10) for fixed z=zz=10 and increasing c. The two “*” are the given data, which are from the solution of (10) with c=0.0800 and 5% error. The right graphs are the converged solutions of (11) for c=0.08:0.04:0.20.

4. Nonnegative least squares problem

The nonnegative least squares problem in (10) for z0 has several formsR(z)TR(z)=miny0R(y)TR(y)whereR(z)data-g(u)-C(:,ssites)z.

This may be written as(12) R(z)TR(z)=(data-g(u))T(data-g(u))-2J(z)(12)

whereJ(z)12zTC(:,ssites)TC(:,ssites)z-zTC(:,ssites)T(data-g(u)).

Figure 1. NLS-Picard algorithm for n = 3.

Figure 1. NLS-Picard algorithm for n = 3.

If C( : , ssites) has full column rank, then the matrix C(:,ssites)TC(:,ssites) is symmetric positive definite (SPD). J(z) is the quadratic functional associated with the linear system (normal equations)0=r(z)C(:,ssites)T(data-g(u))-C(:,ssites)TC(:,ssites)z

When the nonnegative condition is imposed on z0 and the matrix is SPD, then the following are equivalent (see Cryer, Citation1971):J(z)=miny0J(y)(minimum quadratic functional),0r(z)T(y-z)for ally0(variational inequality) and0=r(z)Tzand0r(z)(linear complementarity problem).

Any solution of a variational inequality is unique and depends continuously on the right-hand side. The next theorem is a special case of this.

Theorem 4.1

Consider the nonnegative least squares problem in (10). If C( : , ssites) has full column rank, then there is only one nonnegative solution for each u. Let z=H(u) be the solution for each u in S where S is a bounded set of nonnegative n-vectors. Moreover, if d^(u)-d^(v)K^u-v, then there is a constant C1 such thatC(:,ssites)(H(u)-H(v))C1u-v

Proof

Let z and w be solutions of (10). The variational inequalities for z with y=w, and for w with y=z are0r(z)T(w-z)and0r(w)T(z-w).

Add these to get a contradiction for the case z-w is not a zero vector0(z-w)TC(:,ssites)TC(:,ssites)(z-w)>0.

In order to show the continuity, use the above with z=H(u) and w=H(v):0[C(:,ssites)T(data-g(u))-C(:,ssites)TC(:,ssites)H(u)]T(H(v)-H(u))0[C(:,ssites)T(data-g(v))-C(:,ssites)TC(:,ssites)H(v)]T(H(u)-H(v)).

Add these to obtain for BC(:,ssites)TC(:,ssites)(H(v)-Hu))TB(H(v)-H(u))-(g(v)-g(u))TC(:,ssites)(H(v)-H(u)).

Use the Cauchy inequalityC(:,ssites)(H(v)-H(u))2g(v)-g(u)C(:,ssites)(H(v)-H(u)).

By the definition of g and the assumption on d we have for some C1C(:,ssites)(H(v)-H(u))C1v-u.

The solution of the variational problem for fixed u has the form z=H(u). The assumption that C( : , ssites) has full column rank implies there is a positive constant c0 such that c02wTwwTC(:,ssites)TC(:,ssites)w. This and the above theorem give the following(13) c0H(v)-H(u)C(:,ssites)(H(v)-H(u))C1v-u.(13)

Although in Example 3.2 with n=3 one can estimate c0 and C1, generally these constants are not easily approximated. However, c0 can be estimated by the smallest singular value of the least squares matrix.

5. Convergence proof of the NLS-Picard algorithm

Let the assumptions in the previous theorem hold so that (12) is true for all u,vS. The mapping G(u)A-1(d(H(u))+d^(u)) may not give G(u)S! This appears to be dependent on the particular problem and properties of d^(u). If the bound is chosen so that d^(u) is also nonnegative, then assumption (1) implies A-10 and hence G(u)0. The least squares condition requires the solution to be “close” to the given data at the osites, which suggests the G(u) should remain bounded. The following theorem gives general conditions on the problem.

Theorem 5.1

Let the assumptions of Theorem 4.1 hold for u,vS. Assume there is a solution to the coupled problem (11), uS. Let c0 and C1 be from (12). Ifr^A-1(C1/c0+K^)<1,

then the NLS-Picard algorithm will converge to this solution of (11).

Proof

Let u=G(u) and um+1=G(um).u-um+1=G(u)-G(um)=A-1(d(H(u)-d(H(um))+A-1(d^(u)-d^(um))A-1(C1/c0)u-um+A-1K^u-um=A-1(C1/c0+K^)u-um=r^u-umr^m+1u-u0.

Since r^<1, this implies convergence

6. Applications and sensitivity analysis

Both the NLS-Picard and NLS-Newton methods converge linearly, that is, the errors at the next iteration are bounded by a contractive constant, r^<1, times the error from the current iteration. However, the contractive constant for the NLS-Newton algorithm is significantly smaller than the contractive constant for the NLS-Picard algorithm, and this results in fewer iterations steps for the NLS-Newton algorithm. The calculations in the next two examples can be done by either algorithm. The relevance of the singular values of the least squares matrix to the sensitivity of the calculations with variations in the data and the source and observation sites is demonstrated.

Figure 2. NLS-Newton algorithm with variable c.

Figure 2. NLS-Newton algorithm with variable c.

Example 6.1

Consider the steady-state 1D model in (1). The following calculations were done by the MATLAB code in White (Citation2015, poll1dlsnonla2.m) and with seven observation sites, three source sites, and n=160. The differential Equation (1) was discretized by the upwind finite difference approximations, and the sources terms were the intensities times the approximation of the delta functionalsd(ssites)=[a1a2a3]T/dx

where initially dzeros(n,1). In order to check the accuracy of the finite difference model, calculations were done with n=160,320 and 480, which gave similar results as reported below for n=160. The condition number (ratio of the largest and smallest singular values) of the least squares matrix controls the relative errors in the computed least squares problem. The condition numbers for the n=160,320, and 480 are about the same at 17.0170, 16.5666, and 16.4146, respectively.

For c=0.010 the NLS-Picard algorithm did not converge, but the NLS-Newton algorithm converged in 92 steps. The solid lines in Figure are with no random error in the data, and the dotted lines are from data with 5% random error. For c=0.005 the NLS-Picard algorithm converged in 100 steps with r^.83, and the NLS-Newton algorithm converged in 19 steps with r^.41. If w is changed from 1.0 to 1.1, then the NLS-Newton algorithm converges in 16 steps with r^.28. As c decreases, the iterations required for convergence and r^ decrease.

The dotted lines indicate some uncertainty in the numerical solutions, which can be a function of the physical parameters and the location of the observation sites relative to the source sites. If there is no error in the data, the intensities of the sources are in the vectorz=4.00001.00002.0000.

Use 100 computations with the c=0.001 and 5% random errors. The means and standard deviations of the computed intensities at the three source sites aremean(z)=4.03570.94372.0187andstd(z)=0.25700.36310.1375.

The standard deviation of the intensities at the center source in Figure is large relative to its mean value. The large computed least squares errors are a result of some relatively small singular values in the singular value decomposition of the least squares matrixC(:,ssites)=UΣVT

where the singular values σi are the diagonal components of Σ; U is 7×7, Σ is 7×3 and V is 3×3.

In this example, the third singular value is smallσ=0.28820.09160.0169.

The least square solution of (10) is given by the pseudo inverse of the least squares matrix times data-g(u)d^VΣUTd^=v1u1Td^/σ1+v2u2Td^/σ2+v3u3Td^/σ3.

Because σ3 is relatively small, any errors in d^ will be amplified as well as the components ofv3=0.5778-0.78820.2121T.

This suggests the second component, the center source, of least squares solution will have significant errors. The errors can be decreased if the percentage random error is smaller or by alternative location of the observation sites. In the 1D model this is not too difficult to remedy. However, in the 2D model this is more challenging.

Example 6.2

Consider the steady-state 2D model in (1) with boundary conditions equal to zero up stream and zero normal derivative down stream. The calculations in Figure were done by the MATLAB code in White (Citation2015, testlsnonlf2da3.m) using NLS-Picard and with four variable observation sites and three fixed source sites. The three fixed sources are to the left near x=0, and the four observation sites are indicated by the symbol “*” above the site. The intensities times the approximation of the delta functionals isd(ssites)=[a1a2a3]T/(dxdy)

where initially dzeros(nn,1) and nn=(nx+1)(ny+1). If there is no random error in the data, then the computed source will be8.000016.00004.0000.

The numerical experiments involve varying the location of the observation site at (x, 16) where x=L-ΔL and varying the y-component of the velocity, vely. The velocity term in (1) is (velx,vely)=(1.00,0.10). The base calculation in Figure uses ΔL=30 and vely=0.10 along with 5% random error in the data and c=0.0004 growth coefficient in the nonlinear term. The singular values of the least squares matrix and the mean and standard deviations of the intensities of the three computed sources are recorded for 100 computations.

ΔL=30 and vely=0.10:σ=0.04720.03000.0046mean(z)=7.994716.03414.0040andstd(z)=0.25230.49660.1451.

Changing the y-component of the velocity can cause the concentration to be moved away from the observation sites and cause more uncertainty in the calculations, which is indicated by larger standard deviations and smaller singular values of the least squares matrix. Change vely=0.10 to vely=0.15 and vely=0.05.

ΔL=30 and vely=0.15:σ=0.04240.03020.0002mean(z)=7.956216.03423.6939andstd(z)=0.23990.547930.7650.

Here, the third singular value has significantly decreased and can cause failure of the NLS-Picard algorithm or increased computation errors because of the variation in the data.

ΔL=30 and vely=0.05:σ=0.08530.00040.0001mean(z)=7.788316.60303.9778andstd(z)=0.57442.08840.1167.

In this case the second and third singular values of the least squares matrix decreased, and the standard deviation of the second source increased.

Changing the location of the third observation site can also cause more uncertainty in the calculations, which is indicated by larger standard deviations and smaller singular values of the least squares matrix. Change ΔL=30 to ΔL=20 and ΔL=10 and keep vely=0.10.

ΔL=20 and vely=0.10:σ=0.04720.03000.0005mean(z)=8.001216.03504.5229andstd(z)=0.23500.53214.5265.

Here, the third singular value has significantly decreased and can cause failure of the NLS-Picard algorithm or increased computation errors because of the variation in the data. The standard deviation of the third source has significantly increased.

ΔL=10 and vely=0.10:σ=0.04720.03200.0000396mean(z)=8.000515.9137238.9531andstd(z)=0.23790.5007396.3124.

In this case the third singular values of the least squares matrix have dramatically decreased, and the standard deviation of the third source is very large relative to the mean.

The calculation in Figure has six source sites (near the axes) and nine observation sites (down stream and below the “*”) (White, Citation2015, testlsnonlf2da6.m) uses NLS-Picard. Here the vely is larger with velocity the equal to (1.00, 0.30), and the random error in the data is smaller and equal to 1%:σ=0.07820.06120.04590.03040.01780.0047mean(z)=23.940416.02158.01353.986916.15397.9334andstd(z)=0.85120.44080.20640.05490.90460.2879.

The source closest to the origin is the most difficult to identify; with one percent error in the data the standard deviation is 0.8512 and with five percent error this goes up to 4.2348. This happens despite the relatively close distribution of the singular values from 0.0782 to 0.0047 for a condition number equal to 16.7658.

7. Summary

In order to determine locations and intensities of the source sites from data at the observation sites, there must more observations sites than source sites and the observation sites must be more chosen so that the least squares matrix C( : , ssites) has full column rank. Furthermore, the singular values in the singular value decomposition of the least squares matrix are important in the convergence proof and the location of the observation sites.

Figure 3. Three sources and four observations.

Figure 3. Three sources and four observations.

Figure 4. Six sources and nine observations.

Figure 4. Six sources and nine observations.

Measurement errors in the observations as well as uncertain physical parameters can contribute to significant variation in the computed intensities of the sources. Multiple computations with varying these should be done, and the means and standard deviations should be noted. Large standard deviations from the mean indicate a lack of confidence in the computations. In this case, one must adjust the observation sites, which can be suggested by inspection of the smaller singular values for the least squares matrix and the corresponding columns in V of the singular value decomposition.

Additional information

Funding

No external funding. Some of this research was done before retirement at North Carolina State University.

Notes on contributors

Robert E. White

Professor Robert E. White has published numerous articles on numerical analysis including approximate solutions of partial differential equations, numerical linear algebra, and parallel algorithms. The most recent work is on hazard identification given observation data. He has authored three textbooks. The second edition of Computational Mathematics: Models, Methods and Analysis with MATLAB and MPI will be published in the fall of 2015 by CRC Press/Taylor and Francis.

References

  • Andrle, M., & El Badia, A. (2015). On an inverse problem. Application to a pollution detection problem, II. Inverse Problems in Science and Engineering, 23, 389–412.
  • Berman, A., & Plemmons, R. J. (1994). Nonnegative matrices in the mathematical sciences. Philadelphia, PA: SIAM.
  • Bjorck, A. (1996). Numerical methods for least squares problems. Philadelphia, PA: SIAM.
  • Cryer, C. W. (1971). The solution of a quadratic programming problem using systematic over ralaxation. SIAM Journal on Control and Optimization, 9, 385–392.
  • Hamdi, A. (2007). Identification of point sources in two-dimensional advection-diffusion-reaction equations: Application to pollution sources in a river. Stationary case, Inverse Problems in Science and Engineering, 15, 855–870.
  • Mahar, P. S., & Datta, B. (1997). Optimal monitoring network and ground-water-pollution source identification. Journal of Water Resources Planning and Management, 199, 199–2007.
  • MathWorks Inc. (2015). (123, pp. 199–207). Retrieved from http://www.mathworks.com
  • Meyer, C. D. (2000). Matrix analysis and applied linear algebra. Philadelphia, PA: SIAM.
  • Pao, C. V. (1992). Nonlinear parabolic and elliptic equations. New York, NY: Plenum Press.
  • White, R. E. (2009). Populations with impulsive culling: Identification and control. International Journal of Computer Mathematics, 86, 2143–2164.
  • White, R. E. (2011). Identification of hazards with impulsive sources. International Journal of Computer Mathematics, 88, 762–780.
  • White, R. E. (2015). MATLAB codes for hazard identification. Retrieved from http://www4.ncsu.edu/eos/users/w/white/www/white/hazardid/hazardid.htm

Appendix 1

Details for Example 3.2 with n=3

The least square problem in (11) can be solved for nonnegative z=H(u)z=(6/5)(data(1)+2data(2))-(16/15)cu1(10-u1)-(22/15)cu2(10-u2)-cu3(10-u3).

Next, consider the system in (11). The first two equations give the first and second unknowns in terms of the third unknownu1=-(3-c10)+(3-c10)2+4cu32candu2=-(3-c10)+(3-c10)2+8cu32c.

Insert the formulas for z,u1, and u2 into the third equation in the system to get a nonlinear problem for a single unknown u3 and the single equation(3-c10)2+4cu3+2(3-c10)2+8cu3=3(3-c10)+2c(data(1)+2data(2)).

When c=8/100, this can be solved by Newton’s method to give u3=7.2526. Then solve for the first two unknowns and finally solve for z u1=2.9748,u2=5.4952andz=10.4764.

In the above example, we were able to explicitly solve for nonnegative z in terms of the components of u. The problem Au=d(H(u))+d^(u) is30-103-2-2-24u1u2u3=cu1(10-u1)cu2(10-u2)D(u1,u2)

whereD(u1,u2)(6/5)(data(1)+2data(2))-(16/15)cu1(10-u1)-(22/15)cu2(10-u2).

The equivalent fixed point problem is u=A-1(d(H(u))+d^(u))u1u2u3=(1/18)8234106669cu1(10-u1)cu2(10-u2)D(u1,u2)G(u).

Expand the matrix-vector product to getG(u)f1(.26)+f2(-.13)+2.8f1(-.13)+f2(.07)+5.6f1(-.20)+f2(-.40)+8.4whereficui(10-ui).

Assume u, v are nonnegative and bounded and inS{u|u=u1u2u3T,0ui10,i=1,2,3}.

Then 0fi25c and since z0,G(u)0. If c1, then G(u)<10 and G aps S into S.

The contraction mapping theorem requires G(u) to be contractive, that is,G(u)-G(v)ru-vwherer<1.

One can approximate the differencesΔficui(10-ui)-cvi(10-vi)Δfic10ui-vi.

Then it is possible to estimate r as a function of c G(u)-G(v)(1/18)8234106669Δf1Δf2D(u1,u2)-D(v1,v2)(1/18)Δf1(8+3(-16/15))+Δf2(2+3(-22/15))Δf1(4+6(-16/15))+Δf2(10+6(-22/15))Δf1(6+9(-16/15))+Δf2(6+9(-22/15))c10(10.8/18)u-v,

Let r=c(108/18)<1 and require c<1/6=0.1667.