524
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Ball convergence of a novel Newton-Traub composition for solving equations

& ORCID Icon | (Reviewing Editor)
Article: 1155333 | Received 19 Oct 2015, Accepted 15 Feb 2016, Published online: 08 Mar 2016

Abstract

We present a local convergence analysis of a Newton-Traub composition method in order to approximate a locally unique solution of a nonlinear equation in a Banach space setting. The method was shown to be of convergence order five, if defined on the m-dimensional Euclidean space using Taylor’s expansion and hypotheses reaching up to the fifth derivative (Hueso, Martinez, & Tervel, 2015; Sharma, 2014). We expand the applicability of this method using contractive techniques and hypotheses only on the first Fréchet-derivative of the operator involved. Moreover, we provide computable radius of convergence and error estimates on the distances involved not given in the earlier studies (Hueso et al., 2015; Sharma, 2014). Numerical examples illustrating the theoretical results are also presented in this study.

AMS Subject classifications:

Public Interest Statement

The most commonly used solution methods for finding zeros of the equation F(x)=0 are iterative-i.e. starting from one or several initial approximations a sequence is constructed that converges to a solution of the equation. In the present paper, we present a local convergence analysis of a Newton-Traub composition method in order to approximate a locally unique solution of a nonlinear equation in a Banach space setting. The method was shown to be of convergence order five, if defined on the m-dimensional Euclidean space using Taylor’s expansion and hypotheses reaching up to the fifth derivative. We expand the applicability of this method using contractive techniques and hypotheses only on the first Fréchet-derivative of the operator involved. Moreover, we provide computable radius of convergence and error estimates on the distances involved not given in the earlier studies

1. Introduction

In this paper, we are concerned with the problem of approximating a solution x of the equation(1.1) F(x)=0,(1.1)

where F is a Fréchet-differentiable operator defined on a convex subset D of a Banach space X with values in a Banach space Y.

Many problems in Computational Sciences and other disciplines can be brought in a form like (Equation 1.1) using mathematical modeling (Argyros, Citation2007; Argyros & Hilout, Citation2013; Kantorovich & Akilov, Citation1982; Petkovic, Neta, Petkovic, & Džunič, Citation2013; Traub, Citation1964). It is known that the solutions of these equations can rarely be found in closed form. So, most solution methods for these equations are iterative. Newton-like iterative methods (Amat, Busquier, & Gutiérrez, Citation2003; Argyros, Citation1985,Citation2004,Citation2007; Argyros & Chen, Citation1993; Argyros & Hilout, Citation2012,Citation2013; Candela & Marquina, Citation1990; Chun, Stanica, & Neta, Citation2011; Gutiérrez & Hernández, Citation1998; Hernández, Citation2001,Citation2000; Hernández & Salanova, Citation2000; Hueso et al., Citation2015; Kantorovich & Akilov, Citation1982; Magreñán, Citation2013; Petkovic et al., Citation2013; Rheinboldt, Citation1978; Sharma, Citation2014; Traub, Citation1964; Wang, Citation2013) are famous for approximating a solution of the equation (Equation 1.1). These methods are usually studied based on: semi-local and local convergence. The semi-local convergence matter is, based on the information around an initial point, to give conditions ensuring the convergence of the iterative procedure; while the local one is, based on the information around a solution, to find estimates of the radii of convergence balls (Amat et al., Citation2003; Argyros, Citation1985,Citation2004,Citation2007; Argyros & Chen, Citation1993; Argyros & Hilout, Citation2012,Citation2013; Candela & Marquina, Citation1990; Chun et al., Citation2011; Gutiérrez & Hernández, Citation1998; Hernández, Citation2001,Citation2000; Hernández & Salanova, Citation2000; Hueso et al., Citation2015; Kantorovich & Akilov, Citation1982; Magreñán, Citation2013; Petkovic et al., Citation2013; Rheinboldt, Citation1978; Sharma, citesharma; Traub, Citation1964; Wang, Citation2013).

In this paper, we study the local convergence analysis of the method defined for each n=0,1,2, by(1.2) yn=xn-αF(xn)-1F(xn),zn=xn-[(1+12α)I-12αF(xn)-1F(yn)]F(xn)-1F(xn),xn+1=zn-[βI+γF(xn)-1F(yn)]F(xn)-1F(zn),(1.2)

where x0 is an initial point, I is the identity operator on X, and α,β,γS, with α0,S=R or C. Method (Equation 1.2) was introduced and studied in Sharma (Citation2014) when X=Y=Rm. Sharma showed using Taylors expansions and hypotheses reaching up to the fifth derivative (in this special case) that method (Equation 1.2) is of convergence order five, if β=1+1α and γ=-1α. The efficiency index of method (Equation 1.2) was compared favorably to other competing methods such as Euler’s, Halley’s, super Halley’s, Chebyshev’s (Amat et al., Citation2003; Argyros, Citation1985,Citation2004; Argyros & Chen, Citation1993; Citation2007,Citation2012,Citation2013; Candela & Marquina, Citation1990; Chun et al., Citation2011; Gutiérrez & Hernández, Citation1998; Hernández, Citation2001,Citation2000; Hernández & Salanova, Citation2000; Hueso et al., Citation2015; Kantorovich & Akilov, Citation1982; Magreñán, Citation2013; Petkovic et al., Citation2013; Rheinboldt, Citation1978; Sharma, Citation2014; Traub, Citation1964; Wang, Citation2013). Notice that method (Equation 1.2) uses four functional evaluations per step. Therefore the efficiency index EI=p1m, where p is the order of convergence and m is the number of functional evaluations is EI=514=1.4953. The local convergence of the preceding methods has been shown under hypotheses up to the fifth derivative (or even higher). These hypotheses restrict the applicability of these methods. The hypotheses on the high order derivatives limit the applicability of these methods. As a motivational example, define function f on D=[-12,52) by(1.3) f(x)=x3lnx2+x5-x4,x00,x=0(1.3)

Choose x=1. We also have thatf(x)=3x2lnx2+5x4-4x3+2x2,f(x)=6xlnx2+20x3+12x2+10x

andf(x)=6lnx2+60x2-24x+22.

Notice that f(x) is unbounded on D. Hence, the results in Sharma (Citation2014), cannot apply to show the convergence of method (Equation 1.2) or its special cases requiring hypotheses on the third derivative of function F or higher. Notice that, in-particular there is a plethora of iterative methods for approximating solutions of nonlinear equations (Amat et al., Citation2003; Argyros, Citation1985,Citation2004,Citation2007; Argyros & Chen, Citation1993; Argyros & Hilout, Citation2012,Citation2013; Candela & Marquina, Citation1990; Chun et al., Citation2011; Gutiérrez & Hernández, Citation1998,Citation2000; Hernández, Citation2001,Citation2000; Hueso et al., Citation2015; Kantorovich & Akilov, Citation1982; Magreñán, Citation2013; Petkovic et al., Citation2013; Rheinboldt, Citation1978; Sharma, Citation2014; Traub, Citation1964; Wang, Citation2013). These results show that if the initial point x0 is sufficiently close to the solution x, then the sequence {xn} converges to x. But how close to the solution x the initial guess x0 should be? These local results give no information on the radius of the convergence ball for the corresponding method.

Moreover, notice that the convergence ball of high convergence order methods is usually very small and in general decreases as the convergence order increases. Our approach establishes the local convergence result under hypotheses only on the first derivative. Our approach can give a larger convergence ball than the earlier studies, under weaker hypotheses. The same technique can be used to other methods.

The rest of the paper is organized as follows. In Section 2, we present the local convergence analysis of method (Equation 1.2). The numerical examples are given in the concluding Section 3.

2. Local convergence

We present the local convergence analysis of method (Equation 1.2) in this section. Let L0>0,L>0,M1, and α,β,γS be given parameters with α0. It is convenient for the local convergence analysis of method (Equation 1.2) that follows to define some scalar functions and parameters. Define functions g1,g2,h2,g3,h3 on the interval [0,1L0) byg1(t)=Lt+2M|1-α|2(1-L0t)g2(t)=12(1-L0t)L+L0M(1+g1(t))|α|(1-L0t)t,h2(t)=g2(t)-1,g3(t)=1+M(L0(|β|+|γ|g1(t))t+|β+γ|)(1-L0t)2g2(t),h3(t)=g3(t)-1

and parameters r1 and rA byr1=2(1-M|1-α|)2L0+L,rA=22L0+L.

Suppose that(2.1) M|1-α|<1.(2.1)

Then, we have by (Equation 2.1) that 0<r1<rA<1L0, 0g1(t)<1 for each t[0,r1). Using the definition of functions g1,g2,h2,g3, and h3 we obtain that h2(0)=h3(0)=-1<0 and h2(t)+,h3(t)+ as t1L0-. Then, it follows from the Intermediate Value Theorem that functions h2 and h3 have zeros in the interval (0,1L0). Denote by r2,r3, respectively, the smallest of such zeros. Set(2.2) r=min{r1,r2,r3}.(2.2)

Then, we have that(2.3) 0<r<rA(2.3)

and for each t[0,r)(2.4) 0g1(t)<1(2.4) (2.5) 0g2(t)<1(2.5)

and(2.6) 0g3(t)<1.(2.6)

Denote by U(v,ρ),U¯(v,ρ) the open and closed ball in X , respectively, with center vX and of radius ρ>0. Next, using the above notation we present the local convergence result for method (Equation 1.2) using the preceding notation.

Theorem 2.1

Let F:DXY be a Fréchet-differentiable operator. Suppose that there exist xD,L0>0,L>0,M1,α,β,γS with α0 such that for each x,yD(2.7) M|1-α|<1,F(x)=0,F(x)-1L(Y,X),(2.7) (2.8) F(x)-1(F(x)-F(x))L0x-x,(2.8) (2.9) F(x)-1(F(x)-F(y))Lx-y,(2.9) (2.10) F(x)-1F(x)M,(2.10)

and(2.11) U¯(x,r)D(2.11)

where the radius r is given by Equation (2.2). Then, the sequence {xn} generated by method (Equation 1.2) for x0U(x,r)-{x} is well defined, remains in U(x,r) for each n=0,1,2,, and converges to x. Moreover, the following estimates hold(2.12) yn-xg1(xn-x0)xn-x<xn-x<r,(2.12) (2.13) zn-xg2(xn-x0)xn-x<xn-x(2.13)

and(2.14) xn+1-xg3(xn-x)xn-x<xn-x,(2.14)

where the “g” functions are defined previously. Furthermore, for T[r,2L0) the limit point x is the only solution of the equation F(x)=0 in U¯(x,T)D.

Proof

The proof uses induction to show estimates (Equations 2.12–2.14). In view of (Equation 2.2), the hypothesis x0U(x,r)-{x} and (Equation 2.8), we get that(2.15) F(x)-1(F(x0)-F(x))L0x0-xL0r<1.(2.15)

It follows from Equation (2.15) and the Banach Lemma on invertible operators (Argyros, Citation2007; Argyros & Hilout, Citation2013; Kantorovich & Akilov, Citation1982) that F(x0)-1L(Y,X) and(2.16) F(x0)-1F(x)11-L0x0-x.(2.16)

Hence, y0 is well defined by the first sub-step of method (Equation 1.2) for n=0. We can write by Equation (2.7) that(2.17) F(x0)=F(x0)-F(x)=01F(x+θ(x0-x))(x0-x)dθ.(2.17)

Notice that x+θ(x0-x)-x=θx0-x<r. Hence we deduce that is x+θ(x0-x)U(x,r). Using Equations (2.10) and (2.17) we get that(2.18) F(x)-1F(x0)=01F(x+θ(x0-x))(x0-x)dθMx0-x.(2.18)

Then, using Equations (2.2, 2.3, 2.7, 2.9, 2.16, 2.17) and the first sub-step of method (Equation 1.2) for n=0 that(2.19) y0-x=(x0-x-F(x0)-1F(x))+(1-α)F(x0)-1F(x0))F(x0)-1F(x)01F(x)-1(F(x+θ(x0-x))-F(x0))(x0-x)dθ+|1-α|F(x0)-1F(x)F(x0)-1F(x0)Lx0-x22(1-L0x0-x)+Mx0-x1-L0x0-xg1(x0-x)x0-x<x0-x<r,(2.19)

which shows (Equation 2.12) for n=0 and y0U(x,r).

The second sub-step of method (Equation 1.2) for n=0 can be written as(2.20) z0-x=x0-x-F(x0)-1F(x0)-12αF(x0)-1F(x)×[F(x)-1(F(x0)-F(x))+F(x)-1(F(x)-F(y0))]×(F(x0)-1F(x))(F(x)-1F(x0))(2.20)

Then, using Equations (2.2, 2.3, 2.8, 2.16, 2.18, 2.19, and 2.20) we have in turn that(2.21) z0-xy0-x+12|α|F(x0)-1F(x)×[F(x))-1(F(x0)-F(x))+F(x)-1(F(y0)-F(x))]×F(x0)-1F(x)]F(x)-1F(x0)Lx0-x21-L0x0-x+L0M(1+g1(x0-x)x0-x22|α|(1-L0x0-x)2=g2(x0-x)x0-x<x0-x<r,(2.21)

which shows (Equation 2.13) for n=0 and z0U(x,r).

Then, we also have as in Equation (2.18) that(2.22) F(x)-1F(z0)Mz0-x,(2.22)

since z0U(x,r). In view of (Equation 2.2, 2.4, 2.8, 2.16, 2.19, and 2.22) we get in turn that(2.23) x1-xz0-x+F(x0)-1F(x)[|β|F(x)-1(F(x0)-F(x))+|γ|F(x)-1(F(y0)-F(x))+|β+γ|F(x)-1F(x)]×F(x0)-1F(x)F(x)-1F(z0)z0-x+[|β|L0x0-x+|γ|L0y0-x+|β+γ|]Mz0-x(1-L0x0-x)2[1+M[L0(|β|+|γ|g1(x0-x))+|β+γ|](1-L0x0-x)2g2(x0-x)x0-x=g3(x0-x)x0-x<x0-x<r,(2.23)

which shows (Equation 2.14) for n=0 and x1U(x,r). By simply replacing x0,y0,z0,x1 by xk,yk,zk,xk+1 in the preceding estimates we arrive at (Equations 2.12–2.14). Using the estimate xk+1-x<xk-x<r, we deduce that limkxk=x and xk+1U(x,r).

Finally, to show the uniqueness part, let Q=01F(y+θ(x-y))dθ for some yU¯(x,T) with F(y)=0. Using Equation (2.8) we get thatF(x)-1(Q-F(x))01L0y+θ(x-y)-xdθL02x-y=L02T<1.

It follows that linear operator Q is invertible. Then, from the identity 0=F(x)-F(y)=Q(x-y), we conclude that x=y.

Remark 2.2

(a) The radius rA was obtained by Argyros in (Citation2004) as the convergence radius for Newton’s method under condition (Equations 2.7–2.9). Notice that the convergence radius for Newton’s method given independently by Rheinboldt (Citation1978) and Traub (Citation1964) is given by(2.24) ρ=23L<rA.(2.24)

As an example, let us consider the function f(x)=ex-1. Then x=0. Set D=U(0,1). Then, we have that L0=e-1<l=e, so ρ=0.24252961<rA=0.324947231.

Moreover, the new error bounds (Argyros, Citation2004,Citation2007; Argyros & Hilout, Citation2012,Citation2013) are:xn+1-xL1-L0xn-xxn-x2,

whereas the old ones (Kantorovich & Akilov, Citation1982; Petkovic et al., petkovic)xn+1-xL1-Lxn-xxn-x2.

Clearly, the new error bounds are more precise, if L0<L. Clearly, we do not expect the radius of convergence of method (Equation 1.2) given by r to be larger than rA (see Equation 2.3) . (b) The local results can be used for projection methods such as Arnoldi’s method, the generalized minimum residual method(GMREM), the generalized conjugate method(GCM) for combined Newton/finite projection methods and in connection to the mesh independence principle in order to develop the cheapest and most efficient mesh refinement strategy (Argyros, Citation2004,Citation2007; Argyros & Hilout, Citation2012,Citation2013). (c) The results can be also be used to solve equations where the operator F satisfies the autonomous differential equation (Argyros, Citation2007; Argyros & Hilout, Citation2013; Kantorovich & Akilov, Citation1982; Petkovic et al., Citation2013):F(x)=P(F(x)),

where P is a known continuous operator. Since F(x)=P(F(x))=P(0), we can apply the results without actually knowing the solution x. Let as an example F(x)=ex-1. Then, we can choose P(x)=x+1 and x=0. (d) It is worth noticing that method (Equation 1.2) are not changing if we use the new instead of the old conditions (Hueso et al., Citation2015; Sharma, Citation2014). Moreover, for the error bounds in practice we can use the computational order of convergence (COC)ξ=lnxn+2-xn+1xn+1-xnlnxn+1-xnxn-xn-1,for eachn=1,2,

or the approximate computational order of convergence (ACOC)ξ=lnxn+2-xxn+1-xlnxn+1-xxn-x,for eachn=0,1,2,

instead of the error bounds obtained in Theorem 2.1. (e) In view of (Equation 2.8) and the estimate|F(x)-1F(x)|=|F(x)-1(F(x)-F(x))+I|1+|F(x)-1(F(x)-F(x))|1+L0|x-x|

condition (Equation 2.10) can be dropped and M can be replaced byM(t)=1+L0t

orM(t)=M=2,

since t[0,1L0).

3. Numerical examples

We present numerical examples in this section. We have taken α=0.75,β=1+1α=2.3333 and γ=-1α=-1.333 in our computations.

Example 3.1

Let X=Y=R3,D=U¯(0,1). Define F on D for v=(x,y,z)T by(3.1) F(v)=(ex-1,e-12y2+y,z)T.(3.1)

Then, the Fréchet-derivative is given byF(v)=ex000(e-1)y+10001.

Notice that x=(0,0,0),F(x)=F(x)-1=diag{1,1,1},L0=e-1<L=e,M=2. Then, the parameters arer1=0.16247,rA=0.32495,r2=0.16077,r3=0.08740=r.

Example 3.2

Let X=Y=C[0,1], the space of continuous functions defined on [0, 1] be and equipped with the max norm. Let D=U¯(0,1). Define function F on D by(3.2) F(φ)(x)=φ(x)-501xτφ(τ)3dτ.(3.2)

We have thatF(φ(ξ))(x)=ξ(x)-1501xτφ(τ)2ξ(τ)dτ,foreachξD.

Then, we get that x=0,L0=7.5,L=15,M=2. Then, the parameters arer1=0.033333,rA=0.066667,r2=0.03464,r3=0.01881=r.

Example 3.3

Returning back to the motivation example at the introduction on this paper, we have L=L0=146.6629073,M=2. Then, the parameters arer1=0.0022728,rA=0.0045456,r2=0.0020708,r3=0.001129=r.

Additional information

Funding

The authors received no direct funding for this research.

Notes on contributors

Ioannis K. Argyros

Ioannis K. Argyros was born in 1956 in Athens, Greece. He received a BSc from the University of Athens, Greece; and a MSc and PhD from the University of Georgia, Athens, Georgia, USA, under the supervision of Douglas N Clark. Argyros is currently a full professor of Mathematics at Cameron University, Lawton, OK, USA. He published more than 800 papers and 17 books/ monographs in his area of research, computational mathematics. He is also the editor of 20 peer-reviewed research journals in Mathematics.

Santhosh George

Santhosh George was born in 1967 in Kerala, India. He received his PhD degree in Mathematics from Goa University, under the supervision of M. T. Nair. George is a Professor of Mathematics at National Institute of Technology, Karnataka. Five students have completed their PhD under his guidance. He has many International journal and conference papers to his credit

References

  • Amat, S., Busquier, S., & Gutiérrez, J. M. (2003). Geometric constructions of iterative functions to solve nonlinear equations. Journal of Computational and Applied Mathematics, 157, 197–205.
  • Argyros, I. K. (1985). Quadratic equations and applications to Chandrasekhar’s and related equations. Bulletin of the Australian Mathematical Society, 32, 275–292.
  • Argyros, I. K. (2004). A unifying local-semilocal convergence analysis and applications for two-point Newton-like methods in Banach space. Journal of Mathematical Analysis and Applications, 298, 374–397.
  • Argyros, I. K. (2007). In C. K. Chui & L. Wuytack (Eds.), Computational theory of iterative methods ( Studies in Computational Mathematics, p. 15). New York: Elsevier.
  • Argyros, I. K., & Chen, D. (1993). Results on the Chebyshev method in Banach spaces. Proyecciones, 12, 119–128.
  • Argyros, I. K., & Hilout, S. (2012). Weaker conditions for the convergence of Newton’s method. Journal of Complexity, 28, 364–387.
  • Argyros, I. K., & Hilout, S. (2013). Numerical methods in nonlinear analysis. Hackensack, NJ: World Scientific.
  • Candela, V., & Marquina, A. (1990). Recurrence relations for rational cubic methods II: The Chebyshev method. Computing, 45, 355–367.
  • Chun, C., Stanica, P., & Neta, B. (2011). Third order family of methods in Banach spaces. Computers and Mathematics with Applications, 61, 1665–1675.
  • Guti\’errez, J. M., & Hern\’andez, M. A. (1998). Recurrence relations for the super-Halley method. Computers & Mathematics with Applications, 36, 1–8.
  • Hern\’{a}ndez, M. A. (2001). Chebyshev’s approximation algorithms and applications. Computers & Mathematics with Applications, 41, 433–455.
  • Hern{\’a}ndez, M. A. (2000). Second-derivative-free variant of the Chebyshev method for nonlinear equations. Journal of Optimization Theory and Applications, 104, 501–515.
  • Hernández, M. A. & Salanova, M. A. (2000). Modification of the Kantorovich assumptions for semilocal convergence of the Chebyshev method. Journal of Computational and Applied Mathematics, 126, 131–143.
  • Hueso, J. L., Martinez, E., & Tervel, C. (2015). Convergence, efficiency and dynemics of new fourth and sixth order families of iterative methods for nonlinear systems. Journal of Computational and Applied Mathematics, 275, 412–420.
  • Kantorovich, L. V., & Akilov, G. P. (1982). Functional analysis. Oxford: Pergamon Press.
  • Magre{\~n}{\’a}n, {\’A}. A. (2013). Estudio de la dinámica del método de Newton amortiguado (PhD Thesis). Servicio de Publicaciones, Universidad de La Rioja. Retrieved from http://dialnet.unirioja.es/servlet/tesis?codigo=38821
  • Petkovic, M. S., Neta, B., Petkovic, L., & D\v{z}uni\v{c}, J. (2013). Multipoint methods for solving nonlinear equations. Waltham, MA: Elsevier.
  • Rheinboldt, W. C. (1978). An adaptive continuation process for solving systems of nonlinear equations. In A. N. Tikhonov, et al. (Ed.), Mathematical models and numerical methods (Vol. 3, pp. 129–142). Warsaw: Banach Center.
  • Sharma, J. R. (2014). Some fifth and sixth order iterative methods for solving nonlinear equations. International Journal of Engineering Research and Applications, 4, 268–273.
  • Traub, J. F. (1964). Iterative methods for the solution of equations (Series in Automatic Computation). Englewood Cliffs, NJ: Prentice- Hall.
  • Wang, X., & Kou, J. (2013). Semilocal convergence and R-order for modified Chebyshev--Halley methods. Numerical Algorithms, 64, 105–126.