1. Introduction
In this work, we generate a sequence
for approximating a locally unique solution
of the nonlinear equation
where
F is a Fréchet-differentiable operator defined on a closed convex subset
D of Banach space
with values in a Banach space
. In computational sciences, many problems can be written in the form (
1). See, for example [
1,
2]. The solutions of such equations are rarely attainable in closed form. This shows why most methods for solving these equations are usually iterative in nature. The important part in the construction of an iterative method is to study its convergence analysis. In general, the convergence domain is small. Therefore, it is important to enlarge the convergence domain without using extra hypotheses. Knowledge of the radius of convergence is useful because it gives us the degree of difficulty for obtaining initial points. Another important problem is to find more precise error estimates on
or
. Many authors have studied convergence analysis of iterative methods, see, for example [
1,
2,
3,
4,
5,
6,
7].
The most widely used iterative method for solving (
1) is the quadratically convergent Newton’s method
where
is the inverse of first Fréchet derivative
of the function
In order to accelerate the convergence, researchers have also obtained modified Newton’s or Newton-like methods (see [
4,
6,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17]) and references therein.
There are numerous higher order iterative methods for solving a scalar equation
(see, for example [
2]. Contrary to this fact, higher order methods are rare for multi-dimensional cases, that is, for approximating the solution of
. One possible reason is that the construction of higher order methods for solving systems is a difficult task. Another factual reason is that not every method developed for single equations can be generalized to solve systems of nonlinear equations. Recently, a family of optimal eighth order methods for solving a scalar equation
has been proposed in [
16], which is given by
where
is any optimal fourth order scheme with the base as Newton’s iteration
and
is Newton’s first order divided difference. In particular, they have considered the following optimal fourth order schemes in the second step of (
3):
Ostrowski method (see [
12]):
Ostrowski-like method (see [
12]):
Kung-Traub method (see [
15]):
Motivated by the above methods defined on the real line, we propose the methods that follow but for Banach space valued operators. It can be observed that the above family of eighth order methods can be easily extendable for solving (
1). In view of this, here we study the method (
3) in Banach space. The iterative methods corresponding to the fourth order schemes (
4)–(
6) in the Banach space setting are given as
and
In above each case, we have that
Here
is a first order divided difference on
satisfying
for
and
if
F is differentiable, where
stands for the space of bounded linear operators from
into
. Methods (
7)–(
9) require four inverses and four function evaluations at each step.
The rest of the paper is summarized as follows. In
Section 2, the local convergence, including radius of convergence, computable error bounds and uniqueness results of the proposed methods, is presented. In order to verify the theoretical results of convergence analysis, some numerical examples are presented in
Section 3. Finally, the methods are applied to solve systems of nonlinear equations in
Section 4.
2. Local Convergence
Local convergence analysis of the methods (
7)–(
9) is presented by using some real functions and parameters. Let
be a continuous and increasing function satisfying
. Suppose that equation
has positive solutions. Denote by
the smallest such solution. Let
,
,
and
also be continuous and increasing functions satisfying
Define functions
and
on the interval
by
We have that
and
as
. By applying the Bolzano’s theorem on function
, we deduce that equation
has solutions in the interval
. Let
be the smallest such zero.
Moreover, define function
p and
on the interval
by
and
We get
and
as
Let
be the smallest solution of equation
in the interval
Furthermore, define functions
and
on the interval
by
and
We obtain
and
as
. Let
be the smallest solution of equation
in the interval
. Define functions
q and
on the interval
and functions
and
on the interval
, respectively by
We get
and
as
,
as
. Let
,
be the smallest solutions of equations
,
in the intervals
,
, respectively. Finally, define functions
and
on the interval
by
and
where
We have that
and
as
. Let
be the smallest solution of equation
in the interval
Set
to be the radius of convergence for method (
7). Then, for each
, it follows that
and
Let and stand, respectively for the open and closed balls in with center and of radius .
The local convergence analysis of method (
7), method (
8) and method (
9) is based on the conditions (A):
- (a1)
is continuously Fréchet differentiable and
D is a convex set. The operator
is a divided difference of order one satisfying
and
- (a2)
There exists such that and
- (a3)
There exists function
continuous and increasing with
such that for each
Set
, where
is given in (
11).
- (a4)
There exist continuous and increasing functions
,
,
and
such that for each
and
- (a5)
where
r is given in (
12) for method (
7), by (
30) for method (
8) and by (
31) for method (
9).
- (a6)
There exists
such that
Set
Next, we first present the local convergence analysis of method (
7) based on the conditions (A).
Theorem 1. Assume that the conditions (A) hold. Then, sequence generated for by method (7) is well defined in , remains in for each and converges to α so thatandwhere the functions are defined previously. Moreover, the solution α of equation is unique in . Proof. We shall show assertions (
17)–(
19) using mathematical induction. Let
Then, using
and (
12), we have that
By the Banach perturbation Lemma [
2] and (
20), we deduce that
and
In particular for
,
is well defined by the first substep of method (
7) and (
21) holds for
, since
We get by the first substep of method (
7) for
,
,
, (
13) (for
) and (
12) that
so (
17) holds for
and
We must show the existence of
which shall imply that
is well defined. Using (
12), (
14),
and
, we get in turn that
so
exists and
We can write
Notice that
for each
. Using
and (
24), we get
Then, by (
12), (
13) (for
), (
21), (
22), (
23), (
25) and the second substep of method (
7), we obtain in turn that
which shows (
18) for
and
We must show the existence of
which shall also imply that
is well defined. Using (
12), (
15) and
, we obtain in turn that
so
exists and
Then, using the last substep of method (
7), (
10), (
12), (
13) (for
), (
18), (
23), (
26) and (
27), we get in turn that
which shows (
19) for
and
The induction is completed if
,
replace
,
,
,
in the preceding estimates, respectively. Then, from the estimate
where
, we deduce that
and
Let
for some
such that
. By
and
, we have in turn that
implies that
exists. Then, from the identity
, we conclude that
. □
Next, we shall show the local convergence of method (
8) in an analogous way but functions
,
,
shall be replaced by
,
,
and which are given by
We shall use the same notation for
as in (
12) but notice that
and
correspond to the smallest positive solutions of equations
and
, respectively. Set
The local convergence analysis of method (
8) is given by the following theorem:
Theorem 2. Assume that the conditions hold. Then, the conclusions of Theorem 1 also hold for method (8) with functions , and replacing , and r, respectively. Proof. We have that
as in Theorem 1 and using the second and third substep of method (
8) we get (as in Theorem 1) that
and
□
We define
Denote by
,
, the smallest positive solutions of equations
and
. Set
Then, we have:
Theorem 3. Assume that the conditions hold. Then, the conclusions of Theorem 1 also hold for method (9) with functions , and replacing , and r, respectively. Proof. Notice that from the second and third substep of method (
9) we obtain
and
□
Remark 1. Methods (7)–(9) are not effected, when we use the conditions of the Theorems 1–3 instead of stronger conditions used in ([16], Theorem 1). Moreover, we can compute the computational order of convergence (COC) [18] defined byor the approximate computational order of convergence (ACOC) [9], given byIn this way, we obtain in practice the order of convergence. 4. Applications
Lastly, we apply the methods (
7)–(
9) to solve systems of nonlinear equations in
. The performance is also compared with some existing methods. For example, we choose Newton method (NM), sixth-order methods proposed by Grau et al. [
12] and Sharma and Arora [
15], and eighth-order Triple-Newton Method [
14]. These methods are given as follows:
Grau-Grau-Noguera method:
This method requires two inverses and three function evaluations.
Grau-Grau-Noguera method:
It requires two inverses and three function evaluations.
Sharma-Arora Method:
The method requires one inverse and three function evaluations.
Triple-Newton Method:
This method requires three inverses and three function evaluations.
Example 4. Let us consider the system of nonlinear equations:with initial value towards the required solution of the systems for . Example 5. Next, consider the extended Freudenstein and Roth function [19]:wherewith initial value towards the required solution of the systems for . Computations are performed in the programming package
Mathematica using multiple-precision arithmetic. For every method, we record the number of iterations
needed to converge to the solution such that the stopping criterion
is satisfied. In order to verify the theoretical order of convergence, we calculate the approximate computational order of convergence (ACOC) using the formula (
33). For the computation of divided difference we use the formula (see [
12])
Numerical results are displayed in
Table 4 and
Table 5, which include:
The dimension of the system of equations.
The required number of iterations .
The value of of approximation to the corresponding solution of considered problems, wherein denotes .
The approximate computational order of convergence (ACOC).
From the numerical results shown in
Table 4 and
Table 5 it is clear that the methods possess stable convergence behavior. Moreover, the small values of
, in comparison to the other methods, show the accurate behavior of the presented methods. The computational order of convergence also supports the theoretical order of convergence. Similar numerical tests, carried out for a number of other different problems, confirmed the above conclusions to a large extent.