1. Introduction
In the field of numerical analysis, a significant role is played by numerical methods for solving nonlinear equations. Due to lack of analytical methods, iterative techniques are required to approximate the solutions. One of the foremost objectives to use numerical methods for solving nonlinear transcendental equations is the ability to handle non-analytic and complex functions. Oftentimes, such equations arise in diverse disciplines such as science, engineering, and applied sciences [
1,
2,
3,
4]. For example, in physics, nonlinear equations often describe the behavior of systems with multiple interacting components, such as the Navier-Stokes equations in fluid dynamics. In engineering, nonlinear equations are used to model the behavior of materials under different loads and conditions. The ability to handle large and complex systems is another essential reason to use numerical methods. Nonlinear equations generally describe the behavior of systems with many interacting components, and solving them analytically can be extremely difficult, if not impossible. Numerical methods provide a way to break down these large systems into smaller, more manageable parts and find approximate solutions using iterative techniques.
A plethora of iterative methods are used for solving nonlinear transcendental equations, including fixed point iteration, root-finding methods, and the Newton–Raphson method. Each method has its own robustness and limitations, and the selection of the method depends on the particular equation being solved and the pre-decided accuracy level. For instance, the bisection method is one of the simplest and most robust methods for finding the roots of an equation but has a disadvantage of being slower and diverging for certain types of functions. The Newton–Raphson method, on the other hand, is faster and more accurate, but it requires the derivative of the function and may not converge for certain types of functions.
Moreover, the numerical method to be chosen depends on the specific equation being solved, the interval of the solutions, the number of solutions, and the desired accuracy level. For example, the bisection method is a good choice for finding all solutions in a given interval, while the Newton–Raphson method is better for finding a specific solution with an initial guess. In numerical optimization, root-finding methods are used to find the solutions of nonlinear equations that describe the behavior of the system, which enables the design of algorithms that are more efficient and more robust. There are several root-finding methods for solving nonlinear transcendental equations in research. Some common methods include:
- 1
The bisection method: a simple yet robust method that involves repeatedly bisecting an interval and determining which subinterval a root lies in.
- 2
The Newton–Raphson method: this method uses an initial guess and an iterative process to converge on a root and requires the ability to compute the derivative of the function.
- 3
The Secant method: this method is similar to the Newton–Raphson method but uses the slope of the secant line between two points rather than the derivative of the function.
- 4
Fixed-point iteration: this method involves finding the fixed point of a function using an iterative process. It requires the function to be in a specific form.
- 5
Muller’s method: this method is an extension of the secant method and is used for complex roots.
- 6
Bairstow’s method: this method is used for finding the roots of polynomials with real coefficients, and it is used to find the roots of polynomials of degree greater than two.
- 7
Aitken’s delta-squared method: this method is used for speeding up the convergence of fixed-point iteration method.
- 8
The Hybrid method: as the name suggests, this method combines two or more methods to find the root of the nonlinear equation.
As a workaround, iterative methods have been developed to locate the initial values of solutions to the nonlinear in the form as follows:
where
F is a Fréchet differentiable operator mapping between a Banach space
into a Banach space
, and
D is a convex and open subset of
. The determination of a solution
of the equation, whose analytical form is rarely attainable, is very important in many disciplines [
1,
2,
3,
4]. This is the case since applications are formulated as an equation such as (
1) using mathematical modeling [
1,
2,
3,
5]. This is the explanation of why iterative methods are introduced producing sequences approximating
. There is extensive literature on the convergence of iterative methods motivated by algebraic or geometrical considerations [
3,
5,
6,
7,
8].
A widely used method to solve (
1) is Newton’s (NM), which is defined for each
by
NM uses one function evaluation and one inverse per iteration. It is of convergence order two [
5]. It is always important to develop iterative methods of a higher convergence order as they provide an efficient approximation and more accuracy in finding the solution. There is a plethora of such methods (see [
9,
10,
11,
12,
13,
14] and references therein) proposed by various researchers.
In particular, we investigate the convergence of the fourth convergence order method defined for each
by
where
,
with
are sequences of nonnegative parameters,
k is a natural number, and
The authors in [
8,
9] motivated by the quadrature formula studied the local convergence of this method utilizing the Taylor series expansion of the operator
F in the special case when
, where
m is a natural number. The benefits over the other methods of the same convergence order were also explained in [
8]. The convergence is established under the differentiability assumptions on
,
. However, these results assure the convergence in case the operator is five times differentiable although the method may converge. Let us look at a simple example in the case when
and
.
Then, one can clearly see that the results in [
8,
9] do not apply since
is unbounded at
. Other problems include:
- (1)
The uniqueness of the solution region is not provided.
- (2)
The choice of the starting point is a “shot in the dark ”.
- (3)
There are no estimates on or that can be computed in advance based on the properties of the operator F.
- (4)
The semilocal convergence of the method has not been studied.
- (5)
The derivative higher than one used in the local convergence is not on the method.
It is worth noticing that the aforementioned problems appear in numerous other methods. These problems motivate the writing of this paper. In particular, we positively address all of these problems utilizing the operators on the method and the very general
-continuity conditions on the operator
[
1,
7]. In the case of the semilocal convergence, the concept of the majorizing sequences is employed [
1,
6,
7]. The idea of this paper can also be applied to other methods [
6,
15,
16,
17] analogously since it only depends on the inverse of the operators
and not on the method itself [
12]. Moreover, see the related papers [
18,
19,
20,
21].
The paper is structured as follows: The local convergence in
Section 2 is followed by the semilocal convergence in
Section 3. The numerical applications and concluding remarks appearing in
Section 4 and
Section 5, respectively, complete the paper.
2. Convergence I: Local
We denote the interval by M for brevity.
Suppose:
The function
has the smallest root
, where
Set and .
Define the function
by
The function
has a smallest root
, where
Then, in Theorem 1 the parameter
r given as
is proven to be a radius of convergence for the method (
3).
Set .
It is implied by these definitions that for each
and
The sets , denote, respectively, the open and closed balls in with center and of radius .
The parameter
r and the functions
and
w are connected to the operator
F as follows, provided that
is a solution of the Equation (
1) with
.
for each .
Set
for each .
and
.
The local convergence of the method (
3) follows next based on the terminology and the conditions
–
.
Theorem 1. Suppose the conditions – hold. Then, the sequence is convergent to provided that the starting point .
Proof. We shall establish using induction the assertions
and
with
r,
, and
as previously defined.
By applying the condition
for
, we obatin, in turn, by (
4) and (
5)
The Banach lemma for invertible linear operators [
1,
2,
3,
16] and the estimate (
11) imply that
with
In particular, if
in (
12) the iterate
is well defined, and we can write by the first substep of the method (
3) if
In view of (
4), (
8) (for
), (
12) (for
), (
) and (
13), we have in turn that
Thus, the iterate
and the assertion (
9) holds for
.
Moreover, the iterate
is well defined by the second substep of the method (
3) if
.
Similarly, we first have
so
hence,
That is, the iterate
and the assertion (
10) holds if
.
By switching
,
,
with
,
,
in the previous calculations, the induction for the assertions (
9) and (
10) is terminated. Therefore, the estimate
where
gives
, and the iterate
. □
The uniqueness of the solution region is determined in the next result.
Proposition 1. Suppose:
- (1)
A solution of the equation exists for some .
- (2)
The condition holds on the ball .
- (3)
exists such that
Set .
Then, the equation (1) is uniquely solvable by in the region .
Proof. Let us define the linear operator
T by
It follows by
–
that
thus
. □
Remark 1. We can choose provided that all hypotheses – of the Theorem 1 hold.
3. Convergence II: Semilocal
We still rely on the -continuity of , but a scalar majorizing sequence is also employed.
Let
,
be NCF’s. If
and
, define the sequences
,
by
and
These scalar sequences are shown to be majorizing for the method (
3). However, first, some general convergence conditions are needed for them.
Lemma 1. Suppose that there exists such that for each Then, the sequences , given by the formula (20) are convergent to some . Proof. The Formula (
20) and Condition (
21) imply
Hence, the result follows. □
Remark 2. - (a)
If the function is strictly increasing on the interval ; then, we can choose .
- (b)
If the smallest positive root of the function exists then we can set .
The functions , v and parameter relate to the operators F and provided is such that and .
Suppose:
for each .
Set , where is the smallest positive root of the function .
for each .
and
.
The semilocal convergence follows for the method (
3).
Theorem 2. Suppose that the conditions – hold. Then, the sequence is convergent to some solving the equation and such that Proof. The following assertions are shown using induction.
and
The assertion (
23) holds if
by the choice of
,
, and the first substep of the method (
3). It follows that the iterate
. By switch
, conditions
–
by
,
–
, we obtain
and
We can write by the second substep of the method (
3)
thus,
and
Hence, the iterate
and (
23) holds. We can write by the first substep of the method (
3) in turn that
thus,
Consequently, we obtain
and
Hence, the induction is completed and the iterate
. It follows by Lemma 1 and the condition
that the sequences
,
are Cauchy as convergent. Then, by (
23) and (
24), the sequences
,
are also Cauchy and, as such, they are convergent to some
. Moreover, by letting
in (
28) and using the continuity of the operator
F, we deduce that
. Furthermore, for
an integer, and the estimation
we conclude that (
22) holds by letting
in (
29). □
Next, the uniqueness region is provided.
Proposition 2. Suppose:
- (1)
There exists a solution of the Equation (1) for some . - (2)
The condition holds on the ball .
- (3)
There exists such that
Set .
Then, the equation is uniquely solvable by in the region .
Proof. As in Proposition 1 define the linear operator
for some
with
. Then, it follows in turn by
Thus, we conclude again that . □
Remark 3. - (i)
Under all the conditions of Theorem 2, we can let and .
- (ii)
The condition can be replaced by , where is given in closed form.
4. Examples and Numerical Calculations
Validating and verifying theoretical results, numerical experiments are essential. This section comprises six numerical problems based on three applied science problems to check the theoretical results obtained from preceding sections. Two types of convergence analysis are mainly focused on: semi-local and local.
In order to evaluate the effectiveness of the method (
3), some applications are simulated, and the results are analyzed. In particular, the residual errors, the number of iterations, the convergence radii, and the expected order of convergence are computed. The following formulas used for COC:
or ACOC by:
We observe that the iterations terminate when the error is sufficiently small, according to the following sopping criterion:
- (i)
and
- (ii)
,
where as error tolerance. The stopping criteria ensure that the computed approximations are accurate to a pre-decided level of precision. The numerical examples are stimulated by using Mathematica 11 software.
The first four examples are based on local convergence. Moreover, on the last we apply the method (
3).
Example 1. Let and define F on D for by The first Fréchet derivative is given by Then, we find that , and . Then, taking , , the smallest positive roots of for are and . Then, the radius of convergence is given as .
Example 2. Let . Define F on by Then, clearly . For , , , and . The smallest positive roots of for are and . Then, the radius of convergence is given as .
Example 3. Consider the nonlinear integral equation of mixed Hammerstein-type equation given bywhere . Clearly, . For , , , , and . The smallest positive roots of for are and . Then, the radius of convergence is given as . Example 4. Consider the function defined on by
The unique solution is . Then, we find that for , , , , and . Then, the smallest positive roots of for are and . Then, the radius of convergence is given as .
Example 5. Consider the following nonlinear partial differential equation, also known as problem of molecular interaction and defined bysubject to the following conditions:where . Discretize the PDE (30) by applying the central divided differencewhich further producesa system of nonlinear equations, where , . For instance , we obtain a system of and . The COC, the number of iterations, residual errors, CPU timing, and error difference between two iterations for Example 5 are mentioned in Table 1. Example 6. Let us consider the following the Van der Pol equation, which is defined aswhich governs the flow of current in a vacuum tube, with the boundary conditions , . Further, we consider the partition of the given interval , which is given by If we discretize the above problem (31) by using the second order divided difference for the first and second derivatives, which are given bythen, we obtain a system of nonlinear equations Let us consider and ; so, we have a system of nonlinear equations. The obtained results are depicted in Table 2. Method (3) converges to the following estimated zero: