1. Introduction
There exist in the literature (see, for example, References [
1,
2,
3,
4,
5,
6,
7,
8]) numerous iterative methods without memory, involving or not derivatives, designed to estimate the multiple roots of a nonlinear equation
, but most of them need the knowledge of the multiplicity
m of these roots.
It is well-known that Schröder method [
9]:
is able to converge quadratically to the multiple solution of a nonlinear equation, that is, a value
such that
and
,
, with
m being the multiplicity of the root. This scheme was originally deduced from Newton’s scheme applied on the quotient
, and it is denoted along this manuscript by SM1,
Notice that SM1 requires 3 function evaluations per step. Similarly, the derivative-free Traub-Steffensen’s method for
g,
with
being a real parameter, requires 4 function evaluations per step and is no longer derivative-free. This Traub-Steffensen method on
g is too expensive and is not further considered.
The main advantage of Schröder scheme is its independence of the knowledge of the multiplicity of the nonlinear function, in contrast with the modified Newton’s method for multiple roots,
where
m is the multiplicity of
, that must be known in this case. This scheme was also due to Schröder (also see Reference [
9]), and we denote it by SM2. This scheme is second-order convergent and, therefore, optimal, in the sense of Kung-Traub conjecture, (as it uses two new functional evaluations per iteration; see Reference [
10]). However, it needs the knowledge of the multiplicity, while SM1 does not use it; nevertheless, the main drawback of SM1 scheme is its low efficiency, as it needs to evaluate three nonlinear functions (
,
and
) per iteration.
Our aim in this manuscript is double: from one side, we would like to increase the efficiency of SM1 scheme, holding its ability for finding multiple roots of multiplicity m without knowing m and, from other side, to combine in the same algorithm the capability to find multiple roots with the use of more than one previous iterate. So, we propose an iterative scheme with memory for estimating multiple roots of unknown multiplicity. As far as we know, there is in the literature no iterative procedure satisfying these properties.
In the analysis of the convergence of the proposed scheme, some aspects must be taken into account, as it is an iterative method with memory, so that the error in several previous iterations must be considered and the multiplicity of the root m should also be a key element of the demonstration, although its specific value is not known. Regarding this fact, it should be noticed that for and . So, the Taylor expansions around of f and appearing in the iterative expression should take this information into account.
On the other hand, as our proposed scheme is an iterative procedure that uses three previous iterates for calculating the next one, it is necessary to express the error equation in terms of their corresponding errors and, from it, to deduce its order of convergence. This is made by means of a classical result by Ortega and Rheinboldt [
11], that is presented below.
Theorem 1. Let ψ be an iterative method with memory that generates a sequence of approximations to the root α, and let this sequence converges to α. If there exist a nonzero constant η and positive numbers , , such that the inequalityholds, then the R-order of convergence of the iterative method ψ satisfies the inequalitywhere p is the unique positive root of the equation In this manuscript,
Section 2 is devoted to the design and convergence analysis of the proposed derivative-free iterative method with memory to find multiple roots (without the knowledge of its multiplicity). In
Section 3, its stability is analyzed in order to deduce its dependence on the initial estimations for both simple and multiple roots. In
Section 4, the numerical performance of the method is checked on several test functions, being analyzed, as well as their corresponding basins of attraction, in comparison with existing Schröder methods.
2. Design and Convergence Analysis
Our starting point is the derivative-free scheme with memory due to Traub [
12],
with
,
and
being their initial estimations, with order of convergence
. To estimate the multiple roots of
, we define the auxiliary function
, and then we apply Traub’s method (
1), on
, getting what we call gTM method,
an iterative scheme with memory that is proven to converge to any multiple roots of
f with the same order than original Traub’s scheme, without the knowledge of its multiplicity
m and using two new functional evaluations per iteration.
Theorem 2. Let be an analytic function in the neighborhood of the multiple zero α of f with unknown multiplicity Then, for an initial guess sufficiently close to α, the iteration function gTM defined in (2) has order of convergence , with its error equation beingwhere , and denotes the terms of the error equation with products of powers of , and , whose exponents sum at least 3. Proof. Let
be a multiple zero of
, and
be the error at
-iterate. Expanding
and
of
by using the Taylor’s series expansion, we have
and
where
By using expressions (
3) and (
4), we have
Then,
where
denotes that the neglected terms of the error equation have products of powers of
and
, whose exponents sum at least 3.
and
Therefore, from expression (
2),
and then the order of convergence is the only real root of polynomial
, that is,
, by applying Theorem 1. So, the proof is finished. □
The main advantage of this scheme is its ability for finding simple, as well as multiple, roots of a nonlinear function without the knowledge of the multiplicity, with better efficiency as SM1. Certainly, by using Ostrowski’s efficiency index [
13],
is lower than
, where each index
I is calculated as
, with
p being the order of convergence of the method, and
d the amount of new functional evaluations per iteration.
In the next section, a dynamical analysis is made on this scheme, in order to show its qualitative performance on simple and multiple roots. As it is an iterative method with memory, multidimensional real dynamics must be used.
3. Qualitative Study of the Proposed Iterative Methods with Memory for Multiple Roots
Let us remark that our method uses three previous iterations in order to generate the following one; therefore, it can be expressed in general as
where
,
and
are the initial estimations. By means of the procedure defined in Reference [
14], this method can be described as a discrete real multidimensional dynamical system, and its qualitative behavior can be analyzed.
The qualitative performance of the dynamical system has a key element in the characterization of their fixed points, in terms of stability. In order to calculate the fixed points of
, an auxiliary vectorial function
can be defined, related to
by means of:
Therefore, a fixed point of M is obtained if not only , but also and .
Fixed points
of
M satisfy
and
. So,
,
and
. In the following, we define some basic dynamical concepts as direct extension of those used in complex discrete dynamics analysis (see Reference [
15]).
Let us consider a vectorial rational function obtained by the application of an iterative method on a scalar polynomial . Then, if a fixed point of operator M is different from , where r is a zero of , then the fixed point is called strange. Moreover, the orbit of a point is the set of successive images from by the vector function, that is, . Indeed, a point is called periodic with period p if and , for . We should notice that a fixed point is a 1-periodic point.
It is also known that the qualitative performance of a fixed point of
M is classified in terms of its asymptotical behavior. It can be analyzed by means of the Jacobian matrix
, as is stated in the next result (see, for instance, Reference [
16]).
Theorem 3. Let M from to be of class . Let us also assume that is a k-periodic point. Let be the eigenvalues of the Jacobian matrix at periodic point . Then, it holds that:
- (a)
If all the eigenvalues verify , then is attracting.
- (b)
If one eigenvalue verifies , then is unstable, that is, repelling or saddle.
- (c)
If all the eigenvalues verify , then is repelling.
Moreover, if there exist an eigenvalue
of the Jacobian matrix
evaluated at a fixed point
satisfying
and another one
such that
, then,
is called saddle fixed point. As an extension of the concept in one-dimensional dynamics, if the eigenvalues of
satisfy
for all values of
, then, the fixed point
is not only attracting but also superattracting. Therefore, the method has quadratic convergence, at least on the class of nonlinear functions that derive the rational function (see Reference [
12]).
By considering
an attracting fixed point of
M, its basin of attraction
is defined as the set of preimages of any order
The qualitative performance of different iterative schemes designed for solving nonlinear equations with multiple roots has been studied by different authors (see, for example, References [
17,
18,
19]). It has been made by using discrete complex dynamics, as all these schemes are without memory. In these studies, it has been obtained that, when an iterative method (without memory) designed for finding multiple roots acts on a nonlinear function with both simple and multiple roots, it is quite usual that the basins of attraction of simple roots are narrower than those of multiple roots. Indeed, it may happen that those simple roots define fixed points of the rational function that are repulsive. Therefore, the iterative method should be able to find only multiple roots.
The following qualitative analysis is made on , , so that the capability of the scheme to find both simple and multiple roots (with multiplicity m) is tested.
Theorem 4. The multidimensional rational operator associated with method gTM, when it is applied on polynomial ,Indeed, the rational operator satisfies that there are one strange fixed point with equal components to , that is saddle. Moreover, both fixed points corresponding to the roots of are superattracting. Proof. By definition of the multidimensional dynamical system,
with
, where
. Therefore,
is obtained. In order to calculate the fixed points of
, equation
must be solved. By means of algebraic manipulations, it is reduced to
and
So, the only fixed points are the roots of
and the strange fixed point
that depends on the multiplicity of the root. In order to analyze the stability of the fixed points, we calculate the Jacobian matrix
where
and
where
,
, and
.
The eigenvalues of , when or , are all equal to zero. Then, these are superattracting fixed points.
Regarding the strange fixed point , in order to avoid an indetermination, it is necessary to calculate a simplified rational operator, forcing to ; therefore, the reduced Jacobian matrix has two zero eigenvalues, and the third one is . So, the strange fixed point is always a saddle point and, therefore, is in the boundary of the basins of attraction. □
A very useful tool to visualize the analytical results is the dynamical plane of the system, composed by the set of the different basins of attraction. Here, the dynamical plane of the proposed method gTM is built by calculating the orbit of a mesh of starting points for a fixed value of w in the starting grid. As the iterative schemes needs to be started with three initial estimations, we generate a mesh of dynamical planes, each one of them with a fixed value of w in the interval . In these phase portraits, each point of the mesh is painted in different colors (orange and green in this case), depending on the attractor they converge to (marked as a white star), with a tolerance of . In addition, they appear in black color if the orbit has not reached any attracting fixed point in a maximum of 500 iterations. As the fixed value of w is changed in a vector of values belonging to , it yields to a composition of figures for each multiplicity, giving rise to a kind of contour plots.
In
Figure 1, we show the performance of gTM scheme on
, that is, of rational operator
for simple roots. Observing the behavior for the different plots with the three first iterations varying each in
, the stable feasibility is noticed. The basins of attraction of the roots are the only ones; they are wide, and the only different performance (better than others in terms of the simplicity of the boundary among the basins) is the case
, where the rational function is simplified. In all cases, it is observed that the only possible behavior of method
is the convergence to the roots.
On the other hand, in
Figure 2, we show a very similar performance when one of the roots is double, and the other one is simple. The basins of attraction are equally wide, and this behavior is very similar when other multiplicities have been explored. In addition, in this case can be seen that there is only convergence to the roots, as darker areas are only slower convergence, due to the higher complexity of the boundary of the basins of attraction.
In the next section, the numerical and dynamical performance of our proposed scheme is tested on several nonlinear functions of increasing complexity.
4. Numerical Performance and Dynamical Tests
In this section, we compare three methods, namely SM2 (requiring the knowledge of the multiplicity), SM1, and gTM (derived from Traub’s method). The last two methods do not require the knowledge of the multiplicity, but they do require extra functional evaluations per iteration step (three in case of SM1, two in gTM case).
The methods are compared both qualitatively via the basins of attraction figures and quantitatively via several measures. These measures are: the CPU run-time to run the method on points in a 6 by 6 square centered in the origin. We divided the square by uniformly distributed horizontal and vertical lines and took all points of intersection as initial points for the iterative process. For gTM, a method with memory, we had to take two additional starting points and , where d is the spacing of the lines. Another criterion collected by the code is the average number of iterations per point (AIPP), but, since the methods require different number of functional evaluations per step, we took the average number of functions per point (AFPP). The third criterion is the number of divergent points (DP), which is the number of points for which the method did not converge in 40 iterations using a tolerance of .
The functions used for our comparative study are:
,
,
,
,
,
.
Notice that all but one are polynomials of various degrees and various multiplicities.
In
Figure 3,
Figure 4,
Figure 5,
Figure 6,
Figure 7 and
Figure 8, we have plotted the basins of attraction for the 3 methods for each test function. Each figure has 3 sub-figures, with the left-most being the Schröder method using the multiplicity (SM2), the middle one Schröder method not requiring the knowledge of multiplicity (SM1), and the right-most Traub’s method for multiple roots (gTM).
Based on
Figure 3, it is clear that SM1 and SM2 have similar basins, and gTM has more lobes on the boundary between the two basins. From
Figure 4, we notice that gTM is better than SM1. In the next 3 figures, gTM is best, with wider basins of attraction and narrower black areas of no convergence to the roots. This performance is held even for non-polynomial function
. Moreover, in
Figure 8, it can be noticed that the basins of attraction of method SM2 are wider than our gTM method.
We now refer to the data in
Table 1,
Table 2 and
Table 3. The CPU run-time in seconds is given in
Table 2. It is clear that SM2 is consistently faster than the others. If the multiplicity is not known, then gTM is faster than SM1, except for the first example. On average, gTM is faster than SM1.
The average number of function evaluations per point (see
Table 2) is the highest for SM1 for all examples. Note that the last example is the hardest for all methods. The number of divergent points is the lowest for gTM for examples 1, 3, and 4. SM1 has the most divergent points for the first 6 examples, but, on the last example, gTM performed poorly and became third place overall. The method SM2 was best, on average, for the 3 categories followed by gTM for 2 categories.