Next Article in Journal
Modified Remora Optimization Algorithm with Multistrategies for Global Optimization Problem
Previous Article in Journal
The Dynamics of a Continuous Newton-like Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extended Multi-Step Jarratt-like Schemes of High Order for Equations and Systems

1
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505, USA
2
Department of Computing and Technology, Cameron University, Lawton, OK 73505, USA
3
Department of Computer Science, University of Oklahoma, Norman, OK 73071, USA
4
Facultad de Ingeniería y Ciencias Aplicadas, Universidad de Las Américas, Quito 170124, Ecuador
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(19), 3603; https://doi.org/10.3390/math10193603
Submission received: 4 September 2022 / Revised: 15 September 2022 / Accepted: 18 September 2022 / Published: 2 October 2022
(This article belongs to the Section Computational and Applied Mathematics)

Abstract

:
The local convergence analysis of multi-step, high-order Jarratt-like schemes is extended for solving Banach space valued systems of equations using the derivative instead of up to the ninth derivative as in previous works. Our idea expands the usage of the scheme in cases not considered earlier and can also be utilized in other schemes, too. Experiments test the theoretical results.

1. Introduction

Let Ω 0 be an open subset of a Banach X and H a continuous operator mapping Ω into Y, another Banach space.
Numerous problems from diverse disciplines turn into an equation
H ( x ) = 0 ,
by mathematical modeling [1,2,3,4,5,6,7,8,9,10,11].
We want to find a solution γ in closed form, but this is achieved just in special occasions. This is why iterative schemes are developed generating sequences converging to γ under suitable convergence criteria [2,4,12,13,14,15,16,17,18].
We study the multi-step Jarratt-like scheme developed for x 0 Ω and all n = 0 , 1 , 2 , and for k = 1 , 2 , m 3 , ( m 3 ) by
z 1 = H ( x n ) 1 H ( x n ) , y 1 = x n z 1 , z 2 = H ( x n ) 1 H ( y 1 ) , y 2 = y 1 3 z 2 , z 3 = H ( x n ) 1 H ( y 2 ) z 2 , y 3 = y 1 7 4 z 2 + 1 2 z 3 + 1 4 z 4 , z 4 = H ( x n ) 1 H ( y 2 ) z 3 , z 2 k + 3 = H ( x n ) 1 H ( y k + 2 ) , z 2 k + 4 = H ( x n ) 1 H ( y 2 ) z 2 k + 3 , y k + 3 = y k + 2 2 z 2 k + 3 + z 2 k + 4 .
The 3 m 4 convergence order of (2) is presented in [19] for X = Y = R j from Taylor series expansions with conditions reaching the ninth derivative of H (not appearing in scheme (2)). These restrict the application of scheme (2). Consider the simple illustration for f : Ω : = [ 1 / 2 , 3 / 2 ] R defined by
f ( s ) = s 3 ln ( s 2 ) + s 5 s 3 , s 0 0 , s = 0 .
Then, it is easily seen that f is unbounded. Therefore, the work in [19] cannot guarantee convergence to a solution γ using scheme (2). Notice that the method (2) is of order at least eight if m = 4 , which is optimum in the sense of the Kung–Traub conjecture [20]. Moreover, each step increases the convergence order by three [19]. The computational efficiency and computational cost of the method (2) have been given and discussed in detail (see Section 4 in [19]). It is also shown that it is better from the efficiency point of view when compared with other modified Jarratt-type methods. We also refer the reader to the notable relevant papers by Ignatova et al. [21], Kung et al. [20] and Petkovic [17]. Moreover, no upper bounds on x n γ or results on the uniqueness of γ are given. Motivated by all these, we develop a technique using only H (that appears in (2)), which also gives computable upper bounds on x n γ and a uniqueness result.
Hence, we extend the applicability of scheme (2). Moreover, the Computational Order of Convergence (COC) and the Approximate Computational Order of Convergence (ACOC) are utilized to compute the convergence order, which does not require usage of higher derivatives or divided differences. This is conducted in Section 2. Numerical experiments are given in Section 3. Finally, the conclusion appears in Section 4.

2. Analysis

We develop the real functions ψ 0 , ψ and ψ 1 and real indicators. Next, these functions are connected as majorizing for the operator H (see the conditions ( C 1 ) ( C 4 ) that follow). Set T = [ 0 , + ) .
Suppose that there exist functions:
(a)
ψ 0 : T T so
ψ 0 ( s ) 1 = 0
has a solution denoted by ρ 0 T { 0 } , which is continuous and nondecreasing (CN). Set T 0 = [ 0 , ρ 0 ) .
(b)
ψ : T 0 T , ψ 1 : T 0 T CN so that for
h 1 ( s ) = 0 1 ψ ( ( 1 ζ ) s ) d ζ 1 ψ 0 ( s ) ,
h 2 ( s ) = [ h 1 ( h 1 ( s ) s ) + ( ψ 0 ( h 1 ( s ) s ) + ψ 0 ( s ) ) 0 1 ψ 1 ( ζ h 1 ( s ) s ) d ζ ( 1 ψ 0 ( s ) ) ( 1 ψ 0 ( h 1 ( s ) s ) ) + 2 0 1 ψ 1 ( ζ h 1 ( s ) s ) d ζ 1 ψ 0 ( s ) ] h 1 ( s ) ,
h 3 ( s ) = [ 1 + 7 4 0 1 ψ 1 ( ζ h 1 ( s ) s ) d ζ 1 ψ 0 ( s ) + 1 2 ψ 1 ( h 1 ( s ) s ) h 2 ( s ) s ) ( 1 ψ 0 ( s ) ) 2 + 1 4 ψ 1 ( h 1 ( s ) s ) ψ 1 ( h 2 ( s ) s ) 2 ( 1 ψ 0 ( s ) ) 3 ] h 1 ( s ) ,
for k = 1 , , n 3
h k + 3 ( s ) = [ 1 + 2 0 1 ψ 1 ( ζ h k + 2 ( s ) s ) d ζ 1 ψ 0 ( s ) + ψ 1 ( h 2 ( s ) s ) 0 1 ψ 1 ( ζ h k + 2 ( s ) s ) d ζ ( 1 ψ 0 ( s ) ) 2 ] h k + 1 ( s ) ,
and
h ¯ i ( s ) = h i ( s ) 1 , i = 1 , 2 , , n ,
equations
h ¯ i ( s ) = 0
have the least solutions ρ i , respectively, in T 0 { 0 } .
Then, we show that ρ given by
ρ = min { ρ i }
is a convergence radius for scheme (2).
It follows by these definitions that for each s [ 0 , ρ )
0 ψ 0 ( s ) < 1 , 0 ψ 0 ( h 1 ( s ) s ) < 1
and
0 h i ( s ) < 1 .
Let B [ γ , α ] stand for the closure of ball B ( γ , α ) .
The conditions ( C ) are used to provide the functions ψ as given earlier:
( C 1 )
Equation H ( x ) = 0 has a simple solution γ Ω .
( C 2 )
For all u Ω
H ( γ ) 1 ( H ( γ ) H ( u ) ) ψ 0 ( γ u ) .
Set Ω 0 = Ω B ( γ , ρ 0 ) .
( C 3 )
For all u, v Ω 0
H ( γ ) 1 ( H ( u ) H ( v ) ) ψ ( u v )
and
H ( γ ) 1 H ( v ) ψ 1 ( v γ ) .
( C 4 )
B [ γ , ρ ] Ω .
Next, using conditions ( C ) together with the developed notation, we develop the local result of scheme (2).
Theorem 1.
Suppose conditions ( C ) hold. Then, if x 0 B 0 : = B ( γ , ρ ) { γ } , iteration { x n } developed by scheme (2) exists, stays in B 0 and converges to γ.
Proof. 
Let us choose u B 0 and t n : = x n γ . Then, by (5), (6) and ( C 2 ) , one has
H ( γ ) 1 ( H ( u ) H ( γ ) ) ψ 0 ( u γ ) ψ 0 ( ρ ) < 1 .
By estimation (8) with the lemma of Banach on linear inverse mapping [5], we obtain H ( u ) 1 L ( Y , X ) with
H ( u ) 1 H ( γ ) 1 1 ψ 0 ( u γ ) ·
By (9) for u = x 0 , y i exists. Using (5), (9) (for u = x 0 and n = 0 ), (7) (for i = 1 ), scheme (2) and ( C 3 )
y 1 γ = x 0 γ H ( x 0 ) 1 H ( x 0 ) H ( x 0 ) 1 H ( γ ) 0 1 H ( γ ) 1 H ( γ + ζ ( x 0 γ ) ) H ( x 0 ) ( x 0 γ ) d ζ 0 1 ψ ( ( 1 ζ ) t 0 ) d ζ 1 ψ 0 ( t 0 ) t 0 h 1 ( t 0 ) t 0 t 0 < ρ ,
then, y 1 B ( γ , ρ ) and H ( y 1 ) 1 L ( Y , X ) .
Using (5)–(7) (for i = 2 ), (9) (for u = y 1 ), ( C 3 ), (10) and scheme (2), we obtain in turn
y 2 γ = y 1 γ H ( y 1 ) 1 H ( y 1 ) + H ( y 1 ) 1 H ( x 0 ) H ( y 1 ) H ( x 0 ) 1 H ( y 1 ) 2 H ( x 0 ) 1 H ( y 1 ) [ h 1 ( y 1 γ ) + ( ψ 0 ( y 1 γ ) + ψ 0 ( t 0 ) ) 0 1 ψ 1 ( ζ y 1 γ ) d ζ ( 1 ψ 0 ( y 1 γ ) ) ( 1 ψ 0 ( t 0 ) ) + 2 0 1 ψ 1 ( ζ y 1 γ ) d ζ 1 ψ 0 ( t 0 ) ] y 1 γ h 2 ( t 0 ) t 0 ,
hence, y 2 B ( γ , ρ ) . Similarly, by (5), (7) (for i = 3 ), (10), (11) and scheme (2), we have in turn that
y 3 γ = y 1 γ 7 4 H ( x 0 ) 1 H ( y 1 ) + 1 2 H ( x 0 ) 1 H ( y 2 ) H ( x 0 ) 1 H ( y 1 ) + 1 4 H ( x 0 ) 1 H ( y 2 ) H ( x 0 ) 1 H ( y 2 ) H ( x 0 ) 1 H ( y 1 ) [ 1 + 7 4 0 1 ψ 1 ( ζ y 1 γ ) d ζ 1 ψ 0 ( t 0 ) + 1 2 ψ 1 ( y 1 γ ) ψ 1 ( y 2 γ ) ( 1 ψ 0 ( t 0 ) ) 2 + 1 4 ψ 1 ( y 1 γ ) ψ 1 ( y 2 γ ) 2 ( 1 ψ 0 ( t 0 ) ) 3 ] y 1 γ h 3 ( t 0 ) t 0 t 0 ,
therefore, y 3 B ( γ , ρ ) .
Then, for the rest of the substeps, similarly, we obtain for k = 1 , , m 3
y k + 3 γ = y k + 2 γ 2 H ( x 0 ) 1 H ( y k + 2 ) + H ( x 0 ) 1 H ( y 2 ) H ( x 0 ) 1 H ( y k + 2 ) [ 1 + 2 0 1 ψ 1 ( ζ y k + 2 γ ) d ζ 1 ψ 0 ( t 0 ) + ψ 1 ( y 2 γ ) 0 1 ψ 1 ( ζ y k + 2 γ ) d ζ ( 1 ψ 0 ( t 0 ) ) 2 ] y k + 2 γ h k + 3 ( t 0 ) t 0 t 0 .
By the definition of x 1 = y m and (10)–(13), we deduce
t 1 c t 0 ,
where
c = h 1 ( t 0 ) h 2 ( t 0 ) h m ( t 0 ) [ 0 , 1 ) .
Simply replace x 0 by x j in the preceding calculations to obtain
t j + 1 c t j c j + 1 t 0 < ρ ,
so, lim j + x j = γ and x j + 1 B ( γ , ρ ) . □
We have the uniqueness result:
Proposition 1.
Suppose:
(i) 
Point γ is a simple solution in B ( γ , ρ 0 ) Ω of (1) and
(ii) 
0 1 ψ 0 ( ζ ρ ¯ ) d ζ < 1 for some ρ ¯ ρ 0 .
Then, γ is unique in Ω 1 = Ω B ¯ ( γ , ρ ¯ ) as a solution of (1).
Proof. 
Let q Ω 1 be such that H ( q ) = 0 . Define Q = 0 1 H ( γ + ζ ( q γ ) ) d ζ . Then, in view of ( C 2 ) and (16), we obtain in turn
H ( γ ) 1 ( Q H ( γ ) ) 0 1 ψ 0 ( ζ γ q ) d ζ 0 1 ψ 0 ( ζ ρ ¯ ) d ζ < 1 ,
so, γ = q , since Q 1 L ( Y , X ) and 0 = H ( q ) H ( γ ) = Q ( q γ ) . □
Definition 1.
The COC is defined as
a = ln s n + 1 γ s n γ / ln s n γ s n 1 γ
and, for β n : = s n s n 1 , the ACOC [16] by
b = ln β n + 1 β n / ln β n β n 1 .

3. Applications

We test the conditions ( C ) .
Example 1.
If X = Y = Ω = R , let function H ( s ) = sin s . The conditions ( C ) must be verified. Clearly, γ = 0 solves the equation H ( s ) = 0 . Hence, we choose γ = 0 in conditions ( C ) . In particular, the choice γ = 0 satisfies the condition ( C 1 ) . Concerning the condition ( C 2 ) , we have, in turn, by the definition of H and, since H ( γ ) = 1 and H ( γ ) 1 ( H ( γ ) H ( u ) ) ψ 0 ( γ u ) , that
H ( γ ) 1 ( H ( γ ) H ( u ) ) = | cos γ cos u | ψ 0 ( | γ u | )
provided that ψ 0 ( s ) = s , since there exists c R such that
| cos γ c o s u | = | sin c ( γ u ) | | sin c | | γ u | = | γ u | .
That is the condition ( C 1 ) that is satisfied for this choice of ψ 0 . Moreover, by solving the equation ψ 0 ( s ) 1 = 0 , we obtain ρ 0 = 1 , and we can set Ω 0 = B ( 0 , 1 ) .
Similarly, we have
H ( γ ) 1 ( H ( u ) H ( v ) ) | sin d ( u v ) | | sin d | | u v | | u v |
for some, d R , so we can choose ψ ( s ) = s . Then, the first condition in ( C 2 ) is satisfied. Concerning the second condition in ( C 2 ) , we obtain in turn that
H ( γ ) 1 H ( v ) = | 1 sin v | 1 ,
so, we can choose ψ 1 ( s ) = 1 . Then, we have for k = 3 the resutls of Table 1.
Example 2.
If Ω = B [ 0 , 1 ] , and considering the continuous operator H : Ω S S (with the maximum norm) as
H ( ψ ) ( s ) = ψ ( s ) 5 0 1 s ζ ψ ( ζ ) 3 d ζ ,
where space S contains continuous functions defined on the interval [ 0 , 1 ] , we have that
H ( ψ ( λ ) ) ( s ) = λ ( s ) 15 0 1 s ζ ψ ( ζ ) 2 λ ( ζ ) d ζ , for each λ Ω .
Then, we obtain that γ = 0 , and for k = 3 we obtain the functions ψ 0 ( s ) = ψ ( s ) = 15 2 s , ψ 1 ( s ) = 2 . This way, we obtain the Table 2.
Example 3.
Let Ω = B ( 0 , 1 ) , X = Y = R 3 and γ = ( 0 , 0 , 0 ) T . Let H defined on Ω be
H ( w ) = H ( w 1 , w 2 , w 3 ) = e w 1 1 , e 1 2 w 2 2 + w 2 , w 3 T .
For the point w = ( w 1 , w 2 , w 3 ) T , the derivative H is
H ( w ) = e w 1 0 0 0 ( e 1 ) w 2 + 1 0 0 0 1 .
Taking into account rows with their max. norm and
H ( γ ) = 1 0 0 0 1 0 0 0 1 ,
for k = 3 we obtain through conditions ( C ) ψ 0 ( s ) = ( e 1 ) s , ψ ( s ) = e 1 e 1 s , ψ 1 ( s ) = e 1 e 1 and the radius in Table 3.
Example 4.
Taking into account the academic example in the first section of this work, for k = 3 , we obtain ψ 0 ( s ) = ψ ( s ) = 96.662907 s , ψ 1 ( s ) = 1.0631 . The corresponding radius are in Table 4.
Example 5.
We consider the conventional form of Kepler’s equation
G ( s ) = s σ sin ( s ) τ = 0 ,
where 0 τ π and 0 σ < 1 . In [22], we can find, for distinct values of τ and σ, a numerical study. In this case, we select value τ = 0.1 and σ = 0.27 , so the solution obtained is γ 0.136828 . Since
G ( s ) = 1 σ cos ( s ) ,
we obtain
G ( γ ) 1 ( G ( s ) G ( t ) ) = σ ( cos ( s ) c o s ( t ) ) 1 σ cos ( γ ) = 2 σ sin s + t 2 sin s t 2 1 σ cos ( γ ) σ 1 σ cos ( γ ) s t
and
G ( γ ) 1 G ( s ) = 1 σ cos ( s ) 1 σ cos ( γ ) 1 + σ 1 σ cos ( γ ) ·
Consequently, we have ψ 0 ( s ) = ψ ( s ) = 0.3685888 t and ψ 1 ( s ) = 1.7337327 . The calculated value parameters for k = 3 are given in Table 5.
Example 6.
We analyze the following system of nonlinear equations
ν 2 ν 3 + ν 4 ( ν 2 + ν 3 ) = 0 ν 1 ν 3 + ν 4 ( ν 1 + ν 3 ) = 0 ν 1 ν 2 + ν 4 ( ν 1 + ν 2 ) = 0 ν 1 ν 2 + ν 1 ν 3 + ν 2 ν 3 1 = 0
with ν 1 , ν 2 , ν 3 , ν 4 R . Then, the solution of the previous scheme is showed for γ = ( ν 1 , ν 2 , ν 3 , ν 4 ) T by function H : = ( ν 1 , ν 2 , ν 3 , ν 4 ) : R 4 R 4 defined by
H ( ν 1 , ν 2 , ν 3 , ν 4 ) = ν 2 ν 3 + ν 4 ( ν 2 + ν 3 ) ν 1 ν 3 + ν 4 ( ν 1 + ν 3 ) ν 1 ν 2 + ν 4 ( ν 1 + ν 2 ) ν 1 ν 2 + ν 1 ν 3 + ν 2 ν 3 1
The Fréchet derivative is given by
H ( ν 1 , ν 2 , ν 3 , ν 4 ) = 0 ν 3 + ν 4 ν 2 + ν 4 ν 2 + ν 3 ν 3 + ν 4 0 ν 1 + ν 4 ν 1 + ν 3 ν 2 + ν 4 ν 1 + ν 4 0 ν 1 + ν 2 ν 2 + ν 3 ν 1 + ν 3 ν 1 + ν 2 0
Then, we obtain
ν 1 = 0.57735026 ν 2 = 0.57735026 ν 3 = 0.57735026 ν 4 = 0.28867513
Then, we have that ψ 0 ( s ) = 0.28867513 s , ψ ( s ) = 1.55198152 s , ψ 1 ( s ) = 1.06466589 s and the radii are in Table 6.

4. Conclusions

A technique is introduced involving only H , which also gives computable upper error bounds on x n γ . This way, the method (2) becomes applicable in cases not possible before, since higher hypotheses on H were required. The COC or ACOC are used to determine the convergence order. The technique is not based on method (2) and is very general. Hence, it can be used to extend the applicability of other methods [11,17,20,21]. This will be the focus of our future work.

Author Contributions

Conceptualization, I.K.A., C.A., M.A., J.C. and D.G.; data curation, I.K.A., C.A., M.A., J.C. and D.G.; methodology, I.K.A., C.A., M.A., J.C. and D.G.; project administration, D.G.; formal analysis, I.K.A., C.A., M.A., J.C. and D.G.; investigation, I.K.A., C.A., M.A., J.C. and D.G.; resources, I.K.A., C.A., M.A., J.C. and D.G.; writing—original draft preparation, I.K.A., C.A., M.A., J.C. and D.G.; writing—review and editing, I.K.A., C.A., M.A., J.C. and D.G.; visualization, I.K.A., C.A., M.A., J.C. and D.G.; supervision, I.K.A., C.A., M.A., J.C. and D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Universidad de Las Américas, Quito, Ecuador, grant number FGE.DGS.20.15.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Amat, S.; Argyros, I.K.; Busquier, S.; Hernández–Verón, M.Á.; Martínez, E. On the local convergence study for an efficient k–step iterative method. JCAM 2018, 343, 753–761. [Google Scholar] [CrossRef]
  2. Amorós, C.; Argyros, I.K.; González, D.; Magreñán, Á.A.; Regmi, S.; Sarría, Í. New Improvement of the Domain of Parameters for Newton’s Method. Mathematics 2020, 8, 103. [Google Scholar] [CrossRef] [Green Version]
  3. Argyros, I.K.; George, S. Mathematical Modeling for the Solution of Equations and Systems of Equations with Applications; Nova Publishers: Hauppauge, NY, USA, 2019; Volume IV. [Google Scholar]
  4. Argyros, I.K.; González, D.; Magreñán, Á.A.; Moysi, A.; Regmi, S.; Sicilia, J.A. Convergence and Dynamics of a Higher–Order Method. Symmetry 2020, 12, 420. [Google Scholar]
  5. Argyros, I.K.; Magreñán, Á.A. A Contemporary Study of Iterative Methods; Elsevier: Amsterdam, The Netherlands, 2018. [Google Scholar]
  6. Frontini, M.; Sormani, E. Some variant of Newton’s method with third–order convergence. Appl. Math. Comput. 2003, 140, 419–426. [Google Scholar] [CrossRef]
  7. Grau–Sánchez, M.; Grau, Á.; Noguera, M. Ostrowski type methods for solving systems of nonlinear equations. Appl. Math. Comput. 2011, 218, 2377–2385. [Google Scholar] [CrossRef]
  8. Ortega, J.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equations in Several Variables; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  9. Rheinboldt, W.C. An adaptive continuation process for solving systems of nonlinear equations. Pol. Acad. Sci. Banach Ctr. Publ. 1978, 3, 129–142. [Google Scholar] [CrossRef]
  10. Soleymani, F.; Lotfi, T.; Bakhtiari, P. A multi–step class of iterative methods for nonlinear systems. Optim. Lett. 2014, 8, 1001–1015. [Google Scholar] [CrossRef]
  11. Traub, J.F. Iterative Methods for the Solution of Equations; Prentice-Hall: Englewood Cliffs, NJ, USA, 1964. [Google Scholar]
  12. Cordero, A.; Hueso, J.L.; Martínez, E.; Torregrosa, J.R. A modified Newton–Jarratt’s composition. Numer. Algorithms 2010, 55, 87–99. [Google Scholar] [CrossRef]
  13. Gutiérrez, J.M.; Hernández, M.A. A family of Chebyshev–Halley type methods in Banach spaces. Bull. Aust. Math. Soc. 1997, 55, 113–130. [Google Scholar] [CrossRef] [Green Version]
  14. Homeier, H.H. A modified Newton method with cubic convergence: The multivariable case. J. Comput. Appl. Math. 2004, 169, 161–169. [Google Scholar] [CrossRef] [Green Version]
  15. Jarratt, P. Some fourth order multipoint iterative methods for solving equations. Math. Comp. 1966, 20, 434–437. [Google Scholar] [CrossRef]
  16. Kou, J. A third–order modification of Newton method for systems of nonlinear equations. Appl. Math. Comput. 2007, 191, 117–121. [Google Scholar]
  17. Petkovic, M.S. On a general class of multipoint root–finding methods of high computational efficiency. SIAM J. Numer. Anal. 2011, 49, 1317–1319. [Google Scholar] [CrossRef]
  18. Sharma, J.R.; Guha, R.K.; Sharma, R. An efficient fourth order weighted–Newton method for systems of nonlinear equations. Numer. Algorithms 2013, 62, 307–323. [Google Scholar] [CrossRef]
  19. Ahmad, F.; Tohidi, E.; Ullah, M.Z.; Carrasco, J.A. Higher order multi–step Jarratt–like method for solving systems of nonlinear equations: Application to PDEs and ODEs. Comp. Math. App. 2015, 70, 624–636. [Google Scholar] [CrossRef] [Green Version]
  20. Kung, H.T.; Traub, J.F. Optimal order of one point and multipoint iteration. J. AMC 1974, 21, 643–651. [Google Scholar] [CrossRef]
  21. Ignatova, B.; Kyurkchiev, N.; Iliev, A. Multipoint Algorithms Arising from Optimal in the Sense of Kung–Traub Iterative Procedures for Numerical Solution of Nonlinear Equations. Gen. Math. Notes 2011, 6, 45–79. [Google Scholar]
  22. Danby, J.M.A.; Burkardt, T.M. The solution of Kepler’s equation, I. Celest. Mech. 1983, 31, 95–107. [Google Scholar] [CrossRef]
Table 1. Radius for Example 1.
Table 1. Radius for Example 1.
ρ ρ 0 ρ 1 ρ 2 ρ 3
0.306688…10.666667…0.343549…0.306688…
Table 2. Radius for Example 2.
Table 2. Radius for Example 2.
ρ ρ 0 ρ 1 ρ 2 ρ 3
0.024324…0.133333…0.088888…0.032839…0.024324…
Table 3. Radius for Example 3.
Table 3. Radius for Example 3.
ρ ρ 0 ρ 1 ρ 2 ρ 3
0.115169…0.581977…0.382692…0.148762…0.115169…
Table 4. Radius for Example 4.
Table 4. Radius for Example 4.
ρ ρ 0 ρ 1 ρ 2 ρ 3
0.003149…0.010345…0.006896…0.003463…0.003149…
Table 5. Radius for Example 5.
Table 5. Radius for Example 5.
ρ ρ 0 ρ 1 ρ 2 ρ 3
0.550956…2.713050…1.808700…0.720969…0.550956…
Table 6. Radius for Example 6.
Table 6. Radius for Example 6.
ρ ρ 0 ρ 1 ρ 2 ρ 3
0.391641…0.775990…0.484339…0.413569…0.391641…
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Argyros, I.K.; Argyros, C.; Argyros, M.; Ceballos, J.; González, D. Extended Multi-Step Jarratt-like Schemes of High Order for Equations and Systems. Mathematics 2022, 10, 3603. https://doi.org/10.3390/math10193603

AMA Style

Argyros IK, Argyros C, Argyros M, Ceballos J, González D. Extended Multi-Step Jarratt-like Schemes of High Order for Equations and Systems. Mathematics. 2022; 10(19):3603. https://doi.org/10.3390/math10193603

Chicago/Turabian Style

Argyros, Ioannis K., Chirstopher Argyros, Michael Argyros, Johan Ceballos, and Daniel González. 2022. "Extended Multi-Step Jarratt-like Schemes of High Order for Equations and Systems" Mathematics 10, no. 19: 3603. https://doi.org/10.3390/math10193603

APA Style

Argyros, I. K., Argyros, C., Argyros, M., Ceballos, J., & González, D. (2022). Extended Multi-Step Jarratt-like Schemes of High Order for Equations and Systems. Mathematics, 10(19), 3603. https://doi.org/10.3390/math10193603

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop