Next Article in Journal
Vectorized MATLAB Implementation of the Incremental Minimization Principle for Rate-Independent Dissipative Solids Using FEM: A Constitutive Model of Shape Memory Alloys
Previous Article in Journal
Power Families of Bivariate Proportional Hazard Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparative Analysis of Accelerated Models for Solving Unconstrained Optimization Problems with Application of Khan’s Hybrid Rule

by
Vladimir Rakočević
1,2 and
Milena J. Petrović
3,*
1
Serbian Academy of Sciences and Arts, Kneza Mihajla 35, 11000 Belgrade, Serbia
2
Faculty of Sciences and Mathematics, University of Niš, Višegradska 33, 18106 Niš, Serbia
3
Faculty of Sciences and Mathematics, University of Pristina in Kosovska Mitrovica, Lole Ribara 29, 38220 Kosovska Mitrovica, Serbia
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(23), 4411; https://doi.org/10.3390/math10234411
Submission received: 30 October 2022 / Revised: 17 November 2022 / Accepted: 21 November 2022 / Published: 23 November 2022

Abstract

:
In this paper, we follow a chronological development of gradient descent methods and its accelerated variants later on. We specifically emphasise some contemporary approaches within this research field. Accordingly, a constructive overview over the class of hybrid accelerated models derived from the three-term hybridization process proposed by Khan is presented. Extensive numerical test results illustrate the performance profiles of hybrid and non-hybrid versions of chosen accelerated gradient models regarding the number of iterations, CPU time, and number of function evaluation metrics. Favorable outcomes justify this hybrid approach as an accepted method in developing new efficient optimization schemes.
MSC:
49M15; 49M37; 65B99; 90C26; 90C30; 90C53

1. Class of Accelerated Gradient Descent Methods and Its Benefits

Many contemporary scientific, engineering, medical and problems from various other research areas, are closely related to mathematical optimization theory. Among all others, the unconstrained optimization problems are the most frequently considered [1,2,3,4,5,6,7,8]. Owing to the duality principle, an optimization problem may be viewed as a minimization problem. Unconstrained minimization problems can simply be stated as finding
min f ( x ) , x R n ,
where f : R n R is an objective function, that is solved by the general iteration:
x k + 1 = x k + t k d k .
In (1), x k presents the current iterative point, x k + 1 is the next one; the positive iterative step length value is denoted by t k while d k stands for the k-th search direction vector. As can be observed, two main elements that measure the efficiency and robustness of the iterative rule (1) are the adequately calculated iterative step size t k and the properly chosen iterative search direction d k . Since we are dealing with minimization problems, it is a natural choice to define the search direction so that it fulfils the descent condition, i.e.,
g k T d k < 0 ,
where g k is the gradient of f at the point x k . Apart from that, we use standard notations for the gradient and the Hessian of the objective function f:
g ( x ) = f ( x ) , G ( x ) = 2 f ( x ) , g k = f ( x k ) , G k = 2 f ( x k ) .
Following condition (2), it is easy to conclude that the d g k produces the most certain descent direction, known as the gradient descent direction. Iteration (1) that includes the gradient descent direction is known as the gradient descent method (or GD method)
x k + 1 = x k t k g k .
Step length parameter t k in iterations (1) and (4), is determined either by the exact or by some of the inexact line search procedures. Using the exact line search technique, iterative step length value t k is computed by solving the following minimization task:
f ( x k + t k d k ) = min f ( x k + t d k ) , t > 0 .
It is clear that solving the previous minimization problem in each iterative value presents a time- and resource-consuming task regarding CPU time requirements and the number of required objective function evaluations. For this reason, most contemporary optimization methods use the inexact line search algorithms to calculate iterative step size instead of the exact line search procedure. The convergence properties of line search methods for unconstrained optimization are specifically examined in [9]. Some of the frequently used inexact line search algorithms are weak and strong Wolfe’s algorithms [10], Backtracking algorithm proposed in [11] with Armijo’s rule [12], etc.:
  • Weak Wolfe’s line search: 
    f ( x k + t k d k ) f ( x k ) + δ t k g k T d k
    g ( x k + t k d k ) T d k σ g k T t d k ;
  • Strong Wolfe’s line search: 
    f ( x k + t k d k ) f ( x k ) + δ t k g k T d k
    | g ( x k + t k d k ) T d k | σ g k T t d k ;
  • Backtracking algorithm:
  • Objective function f ( x ) , the direction d k of the search at the point x k and numbers 0 < σ < 0.5 and β ( 0 , 1 ) are required;
  • t = 1 ;
  • f ( x k + t d k ) > f ( x k ) + σ t g k T d k , take t : = t β ;
  • Return t k = t .
Subsequently, the Newton method with line search is given as
x k + 1 = x k t k G k 1 g k .
In (6), G k 1 stands for the inverse of the function Hessian, according to the previously adopted notation. Step length parameter t k is obtained by applying some chosen inexact procedure. Instead of calculating the inverse of the function Hessian, which is often a demanding task, in quasi-Newton methods the adequate approximation of the Hessian (or of its inverse) is used
x k + 1 = x k t k H k g k .
Herein, H k B k 1 and B k is derived as a symmetric positive definite Hessian’s approximation. Updating of { B i } , i N is conducted using the quasi-Newton property of secant equation
B k + 1 s k = y k ,
where parameters s k and y k are the differences in two successive iterative points and iterative gradients, respectively, i.e., s k = x k + 1 x k , y k = g k + 1 g k . In [13], the classification of methods for updating the matrix B k is presented. Therein, the three updating methods are extricated:
  • matrix B k is defined as a scalar matrix, i.e., B k = γ k I , γ k > 0 ;
  • matrix B k is defined as a diagonal matrix, i.e., B k = d i a g ( λ 1 , , λ n ) , λ i > 0 , i = 1 , n ¯ ;
  • matrix B k is defined as a full matrix.
Taking the simplest updating approach, the first one, i.e., B k = γ k I G k , γ k > 0 , the quasi-Newton method (7) is transformed to
x k + 1 = x k t k γ k 1 g k .
In [14], the authors named the iterative methods (8) as a class of accelerated gradient methods. They named it so for their good convergence and performance metrics. Previously in [15], Andrei defined this accelerated iteration by calculating parameter θ k ( = γ k 1 ) as follows:
Algorithm for generating scalar θ k from [15]:
  • Objective function f ( x ) , the direction d k of the search at the point x k and numbers 0 < σ < 0.5 and β ( 0 , 1 ) are required;
  • Apply Backtracking algorithm to calculate t k ( 0 , 1 ] ;
  • Compute z = x k t k g k , g z = f ( z ) , y k = g z g k ;
  • Compute a k = t k g k T g k , b k = t k y k T g k
  • Return θ k = a k b k .
Stanimirović and Miladinović in [14] determined the parameter γ k 1 from (8) on the basis of the second-order Taylor’s expansion, and denoted it as S M method. Results obtained by several researchers on this topic confirmed that this way of deriving the accelerated parameter (as named in [16]), is justifiable regarding the convergence and numerical performance aspects [17,18]. Several forms of this important variable in chosen accelerated gradient schemes are listed as expressions (5)–(9) in [19] and some other approaches are presented in [20,21,22].
Among the various accelerated gradient iterations, for this research we specifically chose three among which Khan’s hybridization three-term process was later applied. The first of the three is the already mentioned S M method, defined by relation (8) and presented in [14]. The second one is the A D D method (i.e., accelerated double direction method) from [16], with the iterative representation
x k + 1 = x k + α k 2 d k α k γ k 1 g k .
In (9), α k > 0 is the iterative step length value, while d k is the second search vector calculated under the assumption d k = 1 , k = 1 , 2 , , by the next procedure:
d k ( t ) = d k * , k m 1 i = 2 m t i 1 d k i + 1 * k m
and d k * is the solution of the problem min x R Φ k ( d ) ,
Φ k ( d ) = f ( x k ) T d + 1 2 γ k + 1 I = g ( x k ) T d + 1 2 γ k + 1 I .
The iterative value of the accelerated parameter γ k , obtained through Taylor’s series of (9), is
γ k + 1 A D D = 2 f ( x k + 1 ) f ( x k ) α k g k T α k d k γ k 1 g k α k d k γ k 1 g k T t k d k γ k 1 g k .
The positive step length value α k of the A D D scheme is derived using the Backtracking algorithm, starting with initial t = 1 . Taking the following substitutions in (9)
α k 2 β k ,
where β k is calculated by a different Backtracking procedure, and
d k g k ,
lead to the ADSS (accelerated double step size) method [17]:
x k + 1 = x k α k γ k 1 g k β k g k = x k α k γ k 1 + β k g k .
Finally, under assumption
α k + β k = 1 ,
A D S S iteration is transformed to T A D S S scheme [18], our third chosen accelerated gradient descent model:
x k + 1 = x k [ α k ( γ k 1 1 ) + 1 ] g k .
As in (9), the iterative step length value α k in (14) is calculated on the basis of Backtracking algorithm. The accelerated parameter of the T A D S S scheme is
γ k + 1 T A D S S = 2 f ( x k + 1 ) f ( x k ) + ψ k g k 2 ψ k 2 g k 2 , ψ k = [ α k γ k 1 α k 2 ) + 1 ] .
In the following proposition, we prove that (8), (9) and (14) iterations are a gradient descending process.
Proposition 1.
The search directions in iterations defined by (8), (9) and (14) fulfil the descending condition (2).
Proof. 
We separately analyze the search directions of all three listed methods.
  • According to the general iteration form (1), the search direction in S M method, defined by relation (8), is d k γ k 1 g k . One of the essential properties of the accelerated parameter γ k is its positiveness. If in some iterative step k of the accelerated gradient algorithms with leading iterative rules (8), (9) and (14) this necessary condition is not fulfilled, then the k-th accelerated scalar value is set to be γ k = 1 . Bearing this fact in mind, we easily conclude that
    g k T d k = g k T ( γ k 1 g k ) = γ k 1 g k 2 < 0 ,
    which confirms that the condition (2) in S M method, defined by (8), is fulfilled.
  • Accelerated double direction A D D scheme (9) contains two search vectors. The first one, denoted as d k , is defined by (10). The second one is of the same form as in the S M iteration, i.e., γ k 1 g k . In the procedure (10), crucial element d k * in deriving vector d k is defined as a solution of the minimization problem (11) that depends on the gradient g k , under the assumption d k = 1 . Thus, the defined vector direction is a relaxed differentiable variant of the procedure for determination of the search vector d k (rule 2) in [23] and accordingly, the d = 0 is globally optimum of the problem (11). Subsequently, we consider only the second direction, which is already performed in the previous item.
  • In T A D S S scheme, vector direction can be seen as [ α k ( γ k 1 1 ) + 1 ] g k . Checking the descent condition (2), we get
    g k T · [ α k ( γ k 1 1 ) + 1 ] g k = [ α k ( γ k 1 1 ) + 1 ] g k 2 < 1 ,
    since γ k 1 > 1 and α k ( 0 , 1 ] according to T A D S S algorithm.
Further on, in Section 2, we analyse Khan’s hybridization rule applied on various gradient methods. Finally, Dolan–Moré representations and parallels among the hybrid and non-hybrid versions of S M , A D D and T A D S S schemes, obtained over large-scale numerical outcomes, are illustrated in Section 3.

2. Three-Term Khan’s Hybridization Principle over the Accelerated Gradient Descent Models

For a nonempty convex subset C of a normed space E , let T : C C be a mapping defined on C . Then, for some sequences u k , v k , z k and y k defined on C , the Picard, Mann and Ishikawa iterative processes [24,25,26] are, respectively, given as:
u 1 = u C , u k + 1 = T u k , k N ,
v 1 = v C , v k + 1 = ( 1 α k ) v k + α k T v k , k N ,
z 1 = z C , z k + 1 = ( 1 α k ) z k + α k T y k , y k = ( 1 β k ) z k + β k T z k . k N
In the listed relations, parameters { α k } , { β k } ( 0 , 1 ) are the sequences of positive numbers, which in the Ishikawa process [26] fulfil the following assumptions
  • 0 α k β k 1 , k 0 ,
  • lim k β k = 0 ,
  • k = 1 α k β k = .
In [27], Khan proposed a new three-term iterative process as follows
x 1 = x R , x k + 1 = T y k , y k = ( 1 α k ) x k + α k T x k , k N
with the sequence of positive numbers { α k } ( 0 , 1 ) , which is considered as a set of constant values, i.e., ( α = α k ( 0 , 1 ) k N , ) in practical numerical tests, as proposed in [27]. Khan confirmed in [27] that the process (16) converges faster than the processes of Picard, Mann and Ishikawa.
Khan developed this iterative process (16) as a hybrid variant of previously mentioned, well-known iterations and with that managed to improve these types of methods. Further, some authors used auspicious aspects of this hybrid rule and applied it to some accelerated gradient optimization methods. The hybridization principle consists of taking the objective accelerated iteration as a guiding operator in (16). As a result, several accelerated-hybridization processes arose [28,29,30,31,32]. We list them below, together with their accelerated parameters.
  • Hybrid accelerated gradient descent method ( H S M ) [28]
    x k + 1 = x k ( α k + 1 ) t k γ k 1 g k ,
    γ k + 1 H S M = 2 γ k γ k f ( x k + 1 ) f ( x k ) + ( α k + 1 ) t k g k 2 ( α k + 1 ) 2 t k 2 g k 2 .
  • Hybrid accelerated double direction method ( H A D D ) [29]
    x k + 1 = x k α t k γ k 1 g k + α t k 2 d k , α ( 1 , 2 ) ,
    γ k + 1 = 2 f ( x k + 1 ) f ( x k ) α g k T t k 2 d k t k γ k 1 g k α 2 t k 2 t k d k γ k 1 g k T t k d k γ k 1 g k .
  • Hybrid accelerated double step size method ( H A D S S ) [30]
    x k + 1 = x k g k α ( t k γ k 1 + p k ) , α α k + 1 ( 1 , 2 ) k ,
    γ k + 1 H A D S S = 2 f ( x k + 1 ) f ( x k ) + α t k γ k 1 + p k g k 2 α 2 t k γ k 1 + p k 2 g k 2 .
  • Hybrid transformed double step size method ( H T A D S S ) [31]
    x k + 1 = x k α ( t k ( γ k 1 1 ) + 1 ) g k , α ( 1 , 2 ) ,
    γ k + 1 H T A D S S = 2 f ( x k + 1 ) f ( x k ) + α φ k g k 2 α 2 φ k 2 g k 2 ,
    where
    φ k = t k ( γ k 1 1 ) + 1 .
  • Hybrid gradient descent method ( H G D ) [32]
    x k + 1 = x k ( α k + 1 ) t k g k , α k ( 0 , 1 ) k .
  • Hybrid accelerated gradient descent method ( H A G D ) [32]
    x k + 1 = x k ( α k + 1 ) θ k t k g k , α ( 0 , 1 ) k , θ k = γ k t k γ k + 1 .
  • Hybrid modified accelerated gradient descent method ( H M A G D ) [32]
    x k + 1 = x k ( α k + 1 ) θ k ( t k + t k 2 t k 3 ) g k , α ( 0 , 1 ) k , θ k = γ k t k γ k + 1 .
  • Hybrid modified improved gradient descent method ( H M I G D ) [32]
    x k + 1 = x k ( α k + 1 ) γ k 1 ( t k + t k 2 t k 3 ) g k , α ( 0 , 1 ) k .
As shown above, from Khan’s hybridization rule at least eight hybrid models have appeared. Convergence properties as well as performance efficiency of these iterative schemes are presented and illustrated in the literature. The leading model of all listed ((17), (19), (21), (23), (26), (27), (28), (29)), and therewith the first one that was developed on the basis of the Khan’s process, is certainly the H S M method from [28]. In paper [28], the authors examined the performance of the defined method for various values of the so-called correction parameter α k ( 0 , 1 ) , i.e., α = α k + 1 ( 1 , 2 ) , which is a necessary factor of all hybrid methods generated through Khan’s hybridization. They experimentally concluded that the H S M method achieves best performance characteristics when the correction parameter is taken closely to its left limit. Later in [33], the authors improved the H S M model by reducing the initial step length parameter in the Backtracking procedure.

3. Dolan–Moré Performance Profiles and Comparisons

In this section, we apply the aspects of Dolan–Moré benchmarking optimization software from [34] on hybrid and non-hybrid variants of chosen accelerated gradient minimization models. In conducted numerical tests, we follow performance metrics regarding the number of iterations, CPU time and number of function evaluations.
For all obtained numerical outcomes, the following points are valid:
  • Codes are written in the visual C++ programming language and run on a Workstation Intel(R) Core(TM) 2.3 GHz.
  • The Backtracking parameters values taken are: σ = 0.0001 and β = 0.8 . These are standard values for the Backtracking parameters applied in various optimization models with Backtracking algorithm [2,14,15,20,21,22,28,29,30,31]. This set of values means that a small portion of the decrease predicted by the linear approximation of the current point is accepted.
  • The stopping criteria are:
    g k 10 6 and | f ( x k + 1 ) f ( x k ) | 1 + | f ( x k ) | 10 16 .
  • Chosen test functions are taken from the unconstrained test functions collection presented in [35]. More precisely, all specific test functions that were used for this research are listed in Listing 1.
Listing 1. Test functions.
Listing 1. Test functions.
1. Extended Penalty
2. Perturbed Quadratic
3. Raydan-1
4. Diagonal 1
5. Diagonal 3
6. Generalized Tridiagonal-1
7. Diagonal 4
8. Extended Himmelblau
9. Quadr. Diag. Perturbed
10. Quadratic QF1
11. Exten. Quadr. Penalty QP1
12. Exten. Quadr. Penalty QP2
13. Quadratic QF2
14. Extended EP1
15. Arwhead
16. Almost Perturbed Quadratic
17. Engval1
18. Quartc
19. Generalized Quartic
20. LIARWHD
21. Diagonal 6
22. Tridia
23. Indef
24. Diagonal 9
25. DIXON3DQ
26. NONSCOMP
27. BIGGSB1
28. Power (Cute)
29. Hager
30. Raydan 2
Further, by i p , s , t p , s and e p , s we denote the number of iterations, the CPU time and the number of function evaluations, respectively, needed for solving problem p when the solver s is applied. The main observation arises from a comparison of performance profiles, considering analyzed metrics, of hybrid and non-hybrid versions of the same scheme. For this investigation, we chose the following three accelerated gradient methods: S M (8), A D D (9) and T A D S S (14). Accordingly, we analyzed their hybrid forms H S M (17), H A D D (19) and H T A D S S (23), as well. So in these tests, the solver s belongs to the set of six elements, s { S M , H S M , A D D , H A D D , T A D S S , H T A D S S } . We specifically chose this set of comparative non-hybrid and hybrid pairs among the others mentioned above in Section 2, since the selected models are the most cited of Khan’s hybrid methods.
According to the benchmark presented in [34] and regarding the comparisons obtained within this paper, for each pair of comparative hybrid and non-hybrid variants we have two solvers, i.e., n s = 2 , where the comparative pairs are:
{ ( S M , H S M ) , ( A D D , H A D D ) , ( T A D S S , H T A D S S ) } .
The total number of experiments for each pair is minimal n p = 210 . Precisely, for the pair ( S M , H S M ) we have conducted numerical tests for 25 test function and 11 different numbers of variables, so n p = 11 · 25 = 275 , the same as for the pair ( T A D S S , H T A D S S ) . The pair ( A D D , H A D D ) included 21 test functions with 10 different numbers of variables, so in this case n p = 210 . In order to apply Dolan–Moré benchmarking optimization software with performance profiles over the chosen accelerated and hybrid models, we use the original outcomes presented in the papers in which these models were generated [28,29,31]. So, since in [28,31] numerical experiments included 25 test functions, while in [29] the number of tested functions is 21, the total number of tests for all test functions and all three pairs of models is n p S M , H S M + n p T A D S S , H T A D S S + n p A D D , H A D D = 760 .
Considering defined parameters, we are now able to expose performance ratios defined for the number of iterations, the CPU time and the number of function evaluations, respectively:
r p , s = i p , s min { i p , s : s { S M , H S M } }             = i p , s min { i p , s : s { A D D , H A D D } } = i p , s min { i p , s : s { T A D S S , H T A D S S } } ,
r p , s = t p , s min { t p , s : s { S M , H S M } }             = t p , s min { t p , s : s { A D D , H A D D } } = t p , s min { t p , s : s { T A D S S , H T A D S S } } ,
r p , s = e p , s min { e p , s : s { S M , H S M } }             = e p , s min { e p , s : s { A D D , H A D D } } = e p , s min { e p , s : s { T A D S S , H T A D S S } } .
As in [34], we define the performance profile for each solver s with respect to all three measured metrics
ρ s ( τ ) = 1 n p s i z e { p P : r p , s τ } ,
which presents the cumulative distribution function. In (30) parameter τ R , while P is the set of problems.
In Figure 1, Figure 2 and Figure 3, we present the performance profiles of S M and H S M regarding the number of iterations, the CPU time and the number of function evaluations, respectively. Comparisons between pairs ( A D D , H A D D ) and ( T A D S S , H T A D S S ), regarding all three tested metrics are similarly illustrated in Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9.
From Figure 1, Figure 2 and Figure 3, we clearly observe that the H S M algorithm outperforms the non-hybrid S M model, with respect to the number of iterations and the CPU time metrics, while regarding the number of evaluations metric, the hybrid and non-hybrid models perform similarly.
From Figure 5 and Figure 6, we see that the hybrid accelerated double direction model shows conspicuously better features regarding CPU time and the number of evaluations metrics. Nevertheless, concerning the number of iterations metric, forerunner A D D is more efficient, as shown in Figure 4.
Finally, comparisons of the T A D S S and the H T A D S S methods are displayed in Figure 7, Figure 8 and Figure 9. From these three presented graphs, we observe that the hybrid version of the transformed double step size method convincingly upgrades its counterpart non-hybrid model. In this case, the dominance of the hybrid variant with respect to all three analyzed metrics is more than evident.
To obtain Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9, a total of 4560 numerical outcomes were included. More precisely, for 6 analyzed methods ( S M , A D D , T A D S S , H S M , H A D D , H T A D S S ) we followed 3 metrics (number of iterations, CPU time, number of function evaluations) on 25 test functions for S M , H S M , T A D S S and H T A D S S solvers and 21 test functions for A D D and H A D D solvers from [35]. For each test function, the tests were conducted for at least 10 different numbers of variables. With that, the execution time for each test is limited by the time-limiter parameter defined in [16].

4. Conclusions

In this research, we present an overview of two gradient method classes: accelerated gradient descent models and its hybrid variants derived from Khan’s three-term iterative rule. This is an important and useful retrospective of one, confirmed efficient approach in defining the robust accelerated methods for solving unconstrained optimization problems. The obtained results, achieved on the basis of comprehensive Dolan–Moré performance profiles [34], conducted on total 4560 numerical outcomes, confirm that Khan’s hybridization rule is justified for use as an applicable technique in generating effective minimization processes. Accordingly, this research paves the way for new possibilities aimed at generating similar hybridization rules and their applications to accelerated gradient schemes.

Author Contributions

Conceptualization, V.R. and M.J.P.; Methodology, V.R.; Software, M.J.P.; Validation, V.R.; Formal analysis, V.R.; Investigation, M.J.P.; Resources, V.R.; Data curation, M.J.P.; Writing–original draft, M.J.P.; Writing–review & editing, V.R.; Visualization, M.J.P.; Supervision, V.R.; Project administration, M.J.P. All authors contributed equally and significantly to the writing of this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by internal-junior project IJ-0202, Faculty of Sciences and Mathematics, University of Priština in Kosovska Mitrovica.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data results are available on request from the corresponding author.

Acknowledgments

The authors gratefully acknowledge support from the project Grant No. 174025 by the Ministry of Education and Science of Republic of Serbia.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Powel, M.J.D. A survey of numerical methods for unconstrained optimization. SIAM Rev. 1970, 12, 79–97. [Google Scholar] [CrossRef]
  2. Andrei, N. Nonlinear Conjugate Gradient Methods for Unconstrained Optimization. In Nonlinear Conjugate Gradient Methods for Unconstrained Optimization; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  3. Nocadal, J.; Wright, S.J. Numerical Optimization. In Numerical Optimization; Springer: New York, NY, USA, 1999. [Google Scholar]
  4. Jacoby, S.L.S.; Kowalik, J.S.; Pizzo, J.T. Iterative Methods for Nonlinear Optimization Problems. In Iterative Methods for Nonlinear Optimization Problems; Prentice-Hall, Inc.: Englewood, NJ, USA, 1977. [Google Scholar]
  5. Deniss, J.E.; Kowalik, J.S.; Schnabel, R.B. Numerical Methods for Unconstrained Optimization and Nonlinear Equations. In Numerical Methods for Unconstrained Optimization and Nonlinear Equations; Prentice-Hall: Englewood Cliffs, NJ, USA, 1983. [Google Scholar]
  6. Fletcher, R. Practical methods of optimization. In Practical Methods of Optimization; Prentice-Hall, Wiley: New York, NY, USA, 2000. [Google Scholar]
  7. Luenberg, D.G.; Ye, Y. Linear and nonlinear programming. In Linear and Nonlinear Programming; Springer Science + Business Media, LLC: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  8. Sun, W.; Yuan, Y.X. Optimization Theory and Methods: Nonlinear Programming. In Optimization Theory and Methods: Nonlinear Programming; Springer: New York, NY, USA, 2006. [Google Scholar]
  9. Shi, Z.J. Convergence of line search methods for unconstrained optimization. Appl. Math. Comput. 2004, 151, 393–405. [Google Scholar] [CrossRef]
  10. Wolfe, P. Convergence conditions for ascent methods. SIAM Rev. 1968, 11, 226–235. [Google Scholar] [CrossRef]
  11. Ortega, J.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equation in Several Variables. In Iterative Solution of Nonlinear Equation in Several Variables; Academic Press: Cambridge, MA, USA, 1970. [Google Scholar]
  12. Armijo, L. Minimization of functions having Lipschitz continuous first partial derivatives. Pac. J. Math. 2008, 16, 1–3. [Google Scholar] [CrossRef] [Green Version]
  13. Brezinski, C. A classification of quasi-Newton methods. Numer. Algor. 2003, 33, 123–135. [Google Scholar] [CrossRef]
  14. Stanimirovic, P.S.; Miladinović, M.B. Accelerated gradient descent methods with line search. Numer. Algor. 2010, 54, 503–520. [Google Scholar] [CrossRef]
  15. Andrei, N. An acceleration of gradient descent algoritham with backtracing for unconstrained optimization. Numer. Algor. 2006, 42, 63–173. [Google Scholar] [CrossRef]
  16. Petrović, M.J.; Stanimirovic, P.S. Accelerated Double Direction Method For Solving Unconstrained Optimization Problems. Math. Probl. Eng. 2014, 2014, 965104. [Google Scholar] [CrossRef] [Green Version]
  17. Petrović, M.J. An accelerated Double Step Size method in unconstrained optimization. Appl. Math. Comput. 2015, 250, 309–319. [Google Scholar] [CrossRef]
  18. Stanimirovic, P.S.; Petrović, M.J.; Milovanović, G.V. A Transformation of Accelerated Double Step Size Method for Unconstrained Optimization. Math. Probl. Eng. 2015, 2015, 283679. [Google Scholar] [CrossRef]
  19. Petrović, M.J.; Valjarević, D.; Ilić, D.; Valjarević, A.; Mladenović, J. An improved modification of accelerated double direction and double step-size optimization schemes. Mathematics 2022, 10, 259. [Google Scholar] [CrossRef]
  20. Barzilai, B.; Borwein, J.M. Two point step-size gradient method. IMA J. Numer. Anal. 1988, 8, 141–148. [Google Scholar] [CrossRef]
  21. Miladinović, M.; Stanimirović, P.S.; Miljković, S. Scalar correction method for solving large scale’ unconstrained minimization problems. J. Optim. Theory Appl. 2011, 151, 304–320. [Google Scholar] [CrossRef]
  22. Andrei, N. A new three-term conjugate gradient algorithm for unconstrained optimization. Numer. Algor. 2014, 68, 305–321. [Google Scholar] [CrossRef]
  23. Djuranović-Miličić, N.I.; Gardašević-Filipović, M. A multi-step curve search algorithm in nonlinear optimization: Nondifferentiable convex case. Facta Univ. Ser. Math. Inform. 2010, 25, 11–24. [Google Scholar]
  24. Picard, E. Memoire sur la theorie des equations aux derivees partielles et la methode des approximations successives. J. Math. Pures Appl. 1890, 6, 145–210. [Google Scholar]
  25. Mann, W.R. Mean value methods in iterations. Proc. Am. Math. Soc. 1953, 4, 506–510. [Google Scholar] [CrossRef]
  26. Ishikawa, S. Fixed points by a new iteration method. Proc. Am. Math. Soc. 1974, 44, 147–150. [Google Scholar] [CrossRef]
  27. Khan, S.H. A Picard-Mann hybrid iterative process. Fixed Point Theory Appl. 2013, 2013, 69. [Google Scholar] [CrossRef]
  28. Petrović, M.J.; Rakočević, V.; Kontrec, N.; Panić, S.; Ilić, D. Hybridization Accel. Gradient Descent Method. Numer. Algor. 2018, 79, 769–786. [Google Scholar] [CrossRef]
  29. Petrović, M.J.; Stanimirović, P.S.; Kontrec, N.; Maldenović, J. Hybrid modification of accelerated double direction method. Math. Probl. Eng. 2018, 2018, 1523267. [Google Scholar] [CrossRef]
  30. Petrović, M.J. Hybridization Rule Applied on Accelerated Double Step Size Optimization Scheme. Filomat 2019, 33, 655–665. [Google Scholar] [CrossRef]
  31. Petrović, M.J.; Rakočević, V.; Valjarević, D.; Ilić, D. A note on hybridization process applied on transformed double step size model. Numer. Algor. 2020, 85, 449–465. [Google Scholar] [CrossRef]
  32. Ivanov, M.J.; Stanimirović, P.S.; Milovanović, G.V.; Djordjević, S.; Brajević, I. Accelerated Multi Step-Size Methods Solving Unconstrained Optimization Problems. Optim. Method Softw. 2020, 85, 449–465. [Google Scholar] [CrossRef]
  33. Petrović, M.J.; Panić, S.; Carerević, M.M. Initial improvement of the hybrid accelerated gradient descent process. Bull. Aust. Math. Soc. 2018, 98, 331–338. [Google Scholar] [CrossRef]
  34. Dolan, E.; Moré, J. Benchmarking optimization software with performance profiles. Math. Program 2002, 91, 201–213. [Google Scholar] [CrossRef]
  35. Andrei, N. An Unconstrained Optimization Test Functions Collection. Adv. Model. Optim. 2008, 10, 147–161. Available online: https://camo.ici.ro/journal/vol10/v10a10.pdf (accessed on 16 November 2022).
Figure 1. Performance profiles for the HSM and SM methods regarding the number of iterations metric.
Figure 1. Performance profiles for the HSM and SM methods regarding the number of iterations metric.
Mathematics 10 04411 g001
Figure 2. Performance profiles for the HSM and SM methods regarding the CPU time metric.
Figure 2. Performance profiles for the HSM and SM methods regarding the CPU time metric.
Mathematics 10 04411 g002
Figure 3. Performance profiles for the HSM and SM methods regarding the number of function evaluations metric.
Figure 3. Performance profiles for the HSM and SM methods regarding the number of function evaluations metric.
Mathematics 10 04411 g003
Figure 4. Performance profiles for the HADD and ADD methods regarding the number of iterations metric.
Figure 4. Performance profiles for the HADD and ADD methods regarding the number of iterations metric.
Mathematics 10 04411 g004
Figure 5. Performance profiles for the HADD and ADD methods regarding the CPU time metric.
Figure 5. Performance profiles for the HADD and ADD methods regarding the CPU time metric.
Mathematics 10 04411 g005
Figure 6. Performance profiles for the HADD and ADD methods regarding the number of function evaluations metric.
Figure 6. Performance profiles for the HADD and ADD methods regarding the number of function evaluations metric.
Mathematics 10 04411 g006
Figure 7. Performance profiles for the HTADSS and TADSS methods regarding the number of iterations metrics.
Figure 7. Performance profiles for the HTADSS and TADSS methods regarding the number of iterations metrics.
Mathematics 10 04411 g007
Figure 8. Performance profiles for the HTADSS and TADSS methods regarding the CPU time metric.
Figure 8. Performance profiles for the HTADSS and TADSS methods regarding the CPU time metric.
Mathematics 10 04411 g008
Figure 9. Performance profiles for the HTADSS and TADSS methods regarding the number of function evaluations metric.
Figure 9. Performance profiles for the HTADSS and TADSS methods regarding the number of function evaluations metric.
Mathematics 10 04411 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rakočević, V.; Petrović, M.J. Comparative Analysis of Accelerated Models for Solving Unconstrained Optimization Problems with Application of Khan’s Hybrid Rule. Mathematics 2022, 10, 4411. https://doi.org/10.3390/math10234411

AMA Style

Rakočević V, Petrović MJ. Comparative Analysis of Accelerated Models for Solving Unconstrained Optimization Problems with Application of Khan’s Hybrid Rule. Mathematics. 2022; 10(23):4411. https://doi.org/10.3390/math10234411

Chicago/Turabian Style

Rakočević, Vladimir, and Milena J. Petrović. 2022. "Comparative Analysis of Accelerated Models for Solving Unconstrained Optimization Problems with Application of Khan’s Hybrid Rule" Mathematics 10, no. 23: 4411. https://doi.org/10.3390/math10234411

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop