Next Article in Journal
Evolution of Tax Exemption Policy and Pricing Strategy Selection in a Competitive Market
Previous Article in Journal
Optimization of Interaction with Counterparties: Selection Game Algorithm under Uncertainty
Previous Article in Special Issue
Convergence Analysis for Yosida Variational Inclusion Problem with Its Corresponding Yosida Resolvent Equation Problem through Inertial Extrapolation Scheme
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

About the Subgradient Method for Equilibrium Problems

by
Abdellatif Moudafi
L.I.S UMR CNRS 7296, Aix Marseille Université, Campus Universitaire de Saint-Jérôme, Avenue Escadrille Normandie-Niemen, 13397 Marseille, France
Mathematics 2024, 12(13), 2081; https://doi.org/10.3390/math12132081
Submission received: 12 June 2024 / Revised: 26 June 2024 / Accepted: 1 July 2024 / Published: 2 July 2024
(This article belongs to the Special Issue Applied Functional Analysis and Applications: 2nd Edition)

Abstract

:
Convergence results of the subgradient algorithm for equilibrium problems were mainly obtained using a Lipschitz continuity assumption on the given bifunctions. In this paper, we first provide a complexity result for monotone equilibrium problems without assuming Lipschitz continuity. Moreover, we give a convergence result of the value of the averaged sequence of iterates beyond Lipschitz continuity. Next, we derive a rate convergence in terms of the distance to the solution set relying on a growth condition. Applications to convex minimization and min–max problems are also stated. These ideas and results deserve to be developed and further refined.

1. Introduction

The subgradient algorithm has recently enjoyed regained popularity, see for example [1,2,3,4]. However, the Lipschitz continuity assumption can be stringent even in a convex setting. It is well known that not only SVM (Support Vector Machine) but also Feed-Forward and Recurrent Neural Networks do not satisfy the Lipschitz continuity assumption, see [5]. Indeed, most l 2 -regularized convex learning problems lack this property, while l 2 regularization and weight decay are ubiquitous in the learning field. The  proposed equilibrium approach will allow us not only to generalize some very recent results in convex minimization [1] and present them for min–max problems in a unified way but also to yield new insights into equilibrium problems. In order to go to the essential to share, we took the same paper outline as in [1] and we assume the reader has some basic knowledge of variational and convex analysis as can be found, for example, in [6,7].
The subgradient algorithm is a powerful tool for constructing algorithms to approximate solutions of optimization and equilibrium problems, the latter being the problem of finding x ¯ C such that
( P ) find x ¯ C , F ( x ¯ , y ) 0 y C ,
where C is a given nonempty closed convex subset of d and F : C × C is a bifunction. The solution set of ( P ) will be denoted by Γ . Such a problem is also known as the equilibrium problem in the sense of Blum and Oettli [8]. It is worth mentioning that variational inequalities, Nash equilibrium, the saddle-point problem, optimization, and many problems arising in applied nonlinear analysis are special cases of equilibrium problems. In this paper, we will be concerned with the convergence analysis of a subgradient method for solving problem ( P ) in which F is not assumed to be Lipschitz continuous. For more simplicity and clarity, we suppose that F verifies the usual conditions:
( A 1 ) F ( x , x ) = 0 f o r a l l x , y C ;
( A 2 ) F i s m o n o t o n e F ( x , y ) + F ( y , x ) 0 for all x , y C ;
( A 3 ) lim sup t 0 F ( t z + ( 1 t ) x , y ) F ( x , y ) f o r a n y x , y , z C ;
( A 4 ) for each x C , y F ( x , y ) is convex and lower semicontinuous.

2. The Main Results

Let us state the subgradient method which generates a sequence ( x k ) k by Algorithm 1:
Algorithm 1 (SGM)
Step 0: Choose x 0 C and set k = 0 .
Step 1: Let x k C and compute η k 2 F ( x k , x k ) , where 2 F stands for the convex subdifferential of F with respect to the second variable.
If η k = 0 , then stop.
Step 2: Update x k + 1 = P C x k α k η k , where α k = β k η k , for all k 0 with β k > 0 .
Now we are in a position to provide a complexity result.
Theorem 1.
Suppose F verifies conditions A 1 to A 4 and note by T the total number of iterations. Set x ¯ T = i = 0 T x i T + 1 , β k = c T + 1 , and c > 0 and let x ¯ be a solution of ( P ) . Then, x k B ( x ¯ , r ) for k = 0 , 1 , · · · , T with r = x 0 x ¯ 2 + c 2 . Moreover, we have
F ( x ¯ , x ¯ T ) r 2 L B 2 r T + 1 ,
where L B is the local Lipschitz continuity constant of F ( x ¯ , · ) on the smallest open convex set O such that the closed ball B ( x ¯ , r ) O .
Proof. 
For all x C , we have
x k + 1 x 2 x k x α k η k 2 x k x 2 2 α k η k , x k x + β k 2 .
On the other hand, η k 2 F ( x k , x k ) implies that F ( x k , x ) η k , x x k , which combined with the monotonicity of F ensures that η k , x k x F ( x , x k ) . Consequently,
x k + 1 x 2 x k x 2 2 β k η k F ( x , x k ) + β k 2 .
Setting x = x ¯ , we have
x k + 1 x ¯ 2 x k x ¯ 2 + β k 2 ,
because β k > 0 together with F ( x ¯ , x k ) 0 . In other words, the sequence ( x k ) k is quasi-Fejér monotone to the solution set Γ .
From the latter inequality, we infer that
x k + 1 x ¯ 2     x 0 x ¯ 2 + i = 0 k β i 2 x 0 x ¯ 2 + c 2 , for 0 k T .
Relation (4) assures that x k B ( x ¯ , r ) for 0 k T .
In light of the consequence of [6]—Proposition 9.13, we have that F ( x ¯ , · ) is locally Lipschitz continuous with constant L B on O which, for 0 k T , in turn gives that η k   L B . This combined with (2) leads to
x k + 1 x ¯ 2 x k x ¯ 2 2 β k L B F ( x ¯ , x k ) + β k 2 .
From which we derive
i = 0 T β i F ( x ¯ , x k ) L B 2 x 0 x ¯ 2 + i = 0 T β i 2 ,
or equivalently
i = 0 T c T + 1 F ( x ¯ , x k ) L B 2 x 0 x ¯ 2 + c 2 .
Using the convexity of the function F ( x ¯ , · ) , we finally obtain
F ( x ¯ , x ¯ T ) = F x ¯ , i = 0 T 1 T + 1 x i i = 0 T 1 T + 1 F ( x ¯ , x i ) r 2 L B 2 r T + 1 .
This completes the proof. □
A convergence result of the values of the averaged sequence of iterates generated by (SGM) is provided in the next Theorem.
Theorem 2.
Suppose that the bifunction F verifies conditions A 1 to A 4 . Set x ¯ k = i = 0 k β i x i i = 0 k β i , σ k = i = 0 k β i , and ( β k ) k l 2 / l 1 and let x ¯ be a solution of ( P ) . Then, the sequence ( x k ) k B ( x ¯ , r ¯ ) where r ¯ = x 0 x ¯ 2 + c ¯ and c ¯ > 0 is such that k = 0 β k 2 c ¯ . Moreover,
k , F ( x ¯ , x ¯ k ) r 2 L B 2 σ k ,
where L B is the local Lipschitz continuity constant of F ( x ¯ , · ) on the smallest open convex set O such that B ( x ¯ , r ) O .
Proof. 
We again have that
x k + 1 x ¯ 2 x k x ¯ 2 2 β k η k F ( x ¯ , x k ) + β k 2 .
Therefore,
x k + 1 x ¯ 2 x 0 x ¯ 2 + i = 0 k β i 2 x 0 x ¯ 2 + c ¯ .
This ensures that ( x k ) k B ( x ¯ , r ¯ ) . Following the same lines as in the proof of Theorem 1, we obtain
x k + 1 x ¯ 2     x k x ¯ 2 2 β k L B F ( x ¯ , x k ) + β k 2 , for k .
This together with the convexity of the function F ( x ¯ , · ) yields to
F ( x ¯ , x ¯ k ) = F x ¯ , i = 0 k β i σ k x i i = 0 k β i σ k F ( x ¯ , x i ) r 2 L B 2 σ k ,
which leads to the announced result. □
Remark 1.
We can obtain more than the convergence of the averaged sequence of the iterates. Actually, we have
lim inf k + F ( x ¯ , x k ) = 0 ,
and the whole sequence ( x k ) k converges to a solution of ( P ) .
Indeed, for all k 0 , the inequality
x k + 1 x ¯ 2 x k x 2 2 β k L B F ( x ¯ , x k ) + β k 2 ,
leads to
k = 0 β k F ( x ¯ , x k ) < + .
Since k = 0 β k = + , this implies that lim inf k + F ( x ¯ , x k ) = 0 .
Now, with the sequence ( x k ) k being quasi-Fejér convergent to the set Γ, namely verifying (3) with ( β k ) k l 2 , this implies its boundedeness. Further, in view of (9), which is still valid for all x C together with both lower semicontinuity and upper hemi-continuity assumptions, we obtain that any cluster point x * of ( x k ) k belongs to Γ. Consequently, the whole sequence converges to x * , see for example [9], and we retrieve the main results in [10].
A convergence result based on a growth property.
Corollary 1.
Suppose in addition to hypotheses A 1 to A 4 that F verifies the following growth property:
t h e r e   e x i s t s γ > 0 s u c h   t h a t F ( x ¯ , x ) γ d i s t 2 ( x , Γ ) f o r   a l l x C a n d   s o m e x ¯ Γ .
Consider the sequence ( β k ) k given by β k = L B γ ( k + 1 ) for all k . Then, for all k , ( x k ) k B ( x ¯ , r ¯ ) , and we have
d i s t 2 ( x k + 1 , Γ ) L B 2 γ 2 ( k + 1 ) ,
where L B is the local Lipschitz continuity constant of F ( x ¯ , · ) on the smallest open convex set O such that B ( x ¯ , r ) O .
Proof. 
The beginning of Theorem 2 ensures that ( x k ) k lies in B ( x ¯ , r ¯ ) and assures again that η k L B . Remember that (7) reads as
x k + 1 x ¯ 2 x k x ¯ 2 2 β k η k F ( x ¯ , x k ) + β k 2 .
By taking x ¯ = p r o j Γ ( x k ) , we obtain
d i s t 2 ( x k + 1 , Γ ) x k + 1 x ¯ 2 d i s t 2 ( x k , Γ ) 2 β k L B F ( x ¯ , x k ) + β k 2 .
This implies
d i s t 2 ( x k + 1 , Γ ) d i s t 2 ( x k , Γ ) 2 k + 1 d i s t 2 ( x k , Γ ) + L B 2 γ 2 ( k + 1 ) 2 ,
or equivalently
( k + 1 ) 2 d i s t 2 ( x k + 1 , Γ ) ( k + 1 ) 2 2 ( k + 1 ) d i s t 2 ( x k , Γ ) + L B 2 γ 2 .
Following the same lines as in the proof of [1]—Corollary 1, for all k , we further have
( k + 1 ) 2 d i s t 2 ( x k + 1 , Γ ) k 2 d i s t 2 ( x k , Γ ) L B 2 γ 2 .
Summing the last inequality for i = 1 to k leads to
( k + 1 ) 2 d i s t 2 ( x k + 1 , Γ ) = i = 0 k ( i + 1 ) 2 d i s t 2 ( x i + 1 , Γ ) i 2 d i s t 2 ( x i , Γ ) L B 2 γ 2 ( k + 1 ) .
From which we derive
d i s t 2 ( x k + 1 , Γ ) L B 2 γ 2 ( k + 1 ) .

3. Applications

In the convex minimization case, (SGM) coincides with the classical subgradient method, and we recover the results obtained in the convex setting in [1]. Indeed, just take F ( x , y ) = f ( y ) f ( x ) , f being a proper convex lower semicontinuous convex function, clearly 2 F ( x k , x k ) = f ( x k ) , and we retrieve
( S G M ) c o m p u t e η k f ( x k ) U p d a t e x k + 1 = P C ( x k α k η k ) .
Theorem 1 reduces to
f ( x ¯ k ) f ( x ¯ ) r 2 L B 2 r T + 1 ,
where x ¯ is a minimizer of f and L B is the local Lipschitz continuity constant of f on the smallest open convex set O such that B ( x ¯ , r ) O . Theorem 2, in turn, leads to the fact that the sequence ( x k ) k B ( x ¯ , r ¯ ) , where r ¯ = x 0 x ¯ 2 + c ¯ and c ¯ > 0 , is such that k = 0 β k 2 c ¯ . Moreover, for all k , we obtain the following convergence result in terms of the suboptimality error:
f ( x ¯ k ) f ( x ¯ ) r 2 L B 2 σ k ,
x ¯ k , σ k , r ¯ , c ¯ were defined in Theorem 2, and ( β k ) k l 2 / l 1 and x ¯ are minimizers of f.
If we assume, in addition, for all x C that
f ( x ) f ( x ¯ ) γ d i s t 2 ( x , Γ ) ,
then
d i s t 2 ( x k + 1 , Γ ) L B 2 γ 2 ( k + 1 ) .
Likewise, by taking F ( v , w ) = L ( z , p ) L ( x , y ) with w = ( z , y ) , v = ( x , p ) and L being a proper closed convex–concave function defined on C × C , C = Q × P with Q , P closed convex sets of n and m , respectively, then clearly
w F ( v , w ) = 1 L ( · , p ) , 2 L ( x , · ) ,
we recover the subgradient algorithm for the saddle function considered in [11] and more recently in [12] and we extend some results in [1] to the convex–concave case. More precisely, (SGM) reduces to
( S G M ) compute ( μ k , γ k ) ( 1 L ( x k , y k ) , 2 ( L ) ( x k , y k ) ) Update ( x k + 1 , y k + 1 ) = P C ( x k , y k ) α k ( μ k , γ k ) .
Theorem 1 reads, in this case, as
L ( x ¯ T , y ¯ ) L ( x ¯ , y ¯ T ) r 2 L B 2 r T + 1 ,
where r = ( x 0 , y 0 ) ( x ¯ , y ¯ ) 2 + c 2 and ( x ¯ , y ¯ ) are a saddle point of L, namely ( x ¯ , y ¯ ) verifying L ( x ¯ , y ) L ( x ¯ , y ¯ ) L ( x , y ¯ ) for all ( x , y ) P × Q and L B being the local Lipschitz continuity constant of L ( · , x ¯ ) L ( x ¯ , · ) on the smallest open convex set O such that B ( x ¯ , r ) O .
Theorem 2, in turn, leads to the sequence ( x k , y k ) k B ( x ¯ , y ¯ ) , r ¯ , where r ¯ = ( x 0 , y 0 ) ( x ¯ , y ¯ ) 2 + c ¯ and c ¯ > 0 is such that k = 0 β k 2 c ¯ . Moreover, for all k , we obtain the following convergence result in terms of a merit function:
L ( x ¯ k , y ¯ ) L ( x ¯ , y ¯ k ) r 2 L B 2 σ k ,
( x ¯ k , y ¯ k ) , σ k , r ¯ , c ¯ were defined in Theorem 2 and β k l 2 / l 1 and ( x ¯ , y ¯ ) are saddle points of L.
Now, if in addition, we assume for all ( x , y ) C that
L ( x , y ¯ ) L ( x ¯ , y ) γ d i s t 2 ( ( x , y ) , Γ ) ,
then
d i s t 2 ( x k + 1 , y k + 1 ) , Γ L B 2 γ 2 ( k + 1 ) .

4. Conclusions

To conclude, the equilibrium approach allowed us not only to generalize some very recent results obtained by X. Li, L. Zhao, D. Zhu, and A. M-Ch.; use them in convex minimization [3]; and present them for min–max problems in a unified way but also to yield new insights into equilibrium problems. More precisely, a complexity result for monotone equilibrium problems was provided without assuming Lipschitz continuity. Moreover, a convergence result of the value of the averaged sequence of iterates beyond Lipschitz continuity was also given. Finally, a rate convergence in terms of the distance to the solution set relying on a growth condition was derived. The proposed results were declined in the convex minimization context as well as in the convex–concave min–max setting.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study.

Acknowledgments

I would like to thank Zoubida Marzaq for her invitation to Saint-Sébastien and wish to thank the “Image & Modèle” team and also the Computer Science System Laboratory (L.I.S.) at Aix Marseille University for their support.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Li, X.; Zhao, L.; Zhu, D.; So, A.M.C. Revisiting subgradient method: Complexity and Convergence Beyond Lipschitz Continuity. arXiv 2023, arXiv:2305.14161. [Google Scholar]
  2. Wang, J.; Xu, H.K. Stepsize Choice for Korpelevich’s and Popov’s extragradient Algorithms for Convex-Concave Minimax Problems. Carpathian J. Math. 2023, 39, 315–323. [Google Scholar] [CrossRef]
  3. Abass, H.A.; Narain, O.K.; Onifade, O.M. Inertial extrapolation method for solving systems of monotone variational inclusion and fixed point problems using Bregman distance approach. Nonlinear Funct. Anal. Appl. 2023, 28, 497–520. [Google Scholar]
  4. Kheawborisut, A.; Khuangsatung, W. A modified Krasnoselskii-type subgradient extragradient algorithm with inertial effects for solving variational inequality problems and fixed point problem. Nonlinear Funct. Anal. Appl. 2024, 29, 393–418. [Google Scholar]
  5. Patel, V.; Berahas, A.S. Gradient Descent in the Absence of Global Lipschitz Continuity of the Gradients. arXiv 2023, arXiv:2210.02418v2. [Google Scholar]
  6. Rockafellar, R.T.; Wets, R.J.-B. Variational Analysis; Springer: Berlin/Heidelberg, Germany, 1998. [Google Scholar]
  7. Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  8. Blum, E.; Oettli, W. From optimization and variational inequalities to equilibrium problems. Math. Stud. 1994, 63, 123–145. [Google Scholar]
  9. Iusem, A.N.; Svaiter, B.F.; Teboulle, M. Entropy-like proximal methods in convex programming. Math. Oper. Res. 1994, 9, 790–814. [Google Scholar] [CrossRef]
  10. Neto, J.X.C.; Santos, P.S.M.; Scheimberg, S.; Souza, S.S. A Subgradient Projected Algorithm for Nonconvex Equilibrium Problems; CLAIO, SBPO: Rio de Janeiro, Brazil, 2012; pp. 3522–3529. [Google Scholar]
  11. Nedic, A.; Ozdaglar, A. Subgradient Methods for Saddle-Point Problems. J. Optim. Theory Appl. 2009, 142, 205–228. [Google Scholar] [CrossRef]
  12. Yen, L.H.; Muu, L.D. A subgradient method for equilibrium problems involving quasiconvex bifunctions. Oper. Res. Lett. 2020, 48, 579–583. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moudafi, A. About the Subgradient Method for Equilibrium Problems. Mathematics 2024, 12, 2081. https://doi.org/10.3390/math12132081

AMA Style

Moudafi A. About the Subgradient Method for Equilibrium Problems. Mathematics. 2024; 12(13):2081. https://doi.org/10.3390/math12132081

Chicago/Turabian Style

Moudafi, Abdellatif. 2024. "About the Subgradient Method for Equilibrium Problems" Mathematics 12, no. 13: 2081. https://doi.org/10.3390/math12132081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop