1. Introduction
The subgradient algorithm has recently enjoyed regained popularity, see for example [
1,
2,
3,
4]. However, the Lipschitz continuity assumption can be stringent even in a convex setting. It is well known that not only SVM (Support Vector Machine) but also Feed-Forward and Recurrent Neural Networks do not satisfy the Lipschitz continuity assumption, see [
5]. Indeed, most 
-regularized convex learning problems lack this property, while 
 regularization and weight decay are ubiquitous in the learning field. The  proposed equilibrium approach will allow us not only to generalize some very recent results in convex minimization [
1] and present them for min–max problems in a unified way but also to yield new insights into equilibrium problems. In order to go to the essential to share, we took the same paper outline as in [
1] and we assume the reader has some basic knowledge of variational and convex analysis as can be found, for example, in [
6,
7].
The subgradient algorithm is a powerful tool for constructing algorithms to approximate solutions of optimization and equilibrium problems, the latter being the problem of finding 
 such that
      
      where 
C is a given nonempty closed convex subset of 
 and 
 is a bifunction. The solution set of 
 will be denoted by 
. Such a problem is also known as the equilibrium problem in the sense of Blum and Oettli [
8]. It is worth mentioning that variational inequalities, Nash equilibrium, the saddle-point problem, optimization, and many problems arising in applied nonlinear analysis are special cases of equilibrium problems. In this paper, we will be concerned with the convergence analysis of a subgradient method for solving problem 
 in which 
F is not assumed to be Lipschitz continuous. For more simplicity and clarity, we suppose that 
F verifies the usual conditions:
;
 for all ;
;
 for each  is convex and lower semicontinuous.
  2. The Main Results
Let us state the subgradient method which generates a sequence 
 by Algorithm 1:
      
| Algorithm 1 (SGM) | 
| Step 0: Choose  and set . | 
| Step 1: Let  and compute , where  stands for the convex subdifferential of F with respect to the second variable. | 
| If , then stop. | 
| Step 2: Update , where , for all  with . | 
Now we are in a position to provide a complexity result.
Theorem 1. Suppose F verifies conditions  to  and note by T the total number of iterations. Set , , and  and let  be a solution of . Then,  for  with . Moreover, we havewhere  is the local Lipschitz continuity constant of  on the smallest open convex set  such that the closed ball .  Proof.  On the other hand, 
 implies that 
, which combined with the monotonicity of 
F ensures that 
. Consequently,
        
Setting 
, we have
        
        because 
 together with 
. In other words, the sequence 
 is quasi-Fejér monotone to the solution set 
.
From the latter inequality, we infer that
        
Relation (4) assures that .
In light of the consequence of [
6]—Proposition 9.13, we have that 
 is locally Lipschitz continuous with constant 
 on 
 which, for  
, in turn gives that 
. This combined with (2) leads to
        
From which we derive
        
        or equivalently
        
Using the convexity of the function 
, we finally obtain
        
This completes the proof.    □
 A convergence result of the values of the averaged sequence of iterates generated by (SGM) is provided in the next Theorem.
Theorem 2. Suppose that the bifunction F verifies conditions  to . Set , , and  and let  be a solution of . Then, the sequence  where  and  is such that . Moreover,where  is the local Lipschitz continuity constant of  on the smallest open convex set  such that .  Proof.  This ensures that 
. Following the same lines as in the proof of Theorem 1, we obtain
        
This together with the convexity of the function 
 yields to
        
        which leads to the announced result.   □
 Remark 1. We can obtain more than the convergence of the averaged sequence of the iterates. Actually, we haveand the whole sequence  converges to a solution of . Indeed, for all , the inequalityleads to Since , this implies that .
Now, with the sequence  being quasi-Fejér convergent to the set Γ, namely verifying (3) with , this implies its boundedeness. Further, in view of (9), which is still valid for all  together with both lower semicontinuity and upper hemi-continuity assumptions, we obtain that any cluster point  of  belongs to Γ. Consequently, the whole sequence converges to , see for example [9], and we retrieve the main results in [10].  A convergence result based on a growth property.
Corollary 1. Suppose in addition to hypotheses  to  that F verifies the following growth property: Consider the sequence  given by  for all . Then, for all , , and we havewhere  is the local Lipschitz continuity constant of  on the smallest open convex set  such that .  Proof.  The beginning of Theorem 2 ensures that 
 lies in 
 and assures again that 
. Remember that (7) reads as
        
By taking 
, we obtain
        
This implies
        
        or equivalently
        
Following the same lines as in the proof of [
1]—Corollary 1, for all 
, we further have
        
Summing the last inequality for 
 to 
k leads to
        
□
   3. Applications
In the convex minimization case, (SGM) coincides with the classical subgradient method, and we recover the results obtained in the convex setting in [
1]. Indeed, just take 
, 
f being a proper convex lower semicontinuous convex function, clearly 
, and we retrieve
      
Theorem 1 reduces to
      
      where 
 is a minimizer of 
f and 
 is the local Lipschitz continuity constant of 
f on the smallest open convex set 
 such that 
. Theorem 2, in turn, leads to the fact that the sequence 
, where 
 and 
, is such that 
. Moreover, for all 
, we obtain the following convergence result in terms of the suboptimality error:
 were defined in Theorem 2, and 
 and 
 are minimizers of 
f.
If we assume, in addition, for all 
 that
      
      then
      
Likewise, by taking 
 with 
, 
 and 
L being a proper closed convex–concave function defined on 
, 
 with 
 closed convex sets of 
 and 
, respectively, then clearly
      
      we recover the subgradient algorithm for the saddle function considered in [
11] and more recently in [
12] and we extend some results in [
1] to the convex–concave case. More precisely, (SGM) reduces to
      
Theorem 1 reads, in this case, as
      
      where 
 and 
 are a saddle point of 
L, namely 
 verifying 
 for all (
 and 
 being the local Lipschitz continuity constant of 
 on the smallest open convex set 
 such that 
.
Theorem 2, in turn, leads to the sequence 
, where 
 and 
 is such that 
. Moreover, for all 
, we obtain the following convergence result in terms of a merit function:
 were defined in Theorem 2 and 
 and 
 are saddle points of 
L.
Now, if in addition, we assume for all 
 that
      
      then