Abstract
In this paper, by making use of some advanced tools from variational analysis and generalized differentiation, we establish necessary optimality conditions for a class of robust fractional minimax programming problems. Sufficient optimality conditions for the considered problem are also obtained by means of generalized convex functions. Additionally, we formulate a dual problem to the primal one and examine duality relations between them. In our results, by using the obtained results, we obtain necessary and sufficient optimality conditions for a class of robust fractional multi-objective optimization problems.
MSC:
90C29; 90C32; 90C46; 90C47
1. Introduction
A fractional multi-objective optimization problem with locally Lipschitzian data in the face of data uncertainty in the constraints is of the form
where is the vector of decision variables, and are uncertain parameters, are uncertainty sets; let be locally Lipschitz functions and let be given functions.
We will treat the problem (1) via a robust optimization approach (see, e.g., [1,2,3,4,5,6,7,8,9]), by following its robust counterpart as shown below,
where is the feasible set of the problem (2), defined by
In what follows, for the sake of convenience, we assume that for all and that for the reference point Furthermore, denote for simplicity.
In this paper, we mainly focus on the study of optimality conditions for a weak Pareto solution of the problem (2) by taking the associated minimax optimization problem into account (see, e.g., [10,11,12,13,14,15,16,17,18,19,20,21] for more related works). Let us consider the following robust fractional minimax programming problem,
where is the feasible set that is the same as defined by the problem (2).
Among others, we first establish the necessary optimality conditions for the problem (4) and its sufficient optimality conditions by means of generalized convex functions. Then, we formulate a dual problem to the problem (4) and examine duality relations between them. Finally, by using the obtained results, we obtain necessary and sufficient optimality conditions for the problem (2), the minimax program-based approach is different from some works in the literature; see, e.g., [22,23,24,25]. The main tools are from variational analysis and generalized differentiation; see, e.g., [26,27].
2. Preliminaries
In this section, we recall some notation and preliminary results that will be used in this paper; see e.g., [26,27]. Let denote the Euclidean space equipped with the usual Euclidean norm The nonnegative orthant of is denoted by As usual, the polar cone of a set is defined by
A given set-valued mapping is said to be closed at if for any sequence and for any sequence one has
Given a multifuction with values in the collection of all the subsets of (and similarly, of course, in infinite dimensions). The limiting construction
is known as the Painlevé–Kuratowski upper/outer limit of F at
Given and define the collection of Fréchet/regular normal cone to at by
If we put
The Mordukhovich/limiting normal cone to at is obtained from regular normal cones by taking the Painlevé–Kurotowski upper limits as
If we put
For an extended real-valued function its domain is defined by
and its epigraph is defined by
Let be finite at Then, the collection of basic subgradients, or the (basic/Mordukhovich/limiting) subdifferential, of at is defined by
We say a function is locally Lipschitz at with rank if there exists such that
where stands for the closed unit ball centered at of radius Furthermore, it also holds that [27] (Theorem 1.22)
The generalized Fermat’s rule is formulated as follows [27] (Proposition 1.30): Let be finite at If is a local minimizer of then
To establish optimality conditions, the following lemmas, which are related to the Mordukhovich/limiting subdifferential calculus, are very useful.
Lemma 1
([27] (Corollary 2.21 and Theorem 4.10 (ii))).
- (i)
- Let be lower semicontinuous around and let all but one of these functions be Lipschitz-continuous around Then,
- (ii)
- Let be lower semicontinuous around for and upper semicontinuous at for suppose that each is Lipschitz continuous around Then, we have the inclusionwhere the equality holds and the maximum functions are lower regular at if each is lower and regular at this point and the sets and are defined as follows:
Lemma 2
([27] (Corollary 4.8)). Let for be Lipschitz continuous around with Then, we have
Lemma 3
([27] Mean Value Inequality (Corollary 4.14 (ii))). If φ is Lipschitz continuous on an open set containing then
where and
For a function being locally Lipschitz continuous at the generalized (in the sense of Clarke) of at in the direction is defined as follows:
In this case, the convexified/Clarke subdifferential of at is the set
which is nonempty, and the Clarke directional derivative is the support function of the Clarke subdifferential, that is,
for each
It follows from [27] that the relationship between the above subdifferentials of at is as follows:
3. Optimality Conditions for Robust Fractional Minimax Programming Problems
In this section, we establish optimality conditions for the robust fractional minimax programming problem (4), which will be pretty helpful for Section 5.
Definition 1.
Let us make some assumptions for functions given in (3). We refer the reader to [28] for more details.
- For a fixed there exists such that the function is upper semicontinuous for each and the functions are Lipschitz of given rank on i.e.,
- The multifunction is closed at for each where the symbol stands for the limiting subdifferential operation with respect to and the notation signifies active indices in at i.e.,with
As usual, in order to obtain the necessary optimality condition of the Karush–Kuhn–Tucker type for a locally optimal solution to the problem (4), the following constraint qualification (see [28] (Definition 3.2)) is needed, which turns into the extended Mangasarian–Fromovitz constraint qualification (see e.g., [29] (p. 72)) in the smooth setting.
Definition 2
([28] (Definition 3.2)). Let Then, the constraint qualification is fulfilled at if
Theorem 1.
Let the be fulfilled at Suppose that is a locally optimal solution to the problem (4); then there are some such that
Proof.
Let be a locally optimal solution to the problem (4). For we consider and as in assumptions and Along with the proof of [28] (Theorem 3.3), we can show that
is closed and compact, and
with the help of Lemma 3 and the relationship between and
Set Because is a locally optimal solution to the problem (4), there exists a neighborhood U of such that
We claim that is also a local minimizer of the following problem
where Actually, it is easy to see that due to Let us justify for Suppose that ; then, Otherwise, shows that
which contradicts (13). If then there exists such that which entails that Therefore, is a local minimizer for
Applying the nonsmooth version of Fermat’s rule (6), we obtain
Moreover, applying the formula for the limiting subdifferential of maximum functions and the limiting subdifferential sum rule for local Lipschitz functions (Lemma 1), we have
Applying the limiting subdifferential of maximum functions and the sum rule to we obtain
where Taking (8) into account, we arrive at
In what follows, we formulate sufficient optimality conditions for the problem (4). For this, we recall the following definition of generalized convexity; see [30,31].
Definition 3.
We say that is generalized convex on at if for any there exists such that
where are defined as in(10).
The following theorem provides sufficient optimality conditions for a globally optimal solution of (4).
Theorem 2.
Let the condition be satisfied at If is generalized convex on at then is a globally optimal solution to the problem (4).
Proof.
Assume to the contrary that is not a globally optimal solution to the problem (4); then, there is such that
By the generalized convex on at we deduce from (16) that for such there is such that
Hence,
Since
Thus, it follows from (18) that for and In addition, due to for and So, we obtain by (20) that
On the other hand, by (17), it holds that
4. Duality for Robust Fractional Minimax Programming Problems
Let
In connection with the robust fractional minimax programming problem (4), we consider a dual problem in the following form:
Here, we denote and
where is defined as in (10) by replacing by
Definition 4.
We say that a point is called a globally optimal solution to the problem (22) if and only if
Theorem 3
(weak duality). Let and let If is generalized convex at then
Proof.
Since there exist and such that
By the generalized convex on at we deduce from (23) that for such x there is such that
It stems from that
We obtain
Since
Thus, we obtain
where the last inequality holds true due to (25). Combining now (26) and (24), we have
This implies that
due to The proof is complete. □
Theorem 4
(strong duality). Let be a locally optimal solution to the problem (4) such that the is satisfied at this point. Then, there exists such that and
Furthermore, if is generalized convex on at any then is globally optimal solution to the problem (22).
Proof.
Thanks to Theorem 1, we can find and such that
Since we have for and Thus, it stems from (28) that for and This entails that
Thus, when is generalized convex on at any we apply the weak duality result in Theorem 3 to conclude that
for any This means that is a globally optimal solution to the problem (22). □
Theorem 5
(converse-like duality). Let be such that If and is generalized convex on at then is a globally optimal solution to the problem (4).
5. Application to Robust Fractional Multi-Objective Optimization Problems
In this section, we will apply the results from Section 3 to study optimality conditions for robust fractional multi-objective optimization problems.
Definition 5.
We say is a weak Pareto solution to the problem (2) if and only if
Theorem 6
([31] (cf. Theorem 5.1)). Suppose that is a weak Pareto solution to the problem (2); then, is an optimal solution of the following problem:
Proof.
Let be a weak Pareto solution to the problem (2). We set
and
On the contrary, assume that is not an optimal solution of the problem (33), and hence there exists such that
The following result is the Karush–Kuhn–Tucker (KKT) necessary conditions for weak Pareto solutions to the problem (2).
Theorem 7
(necessary condition). Let the be fulfilled at If is a weak Pareto solution of problem then there are some such that
Proof.
Theorem 8
(sufficient condition). Let satisfy condition (35). Suppose that is generalized convex on at then is a weak Pareto solution to the problem (2).
Author Contributions
Conceptualization, H.L. and Z.H.; methodology, Z.H. and D.S.K.; investigation, H.L. and Z.H.; resources, Z.H.; writing—original draft preparation, Z.H.; writing—review and editing, H.L. and D.S.K. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by the Fundamental Research Funds for the Central Universities (2412022QD024).
Data Availability Statement
No data were used for the research described in the paper.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Ben-Tal, A.; Nemirovski, A. Robust convex optimization. Math. Oper. Res. 1998, 23, 769–805. [Google Scholar] [CrossRef]
- Ben-Tal, A.; Nemirovski, A. Selected topics in robust convex optimization. Math. Program. 2008, 112, 125–158. [Google Scholar] [CrossRef]
- Ben-Tal, A.; El Ghaoui, L.; Nemirovski, A. Robust Optimization; Princeton University Press: Princeton, NJ, USA, 2009. [Google Scholar]
- Beck, A.; Ben-Tal, A. Duality in robust optimization: Primal worst equals dual best. Oper. Res. Lett. 2009, 37, 1–6. [Google Scholar] [CrossRef]
- Ben-Tal, A.; den Hertog, D.; Vial, J. Deriving robust counterparts of nonlinear uncertain inequalities. Math. Program. 2015, 149, 265–299. [Google Scholar] [CrossRef]
- El Ghaoui, L.; Lebret, H. Robust solution to least-squares problems with uncertain data. SIAM J. Matrix Anal. Appl. 1997, 18, 1035–1064. [Google Scholar] [CrossRef]
- El Ghaoui, L.; Oustry, F.; Lebret, H. Robust solutions to uncertain semidefinite programs. SIAM J. Optim. 1998, 9, 33–52. [Google Scholar] [CrossRef]
- Kouvelis, P.; Yu, G. Robust Discrete Optimization and Its Applications; Springer: New York, NY, USA, 1997. [Google Scholar]
- Lee, G.M.; Son, P.T. On nonsmooth optimality theorems for robust optimization problems. Bull. Korean Math. Soc. 2014, 51, 287–301. [Google Scholar] [CrossRef]
- Antczak, T. Generalized fractional minimax programming with B-(p, r)-invexity. Comput. Math. Appl. 2008, 56, 1505–1525. [Google Scholar] [CrossRef]
- Antczak, T. Nonsmooth minimax programming under locally Lipschitz (Φ, ρ)-invexity. Appl. Math. Comput. 2008, 217, 9606–9624. [Google Scholar] [CrossRef]
- Antczak, T. Parametric approach for approximate efficiency of robust multiobjective fractional programming problems. Math. Methods Appl. Sci. 2021, 44, 11211–11230. [Google Scholar] [CrossRef]
- Bram, J. The Lagrange multiplier theorem for max–min with several constraints. SIAM J. Appl. Math. 1966, 14, 665–667. [Google Scholar] [CrossRef]
- Jayswal, A. Non-differentiable minimax fractional programming with generalized α-univexity. J. Computat. Appl. Math. 2008, 214, 121–135. [Google Scholar] [CrossRef][Green Version]
- Jayswal, A.; Stancu-Minasian, I. Higher-order duality for nondifferentiable minimax programming problem with generalized convexity. Nonlinear Anal. 2011, 74, 616–625. [Google Scholar] [CrossRef]
- Liu, J.C.; Wu, C.S. On minimax fractional optimality conditions with invexity. J. Math. Anal. Appl. 1998, 219, 21–35. [Google Scholar] [CrossRef][Green Version]
- Lai, H.C.; Huang, T.Y. Optimality conditions for a nondifferentiable minimax programming in complex spaces. Nonlinear Anal. 2009, 71, 1205–1212. [Google Scholar] [CrossRef]
- Lai, H.C.; Huang, T.Y. Nondifferentiable minimax fractional programming in complex spaces with parametric duality. J. Global Optimiz. 2012, 53, 243–254. [Google Scholar] [CrossRef]
- Mishra, S.K.; Rueda, N.G. Second-order duality for nondifferentiable minimax programming involving generalized type I functions. J. Optimiz. Theory App. 2006, 130, 477–486. [Google Scholar] [CrossRef]
- Tanimoto, S. Nondifferentiable mathematical programming and convex-concave functions. J. Optimiz. Theory App. 1980, 31, 331–342. [Google Scholar] [CrossRef]
- Zalmai, G.J. Optimality conditions and duality for constrained measurable subset selection problems with minmax objective functions. Optimization 1989, 20, 377–395. [Google Scholar]
- Hong, Z.; Jiao, L.G.; Kim, D.S. On a class of nonsmooth fractional robust multi-objective optimization problems. Part I: Optimality conditions. Appl. Set-Valued Anal. Optim. 2020, 2, 109–121. [Google Scholar]
- Husain, Z.; Jayswal, A.; Ahmad, I. Second order duality for nondifferentiable minimax programming problems with generalized convexity. J. Global Optim. 2009, 44, 593–608. [Google Scholar] [CrossRef]
- Manesh, S.S.; Saraj, M.; Alizadeh, M.; Momeni, M. On robust weakly ε-efficient solutions for multi-objective fractional programming problems under data uncertainty. AIMS Math. 2021, 7, 2331–2347. [Google Scholar] [CrossRef]
- Zalmai, G.J. Parameter-free sufficient optimality conditions and duality models for minmax fractional subset programming problems with generalized (F, ρ, θ)-convex functions. Comput. Math. Appl. 2003, 45, 1507–1535. [Google Scholar] [CrossRef]
- Mordukhovich, B.S. Variational Analysis and Generalized Differentiation, I: Basic Theory; Springer: New York, NY, USA, 2006. [Google Scholar]
- Mordukhovich, B.S. Variational Analysis and Applications; Springer: New York, NY, USA, 2018. [Google Scholar]
- Chuong, T.D. Optimality and duality for robust multiobjective optimization problems. Nonlinear Anal. 2016, 134, 127–143. [Google Scholar] [CrossRef]
- Bonnans, J.F.; Shapiro, A. Perturbation Analysis of Optimization Problems; Springer: New York, NY, USA, 2000. [Google Scholar]
- Chuong, T.D.; Kim, D.S. A class of nonsmooth fractional multiobjective optimization problems. Ann. Oper. Res. 2016, 244, 367–383. [Google Scholar] [CrossRef]
- Chuong, T.D.; Kim, D.S. Nondifferentiable minimax programming problems with applications. Ann. Oper. Res. 2017, 251, 73–87. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).