*Article* **Hybridisation of Swarm Intelligence Algorithms with Multi-Criteria Ordinal Classification: A Strategy to Address Many-Objective Optimisation**

**Alejandro Castellanos 1, Laura Cruz-Reyes 1, Eduardo Fernández 2, Gilberto Rivera 3,\*, Claudia Gomez-Santillan <sup>1</sup> and Nelson Rangel-Valdez <sup>1</sup>**

	- <sup>2</sup> Facultad de Contaduría y Administración, Universidad Autónoma de Coahuila, Torreón 27000, Coahuila, Mexico; eduardo.fernandez@uadec.edu.mx
	- <sup>3</sup> División Multidisciplinaria de Ciudad Universitaria, Universidad Autónoma de Ciudad Juárez, Ciudad Juárez 32579, Chihuahua, Mexico
	- **\*** Correspondence: gilberto.rivera@uacj.mx

**Abstract:** This paper introduces a strategy to enrich swarm intelligence algorithms with the preferences of the Decision Maker (DM) represented in an ordinal classifier based on interval outranking. Ordinal classification is used to bias the search toward the Region of Interest (RoI), the privileged zone of the Pareto frontier containing the most satisfactory solutions according to the DM's preferences. We applied this hybridising strategy to two swarm intelligence algorithms, i.e., Multi-objective Grey Wolf Optimisation and Indicator-based Multi-objective Ant Colony Optimisation for continuous domains. The resulting hybrid algorithms were called GWO-InClass and ACO-InClass. To validate our strategy, we conducted experiments on the DTLZ problems, the most widely studied test suit in the framework of multi-objective optimisation. According to the results, our approach is suitable when many objective functions are treated. GWO-InClass and ACO-InClass demonstrated the capacity of reaching the RoI better than the original metaheuristics that approximate the complete Pareto frontier.

**Keywords:** preference incorporation; ant colony optimisation; grey wolf optimisation; interval outranking; multi-criteria decision analysis

#### **1. Introduction**

Organisations from different sectors and domains often face problems with multiple conflicting objectives to optimise. The scientific literature classifies these problems according to the number of objectives as Multi-objective Optimisation Problems (MOPs) for problems with 2–4 objectives and Many-objective Optimisation Problems (MaOPs) for problems with more than four objectives. Typically, MOPs and MaOPs are addressed through the so-called a posteriori approach, which consists of the following two phases:


Metaheuristic algorithms are promising alternatives to address Phase 1. Multi-objective Evolutionary Algorithms (MOEAs) and Multi-objective Swarm Intelligence Algorithms (MOSIAs) are quite popular because their application does not demand particular mathematical properties on the objective functions, the geometry of the Pareto frontier or the constraints of the problem [1–3].

Even though MOEAs and MOSIAs have been widely used in addressing MOPs, most of them cannot adequately approximate the Pareto frontier for MaOPS. They have to cope

**Citation:** Castellanos, A.; Cruz-Reyes, L.; Fernández, E.; Rivera, G.; Gomez-Santillan, C.; Rangel-Valdez, N. Hybridisation of Swarm Intelligence Algorithms with Multi-Criteria Ordinal Classification: A Strategy to Address Many-Objective Optimisation. *Mathematics* **2022**, *10*, 322. https://doi.org/ 10.3390/math10030322

Academic Editor: Jian Dong

Received: 13 December 2021 Accepted: 18 January 2022 Published: 20 January 2022

**Publisher's Note:** MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

**Copyright:** © 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https:// creativecommons.org/licenses/by/ 4.0/).

with severe difficulties as the number of objectives increases; remarkably, the number of dominance resistant solutions, the high cost to ensure diversity, and the low effectiveness of the genetic operators. Studies on the difficulties and challenges found by MOEAs in addressing MaOPs are discussed by Bechikh et al. [4], López Jaimes and Coello Coello [5], Sudeng and Wattanapongsakorn [6], and Ikeda et al. [7]. One of the main criticisms of the a posteriori approaches is that they require reaching a sufficiently representative approximation of the complete Pareto front before the multi-criteria decision making.

Even supposing a good approximation of the Pareto frontier, the DM has to solve a multi-criteria selection problem in this set to address Phase 2. An alternative is to make a heuristic selection; the DM is supposed to consistently compare solutions on the approximated Pareto frontier until the best compromise is identified. This task is likely to become pretty hard and impractical in problems with many objective functions because of the human mind cognitive limitations (as stated by Miller [8]). Another alternative is to apply a Multi-Criteria Decision Analysis (MCDA) method that articulates the preferences of the DM. Again, this approach assumes that the solution set effectively contains the most preferred solutions, which is questionable under the presence of many objective functions.

As a consequence of the above discussion, there has been an increasing interest in combining MOEAs and MCDA methods in recent years. One way is to consider the preferences to bias the search toward the Region of Interest (RoI). The RoI is the region of the Pareto frontier with the solutions that best match the DM's preferences; accordingly, the best compromise is a solution belonging to the RoI.

The preference incorporation requires considering non-trivial aspects such as defining the model of the DM's preferences, characterising the RoI, and determining the relevance of the solutions [9]. Most methods for preference incorporation admit at least one of the following kinds of information on the preferences [10,11]: weights, e.g., [12]; ranking of solutions, e.g., [13]; ranking of objective functions, e.g., [14]; reference points, e.g., [15]; trade-offs between objective functions, e.g., [16]; desirability thresholds, e.g., [17]; solution classification, e.g., [18]; and pairwise comparisons based on preference relations, e.g., [19,20].

The two alternatives to incorporate the above strategies are the interactive approach and the a priori approach, and both have advantages over the a posteriori one. On the one hand, better solutions are found because there is an increment in the selective pressure toward the RoI [21]. On the other hand, preference incorporation can alleviate the DM's cognitive effort to select the best compromise because the number of candidate solutions is relatively short. Still, the interactive approaches are strongly criticised because they require preference relations with full comparability and transitivity (as the dimensionality increases, these properties become unlikely) [22]. Contrastingly, the a priori preference incorporation does not mandatorily require such properties. However, it demands a model that reflects the DM's preferences about the solutions.

In many real-world MaOPs, the models should support imprecision and vagueness in the DM's preferences. For instance, if the DM is a heterogeneous group (e.g., a board of directors) or an ill-defined entity (e.g., a community in social networks). In those circumstances, the task of eliciting the parameters of a preference model is highly difficult and is only reachable with some level of imprecision. If such imperfect knowledge is not considered, the best compromise could hardly be identified among the existing alternatives. Interval mathematics is a straightforward but effective way to express imprecision [23].

Fernandez et al. [24] introduced an extension of the outranking method by incorporating interval numbers in the preference parameters. This MCDA method can handle incomparability, veto situations, and non-transitive preferences. These properties become critical to address real-world MaOPs because many DMs have non-transitive and noncompensatory preferences. Additionally, the DM feels more comfortable eliciting the values of the parameters as interval numbers than as precise values. If the DM cannot (or not want to) directly give a value for some required parameters, they may use an indirect elicitation method to infer them, e.g., [25].

In light of the above discussions, this paper presents a further analysis to observe how incorporating interval outranking in MOSIAs impacts the performance. We propose embedding multi-criteria ordinal classification based on interval outranking (strictly speaking, the INTERCLASS-nC method [26]) into many-objective optimisation algorithms. Similar to previous approaches based on ordinal classification, e.g., [11], our proposal also requires representative samples of solutions classified by the DM as 'Satisfactory' or 'Dissatisfactory'. However, we extended this notion with two artificial classes (for internal use of the algorithms) that increase the selective pressure toward the RoI. By applying this strategy, we introduce the a priori versions of two relevant a posteriori algorithms based on swarm intelligence; specifically, Multi-objective Grey Wolf Optimisation (MOGWO) [27] and Indicator-based Multi-objective Ant Colony Optimisation for continuous domains (iMOACOR) [28]. The a priori preference incorporation significantly increased the performance of the MOSIAs according to a non-parametric test (Mann–Whitney–Wilcoxon).

The remainder of this paper is organised as follows. Section 2 includes some preliminaries on multi-objective optimisation, interval outranking, ordinal classification, and the MOSIAs taken as baseline. Section 3 details the proposed algorithms with preference incorporation. Section 4 shows the experimental results. Lastly, Section 5 discusses the conclusions and provides some directions for future research.

#### **2. Background**

This section presents an overview of the theoretical foundations. Section 2.1 presents some preliminaries on optimisation with multiple objectives. Section 2.2 briefly describes the baseline versions of the MOSIAs used in this paper. Lastly, Section 2.3 presents the model for multi-criteria ordinal classification based on interval outranking.

#### *2.1. Preliminaries on Multi-Objective Optimisation*

Optimisation refers to finding the values in the decision variables (independent variables) that provoke extreme values of one or more objective functions (dependent variables). It is called 'mono-objective' optimisation if a single function is treated. In contrast, it is called 'multi-objective' optimisation if a few objective functions are treated (typically, up to four). Farina and Amato [29] recognised that most MOEAs are severely affected when they address problems with more than four objective functions, named them 'many-objective' optimisation problems.

Real-world applications often involve optimising several functions that are essentially conflicting [30]. As a consequence, no point is simultaneously optimal in all objective functions.

Here, *x* = *x*1, *x*2, *x*3,..., *xn* represents a solution of a MOP/MaOP: a vector of decision variables that optimises a vector function *f*(*x*) whose components represent the values of the objectives. Equation (1) defines *f*(*x*), where *m* is the number of objectives (dimensionality of the problem), and *n* is the number of decision variables. Applied optimisation models usually add constraints to Equation (1) to reflect real situations.

$$f(\mathbf{x}) = \langle f\_1(\mathbf{x}), f\_2(\mathbf{x}), f\_3(\mathbf{x}), \dots, f\_m(\mathbf{x}) \rangle \quad f\_k: \mathbb{R}^n \to \mathbb{R} \tag{1}$$

Pareto dominance is widely accepted to compare two solutions, determining which of them is better. Thus, Pareto dominance discriminates between solutions by comparing their *f*(*x*). Without loss of generality, let us consider minimising the *m* objectives; the Pareto dominance relation, represented by the symbol -, may be expressed as [31]

$$\begin{aligned} \text{fix} \preccurlyeq y = \left\{ (\mathbf{x}, \mathbf{y}) : \ f\_k(\mathbf{x}) \leqslant f\_k(\mathbf{y}) \,\forall k \in \{1, 2, 3, \dots, m\} \right\} \quad \land\\ f\_k(\mathbf{x}) < f\_k(\mathbf{y}) \; \exists k \in \{1, 2, 3, \dots, m\} \end{aligned} \tag{2}$$

The non-dominated solutions make up the Pareto set, expressed as

$$PS = \{ \mathfrak{x} \in \mathcal{R}\_F : \mathfrak{y} \preccurlyeq \mathfrak{x} \; \nexists \mathfrak{y} \in \mathcal{R}\_F \}, \tag{3}$$

where *RF* is the feasible region.

The Pareto frontier, *PF* = { *f*(*x*) : *x* ∈ *PS*}, is the image of the Pareto set. In the absence of information on the preferences of the DM, a sufficiently representative sample of the Pareto frontier should be calculated.

Identifying a set of Pareto efficient solutions is indeed necessary to solve MOPs and MaOPS. Still, it is not sufficient since the DM must select the best compromise (the solution to implement). The DM chooses the best compromise according to their personal preferences about the objective functions. In practice, the best compromise is the ultimate solution to the problem.

#### *2.2. An Overview of Two Swarm Intelligence Algorithms to Address MaOPs*

In this section, we provide a brief description of MOGWO [27] and iMOACOR [28].

#### 2.2.1. Multi-Objective Grey Wolf Optimisation

Mirjalili et al. [27] proposed MOGWO—Multi-Objective Grey Wolf Optimiser—which extends the mono-objective algorithm Grey Wolf Optimiser (GWO) [32] to treat multiple objectives. This swarm intelligence algorithm is inspired by nature, specifically by the behaviour of grey wolves in tracking and hunting their prey. By analogy, the solution with the best value in the objective function is named the α wolf. The second-best and the third-best solutions are named β and δ wolves, respectively. The remaining solutions are known as Ω wolves. The leaders (α, β, and δ) guide the optimisation process, and the Ω wolves follow the leaders in the search for the global optimum.

Let *n* be the number of decision variables, *ι* be the number of the current iteration, *xι <sup>i</sup>* = *xι <sup>i</sup>*,1, *<sup>x</sup><sup>ι</sup> <sup>i</sup>*,2, *<sup>x</sup><sup>ι</sup> <sup>i</sup>*,3,..., *<sup>x</sup><sup>ι</sup> i*,*n* be the *n*-dimensional location point of the *i*th wolf during the *ι*th iteration, and *x<sup>ι</sup>* = *xι* <sup>1</sup>, *<sup>x</sup><sup>ι</sup>* <sup>2</sup>, *<sup>x</sup><sup>ι</sup>* <sup>3</sup>,..., *<sup>x</sup><sup>ι</sup> h*¯ be the positions of a pack with *h*¯ wolves. The following equation simulates how the Ω wolves are relocated to siege their prey during hunt following the social leadership:

$$\mathbf{x}\_{i}^{\iota+1} = \frac{\mathbf{r}\_{1}^{\iota} + \mathbf{r}\_{2}^{\iota} + \mathbf{r}\_{3}^{\iota}}{3} \quad \forall i \in \{1, 2, 3, \dots, \hbar\}, \tag{4}$$

where x*<sup>ι</sup>* <sup>1</sup>, <sup>x</sup>*<sup>ι</sup>* <sup>2</sup>, and <sup>x</sup>*<sup>ι</sup>* <sup>3</sup> are *n*-dimensional vectors reflecting the influence of the three leader wolves during the *ι*th iteration. Each component of them is calculated as

$$\mathbf{x}\_{1,j}^l = \mathbf{x}\_{\alpha,j} - A\_j \cdot D\_{\alpha,j} \quad \forall j \in \{1, 2, 3, \dots, n\}, \quad \text{where } D\_{\alpha,j} = \left| C\_j \cdot \mathbf{x}\_{\alpha,j} - \mathbf{x}\_{i,j}^l \right|\_{\prime} \tag{5}$$

$$\mathbf{x}\_{2,j}^{\iota} = \mathbf{x}\_{\mathfrak{B},j} - A\_j \cdot D\_{\mathfrak{B},j} \quad \forall j \in \{1, 2, 3, \dots, n\}, \quad \text{where } D\_{\mathfrak{B},j} = \left| C\_j \cdot \mathbf{x}\_{\mathfrak{B},j} - \mathbf{x}\_{i,j}^{\iota} \right|, \tag{6}$$

$$\mathbf{x}\_{3,j}^t = \mathbf{x}\_{\delta,j} - A\_j \cdot D\_{\delta,j} \quad \forall j \in \{1, 2, 3, \dots, n\}, \quad \text{where } D\_{\delta,j} = \left| \mathbf{C}\_j \cdot \mathbf{x}\_{\delta,j} - \mathbf{x}\_{i,j}^t \right|. \tag{7}$$

In Equations (4)–(7), *x*α,*j*, *x*β,*<sup>j</sup>* and *x*δ,*<sup>j</sup>* are the positions of the leader wolves at the *j*th coordinate, *ι* + 1 is the number of the next iteration, and *A* and *D*<sup>α</sup> are coefficient vectors modelling the encircling behaviour of wolves as

$$A\_{\rangle} = 2a\_{\rangle} \cdot r\_{1,\natural} - a\_{\rangle} \,\,\forall j \in \{1, 2, 3, \ldots, n\},\tag{8}$$

and

$$D\_{\mathbf{x},j} = \left| \mathbb{C}\_j \cdot \mathbf{x}\_{\mathbf{a},j} - \mathbf{x}\_{i,j}^t \right| \; \forall j \in \{1, 2, 3, \dots, n\}, \quad \text{where } \mathbb{C}\_j = 2 \cdot r\_{2,j}. \tag{9}$$

In Equations (8) and (9), *a* is an *n*-dimensional vector whose elements linearly decrease from two to zero throughout the run of the algorithm, and *r*<sup>1</sup> and *r*<sup>2</sup> are random

*n*-dimensional vectors with values in [0, 1]. The components of *D*<sup>β</sup> and *D*<sup>δ</sup> are similarly calculated as those of *D*α in Equation (9).

Exploration is promoted by *A* with values greater than one (or less than –1); with that setting, the Ω wolves diverge from the leaders. *C* is another component of GWO that favours exploration, whose components represent weights and are generated at random to emphasise (*Cj* > 1) or de-emphasise (*Cj* < 1) the influence of the α, β and δ wolves in defining the distance in Equations (5)–(7). According to Mirjalili et al. [27], p. 109: "*C* is not linearly decreased in contrast to *A*. The *C* parameter was deliberately required to provide random values at all times in order to emphasise exploration not only during initial iterations but also final iterations".

GWO exploits the search space if |*Aj*| < 1 because, when the components of *A* are in [−1, 1], the position of the *i*th wolf in the next iteration will be located between its current position and the position of the leader. This setting assists Ω wolves to converge toward an estimated position of their prey, provided by *x*α, *x*<sup>β</sup> and *x*δ.

GWO starts the optimisation process by generating solutions at random during the first population. Then, the three best solutions so far are considered as the α, β and δ wolves. In the next iteration, each Ω wolf updates its position by applying Equations (4)–(7). Simultaneously, the components of *a* are linearly decreased in each iteration. Therefore, the pack of wolves tends to widely explore the search space during the first iterations and intensively exploit it during the last iterations. The algorithm stops when a maximum number of iterations is reached, and *x*α is returned as the best solution obtained throughout the optimisation process.

To treat MOPs via GWO, MOGWO integrates the following two extensions into GWO: (*i*) an archive with the Pareto optimal solutions obtained so far, and (*ii*) a criterion for picking the leaders (*x*α, *x*<sup>β</sup> and *x*δ) from the archive.

If the archive is full of non-dominated solutions (note that there is a predefined size) and there is a new solution to be entered; then, a grid technique determines the region in the Pareto frontier that is the most crowded by the archive. A solution is removed from this region at random; then, the new solution may be added to the archive. Thus, the probability of a solution being deleted is proportional to the number of points in each hypercube (region).

Complementarily, the criterion to select *x*α, *x*<sup>β</sup> and *x*<sup>δ</sup> favours the least crowded regions in the Pareto front. The selection is based on a roulette-wheel method with the following probability for each hypercube:

$$p\_l = \frac{\mathcal{C}}{\mathcal{N}\_l} \, \tag{10}$$

where C is a constant (C 1), and *Nl* is the number of solutions in the *l*th hypercube. The two extensions of MOGWO jointly promote the representativeness in the sample of the Pareto frontier.

#### 2.2.2. Indicator-Based Multi-Objective Ant Colony Optimisation for Continuous Domains

Socha and Dorigo [33] proposed ACOR, an extension of Ant Colony Optimisation (ACO) [34] to optimise mono-objective problems with continuous decision variables. In ACOR, the pheromone matrix (*τ*) is an archive that stores the best-so-far solutions. Here, a vector *xl* = *xl*,1, *xl*,2, *xl*,3,..., *xl*,*<sup>n</sup>* represents a solution of a problem with *n* decision variables, and *f*(*xl*) is the objective function to minimise. The *κ* best-evaluated solutions are stored in *τ*, following the implicit order given by *f*(*xl*). The position in *τ* of each *xl* determines a weight (*ωl*) that measures the quality of *xl*, defined in Equation (11).

$$\omega\_{l} = \frac{e^{-\varphi(l)}}{\mathfrak{g} \cdot \kappa \sqrt{2\pi}}, \text{ where } \varphi(l) = \frac{(l-1)^2}{2\mathfrak{g}^2 \kappa^2} \tag{11}$$

Equation (11) defines *ω<sup>l</sup>* as a value of the Gaussian function with standard deviation *ς* · *κ*, mean 1.0, and argument *l*. Here, *κ* is the number of solutions in *τ*, and *ς* is a parameter (0 *ς* 1). The effect of *ς* is to establish the proper balance between the pressures exerted by the best-so-far solution (with values close to one) and the iteration-best solutions (with values close to zero).

Furthermore, a Gaussian kernel (*Gj*) is calculated for each decision variable (1 *j n*) as

$$G\_{\dot{l}}(\mathbf{x}) = \sum\_{l=1}^{\kappa} \omega\_l g\_l^{\dot{l}}(\mathbf{x}),\tag{12}$$

where

$$g\_l^j(\mathbf{x}) = \frac{e^{-\phi\_l(l)}}{s\_l^j \sqrt{2\pi}}, \text{ where } \phi\_l(l) = \frac{\left(\mathbf{x} - \mathbf{x}\_{l,j}\right)^2}{2\left(s\_l^j\right)^2}. \tag{13}$$

According to Equation (13), *g j l* (*x*) is a normal distribution, where *s j <sup>l</sup>* is the standard deviation, and *xl*,*<sup>j</sup>* is the mean. The former is calculated in each iteration as ants construct solutions. Then, *Gj* in Equation (12) is a weighted sum of *g j l* (*x*) ∀*l* ∈ {1, 2, 3, ... *κ*}, which defines the one-dimensional Gaussian function for the *j*th decision variable of the *l*th solution in *τ*.

Ants construct solutions by performing *n* steps. The *i*th ant sets the value for the variable *xi*,*<sup>j</sup>* at the *j*th step. Here, only *Gj*—the resulting Gaussian kernel—is needed. Then, the weights *ω<sup>l</sup>* are computed through Equation (11) and used to sample, by following two phases:

1. One Gaussian function is picked from the Gaussian kernel. The probability *pl* of choosing the *l*th Gaussian function is given by

$$p\_l = \frac{\omega\_l}{\sum\_{r=1}^{\kappa} \omega\_r}.\tag{14}$$

To exploit the synergy among decision variables, *g j l* (*x*) is the single Gaussian function ants use to construct a solution incrementally during a complete iteration.

2. The chosen Gaussian function (*l*) is used to sample new solutions. At the *j*th step, the standard deviation (*s j l* ) is required to calculate the Gaussian function, *g j l* (*x*), picked in Phase 1 (de facto, the sampled Gaussian function will be different in each *j*th construction step). Here, *s j <sup>l</sup>* is dynamically calculated by Equation (15).

$$s\_l^j = \xi \sum\_{r=1}^{\kappa} \frac{\left| \mathcal{X}\_{r,j} - \mathcal{X}\_{l,j} \right|}{\kappa - 1} \tag{15}$$

Equation (15) defines the standard deviation as the average distance from *xl* (the picked solution) to other solutions (*xr*) in *τ*. The parameter *ξ* weights this distance (0 *ξ* 1). The effect of the parameter *ξ* is similar to the pheromone evaporation in ACO, influencing the behaviour of the colony: with low values, exploitation is more promoted than exploration. The value of the *j*th decision variable is inferred by the *i*th ant by following

$$
\boldsymbol{\alpha}\_{i,j} \sim \mathcal{g}\_I^j(\boldsymbol{\pi}).\tag{16}
$$

#### *2.3. Multi-Criteria Ordinal Classification Based on Interval Outranking*

The notion behind outranking is that the credibility of the proposition '*x* is at least as good as *y*'—represented as *σ*(*x*, *y*)—may be calculated by analysing each pair of their criteria scores [35]. ELECTRE is the most representative MCDA method of the outranking approaches. Typically, ELECTRE defines *σ*(*x*, *y*) = *c*(*x*, *y*) · *d*(*x*, *y*), where


Both indexes are calculated in function of a series of parameters that must be appropriately inferred so that *σ* models the preferences of the DM. Interval outranking generalises the classic outranking to the framework of interval numbers, providing support when the values of the preference parameters are imprecisely known.

If the reader is be unfamiliar with interval mathematics, Appendix A compiles the basic notions to understand interval outranking. Note that interval numbers in this paper are written in boldface italic letters.

Let *O* be a set of alternatives (solutions). Each *x* ∈ *O* is evaluated with an *m*dimensional objective function *f*(*x*) = *f*1(*x*), *f*2(*x*), *f*3(*x*),..., *fm*(*x*). Without loss of generality, we suppose that each *fk*(*x*) is a minimising objective and, consequently, the preference of the DM increases as the value of *fk*(*x*) decreases. The parameters of the outranking model are:

• The vector of weights, *wk* = *wk*, *wk* ∀*k* ∈ {1, 2, 3, . . . , *m*}, where *m* ∑ *k*=1 *wk* 1 and *m*

$$\sum\_{k=1}^{\sum \overline{w\_k}} \overline{w\_k} \gg 1;$$


The concordance coalition of two solutions (*x*, *y*), denoted as *Cx*,*<sup>y</sup>* = *k* ∈ {1, 2, 3, ... , *m*} : *P*(*fk*(*y*) *fk*(*x*)) 0.5 , is the subset of objectives favouring the statement '*x* is at least as good as *y*'. The concordance index for the statement '*x* is at least as good as *y*' is the interval number *c***(***x***,** *y***)** = *c*(*x*, *y*), *c*(*x*, *y*) , defined as

$$\underline{\mathcal{L}(\underline{x},\underline{y})} = \begin{cases} \sum\_{k \in \mathcal{C}\_{\underline{x},\underline{y}}} \underline{w}\_{k} & \text{if } \sum\_{k \in \mathcal{C}\_{\underline{x},\underline{y}}} \underline{w}\_{k} + \sum\_{k \in \mathcal{D}\_{\underline{x},\underline{y}}} \overline{w}\_{k} \gg 1, \\\\ 1 - \sum\_{k \in \mathcal{D}\_{\underline{x},\underline{y}}} \overline{w}\_{k} & \text{otherwise,} \end{cases} \tag{17}$$

and

$$\overline{c(x,y)} = \begin{cases} \sum\_{k \in \mathcal{C}\_{x,y}} \overline{w\_k} & \text{if } \sum\_{k \in \mathcal{C}\_{x,y}} \overline{w\_k} + \sum\_{k \in D\_{x,y}} \underline{w\_k} \le 1, \\\\ 1 - \sum\_{k \in D\_{x,y}} \underline{w\_k} & \text{otherwise.} \end{cases} \tag{18}$$

Moreover, the discordance coalition consists of the objectives that are not in the concordance coalition, defined as *Dx*,*<sup>y</sup>* = {1, 2, 3, . . . , *m*} \ *Cx*,*y*. These objectives justify arguments invalidating the outranking relation '*x* is at least as good as *y*'. The value *P fk*(*x*) <sup>−</sup> *fk*(*y*) *vk* models the degree of credibility of the proposition 'the *k*th criterion alone vetoes the statement *x outranks y*'. The discordance index, *d*(*x*, *y*), is calculated appraising the credibility of veto of each objective in *Dx*,*y*, and is expressed as

$$d(\mathbf{x}, \mathbf{y}) = 1 - \max\_{k \in D\_{x, \mathbf{y}}} \left\{ P\left(f\_k(\mathbf{x}) - f\_k(\mathbf{y}) \rhd \mathbf{v}\_k\right) \right\}. \tag{19}$$

Accordingly, *σ*(*x*, *y*) is redefined as

$$\sigma(\mathbf{x}, \mathbf{y}) = \min \left\{ P\left(\mathbf{c}(\mathbf{x}, \mathbf{y}) \geqslant \lambda\right), d(\mathbf{x}, \mathbf{y}) \right\}. \tag{20}$$

By following Equation (20), Fernandez et al. [24] introduced a pair of binary preference relations. First, the crisp outranking relation (S) is presented in Equation (21), where *β* is a threshold on the credibility of '*x* is at least as good as *y*'. Lastly, Equation (22) presents the crisp relation '*x* is preferred to *y*'.

$$\text{tr}\mathbb{S}y = \{ (\mathbf{x}, y) : \sigma(\mathbf{x}, y) \rhd \beta \}\tag{21}$$

$$\text{xPry} = \{(\mathbf{x}, \mathbf{y}) : \mathbf{x} \preccurlyeq \mathbf{y} \lor \text{ (xSy} \land \neg \mathbf{ySx)}\}\tag{22}$$

Fernandez et al. [26] proposed INTERCLASS-nC, an extension of ELECTRE TRInC [36] in the framework of interval outranking to multi-criteria ordinal classification. INTERCLASS-nC is applicable in circumstances in which other MCDA methods fail, specifically when the DM does not want to (or cannot) set precise values for the model parameters (majority threshold, veto thresholds, and weights). INTERCLASS-nC is especially advantageous if the DM has only a vague idea about the boundaries between adjacent classes; nonetheless, they can quickly identify one representative solution at least in each category, which may be characterised by intervals.

INTERCLASS-nC considers the array of classes *C* = -*C*1, *C*2, *C*3,..., *C*, increasingly sorted by preference ( 2), and the subset of reference solutions introduced to characterise each *C*-, *R*- = *r*-,*j* (where 1 *j* |*R*-|, and 1 - ). Additionally, *r*0, *R*1, *R*2, *R*3,..., *R*,*r*+1 is the array of all reference solutions sorted by preference, where *r*<sup>0</sup> is the anti-ideal point and *r*+<sup>1</sup> is the ideal point. Following the interval outranking, the following condition is true for 1 -< :

$$
\neg \exists \mathbf{S} y \; \; \forall \mathbf{x} \in \mathcal{R}\_{\ell \prime} \; \; y \in \mathcal{R}\_{\ell+1} . \tag{23}
$$

Then, Equations (24) and (25) are used to define the categorical credibility indices between an alternative *x* and the category *C*-.

$$\theta(\mathbf{x}, \mathbf{R}\_{\ell}) = \max\_{1 \le \ell \le \lceil R\_{\ell} \rceil} \left\{ \sigma \left( \mathbf{x}, r\_{\ell, \ell} \right) \right\} \tag{24}$$

$$\theta(R\_{\ell'}\mathbf{x}) = \max\_{1 \le \ell \le \lceil R\_{\ell} \rceil} \left\{ \sigma \left( r\_{\ell, \eta'} \mathbf{x} \right) \right\} \tag{25}$$

The crisp relation of interval outranking is extended to compare actions with the sets of characteristic actions as


To suggest a class for a new alternative *x*, INTERCLASS-nC uses the selection function *Sf*(*x*, *R*-) = min{*ϑ*(*x*, *R*-), *ϑ*(*R*-, *x*)} and two heuristics—the ascending rule and the descending rule—that are conjointly used. Each one of these heuristics proposes a class for an alternative *x*. If the classes do not coincide, they define a range of assignments for *x* (any category within such a range is admissible as the class of *x*).

The steps of the ascending rule are the following


The descending rule has the following steps


4. If -= 0, select *C*<sup>1</sup> as a possible class for *x*.

#### **3. Proposed Algorithms**

In this section, we describe how ordinal classification based on interval outranking was embedded in MOGWO and iMOACOR to incorporate the preferences of the DM. The primary strategy is to measure the quality of the solutions in terms of the class suggested by INTERCLASS-nC. Consequently, the selective pressure increases toward the solutions the DM is highly satisfied with. Note that both proposed algorithms are not intended for searching representative approximations of the complete Pareto frontier; instead, they search for the RoI: a relatively short subset of Pareto optimal solutions that best match the DM's preferences. Here, we propose that the RoI is made of solutions classified as 'Highly Satisfactory', and the best compromise—the final prescription chosen by the DM—should be a solution belonging to the RoI. Section 3.1 introduces the Grey Wolf Optimiser with Interval outranking-based ordinal Classification, abbreviated as GWO-InClass. Section 3.2 introduces Ant Colony Optimisation with Interval outranking-based ordinal Classification, abbreviated as ACO-InClass.

#### *3.1. The GWO-InClass Algorithm*

GWO-InClass extends MOGWO by classifying the solutions in each iteration utilising INTERCLASS-nC before picking the leaders from the archive, keeping the solutions ranked in the following order: 'Highly Satisfactory', 'Satisfactory', 'Dissatisfactory', and 'Strongly Dissatisfactory'.

We suggest using an ordinal classifier that has already been validated in an evolutionary algorithm for many-objective optimisation [37]. In this approach, the DM should classify the reference solutions (input) in two classes, 'Dissatisfactory' and 'Satisfactory' (respectively, *C*<sup>1</sup> and *C*2), the minimum number of classes to apply INTERCLASS-nC. Then, the model is extended by artificially adding the classes 'Highly Satisfactory' and 'Strongly Dissatisfactory'. Each new solution *x* generated during the evolutionary search is classified according to the following assumptions:


Algorithm 1 presents an outline of the ordinal classifier proposed by Balderas et al. [37]. A crucial process of GWO-InClass is to update the archive and select the α, β and

δ wolves; this process is presented in Algorithm 2. Here, Lines 1–2 initialise the variables; Lines 3–12 rank the solutions according to the class suggested by Algorithm 1; Lines 13–15 make sure that the archive (A) is not larger than the maximum size allowed by the parameter *h*¯; and, lastly, Lines 16–20 pick three different solutions from the non-empty best-evaluated class to be the leader wolves.

Algorithm 3 provides an algorithmic outline of GWO-InClass. Unlike MOGWO, GWO-InClass needs information about the preferences of the DM, which is exploited to become closer to the RoI. The preferences of the DM are articulated through an interval outranking model. We suggest using the proposal by Fernandez et al. [25] to infer the model parameters that reflect the preferences of the DM. Furthermore, GWO-InClass also needs the reference sets with solutions labelled by the DM as 'Satisfactory' or 'Dissatisfactory'. Initially, synthetic solutions are helpful to perform this task; however, if the DM is not confident about those initial sets, they may additionally have some interactions with GWO-InClass throughout the optimisation process and directly classify the solutions in the archive, updating the reference sets. Binary classification of solutions is one of the least cognitively demanding ways to interact with the DM.

#### **Algorithm 1** Ordinal classifier based on interval outranking

**Input**: Number of objectives (*m*), parameters of the outranking model (*λ*, *β*, *v*, *w*), a solution *x*, representative sets *R*<sup>1</sup> and *R*<sup>2</sup> **Output**: The class for *x* (*class*)

1: *cnc* ← INTERCLASS-nC(*x*, *R*1, *R*2) Descending and ascending rules 2: **if** *cnc* = 'Satisfactory' **then** 3: **if** *x*Pr*y* ∀*y* ∈ *R*<sup>2</sup> **then** 4: *class* ← 'Highly Satisfactory' 5: **else** 6: **if** *y*Pr*x* ∀*y* ∈ *R*<sup>1</sup> **then** 7: *class* ← 'Strongly Dissatisfactory' 8: **return** *class*

#### **Algorithm 2** Selection of the leader wolves

**Input**: Archive ( <sup>A</sup> ), solutions of the current iteration (*x<sup>ι</sup>* ), maximum size of the archive (*h*¯) **Output**: The leader wolves (*x*α, *x*β, *x*δ)

1: R ← -∅, ∅, ∅, ∅ 2: T ← *PS*(A ∪ *<sup>x</sup><sup>ι</sup>* ) Filtering only Pareto efficient solutions, see Equation (3) 3: **for** *xi* ∈ T **do** Main loop for ranking solutions 4: *class* <sup>←</sup> extended\_INTERCLASS-nC(*xi*) Classifying by Algorithm <sup>1</sup> 5: **if** *class* = 'Highly Satisfactory' **then** 6: R<sup>1</sup> ← R<sup>1</sup> ∪ {*xi*} 7: **if** *class* = 'Satisfactory' **then** 8: R<sup>2</sup> ← R<sup>2</sup> ∪ {*xi*} 9: **if** *class* = 'Dissatisfactory' **then** 10: R<sup>3</sup> ← R<sup>3</sup> ∪ {*xi*} 11: **if** *class* = 'Strongly Dissatisfactory' **then** 12: R<sup>4</sup> ← R<sup>4</sup> ∪ {*xi*} 13: **if** |T | > *h*¯ **then** 14: Remove the |T | − *h*¯ worst solutions according to the ranking 15: A←T 16: *leaders* ← *x*α, *x*β, *x*δ 17: **for** *xi* ∈ *leaders* **do** Main loop for picking the leaders 18: *topRanked* ← ⎧ ⎪⎪⎨ ⎪⎪⎩ R<sup>1</sup> if R<sup>1</sup> = ∅ R<sup>2</sup> if R<sup>1</sup> = ∅ ∧ R<sup>2</sup> = ∅ R<sup>3</sup> if R<sup>1</sup> = R<sup>2</sup> = ∅ ∧ R<sup>3</sup> = ∅ R<sup>4</sup> otherwise Choosing the best evaluated set that is not empty 19: *xi* ← roulette\_wheel(*topRanked*) Probabilities calculated by Equation (10) 20: Remove *xi* from R To avoid selecting the same solution as *x*α, *x*<sup>β</sup> or *x*<sup>δ</sup> 21: **return** *x*α, *x*β, *x*<sup>δ</sup>

In Algorithm 3, Lines 1–5 initialise the parameters and generate the initial positions of the wolves. Line 6 picks the first leader wolves. Lines 7–12 present the iterated process of GWO-InClass; here, Line 8 applies Equations (4)–(9) to update the positions of the Ω wolves, following the social leadership of the α, β and δ wolves; Line 10 updates the archive and obtains the latest positions of the leader wolves; and lastly, Lines 11–12 update the parameters and variables to perform the next iteration.

**Algorithm 3** The Grey Wolf Optimiser with Interval outranking-based ordinal Classification

**Input** Number of objectives (*m*), number of decision variables (*n*), parameters of the outranking model (*wk*, *vk* <sup>∀</sup>*<sup>k</sup>* ∈ {1, 2, 3, . . . , *<sup>m</sup>*}, *<sup>λ</sup>*, *<sup>β</sup>*), reference solution sets (*R*1, *<sup>R</sup>*2), maximum size of the archive (¯*h*)

**Output:** An approximation of the RoI (A)


4: A ← ∅ 5: *ι* ← 1 6: (*x*α, *<sup>x</sup>*β, *<sup>x</sup>*δ) <sup>←</sup> get\_leaders(A, *<sup>x</sup>*0, ¯*h*) Algorithm <sup>2</sup> 7: **while** *ι iter*max **do** Main loop of the optimisation process 8: *x<sup>ι</sup> <sup>i</sup>* <sup>←</sup> <sup>x</sup>*ι*−<sup>1</sup> <sup>1</sup> <sup>+</sup>x*ι*−<sup>1</sup> <sup>2</sup> <sup>+</sup>x*ι*−<sup>1</sup> <sup>3</sup> <sup>3</sup> ∀*i* ∈ {1, 2, 3, . . . , ¯*h*} Updating the pack of wolves by applying Equations (4)–(9) 9: Calculate *fk*(*x<sup>ι</sup> i* ) ∀*k* ∈ {1, 2, 3, . . . , *m*}, *i* ∈ {1, 2, 3, . . . , ¯*h*} 10: (*x*α, *<sup>x</sup>*β, *<sup>x</sup>*δ) <sup>←</sup> get\_leaders(A, *<sup>x</sup><sup>ι</sup>* , ¯*h*) Algorithm 2 11: Update *a*, *A*, and *C* 12: *ι* ← *ι* + 1 13: **return** A

#### *3.2. The ACO-InClass Algorithm*

Like GWO-InClass, ACO-InClass stores the best solutions in the pheromone matrix *τ*—its archive—considering the class suggested by Algorithm 1. Figure 1 depicts the pheromone representation used in ACO-InClass; here, *κ* is the size of the archive, *n* is the number of decision variable, *xl* ∀*l* ∈ {1, 2, 3, . . . , *κ*} is the vector of the *l*th solution in the archive, *xl*,*<sup>j</sup>* is the value of the *j*th decision variable of *xl*, *Gj* ∀*j* ∈ {1, 2, 3, . . . , *n*} is the Gaussian kernel for the *j*th decision variable, and *ω<sup>l</sup>* ∀*l* ∈ {1, 2, 3, . . . , *κ*} is the weight of the *l*th solution.


**Figure 1.** Pheromone matrix in ACO-InClass.

The chief difference between iMOACOR and ACO-InClass is the criteria to sort *τ*. Falcón-Cardona and Coello Coello [28] suggested *R*<sup>2</sup> scores [38] as the primary criterion to rank the solutions in *τ*. Contrarily, in this paper, we propose the class indicated by Algorithm 1 as the primary criterion, and *R*<sup>2</sup> scores as the secondary criterion (intra-class solutions are sorted using the *R*<sup>2</sup> metric). The archive is set to store the *κ* top-ranked solutions.

Algorithm 4 provides an algorithmic outline of ACO-InClass. Lines 1 and 7 initialise the parameters and variables of the algorithm; Lines 2–6 generate and evaluate the initial random solutions of the colony. Lines 8–16 present the iterated process of ACO-InClass; here, Lines 9–10 generate the solutions of the colony by following the equations provided in Section 2.2.2, particularly Equations (14)–(16); Lines 11–14 evaluate and rank the solutions; and, lastly, Lines 15–16 update the data structures to perform the next iteration.

#### **Algorithm 4** Ant Colony Optimisation with Interval outranking-based ordinal Classification

**Input** Number of objectives (*m*), number of decision variables (*n*), parameters of the outranking model (*wk*, *vk* <sup>∀</sup>*<sup>k</sup>* ∈ {1, 2, 3, . . . , *<sup>m</sup>*}, *<sup>λ</sup>*, *<sup>β</sup>*), reference solution sets (*R*1, *<sup>R</sup>*2), maximum size of the archive (*κ*)

**Output:** An approximation of the RoI (*τ*)


#### 16: *ι* ← *ι* + 1 17: **return** *τ*

#### **4. Experimental Validation**

We implemented GWO-InClass and ACO-InClass in Java using OpenJDK 11.0.10, on a computer with an Intel Core i7-10510U CPU 1.80 GHz, 16 GB of RAM, and Manjaro 5.10 as operating system. This computer setting applies to all experiments reported in this section.

GWO-InClass has three main parameters to be adjusted: *a*, *C*, and *h*¯. The components of the vector *a* are initially set in two, and are linearly decreased to reach zero in the last iteration. The vector *C* is dynamically generated during each iteration, providing random numbers (cf. [27]). The parameter setting of ACO-InClass is *ς* = 0.1 and *ξ* = 0.5 (cf. [28]). The parameter settings of the reference algorithms—NSGA-III, MOGWO and iMOACOR are also those originally published by the authors in the articles [27,28,39]. A particular case is the size of the population (*h*¯ in GWO-InClass and *κ* in ACO-InClass), which depends on the number of objective functions (*m*): *h*¯ = *κ* = 92 for *m* = 3, *h*¯ = *κ* = 212 for *m* = 5, and *h*¯ = *κ* = 276 for *m* = 10. These values were inspired by the discussion of Deb and Jain [39] on suitable population sizes for evolutionary algorithms. The maximum number of iterations for all the algorithms reported in this section is *iter*max = 1000 for *m* ∈ {3, 5} and *iter*max = 1500 for *m* = 10.

The adjective 'significant' is used in this section if a non-parametric *U* test (also know as Mann–Whitney test or Wilcoxon rank-sum test) with a 0.95-confidence level validates the difference as statistically significant. Furthermore, we performed all tests for statistical significance through STAC [40].

The rest of this section is organised as follows. Section 4.1 presents the test suite used to validate the results, as well as the indicators to measure the quality of the solutions. Section 4.2 describes the results obtained by GWO-InClass. Lastly, Section 4.3 presents the results by ACO-InClass.

#### *4.1. Benchmark Problems and Performance Indicators*

DTLZ [41] has become the standard test suite most broadly accepted to assess the performance of MOEAs and MOSIAs. Accordingly, we ran our algorithms on the nine problems in the DTLZ suite, named DTLZ1–DTLZ9. These problems have continuous decision variables, and are scalable regarding the number of decision variables and objective functions, offering Pareto frontiers with challenging properties (e.g., bias, concavity, convexity, degeneration, multi-frontality, and separability). We explored the dimensionality of each DTLZ problem by considering 3, 5, and 10 objective functions (*m*); accordingly, there are 27 different instances to validate our algorithms. The numbers of position-related variables (*k*) and decision variables (*n*) for each problem are the following:


As GWO-InClass and ACO-InClass consider the DM's preferences, each of the 27 test instances was validated using ten interval outranking models, representing different DMs. Some state-of-the-art studies [42–44] have proposed and used these synthetic DMs to validate a priori optimisation methods run on the DTLZ suite. The primary motivation behind this choice is that an acceptable approximation to the true RoI is known for these synthetic DMs, and these Approximated RoIs (abbreviated as 'A-RoI' from hereon) are in compliance with the interval outranking model described in Section 2.3, which is the keystone of the interval ordinal classifier GWO-InClass and ACO-InClass use (see Algorithm 1).

The A-RoI should contain the most preferred solutions in terms of interval outranking; Rivera et al. [44] followed this underlying principle to calculate the A-RoI as follows (cf. [42,43]):


$$NW(\mathcal{O}) = \arg\min\_{y \in \mathcal{O}} \left\{ \sum\_{\mathbf{x} \in \mathcal{O} \backslash \{y\}} \varrho\_{\mathbf{Pr}}(\mathbf{x}, y) \right\}, \text{ where } \varrho\_{\mathbf{Pr}}(\mathbf{x}, y) = \begin{cases} 1 & \text{if } \mathbf{x} \mathbf{Pr} y\_{\mathbf{y}} \\ 0 & \text{otherwise} \end{cases}$$

3. Then, the A-RoI is finally calculated as

$$A\text{-}RoI(\mathcal{O}) = \arg\max\_{\mathbf{x}\in NW(\mathcal{O})} \left\{ \sum\_{\mathbf{y}\in \mathcal{O}\backslash\{x\}} \varrho\_{\mathbb{S}}(\mathbf{x}, \mathbf{y}) \right\}, \text{ where } \varrho\_{\mathbb{S}}(\mathbf{x}, \mathbf{y}) = \begin{cases} 1 & \text{if } x\text{S}y\_{\prime} \\ 0 & \text{otherwise} \end{cases}$$

The proposed algorithms ran 30 times on each test instance using each synthetic DM (as a consequence, each algorithm was run 300 times per instance). The benchmark algorithms—NSGA-III, MOGWO and iMOACOR—were also run 300 times.

There are several indicators for evaluating the performance of a multi-objective optimiser that approximates the complete Pareto frontier (e.g., hypervolume, spacing, spread, and inverted generational distance). However, these indicators are inadequate for assessing the performance of preference-based multi-objective algorithms because no metric can be directly applied when only a partial Pareto frontier is considered [45]. Furthermore, some attempts to assess the quality of a preferred solution adapt these indicators in an oversimplified way, making the assessments misleading. In line with this notion, we measured the performance via three indicators that consider the solutions that maximise the preferences of the DM. Let *X*∗ be the latest set of solutions of an algorithm, the following three indicators are utilised:

• Minimum Euclidian distance. The distance from the A-RoI to the closest point in *X*∗.


On the one hand, the distance-based indicators measure the quality in terms of the similarity between *X*∗ and the A-RoI; the minimum distance considers the best solution alone, and the average distance considers the overall trend. For these indicators, the lower the values, the closer the approximation. On the other hand, the satisfaction-based indicator measures the quality of the algorithm considering the number of solutions that could potentially become the best compromise solution. For this indicator, greater values are preferred.

#### *4.2. On the Performance of GWO-InClass*

Table 1 shows the results of GWO-InClass in comparison with MOGWO. Given a number of objectives (Column 1), Table 1 presents the average performance in terms of the indicators (referred to in Column 2) obtained by each algorithm (referred to in Column 3) on each DTLZ problem (referred to in Columns 4–12); here, the cells were shaded if the difference is statistically significant: in blue if it is in favour of GWO-InClass, in yellow if it is in favour of MOGWO.


The information in Table 1 may be summarised in the following points:


• Considering DTLZ8 and DTLZ9: The embedding of ordinal classification in MOGWO does not yield any significant benefit. It is worth noting that they are the only problems with side constraints in this benchmark [46].

As a partial conclusion, we would recommend using GWO-InClass instead of MOGWO to address MaOPS. GWO-InClass was at least as good as MOGWO; what is more, it was significantly better on a regular basis. A test for statistical significance supports this hypothesis on the DTLZ test suit.

Additionally, we compared the results of GWO-InClass with NSGA-III, which has been widely accepted by the scientific community as a benchmark algorithm for evolutionary many-objective optimisation. Table 2 presents the results of both algorithms on the DTLZ test suit. Again, the three indicators—the average Euclidean distance, the minimum Euclidean distance, and satisfaction—are considered, and the cells with significant differences are shaded (blue in favour of GWO-InClass, yellow in favour of NSGA-III).

**Table 2.** Average results obtained by GWO-InClass and NSGA-III.


According to Table 2, the worst performance was observed when *m* = 3. Contrarily, GWO-InClass approximated the RoI significantly better than NSGA-III in most of the 10-objective instances. Considering the satisfaction-based indicator, DTLZ1, DTLZ7, and DTLZ9 are especially challenging for GWO-InClass. To conclude, the performance of GWO-InClass becomes competitive as the number of objectives increases according to this standard of the literature.

Lastly, we plotted the results of some single runs of GWO-InClass in Figure 2. We took those runs on the three-objective problems with the results closest to the A-RoI. Although such runs are not representative, they clearly depict how GWO-InClass biases the search toward a privileged region in the Pareto frontier. The best compromise would be a solution belonging to such region (coloured in red in Figure 2).

(**a**) Results of GWO-InClass on DTLZ2

(**b**) Results of GWO-InClass on DTLZ4

(**c**) Results of GWO-InClass on DTLZ5

(**d**) Results of GWO-InClass on DTLZ7

**Figure 2.** Results of GWO-InClass on some 3-objective problems.

#### *4.3. On the Performance of ACO-InClass*

Table 3 presents the results of ACO-InClass in comparison with iMOACOR; its columns should be interpreted with the same meaning provided for Table 1. As a summary of the information provided, let us discuss the following remarks:

• Considering ten objective functions: The advantage of embedding ordinal classification became statistically significant only when *m* = 10. Taking DTLZ1–4 and DTLZ6–8, ACO-InClass outperformed iMOACOR in at least one indicator, performing especially well in DTLZ1, DTLZ3, DTLZ6, and DTLZ8.


We would strongly recommend using ACO-InClass instead of iMOACOR to address MaOPS with about ten objective functions. This insight is relevant because the efficiency of MOSIAs and MOEAs is degraded as the number of objectives increases. The strategy proposed in this paper could become a viable means of mitigating this severe drawback. Still, the DM should be prepared to devote the necessary time to express their preferences.

**Benchmark Problems** *m* **Indicator Algorithm DTLZ1 DTLZ2 DTLZ3 DTLZ4 DTLZ5 DTLZ6 DTLZ7 DTLZ8 DTLZ9** 3 Min. Euclid. iMOACOR 35.33 2.84 506.46 12.77 0.00 5.46 1.30 0.43 14.04 ACO-InClass 47.72 0.19 497.91 7.28 0.00 5.70 1.21 0.44 13.61 Avg. Euclid. iMOACOR 73.25 3.70 633.16 6.21 3.26 5.74 8.16 0.91 14.71 ACO-InClass 79.33 1.33 634.59 14.12 2.99 5.98 7.72 0.91 14.23 Satisfaction iMOACOR 11.62 28.51 24.42 32.82 1.37 2.47 23.37 46.21 0.00 ACO-InClass 11.18 32.21 15.38 31.79 2.08 2.64 24.27 52.14 3.22 5 Min. Euclid. iMOACOR 44.73 0.94 256.49 0.00 0.00 6.50 2.01 7.62 18.28 ACO-InClass 30.93 1.19 200.61 4.29 0.00 6.78 1.88 11.95 17.74 Avg. Euclid. iMOACOR 93.16 3.73 305.69 0.81 0.15 7.59 13.75 51.32 19.28 ACO-InClass 77.75 4.22 275.32 6.79 0.17 7.97 14.47 83.02 18.73 Satisfaction iMOACOR 21.26 18.22 24.56 23.50 16.63 3.99 33.71 60.70 0.00 ACO-InClass 22.91 18.92 22.77 24.55 15.54 3.97 34.43 38.26 3.22 10 Min. Euclid. iMOACOR 20.57 0.39 224.90 6.28 0.09 8.13 7.85 56.96 25.79 ACO-InClass 6.44 0.19 148.08 7.59 0.16 7.09 2.34 1.16 26.22 Avg. Euclid. iMOACOR 53.85 1.23 308.24 6.95 0.21 9.45 26.70 217.89 26.82 ACO-InClass 33.53 1.22 220.34 9.09 0.49 9.12 17.06 2.48 27.34 Satisfaction iMOACOR 28.47 28.61 22.27 40.27 16.13 15.19 48.16 12.24 3.22 ACO-InClass 49.84 47.47 61.03 59.72 11.94 25.63 49.13 87.75 2.21

**Table 3.** Average results obtained by ACO-InClass and iMOACOR.

Furthermore, Table 4 shows the results of ACO-InClass in comparison with those of NSGA-III. The columns of Table 4 should be analogously interpreted as in Table 2. These results together with the statistical tests allow concluding that ACO-InClass is competitive to address 10-objective problems according to the standard established by NSGA-III; in fact, ACO-InClass outperformed NSGA-III in the vast majority of these instances. As a welcome side effect, ACO-InClass can also support the a posteriori decision analysis; this feature would reduce the DM's cognitive effort invested in identifying the best compromise. Adversely, these results clearly imply that ACO-InClass was unsuitable for treating threeobjective instances.

Again, we plotted the results of some runs of ACO-InClass, which are shown in Figure 3. These runs presented the best results on the three-objective instances and clearly depicted how ACO-InClass biases the search toward the A-RoI. Here, the best compromise would be a solution obtained by ACO-InClass (coloured in red in Figure 3).


**Table 4.** Average results obtained by ACO-InClass and NSGA-III.

(**a**) Results of ACO-InClass on DTLZ1

(**b**) Results of ACO-InClass on DTLZ5

(**c**) Results of ACO-InClass on DTLZ6

**Figure 3.** *Cont*.

(**d**) Results of ACO-InClass on DTLZ9

**Figure 3.** Results of ACO-InClass on some 3-objective problems.

#### **5. Conclusions and Directions for Future Research**

This paper introduces a novel strategy to incorporate preferences into swarm intelligence algorithms. Following the taxonomy of Bechikh et al. [4], the proposed strategy falls into the category of 'solution classification'. Initially, the DM should express their preferences about the solutions by classifying them as 'Satisfactory' or 'Dissatisfactory'. The proposed strategy is one of the least cognitively demanding in the framework of solution classification because the DM must merely classify solutions into just two categories.

Then, we suggest that the optimisation algorithms additionally identify two artificial classes—'Highly Satisfactory' and 'Strongly Dissatisfactory'—through an ordinal classifier based on interval outranking to model different levels of intensity in the DM's preferences. Consequently, the classifier increases its ability to discriminate. Here, we hypothesised that swarm intelligence algorithms can obtain the edge by increasing the selective pressure toward the region of the Pareto frontier containing 'Highly Satisfactory' solutions. A straightforward way to achieve it is to consider ordinal classification as the major criterion to rank the solutions in the archive.

Typically, swarm intelligence algorithms use an archive with the best so-far approximation of the Pareto frontier. These solutions are sorted to pick the solution(s) whose patterns will be exploited in the next iteration, hence its relevance.

By applying this strategy, ordinal classification was embedded in two swarm intelligence algorithms, expressly Multi-objective Grey Wolf Optimisation and Indicator-based Multi-objective Ant Colony Optimisation for continuous domains. The extended versions were called Grey Wolf Optimiser with Interval outranking-based ordinal Classification (GWO-InClass) and Ant Colony Optimisation with Interval outranking-based ordinal Classification (ACO-InClass). We used ten synthetic DMs to validate the results; we also considered each problem in the DTLZ test suite and explored different numbers of objective functions (three, five, and ten). The algorithms ran 30 times for each setting.

The impact of our strategy depended on several factors: the number of objectives, the baseline algorithm, and the properties of the test problem. Despite this, our strategy conferred marked benefits when many objective functions were treated. In the case of GWO-InClass, such benefits became significant in six of the nine DTLZ problems when the number of objectives was at least five. In the case of ACO-InClass, the ordinal classifier impacted the performance in seven DTLZ problems only when ten objective functions were considered.

Although our strategy requires that the DM is well-disposed to spend the necessary time to elicit their preferences, such an effort can be favourably compensated when problems with many objective functions are treated; especially, keeping in mind that these problems are still highly challenging for a posteriori algorithms (that approximate the complete Pareto frontier). The embedding of ordinal classification based on interval outranking contributed to coping with these difficulties. Numerical results and tests for statistical significance supported these conclusions.

Perhaps, the major criticism of our approach is that it is only applicable when the preferences of the DM are compatible with the underlying principles of outranking. That is, the DM admits veto effects and has a non-compensatory preference about the objectives.

Further research is needed to draft conclusions with a greater generalisation. First, it is necessary to know the impact of this strategy on other swarm intelligence algorithms (e.g., particle swarm optimisation, artificial bee colony, and elephant herding optimisation). Second, it is also necessary to conduct more experimentation with a deeper analysis to connect the performance with the properties of the problem, providing plausible explanations for those problems in which no advantage was observed (e.g., DTLZ9).

**Author Contributions:** Conceptualization, L.C.-R. and E.F.; methodology, G.R. and C.G.-S.; software, A.C.; validation, A.C. and N.R.-V.; formal analysis, E.F. and G.R.; investigation, A.C. and N.R.-V.; resources, G.R.; data curation, N.R.-V.; writing—original draft preparation, G.R.; writing—review and editing, G.R.; visualization, C.G.-S.; supervision, L.C.-R. and E.F.; project administration, L.C.-R. and C.G.-S. All authors have read and agreed to the published version of the manuscript.

**Funding:** This research received no external funding.

**Institutional Review Board Statement:** Not applicable.

**Informed Consent Statement:** Not applicable.

**Data Availability Statement:** Not applicable.

**Conflicts of Interest:** The authors declare no conflict of interest.

#### **Appendix A. Basic Notions of Interval Mathematics**

An interval number represents a quantity whose precise value is uncertain; yet, the range in which the value lies is known [47]. Moore [48] defines an interval number *E* as *E* = *E*, *E* , where *E* denotes the lower bound and *E* the upper bound of *E*. Note that interval numbers are written in boldface italic letters in this paper.

Considering two interval numbers *E* = *E*, *E* and *D* = *D*, *D* , the basic arithmetic operations are defined as follows:

.


A realisation of an interval number *<sup>E</sup>* is any real number *<sup>e</sup>* <sup>∈</sup> *E*, *E* [49]. Let *e* and *d* be realisations of *E* and *D*, respectively, *E* > *D* if the proposition '*e* is greater than *d*' has greater credibility than '*d* is greater than *e*,' which can be calculated through the possibility function:

$$P(E \gg D) = \begin{cases} 1 & \text{if } p\_{ED} > 1, \\ 0 & \text{if } p\_{ED} < 0, \\ \quad p\_{ED} & \text{otherwise,} \end{cases} \tag{A1}$$

where *pED* = *<sup>E</sup>*−*<sup>D</sup>* (*E*−*E*)+(*D*−*D*) . If *E* and *D* are real numbers *E* and *D*, then

$$P(E \rhd D) = \begin{cases} 1 & \text{if } E \rhd D\_\prime \\ 0 & \text{otherwise.} \end{cases} \tag{A2}$$

The possibility function *P*(*E D*) = *α* is taken as the degree of credibility that the realisation *d* will be smaller than the realisation *e* [50]. The order relations are defined as:


Let us consider *P*(*E D*) = *α*<sup>1</sup> and *P*(*D C*) = *α*2, the possibility function is transitive because

$$
\alpha\_1 \rhd 0.5 \land \alpha\_2 \rhd 0.5 \Rightarrow P(E \rhd \mathcal{C}) \rhd \min\{\alpha\_1, \alpha\_2\},\tag{A3}
$$

as a consequence, the relations and > also meet the transitivity property on interval numbers.

#### **References**

