Next Article in Journal
Polynomial Iterative Learning Control (ILC) Tracking Control Design for Uncertain Repetitive Continuous-Time Linear Systems Applied to an Active Suspension of a Car Seat
Previous Article in Journal
Intelligent Retrieval and Secure Acquisition of Embedded Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Interpolation-Based Evolutionary Algorithm for Bi-Objective Feature Selection in Classification

School of Mechanical, Electrical & Information Engineering, Putian University, Putian 351100, China
Mathematics 2024, 12(16), 2572; https://doi.org/10.3390/math12162572
Submission received: 14 July 2024 / Revised: 4 August 2024 / Accepted: 19 August 2024 / Published: 20 August 2024

Abstract

:
When aimed at minimizing both the classification error and the number of selected features, feature selection can be treated as a bi-objective optimization problem suitable for solving with multi-objective evolutionary algorithms (MOEAs). However, traditional MOEAs may encounter difficulties due to discrete optimization environments and the curse of dimensionality in the feature space, especially for high-dimensional datasets. Therefore, in this paper an interpolation-based evolutionary algorithm (termed IPEA) is proposed for tackling bi-objective feature selection in classification, where an interpolation based initialization method is designed for covering a wide range of search space and exploring the adaptively detected regions of interest. In experiments, IPEA is been compared with four state-of-the-art MOEAs in terms of two widely-used performance metrics on a list of 20 public real-world classification datasets with the dimensionality ranging from low to high. The overall empirical results suggest that IPEA generally performs the best of all tested algorithms, with significantly better search abilities and much lower computational time cost.

1. Introduction

As is known to all, evolutionary algorithms (EAs) have been widely used to solve optimization problems for decades [1], especially for multi-objective optimization problems (MOPs) with multiple contradictory objectives [2], which are then called the multi-objective evolutionary algorithms (MOEAs) [3]. Today, a great variety of MOEAs have been proposed from all around the world; these can be roughly divided into several categories: dominance-based [4,5,6], which adopt the nondominated sorting method for environmental selection; decomposition-based [7,8,9,10], which decompose an MOP into a series of simpler single-objective problems with weight vectors; indicator-based [11,12,13,14], which make use of a certain performance indicators for environmental selection; surrogate-based [15,16,17], which introduce a surrogate model for tackling expensive optimization problems; and multi-task [18,19,20], which generate multiple independent and cooperative tasks for different evolutionary purposes.
What is more, owing to their population-based search capability and ability to function without domain knowledge, MOEAs have also been applied for solving many real-world optimization problems [21,22,23], such as task offloading [24,25], network construction [26,27], community detection [28,29], and the feature selection problem [30,31] that is the focus of this work. To be more specific, feature selection is a common and important data preprocessing technique [32,33] which selects only a subset of useful features for classification rather than all of them. It is especially useful for large-scale datasets or those with high dimensionality [34]. When aimed at minimizing both the ratio of selected features and the ratio of classification errors, feature selection turns out to be a multi-objective (bi-objective) optimization problem, which can be formally described as follows [35]:
m i n i m i z e F ( x ) = ( f 1 ( x ) , f 2 ( x ) , , f M ( x ) ) T s u b j e c t t o x = ( x 1 , x 2 , , x D ) , x i = { 0 , 1 }
where M is the number of objectives, which in bi-objective feature selection is set to 2, D is the total number of features in the decision space, F ( x ) is the objective vector of x , with f i ( x ) denoting the objective value in the f i direction, x = ( x 1 , x 2 , , x D ) is the decision vector of x , where x i = 1 means selecting the ith feature and x i = 0 means not selecting that feature, and f 1 ( x ) denotes the ratio of selected features, which can be further defined as follows:
f 1 ( x ) = i = 1 D x i / D
with the discrete value ranging from 0 to 1 (i.e., 0 , 1 / D , 2 / D , , 1 ). Moreover, given the results in terms of T P (True Positive), T N (True Negative), F P (False Positive), and F N (False Negative), f 2 ( x ) (representing the ratio of classification errors related to the above selected features) can be further defined as follows:
f 2 ( x ) = F P + F N T P + T N + F P + F N .
Despite of the aforementioned advantages, traditional MOEAs still face the curse of dimensionality in tackling bi-objective feature selection, which can cause the total number of features explodes. While many existing works have attempted to deal with this challenge [36,37], most of these have either introduced complicated frameworks or the need for many essential preset parameters. Moreover, it is difficult to balance the algorithm’s performance on both low-dimensional datasets and high-dimensional ones. Therefore, in this paper an interpolation-based EA (termed IPEA) is proposed that is specifically designed for tackling bi-objective feature selection in classification and is suitable for solving both low- and high-dimensional datasets. In IPEA, an interpolation-based initialization method is designed in order to provide a promising hybrid initial population that not only covers a wide range of search space but also allows the algorithm to explore adaptively within the most promising local areas of interest. In addition, a simple and efficient reproduction method is adopted to provide more population diversity and faster algorithm convergence during evolution. Combining the above-mentioned initialization and reproduction processes increases the search ability of IPEA in handling both low- and high-dimensional datasets.
The remainder of this paper is organized as follows: first, the related works are introduced in Section 2; then, the proposed IPEA is detailed in Section 3; the experimental setup is described in Section 4, while the empirical results are presented in Section 5; finally, the paper is concluded in Section 6.

2. Related Works

Evolutionary feature selection [38] can be generally categorized into wrapper-based and filter-based approaches [39,40]. In brief, the wrapper-based approaches [41,42] adopt a classification model such as SVM (Support Vector Machine) or KNN (K-Nearest Neighbors) [43] to evaluate the classification accuracy corresponding to the selected feature subset. By contrast, filter-based approaches [44,45] are independent of any classifier and directly analyze the classification data to explore the explicit or implicit relationships between features and the corresponding classes while ignoring the classification results of the feature subset, which is selected later. Therefore, wrapper-based approaches are relatively more accurate but typically have higher computational costs due to the additional classification process [20,46,47].
In this paper, a wrapper-based approach for evolutionary bi-objective feature selection is chosen. This is a widely used approach that has been discussed in many existing feature selection works over the last several years. For example, in 2020 Xu et al. [37] proposed a segmented initialization method and offspring modification method; however, the key parameter settings are all fixed and cannot be adaptively altered for different classification datasets or high-dimensional feature selection. Subsequently, Xu et al. [35] proposed an evolutionary algorithm named DAEA based on duplication analysis along with an efficient reproduction method; however, this algorithm’s performance has not been tested under the circumstances of high-dimensional feature selection. Following the idea of DAEA, Jiao et al. [48] further improved the method for handling solution duplication and designed a problem reformulation mechanism named PRDH; however, its applicability across different MOEA frameworks remains unconfirmed.
Therefore, this work attempts to design an interpolation based evolutionary algorithm, termed IPEA, which is suitable for tackling both relatively low-dimensional datasets and relatively high-dimensional ones. The framework of IPEA should be not as complicated as many other existing MOEAs designed for feature selection, and the algorithm should be able to effectively adapt to different optimization environment across a wide variety of features while maintaining a good balance between low- and high-dimensional datasets. This represents the major motivation behind this paper.

3. Proposed Algorithm

In this section, the general framework of the proposed algorithm IPEA is first introduced, then its two specially designed components, i.e., the initialization and reproduction processes, are illustrated further.

3.1. General Framework

The general framework of the IPEA is shown in Algorithm 1, where the population size N and number of total features D are input as the primary parameters. The general framework of IPEA is rather similar to traditional MOEAs, but uses the initialization and reproduction processes specially designed in this paper. Moreover, the environmental selection process has been modified by adding the ability to remove duplicated solutions in the current union population P o p , as can be seen in Line 5 of Algorithm 1. Nevertheless, the widely-used truncation method based on nondominated sorting and crowding distances, first introduced in the classic dominance-based algorithm NSGA-II [4], is still adopted in IPEA to select and reserve the N best solutions in P o p . The termination criterion of the IPEA is preset based on counting the number of evaluated objective functions, which is normally equal to the total number of generated solutions during evolution, as described in detail in Section 4.
Algorithm 1  G e n e r a l F r a m e w o r k ( N , D )
  • Input: population size N, total feature number D;
  • Output: final population P o p ;
1:
P o p = I n i t i a l i z e ( N , D ) ; / / Algorithm 2, also invoking Algorithms 3–5
2:
while termination criterion is not reached do
3:
     P o p * = R e p r o d u c e ( P o p , N , D ) ; / / Algorithm 6
4:
     P o p = P o p P o p * ;
5:
     P o p remove duplicated solutions in P o p ;
6:
     P o p select N solutions in P o p by nondominated sorting and crowding distances;
7:
end while
Algorithm 2  I n i t i a l i z e ( N , D )
  • Input: population size N, total feature number D;
  • Output: initial population P o p ;
1:
N i p = [ N ] ; / / number of interpolation positions
2:
T = 1 ( N i p + 1 ) ; / / average interpolation interval
3:
N s p = [ N N i p ] ; / / interpolation subpopulation size
4:
B e s t f 2 = ; / / best objective value in the f 2 direction
5:
for  i = 1 , 2 , , N i p  do
6:
     t 1 = m i n ( N s p , N N s p ( i 1 ) ) ; / / get the smaller one
7:
     t 2 = T i ; / / get the interpolation position
8:
     s u b P o p = N e w S u b P o p ( t 1 , D , t 2 ) ;
9:
     P o p 1 = P o p 1 s u b P o p ;
10:
     t 3 get the best f 2 objective value in s u b P o p ;
11:
    if  B e s t f 2 > t 3  then
12:
         B e s t f 2 = t 3 ;
13:
         B e s t i p = t 2 ;
14:
    end if
15:
end for
16:
P o p 2 = N e w S u b P o p ( N , D , B e s t i p ) ;
17:
P o p 3 = F o r w a r d S u b P o p ( m i n ( N , D ) , D ) ;
18:
P o p 4 = B a c k w a r d S u b P o p ( m i n ( N , D ) , D ) ;
19:
P o p select N best solutions from ( P o p 1 P o p 2 P o p 3 P o p 4 ) by nondominated sorting and crowding distances;

3.2. Initialization Process

The initialization process of the IPEA is shown in Algorithm 2, which also invokes Algorithms 3–5. The initialization process is the major innovation of the IPEA, which is based on the utilization of interpolations in the objective space. Overall, the final initial population P o p is made up of four hybrid subpopulations; while P o p 1 is based on a series of adaptively generated interpolation positions in the objective space, P o p 2 is based on the analysis of the previously generated P o p 1 . In turn, P o p 3 and P o p 4 are respectively inspired by the classic forward and backward search ideas in feature selection. To be more specific, the number of interpolation positions for new subpopulations generated in the objective space is adaptively set in Line 1 of Algorithm 2, the average interval between each two adjacent interpolation positions is accordingly set in Line 2 of Algorithm 2, and the standard interpolation subpopulation size is accordingly set in Line 3 of Algorithm 2. The detailed generation process of each interpolation-based subpopulation is shown from Line 5 to 15 in Algorithm 2, which is based on iteratively searching for the best f 2 objective value in all the subpopulations. During this process, Algorithm 3 is also invoked to generate new subpopulations according to the input interpolation position value. Another two pseudocodes, Algorithms 4 and 5, are respectively invoked to generate the forward search-inspired subpopulation and the backward search-inspired one. As can be seen in Algorithms 4 and 5, the former contains unique solutions with only one selected feature, while the later contains unique solutions with only one unselected feature.
Algorithm 3  N e w S u b P o p ( K , D , I p )
  • Input: subpopulation size K, total feature number D, interpolation position I p ;
  • Output: new subpopulation S u b P o p ;
1:
S u b P o p = Z e r o s ( K , D ) ; / / create a matrix of zeros
2:
for  i = 1 , 2 , , K   do
3:
    for  j = 1 , 2 , , D  do
4:
        if  ρ < I p  then / / ρ is a random probability
5:
            S u b P o p ( i , j ) = 1 ; / / select the jth feature
6:
        end if
7:
    end for
8:
end for
Algorithm 4  F o r w a r d S u b P o p ( K , D )
  • Input: subpopulation size K, total feature number D;
  • Output: new subpopulation S u b P o p ;
1:
H randomly select K unique integers from 1 to D;
2:
S u b P o p = Z e r o s ( K , D ) ; / / create a matrix of zeros
3:
for  i = 1 , 2 , , K  do
4:
     S u b P o p ( i , H ( i ) ) = 1 ; / / select that feature
5:
end for
Algorithm 5  B a c k w a r d S u b P o p ( K , D )
  • Input: subpopulation size K, total feature number D;
  • Output: new subpopulation S u b P o p ;
1:
H randomly select K unique integers from 1 to D;
2:
S u b P o p = O n e s ( K , D ) ; / / create a matrix of ones
3:
for  i = 1 , 2 , , K  do
4:
     S u b P o p ( i , H ( i ) ) = 0 ; / / not select that feature
5:
end for
Figure 1 provides an illustrative example of how the adaptive interpolation-based initialization process in the IPEA works when the population size is set to N = 16 . Thus, according to the previously introduced Algorithm 2, the number of interpolation positions is set to N i p = 4 , the average interpolation interval is set to T = 0.2 , and the standard size of each interpolation based subpopulation is set to N s p = 4 . As can be seen from Figure 1, four interpolation-based subpopulations (together forming P o p 1 ) with an average size of four solutions are generated just around their related interpolation positions, i.e., 0.2, 0.4, 0.6, and 0.8. Apart from these four interpolation positions, two boundary interpolation positions, i.e., 1 D and ( D 1 ) D , are also utilized, which respectively correspond to the previously introduced forward search-inspired subpopulation ( P o p 3 generated by Algorithm 4) and backward search-inspired subpopulation ( P o p 4 generated by Algorithm 5). In the case of Figure 1, the previously introduced best f 2 objective value lies in the subpopulation related to the 0.4 interpolation. Then, another new subpopulation with the full population size is generated by Algorithm 3 with the interpolation position value 0.4 input as the key parameter. In this way, a full population size of new solutions (i.e., P o p 2 ) is initialized around the most promising interpolation position found above. Therefore, combining all four of the above interpolation-related ( P o p 1 P o p 2 P o p 3 P o p 4 ) , as executed in Line 19 of Algorithm 2, covers a wide range of objective space while exploring adaptively within the boundary and promising local areas. It should be noted that in Figure 1 only shows P o p 1 with 16 solutions distributed, while the rest of P o p 2 , P o p 3 , and P o p 4 are not displayed in Figure 1.

3.3. Reproduction Process

The reproduction process of the IPEA is shown in Algorithm 6, which acts as an efficient supplement for the previously introduced interpolation-based initialization process. As can be seen from Algorithm 6, a totally random mating process first is conducted in Line 1 to select N pairs of parent solutions from the current population P o p , indicating that every solution holds the same opportunity for reproduction. Then, the corresponding indexes with different decision variable values are found for each pair of parents; these are used in the later crossover operation. This is key to realizing efficient crossover, i.e., avoiding all invalid crossover operations between the same decision variable values. The adaptive crossover operation is shown from Line 4 to 9, which is based on the dynamic length of t 1 and a random integer limited by t 2 . After crossover, the mutation process is quite simple, adopting the traditional bitwise mutation method widely used in many other MOEAs. The bitwise mutation rate is set to 1 D , implying a relatively delicate and cautious mutation operation within a parent decision vector. Finally, the mutated parent solution is reserved as the expected new offspring.
Algorithm 6  R e p r o d u c e ( P o p , N , D )
  • Input: current population P o p , expected offspring number N, total feature number D;
  • Output: offspring set P o p * ;
1:
P a r s randomly select N pairs of parent solutions from P o p ;
2:
for  i = 1 , 2 , , N   do
3:
     p 1 and p 2 ← get the ith pair of parents from P a r s ;
4:
     t 1 find indexes with different decision variable values between p 1 and p 2 ;
5:
     t 2 get the length of vector t 1 ;
6:
     t 3 randomly select an integer from 1 to t 2 ;
7:
     t 4 randomly select t 3 unique integers from 1 to t 2 ;
8:
     j = t 1 ( t 4 ) ; / / get decision variable indexes for crossover
9:
     p 1 ( j ) = p 2 ( j ) ; / / crossover between two parents
10:
     ρ get a vector of D random probabilities;
11:
     j find indexes with values in ρ smaller than 1 D ;
12:
     p 1 ( j ) = ¬ p 1 ( j ) ; / / mutation within one parent
13:
     P o p * ( i ) = p 1 ;
14:
end for

4. Experimental Setup

4.1. Comparison Algorithms

In this work, four state-of-the-art MOEAs are used as comparison algorithms against the proposed algorithm, IPEA. The compared algorithms are NSGA-II (a nondominated sorting based genetic algorithm) [4], MOEA/D (an MOEA based on decomposition) [7], MOEA/HD (an MOEA based on hierarchical decomposition) [10], and PMEA [13] (a polar metric-based evolutionary algorithm). To be more specific, NSGA-II and MOEA/D are among the most classic and well-known MOEAs, respectively based on dominance and decomposition; MOEA/HD is a recently published MOEA based on both dominance and decomposition, especially designed for solving MOPs with complex Pareto fronts; and PMEA is a recently published MOEA based on a performance indicator called the polar metric, which balances both the population diversity and the algorithm convergence during evolution. Overall, the three mainstream MOEA approaches, i.e., those based on dominance, decomposition and indicator, are all included in the experiments.

4.2. Classification Datasets

A total of 20 open-source classification datasets [49] are used as test problems to check the general optimization performances of all comparison algorithms for bi-objective feature selection. The attribute values of each tested dataset are shown in Table 1, with the total number of features sorted in ascending order. It can be seen from Table 1 that the total number of features for each dataset changes from 60 to 10,509, which covers a wide variety of features ranging from relatively lower to relatively higher dimensionality. Moreover, the number of classification samples changes from 50 to 606 and the number of corresponding classes changes from 2 to 15, suggesting the comprehensiveness of the test instances. In fact, most classification datasets used in this paper originate from real-world scenarios; for example, the “Sonar” dataset contains various patterns obtained by bouncing sonar signals off of a metal cylinder at various angles.

4.3. Performance Indicators

For comprehensive analyses based on empirical data, this paper uses multiple performance indicators to measure the general performance of all compared algorithms on all of the tested datasets. More specifically, the Hypervolume (HV) [50] indicator is used as the main metric to measure general optimization performance regarding both population diversity and algorithm convergence. The reference point for HV is uniformly set to ( 1 , 1 ) in the bi-objective space. In addition, the Inverted Generation Distance Plus (IGD+) [51] indicator, which also covers the measurement of both diversity and convergence, is used to supplement the HV. The preference of IGD+ is highly related to the choice of reference points. Thus, for the sake of fairness, in this paper the reference points of IGD+ are set to the combination of all final nondominated solutions obtained by all of the tested algorithms. Normally, a greater HV value is preferable, while a smaller IGD+ value is preferable. Lastly, Wilcoxon’s test with a 5 % significance level is adopted to check the significant differences between each two compared algorithms; results below the level of significance are prefixed by the symbol ★.

4.4. Parameter Settings

In the experiments, all of the compared algorithms adopt the same parameter settings stated in the original papers. All algorithms are coded on an open-source MATLAB platform called PlatEMO [52]. Prior to evolution, each classification dataset is randomly divided into training and test subsets in a proportion of 70 / 30 according to the stratified split process [35]. Then, a KNN ( K = 5 ) classification model is utilized with 10-fold cross-validation in order to avoid feature selection bias [53]. Lastly, each experimental run was independently executed 20 times, the population size for each algorithm was set to 100, and the termination criterion for each algorithm (i.e., the number of objective function evaluations) was set to 10,000, which is about 100 evolutionary generations.

5. Experimental Studies

In this section, the empirical results are first analyzed in terms of the HV and IGD+ metrics, shown in Table 2 and Table 3, respectively. In addition, the minimum classification error obtained by each algorithm along with the corresponding number of selected features are studied in Table 4. Then, the final nondominated solution distributions in the objective space, shown in Figure 2, are further analyzed. Finally, the computational time cost is analyzed by recording the experimental running time, with the results shown in Table 5.

5.1. Empirical Result Analyses

The general performance of each algorithm on all classification datasets in terms of the HV and IGD+ metrics is shown in Table 2 and Table 3, respectively, with the mean metric values and corresponding standard deviations displayed in each two rows. First, it can be seen from Table 2 and Table 3 that the proposed algorithm IPEA has the best performance of all the compared algorithms on all tested datasets, and the advantages of IPEA over all the other algorithms are rather significant on almost all of the tested datasets. Because both the HV and IGD+ metrics take the population diversity and convergence into consideration, the success of the IPEA in terms of these two metrics obviously suggests its superiority in both diversity and convergence. Moreover, the excellent performance of the IPEA on all of the datasets implies its great advantages and outstandingly effective search ability in tackling bi-objective feature selection not only for low-dimensional datasets but also for high-dimensional ones. In fact, the advantages of the IPEA in tackling the high-dimensional datasets appear to be even more obvious, showing huge differences in magnitude in the results shown in Table 3 due to the sparsity of feature space and the difficulty of finding more optimal solutions.

5.2. Classification Performance Analyses

The empirical results of each algorithm run on each dataset in terms of classification performance are shown in Table 4, with the mean minimum classification error values and corresponding mean number of selected features displayed in each two rows. First, it can be seen from Table 4 that the proposed IPEA has the best performance of all compared algorithms on all tested datasets in terms of the classification results. Again, the advantages of the IPEA over the other algorithms are quite significant on almost all the datasets, with the exception of the first two lower-dimensional ones, which do not fully challenge the IPEA’s search abilities. The minimum classification error (representing the best classification accuracy) obtained by the IPEA is generally much better than the other algorithms on each dataset. Moreover, because the corresponding number of selected features (shown as the NSF values in Table 4) generally affects the classification time and cost (as more selected features means more time and cost), the obviously superior NSF values obtained by the IPEA imply excellent computational efficiency, which actually proved in Section 5.4 below.
The implementation process of the whole wrapper-based evolutionary feature selection experiment is described as follows. At first, a population of N initial solutions containing different subsets of the selected features is evolved generation-by-generation. During this process, the selected features in each solution are only applied to training data that have been previously split from the tested datasets in a proportion of 70 / 30 ; meanwhile a KNN ( K = 5 ) model and 10-fold cross-validation are executed for classification, returning the f 1 and f 2 objective values for the current solutions. When terminating the iterative evolution process, a set of nondominated solutions is selected from the last population, which is then applied to the test data to obtain the final classification results shown in Table 4 and the HV and IGD+ performance results shown in Table 2 and Table 3.

5.3. Nondominated Solution Distributions

The final nondominated solution distributions in the objective space obtained by each algorithm on each dataset are shown in Figure 2, along with 20 subfigures showing the final obtained nondominated solutions on the training data during evolution. For the sake of fairness, we choose the run with median HV performance from among 20 independent runs on each dataset. Because the nondominated solutions cannot be used to calculate the mean values, as there is no mean nondominated solution, we use the median performance here instead of the mean performance. Overall, it can be seen from Figure 2 that the IPEA obviously performs the best among all the algorithms, with more diverse solution distributions and more converged objective values in both the f 1 and f 2 directions.
To be more specific, the nondominated solution distributions of all the algorithms compared in the first four subfigures relate to relatively low-dimensional datasets. In Figure 2a–d the results appear quite similar, although the IPEA still performs the best. However, in the rest of the subfigures, i.e., Figure 2e–t, which are related to relatively high-dimensional datasets, the advantages of the IPEA in terms of the nondominated solution distributions become tremendous, leaving all the other algorithms’ nondominated solution distributions far behind.
Moreover, it can be observed from Figure 2 that the nondominated solution distributions are relatively sparse on the high-dimensional datasets compared to the low-dimensional datasets. This is because the feature space or search space becomes sparser as the number of total features increases, making it much more difficult for MOEAs to find enough optimal solutions during evolution. This phenomenon also happens to the proposed IPEA; however, the negative impact is been greatly reduced compared with the other algorithms.

5.4. Computational Time Analyses

The mean computational time cost for each algorithm run on each dataset is recorded in seconds, with the results shown in Table 5. It can be seen from Table 5 that the mean computational time cost of the IPEA is always the smallest among all the algorithms on each dataset, with significant advantages over all other algorithms. Moreover, the advantages of the IPEA in terms of the computational time are even more significant on the relatively high-dimensional datasets, generally incurring only half the other algorithms’ mean computational time cost. In fact, if only considering the MOEA framework, the theoretical time complexity of IPEA is not very different from that of many other traditional MOEAs, as the general framework of the IPEA inherits the basic ideas of most MOEAs. The worst theoretical time complexity of the IPEA can be estimated as O ( M N 2 ) with M as the number of objectives and N as the population size, which mainly comes from the process of nondominated sorting. However, the real running time of the IPEA on each dataset is much smaller than the theoretical expectations. This is mainly because the IPEA saves a great deal of time during the classification process owing to its better selection of feature subsets, with much smaller numbers of selected features being used for classification. Therefore, the computational efficiency of the IPEA regarding the whole evolutionary feature selection process is rather time-saving, with significant efficiency advantages over the compared algorithms on each tested dataset, including both the relatively low-dimension and relatively high-dimension ones.

6. Conclusions

In this work, an interpolation-based EA (termed the IPEA) is proposed for tackling bi-objective feature selection on both low-dimensional and high-dimensional classification datasets. The design of the IPEA incorporates an interpolation-based initialization method in order to provide a promising hybrid initial population, allowing it to cover a wide range of search space while adaptively exploring the regions of interest. Moreover, an efficient reproduction method is adopted to provide greater population diversity and better algorithm convergence during evolution. The framework of the IPEA is simple but effective, and can adapt to different kinds of optimization environments across a wide variety of features. The outstanding performance and significant advantages of the IPEA have been verified in the experiments and comprehensively analyzed. When compared with four state-of-the-art MOEAs on a list of 20 public real-world classification datasets, the IPEA always performs best in terms of two widely-used performance metrics, showing excellent search abilities. Furthermore, the IPEA also shows the best performance in terms of the final nondominated solution distributions in the objective space, and is proven to be the most computationally efficient among all tested algorithms in terms of the real running time recorded on each tested dataset.
In future work, it is planned to study the feasibility of the proposed algorithm IPEA on additional kinds of discrete optimization problems, such as neural network construction and community node detection.

Funding

This research was funded by National Natural Science Foundation of China (grant number 62103209), the Natural Science Foundation of Fujian Province (grant number 2020J05213), the Scientific Research Project of Putian Science and Technology Bureau (grant number 2021ZP07), the Startup Fund for Advanced Talents of Putian University (grant number 2019002), and the Research Projects of Putian University (grant number JG202306).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China under Grant 62103209, by the Natural Science Foundation of Fujian Province under Grant 2020J05213, by the Scientific Research Project of Putian Science and Technology Bureau under Grant 2021ZP07, by the Startup Fund for Advanced Talents of Putian University under Grant 2019002, and by the Research Projects of Putian University under Grant JG202306.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Eiben, A.E.; Smith, J.E. What is an evolutionary algorithm? In Introduction to Evolutionary Computing; Springer: Berlin/Heidelberg, Germany, 2015; pp. 25–48. [Google Scholar]
  2. Coello, C.A.C.; Lamont, G.B.; Van Veldhuizen, D.A. Evolutionary Algorithms for Solving Multi-Objective Problems; Springer: New York, NY, USA, 2007; Volume 5. [Google Scholar]
  3. Zhou, A.; Qu, B.Y.; Li, H.; Zhao, S.Z.; Suganthan, P.N.; Zhang, Q. Multiobjective evolutionary algorithms: A survey of the state of the art. Swarm Evol. Comput. 2011, 1, 32–49. [Google Scholar]
  4. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  5. Deb, K.; Jain, H. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints. IEEE Trans. Evol. Comput. 2014, 18, 577–601. [Google Scholar] [CrossRef]
  6. Tian, Y.; Cheng, R.; Zhang, X.; Su, Y.; Jin, Y. A Strengthened Dominance Relation Considering Convergence and Diversity for Evolutionary Many-Objective Optimization. IEEE Trans. Evol. Comput. 2019, 23, 331–345. [Google Scholar]
  7. Zhang, Q.; Li, H. MOEA/D: A Multiobjective Evolutionary Algorithm Based on Decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  8. Li, H.; Zhang, Q. Multiobjective Optimization Problems With Complicated Pareto Sets, MOEA/D and NSGA-II. IEEE Trans. Evol. Comput. 2009, 13, 284–302. [Google Scholar] [CrossRef]
  9. Li, K.; Zhang, Q.; Kwong, S.; Li, M.; Wang, R. Stable Matching-Based Selection in Evolutionary Multiobjective Optimization. IEEE Trans. Evol. Comput. 2014, 18, 909–923. [Google Scholar] [CrossRef]
  10. Xu, H.; Zeng, W.; Zhang, D.; Zeng, X. MOEA/HD: A Multiobjective Evolutionary Algorithm Based on Hierarchical Decomposition. IEEE Trans. Cybern. 2019, 49, 517–526. [Google Scholar] [CrossRef]
  11. Bader, J.; Zitzler, E. HypE: An Algorithm for Fast Hypervolume-Based Many-Objective Optimization. Evol. Comput. 2011, 19, 45–76. [Google Scholar] [CrossRef]
  12. Liang, Z.; Luo, T.; Hu, K.; Ma, X.; Zhu, Z. An Indicator-Based Many-Objective Evolutionary Algorithm With Boundary Protection. IEEE Trans. Cybern. 2021, 51, 4553–4566. [Google Scholar] [CrossRef]
  13. Xu, H.; Zeng, W.; Zeng, X.; Yen, G.G. A Polar-Metric-Based Evolutionary Algorithm. IEEE Trans. Cybern. 2021, 51, 3429–3440. [Google Scholar] [CrossRef] [PubMed]
  14. Lopes, C.L.d.V.; Martins, F.V.C.; Wanner, E.F.; Deb, K. Analyzing Dominance Move (MIP-DoM) Indicator for Multiobjective and Many-Objective Optimization. IEEE Trans. Evol. Comput. 2022, 26, 476–489. [Google Scholar] [CrossRef]
  15. Wang, H.; Jin, Y.; Sun, C.; Doherty, J. Offline data-driven evolutionary optimization using selective surrogate ensembles. IEEE Trans. Evol. Comput. 2018, 23, 203–216. [Google Scholar]
  16. Lin, Q.; Wu, X.; Ma, L.; Li, J.; Gong, M.; Coello, C.A.C. An Ensemble Surrogate-Based Framework for Expensive Multiobjective Evolutionary Optimization. IEEE Trans. Evol. Comput. 2022, 26, 631–645. [Google Scholar] [CrossRef]
  17. Sonoda, T.; Nakata, M. Multiple Classifiers-Assisted Evolutionary Algorithm Based on Decomposition for High-Dimensional Multiobjective Problems. IEEE Trans. Evol. Comput. 2022, 26, 1581–1595. [Google Scholar] [CrossRef]
  18. Da, B.; Gupta, A.; Ong, Y.S.; Feng, L. Evolutionary multitasking across single and multi-objective formulations for improved problem solving. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1695–1701. [Google Scholar]
  19. Gupta, A.; Ong, Y.S.; Feng, L.; Tan, K.C. Multiobjective Multifactorial Optimization in Evolutionary Multitasking. IEEE Trans. Cybern. 2017, 47, 1652–1665. [Google Scholar]
  20. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. Evolutionary Multitasking for Feature Selection in High-Dimensional Classification via Particle Swarm Optimization. IEEE Trans. Evol. Comput. 2022, 26, 446–460. [Google Scholar] [CrossRef]
  21. Cao, F.; Tang, Z.; Zhu, C.; Zhao, X. An Efficient Hybrid Multi-Objective Optimization Method Coupling Global Evolutionary and Local Gradient Searches for Solving Aerodynamic Optimization Problems. Mathematics 2023, 11, 3844. [Google Scholar] [CrossRef]
  22. Garces-Jimenez, A.; Gomez-Pulido, J.M.; Gallego-Salvador, N.; Garcia-Tejedor, A.J. Genetic and Swarm Algorithms for Optimizing the Control of Building HVAC Systems Using Real Data: A Comparative Study. Mathematics 2021, 9, 2181. [Google Scholar] [CrossRef]
  23. Ramos-Pérez, J.M.; Miranda, G.; Segredo, E.; León, C.; Rodríguez-León, C. Application of Multi-Objective Evolutionary Algorithms for Planning Healthy and Balanced School Lunches. Mathematics 2021, 9, 80. [Google Scholar] [CrossRef]
  24. Long, S.; Zhang, Y.; Deng, Q.; Pei, T.; Ouyang, J.; Xia, Z. An Efficient Task Offloading Approach Based on Multi-Objective Evolutionary Algorithm in Cloud-Edge Collaborative Environment. IEEE Trans. Netw. Sci. Eng. 2023, 10, 645–657. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Ma, S.; Jiang, X. Research on Multi-Objective Multi-Robot Task Allocation by Lin-Kernighan-Helsgaun Guided Evolutionary Algorithms. Mathematics 2022, 10, 4714. [Google Scholar] [CrossRef]
  26. Xue, Y.; Chen, C.; Slowik, A. Neural Architecture Search Based on a Multi-Objective Evolutionary Algorithm With Probability Stack. IEEE Trans. Evol. Comput. 2023, 27, 778–786. [Google Scholar] [CrossRef]
  27. Ponti, A.; Candelieri, A.; Giordani, I.; Archetti, F. Intrusion Detection in Networks by Wasserstein Enabled Many-Objective Evolutionary Algorithms. Mathematics 2023, 11, 2342. [Google Scholar] [CrossRef]
  28. Zhu, W.; Li, H.; Wei, W. A Two-Stage Multi-Objective Evolutionary Algorithm for Community Detection in Complex Networks. Mathematics 2023, 11, 2702. [Google Scholar] [CrossRef]
  29. Gao, C.; Yin, Z.; Wang, Z.; Li, X.; Li, X. Multilayer Network Community Detection: A Novel Multi-Objective Evolutionary Algorithm Based on Consensus Prior Information [Feature]. IEEE Comput. Intell. Mag. 2023, 18, 46–59. [Google Scholar] [CrossRef]
  30. Xu, H.; Xue, B.; Zhang, M. A Bi-Search Evolutionary Algorithm for High-Dimensional Bi-Objective Feature Selection. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 1–14. [Google Scholar] [CrossRef]
  31. Xu, H.; Xue, B.; Zhang, M. Probe Population Based Initialization and Genetic Pool Based Reproduction for Evolutionary Bi-Objective Feature Selection. IEEE Trans. Evol. Comput. 2024, 1. [Google Scholar] [CrossRef]
  32. Nguyen, B.H.; Xue, B.; Andreae, P.; Ishibuchi, H.; Zhang, M. Multiple Reference Points-Based Decomposition for Multiobjective Feature Selection in Classification: Static and Dynamic Mechanisms. IEEE Trans. Evol. Comput. 2020, 24, 170–184. [Google Scholar] [CrossRef]
  33. Khalid, S.; Khalil, T.; Nasreen, S. A survey of feature selection and feature extraction techniques in machine learning. In Proceedings of the 2014 Science and Information Conference, London, UK, 27–29 August 2014; pp. 372–378. [Google Scholar] [CrossRef]
  34. Cheng, F.; Cui, J.; Wang, Q.; Zhang, L. A Variable Granularity Search-Based Multiobjective Feature Selection Algorithm for High-Dimensional Data Classification. IEEE Trans. Evol. Comput. 2023, 27, 266–280. [Google Scholar] [CrossRef]
  35. Xu, H.; Xue, B.; Zhang, M. A Duplication Analysis-Based Evolutionary Algorithm for Biobjective Feature Selection. IEEE Trans. Evol. Comput. 2021, 25, 205–218. [Google Scholar] [CrossRef]
  36. Cheng, F.; Chu, F.; Xu, Y.; Zhang, L. A Steering-Matrix-Based Multiobjective Evolutionary Algorithm for High-Dimensional Feature Selection. IEEE Trans. Cybern. 2022, 52, 9695–9708. [Google Scholar] [CrossRef] [PubMed]
  37. Xu, H.; Xue, B.; Zhang, M. Segmented Initialization and Offspring Modification in Evolutionary Algorithms for Bi-Objective Feature Selection. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference, GECCO’20, Cancun, Mexico, 8–12 July 2020; pp. 444–452. [Google Scholar]
  38. De La Iglesia, B. Evolutionary computation for feature selection in classification problems. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2013, 3, 381–407. [Google Scholar]
  39. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput. 2015, 20, 606–626. [Google Scholar]
  40. Dokeroglu, T.; Deniz, A.; Kiziloz, H.E. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 2022, 494, 269–296. [Google Scholar]
  41. Mukhopadhyay, A.; Maulik, U. An SVM-wrapped multiobjective evolutionary feature selection approach for identifying cancer-microRNA markers. IEEE Trans. Nanobiosci. 2013, 12, 275–281. [Google Scholar]
  42. Vignolo, L.D.; Milone, D.H.; Scharcanski, J. Feature selection for face recognition based on multi-objective evolutionary wrappers. Expert Syst. Appl. 2013, 40, 5077–5084. [Google Scholar]
  43. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  44. Lazar, C.; Taminau, J.; Meganck, S.; Steenhoff, D.; Coletta, A.; Molter, C.; de Schaetzen, V.; Duque, R.; Bersini, H.; Nowe, A. A survey on filter techniques for feature selection in gene expression microarray analysis. IEEE/ACM Trans. Comput. Biol. Bioinform. TCBB 2012, 9, 1106–1119. [Google Scholar]
  45. Xue, B.; Cervante, L.; Shang, L.; Browne, W.N.; Zhang, M. Multi-objective evolutionary algorithms for filter based feature selection in classification. Int. J. Artif. Intell. Tools 2013, 22, 1350024. [Google Scholar]
  46. Xue, B.; Zhang, M.; Browne, W.N. Particle swarm optimisation for feature selection in classification: Novel initialisation and updating mechanisms. Appl. Soft Comput. 2014, 18, 261–276. [Google Scholar]
  47. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. An Evolutionary Multitasking-Based Feature Selection Method for High-Dimensional Classification. IEEE Trans. Cybern. 2022, 52, 7172–7186. [Google Scholar] [CrossRef] [PubMed]
  48. Jiao, R.; Xue, B.; Zhang, M. Solving Multi-objective Feature Selection Problems in Classification via Problem Reformulation and Duplication Handling. IEEE Trans. Evol. Comput. 2022, 28, 846–860. [Google Scholar] [CrossRef]
  49. Kelly, M.; Longjohn, R.; Nottingham, K. The UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu (accessed on 18 August 2024).
  50. While, L.; Hingston, P.; Barone, L.; Huband, S. A faster algorithm for calculating Hypervolume. IEEE Trans. Evol. Comput. 2006, 10, 29–38. [Google Scholar]
  51. Ishibuchi, H.; Imada, R.; Masuyama, N.; Nojima, Y. Comparison of Hypervolume, IGD and IGD+ from the Viewpoint of Optimal Distributions of Solutions. In Proceedings of the Evolutionary Multi-Criterion Optimization 2019, East Lansing, MI, USA, 10–13 March 2019; Deb, K., Goodman, E., Coello Coello, C.A., Klamroth, K., Miettinen, K., Mostaghim, S., Reed, P., Eds.; Springer: Cham, Switzerland, 2019; pp. 332–345. [Google Scholar]
  52. Tian, Y.; Cheng, R.; Zhang, X.; Jin, Y. PlatEMO: A MATLAB Platform for Evolutionary Multi-Objective Optimization. IEEE Comput. Intell. Mag. 2017, 12, 73–87. [Google Scholar]
  53. Tran, B.; Xue, B.; Zhang, M.; Nguyen, S. Investigation on particle swarm optimisation for feature selection on high-dimensional data: Local search and selection bias. Connect. Sci. 2016, 28, 270–294. [Google Scholar]
Figure 1. An example of how the adaptive interpolation-based initialization process in the IPEA works when the population size is set to N = 16 .
Figure 1. An example of how the adaptive interpolation-based initialization process in the IPEA works when the population size is set to N = 16 .
Mathematics 12 02572 g001
Figure 2. Nondominated solution distributions in the objective space corresponding to the runs of median HV performances obtained by each algorithm on each dataset. (a) Sonar. (b) HillValley. (c) Synthetic. (d) Arrhythmia. (e) Yale. (f) Colon. (g) SRBCT. (h) AR10P. (i) PIE10P. (j) Leukemia1. (k) Tumor9. (l) Brain1. (m) Leukemia2. (n) CNS. (o) ALLAML. (p) Nci9. (q) Pixraw10P. (r) Orlraws10P. (s) Brain2. (t) Prostate.
Figure 2. Nondominated solution distributions in the objective space corresponding to the runs of median HV performances obtained by each algorithm on each dataset. (a) Sonar. (b) HillValley. (c) Synthetic. (d) Arrhythmia. (e) Yale. (f) Colon. (g) SRBCT. (h) AR10P. (i) PIE10P. (j) Leukemia1. (k) Tumor9. (l) Brain1. (m) Leukemia2. (n) CNS. (o) ALLAML. (p) Nci9. (q) Pixraw10P. (r) Orlraws10P. (s) Brain2. (t) Prostate.
Mathematics 12 02572 g002
Table 1. Attributes of each classification dataset used as test problems.
Table 1. Attributes of each classification dataset used as test problems.
No.DatasetFeatureSampleClass
1Sonar602082
2HillValley1006062
3Synthetic1002005
4Arrhythmia27845213
5Yale102416515
6Colon2000622
7SRBCT2308834
8AR10P240013010
9PIE10P242021010
10Leukemia15327723
11Tumor95726609
12Brain15920905
13Leukemia27070722
14CNS7129602
15ALLAML7129722
16Nci99712609
17Pixraw10P10,00010010
18Orlraws10P10,30410010
19Brain210,367504
20Prostate10,5091022
Table 2. Mean HV performance of each algorithm on each dataset, with the best results marked in gray (the greater the better) and insignificant differences prefixed by ★.
Table 2. Mean HV performance of each algorithm on each dataset, with the best results marked in gray (the greater the better) and insignificant differences prefixed by ★.
DatasetIPEANSGA-IIMOEA/DMOEA/HDPMEA
Sonar8.072e−01★ 7.955e−01★ 7.940e−01★ 8.027e−017.740e−01
±2.36e−02±2.59e−02±2.87e−02±2.53e−02±3.09e−02
HillValley6.310e−015.849e−01★ 6.259e−01★ 6.259e−015.819e−01
±1.08e−02±2.51e−02±1.04e−02±8.54e−03±2.18e−02
Synthetic4.343e−013.846e−013.852e−013.956e−013.860e−01
±3.76e−02±3.88e−02±4.16e−02±5.93e−02±3.74e−02
Arrhythmia6.932e−016.308e−016.655e−014.826e−014.715e−01
±1.64e−02±2.66e−02±1.58e−02±1.69e−02±4.16e−02
Yale7.043e−014.878e−015.127e−014.920e−014.868e−01
±2.72e−02±1.34e−02±2.62e−02±3.39e−02±1.92e−02
Colon8.990e−015.500e−016.078e−015.239e−015.387e−01
±3.06e−02±2.65e−02±4.78e−02±3.30e−02±2.68e−02
SRBCT9.230e−012.841e−013.178e−012.554e−012.759e−01
±5.64e−02±2.05e−03±2.03e−03±1.81e−03±2.03e−03
AR10P7.778e−013.631e−013.709e−013.449e−013.536e−01
±3.52e−02±2.01e−02±2.26e−02±1.75e−02±1.48e−02
PIE10P9.347e−016.023e−016.458e−015.883e−015.902e−01
±2.07e−02±1.06e−02±1.13e−02±1.22e−02±9.87e−03
Leukemia19.647e−015.290e−015.431e−015.132e−015.160e−01
±2.43e−02±1.81e−02±3.10e−02±2.40e−02±1.93e−02
Tumor94.897e−012.790e−012.832e−012.673e−012.789e−01
±5.65e−02±2.80e−02±2.54e−02±2.35e−02±2.24e−02
Brain17.844e−014.718e−014.906e−014.535e−014.698e−01
±4.93e−02±3.11e−03±1.09e−02±1.00e−02±2.77e−03
Leukemia29.337e−015.360e−015.450e−015.126e−015.270e−01
±6.48e−02±8.95e−03±1.94e−02±1.79e−02±1.55e−02
CNS6.842e−013.781e−013.743e−013.707e−013.663e−01
±6.93e−02±3.28e−02±3.32e−02±3.25e−02±3.52e−02
ALLAML9.813e−015.205e−015.358e−015.060e−015.148e−01
±3.66e−02±1.52e−02±1.34e−02±1.44e−02±1.52e−02
Nci94.712e−012.406e−012.616e−012.370e−012.447e−01
±8.13e−02±2.54e−02±2.94e−02±2.21e−02±2.46e−02
Pixraw10P9.802e−015.795e−015.911e−015.640e−015.778e−01
±2.83e−02±2.22e−03±9.88e−03±7.64e−03±3.00e−03
Orlraws10P9.605e−015.390e−015.447e−015.297e−015.392e−01
±3.28e−02±7.53e−03±9.58e−03±8.00e−03±6.12e−03
Brain27.181e−013.903e−013.824e−013.782e−013.790e−01
±5.65e−02±2.15e−02±2.43e−02±2.12e−02±1.80e−02
Prostate9.486e−014.629e−014.599e−014.559e−014.598e−01
±2.31e−02±1.29e−02±1.51e−02±1.16e−02±2.09e−02
Table 3. Mean IGD+ performance of each algorithm on each dataset, with the best results marked in gray (the smaller the better) and insignificant differences prefixed by ★.
Table 3. Mean IGD+ performance of each algorithm on each dataset, with the best results marked in gray (the smaller the better) and insignificant differences prefixed by ★.
DatasetIPEANSGA-IIMOEA/DMOEA/HDPMEA
Sonar5.352e−02★ 6.239e−02★ 6.637e−02★ 5.961e−027.886e−02
±1.54e−02±2.31e−02±2.18e−02±1.70e−02±2.94e−02
HillValley2.504e−026.517e−02★ 2.614e−02★ 2.688e−027.049e−02
±6.73e−03±2.45e−02±9.07e−03±1.06e−02±2.43e−02
Synthetic9.448e−021.408e−011.341e−011.328e−011.402e−01
±2.35e−02±3.14e−02±3.52e−02±4.19e−02±3.06e−02
Arrhythmia1.767e−026.646e−023.616e−022.092e−012.317e−01
±7.17e−03±1.92e−02±8.31e−03±1.40e−02±4.63e−02
Yale2.232e−023.175e−012.643e−013.197e−013.166e−01
±1.19e−02±1.06e−02±1.39e−02±1.47e−02±8.91e−03
Colon6.077e−023.889e−013.241e−014.155e−014.011e−01
±2.96e−02±1.69e−02±3.84e−02±2.42e−02±2.01e−02
SRBCT3.704e−026.225e−015.751e−016.693e−016.352e−01
±3.28e−02±3.15e−03±2.55e−03±3.12e−03±3.24e−03
AR10P5.458e−024.639e−014.353e−014.810e−014.725e−01
±2.13e−02±1.59e−02±1.92e−02±1.51e−02±1.27e−02
PIE10P1.366e−023.730e−013.164e−013.867e−013.806e−01
±8.19e−03±8.95e−03±1.00e−02±9.94e−03±6.47e−03
Leukemia11.942e−024.343e−014.077e−014.502e−014.415e−01
±1.34e−02±7.64e−03±2.30e−02±1.30e−02±9.49e−03
Tumor99.052e−024.515e−014.323e−014.650e−014.516e−01
±5.05e−02±1.79e−02±1.43e−02±1.20e−02±1.36e−02
Brain16.266e−024.300e−014.036e−014.558e−014.328e−01
±3.95e−02±4.38e−03±1.53e−02±1.41e−02±3.90e−03
Leukemia27.277e−024.499e−014.349e−014.738e−014.571e−01
±7.13e−02±6.09e−03±1.71e−02±1.31e−02±1.05e−02
CNS1.806e−015.032e−014.974e−015.127e−015.165e−01
±7.62e−02±2.82e−02±3.43e−02±2.91e−02±3.02e−02
ALLAML2.056e−024.608e−014.411e−014.758e−014.636e−01
±4.03e−02±1.01e−02±9.58e−03±1.06e−02±9.55e−03
Nci99.879e−024.775e−014.513e−015.029e−014.756e−01
±6.21e−02±1.43e−02±1.49e−02±1.07e−02±1.19e−02
Pixraw10P1.172e−024.431e−014.300e−014.607e−014.450e−01
±1.88e−02±2.51e−03±1.12e−02±8.65e−03±3.40e−03
Orlraws10P1.684e−024.491e−014.386e−014.603e−014.498e−01
±1.71e−02±3.95e−03±4.68e−03±5.75e−03±3.23e−03
Brain21.001e−014.813e−014.828e−014.934e−014.888e−01
±4.86e−02±1.33e−02±1.94e−02±1.34e−02±1.11e−02
Prostate1.465e−024.890e−014.865e−014.968e−014.929e−01
±1.95e−02±1.09e−02±1.20e−02±9.49e−03±1.55e−02
Table 4. Mean minimum classification errors obtained by each algorithm on each dataset, with the best results marked in gray (the smaller the better) and insignificant differences prefixed by ★. In addition, the mean rounded number of selected features (NSF) related to the obtained minimum classification error is shown beneath as NSF.
Table 4. Mean minimum classification errors obtained by each algorithm on each dataset, with the best results marked in gray (the smaller the better) and insignificant differences prefixed by ★. In addition, the mean rounded number of selected features (NSF) related to the obtained minimum classification error is shown beneath as NSF.
DatasetIPEANSGA-IIMOEA/DMOEA/HDPMEA
Sonar1.957e−01★ 2.000e−01★ 2.065e−01★ 2.032e−01★ 2.097e−01
NSF: 6NSF: 7NSF: 6NSF: 5NSF: 7
HillValley3.984e−014.211e−01★ 4.048e−01★ 4.038e−014.218e−01
NSF: 4NSF: 10NSF: 3NSF: 4NSF: 11
Synthetic6.161e−016.439e−016.600e−016.433e−016.406e−01
NSF: 8NSF: 15NSF: 9NSF: 12NSF: 15
Arrhythmia3.317e−013.775e−013.530e−014.650e−014.566e−01
NSF: 8NSF: 17NSF: 12NSF: 51NSF: 58
Yale3.185e−013.519e−013.711e−013.504e−013.459e−01
NSF: 10NSF: 329NSF: 265NSF: 332NSF: 323
Colon1.088e−012.070e−011.947e−012.246e−012.123e−01
NSF: 3NSF: 703NSF: 552NSF: 735NSF: 720
SRBCT7.600e−026.400e−016.400e−016.400e−016.400e−01
NSF: 4NSF: 810NSF: 610NSF: 990NSF: 866
AR10P2.408e−014.958e−015.150e−015.150e−015.075e−01
NSF: 7NSF: 909NSF: 780NSF: 930NSF: 931
PIE10P7.389e−029.722e−021.044e−011.028e−011.050e−01
NSF: 11NSF: 917NSF: 767NSF: 933NSF: 927
Leukemia14.697e−021.636e−011.833e−011.712e−011.773e−01
NSF: 3NSF: 2236NSF: 2052NSF: 2308NSF: 2251
Tumor95.537e−016.019e−016.148e−016.037e−015.944e−01
NSF: 6NSF: 2455NSF: 2329NSF: 2517NSF: 2466
Brain12.370e−012.593e−012.593e−012.593e−012.593e−01
NSF: 4NSF: 2499NSF: 2321NSF: 2640NSF: 2512
Leukemia26.818e−021.318e−011.470e−011.439e−011.500e−01
NSF: 2NSF: 3056NSF: 2908NSF: 3181NSF: 3067
CNS3.537e−014.056e−014.370e−014.167e−014.352e−01
NSF: 2NSF: 3104NSF: 2956NSF: 3172NSF: 3122
ALLAML1.818e−021.576e−011.652e−011.606e−011.621e−01
NSF: 2NSF: 3081NSF: 2957NSF: 3181NSF: 3077
Nci95.860e−016.491e−016.421e−016.404e−016.526e−01
NSF: 6NSF: 4278NSF: 4092NSF: 4606NSF: 4292
Pixraw10P1.778e−023.333e−023.333e−023.333e−023.333e−02
NSF: 2NSF: 4428NSF: 4277NSF: 4628NSF: 4455
Orlraws10P3.667e−021.033e−011.100e−011.044e−011.022e−01
NSF: 7NSF: 4582NSF: 4463NSF: 4702NSF: 4590
Brain23.089e−013.778e−013.978e−013.933e−013.844e−01
NSF: 4NSF: 4637NSF: 4572NSF: 4733NSF: 4656
Prostate5.484e−022.462e−012.591e−012.452e−012.462e−01
NSF: 3NSF: 4718NSF: 4585NSF: 4797NSF: 4721
Table 5. Mean computational time cost in seconds for each algorithm run on each dataset, with the best results marked in gray (the smaller the better) and insignificant differences prefixed by ★.
Table 5. Mean computational time cost in seconds for each algorithm run on each dataset, with the best results marked in gray (the smaller the better) and insignificant differences prefixed by ★.
DatasetIPEANSGA-IIMOEA/DMOEA/HDPMEA
Sonar3.881e+013.923e+014.050e+013.944e+013.937e+01
±3.25e−01±4.31e−01±3.27e−01±4.60e−01±3.51e−01
HillValley1.455e+021.723e+021.506e+021.634e+021.758e+02
±2.17e+00±7.25e+00±1.46e+00±4.00e+00±3.90e+00
Synthetic3.990e+014.278e+014.311e+014.353e+014.257e+01
±5.41e−01±1.25e+00±6.22e−01±8.03e−01±6.76e−01
Arrhythmia1.333e+021.571e+021.516e+021.957e+021.887e+02
±1.21e+00±1.18e+01±6.88e+00±5.01e+00±7.70e+00
Yale2.251e+023.871e+023.698e+024.042e+023.789e+02
±4.75e+00±9.78e+00±9.55e+00±1.49e+01±1.05e+01
Colon8.224e+011.926e+021.666e+021.946e+021.823e+02
±3.08e+00±6.56e+00±7.77e+00±8.58e+00±7.88e+00
SRBCT2.754e+024.754e+024.368e+025.117e+024.727e+02
±3.46e+00±6.34e+00±5.39e+00±7.46e+00±5.93e+00
AR10P6.210e+021.087e+031.053e+031.119e+031.075e+03
±1.78e+01±2.46e+01±3.21e+01±3.56e+01±3.68e+01
PIE10P1.167e+032.145e+032.087e+032.309e+032.113e+03
±2.69e+01±1.28e+02±1.08e+02±1.18e+02±8.12e+01
Leukemia18.660e+021.586e+031.499e+031.614e+031.499e+03
±3.40e+01±1.12e+02±1.00e+02±1.13e+02±5.18e+01
Tumor97.800e+021.441e+031.379e+031.458e+031.338e+03
±4.19e+01±1.05e+02±1.09e+02±1.07e+02±5.66e+01
Brain11.269e+032.641e+032.542e+032.668e+032.408e+03
±5.98e+01±2.46e+02±2.36e+02±2.35e+02±4.55e+01
Leukemia21.210e+032.529e+032.402e+032.541e+032.286e+03
±5.99e+01±2.35e+02±2.17e+02±2.37e+02±4.97e+01
CNS1.022e+031.874e+031.792e+031.905e+031.769e+03
±5.68e+01±1.34e+02±1.42e+02±1.50e+02±5.26e+01
ALLAML1.241e+032.579e+032.461e+032.598e+032.306e+03
±6.69e+01±2.69e+02±2.53e+02±2.61e+02±1.78e+01
Nci91.490e+032.924e+032.832e+033.042e+032.678e+03
±8.17e+01±2.76e+02±3.03e+02±3.01e+02±2.47e+01
Pixraw10P2.721e+035.213e+035.096e+035.320e+034.774e+03
±1.87e+02±4.87e+02±4.98e+02±5.02e+02±4.74e+01
Orlraws10P2.841e+035.396e+035.272e+034.905e+034.833e+03
±1.78e+02±5.12e+02±5.59e+02±1.51e+02±1.25e+02
Brain21.438e+032.469e+032.313e+032.493e+032.438e+03
±1.05e+02±6.16e+01±5.46e+01±7.12e+01±5.03e+01
Prostate3.098e+035.129e+034.943e+035.198e+035.107e+03
±2.09e+02±1.59e+02±1.24e+02±1.78e+02±1.31e+02
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, H. An Interpolation-Based Evolutionary Algorithm for Bi-Objective Feature Selection in Classification. Mathematics 2024, 12, 2572. https://doi.org/10.3390/math12162572

AMA Style

Xu H. An Interpolation-Based Evolutionary Algorithm for Bi-Objective Feature Selection in Classification. Mathematics. 2024; 12(16):2572. https://doi.org/10.3390/math12162572

Chicago/Turabian Style

Xu, Hang. 2024. "An Interpolation-Based Evolutionary Algorithm for Bi-Objective Feature Selection in Classification" Mathematics 12, no. 16: 2572. https://doi.org/10.3390/math12162572

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop