You are currently viewing a new version of our website. To view the old version click .
Symmetry
  • Article
  • Open Access

14 October 2025

A Dynamic Surrogate-Assisted Hybrid Breeding Algorithm for High-Dimensional Imbalanced Feature Selection

,
and
School of Computer Science, Hubei University of Technology, Wuhan 430068, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Computer

Abstract

With the growing complexity of high-dimensional imbalanced datasets in critical fields such as medical diagnosis and bioinformatics, feature selection has become essential to reduce computational costs, alleviate model bias, and improve classification performance. DS-IHBO, a dynamic surrogate-assisted feature selection algorithm integrating relevance-based redundant feature filtering and an improved hybrid breeding algorithm, is presented in this paper. Departing from traditional surrogate-assisted approaches that use static approximations, DS-IHBO employs a dynamic surrogate switching mechanism capable of adapting to diverse data distributions and imbalance ratios through multiple surrogate units built via clustering. It enhances the hybrid breeding algorithm with asymmetric stratified population initialization, adaptive differential operators, and t-distribution mutation strategies to strengthen its global exploration and convergence accuracy. Tests on 12 real-world imbalanced datasets (4–98% imbalance) show that DS-IHBO achieves a 3.48% improvement in accuracy, a 4.80% improvement in F1 score, and an 83.85% reduction in computational time compared with leading methods. These results demonstrate its effectiveness for high-dimensional imbalanced feature selection and strong potential for real-world applications.

1. Introduction

With the rapid development of information technology and the advent of the big data era, various types of data now exhibit characteristics of high dimensionality, complexity, and diversity. Among these types of data, high-dimensional imbalanced data are commonly found in numerous fields, including financial fraud detection [], medical diagnosis [], bioinformatics [], and intrusion detection []. Such data possess distinct properties in terms of sample size and feature dimensionality. For instance, credit card fraud detection datasets contain significantly fewer fraudulent samples than normal transactions, and these transaction records often exhibit highly dimensional feature spaces []. Similarly, in the medical field, diagnostic data for rare diseases typically include few samples but numerous features []. These data not only fail to effectively convey class label information but also provide misleading signals for subsequent modeling.
Feature selection (FS) is a crucial dimensionality-reduction technique that has been widely adopted across various domains, such as natural language processing and computer vision [,,]. The FS workflow typically involves three main steps. First, a subset of features is selected. Second, this subset is evaluated using an evaluation function. Third, the selection is iteratively refined until a termination condition is met or a maximum number of iterations is achieved []. FS constitutes a quintessential NP-hard combinatorial problem, where the solution space grows exponentially with respect to the dimensionality of features. Some features that appear to have weak correlations with the target individually might contribute significantly to classification tasks when combined in specific ways. When existing FS methods are applied to high-dimensional imbalanced datasets, several issues arise in terms of feature evaluation metrics. Some methods have moved beyond traditional class-related indicators, instead assessing feature performance based on distributional differences across classes [,]. However, they still tend to incur FS bias, lack a dynamic trade-off mechanism, overlook the synergistic effects among features, and exhibit substantially increased computational complexity.
Evolutionary algorithms (EAs), including particle swarm optimization (PSO) [], genetic algorithm (GA) [], and other advanced variants [,,], have been widely applied across numerous disciplines. This is due to their capability of optimizing non-differentiable and non-continuous objective functions. Sadeeq et al. [] proposed Cauchy Artificial Rabbits Optimization (CARO) to improve ARO’s search ability in constrained environments and solve complex constrained optimization problems in power system engineering. Sun et al. [] developed the three-stage FSUHO approach to find optimal feature subsets in high-dimensional data and address overlooked feature–feature/feature–class interactions. Among these EAs, the hybrid breeding optimization algorithm (HBO) [] stands out due to its powerful search ability and co-evolution mechanism. Drawing on hybrid vigor theory, which implies a form of symmetric enhancement between parental traits, HBO has been successfully applied to intrusion detection [], FS [], and bandwidth selection [] problems. Nevertheless, when applied to high-dimensional imbalanced FS problems, HBO and its variants still experience slow convergence rates and limited robustness, underscoring the need for improved symmetry between model capacity and problem complexity.
Surrogate-assisted EAs have demonstrated remarkable effectiveness in computationally intensive applications, and their integration into FS provides a promising direction for reducing computational costs in high-dimensional scenarios. Brownlee et al. [] employed a Markov network as a surrogate model and adopted GA to search within this network. Chen et al. [] developed a PSO-based FS approach using correlation-guided updating and surrogate-based particle selection to enhance search efficiency. Nguyen et al. [] proposed a surrogate-assisted FS method that constructs a representative sample set from the original data to reduce computational cost. Song et al. [] introduced a cooperative feature and sample grouping strategy combined with surrogate-assisted PSO to tackle high-dimensional FS. These methods can be broadly categorized into two groups according to their surrogate model construction strategies: the first two focus on building surrogate models for the objective function (i.e., fitness), which are referred to as objective regression surrogates; the latter two utilize a small number of representative samples as a substitute for the entire dataset, and are thus termed sample surrogates. In turn, these findings provide solid empirical and methodological support for the successful application of surrogate models in FS. However, most surrogate models still lack the capacity to capture high-dimensional nonlinear feature interactions, especially the weak signals from minority classes, which consequently degrade the classification performance of the resulting feature subsets.
To address these drawbacks, this paper investigates an improved HBO (IHBO) algorithm based on filtering and dynamic surrogate-assisted strategies, called DS-IHBO, for high-dimensional imbalanced FS in classification problems. The main contributions of this paper are as follows:
  • IHBO, a binary variant of HBO, is developed through the incorporation of hierarchical population initialization, t-distribution mutation disturbance, and an adaptive difference operator selection strategy into HBO to enhance its performance. These improvements enable the algorithm to escape the local optimum and improve exploration in high-dimensional spaces, leading to a significant boost in its global search and optimization performance.
  • DS-IHBO, a hybrid FS method, is proposed. In this method, irrelevant features are removed based on an importance index; IHBO is utilized for global optimization, and a fast dynamic surrogate-assisted strategy is incorporated. The original dataset is used to evaluate the fitness of the global optimal solution in each iteration, while the surrogate unit is dynamically switched to assess other solutions. This design makes it particularly well-suited for real-world applications such as disease detection in medical scenes.
  • The proposed methods are validated on twelve datasets. The experimental results demonstrate that the proposed approach significantly outperforms six state-of-the-art techniques in key classification performance metrics and computational cost, showcasing its efficacy in handling high-dimensional, imbalanced, and complex data.
The structure of this paper is organized as follows. Section 2 presents the related work and foundational problem definitions. Section 3 details the proposed methodology. Experimental validation of the approach is provided in Section 4, while Section 5 concludes with a conclusion and discussion.

3. Proposed Approach

This section first presents the workflow of DS-IHBO, followed by a detailed explanation of the individual modules that comprise DS-IHBO.

3.1. The Workflow of DS-IHBO

The DS-IHBO FS algorithm consists of three interconnected modules: First, Module I eliminates irrelevant features to form a streamlined feature subset. Second, taking the feature subset derived from Module I as the foundational search space, Module II (the improved HBO algorithm) performs evolutionary search. During this process, Module II and Module III (the dynamic surrogate-assisted mechanism) function in parallel synergy: Module III initially constructs multiple distinct surrogate units, and subsequently dynamically selects the most appropriate surrogate unit in real time during iterations to evaluate individual updates within Module II. After a specified number of iterations, Module II switches to evaluating all individuals using the original dataset. This process continues until the iteration terminates, ultimately outputting the feature subset corresponding to the globally optimal individual. Figure 2 shows its basic workflow.
Figure 2. The workflow of DS-IHBO.

3.2. Module I: Removing Irrelevant Features

Building upon the SU measure, the interdependence U i j between features F i and F j can be quantified using Equation (5). However, solely considering feature–class relationships may overlook inter-feature redundancies. Inspired by max-relevance and min-redundancy (mRMR) [], a relevance-redundancy index ( ψ i ) is proposed to assess feature importance. The computation involves: (1) selecting the target feature F i ; (2) establishing the remaining feature subset L; and (3) calculating ψ i for F i via Equation (6).
ψ i = U ( x i , c ) 1 m 1 x i L U ( x i , x j )
where m 1 denotes the number of features in L. This produces a vector encoding each feature’s class-relevance while incorporating redundancy considerations. The min-max normalization is used to normalize the relevance-redundancy index to [ 0 , 1 ] , as shown in Equation (7).
ψ i = ψ i min ( Ψ ) max ( Ψ ) min ( Ψ ) , Ψ = { ψ 1 , ψ 2 , , ψ m } .
For a dataset S with D features F = { f 1 , f 2 , , f D } , the value ψ i for each feature f i (where i = 1 , 2 , , D ) is first calculated. Features with normalized relevance-redundancy indices ψ i above the specified threshold ρ 0 are identified as strongly relevant features and aggregated into a subset F . The threshold ρ 0 is crucial in determining these strongly relevance features. Traditionally, ρ 0 is set to a very small constant. However, in some datasets, all feature relevance values are relatively high. Taking the Yale_64 dataset as an instance, where typical relevance values approximate 0.5, the threshold ρ 0 is mathematically established following Song et al. [] in Equation (8).
ρ 0 = min 0.1 · ψ max , ψ D / log D th
where · represents the rounding down function. Two constraints are designed in Equation (8) to balance “avoiding missing effective features” and “removing irrelevant features”:
Constraint 1:  ψ max is the maximum value of the normalized relevance-redundancy index (ranging from 0 to 1) in the dataset. Multiplying by 0.1 ensures that ρ 0 does not exceed 10% of the most relevant feature’s index. This avoids setting an excessively high threshold (e.g., in datasets where most ψ i are low, a fixed threshold like 0.2 might filter out all valid features). For example, if  ψ max = 0.5 , this term limits ρ 0 to at most 0.05, preventing over-filtering.
Constraint 2:  D / log D represents the “recommended number of candidate features” derived from the dataset dimension D (a common empirical rule in high-dimensional FS to balance computational efficiency and information retention). ψ D / log D th is the ψ i value of the feature ranked at this “recommended number” (sorted in descending order of ψ i ). This constraint ensures that ρ 0 is aligned with the dataset’s dimension, avoiding excessively low thresholds (e.g., in high-dimensional datasets with D = 1000 , 1000 / log 1000 144 , so ρ 0 is at least the ψ i of the 144th feature, preventing retaining too many redundant features).
By taking the minimum of the two constraints, ρ 0 automatically adapts to the dataset’s relevance level and dimension, ensuring it is neither too high (missing valid features) nor too low (retaining noise).

3.3. Module II: The Improved HBO Algorithm

The steps of the IHBO algorithm are as follows:
(i) Population initialization with asymmetric feature stratification: Traditional random initialization often exhibits strong distributional symmetry, leading to populations concentrated in limited regions of the search space and consequently reducing population diversity. To break this symmetry and introduce controlled asymmetry, a novel initialization mechanism based on the relevance-redundancy index is proposed, as detailed in Algorithm 1.
Specifically, features are first divided into three groups reflecting asymmetric relevance levels: the top 10% as high-relevance, the next 10–50% as mid-relevance, and the remaining 50% as low-relevance. This stratified division intentionally introduces structural asymmetry to prioritize features with high discriminative power. For each individual, the number of features to be initialized, denoted by μ , is randomly drawn from the interval [ 0 , 0.3 D ] , where D is the total feature dimension. All high-relevance features are symmetrically included to ensure essential information retention, while mid- and low-relevance features are asymmetrically sampled: θ · μ features from the mid-relevance group and ( 1 θ ) · μ from the low-relevance group, where θ follows a uniform distribution over [ 0.7 , 1 ] . This asymmetric sampling strategy effectively balances diversity and relevance, reduces redundancy, and promotes symmetric compactness within the selected feature subsets. The process repeats until a population of size N P is constructed.
Algorithm 1 Population initialization strategy
  1:
Input:  N P : population size; D: feature dimension
  2:
Output: P: initial population
  3:
for  i = 1 , 2 , , D  do
  4:
   Calculate relevance-redundancy index (i)
  5:
end for
  6:
Sort features according index (i)
  7:
High-relevance in the top 10%; mid-relevance in the top 10–50%; low-relevance in the bottom 50%
  8:
for  n = 1 , 2 , , N P  do
  9:
    μ Randomly generate in [ 0 , 0.3 · D ]
10:
   Select all high-relevance features
11:
    θ Randomly generate in [ 0.7 , 1 ]
12:
   Randomly select θ · μ mid-relevance features and ( 1 θ ) · μ low-relevance features
13:
   for  i = 1 , 2 , , D  do
14:
     if i is selected then
15:
         F ( i ) = 1
16:
     end if
17:
   end for
18:
end for
19:
return initial population P
(ii) An update strategy for the maintainer line: In HBO, the maintainer line represents a group of individuals with optimal fitness values. It is used to cross with the sterile line to generate improved sterile individuals with superior traits, thereby guiding the direction of population updates. However, the original HBO does not establish an effective strategy for updating the maintainer line. To address this limitation, this paper proposes an improvement to the maintainer line’s update based on a dynamic selection strategy of adaptive differential operators. Given that differential operators have multiple variants, and each is tailored to distinct types of search tasks, the objective is to integrate the advantages of these diverse differential variants. Accordingly, three improved differential operator variants are proposed: a global differential operator for the global search phase, a transitional one for the transition from global to local search, and a local one for the local search phase. These three differential operators are denoted as D E 1 , D E 2 , and  D E 3 , respectively, as shown in Equations (9)–(11).
D E 1 : x i j ( t ) + r 1 · x p 1 j ( t ) x p 2 j ( t ) + ( 1 r 1 ) · x p 3 j ( t ) x i j ( t )
where r 1 and r 2 are random values in the range of 0 to 1, p m represents randomly selected individuals from the population. They satisfy conditions ( p m i ) and ( p i p j ,   i j ,   1 i ,   j 4 ) . The term r 1 · ( x p 1 j ( t ) x p 2 j ( t ) ) creates a perturbation vector based on the difference between two randomly chosen individuals. This encourages the search to span wide areas of the solution space. The second term, ( 1 r 1 ) · ( x p 3 j ( t ) x i j ( t ) ) , adds another layer of random perturbation relative to the current individual’s position, further enhancing diversity and preventing premature convergence. The strength of this operator lies in its ability to discover new and potentially promising regions, which is critical in the early stages.
D E 2 : x i j ( t ) + F · x p 1 j ( t ) x p 2 j ( t ) + F · x best j ( t ) x i j ( t )
where x best j ( t ) represents the value of the j-th dimension of the best solution in the current generation. D E 2 incorporates both a random perturbation term ( x p 1 j ( t ) x p 2 j ( t ) ) for exploration and a term that guides the search toward the best solution ( x best j ( t ) x i j ( t ) ) for exploitation. It provides a smooth transition from a global, explorative search to a local, exploitative one.
D E 3 : x best j ( t ) + r 2 · x p 1 j ( t ) x p 2 j ( t ) + ( 1 r 2 ) · x p 3 j ( t ) x p 4 j ( t )
where F is the scaling factor that controls the transition of the population from global search to local search. The search is anchored to x best j ( t ) , the best solution found so far. The subsequent perturbation terms are relatively small variations created from the differences in random individuals. This structure ensures that the search is concentrated in the promising region around the current best solution, allowing for fine-tuning and convergence towards a refined optimum. This is essential in the later stages of the algorithm.
To enable the algorithm to use different selection probabilities for differential strategies in different optimization stages, the probability selection strategies are defined as follows. In the early stage of optimization, a higher probability is assigned to the exploratory strategy D E 1 . In the middle stage, the probability gradually shifts to the balancing strategy D E 2 . In the later stage of optimization, a higher probability is assigned to the exploitative strategy D E 3 to refine the solutions. This paper introduces an adaptive probability generation mechanism to dynamically produce the selection probabilities of the three differential strategies, using the roulette_wheel_selection  ( p i ) algorithm to determine D E i . As shown in Equations (12)–(17).
s 1 = p e a k 1 + exp ( t T / 6 ) T / 25 + s min
s 2 = p e a k · exp ( t T / 2 ) 2 10 · T + s min
s 3 = p e a k 1 + exp ( t 5 · T / 6 ) T / 25 + s min
p e a k = s max s min
S = s 1 + s 2 + s 3
p i = s i S , i = 1 , 2 , 3
where s min and s max are boundary correction parameters to prevent values before normalization from being too small or too large. p i ( i = 1 , 2 , 3 ) represents the normalized probability that the i-th difference operator is selected, t and T represent the current iteration number and the maximum iteration number, respectively. Figure 3 demonstrates the trend of adaptive adjustment for the selection probabilities of the three differential strategies as the number of iterations increases.
Figure 3. The probability transfer curve of differential operators.
As shown in Figure 3, in the early stage of evolution, the algorithm focuses on exploring the global search space to find more promising regions. Therefore, the probability of selecting the global exploration strategy D E 1 is higher. As evolution progresses, the optimization process gradually transitions from global search to local search. To smoothly transition between these two stages, the algorithm gradually decreases the probability of selecting the global exploration strategy D E 1 and increases the probability of selecting the balancing strategy D E 2 . In the later stage of evolution, the selection gradually shifts to the local exploitation strategy D E 3 to enhance the search in the vicinity of the optimal solution.
(iii) Modifications to the hybridization and selfing phases: The objective of the hybridization phase is to update the sterile line. However, the original update strategy does not fully consider the difference between the exploration phase in the early stage and the exploitation phase in the later stage. Therefore, this paper introduces a t-distribution mutation disturbance strategy to improve the hybridization phase. The t-distribution, also known as the Student’s t-distribution, is characterized by its degree of freedom ( d f ) and mutation scale ( σ ). The degree of freedom controls the shape of the distribution, mainly affecting the thickness of the tails, while σ determines the width of the distribution. Based on the characteristics of t-distribution, this paper improves the hybridization phase as shown in Equations (18)–(21).
X new ( i ) d ( t ) = t · X s d ( t ) + ( 1 t ) · X m d ( t )
f ( t d f , σ ) = Γ d f + 1 2 Γ d f 2 d f π σ 1 + 1 d f t σ 2 d f + 1 2
d f = 2 + t T · 28
σ = ( σ max σ min ) · 1 t T g r + σ min
The random number r of the original hybridization stage is replaced by random sampling of t-distribution. X new ( i ) d ( t ) represents a new sterile individual generated by hybridization, where each dimension is disturbed by the random variables generated by t-distribution. Γ represents the Gamma function. σ max and σ min denote the maximum and minimum values of the mutation scale, which control the range of generated random numbers. The larger the mutation scale, the larger the range of generated random numbers, and vice versa. g r is the growth rate, which controls the curvature of the σ change curve.
Figure 4 shows the changes in the shape of the t-distribution under different degrees of freedom and mutation scales. When the degree of freedom is one, the t-distribution becomes a Cauchy distribution with heavy tails. As the degree of freedom increases, the heavy tails lessen, gradually approaching a normal distribution. A larger mutation scale results in a wider t-distribution curve, indicating more dispersed data points, while a smaller mutation scale narrows the distribution, concentrating the data points. Therefore, this paper leverages the characteristics of the t-distribution by combining different degrees of freedom and mutation scales to control its shape, thereby managing the global and local update strategies of individuals. Specifically, in the initial iteration stages, smaller degrees of freedom and larger mutation scales make the t-distribution similar to the Cauchy distribution with dispersed data, generating larger perturbations that favor global search. As iterations progress, increasing degrees of freedom and decreasing mutation scales cause the t-distribution to approach a standard normal distribution with smaller variance, resulting in smaller perturbations and a preference for local search.
Figure 4. The shape of t-distribution corresponding to different degrees of freedom and mutation scales.
The selfing phase focuses on updating the restorer line. When a restorer individual reaches the maximum number of selfings, it is considered to have fallen into a local optimum. At this point, the renewal operation is triggered for the individual, which plays a decisive role in the transition between selfing and renewal. Instead of setting a constant value in the original HBO algorithm, this paper introduces an enhanced selfing upper bound, as defined in Equation (22).
S C = S C min + S C max S C min · 1 t T 2
where S C min and S C max represent the minimum and maximum bounds for selfing count, respectively. In the early iterations, the algorithm is in the global search phase, and most individuals could perform effective updates within a few cycles. Therefore, S C is set to a large value. If an individual reaches this bound early, it indicates that it has fallen into a local optimum, which will trigger the renewal operation. As optimization progresses to later iterations, the possibility of an individual falling into a local optimum increases. Therefore, setting S C to a smaller value in later stages helps an individual escape the local optimum more quickly.
Collectively, these three improvements are specifically designed to address the core challenges of high-dimensional FS. The asymmetric population initialization addresses the massive search space and feature redundancy: a relevance-redundancy index guides initial searches toward promising feature combinations, effectively pruning unpromising directions from the outset. The adaptive update strategy for the maintainer line delivers the required dynamic balance between exploration and exploitation. Broad exploration in the early stage is critical for navigating the vast feature space, while focused exploitation in the later stage refines the optimal feature subset. Modifications to the hybridization and selfing phases act as mechanisms to preserve population diversity and avoid local optima. The adaptive shape of the t-distribution enables large perturbations for global exploration when required, and the dynamic selfing count ensures the renewal of stagnated individuals, preventing premature convergence.
To analyze the impact of the proposed improved strategy on algorithm time complexity, a comprehensive comparison between HBO and IHBO is conducted at each stage. With population size N, problem dimension D, and maximum iteration number T assumed, the results of the comparative analysis are shown in Table 3.
Table 3. Time complexity comparison.
Based on the analysis, IHBO’s time complexity in the population initialization stage is approximately D times that of HBO. However, since population initialization occurs only once, its impact on overall time complexity is relatively minor. In the population update stage, IHBO adds only the maintainer line update process compared with HBO, which increases time complexity by approximately 1/3 but significantly enhances the utilization rate of superior individuals and population diversity. Regarding population sorting and best solution update, both algorithms exhibit identical time complexity. Overall, IHBO marginally increases computational overhead only during population initialization and maintainer line update stages. When HBO’s population size is set to approximately 1.5 times that of IHBO, comparable total time complexity can be achieved.

3.4. Module III: Dynamic Surrogate-Assisted Mechanism

(i) Generation of surrogate units by density-adaptive symmetric sampling: To effectively handle imbalanced data distributions while maintaining representative symmetry between majority and minority classes, an adaptive K-Nearest Neighbors (KNN) algorithm is introduced. This method dynamically adjusts the number of neighbors K based on local density to discriminate between class-central and class-boundary samples. For a given sample s i , the local density D i is computed as the average distance to its k nearest neighbors, as formalized in Equation (23).
D i = 1 k j = 1 k d ( s i , s j )
Based on D i , the number of neighbors K i is determined by a threshold D threshold , typically the median of all D i in Equation (24).
K i = K max , if D i D threshold K min , if D i > D threshold
Subsequently, the algorithm determines whether the K i nearest neighbors of a given sample belong to the same class; if so, the sample is classified as central, otherwise as boundary, as illustrated in Figure 5.
Figure 5. Boundary division based on adaptive KNN. The dashed lines represent the boundary of the class, and different colors represent different classes.
In imbalanced datasets, the average number of samples per class, denoted as N ¯ , is first calculated. Classes with more than N ¯ samples are defined as majority classes, while those with N ¯ or fewer are identified as minority classes. The adaptive KNN algorithm is then applied to distinguish between central samples and boundary samples.
Since majority classes contain substantially more samples than minority ones, using all majority class samples for training would significantly increase computational cost. To mitigate redundancy, all boundary samples are retained in their entirety, as they contain critical information for classification models, while 40% of central samples are randomly sampled, given that they are typically redundant.
Although minority classes have fewer samples, they are crucial for the model. To prevent their influence from being overshadowed by majority classes, the average number of samples N maj across all majority classes is first calculated. The SMOTE [] oversampling technique is then applied to expand minority classes to N maj . All boundary samples are retained, and 40% of the central samples are randomly selected. Representative samples from both majority and minority classes are combined to form the final set of representative samples. Based on these samples, multiple surrogate units are constructed.
The total number of samples in each dataset is denoted as N. To balance accuracy and computational cost, six surrogate units with different sample sizes are constructed as follows:
Step 1.
Calculate the target sample sizes for the six surrogate units based on preset proportions of the original dataset size N: 0.75N, 0.65N, 0.55N, 0.45N, 0.35N, and 0.25N;
Step 2.
Based on the representative sample set, calculate the sample size for each category and determine the proportion of each category: category sample size/total sample size. Set this as the target category proportion to ensure that subsequent surrogate unit construction selects samples from each category according to the same proportion;
Step 3.
Apply agglomerative clustering algorithm to partition the representative sample set into six clusters;
Step 4.
Each surrogate unit selects a certain number of target samples evenly from the six clusters. For example, in the first surrogate unit, the algorithm needs to select 0.75N/6 samples from each cluster;
Step 5.
Traverse the six clusters. Within each cluster, select samples for each category based on the required number of category samples (target category proportion × target sample number from the previous step). If the cluster contains no samples of a particular category, skip that category. If the number of samples for a certain category within the cluster is less than or equal to the required category sample number, select all samples; otherwise, calculate the centroid of that category’s samples in the current cluster and prioritize selecting the required number of samples that are farthest from the centroid (i.e., most representative);
Step 6.
After completing sample selection for each cluster, check whether the proportion of each category in the surrogate unit conforms to the target category proportion. For categories that do not conform, perform global random supplementation from the remaining unselected samples until reaching the target category proportion (i.e., achieving the target sample number for each surrogate unit).
(ii) Dynamic symmetry-preserving surrogate assistance: The surrogate unit selection process begins with population initialization, where the original dataset S serves as the evaluation benchmark. The global optimal individual x g b e s t , possessing the optimal ground truth fitness f 0 , is first identified. Subsequently, l surrogate units generate corresponding fitness values { f 1 , f 2 , , f l } for x g b e s t . By comparing the prediction errors | f i f 0 | , the most accurate surrogate unit (with minimal deviation) is selected for feature subset evaluation in subsequent iterations.
During the initial T s iterations, the feature subsets are evaluated using the surrogate unit, which is periodically updated to ensure alignment with the true dataset distribution. However, overly frequent updates may impair the population’s ability to adapt to evolving search environments. Consequently, the surrogate unit is only refreshed when the global optimum individual x g b e s t t shows no improvement in its real fitness after T n consecutive iterations. The surrogate selection mechanism follows the initialization approach described above, with the current x g b e s t t replacing the previous x g b e s t as the reference solution. Upon completing the surrogate phase (i.e., after  T s iterations), candidate solutions are evaluated against the original dataset. Our investigation of six T s values (0, 25, 50, 75, and 100) revealed 75 as the optimal setting. The complete dynamic surrogate-assisted IHBO method is summarized in Algorithm 2.
Algorithm 2 Dynamic surrogate combining with IHBO
  1:
Input: Surrogate units: S 1 , S 2 , S 3 , S 4 , S 5 , S 6 .
  2:
Output: A feature subset.
  3:
Initialize IHBO Population, find x g b e s t using the original dataset.
  4:
Select the most appropriate surrogate unit based on x g b e s t .
  5:
while T m a x is not reached do
  6:
   if  T T s  then
  7:
     Maintainer-update();
  8:
     Hybridization();
  9:
     Selfing();
10:
     Evaluate individuals using surrogate unit;
11:
     Update x g b e s t , evaluate x g b e s t using original dataset.
12:
     if  x g b e s t ’s real fitness is not improved for T u iterations then
13:
        Select the most appropriate surrogate unit based on x g b e s t .
14:
     end if
15:
   else
16:
     Maintainer-update();
17:
     Hybridization();
18:
     Selfing();
19:
     Evaluate individuals using original dataset;
20:
     Update x g b e s t .
21:
   end if
22:
    T = T + 1
23:
end while
24:
return  x g b e s t as the final feature subset.
The computational complexity of the proposed algorithm mainly includes the following aspects: (1) The complexity of filtering out irrelevant features in Module I is O ( D · S ) ; (2) Module II and Module III are executed in parallel. The complexity of implementing IHBO is O ( T · ( N + N · D + 1 ) + m a x ( D · log D , N · D ) ) , and thus, the complexity after adopting the dynamic surrogate mechanism is approximately O ( T s · N · D + ( T T s ) · N · S · D ) . In the above process, D denotes the number of features in the original data, S denotes the number of samples in the original data, D denotes the number of reduced features after the first stage, T denotes the number of iterations, T s denotes the number of surrogate iterations, and N denotes the population size.

4. Experiments

This section first presents ablation experiments on Module I and Module III to verify the effectiveness of their core designs. Second, it introduces the fitness function and datasets, and details five state-of-the-art comparative methods along with their parameter configurations. It then presents comprehensive experimental comparisons between DS-IHBO and the five FS methods on 12 publicly available datasets, using five evaluation metrics. Finally, it includes sensitivity analysis on the key parameters of DS-IHBO in the surrogate phase, including ablation verification of Module III.

4.1. Datasets

To demonstrate the advantages of the proposed algorithm, comprehensive comparisons are conducted with six established methods across 12 datasets—all of which are open-source and sourced from http://archive.ics.uci.edu/ml (accessed on 4 September 2025), https://jundongl.github.io/scikit-feature/datasets.html (accessed on 4 September 2025) and https://ckzixf.github.io/dataset.html (accessed on 4 September 2025). These datasets encompass key domains, including medical diagnosis (DLBCL, Parkinson, GLIOMA, Lung2, BrainTumor1, Lymphoma, 11_Tumor), bioinformatics (Musk2), computer vision (USPS), text mining (RELATHE), and industrial and Web (PCMAC, InternetAd). The datasets highlight the following characteristics: (1) inclusion of both binary classification and multi-classification; (2) large sample sizes: The USPS dataset contains the highest number of samples, with a total of 9298 instances; (3) high-dimensionality: The 11_Tumor dataset includes 12,533 features; (4) class imbalance: all datasets are imbalanced, with the Lung2 dataset exhibiting the most significant imbalance ratio of 4.32%. Table 4 presents the essential information of the datasets used in this study, where Max indicates the number of majority class samples, Min indicates the number of minority class samples, and the imbalance rate indicates the ratio of Min to Max.
Table 4. Essential information of the 12 datasets.
To ensure the quality and consistency of experimental data, data preprocessing is performed for each dataset. This process included the following steps: missing values are imputed using the mean of the corresponding feature columns; categorical features are converted to integer format via one-hot encoding; and all feature values are scaled to a unified standard range using min-max normalization.

4.2. Ablation Experiments for Module I

This section analyzes the algorithm without the filter module (denoted as DS-IHBO/Module I). The accuracy (Acc) and running time of this algorithm are compared with those of the proposed DS-IHBO in Table 5. On the high-dimensional USPS dataset, DS-IHBO dynamically adapts threshold ρ 0 to eliminate redundant features while retaining classification-critical core features in the filtering stage. This slightly improves accuracy from 96.21% to 96.57% and reduces running time to 6872.09 s (a decrease of approximately 34%), further validating the filtering stage’s necessity. In contrast, DS-IHBO/Module I exhibits limitations. Redundant noise in full features interferes with classifier decisions, leading to consistently lower Acc than DS-IHBO. Additionally, the high-dimensional space increases IHBO’s evolutionary burden: more feature combinations require fitness evaluation per iteration, and the search tends to fall into redundant-feature-dominated local optima, ultimately reducing efficiency.
Table 5. Acc and running time (unit: s) obtained by DS-IHBO/Module I and DS-IHBO.

4.3. Ablation Experiments for Module II

The IHBO algorithm integrates four key strategies: stratified population initialization (HBO_IN), adaptive difference operator-based maintainer line update (HBO_MA), t-distribution-driven mutation for hybridization (HBO_HY), and enhanced selfing upper bound for selfing phase (HBO_SE). Each addresses specific algorithmic weaknesses, and their combination aims to boost overall performance.
For fair comparison and to validate the effectiveness of each strategy, 30 independent runs of each algorithm are performed on every test function (across dimensions) to reduce the impact of randomness. Maximum iterations are uniformly set to 1000. For consistent fitness evaluations, IHBO and HBO_MA used a population size of 40, while the original HBO and the other three variants used 60. Performance is evaluated using five metrics: best/worst/average fitness (best, worst, avg), standard deviation (std), and average runtime (time). Table 6, Table 7, Table 8 and Table 9 present the results for all algorithms on CEC2022 benchmarks (10D and 20D, respectively).
Table 6. Performance comparison of IHBO and other variants on CEC2022 (Dim = 10) (F1–F6).
Table 7. Performance comparison of IHBO and other variants on CEC2022 (Dim = 10) (F7–F12).
Table 8. Performance comparison of IHBO and other variants on CEC2022 (Dim = 20) (F1–F6).
Table 9. Performance comparison of IHBO and other variants on CEC2022 (Dim = 20) (F7–F12).
Table 6, Table 7, Table 8 and Table 9 show IHBO outperforms others on unimodal function F1 (10D/20D), confirming multi-strategy integration enhances local search. It achieved optimal F1 results across best/worst/avg metrics, with high stability (std: 1.32 × 10−13 for 10D, 3.40 × 10−11 for 20D) and improved consistency at higher dimensions. For multimodal functions (F2–F5), IHBO had the best avg fitness on F3/F5 (10D) and F2/F3/F4 (20D). Notably, it had a std of 0 for F5 (10D) and F3 (20D), proving its ability to escape local optima via balanced exploration/exploitation. IHBO also outperformed competitors on F2/F3/F5 (both dimensions) in worst fitness, except for F4 (10D/20D). HBO_IN’s stratified initialization boosted population diversity, seen in its strong F5 performance (both dimensions). For hybrid functions (F6–F8), HBO_MA performed best on F6 (10D/20D). While IHBO lagged slightly (e.g., 1.84 × 103 vs. 1.81 × 103 for F6 10D), it remained competitive thanks to adaptive differential operator selection. On composite functions (F9–F12), IHBO delivered the best avg fitness for F9 (10D), F11 (10D/20D), and F12 (10D/20D), showing strong global optimization in complex high-dimensional tasks. It also had the best worst fitness for F9/F12 (20D). Though HBO_HY is less stable than original HBO at 10D, it improved at 20D (minimal std across F9–F12). IHBO maintained the lowest overall std, highlighting multi-strategy integration’s value.
The performance advantages of IHBO over other variants are statistically validated by the Friedman test results in Table 10. In both 10-dimensional and 20-dimensional experimental scenarios, IHBO ranks first and significantly outperforms other variants. In terms of statistical significance, the p-value is 0.0001 for the 10-dimensional scenario and 3.2 × 10−5 for the 20-dimensional scenario. Both values are far less than the 0.05 significance level, indicating that the performance differences among the algorithms are not random or accidental but exhibit extremely strong statistical reliability.
Table 10. Friedman test results obtained by five variants (Dim = 10/20).
For computational efficiency, HBO_HY had the shortest runtime in most cases, due to t-distribution perturbation accelerating convergence. HBO_IN matched original HBO’s runtime while outperforming it on F4/F6/F8/F10, thanks to high-quality initialization. Notably, IHBO had no significant overhead vs. original HBO (consistent with Section 3.3’s time complexity analysis), and runtime gaps narrowed further at higher dimensions (e.g., F7/F8 20D).

4.4. Fitness Function

In imbalanced high-dimensional FS tasks, algorithms are expected to identify minimal feature subsets that achieve maximal F1 scores. Consequently, the F1 score and size of feature subsets constitute the primary optimization objectives. This paper employs the fitness function shown in Equation (25).
F i t n e s s = λ · e r r o r + ( 1 λ ) · d D
where e r r o r represents F1 loss (i.e., the complement of the F1 score), d and D denote the size of the selected feature subset and the total number of features, respectively. λ serves as a weighting factor balancing the influence of F1 loss and feature subset size, typically set to 0.99.

4.5. Competitive Methods and Parameter Settings

This paper benchmarks DS-IHBO against five advanced FS methods, with several implementing surrogate-assisted methods, methods suitable for high-dimensional and unbalanced data FS, and FS methods guided by feature correlation as follows:
  • A hybrid artificial immune optimization for high-dimensional FS (HFSIA) [];
  • Surrogate sample-assisted PSO for FS on high-dimensional data (SS-PSO) [];
  • A Surrogate-Assisted Multi-Phase Ensemble FS Algorithm With Particle Swarm Optimization in Imbalanced Data (SMEFS-PSO) [];
  • IMOABC: An efficient multi-objective filter–wrapper hybrid approach for high-dimensional FS (IMOABC) [];
  • Correlation-guided updating strategy for FS in classification with surrogate-assisted PSO (CUS-SPSO) [];
All computational methods discussed in this study are implemented in Python 3.7. Performance metrics are acquired on a desktop system equipped with an Intel(R) Core(TM) i5-12400F processor (2.5GHz) with 32GB RAM. The configuration parameters for the benchmark algorithms followed the recommendations of their original publications, as detailed in Table 11. Reported data represent mean values from 30 independent executions of each algorithm. To prevent FS bias, the dataset is partitioned into two independent subsets prior to FS: 70% as the training set and 30% as the test set. All algorithms are executed on identical training sets, and the optimal feature subsets obtained are evaluated in the same test set. Within the training set, five-fold cross-validation is employed to compute the fitness value of candidate solutions. The test set is kept completely isolated from the FS process.
Table 11. Parameter settings.
As both our proposed technique and the five compared evolutionary FS approaches adopt the wrapper methodology, individual fitness evaluation requires classifier involvement. The support vector machine (SVM) is adopted to assess the fitness of feature subsets, as it has been extensively employed in numerous studies []. The parameter C of the SVM is set to 4. To maintain temporal comparison equity, identical classification models are employed across all methods.

4.6. Evaluation Metrics

Classification accuracy has been widely used to evaluate the performance of FS algorithms. However, accuracy alone does not effectively assess an algorithm’s ability to handle imbalanced data. Therefore, this paper adds F1 score and G-mean for the classification task. The F1 score represents a balanced metric combining precision and recall through their harmonic mean, serving as a robust indicator of classification performance with special emphasis on minority class identification. The G-mean assesses the equilibrium between sensitivity across different classes, establishing itself as an essential evaluation criterion for datasets exhibiting class imbalance. Since G-mean lacks a unified standard definition for multi-class problems, in which inconsistent implementations may lead to divergent results, it is exclusively employed for evaluating binary classification datasets in this study.

4.7. Result Comparison and Analyses

The experimental evaluation assesses DS-IHBO against five competing algorithms on 12 datasets using five performance metrics: classification accuracy ( A c c ), F1 score, G-mean value, selected feature count ( d ), and computational time ( t ). Additionally, a t-test with a confidence level of 0.05 is employed to assess the significance of differences between the two algorithms. In the results, “+” indicates that the performance of DS-IHBO is significantly superior to that of the comparative algorithm; “−” indicates that the performance of the selected comparative algorithm is significantly superior to DS-IHBO; while “=” indicates no significant difference between them.
(1) Acc, F1 and G-mean values: As evidenced in Table 12, DS-IHBO achieves superior accuracy compared with the five benchmark algorithms on 9 out of 12 datasets. For the remaining three datasets, Musk2, BrainTumor1, and USPS, DS-IHBO exhibits minimal performance gaps of approximately 0.49%, 0.98%, and 0.18%, respectively, relative to the top-performing methods. Furthermore, t-test results demonstrate that (1) DS-IHBO significantly outperforms HFSIA across all datasets, potentially attributable to HFSIA’s sensitivity to dataset-specific parameter tuning; suboptimal parameter selection in its mutation strategy may compromise its effectiveness; (2) DS-IHBO delivers significantly higher accuracy than SS-PSO on all datasets except Musk2 and USPS. This advantage may stem from feature information loss caused by SS-PSO’s suboptimal clustering strategy and compounded by inappropriate surrogate model selection, propagating errors to final classification outcomes; (3) with the exception of Musk2, DS-IHBO significantly surpasses IMOABC in accuracy across all datasets. This discrepancy may arise from IMOABC’s reliance on Fisher Score, which assumes feature–target correlations reliably predict classification importance, an assumption potentially invalidated in complex datasets; (4) DS-IHBO achieves statistically significant superiority over CUS-SPSO on all datasets except Musk2. A key limiting factor for CUS-SPSO appears to be its use of Euclidean distance for fitness prediction in discrete-variable optimization problems, which can substantially impair prediction precision.
Table 12. Average Acc and F1 values obtained by six algorithms (std).
As evidenced by Table 12: (1) DS-IHBO achieves the highest F1 scores on all datasets except Lung2, USPS, and 11_Tumor. Notably, it significantly outperforms all five competing methods on the PCMAC, RELATHE, Parkinson, and Lymphoma datasets; (2) effective classification necessitates precise delineation of inter-class boundaries. Algorithms failing to construct robust decision boundaries incur higher misclassification rates. For instance, in the Lymphoma dataset, DS-IHBO’s integration of class-imbalance handling mechanisms within its surrogate model enables more adaptive boundary construction aligned with dataset characteristics; (3) comparative analysis reveals greater performance volatility among algorithms on datasets with higher imbalance ratios (e.g., BrainTumor1), where diverging F1 scores indicate amplified classification challenges. Critically, DS-IHBO consistently achieves the highest F1 score with the smallest standard deviation under such conditions, demonstrating exceptional stability and imbalance robustness. Figure 6 shows the visual comparison of six methods on accuracy and F1 score.
Figure 6. Comparison between DS-IHBO and other algorithms in average accuracy and F1 score visualization.
The Friedman test at the 0.05 significance level and the Finner test with an α value of 0.05 are employed to validate the performance of the proposed algorithm. Table 13 and Table 14 present the test results for the six algorithms, where DS-IHBO is selected as the reference method. In the tables, a smaller p-value from the Friedman test indicates higher confidence in the superiority of an algorithm over the others, while a smaller ranking value denotes superior performance across all datasets. The Friedman test results reveal that DS-IHBO achieved the lowest ranking value, significantly outperforming the other five comparative algorithms with statistical significance. The Finner test results indicate no significant difference between DS-IHBO and SMEFS-PSO, but DS-IHBO is significantly superior to the other four comparative algorithms. To strengthen the evidence for pairwise comparisons and quantify the magnitude of differences, the Wilcoxon signed-rank test is further performed in Table 15 and Table 16, and effect sizes are calculated to assess the practical significance of the differences. DS-IHBO is significantly superior to the aforementioned four algorithms in both accuracy and F1 score, with all effect sizes being highly significant. Although there are no statistically significant differences in classification performance between DS-IHBO and SMEFS-PSO, DS-IHBO showed better balance when considering the comprehensive objectives of FS. Subsequent experiments demonstrate that under comparable classification performance, DS-IHBO achieved higher computational efficiency and better feature dimensionality reduction performance.
Table 13. Friedman and Finner test results obtained by six algorithms (Acc).
Table 14. Friedman and Finner test results obtained by six algorithms (F1).
Table 15. Wilcoxon signed-rank test results obtained by six algorithms (Acc).
Table 16. Wilcoxon signed-rank test results obtained by six algorithms (F1).
Table 17 presents the imbalance robustness performance of six algorithms on binary classification problems. DS-IHBO consistently achieved optimal or co-optimal G-mean values across all six datasets. Notably, it significantly outperformed other algorithms on PCMAC, RELATHE, and Parkinson datasets (with improvements of 0.23%, 1.70%, and 1.21%, respectively). For complex datasets (e.g., RELATHE), DS-IHBO substantially exceeds the second-best algorithm (IMOABC). It also exhibits outstanding performance on noise-sensitive datasets (e.g., Parkinson). Furthermore, DS-IHBO maintained lower standard deviations on 5/6 datasets. Particularly on PCMAC, its standard deviation (0.79) is significantly lower than SS-PSO (1.97) and HFSIA (2.08), demonstrating enhanced algorithmic robustness.
Table 17. Average G-mean values obtained by six algorithms (std).
(2) Convergent curves of fitness value: As shown in Figure 7 and Figure 8, DS-IHBO exhibits rapid convergence in early iterations. For example, it achieves over 90% accuracy on the PCMAC dataset within 50 generations, and its convergence curves remain smooth in later iterations for datasets like RELATHE and Lymphoma. Theoretically, this performance stems from the synergy of its three modules: Module I narrows the search space via dimensionality reduction to lay the foundation for fast convergence; Module II maintains population diversity and balances exploration-exploitation through asymmetric initialization and adaptive differential operators, avoiding premature convergence to local optima; Module III reduces redundant evaluations on the original dataset to sustain convergence efficiency. To highlight the advantages of dynamic surrogate technology, DS-IHBO is compared with two baseline algorithms: (1) Static surrogate-based methods (e.g., CUS-SPSO, SMEFS-PSO) rely on fixed surrogate units, which easily introduce evaluation bias due to data distribution changes. For instance, CUS-SPSO shows obvious fitness discrepancies between training and test sets on the USPS dataset, leading to overfitting—while DS-IHBO’s dynamic surrogate updates effectively capture feature interactions and avoid such issues. (2) Surrogate-free EAs (e.g., HFSIA, IMOABC) require full-sample fitness evaluation, significantly increasing computational cost. For example, DS-IHBO’s runtime on PCMAC is only 19.3% of SMEFS-PSO’s. Further, HFSIA fails to maintain population diversity and falls into local optima on the GLIOMA dataset due to resource-intensive full-sample evaluations, whereas DS-IHBO avoids this via surrogate-assisted rapid exploration.
Figure 7. Evolution curves on the first six datasets.
Figure 8. Evolution curves on the last six datasets.
(3) d and t values: Analysis of feature subset size (Table 18): (1) DS-IHBO selects the smallest feature subsets in half of the datasets; (2) HFSIA selects fewer features primarily due to its fixed filter threshold (200 features). Although achieving compact subsets, HFSIA exhibits inferior classification performance and imbalance robustness; (3) IMOABC attains the minimal feature count on PCMAC and Lung2, likely attributable to its tri-objective optimization framework: feature error rate minimization, feature subset proportion reduction, and inter-feature distance maximization.
Table 18. Average d values obtained by six algorithms.
To analyze the trade-off between performance gains and computational efficiency, the t values of these algorithms are compared in Table 19. Key observations include (1) DS-IHBO demonstrates significantly lower t values (computation time) than competitors on 10/12 datasets, owing to its efficient dynamic surrogate assistance. While not dominant on Musk2 and PCMAC, it maintains superior accuracy, F1 score, and G-mean; (2) HFSIA, SMEFS-PSO, and IMOABC—integrating filter-surrogate hybrid techniques—incur lower computational costs than pure evolutionary or surrogate methods. Notably, SMEFS-PSO’s local search phase introduces overhead in high-dimensional spaces, causing efficiency variance across scenarios; (3) CUS-SPSO exhibits suboptimal runtime performance versus peers, resulting from computation-intensive operations: particle selection strategies at each iteration increasing search complexity; pairwise feature correlation calculations during updates, particularly costly in high-dimensional data.
Table 19. Average t values obtained by six algorithms (unit: s).

4.8. Parameter Sensitivity Analysis

The performance of the dynamic surrogate-assisted model depends significantly on two key parameters: the number of iterations T s during the surrogate evaluation phase and the surrogate unit update trigger threshold T n . To determine the optimal parameter configuration, a comprehensive sensitivity analysis is conducted on these two parameters.

4.8.1. Parameter of Surrogate Model Iteration Count ( T s )

Five different values of T s (0, 25, 50, 75 and 100) are tested to evaluate the impact of the surrogate model’s iteration count on algorithm performance. T s = 0 indicates no use of the surrogate model at all, while higher T s values correspond to longer surrogate evaluation phases. Analysis of F1 score boxplots (Figure 9) reveals that on datasets such as DLBCL, PCMAC, and Musk2, T s = 75 achieved the best F1 scores. Its median value is significantly higher than those of other settings. Particularly on the DLBCL dataset, T s = 75 yielded an excellent F1 score of nearly 99%, significantly outperforming T s = 0 (no surrogate model) at 96%. This improvement is attributed to the surrogate model’s mechanism for handling imbalanced classes. T s = 75 generally exhibited smaller variance across most datasets, indicating superior algorithm stability at this setting. In contrast, T s = 0 (no surrogate model) showed larger performance fluctuations on many datasets. Performance showed a significant upward trend as T s increased from 0 to 75. However, when T s = 100, performance degradation is observed on some datasets, suggesting that an excessively long surrogate evaluation phase may introduce errors. Runtime analysis (Figure 10) highlights crucial computational efficiency characteristics: T s = 0 incurred the longest runtime, exceeding 10,000 s on large-scale complex datasets like PCMAC and Musk2. T s = 75 achieved favorable time efficiency on most datasets, reducing runtime by 60% to 80% compared with T s = 0. The runtime exhibited a steady decreasing trend from T s = 0 to T s = 100. Balancing the impact on F1 performance and runtime, T s = 75 is ultimately set to 75.
Figure 9. Box plots of F1 values obtained by DS-IHBO under different T s values.
Figure 10. Running time of DS-IHBO under different T s values.

4.8.2. Dynamic Surrogate Switching Parameter ( T n )

Three different values of T n (5, 6, and 7) are tested to evaluate the impact of surrogate unit update frequency on algorithm performance. The comprehensive performance analysis (Figure 11) shows that T n = 6 achieved the highest accuracy rates on most datasets. Its accuracy advantage is particularly pronounced on complex datasets such as DLBCL and PCMAC. T n = 6 demonstrated the most balanced performance in terms of F1 score, maintaining high performance levels across all datasets. T n = 5 triggered updates too frequently, potentially causing excessive switching between surrogate units and adversely affecting convergence stability. T n = 7 triggered updates too infrequently, which might result in delayed responses to rapid environmental changes, potentially missing optimal opportunities for surrogate unit switching. T n = 6 exhibited a good balance in runtime. It avoided the computational burden associated with overly frequent updates ( T n = 5) and mitigated the risk of slower convergence caused by delayed updates ( T n = 7). T n = 6 provides an appropriate update frequency. It enables timely switching to a more suitable surrogate unit when the algorithm becomes trapped in local optimum. This moderate update frequency ensures the surrogate model adapts effectively to environmental changes during the search process while reducing computational cost.
Figure 11. Acc, F1, and running time of DS-IHBO under different T n values.
The dynamic surrogate-assisted DS-IHBO demonstrates significant improvements over IHBO in classification accuracy, imbalance handling, and computational efficiency. This enhancement stems from four key mechanisms: (1) incorporating full-sample evaluation after 70% of maximum iterations ( 0.75 · T m ) guarantees the final feature subset’s classification quality; (2) the adaptive surrogate selection mechanism continuously updates the most appropriate surrogate model during evolution, maintaining prediction reliability and boosting performance; (3) surrogate sample reduction substantially decreases training set dimensions, dramatically lowering individual evaluation costs; and (4) integrated adaptive KNN with minority-focused sampling prioritizes underrepresented classes during surrogate construction, effectively addressing class imbalance while strengthening minority class recognition.

5. Conclusions and Discussion

5.1. Conclusions

DS-IHBO, an enhanced FS approach derived from the IHBO algorithm, is presented in this paper. The proposed method integrates a dynamic surrogate-assisted mechanism with the IHBO algorithm to address the challenges of high-dimensional imbalanced FS while optimizing the search capability of HBO. Through strategic partitioning of complete sample sets into smaller surrogate units, our approach achieves a substantial reduction in computational overhead for fitness evaluation. Furthermore, the ensemble system continuously refines its surrogate units by comparing predicted and actual fitness values, thereby maintaining high prediction accuracy throughout the optimization process. The performance of our proposed FS method is assessed through comparative analysis with five EAs-based FS techniques (HFSIA, SS-PSO, SMEFS-PSO, IMOABC, and CUS-SPSO), which represent diverse strategies including high-dimensional processing, imbalance handling, surrogate modeling, and correlation-guided selection. Comparative analysis revealed that the DS-IHBO method consistently evolves high-quality and complementary feature subsets with superior classification accuracy while maintaining significantly lower computational overhead compared with competing approaches. Compared with IHBO without a dynamic surrogate, DS-IHBO achieved higher A c c values and lower t values on all datasets, as well as the highest F1 scores and G-mean values. These results significantly verify the efficiency and robustness of the algorithm. In summary, the proposed approach effectively enhances the HBO algorithm’s FS capability, yielding optimal feature combinations for classification tasks.

5.2. Limitations and Future Work

This paper primarily employs swarm intelligence algorithms to identify optimal feature subsets. Although the proposed DS-IHBO algorithm enhances the exploration of the search space and prevents entrapment in local optima, its convergence accuracy requires further improvement. Future research directions will include (1) expanding the application of the DS-IHBO algorithm to other high-dimensional optimization scenarios (e.g., image classification with high-resolution feature maps) to verify its generalizability; (2) exploring the integration of the proposed method with more advanced surrogate models (such as hybrid surrogate models) to further improve the efficiency of high-dimensional FS.

Author Contributions

Conceptualization, Y.M. and B.L.; methodology and software, Z.Y. and B.L.; validation, Y.M., B.L. and Z.Y.; formal analysis, B.L. and Z.Y.; investigation, B.L. and Y.M.; resources, Y.M. and Z.Y.; data curation, B.L. and Z.Y.; writing—original draft, B.L.; writing—review and editing, Y.M., B.L. and Z.Y.; visualization, Z.Y.; supervision and project administration, Y.M.; funding acquisition, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (Grant Nos. 62376089, 62302153, 62302154), the Key Research and Development Program of Hubei Province, China (Grant No. 2023BEB024), the Young and Middle-aged Scientific and Technological Innovation Team Plan in Higher Education Institutions in Hubei Province, China (Grant No. T2023007), and the National Natural Science Foundation of China (Grant No. U23A20318).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data are contained within this article.

Acknowledgments

We greatly appreciate the efforts made by the reviewers and editorial team for our article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tudisco, A.; Volpe, D.; Ranieri, G.; Curato, G.; Ricossa, D.; Graziano, M.; Corbelletto, D. Evaluating the computational advantages of the variational quantum circuit model in financial fraud detection. IEEE Access 2024, 12, 102918–102940. [Google Scholar] [CrossRef]
  2. Fu, G.H.; Wang, J.B.; Lin, W. An adaptive loss backward feature elimination method for class-imbalanced and mixed-type data in medical diagnosis. Chemom. Intell. Lab. Syst. 2023, 236, 104809. [Google Scholar] [CrossRef]
  3. Nssibi, M.; Manita, G.; Chhabra, A.; Mirjalili, S.; Korbaa, O. Gene selection for high dimensional biological datasets using hybrid island binary artificial bee colony with chaos game optimization. Artif. Intell. Rev. 2024, 57, 51. [Google Scholar] [CrossRef]
  4. Arafah, M.; Phillips, I.; Adnane, A.; Hadi, W.; Alauthman, M.; Al-Banna, A.K. Anomaly-based network intrusion detection using denoising autoencoder and Wasserstein GAN synthetic attacks. Appl. Soft Comput. 2025, 168, 112455. [Google Scholar] [CrossRef]
  5. Barbieri, M.C.; Grisci, B.I.; Dorn, M. Analysis and comparison of feature selection methods towards performance and stability. Expert Syst. Appl. 2024, 249, 123667. [Google Scholar] [CrossRef]
  6. Li, M.; Zhao, Y.; Zhang, F.; Luo, B.; Yang, C.; Gui, W.; Chang, K. Multi-scale feature selection network for lightweight image super-resolution. Neural Netw. 2024, 169, 352–364. [Google Scholar] [CrossRef] [PubMed]
  7. Sharif, M.; Tanvir, U.; Munir, E.U.; Khan, M.A.; Yasmin, M. Brain tumor segmentation and classification by improved binomial thresholding and multi-features selection. J. Ambient Intell. Humaniz. Comput. 2024, 15, 1063–1082. [Google Scholar] [CrossRef]
  8. Nguyen, B.H.; Xue, B.; Zhang, M. A constrained competitive swarm optimizer with an SVM-based surrogate model for feature selection. IEEE Trans. Evol. Comput. 2024, 28, 2–16. [Google Scholar] [CrossRef]
  9. Hakami, A. Strategies for overcoming data scarcity, imbalance, and feature selection challenges in machine learning models for predictive maintenance. Sci. Rep. 2024, 14, 9645. [Google Scholar] [CrossRef]
  10. Kim, J.; Kang, J.; Sohn, M. Ensemble learning-based filter-centric hybrid feature selection framework for high-dimensional imbalanced data. Knowl.-Based Syst. 2021, 220, 106901. [Google Scholar] [CrossRef]
  11. Luo, X.; Chen, J.; Yuan, Y.; Wang, Z. Pseudo gradient-adjusted particle swarm optimization for accurate adaptive latent factor analysis. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 2213–2226. [Google Scholar] [CrossRef]
  12. Qi, W.; Zhang, N.; Zong, G.; Su, S.F.; Yan, H.; Yeh, R.H. Event-triggered SMC for networked Markov jumping systems with channel fading and applications: Genetic algorithm. IEEE Trans. Cybern. 2023, 53, 6503–6515. [Google Scholar] [CrossRef] [PubMed]
  13. Hou, Y.; Guo, X.; Han, H.; Wang, J.; Du, Y. Adaptive ant colony optimization algorithm based on real-time logistics features for instant delivery. IEEE Trans. Cybern. 2024, 54, 6358–6370. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, L.; Chen, X. Elite-driven grey wolf optimization for global optimization and its application to feature selection. Swarm Evol. Comput. 2025, 92, 101795. [Google Scholar] [CrossRef]
  15. Yu, F.; Yin, L.; Zeng, B.; Lu, C.; Xiao, Z. A self-learning discrete artificial bee colony algorithm for energy-efficient distributed heterogeneous L-R fuzzy welding shop scheduling problem. IEEE Trans. Fuzzy Syst. 2024, 32, 3753–3764. [Google Scholar] [CrossRef]
  16. Sadeeq, H.T. Cauchy Operator Boosted Artificial Rabbits Optimization for Solving Power System Problems. Eng 2025, 6, 174. [Google Scholar] [CrossRef]
  17. Sun, L.; Sun, S.; Ding, W.; Huang, X.; Fan, P.; Li, K.; Chen, L. Feature selection using symmetric uncertainty and hybrid optimization for high-dimensional data. Int. J. Mach. Learn. Cybern. 2023, 14, 4339–4360. [Google Scholar] [CrossRef]
  18. Ye, Z.; Ma, L.; Chen, H. A hybrid rice optimization algorithm. In Proceedings of the 11th International Conference on Computer Science & Education (ICCSE), Nagoya, Japan, 23–25 August 2016; pp. 169–174. [Google Scholar]
  19. Ye, Z.; Luo, J.; Zhou, W.; Wang, M.; He, Q. An ensemble framework with improved hybrid breeding optimization-based feature selection for intrusion detection. Future Gener. Comput. Syst. 2024, 151, 124–136. [Google Scholar] [CrossRef]
  20. Mei, M.; Zhang, S.; Ye, Z.; Wang, M.; Zhou, W.; Yang, J.; Zhang, J.; Yan, L.; Shen, J. A cooperative hybrid breeding swarm intelligence algorithm for feature selection. Pattern Recognit. 2026, 169, 111901. [Google Scholar] [CrossRef]
  21. Wei, C.; Peng, B.; Li, C.; Liu, Y.; Ye, Z.; Zuo, Z. A two-stage optimized robust kernel density estimation for Bayesian classification with outliers. Int. J. Mach. Learn. Cybern. 2025, 1–25. [Google Scholar] [CrossRef]
  22. Brownlee, A.E.I.; Regnier-Coudert, O.; McCall, J.A.W.; Massie, S.; Stulajter, S. An application of a GA with Markov network surrogate to feature selection. Int. J. Syst. Sci. 2013, 44, 2039–2056. [Google Scholar] [CrossRef]
  23. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. Correlation-guided updating strategy for feature selection in classification with surrogate-assisted particle swarm optimization. IEEE Trans. Evol. Comput. 2022, 26, 1015–1029. [Google Scholar] [CrossRef]
  24. Nguyen, H.B.; Xue, B.; Andreae, P. Surrogate-model based particle swarm optimisation with local search for feature selection in classification. In Applications of Evolutionary Computation; Squillero, G., Burelli, P., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 487–505. [Google Scholar]
  25. Song, X.F.; Zhang, Y.; Gong, D.W.; Gao, X.Z. A fast hybrid feature selection based on correlation-guided clustering and particle swarm optimization for high-dimensional data. IEEE Trans. Cybern. 2022, 52, 9573–9586. [Google Scholar] [CrossRef] [PubMed]
  26. Al-Rawashdeh, R.; Aljughaiman, A.; Albuali, A.; Alsenani, Y.; Alnaeem, M. Enhancing DoS detection in WSNs using enhanced ant colony optimization algorithm. IEEE Access 2024, 12, 134651–134671. [Google Scholar] [CrossRef]
  27. Miao, F.; Wu, Y.; Yan, G.; Si, X. A memory interaction quadratic interpolation whale optimization algorithm based on reverse information correction for high-dimensional feature selection. Appl. Soft Comput. 2024, 164, 111979. [Google Scholar] [CrossRef]
  28. Ahadzadeh, B.; Abdar, M.; Safara, F.; Khosravi, A.; Menhaj, M.B.; Suganthan, P.N. SFE: A simple, fast, and efficient feature selection algorithm for high-dimensional data. IEEE Trans. Evol. Comput. 2023, 27, 1896–1911. [Google Scholar] [CrossRef]
  29. Pan, J.S.; Shi, H.J.; Chu, S.C.; Hu, P.; Shehadeh, H.A. Parallel Binary Rafflesia Optimization Algorithm and Its Application in Feature Selection Problem. Symmetry 2023, 15, 1073. [Google Scholar] [CrossRef]
  30. Li, L.; Xuan, M.; Lin, Q.; Jiang, M.; Ming, Z.; Tan, K.C. An evolutionary multitasking algorithm with multiple filtering for high-dimensional feature selection. IEEE Trans. Evol. Comput. 2023, 27, 802–816. [Google Scholar] [CrossRef]
  31. Ying, W.; Wang, D.; Chen, H.; Fu, Y. Feature selection as deep sequential generative learning. ACM Trans. Knowl. Discov. Data 2024, 18, 1–21. [Google Scholar] [CrossRef]
  32. Pan, H.; Chen, S.; Xiong, H. A high-dimensional feature selection method based on modified Gray Wolf optimization. Appl. Soft Comput. 2023, 135, 110031. [Google Scholar] [CrossRef]
  33. Jiang, Z.; Zhang, Y.; Wang, J. A multi-surrogate-assisted dual-layer ensemble feature selection algorithm. Appl. Soft Comput. 2021, 110, 107625. [Google Scholar] [CrossRef]
  34. Gao, J.; Wang, Z.; Jin, T.; Cheng, J.; Lei, Z.; Gao, S. Information gain ratio-based subfeature grouping empowers particle swarm optimization for feature selection. Knowl.-Based Syst. 2024, 286, 111380. [Google Scholar] [CrossRef]
  35. Lin, H.; Wang, C.; Hao, Q. A novel personality detection method based on high-dimensional psycholinguistic features and improved distributed Gray Wolf optimizer for feature selection. Inf. Process. Manag. 2023, 60, 103217. [Google Scholar] [CrossRef]
  36. Song, X.; Jiang, Z.; Zhang, Y.; Peng, C.; Guo, Y. A surrogate-assisted multi-phase ensemble feature selection algorithm with particle swarm optimization in imbalanced data. IEEE Trans. Emerg. Top. Comput. Intell. 2025, 1–16. [Google Scholar] [CrossRef]
  37. Song, X.; Zhang, Y.; Gong, D.; Liu, H.; Zhang, W. Surrogate sample-assisted particle swarm optimization for feature selection on high-dimensional data. IEEE Trans. Evol. Comput. 2023, 27, 595–609. [Google Scholar] [CrossRef]
  38. Liu, S.; Wang, H.; Peng, W.; Yao, W. A Surrogate-Assisted Evolutionary Feature Selection Algorithm With Parallel Random Grouping for High-Dimensional Classification. IEEE Trans. Evol. Comput. 2022, 26, 1087–1101. [Google Scholar]
  39. Liu, S.; Wang, H.; Peng, W.; Yao, W. Surrogate-assisted evolutionary algorithms for expensive combinatorial optimization: A survey. Complex Intell. Syst. 2024, 10, 5933–5949. [Google Scholar] [CrossRef]
  40. Chu, S.C.; Yuan, X.; Pan, J.S.; Lin, B.S.; Lee, Z.J. A multi-strategy surrogate-assisted social learning particle swarm optimization for expensive optimization and applications. Appl. Soft Comput. 2024, 162, 111876. [Google Scholar] [CrossRef]
  41. Yu, K.; Sun, S.; Liang, J.; Chen, K.; Qu, B.; Yue, C.; Suganthan, P.N. A Space Transformation-Based Multiform Approach for Multiobjective Feature Selection in High-Dimensional Classification. IEEE Trans. Syst. Man Cybern. Syst. 2024, 54, 7305–7317. [Google Scholar]
  42. Hu, P.; Zhu, J. A filter-wrapper model for high-dimensional feature selection based on evolutionary computation. Appl. Intell. 2025, 55, 581. [Google Scholar] [CrossRef]
  43. Gong, H.; Li, Y.; Zhang, J.; Zhang, B.; Wang, X. A new filter feature selection algorithm for classification task by ensembling Pearson correlation coefficient and mutual information. Eng. Appl. Artif. Intell. 2024, 131, 107865. [Google Scholar] [CrossRef]
  44. Wang, Q.; Jiang, H.; Ren, J.; Liu, H.; Wang, X.; Zhang, B. An intrusion detection algorithm based on joint symmetric uncertainty and hyperparameter optimized fusion neural network. Expert Syst. Appl. 2024, 244, 123014. [Google Scholar] [CrossRef]
  45. Ramírez-Gallego, S.; Lastra, I.; Martínez-Rego, D.; Bolón-Canedo, V.; Benítez, J.M.; Herrera, F.; Alonso-Betanzos, A. Fast-mRMR: Fast minimum redundancy maximum relevance algorithm for high-dimensional big data. Int. J. Intell. Syst. 2017, 32, 134–152. [Google Scholar] [CrossRef]
  46. Liaw, L.C.M.; Tan, S.C.; Goh, P.Y.; Lim, C.P. A histogram SMOTE-based sampling algorithm with incremental learning for imbalanced data classification. Inf. Sci. 2025, 686, 121193. [Google Scholar] [CrossRef]
  47. Zhu, Y.; Li, W.; Li, T. A hybrid artificial immune optimization for high-dimensional feature selection. Knowl.-Based Syst. 2023, 260, 110111. [Google Scholar] [CrossRef]
  48. Li, J.; Luo, T.; Zhang, B.; Chen, M.; Zhou, J. IMOABC: An efficient multi-objective filter–wrapper hybrid approach for high-dimensional feature selection. J. King Saud Univ. Comput. Inf. Sci. 2024, 36, 102205. [Google Scholar] [CrossRef]
  49. Wang, H.; Li, G.; Wang, Z. Fast SVM classifier for large-scale classification problems. Inf. Sci. 2023, 642, 119136. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.