Next Article in Journal
Complexity, Uncertainty, and Entropy: Applications to Food Sensory Perception and Other Complex Phenomena
Next Article in Special Issue
AlphaRouter: Bridging the Gap Between Reinforcement Learning and Optimization for Vehicle Routing with Monte Carlo Tree Searches
Previous Article in Journal
Solving Flexible Job-Shop Scheduling Problems Based on Quantum Computing
Previous Article in Special Issue
Adaptive CoCoLasso for High-Dimensional Measurement Error Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Regularized Feature Selection for Class-Specific and Global Feature Associations

1
College of Computer Science, Nankai University, Tianjin 300350, China
2
School of Mathematics and Statistics Science, Ludong University, Yantai 264025, China
3
Information and Navigation College, Air Force Engineering University, Xi’an 710077, China
*
Authors to whom correspondence should be addressed.
Entropy 2025, 27(2), 190; https://doi.org/10.3390/e27020190
Submission received: 14 January 2025 / Revised: 3 February 2025 / Accepted: 11 February 2025 / Published: 13 February 2025
(This article belongs to the Special Issue Information-Theoretic Methods in Data Analytics)

Abstract

:
Understanding feature associations is vital for selecting the most informative features. Existing methods primarily focus on global feature associations, which capture overall relationships across all samples. However, they often overlook class-specific feature interactions, which are essential for capturing locality features that may only be significant within certain classes. In this paper, we propose Dual-Regularized Feature Selection (DRFS), which incorporates two feature association regularizers to address both class-specific and global feature relationships. The class-specific regularizer captures the local geometric structure of features within each class. Meanwhile, the global regularizer utilizes a global feature similarity matrix to eliminate redundant features across classes. By combining these regularizers, DRFS selects features that preserve both local interactions within each class and global discriminative power, with each regularizer complementing the other to enhance feature selection. Experimental results on eight public real-world datasets demonstrate that DRFS outperforms existing methods in classification accuracy.

1. Introduction

Feature selection (FS) aims to identify the most important features to improve model performance and has been extensively studied [1,2,3]. By removing redundant, noisy, and irrelevant features, FS enhances model performance, reduces computational complexity, and minimizes storage requirements [4,5]. Due to these benefits, FS methods have been successfully applied across a wide range of fields, including bioinformatics [6,7], computer vision [8], multi-view learning [9,10], and recommendation systems [11].
A critical aspect of FS involves understanding feature associations, which can reveal redundancies or meaningful patterns in data [12,13]. Redundant features, characterized by overlapping information, often degrade model performance [14]. In contrast, meaningful feature associations uncover inherent patterns that are essential for accurate predictions, such as the relationship between temperature and precipitation in weather forecasting. Existing FS methods generally emphasize global feature associations, which capture overall relationships across the entire dataset. These methods generally construct feature similarity matrices and apply regularization techniques to guide the selection process. For example, some approaches utilize cosine similarity matrices to model feature associations and introduce global redundancy regularization to eliminate redundant features [15,16]. Other approaches consider manifold structures in the feature space to select features that preserve these structures [12,17,18]. These methods, as illustrated in Figure 1a, often focus on global interactions across the features, emphasizing paired feature associations or preserving global feature manifolds.
While global feature associations have significantly advanced feature selection, they overlook class-specific interactions that are essential for capturing patterns unique to individual classes. Unlike global associations, which capture relationships across the entire dataset, class-specific associations highlight feature relationships that are meaningful within specific classes but may not generalize to others. Disregarding these localized interactions can result in suboptimal feature selection, as key features relevant to particular classes may be overlooked. The limitations of focusing solely on global associations become apparent in various applications. For example, in medical diagnoses involving multi-class diseases, global associations might highlight features like age and blood pressure as significant across all conditions. However, class-specific interactions, such as the combination of elevated blood sugar and cholesterol levels for diabetes, or the relationship between heart rate variability and blood pressure for cardiovascular diseases, are unique to specific classes. Overlooking such interactions can obscure important class-specific insights, resulting in less effective feature selection.
To address these challenges, we propose Dual-Regularized Feature Selection (DRFS), a novel approach that considers class-specific and global feature associations. By introducing separate regularization terms for global and class-specific associations, our approach aims to enhance the selection of features that are relevant across the entire dataset while capturing unique patterns within individual classes. The class-specific regularizer captures feature interactions unique to each class, while the global regularizer helps reduce redundancy across features by considering the entire dataset. This dual consideration not only improves model performance by selecting more informative features but also enhances interpretability by providing insights into both general and class-specific patterns. As shown in Figure 1b, our method constructs a feature weight matrix that considers class-specific interactions independently while addressing global redundancies. Notably, the class-specific feature similarity matrices exhibit unique structures that differ from one another and from the global matrix, reflecting their localized characteristics. Moreover, the flexibility of DRFS allows it to be integrated with various FS models. In this study, we validate DRFS within a sparse regression framework, which minimizes the difference between selected features and sample labels while incorporating both class-specific and global regularization. Our main contributions are as follows:
  • We propose a novel dual-regularization approach for FS, balancing class-specific feature interactions and global redundancy elimination.
  • We theoretically show that our dual regularization effectively preserves class-specific feature geometry while addressing global redundancies.
  • Extensive experiments on eight real-world datasets show that DRFS consistently outperforms existing FS methods in classification accuracy.
The rest of the paper is organized as follows: Section 2 reviews related work and introduces notation. Section 3 presents the proposed DRFS method. Section 4 details the optimization algorithm, including its complexity and convergence analysis. Section 5 presents the experimental results and analysis. Section 6 concludes the paper.

2. Related Works

FS has been studied for many years [1,19,20]. Traditional FS methods are generally categorized into three approaches: filter, wrapper, and embedded methods [1]. Filter methods rely on statistical measures, such as information gain and correlation coefficients, to evaluate and select features [4,5,21]. These methods are among the simplest and fastest techniques for FS. However, they may overlook relationships between features, as they rely solely on individual feature statistics rather than considering feature interactions. Wrapper methods, in contrast, approach FS as a search problem. They evaluate the performance of different feature subsets by training models on each subset, ultimately selecting the optimal one [22,23]. While wrapper methods provide a better understanding of feature relationships compared to filter methods, this advantage comes at a high computational cost, making them less suitable for high-dimensional datasets. Embedded methods integrate feature selection directly into the model training process, selecting the optimal subset while minimizing redundancy and noise [2,13,24]. Many embedded methods employ sparse learning techniques to constrain feature weights during training  [25,26]. These embedded feature selection algorithms have achieved very good performance in many fields [9]. The method proposed in this work belongs to the category of embedded feature selection methods. Below, we provide a brief overview of related works.
Recently, significant attention has been directed towards leveraging feature associations to enhance FS methods. These approaches generally follow two key steps: (1) constructing a feature similarity matrix to capture feature associations, and (2) designing models to select features based on the constructed matrix. For example, Chen et al. introduced the uncorrelated Lasso, which aims to select uncorrelated features by utilizing a squared cosine similarity matrix and l 1 -norm minimization [15]. Building on this, Xu et al. proposed a global redundancy minimization approach (GRMOR), employing orthogonal regression to identify discriminative features [16]. Zhang et al. proposed Interacted Lasso, capturing feature interactions using hypergraphs [27]. Similarly, Cui et al. proposed the structurally interacting elastic net (InElasticNet), which incorporates an information-theoretic criterion for measuring feature associations [28]. They further extended their work to focus on joint feature relevance through the fused Lasso (InFusedLasso) [29]. These methods, predominantly based on Lasso-like frameworks, are effective but limited by their reliance on linear association measures.
Feature manifold learning has recently emerged as a promising direction for feature selection. Xu et al. utilized Pearson correlation coefficients and mutual information to construct feature associations and proposed a redundancy minimization regularization (SFSRM) [26]. Lai et al. improved upon this by using Gaussian functions to refine feature similarity construction and introduced the new regularization term (AGLRM) [30]. Shang et al. explored manifold-preserving methods, assuming that manifold structures exist in the feature space. They proposed dual-graph regularized feature selection (NSSRD) [17] and NNSAFS [18], employing parameter-free and Gaussian-based techniques for feature graph construction. In addition, Roffo et al. proposed infinite feature selection (Inf-FS), which selects features by using the power series of the similarity matrix [31]. They used the Fisher criterion, normalized mutual information, and normalized standard deviation to calculate feature associations. Recently, Cohen et al. proposed a filter FS method called Manifold-based Feature Selection (ManiFeSt), which constructs feature similarity representations for each class and computes feature scores using mean and difference operators [12].
A detailed summary of these methods is presented in Table 1. As shown in Table 1, existing methods mostly consider global feature associations, but not class-specific feature associations. Although ManiFeSt considers class-specific feature associations, it lacks consideration of global feature associations. Our method not only considers global associations, but also class feature associations. Specifically, these methods calculate feature associations in the entire sample space. Additionally, ManiFeSt effectively leverages the feature manifold of each class to improve FS performance. While our method also preserves the structure of individual classes, it differs from ManiFeSt in two key aspects. First, we construct feature similarity matrices using k-nearest neighbor graphs, focusing on the most relevant feature interactions. Unlike ManiFeSt, which computes all pairwise feature similarities, our approach better captures the local manifold structure. Second, we incorporate global feature associations to eliminate redundancy, enabling the selection of more discriminative features. These distinctions allow our model to achieve more effective feature selection.
Notation: For a matrix X R n × d , its ( i , j ) -th element, i-th column, and j-th row are denoted as X i j , x i , and  x j , respectively. The inverse, transpose, and trace of X are X 1 , X T , and  Tr ( X ) , respectively. The 2 , 1 -norm is X 2 , 1 = i n x i 2 . The Frobenius norm of X is given by X F 2 = i = 1 d x i 2 2 . Let n, d, and r denote the number of samples, features, and classes, respectively. The data matrix is X R n × d , where x i R 1 × d represents the i-th sample, and  x j R n represents the j-th feature. The label matrix is Y R n × r . The feature similarity matrix for class l is M ( l ) R d × d , and the global feature similarity matrix is M R d × d .

3. Methods

We propose the novel Dual-Regularized Feature Selection (DRFS) model with two complementary regularizations: class-specific feature interaction preservation and global redundancy elimination. The first regularization ensures the preservation of local feature interactions within each class, capturing the underlying feature manifold structure. The second regularization addresses global feature associations, aiming to eliminate redundant features. These complementary strategies combine to select features that retain local geometric relationships and enhance global discriminative power, as shown in Figure 2. To validate the effectiveness of the proposed regularizations, we use a sparse regression model as the baseline. Notably, our regularization technique is designed to function as an adaptable plugin, compatible with various FS models.

3.1. Class-Specific Feature Interaction Preservation

Class-specific feature interaction preservation regularization is developed in two stages. First, we construct class-specific feature similarity matrices to capture the local manifold structure. Next, we introduce a regularization strategy based on these matrices to preserve the local feature geometry.
In the first stage, we model the geometric structures of features for each class. Given a dataset X with r classes, we partition the data by class labels, such that X = { X ( 1 ) , X ( 2 ) , , X ( r ) } , where X ( l ) R n ( l ) × d represents the sample matrix for class l with n ( l ) samples. i = 1 l n ( l ) = n is the total number of samples. For each class l, we define a feature similarity matrix M ( l ) R d × d that quantifies the relationships between features. To capture non-linear interactions between features, we use the Radial Basis Function (RBF) kernel, which is commonly employed for manifold learning [12]. To model local feature geometric structure, we define the feature similarity matrix for class l as:
M i j ( l ) = e x i ( l ) x j ( l ) 2 2 / σ l 2 , x i ( l ) N k ( x j ( l ) ) or x j ( l ) N k ( x i ( l ) ) 0 , otherwise ,
where x i ( l ) and x j ( l ) represent the i-th and j-th features of class l, and  N k ( x i ( l ) ) is the set of k-nearest neighbors of feature x i ( l ) . M i j ( l ) reflects the relationship between features i and j within class l. The k-nearest neighbors approach captures the most relevant features, providing a more accurate representation of the feature manifold compared to pairwise similarity, as shown in [12]. Moreover, the storage complexity of our approach is O ( d k ) , which is more efficient than the O ( d 2 ) required by pairwise methods.
In the second stage, we apply the feature similarity matrix to solve the following optimization problem, which helps preserve the local feature interactions:
min w l 0 Tr ( w l L M ( l ) w l ) = i , j d ( W i l W j l ) 2 M i j ( l ) ,
where L M ( l ) = D ( l ) M ( l ) is the Laplacian matrix of M ( l ) , and  D ( l ) is the degree matrix with entries D i i ( l ) = j = 1 d M i j ( l ) . The vector w l represents the feature weights for class l. If M i j ( l ) is large, indicating that features i and j are similar in class l, their contributions and corresponding weights w i l and w j l should be similar, ensuring that highly correlated interactive features within a specific class are selected simultaneously.
For all r classes, we aggregate the class-specific regularization terms as follows:
min W 0 Ω l o c a l ( W ) = l = 1 r Tr ( w l L M ( l ) w l ) .
where W = [ w 1 , , w r ] is the feature weight matrix, which is solved in a column-wise fashion.

3.2. Global Feature Redundancy Elimination

While class-specific regularization captures interactions within classes, highly correlated features may generate redundancy. To address this, we propose a global feature redundancy elimination regularization that removes redundancy across classes. First, we construct a global feature similarity matrix M R d × d to represent feature associations across the entire feature space. The global feature similarity matrix is defined as
M i j = e x i x j 2 2 / σ 2 , x i N k ( x j ) or x j N k ( x i ) 0 , otherwise ,
where N k ( x i ) denotes the set of k-nearest neighbors of feature x i . This ensures that only the most correlated feature associations are captured. In contrast to the local feature interactions within a specific class, as measured in (1), the global matrix in (4) quantifies feature redundancy across all classes.
Interestingly, consider two redundancy features i and j, which exhibit consistently high similarity across all classes. In other words, for every class l, the condition x i ( l ) N k ( x j ( l ) ) or x j ( l ) N k ( x i ( l ) ) holds. Consequently, the global condition x i N k ( x j ) or x j N k ( x i ) is also satisfied. We now establish the connection between the global feature similarity in (4) and class-specific similarities in (1) in the following theorem.
Theorem 1. 
Let σ = σ 1 = = σ r and assume that features i and j exhibit redundancy in each class l. Then, the global similarity between features can be expressed as
M i j = M i j ( 1 ) M i j ( 2 ) M i j ( r ) .
Proof. 
By exploiting the decomposition of the Euclidean distance across classes, we have
M i j = e x i x j 2 2 / σ 2 = e l r x i ( l ) x j ( l ) 2 2 / σ 2 = Π l r e x i ( l ) x j ( l ) 2 2 / σ 2 = Π l r M i j ( l )
This completes the proof.    □
From Theorem 1, the global feature similarity matrix aggregates local feature associations across all classes, assigning high values to redundant features. In practice, the assumptions in Theorem 1 are challenging to satisfy due to the computation of feature similarity matrices using k-nearest neighbors with varying σ l . Nonetheless, the aggregation approach described in Theorem 1 still ensures that redundant feature associations are preserved in M , either across all classes or within specific classes.
To eliminate global redundancy, we introduce the following regularization term:
min W 0 Ω g l o b a l ( W ) = p = 1 r q p r w p T Mw q ,
where w p and w q represent the p and q columns of W , respectively. Minimizing (7) reduces redundancy by encouraging the model to eliminate features that exhibit high correlation across different classes. To better understand the effect of minimizing p = 1 r q p r w p T Mw q , we expand (7) as follows:
p = 1 r q p r w p T Mw q = p = 1 r q p r ( i d j d W i p W j q M i j ) = i , j d ( p = 1 r q p r W i p W j q M i j ) = i , j d ( p = 1 r W i p ( q p r W j q ) M i j ) = i , j d ( | w i | | w j | M i j p = 1 r W i p W j p M i j ) = i , j d ( | w i | | w j | M i j ) + i , j d ( p = 1 r W i p W j p ( M i j ) ) ,
where | w i | = p = 1 r W i p represents the weight accumulation of w , which indicates the importance of feature i. Minimizing the formula above is equivalent to minimizing the two components separately. We analyze each term as follows. (1) For the first term: If M i j is large, it indicates that features i and j are highly correlated. In this case, if  | w i | > | w j | , the weight | w j | will decrease to minimize this term. When features i and j are redundant, the model reduces the weight of one feature to eliminate redundancy. (2) For the second term: If M i j is large (i.e., M i j is small), it indicates that the features i and j are less correlated. To minimize this term, the weights w i p and w j p for class p will differ, emphasizing the distinction between two features. When the features are not redundant, they should contribute to different classes to enhance the model’s discriminative ability. These observations highlight the significance of our approach. It clearly demonstrates that (7) effectively removes redundant features between classes and within classes.

3.3. Unified Objective Function

We introduce the dual-regularization term in (9), which consists of the class-specific regularization term, Ω l o c a l ( W ) , and the global regularization term, Ω g l o b a l ( W ) :
R ( W ) = min W 0 1 r Ω l o c a l ( W ) + 1 r ( r 1 ) Ω g l o b a l ( W ) = 1 r l T r ( w l T L M ( l ) w l ) + 1 r ( r 1 ) p q p r w p T Mw q .
Here, 1 r and 1 r ( r 1 ) are normalization factors that balance the contributions of the class-specific and global regularization terms. The dual-regularization term, R ( W ) , effectively preserves the local feature structures while simultaneously reducing global feature redundancy. This regularization term can be easily integrated into existing FS models, enriching them with additional feature association information. In this work, we apply it to the following sparse regression model, a simple yet widely used approach. The final objective function is defined as
min W 0 XW Y F 2 + α W 2 , 1 + β ( 1 r Ω l o c a l ( W ) + 1 r ( r 1 ) Ω g l o b a l ( W ) ) .
where Y is the label matrix and W 2 , 1 is the sparse norm used to enforce sparsity in W . The constraint W 0 facilitates the selection of features with meaningful physical interpretations [24]. The parameter α controls the sparsity level, and  β determines the strength of the dual-regularization term. Note that we use β to scale both the class-specific and global regularization terms. The normalization factors 1 r and 1 r ( r 1 ) ensure these terms have comparable magnitudes, thus simplifying the model by reducing the number of parameters.

4. Optimization

In this section, we propose an efficient algorithm to solve the objective function in (10), leveraging the Augmented Lagrangian Multiplier (ALM) method [32]. First, we introduce a slack variable Z and reformulate (10) as follows:
min W XW Y F 2 + α W 2 , 1 + β ( 1 r Ω l o c a l ( Z ) + 1 r ( r 1 ) Ω g l o b a l ( Z ) )       s . t . W 0 , Z = W , Z 0 .
The corresponding augmented Lagrangian function is
min W 0 , Z 0 XW Y F 2 + α W 2 , 1 + μ W Z + Λ μ F 2 + β ( 1 r Ω l o c a l ( Z ) + 1 r ( r 1 ) Ω g l o b a l ( Z ) ) ,
where Λ is the Lagrange multiplier and μ is the penalty parameter. We develop an alternatively iterative algorithm to solve it.

4.1. Optimize W

When Z is fixed, problem (12) can be transformed to
min W 0 XW Y F 2 + μ W Z + Λ μ F 2 + α W 2 , 1 .
The corresponding Lagrangian function is
min W XW Y F 2 + μ W Z + Λ μ F 2 + α T r ( W T DW ) + T r ( Σ W ) ,
where Σ is Lagrange multipliers for the constrains W 0 , and  D is the diagonal matrix with D i i = 1 / 2 W i 2 . Setting the derivative with respect to W to zero, we have
( X T X + μ I + α D ) W X T Y μ Z + Λ + Σ = 0 .
Using the Karush–Kuhn–Tucker condition Σ i j W i j , we have
( ( X T X + μ I + α D ) W ) i j W i j = ( X T Y + μ Z Λ ) i j W i j
To solve this, we decompose any matrix O into two non-negative parts, O = O + O . Then, the update rule for W is
W i j W i j ( X T X + μ I + α D ) W + ( X T Y + μ Z Λ ) + i j ( X T X + μ I + α D ) + W + ( X T Y + μ Z Λ ) i j

4.2. Optimize Z

With fixed W , problem (12) becomes
min Z β ( 1 r Ω l o c a l ( Z ) + 1 r ( r 1 ) Ω g l o b a l ( Z ) ) + μ W Z + Λ μ F 2 .
The above formula can be reformulated as
min z l β r l T r ( z l T L M ( l ) z l ) + β r ( r 1 ) l r q l r Z l T Mz q + μ l w l z l + Λ l μ 2 2 .
Taking the derivative with respect to z l and setting it to zero, we have
2 β r L M ( l ) z l + 2 β r ( r 1 ) q l r Mz q + 2 μ ( z l w l Λ l μ ) = 0
Then, we have
( β r L M ( l ) + μ I ) z l = μ w l + Λ l β r ( r 1 ) q l r Mz q
The solution of z l is
z l = ( β r L M ( l ) + μ I ) 1 ( μ w l + Λ l β r ( r 1 ) q l r Mz q )
To guarantee that the solution Z remains within the feasible region ( Z 0 ), an element-wise projection is applied:
z l = max ( z l , 0 )
where the max ( · ) operation is applied element-wise, projecting z l into the feasible region.
The detailed steps for DRFS are outlined in Algorithm 1.
Algorithm 1 DRFS
Input: Data matrix X , label matrix Y , α ,   β .
Initialize μ = 1 ,   ρ = 1.1 ,   μ m a x = 10 8 .
Output: Feature weight matrix W .
  1:
Compute feature similarity matrices M ( 1 ) ,   ,   M ( r ) , M
  2:
while not converaged do
  3:
   Update W by W i j W i j ( X T X + μ I + λ D ) W + ( X T Y μ Z + Λ ) + i j ( X T X + μ I + λ D ) + W + ( X T Y μ Z + Λ ) i j ;
  4:
   Compute diagonal matrix D as D i i = 1 / 2 W i 2 ;
  5:
   for  l = 1   to r do
  6:
       Update z l by z l = 1 r L M ( l ) + μ I 1 ( μ w l + Λ l 1 r ( r 1 ) k l r Mz k ) ;
  7:
        z l = max ( z l , 0 ) ;
  8:
   end for
  9:
   Update Λ Λ + μ ( W Z ) ;
10:
   Update μ = min ( ρ μ , μ m a x ) ;
11:
end while

4.3. Complexity and Convergence

DRFS is solved using Algorithm 1, which primarily involves optimizing W and Z . The computational complexity of optimizing W is O ( d 2 n ) , while the most computationally expensive step is the optimization of z l , requiring matrix inversion with a complexity of O ( d 3 ) . Thus, the overall time complexity is O ( d 2 n + d 3 r ) . In practice, the number of classes r is much smaller than both the number of samples n and the number of features d (i.e., r d , r n ). As a result, the computational overhead introduced by r is relatively small.
Problems (13) and (18) are convex, ensuring reliable optimization. The convergence of W , updated via (17), has been theoretically established in [24]. The convergence of D is supported by the analysis in [25]. Additionally, the variable Z has a closed-form solution, ensuring its optimality in each iteration. However, it is difficult to prove the overall convergence due to the complexity of the ALM method. Fortunately, we have empirically validated the convergence of our method, as demonstrated in Section 5.6.

5. Experiments

In this section, we design four experiments to evaluate the effectiveness of DRFS, focusing on the following aspects: (1) The performance of DRFS with SVM and 1NN classifiers. (2) Ablation studies to validate the role of class-specific and global regularization terms. (3) Sensitivity analysis to investigate the impact of the parameters α and β , as well as the sensitivity to class-specific and global regularization terms. (4) A convergence analysis of Algorithm 1.

5.1. Datasets

We conduct experiments on eight public benchmark datasets, including four bioinformatics datasets (lung_discrete [33], SRBCT [34], LUNG [35], GLIOMA), one spoken letter recognition dataset (Isolet [36]), and three image datasets (mfeat [37], Yale [38], warpPIE10P [38]). The details of all the datasets are summarized in Table 2. The lung_discrete dataset contains 73 samples described by 325 gene expressions. The Isolet dataset consists of 1560 spoken letter samples, each represented by 617 features. The mfeat dataset consists of 2000 samples of handwritten digits (0–9), characterized by 649 features. The Yale dataset includes 165 facial images of size 32 × 32, with each pixel treated as a feature. The warpPIE10P dataset comprises 210 images of 10 individuals, each represented by 2420 features (44 × 55). The SRBCT dataset consists of 83 samples, each described by 2308 gene expression features. The LUNG dataset includes 203 samples, each characterized by 3312 gene expression features. The GLIOMA dataset contains 50 samples, with each represented by 4434 gene expression features. The datasets used for the experiments are derived from the scikit-feature selection repository (https://jundongl.github.io/scikit-feature/datasets.html, accessed on 13 January 2025) and the UCI machine learning repository (https://archive.ics.uci.edu/ml/datasets, accessed on 13 January 2025).

5.2. Experiment Settings

We evaluate the performance of DRFS by comparing it with several state-of-the-art feature selection methods. Specifically, we include three traditional filter-based methods: Fisher Score [5], Gini Index [21], and RelieF [4]. Additionally, we compare DRFS with RJFWL [24], a non-negative constrained sparse regression method. Most importantly, we compare DRFS with three feature graph-based methods: GRMOR [16], Inf-FS [31], and ManiFeSt [12]. GRMOR [16] computes feature similarity using squared cosine similarities to identify and eliminate redundant features. Inf-FS [31] adopts a weighted feature similarity calculation approach and leverages the power series of matrices or Markov chains to determine feature scores. ManiFeSt [12] calculates feature similarity matrices under different classes and calculates feature scores based on manifold learning. The code for Fisher, Gini Index, and RelieF is available at https://github.com/jundongl/scikit-feature, accessed on 13 January 2025. The code for Inf-FS can be found at https://github.com/giorgioroffo/Infinite-Feature-Selection, accessed on 13 January 2025. The code for ManiFeSt is available at https://github.com/DavidCohen2/ManiFeSt, acessed on 12 February 2025. Our code is available at https://github.com/Wangchenchen233/DRFS, accessed on 13 January 2025.
We evaluate the performance of feature selection using classification accuracy on the selected features as the evaluation metric. We employ 5-fold cross-validation to assess performance with linear SVM and 1-Nearest Neighbor (1-NN) as classifiers. The parameter C for the linear SVM is fixed at 1 across all experiments, as in [39]. For all datasets, the range of selected features is set to { 10 , 20 , 30 , 40 , 50 , 60 , 70 , 80 , 90 , 100 } . For ManiFeSt, we use a naive one-vs-one extension for the multi-class setting. For global feature similarity matrix, we set σ 2 = m e a n ( x i x j 2 2 ) ,   i ,   j = 1 ,   ,   d . Similarly, for class-specific feature similarities, we set σ l 2 = m e a n ( x i ( l ) x j ( l ) 2 2 ) ,   i ,   j = 1 ,   ,   d . The number of feature neighbors k is set to 10. For DRFS, we set the parameter search interval to { 0.001 , 0.01 , 0.1 , 1 , 10 , 100 , 1000 } . For RJFWL and GRMOR, the parameters are adjusted within the interval { 0.001 , 0.01 , 0.1 , 1 , 10 , 100 , 1000 } . For Inf-FS, we adjust the parameters at { 0.2 , 0.4 , 0.6 , 0.8 } . All experiments are conducted on a PC with an Intel(R) Core(TM) i9-7900X CPU @ 3.3 GHz and 64 GB RAM.

5.3. Results and Analysis

Figure 3 and Figure 4 illustrate the classification accuracy results achieved by the SVM and 1-NN classifiers, respectively. Table 3 and Table 4 summarize the best performance achieved by each method. Overall, DRFS consistently outperforms competing approaches across multiple datasets and classifiers. When compared with traditional filter-based methods such as Fisher Score, Gini Index, and RelieF, DRFS demonstrates superior classification accuracy. By incorporating feature association regularization, DRFS also outperforms the sparse regression-based method RJFWL, showcasing its ability to effectively capture feature interactions.
Compared to other feature association methods, DRFS achieves the best overall performance. Notably, DRFS outperforms GRMOR, which focuses solely on global feature redundancy minimization. For instance, on the lung_discrete and Yale datasets, DRFS improves classification accuracy by 5.3% and 5.5% with the SVM classifier, respectively. Furthermore, DRFS outperforms other advanced graph-based and manifold learning methods, including Inf-FS and ManiFeSt. For instance, as shown in Table 4, DRFS achieves higher classification accuracy than ManiFeSt on all datasets with the 1-NN classifier, outperforming Inf-FS on all datasets except for LUNG. This result underscores DRFS’s ability to leverage both global and local feature associations effectively, resulting in improved feature selection and classification outcomes.
DRFS achieves the best performance on biological datasets, such as the lung_discrete and GLIOMA datasets, as shown in Figure 3 and Figure 4. Compared to the optimal results in Table 3, DRFS achieves a 4% improvement on the GLIOMA dataset. Similarly, compared to Table 4, DRFS improves by 2.76% on the lung_discrete dataset. On balanced datasets such as Isolet, mfeat, Yale, and warpPIE10P, DRFS delivers the best results when using the SVM classifier. For unbalanced datasets, including lung_discrete, SRBCT, and GLIOMA, DRFS consistently outperforms all other methods. However, on the particularly imbalanced LUNG dataset, DRFS achieves only the second-best performance, highlighting a potential limitation in handling extreme data imbalance.
To statistically compare the performance of the methods, we apply the Friedman test, followed by the Bonferroni–Dunn post hoc analysis. The results of the Friedman test, presented in Table 5, reject the null hypothesis—that all methods perform equivalently—at a significance level of 0.05. Figure 5 displays the critical distance (CD) diagrams, with a critical distance of 3.7121, which indicates that the methods are statistically similar within this threshold. It is noteworthy that DRFS consistently ranks first when evaluated using both the SVM and 1NN classifiers, underscoring its superior performance across these classifiers.

5.4. Ablation Study

Table 6 and Table 7 evaluate the impact of the two regularization terms in DRFS on classification performance. These regularization terms are class-specific regularization and global regularization, which together form four different module configurations. In Table 6 and Table 7, the column labeled “Modules” indicates the specific combinations of these regularization terms used in each model. A check mark (√) signifies that the corresponding regularization term has been selected for that model configuration.
The results demonstrate that the combination of both regularization terms consistently outperforms configurations where only one term is used or no regularization is applied. Using a single regularization term results in notable performance gains compared to the absence of regularization, indicating that even partial regularization contributes to effective feature selection. For instance, combining both terms results in relative improvements of 8.5% and 12.9% on the lung_discrete and Yale datasets with SVM, respectively, compared to no regularization. Similarly, with the 1-NN classifier, combining both regularization terms achieves improvements of 9.2% and 10.6% on the same datasets. Interestingly, on the mfeat dataset, the class-specific regularization term alone achieves performance comparable to the combined approach. This observation underscores the critical role of capturing class-specific feature associations, a key contribution of DRFS. These findings highlight the importance of incorporating both global and class-specific regularization terms to enhance classification accuracy.

5.5. Parameter Sensitivity

DRFS introduces two key parameters, α and β , that significantly influence its performance. The parameter α impacts performance by controlling the sparsity of the feature weight matrix. The parameter β governs the strength of the feature association regularization, influencing the model’s ability to capture feature interactions. To evaluate their sensitivity, we fix the number of selected features to 50 and examine their impact on classification accuracy. The results are illustrated in Figure 6. It is evident that optimal performance arises from a balanced interplay between α and β . Notably, the parameter α exerts a more pronounced effect on model performance. When α is set too high, the feature selection matrix becomes overly sparse, leading to the selection of only a few features. Consequently, the model may lack sufficient information for effective learning. Conversely, when α is too low, the model selects many features, including redundant or irrelevant ones, which degrades performance. In comparison, the model’s performance shows weaker sensitivity to variations in β . This relative robustness can be attributed to the fact that β regulates feature association regularization, which refines feature relationships rather than directly influencing sparsity. These findings highlight the importance of careful parameter tuning in DRFS.
DRFS regulates the strength of class-specific and global regularizations through the normalization coefficients 1 r and 1 r ( r 1 ) , respectively. To further investigate the effect of these regularization terms on model performance, we introduce two additional regularization parameters, λ 1 and λ 2 , as represented in the following optimization formulation:
min W 0 XW Y F 2 + α W 2 , 1 + λ 1 r Ω l o c a l ( W ) + λ 2 r ( r 1 ) Ω g l o b a l ( W ) .
Figure 7 illustrates the influence of the parameters λ 1 and λ 2 on DRFS performance, where we set α = 1 and the number of selected features to 50 for consistency. The results indicate that DRFS exhibits a general insensitivity to the variations in these parameters across most datasets, except for the SRBCT and GLIOMA datasets. Specifically, a distinct pattern emerges on the SRBCT dataset, where the model’s performance improves as λ 1 decreases and λ 2 increases. In contrast, on the GLIOMA dataset, larger values of λ 1 yield better performance. These observations suggest that the nature of intra-class feature associations and inter-class feature redundancy varies across different datasets. Since DRFS exhibits low sensitivity to λ 1 and λ 2 on most datasets, we recommend using the strategy in Equation (10) for efficient parameter adjustment.
Figure 8 illustrates the impact of the number of feature neighbors, k, on classification performance across all datasets. The parameter k controls the neighborhood size used to calculate feature similarities, playing a critical role in representing feature associations. As shown in Figure 8, the effect of k on classification accuracy is closely related to the number of selected features. Specifically, the impact of k is more pronounced when fewer features are selected. For example, on the Isolet, mfeat, Yale, and warpPIE10P datasets, when only 10 features are selected, accuracy decreases as k increases. This suggests that with smaller feature sets, a larger neighborhood may introduce more feature associations affecting model, which can degrade model performance. Moreover, we observe that the SRBCT and GLIOMA datasets exhibit higher sensitivity to variations in k. This may be attributed to the inherent complexity of these datasets, which could lead to more intricate feature interactions. As evidenced by the parameter sensitivity analysis in Figure 6 and Figure 7, these datasets demonstrate more noticeable variations in performance. In this study, k is uniformly set to 10 to ensure consistency in capturing feature associations across all datasets.

5.6. Convergence Study

In Section 4.3, we provided a theoretical analysis of the convergence of Algorithm 1. In this subsection, we further investigate the convergence behavior of the objective function in (12) across iterations. Understanding the convergence of the DRFS algorithm is crucial to ensure its efficiency and stability when optimizing the feature selection process. Figure 9 shows the convergence curve of DRFS for various datasets. As shown, DRFS exhibits consistent and steady convergence, with the objective function value decreasing progressively until stabilization. In most cases, the algorithm reaches a stable solution within 30 iterations, highlighting its computational efficiency.

6. Conclusions

In this paper, we addressed the limitations of existing feature selection methods focusing solely on global feature associations. We highlighted the importance of incorporating class-specific feature interactions, which are essential for capturing patterns unique to individual classes. By proposing the Dual-Regularized Feature Selection (DRFS) method, we introduced a novel approach that simultaneously considers both class-specific and global feature associations. Our dual-regularization strategy ensures that important class-specific features are preserved while minimizing redundant global features. Importantly, our regularization technique is designed to function as an adaptable plugin, compatible with various feature selection models. The experimental results demonstrated that DRFS outperforms existing methods in classification tasks, offering a more effective and robust solution for feature selection.
However, our method has some limitations. Due to the fine-grained consideration of feature interactions, DRFS faces challenges in terms of storage and computation when dealing with very high dimensional features. Additionally, since our model is a shallow model, it may struggle to capture highly complex, non-linear relationships between features. In future work, we plan to address these challenges by optimizing the algorithm for high-dimensional data and exploring its integration with deep learning models to better capture non-linear interactions.

Author Contributions

Conceptualization, methodology, writing—original draft preparation, C.W.; validation, writing—review and editing, J.W. (Jun Wang), Y.L., C.P. and J.W. (Jinmao Wei); resources, J.W. (Jinmao Wei); supervision, J.W. (Jun Wang) and J.W. (Jinmao Wei); funding acquisition, J.W. (Jun Wang) and J.W. (Jinmao Wei). All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Key R&D Programs of China (2021YFC2100800, 2021YFC2100801), CAAI-MindSpore Open Fund, the National Natural Science Foundation of China (62106091), and Shandong Provincial Natural Science Foundation (ZR2021MF054, ZR2022MF247).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in scikit-feature selection repository (https://jundongl.github.io/scikit-feature/datasets.html, accessed on 13 January 2025) or UCI machine learning repository (https://archive.ics.uci.edu/ml/datasets, accessed on 13 January 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2018, 50, 94. [Google Scholar] [CrossRef]
  2. Yue, H.; Li, J.; Liu, H. Second-Order Unsupervised Feature Selection via Knowledge Contrastive Distillation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 15577–15587. [Google Scholar] [CrossRef]
  3. Xu, Y.; Wang, J.; Wei, J. To avoid the pitfall of missing labels in feature selection: A generative model gives the answer. In Proceedings of the Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 6534–6541. [Google Scholar]
  4. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef]
  5. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification; Wiley: Hoboken, NJ, USA, 2012. [Google Scholar]
  6. Yamada, M.; Tang, J.; Lugo-Martinez, J.; Hodzic, E.; Shrestha, R.; Saha, A.; Ouyang, H.; Yin, D.; Mamitsuka, H.; Sahinalp, C.; et al. Ultra high-dimensional nonlinear feature selection for big biological data. IEEE Trans. Knowl. Data Eng. 2018, 30, 1352–1365. [Google Scholar] [CrossRef]
  7. Ang, J.C.; Mirzal, A.; Haron, H.; Hamed, H.N.A. Supervised, unsupervised, and semi-supervised feature selection: A review on gene selection. IEEE/ACM Trans. Comput. Biol. Bioinform. 2015, 13, 971–989. [Google Scholar] [CrossRef]
  8. Zhang, X.; Xu, M.; Zhou, X. RealNet: A feature selection network with realistic synthetic anomaly for anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16699–16708. [Google Scholar]
  9. Zhang, C.; Fang, Y.; Liang, X.; Wu, X.; Jiang, B. Efficient multi-view unsupervised feature selection with adaptive structure learning and inference. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24), Jeju, Republic of Korea, 3–9 August 2024; pp. 5443–5452. [Google Scholar]
  10. Zhang, R.; Nie, F.; Li, X.; Wei, X. Feature selection with multi-view data: A survey. Inf. Fusion 2019, 50, 158–167. [Google Scholar] [CrossRef]
  11. Wang, Y.; Zhao, X.; Xu, T.; Wu, X. Autofield: Automating feature selection in deep recommender systems. In Proceedings of the ACM Web Conference, Lyon, France, 25–29 April 2022; pp. 1977–1986. [Google Scholar]
  12. Cohen, D.; Shnitzer, T.; Kluger, Y.; Talmon, R. Few-Sample Feature Selection via Feature Manifold Learning. In Proceedings of the International Conference on Machine Learning (ICML’23), Zhuhai, China, 17–20 February 2023. [Google Scholar]
  13. Wang, C.; Wang, J.; Gu, Z.; Wei, J.M.; Liu, J. Unsupervised feature selection by learning exponential weights. Pattern Recogn. 2024, 148, 110183. [Google Scholar] [CrossRef]
  14. Wang, J.; Wei, J.M.; Yang, Z.; Wang, S.Q. Feature selection by maximizing independent classification information. IEEE Trans. Knowl. Data Eng. 2017, 29, 828–841. [Google Scholar] [CrossRef]
  15. Chen, S.B.; Ding, C.; Luo, B.; Xie, Y. Uncorrelated lasso. In Proceedings of the AAAI Conference on Artificial Intelligence, Bellevue, WA, USA, 14–18 July 2013; Volume 27, pp. 166–172. [Google Scholar]
  16. Xu, X.; Wu, X.; Wei, F.; Zhong, W.; Nie, F. A general framework for feature selection under orthogonal regression with global redundancy minimization. IEEE Trans. Knowl. Data Eng. 2021, 34, 5056–5069. [Google Scholar] [CrossRef]
  17. Shang, R.; Wang, W.; Stolkin, R.; Jiao, L. Non-negative spectral learning and sparse regression-based dual-graph regularized feature selection. IEEE Trans. Cybern. 2017, 48, 793–806. [Google Scholar] [CrossRef]
  18. Shang, R.; Zhang, W.; Lu, M.; Jiao, L.; Li, Y. Feature selection based on non-negative spectral feature learning and adaptive rank constraint. Knowl.-Based Syst. 2022, 236, 107749. [Google Scholar] [CrossRef]
  19. Xu, Y.; Wang, J.; An, S.; Wei, J.; Ruan, J. Semi-supervised multi-label feature selection by preserving feature-label space consistency. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino, Italy, 22–26 October 2018; pp. 783–792. [Google Scholar]
  20. Brezočnik, L.; Fister, I., Jr.; Podgorelec, V. Swarm intelligence algorithms for feature selection: A review. Appl. Sci. 2018, 8, 1521. [Google Scholar] [CrossRef]
  21. Shang, W.; Huang, H.; Zhu, H.; Lin, Y.; Qu, Y.; Wang, Z. A novel feature selection algorithm for text categorization. Expert Syst. Appl. 2007, 33, 1–5. [Google Scholar] [CrossRef]
  22. El Aboudi, N.; Benhlima, L. Review on wrapper feature selection approaches. In Proceedings of the 2016 International Conference on Engineering & MIS (ICEMIS), Agadir, Morocco, 22–24 September 2016; pp. 1–5. [Google Scholar]
  23. Maldonado, J.; Riff, M.C.; Neveu, B. A review of recent approaches on wrapper feature selection for intrusion detection. Expert Syst. Appl. 2022, 198, 116822. [Google Scholar] [CrossRef]
  24. Yan, H.; Yang, J.; Yang, J. Robust joint feature weights learning framework. IEEE Trans. Knowl. Data Eng. 2016, 28, 1327–1339. [Google Scholar] [CrossRef]
  25. Nie, F.; Huang, H.; Cai, X.; Ding, C. Efficient and robust feature selection via joint 2,1-norms minimization. In Proceedings of the Advances in Neural Information Processing Systems 23 (NIPS 2010), Vancouver, BC, Canada, 6–9 December 2010. [Google Scholar]
  26. Xu, S.; Dai, J.; Shi, H. Semi-supervised feature selection based on least square regression with redundancy minimization. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar]
  27. Zhang, Z.; Tian, Y.; Bai, L.; Xiahou, J.; Hancock, E. High-order covariate interacted Lasso for feature selection. Pattern Recognit. Lett. 2017, 87, 139–146. [Google Scholar] [CrossRef]
  28. Cui, L.; Bai, L.; Zhang, Z.; Wang, Y.; Hancock, E.R. Identifying the most informative features using a structurally interacting elastic net. Neurocomputing 2019, 336, 13–26. [Google Scholar] [CrossRef]
  29. Cui, L.; Bai, L.; Wang, Y.; Philip, S.Y.; Hancock, E.R. Fused lasso for feature selection using structural information. Pattern Recogn. 2021, 119, 108058. [Google Scholar] [CrossRef]
  30. Lai, J.; Chen, H.; Li, T.; Yang, X. Adaptive graph learning for semi-supervised feature selection with redundancy minimization. Inf. Sci. 2022, 609, 465–488. [Google Scholar] [CrossRef]
  31. Roffo, G.; Melzi, S.; Castellani, U.; Vinciarelli, A.; Cristani, M. Infinite Feature Selection: A Graph-based Feature Filtering Approach. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4396–4410. [Google Scholar] [CrossRef]
  32. Bertsekas, D.P. Constrained Optimization and Lagrange Multiplier Methods; Academic Press: Cambridge, MA, USA, 2014. [Google Scholar]
  33. Garber, M.E.; Troyanskaya, O.G.; Schluens, K.; Petersen, S.; Thaesler, Z.; Pacyna-Gengelbach, M.; Van De Rijn, M.; Rosen, G.D.; Perou, C.M.; Whyte, R.I.; et al. Diversity of gene expression in adenocarcinoma of the lung. Proc. Natl. Acad. Sci. USA 2001, 98, 13784–13789. [Google Scholar] [CrossRef]
  34. Khan, J.; Wei, J.S.; Ringner, M.; Saal, L.H.; Ladanyi, M.; Westermann, F.; Berthold, F.; Schwab, M.; Antonescu, C.R.; Peterson, C.; et al. Classification and diagnostic prediction of cancers using gene expression profiling and artificial neural networks. Nat. Med. 2001, 7, 673–679. [Google Scholar] [CrossRef]
  35. Singh, D.; Febbo, P.G.; Ross, K.; Jackson, D.G.; Manola, J.; Ladd, C.; Tamayo, P.; Renshaw, A.A.; D’Amico, A.V.; Richie, J.P.; et al. Gene expression correlates of clinical prostate cancer behavior. Cancer Cell 2002, 1, 203–209. [Google Scholar] [CrossRef]
  36. Fanty, M.; Cole, R. Spoken letter recognition. Adv. Neural Inf. Process. Syst. 1990, 3, 220–226. [Google Scholar]
  37. Breukelen, M.V.; Duin, R.P.W.; Tax, D.M.J.; Hartog, J.E.D. Handwritten Digit Recognition by Combined Classifiers. Kybernetika 1998, 34, 381–386. [Google Scholar]
  38. Cai, D.; Zhang, C.; He, X. Unsupervised feature selection for multi-cluster data. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 24–28 July 2010; pp. 333–342. [Google Scholar]
  39. Ming, D.; Ding, C. Robust flexible feature selection via exclusive L21 regularization. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3158–3164. [Google Scholar]
Figure 1. Comparison of global feature association-based methods (a) and our approach (b). (a) Existing methods compute feature associations across the entire dataset to preserve the global feature manifold or eliminate redundant features. (b) Our method retains class-specific feature manifolds while removing global feature redundancies.
Figure 1. Comparison of global feature association-based methods (a) and our approach (b). (a) Existing methods compute feature associations across the entire dataset to preserve the global feature manifold or eliminate redundant features. (b) Our method retains class-specific feature manifolds while removing global feature redundancies.
Entropy 27 00190 g001
Figure 2. An overview of the proposed DRFS method.
Figure 2. An overview of the proposed DRFS method.
Entropy 27 00190 g002
Figure 3. Classification accuracy of the SVM classifier with different number of selected features on eight datasets.
Figure 3. Classification accuracy of the SVM classifier with different number of selected features on eight datasets.
Entropy 27 00190 g003
Figure 4. Classification accuracy of the 1-NN classifier with different number of selected features on eight datasets.
Figure 4. Classification accuracy of the 1-NN classifier with different number of selected features on eight datasets.
Entropy 27 00190 g004
Figure 5. Critical difference diagram by the Bonferroni–Dunn post hoc test (significance level of 0.05).
Figure 5. Critical difference diagram by the Bonferroni–Dunn post hoc test (significance level of 0.05).
Entropy 27 00190 g005
Figure 6. Parameter sensitivity study of DRFS with respect to α and β across all datasets. The parameters α and β correspond to the sparse norm and dual regularizations, respectively.
Figure 6. Parameter sensitivity study of DRFS with respect to α and β across all datasets. The parameters α and β correspond to the sparse norm and dual regularizations, respectively.
Entropy 27 00190 g006
Figure 7. Parameter sensitivity study of the class-specific and global regularizations across all datasets. The parameters λ 1 and λ 2 correspond to the class-specific and global regularizations, respectively.
Figure 7. Parameter sensitivity study of the class-specific and global regularizations across all datasets. The parameters λ 1 and λ 2 correspond to the class-specific and global regularizations, respectively.
Entropy 27 00190 g007
Figure 8. Accuracy of DRFS with varying feature neighbors k. “# Feas” denotes the number of selected features.
Figure 8. Accuracy of DRFS with varying feature neighbors k. “# Feas” denotes the number of selected features.
Entropy 27 00190 g008
Figure 9. Convergence curves of DRFS across all datasets.
Figure 9. Convergence curves of DRFS across all datasets.
Entropy 27 00190 g009
Table 1. Summary of key feature similarity-based FS methods. For ‘Task’, ‘S’ for supervised, ‘SE’ for semi-supervised, and ‘U’ for unsupervised.
Table 1. Summary of key feature similarity-based FS methods. For ‘Task’, ‘S’ for supervised, ‘SE’ for semi-supervised, and ‘U’ for unsupervised.
MethodGlobal
Association
Class-Specific
Association
Similarity
Measure
Task
UnLasso [15]×Square cosine similarityS
GRMOR [16]×Square cosine similarityS
InteractedLasso [27]×HypergraphS
InElasticNet [28]×Information theoryS
InFusedLasso [29]×Information theoryS
SFSRM_MI [26]×Information theorySE
SFSRM_P [26]×Pearson correlation coefficientSE
AGLRM [30]×Gaussian functionSE
NSSRD [17]×Gaussian functionU
NSSRD_PF [17]×Parameter-free methodsU
NNSAFS [18]×Parameter-free methodsU
ManiFeSt [12]×Gaussian functionS
Inf-FS [31]×Weighted strategyS/U
DRFS (ours)Gaussian functionS
Table 2. Dataset description.
Table 2. Dataset description.
Dataset# Features# Samples# ClassesClass DistributionType
lung_discrete325737[6, 5, 5, 16, 7, 13, 21]Bioinformatics
Isolet61715602660 samples per classSpoken Letter
mfeat649200010200 samples per classImage
Yale10241651511 samples per classImage
warpPIE10P24202101021 samples per classImage
SRBCT2308834[29, 11, 18, 25]Bioinformatics
LUNG33122035[139, 17, 21, 20, 6]Bioinformatics
GLIOMA4434504[14, 7, 14, 15]Bioinformatics
Table 3. Comparison of optimal classification accuracy across eight datasets using the SVM classifier. The best-performing method for each dataset is highlighted in bold, while the second-best result is underlined.
Table 3. Comparison of optimal classification accuracy across eight datasets using the SVM classifier. The best-performing method for each dataset is highlighted in bold, while the second-best result is underlined.
  Dataset  AllFeaFisher
[5]
Gini Index
[21]
RelieF
[4]
RJFWL
[24]
GRMOR
[16]
Inf-FS
[31]
ManiFeSt
[12]
DRFS
(ours)
lung_discrete0.85140.93330.79430.87710.89240.89140.89240.84950.9448
Isolet0.96350.83850.76600.79100.86220.89940.91410.79550.9179
mfeat0.98700.98750.96350.97900.98500.98200.97600.94800.9870
Yale0.75150.76360.56970.68480.66060.71520.65450.42420.7697
warpPIE10P1.00000.97620.94290.96670.99521.00000.98101.00001.0000
SRBCT1.00001.00001.00000.96471.00000.96471.00000.77281.0000
LUNG0.96590.95100.93600.97050.94610.96570.96070.85720.9607
GLIOMA0.76000.82000.88000.82000.86000.88000.80000.64000.9200
Table 4. Comparison of optimal classification accuracy across eight datasets using the 1-NN classifier. The best-performing method of each dataset is highlighted in bold, while the second-best result is underlined.
Table 4. Comparison of optimal classification accuracy across eight datasets using the 1-NN classifier. The best-performing method of each dataset is highlighted in bold, while the second-best result is underlined.
  Dataset  AllFeaFisher
[5]
Gini Index
[21]
RelieF
[4]
RJFWL
[24]
GRMOR
[16]
Inf-FS
[31]
ManiFeSt
[12]
DRFS
(ours)
lung_discrete0.83620.88950.76760.87810.83710.84950.85050.80860.9171
Isolet0.86860.80380.66670.67370.75830.81670.83140.63530.8718
mfeat0.98100.98450.93150.95900.97750.97950.97850.91250.9855
Yale0.66060.69090.56970.57580.60000.67880.61210.35150.6545
warpPIE10P0.95710.97620.85710.96670.99050.97620.97140.94291.0000
SRBCT0.86841.00000.97650.89261.00000.92940.94040.63971.0000
LUNG0.95100.94600.90160.93120.93130.95570.96090.82300.9559
GLIOMA0.78000.80000.84000.80000.82000.84000.82000.50000.8600
Table 5. Friedman test statistics F F and corresponding p-value for each metric. The critical value is 14.07 at a significance level of α = 0.05 .
Table 5. Friedman test statistics F F and corresponding p-value for each metric. The critical value is 14.07 at a significance level of α = 0.05 .
Metric F F p-Value
SVM25.57 6.02 × 10 4
1-NN37.49 3.78 × 10 6
Table 6. Ablation study showing average SVM classification performance with selected features in the range [ 10 ,   20 ,   ,   100 ] .
Table 6. Ablation study showing average SVM classification performance with selected features in the range [ 10 ,   20 ,   ,   100 ] .
ClassifierRegularizationModules
Class-Specific
Global
××
××
SVMlung_discrete0.83800.91150.91710.9226
Isolet0.76830.83100.83180.8320
mfeat0.97290.97520.97660.9766
Yale0.57940.69820.70790.7091
warpPIE10P0.97900.98140.98710.9881
SRBCT0.99530.99650.99761.0000
LUNG0.92480.92570.93310.9346
GLIOMA0.82400.84600.86200.8740
Table 7. Ablation study showing average 1-NN classification performance with selected features in the range [ 10 ,   20 ,   ,   100 ] .
Table 7. Ablation study showing average 1-NN classification performance with selected features in the range [ 10 ,   20 ,   ,   100 ] .
ClassifierRegularizationModules
Class-Specific
Global
××
××
1-NNlung_discrete0.80200.87700.88950.8937
Isolet0.68860.77440.77850.7790
mfeat0.96660.97580.97800.9780
Yale0.51030.60910.61090.6164
warpPIE10P0.97100.98240.98290.9838
SRBCT0.99290.99880.99881.0000
LUNG0.90070.89130.92920.9307
GLIOMA0.78600.78000.83200.8500
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, C.; Wang, J.; Li, Y.; Piao, C.; Wei, J. Dual-Regularized Feature Selection for Class-Specific and Global Feature Associations. Entropy 2025, 27, 190. https://doi.org/10.3390/e27020190

AMA Style

Wang C, Wang J, Li Y, Piao C, Wei J. Dual-Regularized Feature Selection for Class-Specific and Global Feature Associations. Entropy. 2025; 27(2):190. https://doi.org/10.3390/e27020190

Chicago/Turabian Style

Wang, Chenchen, Jun Wang, Yanfei Li, Chengkai Piao, and Jinmao Wei. 2025. "Dual-Regularized Feature Selection for Class-Specific and Global Feature Associations" Entropy 27, no. 2: 190. https://doi.org/10.3390/e27020190

APA Style

Wang, C., Wang, J., Li, Y., Piao, C., & Wei, J. (2025). Dual-Regularized Feature Selection for Class-Specific and Global Feature Associations. Entropy, 27(2), 190. https://doi.org/10.3390/e27020190

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop