Next Article in Journal
Bioactivation of Isoxazole-Containing Bromodomain and Extra-Terminal Domain (BET) Inhibitors
Previous Article in Journal
An Improvement of SPME-Based Sampling Technique to Collect Volatile Organic Compounds from Quercus ilex at the Environmental Level
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Ranking and Screening for Class-Imbalanced Metabolomics Data Based on Rank Aggregation Coupled with Re-Balance

1
School of Science, Kunming University of Science and Technology, Kunming 650500, China
2
Faculty of Agriculture and Food, Kunming University of Science and Technology, Kunming 650500, China
*
Authors to whom correspondence should be addressed.
Metabolites 2021, 11(6), 389; https://doi.org/10.3390/metabo11060389
Submission received: 9 April 2021 / Revised: 31 May 2021 / Accepted: 8 June 2021 / Published: 14 June 2021
(This article belongs to the Section Bioinformatics and Data Analysis)

Abstract

:
Feature screening is an important and challenging topic in current class-imbalance learning. Most of the existing feature screening algorithms in class-imbalance learning are based on filtering techniques. However, the variable rankings obtained by various filtering techniques are generally different, and this inconsistency among different variable ranking methods is usually ignored in practice. To address this problem, we propose a simple strategy called rank aggregation with re-balance (RAR) for finding key variables from class-imbalanced data. RAR fuses each rank to generate a synthetic rank that takes every ranking into account. The class-imbalanced data are modified via different re-sampling procedures, and RAR is performed in this balanced situation. Five class-imbalanced real datasets and their re-balanced ones are employed to test the RAR’s performance, and RAR is compared with several popular feature screening methods. The result shows that RAR is highly competitive and almost better than single filtering screening in terms of several assessing metrics. Performing re-balanced pretreatment is hugely effective in rank aggregation when the data are class-imbalanced.

1. Introduction

Datasets with imbalanced distribution are quite common in classification. In the settings of binary category, a dataset is called “imbalanced” if the number of one class is far larger than the others in the training data. Generally, the majority class is called negative while the minority class is called positive. Thus, the number of positive instances is often much lower than that of negative ones.
A hindrance in class-imbalance learning is that standard classifiers are often biased towards the majority classes. Therefore, there is a higher misclassification rate in the minority instances [1,2]. Re-sampling is the standard strategy to deal with class-imbalance learning tasks. Many studies [2,3,4] have shown that re-sampling the dataset is an effective way to enhance the overall performance of the classification for several types of classifiers. Re-sampling methods concentrate on modifying the training set to make it suitable for a standard classifier. There are generally three types of re-sampling strategies to balance the class distribution: over-sampling, under-sampling, and hybrid sampling.
  • Over-sampling adds a set sampled from the minority class. Randomly duplicating the minority instances, SMOTE [5] and smoothed bootstrap [6] are three widely used over-sampling methods.
  • Under-sampling removes some of the data points from the majority class to alleviate the harms of imbalanced distribution. Random under-sampling (RUS) is a simple but effective way to randomly remove part of the majority class.
  • Hybrid-sampling is a combination of over-sampling and under-sampling.
Let D be a dataset with p features x 1 , x 2 , , x p , the target of feature screening is to extract a part of features x 1 , x 2 , , x m such that m < < p and these selected features satisfy the specified conditions of the task at hand [7]. For instance, the target is to select the subset of candidate features to maximize classifier accuracy in a classification setting. In the past two decades, many papers in studies have adopted the feature screening methods [8,9,10]. Feature screening has many advantages such as reducing susceptibility to over-fitting, training models faster and offsetting the pernicious effects of the curse of dimensionality [8]. The disadvantage of feature screening is that some crucial features may be omitted, thus harming classification performance.
Filtering [11], wrapping [12], and embedding [13] are three kinds of approaches for feature screening. Filter algorithms screen top-ranked variables via a certain metric. Wrapper methods perform a search in all the combinations to find the best subsets of all features. Generally, a complete search is often time-consuming and greedy, so the heuristic technique is frequently utilized to explore the solutions. Embedded algorithms screen important variables while building the classifier. Of all the three types of feature screening, filter methods are the simplest and the most frequently used to solve real-world imbalanced problems [14] in class-imbalance learning community. Many metrics have been utilized to perform filtering feature screening algorithms, such as t test, Fisher score [15], Hellinger distance [16], Relief [17], ReliefF [18], information gain [19], Gini index [20], A U C R O C [21], A U C P R C [22], geometric mean [23], F-measure [24], and R-value [25].
Ensemble feature selection has been widely applied to the field of classification [26], such as Nazrul et al. [27] provided an ensemble feature selection method using feature–class and feature mutual information to select an optimal subset of features by combining multiple subsets of features. Yang et al. [28] proposed an ensemble-based wrapper approach for feature selection from data with highly imbalanced class distribution. Nowadays, feature selection methods are popular in metabolomics data analysis. In order to resolve the problem of filtering the discriminative metabolites from high-dimension metabolomics data, Lin et al. [29] proposed a mutual information (MI)-SVM-RFE method that filters out noise and non-informative variables by means of artificial variables and MI, then conducts SVM-RFE to select the most discriminative features. Fu et al. [30] proposed two feature selection algorithms that, by minimizing the overlap degree between the majority and the minority, are effective in recognizing key features and control false discoveries for class-imbalanced metabolomics data. The above feature screening methods are usually established for balanced datasets, but they are also directly utilized in class-imbalance situations.
Different filtered approaches give different feature rankings because of their different theories, even when just counting top-ranked features. Motivated by this problem, we propose a simple strategy called rank aggregation with re-balance (RAR) to combine all methods’ ranking results in this study. It is an essential tool to fuse each rank to generate a synthetic rank that takes every ranking into account for class-imbalanced data. Different from the general feature selection methods, the proposed method combines different feature selection methods rather than simply accepting the result of one method, which can enhance the stability of the algorithm. At the same time, the great performances of the experiments in balanced and imbalanced metabolomics datasets verify the strong generalization abilities of RAR.

2. Results

2.1. Kendall’s τ Rank Correlation of Eight Filtering Methods on Class-Imbalanced Data

Each filtered method above can be employed to perform feature screening. However, we noted that different filtering feature screening techniques may give different rankings, especially when the data are extremely class-imbalanced. In this section, we compare methods using Kendall’s τ rank correlation [31].
The Kendall’s τ rank correlation of eight filtering methods (t test, Fisher score, Hellinger distance, Relief, ReliefF, Information gain, Gini index, and R-value) are computed with simulated data that are generated by multivariate normal distributions, namely, X | ( y = 0 ) N p ( μ 0 , Σ ) and X | ( y = 1 ) N p ( μ 1 , Σ ) , where the label y = 0 denotes the majority class and y = 1 minority class, respectively. The predictors in two classes have the same covariance matrix Σ , which is set to be a unit matrix for the purpose of simplicity. Two cases are considered in this study. In case one, the number of p = 8 , and eight variables are all set to be key features. The difference of mean values μ 1 μ 0 = [ 2.4 , 2.2 , 2 , 1.8 , 1.6 , 1.4 , 1.2 , 1 ] . In case two, p = 16 , and the first eight variables are set to be the same with case one, but another eight irrelevant predictors are added. The number of total instances is set to 960. The negative to the positive ratios here are set be 1:1, 3:1, 9:1, 31:1, and 95:1, respectively. There are 28 Kendall’s τ rank correlation coefficients among 8 filtering methods, and the mean of these coefficients (with 100 repeats) is shown in Figure 1. As stated above, τ = 1 if all pairs are concordant. Whereas the maximum of τ is 0.88 in case one (left, Figure 1) where there are no irrelevant predictors, and 0.76 in case two (right, Figure 1) where one-half of features are irrelevant variables. The two maximal τ values are reached when two classes are exactly balanced, and τ reduces as the imbalance ratio increases in two cases. It indicates that these filtering methods probably generate different feature rankings, and such differences tend to be intensified when the class imbalanced ratio increases. Consequently, it is hard to say that a filtering approach is better or worse than another one, and it is a big risk to just depend on a single filter algorithm to make decisions. We have known that such a difference will occur due to the different principles of the filtering methods, but we also presume that class imbalance intensifies this difference. A natural way to combat this challenge may combine each filtering approach’s information and relieve the effect of class imbalance. This is the motivation for why we propose the strategy of rank aggregation with re-balance.

2.2. Rank Aggregation(RA) on Original Balanced Data

In our computation, eight filtering methods—t test, Fisher score, Hellinger distance, Relief, ReliefF, Information gain, Gini index, and R-value—are aggregated to generate an incorporative rank. Rank aggregation is firstly tested with the original balanced dataset “NPC”. Artificial rebalancing is unnecessary, and just case 1 (no resampling) is performed. Rank aggregation is compared with eight filtering methods: t test, Fisher score, Hellinger distance, Relief, ReliefF, Information gain, Gini index, and R-value. G m e a n , F 1 , A U C R O C and A U C P R C are utilized as evaluation measurements. The rank lists ordered by their importance are shown on the x-axis in Figure 2. The top seven features are selected according to all four assessment metrics.

2.3. Rank Aggregation with Re-Balance (RAR) on Imbalanced Data

Figure 3, Figure 4, Figure 5 and Figure 6 show the aggregated rank lists on seven cases with the datasets “TBI”, “CHD2-1”, “CHD2-2”, and “ATR”, respectively. Rank aggregation combines each ranking into a list reflective of the overall preference, and each subgraph of four figures shows the aggregation results based on the CE algorithm. The x-axis is the optimal list obtained by the rank aggregation algorithm. The y-axis also ranks, and the gray line is the rank of the original data; the black line is their average rank; and the red line is the aggregate result of the CE algorithm. The order of the x-axis rank is based on the aggregate ranks obtained by the red line. The performances measured by G m e a n , F 1 , A U C R O C , and A U C P R C are given in Table 1, Table 2, Table 3 and Table 4, respectively.

3. Discussion

Table 1 and Table 2 show that RA reached the maximal values of G m e a n and F 1 . It can be seen from Table 3 and Table 4 that RA and ReliefF obtained the maximal values of A U C R O C and A U C P R C . Therefore, RA outperformed single filtering methods when assessed with G m e a n , F 1 , A U C R O C , and A U C P R C . The NPC dataset had a completely balanced distribution, and RA worked well on it. Thus, rank aggregation is necessary to integrate different results, even if in a totally balanced situation, and a consensual feature ranking list is provided.
Aggregation ranking lists in Figure 3, Figure 4, Figure 5 and Figure 6 tell us the order of importance of each feature. Though the rank lists derived from different subsampling methods were not the same, the top features were approximately consistent. After obtaining the rank list, another task is to figure out how many features should be considered as key variables. In this computation, we performed 5-fold cross-validation [32] to find the optimal number of key features. As recent studies have showed that A U C P R C is more informative in imbalanced learning [32,33], A U C P R C was employed as perfomance metric in this section, and random forest classifier was utilized to implement classification. Namely, the value of A U C P R C was calculated, as the top k ranked features were used each time, where k varies from 1 to p (see Figure 7, Figure 8, Figure 9 and Figure 10). We chose the optimal k- value such that the random forest classifier had the maximal A U C P R C in identifying classification. It can be seen from Table 1, Table 2, Table 3 and Table 4 that the optimal number of important features varied greatly under different re-balanced strategies. One possible reason is that the artificial data generated by different subsampling have difference to some extent. Another possible reason is the measurement changes sightly as the number of candidate features changes. This seems to be true from the Figure 7, Figure 8, Figure 9 and Figure 10 where each curve tends to be flat as the changes in the number of features used for classification. It also noted that the A U C P R C under no re-sampling (case 1) was generally lower than that under six re-sampling methods.
Table 1, Table 2, Table 3 and Table 4 report the results of these real datasets with the assessing metrics G m e a n , F 1 , A U C R O C , and A U C P R C , respectively. We can perform comparisons in several aspects. Original imbalanced datasets are employed in case 1 from Table 1, Table 2, Table 3 and Table 4 (except NPC dataset). Of all the 16 “no re-balance” situations, the aggregation rank method reached the maximal measures in 12 situations compared with the other 8 filtering methods (t test, Fisher score, Hellinger distance, Relief, ReliefF, information gain, Gini, and R-value). It indicates that aggregation rank was better than a single filtering rank with the proportion of 75.00% when the data are class-imbalanced. If the original dataset NPC was counted in, this proportion was 80.00%. Therefore, rank aggregation is generally superior to single filtering methods, no matter how the data are balanced or imbalanced.
Re-balanced datasets were artificially generated and utilized in cases 2–7 from Table 1, Table 2, Table 3 and Table 4. Of all the 96 scenarios with re-balance, the aggregation rank method reached the maximal measures in 83 scenarios compared to the other eight filtering methods. It means that aggregation rank outperformed single filtering rank with the proportion of 86.46% when the class-imbalanced data were treated with re-balance strategies. Thus, performing aggregation rank is extremely effective in dealing with class-imbalanced data.
Rank aggregation was performed on both imbalanced datasets (RA) and re-balanced datasets (RAR). Of all the 96 scenarios with re-balance (cases 2–7 in Table 1, Table 2, Table 3 and Table 4), there were 93 situations whose measurements were equal or greater than those from case 1 (no re-balance). It shows that aggregation rank with re-balance strategies performed better with the proportion of 96.88% than that with original class-imbalanced data. Therefore, performing re-balance can play a crucial role in improving the performance of rank aggregation when the data are class-imbalanced.
Figure 7, Figure 8, Figure 9 and Figure 10 show the A U C P R C curves of seven cases on four imbalanced datasets. A U C P R C from re-balanced data (cases 2–7) was generally higher than that from imbalanced data (case 1). In other words, the performance can be promoted after re-sampling to balance the imbalanced data artificially. Case 5 and case 6 are two under-sampling methods, and the A U C P R C was generally lower than that from over-sampling or hybrid sampling (cases 2–4). The possible reason is that some of the useful information is missed in doing under-sampling when the size of the minority instances is too small (see Table 5). Therefore, one should be cautious about using under-sampling in practice.
In sum, different filter methods generate different rankings. Rank aggregation is necessary to integrate different results and provide a consensual feature ranking list. Class-imbalance usually leads to degraded performance from a filtering method on feature importance ranking. This harmfulness can be alleviated via different re-balance strategies in sample space.

4. Materials and Methods

4.1. Notations

The notations used in this study are listed below:
Da dataset with two classes C 1 and C 2
C 1 the minority (positive) class
C 2 the majority (negative) class
nthe size of the total instances in D
pthe number of the features in D
n k the size of C k , k = 1 , 2
| D | the number of samples in D
x j the jth feature, j = 1 , 2 , , p
x i the ith instance, i = 1 , 2 , , n
μ k j the expectation of jth feature in C k , k = 1 , 2 ; j = 1 , 2 , , p
σ k j 2 the variance of jth feature in C k , k = 1 , 2 ; j = 1 , 2 , , p
x ¯ k j the sample mean of jth feature in C k , k = 1 , 2 ; j = 1 , 2 , , p
s k j 2 the sample variance of jth feature in C k , k = 1 , 2 ; j = 1 , 2 , , p

4.2. Eight Filtering Methods

4.2.1. t Test

Feature screening using the t test statistic [34] is similar to performing a hypothesis test (the null hypothesis is that there is no difference in the means) on the class’s distribution, and its significance indicates the difference between majority and minority classes. The lower the p value of this t test, the higher the majority and minority classes’ significant difference. Consequently, the considered feature is more relevant to the separation of two classes.

4.2.2. Fisher Score

Fisher score [35] is simple and generally quite effective, which can be a criterion of feature screening. Fisher score of a single feature is defined as follows:
F i s h e r ( x j ) = μ 1 j μ 2 j σ 1 j 2 + σ 2 j 2 , j = 1 , 2 , , p ,
μ 1 j , μ 2 j , σ 2 j 2 , and σ 2 j 2 can be replaced by their corresponding sample statistics in computation, namely,
F i s h e r ( x j ) = x ¯ 1 j x ¯ 2 j s 1 j 2 + s 2 j 2 , j = 1 , 2 , , p
A feature with a large Fisher score is more crucial for discriminating the two categories.

4.2.3. Hellinger Distance

Hellinger distance can be used to measure a distributional divergence [36]. Denoted the two normal distributions by P and Q, Hellinger distance is calculated as follows:
D H 2 ( P , Q ) = 2 2 2 σ 1 σ 2 σ 1 2 + σ 2 2 e x p ( μ 1 μ 2 ) 2 4 ( σ 1 2 + σ 2 2 ) ,
where μ 1 , σ 1 2 , μ 2 , and σ 2 2 are the expectation and variance of P and Q, respectively, and their corresponding sample statistics are used in practice [37]. The larger the Hellinger distance is, the more divergent the two distributions are.

4.2.4. Relief and ReliefF

The Relief is an iteration method that tries to give each feature a score to indicate its level of relevance to the response [37,38]. Let x i be an instance; n e a r h i t i and n e a r m i s s i be its two nearest neighbors from the same class and the other class by the Euclidean distance, respectively. The score vector s = ( s 1 , s 2 , , s p ) T is refreshed as follows:
s j s j ( x i j n e a r h i t i j ) 2 + ( x i j n e a r m i s s i j ) 2 , j = 1 , 2 , , p ,
where x i j , n e a r h i t i j and n e a r m i s s i j are the jth element of x i , n e a r h i t i and n e a r m i s s i , respectively. A feature with a higher score is more crucial to the response. Though ReliefF [39] is originally developed for dealing with multi-class and noise datasets, it can be applied to binary classification cases. Compared with Relief that searches one nearest instance from the same class and one from the other class in updating the weights, ReliefF finds k nearest neighbors. Similarly, a feature with a higher score is more important to the response.

4.2.5. Information Gain (IG)

Information gain [40] is the measurement of informational theory and can be utilized to assess the importance of a given feature. In the settings of binary classification, the information entropy of the set D is defined as follows:
E n t ( D ) = k = 1 2 n k n l o g 2 n k n
Assuming that a discrete feature (attribute) x has V different values { x 1 , x 2 , , x V } , and D v is the subset of instance set D satisfying x = x v . The information gain of the variable x is
I G ( D , x ) = E n t ( D ) v = 1 V | D v | n E n t ( D v ) ,
The larger the information gain is, the more important the feature is for separating the classes. A continuous feature should be discretized before using the IG metric.

4.2.6. Gini Index

Gini index [41] fits binary digits, continuous numerical values, ordinal numbers, etc. It is a non-purity split method. Gini index of D is defined as follows:
G i n i ( D ) = 1 k = 1 2 p k 2 ,
where p k is the probability that any instance belongs to C k , and it is replaced with n k / n , ( k = 1 , 2 ) in practice. If we divide D into M subsets D 1 , D 2 , , D M , the Gini index after splitting is:
G i n i s p l i t ( D ) = m = 1 M | D m | n G i n i ( D m )
The smaller the Gini index is, the more important the feature is.

4.2.7. R-Value

R-value [30,42] indicates the degree of overlap for the class-imbalanced dataset. R-value for a dataset D is defined as follows:
R ( D ) = 1 n k = 1 2 m = 1 C k λ ( k N N P k m , D C k θ ) ,
where
λ ( x ) = 1 , i f x > 0 , 0 , o t h e r w i s e
where k N N ( P , C i ) is the subset of k nearest neighbors of instance P that belong to the set of instances C i , and θ is the threshold generally set to be k / 2 [43]. The smaller the R-value is, the more important the feature is for discriminating the categories.

4.3. Four Evaluation Metrics

4.3.1. Geometric Mean and F-Measure

True positive, true negative, false positive, and false negative are denoted by T P , T N , F P , and F N , respectively. Some common metrics are listed below:
T P R = r e c a l l = T P T P + F N ; T N R = T N T N + F P ; F P R = F P F P + T N ; p r e c i s i o n = T P T P + F P ; G m e a n = T P R × T N R ; F 1 = 2 p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
The range of both G m e a n and F 1 is [ 0 , 1 ] . The larger they are, the better the classifer works.

4.3.2. A U C R O C and A U C P R C

A U C R O C is the area under the receiver operating characteristic curve (ROC) [44]. A U C P R C is the area under the precision recall curve (PRC) [45]. Both A U C R O C and A U C P R C range from 0 to 1, and the larger they are, the better the classifier is built for the imbalanced learning. More details on A U C R O C or A U C P R C can be found in our previous studies [37,46].
G m e a n , F 1 , A U C R O C , and A U C P R C are more widely used than the metric A c c u r a c y in class-imbalance learning. These metrics actually pay more attention to the minority samples.

4.4. Kendall’S τ Rank Correlation

Kendall’s τ rank correlation statistic [47] can be applied to calculate the degree of comparability between the feature rankings of two filtering techniques. Let the two feature rankings generated by two filters be
f 1 : r 11 , r 21 , r 31 , , r p 1 , f 2 : r 12 , r 22 , r 32 , , r p 2 ,
and there are no ties in each of ranking list. Then Kendall’s τ is calculated as follows,
τ ( f 1 , f 2 ) = i < j s g n ( r i 1 r j 1 ) s g n ( r i 2 r j 2 ) p ( p 1 ) / 2 ,
where s g n ( x ) is the sign function, namely it equals 1 if x is positive and 1 if x negative. A pair of ( i , j ) is called concordant if r i 1 > r j 1 and r i 2 > r j 2 or r i 1 < r j 1 and r i 2 < r j 2 . Otherwise, they are considered discordant. The numerator i < j s g n ( r i 1 r j 1 ) s g n ( r i 2 r j 2 ) is the difference between the number of concordant pairs and the number of discordant pairs, and the denominator p ( p 1 ) / 2 is the number of all distinct pairs of p elements. The range of τ is [ 1 , 1 ] . If τ = 0 , the correlation of two rankings is weak; if τ = 1 , then all pairs will be discordant, and the two rankings are exactly opposite; if τ = 1 , then all pairs are exactly concordant [48].

4.5. Rank Aggregation with Re-Balance for Class-Imbalanced Data

As mentioned above, there are differences among the ranks from different filtering methods, but we assume that they are equal in match, namely, no one is better or worse than another. Rank aggregation (RA) is a greatly intuitive metric that computes the absolute differences between the ranks of all individual features [49]. Rank aggregation with re-balance (RAR) consists of two stages for class-imbalanced data and is illustrated in Figure 11. In sample space, the data are artificially balanced by generating new instances of the minority class or (and) removing some of the majority class instances. In feature space, m rank lists are first computed using m different filtering methods. Each rank list is the full permutation of all the features. Then, they are merged to be an aggregated rank. Feature screening and classification can be performed according to this aggregated rank.

4.5.1. Rank Aggregation

As mentioned above, different filter techniques will give different feature ranking results. The rank aggregation method [34,50] combines all the rankings together, by aggregating all feature ranking lists generated from different filtering methods.
RA is to find an optimal ranking δ * such that
δ * = arg min i = 1 K w i d ( δ , f i ) ,
where f i is the ith feature ranking list, δ represents a ranking list with the same length of f i , d is a distance function, and w i is the important weight related with list f i . In this study, d is chosen to be the Spearman’s foot rule distance [50]:
d ( δ , L M ) = t L M δ | M ( r δ ( t ) ) M ( r L M ( t ) ) | × | r δ ( t ) r L M ( t ) |
L M = { A 1 M , . . . , A m M } denotes an ordered list of top m algorithms produced by the validation measure M. Let M ( 1 ) , . . . , M ( m ) be the scores for the top m algorithms in L M , where M ( 1 ) is the best score given by measure M and so on. Let r M ( A ) be the rank of A under M (1 means “best”) if A is within top m, and be equal to m + 1 ; otherwise, r δ ( A ) is defined likewise.
The optimization of the objective (12) is achieved by using the Monte Carlo cross-entropy (CE) algorithm [51,52]. CE Monte Carlo algorithm is a stochastic search method, which produces a “better” sample in the future, which is concentrated around an x that corresponds to an optimal δ * [50].

4.5.2. Strategies to Generate New Samples

Before performing rank aggregation, the training instances are to be modified to produce a more balanced class distribution. To achieve this task, new minority or (and) majority class samples need to be generated or drawn from the original dataset. We employ the following three strategies to gain new samples:

Randomly Sampling

In the over-sampling, some (all) the minority class instances are randomly duplicated; in the under-sampling, a portion of majority samples are randomly removed.

SMOTE

Synthetic minority over-sampling technique (SMOTE) is a popular over-sampling algorithm [5]. Figure 12 illustrates how to generate new samples according to the selected point x i in SMOTE. The five selected nearest neighbors of x i are x i 1 to x i 5 . x i 1 to x i 5 are the synthetic data points created by the randomized interpolation. Namely,
x i h = x i + u h ( x i h x i ) , h = 1 , 2 , , 5 ,
where u h is a random number between 0 and 1. The above operation can be repeated to obtain requested synthetic minority instances.

Smoothed Bootstrap

Smoothed bootstrap technique repeatedly bootstraps the data from the two classes and employs smoothed kernel functions to generate new approximately balanced samples [53]. A new instance is generated by performing the following three steps:
step
1: choose y = k { 1 , 2 } with probability 0.5 ;
step
2: choose ( x i , y i ) in the original daa set such that y i = k with probability 1 / n k ;
step
3: sample x from a probability distribution K H k ( · , x i ) , which is centered at x i and depends on the smoothing matrix H k .
In brief, smoothed bootstrap firstly draws randomly from the original dataset an instance from one of the two categories, then generates a new instance in its neighborhood.

4.6. Experiment and Assessing Metrics

As shown in Table 5, five metabolomics datasets were employed to test our algorithm. NPC is a nasopharyngeal carcinoma dataset [32,54] that is exactly balanced. In this study, NPC was utilized to investigate the performance of rank aggregation strategy on original balanced data, which included 100 patients with nasopharyngeal carcinoma and 100 healthy controls. Traumatic brain injury (TBI) is from our previous studies [32,55], which reports the serum metabolic profiling of TBI patients with (or without) cognitive impairment (CI). The TBI dataset included 73 TBI patients with CI and 31 TBI patients without CI. CHD2-1 and CHD2-2 datasets are actually from the same experiment about coronary heart disease (CHD) [30]. The CHD2-1 dataset contains 21 patients with CHD, and the CHD2-2 dataset contains 16 patients with coronary heart disease associated with type 2 diabetes mellitus (CHD-T2DM), which are compared with a control group of 51 healthy adults. ATR is an Acori Tatarinowii Rhizoma dataset, which included 21 samples collected from Sichuan Province, and 8 samples were from Anhui Province in China [56]. Table 5 lists the summary of five datasets; included are the numbers of attributes, total instances, the majority, the minority instances, and the imbalance ratio. The NPC dataset was utilized to test the performance of rank aggregation under original balanced distribution. The other four imbalanced data sets were used to evaluate the RAR algorithm with artificially re-balanced data.
This section shows the efficacy of the proposed RAR algorithm on one original balanced dataset and four class-imbalanced datasets and compares it with other filtering feature screening methods via several assessing metrics. Rank aggregation was performed under the following seven situations:
  • Case 1 : no re-sampling: the original datasets are directly utilized to perform rank aggregation. Denoted case 1 by “RA” because there is no re-sampling in it.
  • Case 2 : hybrid-sampling A: some instances of the majority class are randomly eliminated, and new synthetic minority examples are generated by SMOTE. The size of the remaining majority is equal to the size of the (original plus new generated) minority class.
  • Case 3 : hybrid-sampling B: A new synthetic dataset is generated according to the smoothed bootstrap re-sampling technique. The sizes of the majority and minority classes are approximately equal.
  • Case 4 : over-sampling A: new minority class instances are randomly duplicated according to the original minority group.
  • Case 5 : over-sampling B: new synthetic minority examples are generated on the basis of the smoothed bootstrap re-sampling technique.
  • Case 6 : under-sampling A: some instances from the majority class are randomly removed so that the size of the remaining majority class is equal to the size of the minority.
  • Case 7 : under-sampling B: new synthetic majority examples are generated according to the smoothed bootstrap re-sampling technique.
Note that NPC is balanced, and just case 1 is performed on it. Table 6 lists the summary of the six re-balanced strategies. In this study, G m e a n , F 1 , A U C R O C , and A U C P R C are employed to assess the performance of RA or RAR algorithm on five datasets under seven cases.

5. Conclusions

In this paper, we propose a simple but effective strategy called RAR for feature screening of class-imbalanced data by aggregating rankings from individual filtering algorithms and modifying the class-imbalanced data with various re-sampling methods to provide balanced or more adequate data. RAR can address the problem of inconsistency between different feature ranking methods to a large extent. The results on real datasets show that RAR is highly competitive and almost better than single filtering screening in terms of geometric mean, F-measure, A U C R O C , and A U C P R C . After performing re-balanced pretreatment, the performance of rank aggregation can be highly improved, so re-sampling to balance the classes is extremely useful in rank aggregation when the data are class-imbalanced in metabolomics. Our proposed method serves as a reference for future research on feature selection for the diagnosis of diseases.
Rank aggregation is a general idea to investigate the importance of features. In this study, rankings from eight filtering algorithms are employed to generate the aggregated rank. There are many other filter techniques, such as Chi-squared, power, Kolmogorov–Smirnov statistic, and signal-to-noise ratio [57], which are all widely utilized in class-imbalance learning. In addition, considering that a re-sampling method can also generate a rank list, rank aggregation can be performed according to the various re-sampling algorithms rather than different filtering methods. Further, if necessary, ensemble multiple rank aggregations could be performed to combine those aggregated rankings derived from different algorithms. Finally, although RAR is used in the metabolomics datasets in this study, it is potentially available for handing high-dimensional imbalanced data from other fields, such as economics and biology.

Author Contributions

Conceptualization, G.-H.F.; Data curation, G.-H.F. and J.-B.W.; Funding acquisition, L.-Z.Y.; Software, J.-B.W.; Writing—original draft, M.-J.Z.; Writing—review and editing, G.-H.F. and L.-Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is financially supported by the National Natural Science Foundation of China (Grant Nos. 11761041 and 21775058).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets for this article are available from the corresponding author.

Acknowledgments

The authors sincerely thank the academic editor and four anonymous reviewers for their constructive comments that led to the current improved version of the paper. We would like to thank the National Natural Science Foundation of China for its financial support (Grant Nos. 11761041 and 21775058).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Brodley, C.; Friedl, M. Identifying mislabeled training data. J. Artif. Intell. Res. 1999, 11, 131–167. [Google Scholar] [CrossRef]
  2. Chawla, N. Data mining for imbalanced datasets: An overview. In Data Mining and Knowledge Discovery Handbook; Springer: Berlin/Heidelberg, Germany, 2009; pp. 875–886. [Google Scholar]
  3. Krawczyk, B. Learning from imbalanced data: Open challenges and future directions. Prog. Artif. Intell. 2016, 5, 221–232. [Google Scholar] [CrossRef] [Green Version]
  4. Cordón, I.; García, S.; Fernández, A.; Herrera, F. Imbalance: Oversampling algorithms for imbalanced classification in R. Knowl. Based Syst. 2018, 161, 329–341. [Google Scholar] [CrossRef]
  5. Chawla, N.; Bowyer, K.; Hall, L.; Kegelmeyer, W. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  6. Lunardon, N.; Menardi, G.; Torelli, N. ROSE: A Package for Binary Imbalanced Learning. R J. 2014, 6, 79–89. [Google Scholar] [CrossRef] [Green Version]
  7. Hulse, J.V.; Khoshgoftaar, T.; Napolitano, A.; Wald, R. Feature selection with high-dimensional imbalanced data. In Proceedings of the 2009 IEEE International Conference on Data Mining Workshops, Miami, FL, USA, 6 December 2009; pp. 507–514. [Google Scholar]
  8. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  9. Saeys, Y.; Inza, I.; Larrañaga, P. A review of feature selection techniques in bioinformatics. Bioinformatics 2007, 23, 2507–2517. [Google Scholar] [CrossRef] [Green Version]
  10. Yun, Y.H.; Li, H.D.; Deng, B.C.; Cao, D.S. An overview of variable selection methods in multivariate analysis of near-infrared spectra. TrAC Trends Anal. Chem. 2019, 113, 102–115. [Google Scholar] [CrossRef]
  11. Su, X.; Khoshgoftaar, T. A survey of collaborative filtering techniques. Adv. Artif. Intell. 2009, 2009, 421425. [Google Scholar] [CrossRef]
  12. Ambjørn, J.; Janik, R.; Kristjansen, C. Wrapping interactions and a new source of corrections to the spin-chain/string duality. Nucl. Phys. B 2006, 736, 288–301. [Google Scholar] [CrossRef] [Green Version]
  13. Higman, G.; Neumann, B.; Neuman, H. Embedding theorems for groups. J. Lond. Math. Soc. 1949, 1, 247–254. [Google Scholar] [CrossRef]
  14. Guo, H.; Li, Y.; Shang, J.; Gu, M.; Huang, Y.; Gong, B. Learning from class-imbalanced data: Review of methods and applications. Expert Syst. Appl. 2017, 73, 220–239. [Google Scholar]
  15. Gu, Q.; Li, Z.; Han, J. Generalized fisher score for feature selection. arXiv 2012, arXiv:1202.3725. [Google Scholar]
  16. Yin, L.; Ge, Y.; Xiao, K.; Wang, X.; Quan, X. Feature selection for high-dimensional imbalanced data. Neurocomputing 2013, 105, 3–11. [Google Scholar] [CrossRef]
  17. Spolaôr, N.; Cherman, E.; Monard, M.; Lee, H. ReliefF for multi-label feature selection. In Proceedings of the 2013 Brazilian Conference on Intelligent Systems, Fortaleza, Brazil, 19–24 October 2013; pp. 6–11. [Google Scholar]
  18. Kira, K.; Rendell, L. The feature selection problem: Traditional methods and a new algorithm. Aaai 1992, 2, 129–134. [Google Scholar]
  19. Lee, C.; Lee, G. Information gain and divergence-based feature selection for machine learning-based text categorization. Inf. Process. Manag. 2006, 42, 155–165. [Google Scholar] [CrossRef]
  20. Lerman, R.; Yitzhaki, S. A note on the calculation and interpretation of the Gini index. Econ. Lett. 1984, 15, 363–368. [Google Scholar] [CrossRef]
  21. Lobo, J.; Jiménez-Valverde, A.; Real, R. AUC: A misleading measure of the performance of predictive distribution models. Glob. Ecol. Biogeogr. 2008, 17, 145–151. [Google Scholar] [CrossRef]
  22. Boyd, K.; Eng, K.; Page, C. Area under the precision-recall curve: Point estimates and confidence intervals. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2013; pp. 451–466. [Google Scholar]
  23. Altidor, W.; Khoshgoftaar, T.; Napolitano, A. Wrapper-based feature ranking for software engineering metrics. In Proceedings of the 2009 International Conference on Machine Learning and Applications, Miami, FL, USA, 13–15 December 2009; pp. 241–246. [Google Scholar]
  24. Pillai, I.; Fumera, G.; Roli, F. F-measure optimisation in multi-label classifiers. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; pp. 2424–2427. [Google Scholar]
  25. Lee, J.; Batnyam, N.; Oh, S. RFS: Efficient feature selection method based on R-value. Comput. Biol. Med. 2013, 43, 91–99. [Google Scholar] [CrossRef] [PubMed]
  26. Ali, M.; Ali, S.I.; Kim, D.; Hur, T.; Bang, J.; Lee, S.; Kang, B.H.; Hussain, M.; Zhou, F. UEFS: An efficient and comprehensive ensemble-based feature selection methodology to select informative features. PLoS ONE 2018, 13, e0202705. [Google Scholar] [CrossRef] [PubMed]
  27. Hoque, N.; Singh, M.; Bhattacharyya, D.K. EFS-MI: An ensemble feature selection method for classification. Complex Intell. Syst. 2018, 4, 105–118. [Google Scholar] [CrossRef]
  28. Yang, P.; Liu, W.; Zhou, B.B.; Chawla, S.; Zomaya, A.Y. Ensemble-Based Wrapper Methods for Feature Selection and Class Imbalance Learning; Springer: Berlin/Heidelberg, Germany, 2013; pp. 544–555. [Google Scholar]
  29. Lin, X.; Yang, F.; Zhou, L.; Yin, P.; Kong, H.; Xing, W.; Lu, X.; Jia, L.; Wang, Q.; Xu, G. A support vector machine-recursive feature elimination feature selection method based on artificial contrast variables and mutual information. J. Chromatogr. B 2012, 910, 149–155. [Google Scholar] [CrossRef] [PubMed]
  30. Fu, G.H.; Wu, Y.J.; Zong, M.J.; Yi, L.Z. Feature selection and classification by minimizing overlap degree for class-imbalanced data in metabolomics. Chemom. Intell. Lab. Syst. 2020, 196, 103906. [Google Scholar] [CrossRef]
  31. Sen, P. Estimates of the regression coefficient based on Kendall’s tau. J. Am. Stat. Assoc. 1968, 63, 1379–1389. [Google Scholar] [CrossRef]
  32. Fu, G.H.; Xu, F.; Zhang, B.Y.; Yi, L.Z. Stable variable selection of class-imbalanced data with precision-recall criterion. Chemom. Intell. Lab. Syst. 2017, 171, 241–250. [Google Scholar] [CrossRef]
  33. Takaya, S.; Marc, R.; Guy, B. The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar]
  34. Yun, Y.H.; Deng, B.C.; Cao, D.S.; Wang, W.T.; Liang, Y.Z. Variable importance analysis based on rank aggregation with applications in metabolomics for biomarker discovery. Anal. Chim. Acta 2016, 911, 27–34. [Google Scholar] [CrossRef]
  35. Weston, J.; Mukherjee, S.; Chapelle, O. Feature selection for SVMs. In Proceedings of the Advances in Neural information Processing Systems, Vancouver, BC, Canada, 3–8 December 2001; pp. 668–674. [Google Scholar]
  36. Kailath, T. The Divergence and Bhattacharyya Distance Measures in Signal Selection. IEEE Trans. Commun. Technol. 1967, 15, 52–60. [Google Scholar] [CrossRef]
  37. Fu, G.H.; Wu, Y.J.; Zong, M.J.; Pan, J. Hellinger distance-based stable sparse feature selection for high-dimensional class-imbalanced data. BMC Bioinform. 2020, 21, 121. [Google Scholar] [CrossRef] [Green Version]
  38. Robnik-Šikonja, M.; Kononenko, I. Theoretical and Empirical Analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  39. Kononenko, I. Estimating attributes: Analysis and extensions of RELIEF. In European Conference on Machine Learning; Springer: Berlin/Heidelberg, Germany, 1994; pp. 171–182. [Google Scholar]
  40. Yang, Y.; Pedersen, J. A comparative study on feature selection in text categorization. Icml 1997, 97, 35. [Google Scholar]
  41. Shang, W.; Huang, H.; Zhu, H.; Lin, Y.; Qu, Y.; Wang, Z. A novel feature selection algorithm for text categorization. Expert Syst. Appl. 2007, 33, 1–5. [Google Scholar] [CrossRef]
  42. Borsos, Z.; Lemnaru, C.; Potolea, R. Dealing with overlap and imbalance: A new metric and approach. Pattern Anal. Appl. 2018, 21, 381–395. [Google Scholar] [CrossRef]
  43. Oh, S. A new dataset evaluation method based on category overlap. Comput. Biol. Med. 2011, 41, 115–122. [Google Scholar] [CrossRef]
  44. Provost, F.; Fawcett, T. Robust classification for imprecise environments. Mach. Learn. 2001, 42, 203–231. [Google Scholar] [CrossRef]
  45. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar]
  46. Fu, G.H.; Yi, L.Z.; Pan, J. Tuning model parameters in class-imbalanced learning with precision-recall curve. Biom. J. 2019, 61, 652–664. [Google Scholar] [CrossRef]
  47. Kendall, M.G. A New Measure of Rank Correlation. Biometrika 1938, 30, 81–93. [Google Scholar] [CrossRef]
  48. Shieh, G. A weighted Kendall’s tau statistic. Stat. Probab. Lett. 1998, 39, 17–24. [Google Scholar] [CrossRef]
  49. Pihur, V. Statistical Methods for High-Dimensional Genomics Data Analysis; University of Louisville: Louisville, KY, USA, 2009. [Google Scholar]
  50. Pihur, V.; Datta, S.; Datta, S. RankAggreg, an R package for weighted rank aggregation. BMC Bioinform. 2009, 10, 62. [Google Scholar] [CrossRef] [Green Version]
  51. Pihur, V.; Datta, S.; Datta, S. Weighted rank aggregation of cluster validation measures: A Monte Carlo cross-entropy approach. Bioinformatics 2007, 23, 1607–1615. [Google Scholar] [CrossRef] [Green Version]
  52. Pihur, V.; Datta, S.; Datta, S. Finding common genes in multiple cancer types through meta–analysis of microarray experiments: A rank aggregation approach. Genomics 2008, 92, 400–403. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Menardi, G.; Torelli, N. Training and assessing classification rules with imbalanced data. Data Min. Knowl. Discov. 2014, 28, 92–122. [Google Scholar] [CrossRef]
  54. Fu, G.H.; Yi, L.Z.; Pan, J. LASSO-based false-positive selection for class-imbalanced data in metabolomics. J. Chemom. 2019, 33. [Google Scholar] [CrossRef]
  55. Fu, G.H.; Zhang, B.Y.; Kou, H.D.; Yi, L.Z. Stable biomarker screening and classification by subsampling-based sparse regularization coupled with support vector machines in metabolomics. Chemom. Intell. Lab. Syst. 2017, 160, 22–31. [Google Scholar] [CrossRef]
  56. Ma, S.S.; Zhang, B.Y.; Chen, L.; Zhang, X.J.; Ren, D.B.; Yi, L.Z. Discrimination of Acori Tatarinowii Rhizoma from two habitats based on GC-MS fingerprinting and LASSO-PLS-DA. J. Cent. South Univ. 2018, 25, 1063–1075. [Google Scholar] [CrossRef]
  57. Fernández, A.; García, S.; Galar, M.; Prati, R.C.; Krawczyk, B.; Herrera, F. Dimensionality Reduction for Imbalanced Learning. In Learning from Imbalanced Data Sets; Springer: Cham, Switzerland, 2018; pp. 227–251. [Google Scholar] [CrossRef]
Figure 1. Kendall’s τ rank correlation coefficient under different imbalance ratios (with 100 repeats). Left: eight key variables; right: eight key plus eight irrelevant variables.
Figure 1. Kendall’s τ rank correlation coefficient under different imbalance ratios (with 100 repeats). Left: eight key variables; right: eight key plus eight irrelevant variables.
Metabolites 11 00389 g001
Figure 2. Rank lists of rank aggregation with the dataset “NPC” in case 1.
Figure 2. Rank lists of rank aggregation with the dataset “NPC” in case 1.
Metabolites 11 00389 g002
Figure 3. Rank lists of rank aggregation with the dataset “TBI” on seven cases.
Figure 3. Rank lists of rank aggregation with the dataset “TBI” on seven cases.
Metabolites 11 00389 g003
Figure 4. Rank lists of rank aggregation with the dataset “CHD2-1” on seven cases.
Figure 4. Rank lists of rank aggregation with the dataset “CHD2-1” on seven cases.
Metabolites 11 00389 g004
Figure 5. Rank lists of rank aggregation with the dataset “CHD2-2” on seven cases.
Figure 5. Rank lists of rank aggregation with the dataset “CHD2-2” on seven cases.
Metabolites 11 00389 g005
Figure 6. Rank lists of rank aggregation with the dataset “ATR” on seven cases.
Figure 6. Rank lists of rank aggregation with the dataset “ATR” on seven cases.
Metabolites 11 00389 g006
Figure 7. A U C P R C when top k features are used with the dataset “TBI” on seven cases.
Figure 7. A U C P R C when top k features are used with the dataset “TBI” on seven cases.
Metabolites 11 00389 g007
Figure 8. A U C P R C when top k features are used with the dataset “CHD2-1” on seven cases.
Figure 8. A U C P R C when top k features are used with the dataset “CHD2-1” on seven cases.
Metabolites 11 00389 g008
Figure 9. A U C P R C when top k features are used with the dataset “CHD2-2” on seven cases.
Figure 9. A U C P R C when top k features are used with the dataset “CHD2-2” on seven cases.
Metabolites 11 00389 g009
Figure 10. A U C P R C when top k features are used with the dataset “ATR” on seven cases.
Figure 10. A U C P R C when top k features are used with the dataset “ATR” on seven cases.
Metabolites 11 00389 g010
Figure 11. The frame of rank aggregation with re-balance.
Figure 11. The frame of rank aggregation with re-balance.
Metabolites 11 00389 g011
Figure 12. An illustration of how to create the synthetic data points in the SMOTE algorithm.
Figure 12. An illustration of how to create the synthetic data points in the SMOTE algorithm.
Metabolites 11 00389 g012
Table 1. G m e a n from rank aggregation and eight filtering techniques (The best result is in bold).
Table 1. G m e a n from rank aggregation and eight filtering techniques (The best result is in bold).
DatasetResamplingNo.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 171.000.870.970.950.950.950.951.000.92
TBICase 160.960.410.680.680.700.880.720.580.63
Case 2111.000.840.880.950.900.940.860.900.90
Case 311.000.950.840.950.900.800.950.901.00
Case 4101.000.961.000.890.930.970.930.851.00
Case 571.000.930.900.970.970.930.970.860.93
Case 6121.000.750.830.820.830.910.850.711.00
Case 7291.000.710.700.650.820.850.780.710.71
CHD2-1Case 1100.870.670.710.870.000.770.770.470.87
Case 2110.940.850.851.000.870.930.940.850.91
Case 361.000.861.001.001.000.931.000.931.00
Case 4311.001.001.001.001.001.001.000.951.00
Case 5371.001.001.001.001.001.001.001.001.00
Case 6110.870.610.750.870.870.890.870.500.87
Case 7100.870.610.710.870.870.870.870.710.71
CHD2-2Case 130.580.480.580.480.580.550.680.000.73
Case 2341.001.001.001.001.001.001.001.001.00
Case 3141.001.001.001.001.001.001.001.001.00
Case 4241.001.001.001.001.001.001.001.001.00
Case 5281.001.001.001.001.001.001.001.001.00
Case 671.001.000.820.821.000.820.820.470.82
Case 740.860.821.000.821.000.611.000.000.75
ATRCase 1251.001.001.001.001.001.001.000.501.00
Case 291.001.001.001.001.001.001.001.001.00
Case 381.001.001.001.001.001.001.001.001.00
Case 421.001.001.001.001.001.001.000.891.00
Case 521.001.001.001.001.001.001.000.801.00
Case 691.001.001.001.001.001.001.000.711.00
Case 7101.001.001.001.001.001.001.001.001.00
Table 2. F 1 from rank aggregation and eight filtering techniques (The best result is in bold).
Table 2. F 1 from rank aggregation and eight filtering techniques (The best result is in bold).
DatasetResamplingNO.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 181.000.880.950.950.971.000.950.950.92
TBICase 1110.930.820.860.870.900.900.880.830.88
Case 2131.000.840.900.940.950.940.840.870.86
Case 311.000.870.871.001.000.960.950.831.00
Case 471.000.910.970.930.970.900.930.891.00
Case 5101.000.870.930.930.930.930.970.880.93
Case 6191.000.730.800.860.860.770.800.710.91
Case 7111.000.830.920.800.860.860.750.500.86
CHD2-1Case 1100.880.870.880.870.870.870.870.830.87
Case 2110.940.940.710.930.890.930.880.670.89
Case 361.000.891.001.001.000.930.940.940.92
Case 4311.000.950.950.960.900.960.900.860.95
Case 5371.001.000.910.950.951.000.950.860.95
Case 6110.890.500.750.730.891.000.830.550.67
Case 7100.860.730.570.800.750.860.670.550.57
CHD2-2Case 130.910.870.860.950.870.910.900.870.91
Case 2341.001.001.000.880.920.931.000.880.93
Case 3141.001.000.920.860.921.000.930.860.93
Case 4241.000.900.910.950.950.951.000.950.95
Case 5281.000.950.950.950.951.000.960.951.00
Case 670.860.570.800.800.750.860.860.500.86
Case 740.860.800.750.570.860.860.860.500.67
ATRCase 1251.001.000.891.001.001.001.000.891.00
Case 291.001.001.001.001.001.001.001.001.00
Case 381.001.001.001.001.001.001.001.001.00
Case 421.001.001.001.001.001.001.000.891.00
Case 521.001.001.001.001.001.001.001.001.00
Case 691.001.001.001.001.001.001.000.671.00
Case 7101.001.001.001.001.001.001.000.671.00
Table 3. A U C R O C from rank aggregation and eight filtering techniques (The best result is in bold).
Table 3. A U C R O C from rank aggregation and eight filtering techniques (The best result is in bold).
DatasetResamplingNO.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 1160.960.900.910.940.930.960.930.930.95
TBICase 130.700.480.610.520.610.650.580.490.67
Case 2110.890.800.810.780.820.870.790.720.83
Case 310.950.830.740.940.850.830.850.760.95
Case 4280.910.850.860.870.880.860.860.840.89
Case 5260.930.910.920.900.900.920.900.880.90
Case 6220.770.650.680.690.730.740.730.500.71
Case 7250.710.630.560.680.530.610.610.350.66
CHD2-1Case 1100.600.480.510.610.500.590.600.450.59
Case 2110.760.660.600.750.770.730.720.610.73
Case 360.920.850.810.920.900.810.880.800.79
Case 4310.860.810.790.860.870.820.840.750.84
Case 5370.870.850.840.860.860.810.860.900.82
Case 6110.640.570.450.500.570.570.640.360.62
Case 7100.600.520.550.520.480.520.600.360.52
CHD2-2Case 130.600.520.490.540.500.520.550.390.57
Case 2340.860.790.840.880.730.800.800.700.79
Case 3140.900.850.840.820.900.680.880.810.84
Case 4240.930.880.880.880.900.860.910.870.93
Case 5280.910.880.900.880.850.910.890.890.90
Case 670.630.530.560.660.590.440.560.280.59
Case 740.660.560.630.500.660.660.630.220.56
ATRCase 1250.880.980.510.850.850.760.850.500.76
Case 290.960.790.860.960.960.930.960.750.93
Case 380.971.000.900.970.900.970.970.800.93
Case 420.980.930.830.880.980.950.980.810.95
Case 520.980.810.860.760.930.930.950.710.88
Case 690.940.810.750.880.940.880.810.190.81
Case 7100.940.810.630.690.630.940.880.190.75
Table 4. A U C P R C from rank aggregation and eight filtering techniques (The best result is in bold).
Table 4. A U C P R C from rank aggregation and eight filtering techniques (The best result is in bold).
DatasetResamplingNO.RARARt TestFisherHellingerReliefReliefFIGGiniR-Value
NPCCase 1150.960.870.900.910.920.960.920.910.91
TBICase 180.620.270.580.400.470.560.470.260.46
Case 2180.850.790.810.780.860.840.810.650.75
Case 310.930.770.600.910.850.740.850.740.89
Case 4270.890.840.810.830.850.860.830.810.85
Case 5130.910.810.820.810.820.900.860.790.81
Case 6300.780.610.690.640.620.660.670.500.70
Case 7240.710.650.540.640.590.570.620.410.58
CHD2-1Case 1100.610.330.300.520.280.450.430.230.45
Case 2110.770.560.580.690.650.640.680.550.65
Case 360.880.790.880.810.770.810.860.790.86
Case 4310.860.810.800.800.820.830.850.730.80
Case 5370.870.810.750.840.850.820.840.780.82
Case 6110.660.460.450.480.650.530.550.400.50
Case 7100.630.440.530.520.550.560.560.400.48
CHD2-2Case 130.450.220.330.270.320.310.270.210.26
Case 2340.870.680.780.760.820.850.860.730.80
Case 3140.870.810.740.810.860.760.830.790.78
Case 4240.910.840.900.870.850.850.870.850.81
Case 5280.900.860.880.800.880.860.890.790.79
Case 670.710.420.620.640.660.450.570.350.63
Case 740.640.640.570.500.630.550.660.400.60
ATRCase 1250.930.750.520.520.820.820.820.250.60
Case 291.000.810.880.920.940.940.940.680.88
Case 381.001.000.781.000.880.931.000.820.88
Case 420.950.840.790.910.910.880.910.700.91
Case 520.950.790.780.780.880.900.950.610.88
Case 691.000.890.690.850.890.760.850.360.80
Case 7101.000.610.540.640.540.800.890.370.80
Table 5. The summary of five datasets.
Table 5. The summary of five datasets.
DatasetsAttributesInstancesMajorityMinorityRatio
NPC242001001001.00
TBI4210473312.35
CHD2-1507251212.43
CHD2-2506751163.19
ATR104292182.63
Table 6. Re-balanced strategies.
Table 6. Re-balanced strategies.
MethodsRe-Sampling ProcessAlgorithm Process
Under-SamplingOver-SamplingHybridSMOTERandomSmoothed Bootstrap
Case 2 YesYes
Case 3 Yes Yes
Case 4 Yes Yes
Case 5 Yes Yes
Case 6Yes Yes
Case 7Yes Yes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fu, G.-H.; Wang, J.-B.; Zong, M.-J.; Yi, L.-Z. Feature Ranking and Screening for Class-Imbalanced Metabolomics Data Based on Rank Aggregation Coupled with Re-Balance. Metabolites 2021, 11, 389. https://doi.org/10.3390/metabo11060389

AMA Style

Fu G-H, Wang J-B, Zong M-J, Yi L-Z. Feature Ranking and Screening for Class-Imbalanced Metabolomics Data Based on Rank Aggregation Coupled with Re-Balance. Metabolites. 2021; 11(6):389. https://doi.org/10.3390/metabo11060389

Chicago/Turabian Style

Fu, Guang-Hui, Jia-Bao Wang, Min-Jie Zong, and Lun-Zhao Yi. 2021. "Feature Ranking and Screening for Class-Imbalanced Metabolomics Data Based on Rank Aggregation Coupled with Re-Balance" Metabolites 11, no. 6: 389. https://doi.org/10.3390/metabo11060389

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop