Next Article in Journal
A Classification of Compact Cohomogeneity One Locally Conformal Kähler Manifolds
Previous Article in Journal
Improved Snake Optimizer Using Sobol Sequential Nonlinear Factors and Different Learning Strategies and Its Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Imbalanced Data Classification Based on Improved Random-SMOTE and Feature Standard Deviation

1
School of Science, Zhejiang Sci-Tech University, Hangzhou 310018, China
2
School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
3
Longgang Research Institute, Zhejiang Sci-Tech University, Longgang 325000, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(11), 1709; https://doi.org/10.3390/math12111709
Submission received: 25 April 2024 / Revised: 16 May 2024 / Accepted: 27 May 2024 / Published: 30 May 2024

Abstract

:
Oversampling techniques are widely used to rebalance imbalanced datasets. However, most of the oversampling methods may introduce noise and fuzzy boundaries for dataset classification, leading to the overfitting phenomenon. To solve this problem, we propose a new method (FSDR-SMOTE) based on Random-SMOTE and Feature Standard Deviation for rebalancing imbalanced datasets. The method first removes noisy samples based on the Tukey criterion and then calculates the feature standard deviation reflecting the degree of data discretization to detect the sample location, and classifies the samples into boundary samples and safety samples. Secondly, the K-means clustering algorithm is employed to partition the minority class samples into several sub-clusters. Within each sub-cluster, new samples are generated based on random samples, boundary samples, and the corresponding sub-cluster center. The experimental results show that the average evaluation value obtained by FSDR-SMOTE is 93.31% (93.16%, and 86.53%) in terms of the F-measure (G-mean, and MCC) on the 20 benchmark datasets selected from the UCI machine learning library.

1. Introduction

In the field of machine learning, the problem of imbalanced data classification [1] has attracted great interests. In many real-world scenarios, due to a variety of factors such as bias in data collection [2], inherent rarity of categories [3], differences in the frequency of events [4,5], etc., research often needs to deal with a large number of imbalanced datasets, where the number of samples of one class is much larger than that of other classes. However, traditional classification algorithms usually assume that the number of samples in each class is balanced during the training phase, and thus traditional classification algorithms often exhibit degraded performance when dealing with imbalanced data. Specifically, due to the higher number of samples in the majority class, the model is more likely to learn the features of the majority class during the training process, resulting in the classifier being biased towards the majority class during prediction [6]. Bias toward majority class samples makes the classifier less capable of identifying minority class samples, and may even completely ignore the importance of the minority class, which can lead to misclassification in practical applications [7,8]. Therefore, an in-depth study of imbalanced classification algorithms is crucial to improve the performance and generalization ability of classifiers, especially when it comes to minority class samples with practical applications.
Currently, the existing methods for addressing imbalances in datasets are categorized into data-level and algorithm-level methods. Data-centric methods involve resampling techniques, such as augmenting the minority class with oversampling strategies [9], reducing the number of the majority class samples with undersampling strategies [10], and achieving balance with hybrid sampling methods [11]. In addition, algorithm-level methods dynamically focus on difficult-to-classify samples by adjusting category weights, implementing cost-sensitive learning [12], using ensemble learning [13,14], and so on.
Oversampling is one of the effective means of solving the problem of imbalanced data classification because oversampling can rebalance the dataset while maintaining the data characteristics. Among oversampling techniques, randomly replicating the minority class samples (ROS) [15] is the simplest and most straightforward oversampling method; however, ROS may introduce noisy instances and overfitting risks since the data distribution is not taken into account. In contrast, the oversampling technique of synthetizing minority class samples [16,17] can improve the classification accuracy of the model to some extent, but it still suffers from these problems mentioned above. To solve these problems, researchers have invested a lot of effort in the data preprocessing stage, especially in the noise filtering phase. For example, Nawaf et al. [18] used Tukey’s rule to remove outliers and noisy samples from the original dataset. Liang [19] combined support vector machine and nearest neighbor algorithms to deal with noise to reduce its effect on the oversampling phase. Therefore, when dealing with imbalanced data, fully considering the characteristics of data distribution plays a crucial role in improving algorithm performance.
To further enhance the effectiveness of synthesizing new samples, some existing oversampling methods focus on the minority class instances located near the decision boundary after denoising. For instance, based on the confusing information of the samples, Zhang et al. [20] classified the minority class samples into three types: noise samples, boundary samples, and safe samples, where only boundary samples are selected for generating new synthetic samples. In this way, the newly generated samples are closer to the decision boundary, which can improve the performance of the classifier. In addition, Cheng et al. [21] used Natural Neighbors (NaN) to calculate the local density of the samples and filtered the representative samples with maximum local density, aiming at changing the selection strategy for resampling samples. The common goal of all these methods is to improve the quality and effectiveness of synthetic samples and then optimize the performance of the classifier in the imbalanced data classification problem. However, for identifying the minority class samples located near the decision boundary, most of the research tends to focus on the relationships between samples, without considering the different characteristics of them.
In addition to changing the distribution region of the synthetic samples by adjusting their weights, comprehensive strategies for generating new samples are particularly important when improving the performance of oversampling algorithms. For example, stochastic linear interpolation is a common method for generating a new sample, which is based on a stochastic function and two target samples. Although the linear interpolation method is more effective in most cases, it may trigger the problem of redundancy in some regions with high sample density, leading to the ineffectiveness of the synthesized samples. To address this issue, Dong et al. [22] proposed the Random-SMOTE method, which interpolates within a triangular region consisting of the cluster center and two randomly selected neighbor samples. Random-SMOTE can enhance the diversity of synthetic samples and avoid generating redundant samples. Although Random-SMOTE can alleviate the problem of sample redundancy to a certain extent, Random-SMOTE does not fully consider the imbalance within the sample classes, resulting in the samples in the sparse region still carrying limited classification information.
To address the above problems, this paper proposes an imbalanced data classification method (FSDR-SMOTE) based on Random-SMOTE and feature standard deviation. In FSDR-SMOTE, noisy samples are firstly removed based on the Tukey criterion, and the feature standard deviation reflecting the degree of data discretization is utilized to detect the sample locations. Furthermore, the improved Random- SMOTE method is proposed to enhance the stability and accuracy of the algorithm when maintaining the diversity of the synthesized samples. The main contributions of this work are summarized as follows:
  • A new strategy based on feature standard deviation is proposed to detect the locations of samples. Feature standard deviation strategy utilizes standard deviation in feature dimensions among different samples to construct a boundary sample set, aiming to optimize the position of the decision boundary.
  • An improved three-point interpolation method proposed in Random-SMOTE is presented to synthesize samples, where both inter-class and intra-class attributes are taken into account, with the purpose of avoiding overfitting and optimizing the decision boundary.
  • A new method (FSDR-SMOTE) is proposed to deal with imbalanced data. The FSDR-SMOTE can enhance the diversity of synthetic samples in the minority class, reduce the probability of generating noisy samples, and improve the classification performance of the classifier.

2. Related Work

In recent years, many research efforts have been focused on how to improve classification accuracy of algorithms by utilizing useful information carried by the minority class samples effectively [23,24]. The oversampling method is one of the most representative methods for solving the imbalanced data classification problem, which can realize the balance between samples of different classes by synthesizing new minority class samples.
As one of the representative oversampling methods, the SMOTE was proposed by Chawla et al. [9]. In SMOTE, new samples are generated by linearly interpolating between a sample x i chosen arbitrarily from the minority class and any one x j of the k-nearest neighbors of that sample. When dealing with an imbalanced data classification task, the SMOTE can mitigate the overfitting problem and improve the classification performance of the classifier. The procedure for synthesizing minority class samples is as follows:
x n e w = x i + r a n d ( 0 , 1 ) × ( x i x j )
where r a n d ( 0 , 1 ) denotes a random number chosen from ( 0 , 1 ) .
Although the SMOTE can alleviate the overfitting phenomenon effectively, the characteristic of data distribution has not been taken into account. To address this issue, HE et al. [25] proposed an adaptive integrated oversampling method, named ADASYN, which assigns minority class samples with different weights. Borderline-SMOTE [26] proposed by Han et al. can increase the number of borderline minority class samples to alleviate the impact of class overlap; however, it is difficult to guarantee the diversity of synthesized new samples.
The SMOTE-Tomek [27] is a hybrid sampling method used for balancing datasets. Here are the steps involved: First, the SMOTE algorithm is employed to oversample the minority class samples in the dataset. Next, the nearest neighbors of all majority class samples in the dataset are calculated. If the nearest neighbor of a majority class sample is a minority class sample, a Tomek link is established. Finally, all Tomek links are removed, aiming at obtaining a balanced dataset. The schematic diagram of the SMOTE-Tomek method is shown in Figure 1.
SVMSMOTE [28] focuses on generating samples of the minority class along the decision boundary. The process of creating synthetic samples with SVMSMOTE is as follows. First, a Support Vector Machine (SVM) [29] is trained based on the dataset, which can produce multiple support vectors. Then, for each support vector, the k-nearest neighbors from the minority class samples are found. If the number of majority class samples is less than half of its nearest neighbors, the extrapolation method is employed to synthesize minority class samples for extending the area of the minority class. Conversely, the interpolation method is utilized to synthesize minority class samples.
RN-SMOTE [30] is an oversampling method that integrates SMOTE and DBSCAN (Density-Based Spatial Clustering of Applications with Noise) [31]. Firstly, RN-SMOTE employs SMOTE for oversampling to achieve a balanced dataset, which may contain some noise samples. Subsequently, DBSCAN clustering is utilized to eliminate the noise within the minority class samples, with the purpose of obtaining a clean dataset. Finally, SMOTE is reapplied to oversample the now balanced dataset. The main process of RN-SMOTE is shown in Figure 2, where the bars represent the quantity of samples [30].
In order to improve the strategy of linear interpolation synthetic samples between two points in SMOTE, Dong et al. [22] proposed Random-SMOTE based on the SMOTE. The main idea of Random-SMOTE is to construct a triangular region, and then generate new samples randomly within the region. New generated samples are closer to the true distribution of samples, which can help to improve the classification performance of the classifier biasing towards the minority class samples. The process of Random-SMOTE for synthesizing minority class samples is shown in Figure 3.
In the Random-SMOTE algorithm, for each sample x in the minority class, two samples y i and y j are selected randomly from them to form a triangular region. First of all, two temporary samples, namely t i and t j , are generated by using the following method:
t k = y i + r a n d ( 0 , 1 ) × ( y i y j ) , k = i , j
Then, the linear interpolation strategy is utilized based on sample x and temporary samples ( t i and t j ) to generate two new minority class samples ( p i and p j ), shown as Equation (3):
p k = x + r a n d ( 0 , 1 ) × ( x t k ) , k = i , j
Finally, the newly synthesized samples are added to the original dataset. In this way, the process for synthesizing new samples is repeated until the dataset is balanced.
Studies have shown that the Random-SMOTE method can improve the quality of newly synthesized samples to some extent; however, there are still some shortcomings. First, the characteristic of the distribution of samples is not taken into account, which may reduce the diversity of newly synthesized samples. Second, the characteristics of the samples are not taken seriously in the Random-SMOTE method when synthesizing new samples, which may lead to the introduction of noise samples and the reduction in performance of the classifier. Therefore, this paper proposes a new oversampling method, namely FSDR-SMOTE, where different features of samples are taken into account to identify the sample locations when synthesizing new samples.

3. Method

To deal with imbalanced data classification effectively, an improved Random-SMOTE method based on feature standard deviation and adaptive denoising, namely FSDR-SMOTE, is proposed in this section. The main steps of FSDR-SMOTE include three parts. The first part is data pre-processing, where the Tukey rule [32] and K-means [33] algorithms are employed in minority class samples for denoising and clustering, respectively. Based on feature standard deviation, the minority class samples are divided into the boundary samples and safety samples in the second part. Finally, the process of synthesizing new samples in the Random-SMOTE is improved to reduce the influence caused by intra-class imbalance. The flowchart of FSDR-SMOTE is shown in Figure 4, and the pseudo-code for FSDR-SMOTE is given in Algorithm 1.
Algorithm 1 FSDR-SMOTE
Input: 
Majority Sample Set X m a j ; Minority Sample Set X m i n ; Number of New Synthetic Samples N; Constant in Tukey rule r; Constant of the characteristic standard deviation discriminant boundary sample k.
Output: 
Synthetic minority class samples X g e n .
  1:
X g e n ;
  2:
for  x i X m i n  do
  3:
    Calculate the distance d between x i and x ¯ m i n , sort in ascending order;
  4:
    Calculate the u b and l b according to Equations (4)–(8);
  5:
    if  d i > u b  then
  6:
        remove x i from X m i n ;
  7:
    end if
  8:
end for
  9:
X m i n c X m i n ;
10:
Use K-means to divide X m i n c into P clusters, denoted as L 1 , L 2 , , L P , and get C = { c i } , i = 1 , 2 , 3 , , p , where c i is the cluster center of each subcluster;
11:
for  x i X m i n c  do
12:
      Calculate σ j according to Equation (10);
13:
      Boundary sample sets B j are screened according to Equation (11);
14:
end for
15:
B B j ;
16:
for n in N do
17:
      Randomly select sample x i in set B and record the cluster L i to which x i belongs and the center c i of that cluster;
18:
      while True do
19:
            Randomly select a sample y i from cluster L i ;
20:
            if  d i s t ( x i , c i ) > d i s t ( x i , y i )  then
21:
                 break
22:
            end if
23:
      end while
24:
      New sample s i is generated from the three points x i , y i , and  c i according to the synthesis rule of Equations (13) and (14);
25:
       X g e n s i ;
26:
end for

3.1. Data Preprocessing

Tukey’s rule [32] is an effective method for identifying outliers in multivariate data. When the dataset X = x 1 , x 2 , x 3 , , x n , where X R and x 1 x 2 x 3 x n . The interquartile range (IQR) can be obtained based on the difference between the first quartile (Q1) and the third quartile (Q3), as shown in Equations (4) and (5), respectively. Based on IQR, the degree of dispersion of the data distribution can be reflected, as shown in Equation (6). Furthermore, l b and u b mean the lower bound and upper bound of IQR, as shown in Equations (7) and (8), respectively.
Q 1 = x i i = r o u n d ( ( N + 1 ) × 0.25 )
Q 3 = x j j = r o u n d ( ( N + 1 ) × 0.75 )
I Q R = Q 3 Q 1
l b = Q 3 r × I Q R
u b = Q 1 + r × I Q R
In the above process, the i represents the index of the data point, N represents the total number of data points, and the default value of r proposed by Tukey is 1.5.
In our work, the Tukey criterion is improved to detect the noise effectively in imbalanced data classification task. For a dataset X m i n = { x m i n 1 , x m i n 2 , x m i n 3 , , x m i n m } containing m minority class samples, the mean x ¯ m i n at each dimension of the minority samples is taken as the center of them. The Euclidean distance d j = | x j x ¯ m i n | between each minority class sample x j and the mean x ¯ m i n is calculated and then sorted in ascending order. In particular, the upper and lower bounds of Tukey’s criterion are calculated using Equations (4)–(8). For a sample x s , if d s < l b , the sample is retained because d s is close to the minority class center x ¯ m i n . For a sample that exceeds the upper bound, i.e., d s > l b , d s is defined as noise and removed from the dataset X m i n , aiming at obtaining the noise-reduced minority class sample set ( X m i n c ). By employing the k-means clustering algorithm, our work can divide the noise-reduced minority class samples into several clusters, where the center of each cluster is recorded.

3.2. Boundary Sample Screening

Research suggests that the generalization performance of most oversampling methods is affected since they do not pay enough attention to samples in the boundary region. To address this issue, the relative positions of boundary samples in each sub-cluster consisted of the minority class samples are taken into account. In this way, the oversampling method can be optimized effectively by improving the robustness and accuracy of the classifier in our work. For an n-dimensional minority sample set X = { x 1 , x 2 , x 3 , , x m } , sample x i = { a 1 i , a 2 i , a 3 i , , a n i } containing m samples, where i = 1 , 2 , , m .
a ¯ j = k = 1 m a j k m
σ j = i = 1 m ( a j i a ¯ j ) 2 m , j = 1 , 2 , , n
B j = { x i | a j i a ¯ j | > k · σ j } , i = 1 , 2 , , m
B = j = 1 n B j
where a j i denotes the value of sample i in the jth dimension; a ¯ j denotes the mean of the sample in the jth dimension; σ j denotes the standard deviation of the sample in the jth dimension; and k is a set of parameters.
Furthermore, the mean value and the standard deviation of all samples in each dimension can be calculated by Equations (9) and (10), respectively. Samples that satisfy Equation (11) are put into B j , and then form the boundary sample set B, which will be used for replacing the rest of the samples in the secure sample set S.

3.3. Synthesis of New Samples

In this section, the K-means clustering algorithm is employed in the minority class samples to obtain P clusters, where the clustering center is denoted as C = { c i } , where i = 1 , 2 , 3 , , P , and i denotes the index of the cluster. Firstly, a sample x l is randomly selected from the boundary sample set B, where the subscript l indicates that the sample belongs to the lth cluster. Secondly, the other sample y l is randomly selected from the lth cluster (excluding the cluster center c l ), which satisfies d i s t ( x l , y l ) < d i s t ( x l , c l ) , where d i s t ( A , B ) denotes the Euclidean distance between the sample A and B. Next, the temporary sample t l is generated by using Equation (13). Finally, a new minority class sample s l is generated by using Equation (14). In this way, a balanced dataset can be obtained by adding the synthetic new sample s l to the original dataset and repeating the above steps.
t l = x l + r a n d ( 0 , 1 ) × ( x l y l )
s l = t l + r a n d ( 0 , 1 ) × ( c l t l )

4. Experimental Results and Analysis

4.1. Benchmark Dataset

To validate the effectiveness of the FSDR-SMOTE algorithm, 20 imbalanced datasets were selected from the UCI machine learning dataset. For the multi-class datasets in these 20 datasets, a specific class is regarded as the minority class and the rest of the classes are merged into the majority classes. The details of these 20 datasets are given in Table 1, where the ratio of the number of the majority class samples to that of the minority class samples is known as the imbalance ratio (IR). The number of samples in 20 imbalanced datasets ranges from 132 to 13611, and the IR ranges from 1.25 to 22.69, which can cover a comprehensive evaluation in terms of algorithm performance.

4.2. Evaluation Indicator

When dealing with the two-class imbalanced data classification problem, four evaluation metrics, such as F-measure, G-mean, AUC [34], and MCC [35], are usually calculated based on the confusion matrix for evaluating the classification performance of the oversampling method. The confusion matrix [36] is shown in Table 2, where TP denotes that positive class samples are predicted as positive class samples, FN denotes that positive class samples are predicted as negative class samples, FP denotes that negative class samples are predicted as positive class samples, and TN denotes that negative class samples are predicted as negative class samples.
The formula for calculating the relevant assessment indicators is set out below:
A c c u r a c y = T P + T N T P + F P + F N + T N
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F m e a s u r e = ( 1 + β 2 ) × R e c a l l × P r e c i s i o n β 2 × R e c a l l + P r e c i s i o n
G m e a n = T P R × T N R
M C C = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
Among them, F-measure can comprehensively denote the performance of the classifier on recognizing the minority class samples; G-mean can reflect the comprehensive performance of the classifier on positive and negative classes; AUC is not easily affected by the imbalance ratio and evaluates the overall performance of the classifier objectively. As another effective evaluation index, MCC can provide a more comprehensive performance evaluation since MCC takes these four classification results into account.
In our work, the Wilcoxon signed-rank test (the significance level α = 0.05 ) is employed to assess whether there is a significant difference in performance between the FSDR-SMOTE method and competitors.

4.3. Experimental Setup and Environment

In this subsection, the proposed FSDR-SMOTE is firstly compared with three classical oversampling methods, including SMOTE [9], ADASYN [25], and Borderline-SMOTE [26]. Furthermore, it is also compared with four other representative oversampling methods, which are SMOTE-Tomek [27], SVMSMOTE [28], RN-SMOTE [30], and Random-SMOTE [23]. At the same time, the Random Forest classifier based on a grid search algorithm is employed to optimize its parameters. The 5-fold cross-validation strategy is adopted to ensure the reliability and statistical validity of the experimental results in our work.
All experiments are conducted on a computer with Windows 10, AMD Ryzen 5-3600 processor, NVIDIA GTX 3060 graphics card, 16 GB of RAM, and the programming language is Python v3.9, which is implemented using the PyCharm platform.

4.4. Comparison of Sampling Effects of Different Methods

Figure 5 illustrates the sampling effects of the proposed FSDR-SMOTE algorithm and seven competitors on the Haberman dataset. In Figure 5, blue points represent original majority class samples; red points denote original minority class samples; green points symbolize newly synthesized samples.
It can be seen from Figure 5a that the original dataset contains some noise samples that are located at a significant spatial distance from the minority class samples. Figure 5b–g represents the sampling effects obtained by SMOTE, ADASYN, Borderline-SMOTE, SMOTE-Tomek, SVMSMOTE, and Random-SMOTE, respectively. The ADASYN and Borderline-SMOTE algorithms, as shown in Figure 5, have synthesized more samples in the border area. In contrast, the SMOTE, SMOTE-Tomek, SVMSMOTE, and Random-SMOTE algorithms have generated more samples in the non-border area.
Figure 5h,i displays the sampling renderings of RN-SMOTE and the FSDR-SMOTE proposed in this paper, respectively. Unlike the first six comparison algorithms, both RN-SMOTE and FSDR-SMOTE have taken noise into account in data preprocessing, which can effectively filter the noise in the minority class. Furthermore, the FSDR-SMOTE algorithm can effectively synthesize samples in the minority class boundary area, thereby mitigating the imbalance problem within the sample class.

4.5. Experimental Results

Table 3 shows the comparison of F-measure, G-mean, and MCC indicators of eight methods combined with the Random Forest [37] classifier on 20 imbalanced datasets. The optimal values obtained by different methods on each evaluation metric are shown in bold. As can be seen in Table 3, FSDR-SMOTE outperforms the comparison algorithms in most cases. Specifically, under the F-measure evaluation value, FSDR-SMOTE achieved the best results on 14 out of 20 datasets. Under the G-mean evaluation index, FSDR-SMOTE has the best results on 11 out of 20 datasets. FSDR-SMOTE also achieved 12 best results under the MCC evaluation index. Compared to the optimal results of the other seven comparative algorithms, the F-measure (G-mean and MCC) evaluation metric obtained by the FSDR-SMOTE method has improved by 6.95% (6.23% and 12.35%) on the Hayes-Roth dataset. Meanwhile, on the Wholesale, Haberman, and Blood datasets, the performance of FSDR-SMOTE by improved more than 1% compared with the other seven methods in terms of the F-measure, G-mean, and MCC evaluation.
In addition, in datasets that did not achieve optimal evaluation results, such as Rice, Qsar, Yeast1, etc., the evaluation results obtained by FSDR-SMOTE differed from the optimal results by less than 0.5%. It can be seen from the comprehensive results obtained by the FSDR-SMOTE algorithm that it has certain advantages compared with the seven comparison algorithms.
To further analyze the significant difference between the FSDR-SMOTE and these competitors, the Wilcoxon signed rank test ( α = 0.05 ) is performed on the results of F-measure and MCC, with the null hypothesis that all algorithms have the same effect. Obtained test results are shown in Table 4.
It can be seen from Table 4 that the null hypothesis is accepted by FSDR-SMOTE and SMOTE-Tomek, whereas it is rejected by the remaining six algorithms. This means that FSDR-SMOTE significantly outperforms these six algorithms in a statistically significant way. Although FSDR-SMOTE does not obtain significant superiority compared with SMOTE-Tomek, it achieves a more positive rank R+ on both key metrics, i.e., F-measure and MCC, which suggests that FSDR-SMOTE outperforms SMOTE-Tomek more often on these 20 datasets in the comparison experiments.
To visualize the advantages of the FSDR-SMOTE method, the ROC curves of the eight algorithms on six different datasets were plotted. In Figure 6, on the Hayes-Roth, Australian, Haberman, and Yeast1 datasets, the ROC curve clearly shows that the FSDR-SMOTE method consistently outperforms other comparison algorithms. This indicates that the FSDR-SMOTE algorithm has achieved higher AUC values on these datasets, demonstrating its superior classification effects compared to the other seven algorithms. In conclusion, the combination of the FSDR-SMOTE algorithm and the Random Forest classifier exhibits better performance on most datasets, thereby proving the algorithm’s stability.
Furthermore, the proposed FSDR-SMOTE is combined with different classifiers to validate the model performance. These classifiers include Random Forest [37], Adaboost [38], KNN [39], Gaussian Naive Bayes (BC) [40], and Logistic Regression (LR) [41]. Figure 7 shows the experimental curves obtained by FSDR-SMOTE with five different classifiers on 20 datasets in terms of four metrics, i.e., F-measure, G-mean, AUC, and MCC. In Figure 7, the abscissa represents 20 datasets, which are numbered according to the order in Table 1. It can be seen that FSDR-SMOTE combined with Random Forest achieves optimal or near-optimal classification accuracies on most of these 20 datasets, which can also show the compatibility of FSDR-SMOTE with different classifiers.

4.6. Ablation Study

To further investigate the effectiveness of data preprocessing techniques in the FSDR-SMOTE method, we conducted ablation studies on eight imbalanced datasets, namely, Haberman, Blood, Wisconsin, German, Australian, Qsar, Yeast1, and Abalone11.
Figure 8 intuitively illustrates the changes in the number of minority class samples across these eight different datasets before and after data preprocessing. As can be inferred from the figure, the application of the improved Tukey rule for data preprocessing effectively reduces the number of minority class samples, primarily by removing noise samples. Notably, after preprocessing, the proportion of minority class samples within the original minority class samples remains between 84.85% and 89.54%. This indicates that the FSDR-SMOTE method successfully eliminates samples above the threshold while effectively retaining most of the genuine and significant data, thereby minimizing the impact on minority class diversity.
Figure 9 presents the evaluation results using F-measure, G-mean, AUC, and MCC. It is evident from the figure that all evaluation metrics of the FSDR-SMOTE method outperform those of the competitors on the eight imbalanced datasets without preprocessing. These results demonstrate that for most datasets, the improvement in evaluation metrics exceeds 1%, particularly on the Blood and Australian datasets, where the improvement in MCC reaches as high as 6.05% and 7.19%, respectively. These substantial improvements underscore the effectiveness of employing the improved Tukey rule for data preprocessing in the FSDR-SMOTE method.

5. Discussion

The classic oversampling method, SMOTE, can effectively enhance the classification accuracy of minority class samples in imbalanced datasets. However, SMOTE is not sensitive to noise and is prone to intra-class imbalance of samples, potentially introducing noise and generating redundant samples. Effective data preprocessing is a crucial step in addressing the noise problem, but there is a risk of inadvertently deleting real data in this process. Thus, our research focuses on effectively eliminating noise data while preserving most of the real data. In the FSDR-SMOTE method, we utilize the improved Tukey rule [32] for data preprocessing. The experimental results showed that the preprocessed dataset retained at least 84.85% of minority class samples. After using an improved Tukey rule for data preprocessing on eight datasets of ablation studies, the F-measure (G-mean, AUC, and MCC) evaluation values increased by 1.59% (1.78%, 1.70%, and 3.38%), respectively.
Additionally, our proposed FSDR-SMOTE method takes into account the imbalance within sample classes. During the sample synthesis process, we combined K-means [33] clustering with an improved three-point interpolation strategy in Random-SMOTE [23]. We conducted comparative experiments to evaluate the FSDR-SMOTE method against seven other existing sampling methods (SMOTE [9], ADASYN [25], Borderline-SMOTE [26], SMOTE-Tomek [27], SVMSMOTE [28], RN-SMOTE [30], and Random-SMOTE [23]). FSDR-SMOTE achieved an average improvement of 0.72% (0.48% and 0.98%) over the second-best results across the 20 datasets in terms of F-measure (G-mean and MCC), respectively. Thereby, fully considering the distribution characteristics of samples and adopting a more efficient sample synthesis strategy are keys to enhancing model classification performance.

6. Conclusions

In this paper, a new oversampling method (FSDR-SMOTE) is proposed to deal with the imbalanced data classification problem. In FSDR-SMOTE, the noisy samples are firstly removed based on the improved Tukey criterion in the data preprocessing stage. Secondly, a new feature standard deviation method is proposed to divide these samples into boundary samples and safety ones in view of the impact of different distributions of synthetic samples on the model performance. Finally, the strategy of synthesizing new samples in Random-SMOTE is improved. The experimental results show that the average evaluation values of F-measure, G-mean, and MCC obtain by FSDR-SMOTE on 20 imbalanced datasets reach 93.31%, 93.16%, and 86.53%, respectively. The average values of all evaluations are superior to the other seven comparison methods. Furthermore, FSDR-SMOTE was combined with Random Forest and the other four classifiers to validate the compatibility of FSDR-SMOTE.
However, FSDR-SMOTE performs poorly in dealing with datasets with a higher imbalance ratio (IR > 8). Therefore, our future work will devote to constructing new hybrid strategies for optimizing the FSDR-SMOTE method proposed in this paper.

Author Contributions

Conceptualization, Y.Z. and L.D.; methodology, Y.Z. and B.W.; software, Y.Z.; validation, Y.Z.; formal analysis, Y.Z. and L.D.; investigation, Y.Z.; resources, Y.Z.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, L.D. and B.W.; visualization, Y.Z.; supervision, L.D. and B.W.; project administration, L.D. and B.W.; funding acquisition, L.D. and B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by grants from the National Natural Science Foundation of China (61806204), the Basic Public Welfare Research Project of Zhejiang Province (LGF22F020020), and the Research Fund Project of Zhejiang Sci-Tech University Longgang Research Institute (LGYJY2023003).

Data Availability Statement

The data used in this article come from the UCI Machine Learning Repository, http://archive.ics.uci.edu/ (accessed on 7 March 2024).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tomašev, N.; Mladenić, D. Class imbalance and the curse of minority hubs. Knowl. Based Syst. 2013, 53, 157–172. [Google Scholar] [CrossRef]
  2. Vasighizaker, A.; Jalili, S. C-PUGP: A cluster-based positive unlabeled learning method for disease gene prediction and prioritization. Comput. Biol. Chem. 2018, 76, 23–31. [Google Scholar] [CrossRef] [PubMed]
  3. Jurgovsky, J.; Granitzer, M.; Ziegler, K.; Calabretto, S.; Portier, P.E.; He-Guelton, L.; Caelen, O. Sequence classification for credit-card fraud detection. Expert Syst. Appl. 2018, 100, 234–245. [Google Scholar] [CrossRef]
  4. Malhotra, R.; Kamal, S. An empirical study to investigate oversampling methods for improving software defect prediction using imbalanced data. Neurocomputing 2019, 343, 120–140. [Google Scholar] [CrossRef]
  5. Zhou, X.; Hu, Y.; Liang, W.; Ma, J.; Jin, Q. Variational LSTM enhanced anomaly detection for industrial big data. IEEE Trans. Ind. Inform. 2020, 17, 3469–3477. [Google Scholar] [CrossRef]
  6. Tao, X.; Li, Q.; Guo, W.; Ren, C.; Li, C.; Liu, R.; Zou, J. Self-adaptive cost weights-based support vector machine cost-sensitive ensemble for imbalanced data classification. Inf. Sci. 2019, 487, 31–56. [Google Scholar] [CrossRef]
  7. Daneshfar, F.; Aghajani, M.J. Enhanced text classification through an improved discrete laying chicken algorithm. Expert Syst. 2024, e13553. [Google Scholar] [CrossRef]
  8. Revathy, V.; Pillai, A.S.; Daneshfar, F. LyEmoBERT: Classification of lyrics’ emotion and recommendation using a pre-trained model. Procedia Comput. Sci. 2023, 218, 1196–1208. [Google Scholar] [CrossRef]
  9. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  10. Nadi, A.A.; Gildeh, B.S.; Kazempoor, J.; Tran, K.D.; Tran, K.P. Cost-effective optimization strategies and sampling plan for Weibull quantiles under type-II censoring. Appl. Math. Model. 2023, 116, 16–31. [Google Scholar] [CrossRef]
  11. Tao, X.; Chen, W.; Li, X.; Zhang, X.; Li, Y.; Guo, J. The ensemble of density-sensitive SVDD classifier based on maximum soft margin for imbalanced datasets. Knowl. Based Syst. 2021, 219, 106897. [Google Scholar] [CrossRef]
  12. Li, W.; Sun, S.; Zhang, S.; Zhang, H.; Shi, Y. Cost-Sensitive Approach to Improve the HTTP Traffic Detection Performance on Imbalanced Data. Secur. Commun. Netw. 2021, 2021, 6674325. [Google Scholar] [CrossRef]
  13. Li, J.; Zhu, Q. A boosting self-training framework based on instance generation with natural neighbors for K nearest neighbor. Appl. Intell. 2020, 50, 3535–3553. [Google Scholar] [CrossRef]
  14. Xia, S.; Wang, G.; Chen, Z.; Duan, Y. Complete random forest based class noise filtering learning for improving the generalizability of classifiers. IEEE Trans. Knowl. Data Eng. 2018, 31, 2063–2078. [Google Scholar] [CrossRef]
  15. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  16. Wei, G.; Mu, W.; Song, Y.; Dou, J. An improved and random synthetic minority oversampling technique for imbalanced data. Knowl. Based Syst. 2022, 248, 108839. [Google Scholar] [CrossRef]
  17. Meng, D.; Li, Y. An imbalanced learning method by combining SMOTE with Center Offset Factor. Appl. Soft Comput. 2022, 120, 108618. [Google Scholar] [CrossRef]
  18. Shrifan, N.H.; Akbar, M.F.; Isa, N.A.M. An adaptive outlier removal aided k-means clustering algorithm. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 6365–6376. [Google Scholar] [CrossRef]
  19. Liang, X.; Jiang, A.; Li, T.; Xue, Y.; Wang, G. LR-SMOTE—An improved unbalanced data set oversampling based on K-means and SVM. Knowl. Based Syst. 2020, 196, 105845. [Google Scholar] [CrossRef]
  20. Zhang, A.; Yu, H.; Zhou, S.; Huan, Z.; Yang, X. Instance weighted SMOTE by indirectly exploring the data distribution. Knowl. Based Syst. 2022, 249, 108919. [Google Scholar] [CrossRef]
  21. Cheng, D.; Zhu, Q.; Huang, J.; Yang, L.; Wu, Q. Natural neighbor-based clustering algorithm with local representatives. Knowl. Based Syst. 2017, 123, 238–253. [Google Scholar] [CrossRef]
  22. Dong, Y.; Wang, X. A new over-sampling approach: Random-SMOTE for learning from imbalanced data sets. In Proceedings of the Knowledge Science, Engineering and Management: 5th International Conference, KSEM 2011, Irvine, CA, USA, 12–14 December 2011; Proceedings 5; Springer: Berlin/Heidelberg, Germany, 2011; pp. 343–352. [Google Scholar] [CrossRef]
  23. Bader-El-Den, M.; Teitei, E.; Perry, T. Biased random forest for dealing with the class imbalance problem. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 2163–2172. [Google Scholar] [CrossRef] [PubMed]
  24. Rekha, G.; Tyagi, A.K.; Sreenath, N.; Mishra, S. Class imbalanced data: Open issues and future research directions. In Proceedings of the 2021 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 27–29 January 2021; pp. 1–6. [Google Scholar] [CrossRef]
  25. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 1322–1328. [Google Scholar] [CrossRef]
  26. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, during 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar] [CrossRef]
  27. Batista, G.E.; Prati, R.C.; Monard, M.C. A study of the behavior of several methods for balancing machine learning training data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  28. Nguyen, H.M.; Cooper, E.W.; Kamei, K. Borderline over-sampling for imbalanced data classification. Int. J. Knowl. Eng. Soft Data Paradig. 2011, 3, 4–21. [Google Scholar] [CrossRef]
  29. Huang, S.; Cai, N.; Pacheco, P.P.; Narrandes, S.; Wang, Y.; Xu, W. Applications of support vector machine (SVM) learning in cancer genomics. Cancer Genom. Proteom. 2018, 15, 41–51. [Google Scholar]
  30. Arafa, A.; El-Fishawy, N.; Badawy, M.; Radad, M. RN-SMOTE: Reduced Noise SMOTE based on DBSCAN for enhancing imbalanced data classification. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 5059–5074. [Google Scholar] [CrossRef]
  31. Schubert, E.; Sander, J.; Ester, M.; Kriegel, H.P.; Xu, X. DBSCAN revisited, revisited: Why and how you should (still) use DBSCAN. ACM Trans. Database Syst. (TODS) 2017, 42, 1–21. [Google Scholar] [CrossRef]
  32. Huyghues-Beaufond, N.; Tindemans, S.; Falugi, P.; Sun, M.; Strbac, G. Robust and automatic data cleansing method for short-term load forecasting of distribution feeders. Appl. Energy 2020, 261, 114405. [Google Scholar] [CrossRef]
  33. Ahmed, M.; Seraj, R.; Islam, S.M.S. The k-means algorithm: A comprehensive survey and performance evaluation. Electronics 2020, 9, 1295. [Google Scholar] [CrossRef]
  34. Su, C.T.; Chen, L.S.; Yih, Y. Knowledge acquisition through information granulation for imbalanced data. Expert Syst. Appl. 2006, 31, 531–541. [Google Scholar] [CrossRef]
  35. Zhu, Q. On the performance of Matthews correlation coefficient (MCC) for imbalanced dataset. Pattern Recognit. Lett. 2020, 136, 71–80. [Google Scholar] [CrossRef]
  36. Visa, S.; Ramsay, B.; Ralescu, A.L.; Van Der Knaap, E. Confusion matrix-based feature selection. Maics 2011, 710, 120–127. [Google Scholar]
  37. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  38. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  39. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  40. Venkata, P.; Pandya, V. Data mining model and Gaussian Naive Bayes based fault diagnostic analysis of modern power system networks. Mater. Today Proc. 2022, 62, 7156–7161. [Google Scholar] [CrossRef]
  41. Hosmer, D.W., Jr.; Lemeshow, S.; Sturdivant, R.X. Applied Logistic Regression; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
Figure 1. Schematic diagram of SMOTE-Tomek method.
Figure 1. Schematic diagram of SMOTE-Tomek method.
Mathematics 12 01709 g001
Figure 2. Overview of the RN-SMOTE method [30].
Figure 2. Overview of the RN-SMOTE method [30].
Mathematics 12 01709 g002
Figure 3. Schematic diagram of Random-SMOTE for synthesizing the new sample.
Figure 3. Schematic diagram of Random-SMOTE for synthesizing the new sample.
Mathematics 12 01709 g003
Figure 4. The flowchart of the FSDR-SMOTE.
Figure 4. The flowchart of the FSDR-SMOTE.
Mathematics 12 01709 g004
Figure 5. Sampling effects of different algorithms on the dataset Haberman. Red dots denote minority samples, blue dots denote majority samples, and green dots are new synthetic samples.
Figure 5. Sampling effects of different algorithms on the dataset Haberman. Red dots denote minority samples, blue dots denote majority samples, and green dots are new synthetic samples.
Mathematics 12 01709 g005
Figure 6. ROC curves of eight algorithms on six datasets.
Figure 6. ROC curves of eight algorithms on six datasets.
Mathematics 12 01709 g006
Figure 7. The different evaluation index results obtained by FSDR-SMOTE with different classifiers.
Figure 7. The different evaluation index results obtained by FSDR-SMOTE with different classifiers.
Mathematics 12 01709 g007
Figure 8. Sample size of different datasets before and after preprocessing.
Figure 8. Sample size of different datasets before and after preprocessing.
Mathematics 12 01709 g008
Figure 9. Comparison of evaluation indicators for dataset ablation study.
Figure 9. Comparison of evaluation indicators for dataset ablation study.
Mathematics 12 01709 g009
Table 1. Dataset information.
Table 1. Dataset information.
DatasetSizeAttributesIRMinorityMajority
ClassNumberClassNumber
Australian690151.2513072383
Rice381071.34Cammeo1630Osmancik2180
Hayes-Roth13251.59051181
Wisconsin683101.8642392444
Qsar1055411.97RB355NRB700
Wholesale44082.121421298
German1000202.3323001700
Yeast11484102.46NUC429remainder1055
Haberman30632.782811225
Blood74853.211780570
Ecoli133683.31im77remainder259
Glass621496.38629remainder185
Dry_bean413,611157.35CALI1630remainder11,981
Abalone11417787.5811487remainder3690
Yeast41484108.1ME3163remainder1321
Ecoli433688.6imU35remainder301
Yeast4vs05679528109.35ME251MIT, ME3, EXC, VAC, ERL477
Climate5402110.740461494
Pageblocks254731015.642329remainder5144
Pageblocks154731022.693, 4, 52311, 25242
Table 2. Confusion matrix.
Table 2. Confusion matrix.
Predicted PositivePredicted Negative
PositiveTrue Positive (TP)False Positive (FP)
NegativeFalse Negative (FN)True Negative (TN)
Table 3. Experimental results obtained by the FSDR-SMOTE and seven comparison methods.
Table 3. Experimental results obtained by the FSDR-SMOTE and seven comparison methods.
DatasetEISMOTEADASYNBSSTKSVMSRNSRSFSDS
AustralianF10.87300.86400.86210.91180.87750.88990.88050.9162
G0.87550.86060.86640.91300.88070.89270.88570.9180
MCC0.75120.72260.73520.82670.76310.78680.77290.8367
RiceF10.92930.90920.91900.95190.92070.93010.92670.9463
G0.92970.91320.92130.95180.92200.93040.93050.9465
MCC0.85950.82880.84580.90360.84510.86080.86120.8932
Hayes-RothF10.84700.84700.84060.82940.84330.83480.81890.9165
G0.85000.82780.83460.83650.84550.83980.83300.9123
MCC0.71160.67410.67830.68230.69850.69030.67730.8351
WisconsinF10.97740.97790.98020.97950.97760.98150.97650.9818
G0.97750.97840.98020.97980.97780.98130.97850.9818
MCC0.95530.95840.96130.96030.95650.96320.95780.9636
QsarF10.90100.90510.90680.91490.90310.91740.89580.9182
G0.90140.90900.90840.91470.90550.91770.90440.9164
MCC0.80280.81980.81820.82990.81240.83510.80880.8340
WholesaleF10.93510.92210.93390.94360.92680.93720.93330.9566
G0.93730.92390.93680.94520.92780.93830.93960.9566
MCC0.87600.85000.87620.89140.85790.87840.88030.9147
GermanF10.84000.83800.83920.84820.83070.84450.83260.8591
G0.83810.83730.83970.84810.83030.84280.84400.8424
MCC0.67700.67530.68000.69650.66160.68690.68610.7049
Yeast1F10.84140.82800.83340.85090.83600.85610.82240.8561
G0.84440.83650.83780.85380.83760.85820.83810.8526
MCC0.69050.67760.67840.70970.67600.71690.67990.7014
HabermanF10.77720.75420.76660.79360.76750.76290.76760.8092
G0.78260.76340.77180.80310.76800.76810.79150.8073
MCC0.57180.53220.54940.61310.53930.54500.59410.6162
BloodF10.75420.75750.77700.79500.78840.78410.80010.8290
G0.75650.76060.77470.78640.78260.78210.81960.8245
MCC0.51620.52550.55140.57710.56950.56590.63560.6524
Ecoli1F10.93970.92320.93630.94090.92290.93150.91920.9453
G0.94140.92630.93830.94310.92690.93320.92660.9472
MCC0.88800.85800.88260.88980.85860.87360.86220.8956
Glass6F10.99720.99700.99730.99630.99680.99780.99510.9985
G0.99720.99700.99730.99640.99660.99780.99510.9985
MCC0.99470.99420.99460.99270.99150.99510.99100.9971
Dry_bean4F10.98970.98790.99080.99060.98980.99040.98830.9941
G0.98980.98800.99090.99060.98990.99040.98950.9942
MCC0.97960.97620.98190.98120.98000.98090.97960.9883
Abalone11F10.92020.92200.92630.92000.92500.91960.84440.8916
G0.92190.92440.92760.92190.92520.92150.85920.8896
MCC0.84580.85210.85680.84580.85040.84540.73580.7802
Yeast4F10.97130.97040.97440.97530.97390.97300.96980.9750
G0.97140.97080.97460.97550.97400.97320.97230.9750
MCC0.94300.94280.95020.95150.94820.94710.94690.9502
Ecoli4F10.95100.95580.95450.95440.95580.95590.94860.9623
G0.95220.95680.95540.95550.95720.95560.94980.9628
MCC0.90640.91650.91410.91280.91910.91420.91420.9265
Yeast4vs05679F10.95890.95300.95870.96180.96490.95610.91490.9499
G0.95900.95380.95960.96240.96540.95780.92160.9495
MCC0.91880.90930.92100.92560.93180.91560.85410.9009
ClimateF10.98060.97630.97240.97820.98100.98040.96700.9753
G0.98090.97650.97250.97820.97290.97990.96990.9756
MCC0.96250.95350.94490.95700.94860.96090.94110.9508
Pageblocks2F10.99250.99230.99390.99230.99400.99190.99190.9948
G0.99250.99220.99380.99230.99400.99190.99300.9947
MCC0.98490.98450.98770.98470.98810.98370.98580.9895
Pageblocks1F10.98810.98730.98850.98870.98820.98820.97920.9870
G0.98820.98740.98850.98880.98820.98820.98080.9870
MCC0.97660.97510.97720.97780.97660.97670.96410.9740
EI: Evaluation Indicators; F1: F-measure; G: G-mean; BS: Borderline-SMOTE; STK: SMOTE-Tomek; SVMS: SVMSMOTE; RNS: RN-SMOTE; RS: Random-SMOTE; FSDS: FSDR-SMOTE. The best results are in bold.
Table 4. The results obtained by FSDR-SMOTE and competitors on Wilcoxon cosigned-rank test.
Table 4. The results obtained by FSDR-SMOTE and competitors on Wilcoxon cosigned-rank test.
MethodsSMOTEADASYNBSSTKSVMSRNSRS
EIF1
R+177188181156175169210
R−3322295435410
Assumingrejectrejectrejectacceptrejectrejectreject
SelectFSDSFSDSFSDSBothFSDSFSDSFSDS
EIMCC
R+172184176136179158210
R−3826347431520
Assumingrejectrejectrejectacceptrejectrejectreject
SelectFSDSFSDSFSDSBothFSDSFSDSFSDS
EI: Evaluation Indicators; F1: F-measure; BS: Borderline-SMOTE; STK: SMOTE-Tomek; SVMS: SVMSMOTE; RNS: RN-SMOTE; RS: Random-SMOTE; FSDS: FSDR-SMOTE.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Deng, L.; Wei, B. Imbalanced Data Classification Based on Improved Random-SMOTE and Feature Standard Deviation. Mathematics 2024, 12, 1709. https://doi.org/10.3390/math12111709

AMA Style

Zhang Y, Deng L, Wei B. Imbalanced Data Classification Based on Improved Random-SMOTE and Feature Standard Deviation. Mathematics. 2024; 12(11):1709. https://doi.org/10.3390/math12111709

Chicago/Turabian Style

Zhang, Ying, Li Deng, and Bo Wei. 2024. "Imbalanced Data Classification Based on Improved Random-SMOTE and Feature Standard Deviation" Mathematics 12, no. 11: 1709. https://doi.org/10.3390/math12111709

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop