Next Article in Journal
Combining Fractional Derivatives and Machine Learning: A Review
Next Article in Special Issue
Kernel Partial Least Squares Feature Selection Based on Maximum Weight Minimum Redundancy
Previous Article in Journal
Domain Adaptation Principal Component Analysis: Base Linear Method for Learning with Out-of-Distribution Data
Previous Article in Special Issue
The Determination Method of Satisfactory Consistency of the Interval Number Pairwise Comparisons Matrix Based on Submatrix
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Relative Density-Based Intuitionistic Fuzzy SVM for Class Imbalance Learning

1
School of Mathematics and Statistics, Xi’dian University, Xi’an 710071, China
2
School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 34; https://doi.org/10.3390/e25010034
Submission received: 26 November 2022 / Revised: 20 December 2022 / Accepted: 21 December 2022 / Published: 24 December 2022

Abstract

:
The support vector machine (SVM) has been combined with the intuitionistic fuzzy set to suppress the negative impact of noises and outliers in classification. However, it has some inherent defects, resulting in the inaccurate prior distribution estimation for datasets, especially the imbalanced datasets with non-normally distributed data, further reducing the performance of the classification model for imbalance learning. To solve these problems, we propose a novel relative density-based intuitionistic fuzzy support vector machine (RIFSVM) algorithm for imbalanced learning in the presence of noise and outliers. In our proposed algorithm, the relative density, which is estimated by adopting the k-nearest-neighbor distances, is used to calculate the intuitionistic fuzzy numbers. The fuzzy values of the majority class instances are designed by multiplying the score function of the intuitionistic fuzzy number by the imbalance ratio, and the fuzzy values of minority class instances are assigned the intuitionistic fuzzy membership degree. With the help of the strong capture ability of the relative density to prior information and the strong recognition ability of the intuitionistic fuzzy score function to noises and outliers, the proposed RIFSVM not only reduces the influence of class imbalance but also suppresses the impact of noises and outliers, and further improves the classification performance. Experiments on the synthetic and public imbalanced datasets show that our approach has better performance in terms of G-Means, F-Measures, and AUC than the other class imbalance classification algorithms.

1. Introduction

The class imbalance learning problem in binary classification occurs when the number of one category is significantly greater than that of the other category [1,2]. The imbalance datasets exist in various application domains, such as biological recognition  [3,4], medical diagnosis [5,6], fault diagnosis [7], credit card fraud detection [8,9], and text categorization [10,11], etc. When tackling imbalanced datasets, due to the main role of the majority class, the traditional classification methods designed for balanced datasets may not always achieve good classification performance for the minority class. Therefore, many improved classical algorithms and novel algorithms [12] have emerged to deal with imbalanced classification. As one of the most classic classification algorithms, SVM [13,14,15] shows relatively more robustness than other methods in imbalanced classification problems, but it is still unsatisfactory. Since the traditional SVM considers all instances equally and ignores the difference between the majority and minority classes, thus the decision surface may be biased toward the majority class instances when handling imbalanced datasets [16,17], especially when there exist noises and outliers.
On the one hand, the fuzzy support vector machine (FSVM) was originally designed to deal with the problem of outliers and noises [18]. The FSVM algorithm uses fuzzy membership functions to assign different fuzzy membership values (MVs) for each instance and redefines the SVM. It allows different instances to make different contributions to the generation of the separated hyperplane. Strategies for calculating fuzzy MVs are very important for the performance of an FSVM method. Currently, the more commonly used method is to give fuzzy value to instances only according to their membership degree of a certain class not the relationship between classes. It results in inaccurate sample distribution information. Based on the intuitionistic fuzzy set [19], Ha and Wang [20] proposed an intuitionistic fuzzy support vector machine (IFSVM), the fuzzy value of which is calculated based on the membership and non-membership degrees determined by the distribution information of the instances. Then, based on intuitionistic fuzzy sets and kernel functions, Ha et al. proposed a new fuzzy support vector machine [21], which is superior in dealing with outliers and noise. Salim et al. [22] introduced the intuitionistic fuzzy number into the twin support vector machine, which evaluated the contribution of instances to the separated hyperplane according to the score function of instances, thus improving the disadvantage that the support vector far from the center of the class is given a lower fuzzy value. However, we know that the contribution of a sample to the classification hyperplane cannot be accurately described only by its distance from the sample center. Moreover, class centers are obtained based on all instances, so outliers may cause class center offset. In addition, the non-membership degree of sparse instances will also be inaccurate due to improper parameter selection.
Many previous works have shown that FSVMs outperform standard SVM in terms of accuracy and robustness, but they are still affected by the imbalanced data distribution. The state-of-the-art techniques to develop the performance of imbalanced data classification are generally called class imbalance learning [23,24] techniques. These techniques can be broadly classified as external and internal. The external techniques deal with the data sets before training a classifier, and the internal techniques create or modify the SVM. We will focus on the latter in this paper. To address the problem of imbalanced datasets classification disturbed by outlier and noise. Batuwita and Palade [24] combine FSVM with cost-sensitive learning to obtain the FSVM-CIL serial algorithms. The six membership functions of it are obtained by combining three different methods with two kinds of decaying functions. Most of them calculate the importance of each instance by Euclidean distance, so the membership functions using this metric are sensitive to the dimension of the data distribution in feature space. Liu [25] modified the distance measure and proposed the Gaussian fuzzy support vector machine (GFSVM), which is an extension of the FSVM-CIL. Experimental results indicated that this improved approach has better performance in class imbalanced learning. Based on the principle of space consistency, Tao et al. presented affinity and class probability-based FSVM (ACFSVM) [26], which can suppress the impact of noises and outliers in the majority class instances on the model. Deep et al. [27] presented an entropy-based entropy-based fuzzy twin support vector machine (EFTWSVM) for imbalanced datasets, which uses the information entropy of instances to determine the membership degree of each instance in the fuzzy twin support vector machine. Richhariya et al. [28] presented a robust fuzzy least squares twin SVM (RFLSTSVM), which employs the 2-norm of the slack variables to make the optimization problem strongly convex and uses the imbalance ratio of instances in calculating the membership degree of samples. Based on the relative density which is a more robust prior information extraction method, Yu et al. [29] presented two fuzzy SVM algorithms to tackle the problem of class imbalance learning, one is the fuzzy SVM algorithm based on the within-class relative density information (FSVM-WD), and the other is fuzzy SVM based on between-class relative density information (FSVM-WD). These two methods overcome the inherent defect that traditional methods are not accurate in the estimation of prior data distribution. Most of these methods set the fuzzy value of minority class instances to 1 to highlight their importance, but these do not consider the distribution characteristics of minority class instance.
In this paper, we propose a new relative density-based intuitionistic FSVM termed RIFSVM to deal with both the problem of class imbalance and noise/outlier. In our proposed model, the relative density obtained by the K-nearest neighbors-based probability density estimation (KNN-PDE) method [30,31] is utilized to calculate the membership degree and non-membership degree of instances. Because the relative density is adaptive for different data distribution types, and can well reflect the prior distribution information of data, it is more robust than the Euclidean distance-based measure. In addition, this method does not need to obtain the density information of all sample points, but only needs to calculate the distance between k neighbors of each sample and estimate its probability density distribution in feature space. Then, for instances of the majority class, we calculate the score function according to their membership degree and non-membership degree, and then set the product of the score function and the imbalance ratio as the final fuzzy value. For the instances of minority classes, the membership degree is directly set to the fuzzy value, which avoids the situation that the non-membership of minority classes is too high due to sparsity, so some points are misjudged as noise. Comprehensive experimental results on the synthetic and benchmark imbalanced datasets demonstrate the superiority of the RIFSVM to the existing and state-of-the-art class imbalance classification algorithms.
The remainder of this paper is structured as follows: Fuzzy-type support vector machines and fuzzy membership functions summarised in Section 2. Our proposed algorithm is detailedly described in Section 3. The experimental results and analysis to validate the validity of the algorithm are reported in Section 4. Finally, we conclude this paper and indicate future work in Section 5.

2. Fuzzy-Type Support Vector Machines and Fuzzy Membership Functions

2.1. Fuzzy-Type Support Vector Machines (FSVMs)

Unlike standard SVM, FSVM [18,32] assigns the corresponding membership degree to each instance according to the specified membership function. Assume that we have a binary classification problem with the dataset T = { ( x i , y i ) i = 1 , 2 , , N } , where x i R d represents an d-dimensional input vector, and y i { 1 , 1 } refers to the corresponding class label. Then, for the given T s = { ( x i , y i , s i ) i = 1 , 2 , , N } , where s i denotes the fuzzy membership value of the i-th instance, which indicates the significance of the associated instance, the goal of FSVM is to learn parameters ( w , b ) of a hypothesis h ( x ) = w , φ ( x i ) + b from the optimization problem:
min w , b 1 2 w 2 + C i = 1 N s i ξ i s . t . y i ( w T φ ( x i ) + b ) + ξ i 1 , ξ i 0 , i = 1 , 2 , , N ,
where w H , b R , tradeoff parameters C > 0 , 0 s i 1 and φ is a feature mapping that maps the instance space into a high feature space. The slack variable ξ i denotes the measured error of the instance x i in FSVM, and the membership s i of a data point x i is incorporated into the objective function in the FSVM optimization problem (1), which is the only difference between the original SVM optimization problem and it.
The above optimization problem can be solved by solving the quadratic programming of the following dual form:
max α i = 1 N α i 1 2 i = 1 N j = 1 N α i α j y i y j K ( x i , x j ) s . t . i = 1 N y i α i = 0 , 0 α i s i C , i = 1 , N ,
where α = [ α 1 , α 2 , α N ] is the vector of Lagrange multipliers, and K ( x i , x j ) being a kernel function, i.e., the inner product of the feature vectors x i and x j in the feature space φ ( x i ) , φ ( x j ) . It is observed that the upper bound on the value of α i differs between the dual optimization problems for the standard SVM and FSVM.
The optimal values α * can be obtained by solving the optimization (2), and the optimal weight vector w * and b * can be expressed as [13]:
w * = i = 1 N α * i y i φ ( x i )
and
b * = y i i = 1 N α * i y i K ( x i , x j )
respectively. Then the decision function can be expressed as:
h ( x ) = s i g n ( i = 1 N α * i y i K ( x i , x ) + b * ) .
The fuzzy value s i represents the weight of parameter ξ i in the objective function, reflecting the importance of the corresponding instance to the classification hyperplane. Then, the classification hyperplane can be made more reasonable by giving small fuzzy MVs for the noises and outliers. Obviously, FSVM can effectively handle outliers and noise with the help of well-defined membership functions. Therefore, defining an appropriate fuzzy membership function (MF) becomes a key issue to improve the generalization performance of FSVM.

2.2. Common Fuzzy MFs and Their Limitations for Class Imbalance Learning

Here, we briefly discuss some classical and new informative heuristic membership functions for general purposes, such as linear and exponential fuzzy functions [24], and the Gaussian fuzzy [33] function:
μ l i n ( x i ) = 1 d i max ( d i ) + Δ ;
μ exp ( x i ) = 2 1 + exp ( β d i ) , β [ 0 , 1 ] ;
μ g a u ( x i ) = exp ( d i μ 2 2 σ 2 ) ,
where d i indicates the Euclidean distance between the instance x i and the reference objects, Δ is a small positive value and β is a parameter that determines the degree of the decay, and μ and σ being the separate mean and standard deviations to be tuned during the training process.
The popular three distance reference objects contain within-class centroid, real separating hyperplane, and estimated sphere centroid. Relevant experiments in [29] show that when the datasets are approximatively subject to the standardized normal distribution, the within-class centroid reference is reliable. However, it is not a reliable reference anymore for the datasets with non-normal distribution, e.g., the datasets with small disjunctions or manifold distribution. The estimated sphere centroid reference has the same drawbacks, especially on the highly imbalanced datasets. The real separating hyperplane reference is significantly better than the two reference object. However, it is still sensitive to the degree of imbalance due to the initial separation hyperplane would be biased towards the minority class. That is, the classification model obtained by is still biased toward one of the classes.
The fuzzy value of these methods are based on the MVs of instances to their own classes, ignoring the uncertainty of instances. However, only assigning fuzzy values to instances according to the membership degree of a certain class of sample points cannot accurately describe the distribution information of instances. Therefore, we are interested in intuitionistic fuzzy sets proposed by Ha et al. [21], which considers both the membership and the non-membership degrees of instances.

2.3. Intuitionistic Fuzzy MF and Its Limitation for Class Imbalance Learning

Definition 1.
Let X be a nonempty set, then an intuitionistic fuzzy set in a universe X, denoted A, is defined as:
A = { ( x , μ A ( x ) , ν A ( x ) ) | x X } ,
where μ A : X [ 0 , 1 ] and ν A : X [ 0 , 1 ] are the degrees of membership and nonmembership functions of x X respectively, and 0 μ A ( x ) + ν A ( x ) 1 .
The score function of x X can be determined as follows:
H ( x ) = 1 ν A ( x ) 2 μ A ( x ) ν A ( x ) ,
The score function calculates the score values of instances based on their membership and non-membership values. These values can be used to compare the levels of membership among several examples of the same class.
For a binary classification problem, consider intuitionistic fuzzy sets. The instances can be converted into an intuitionistic fuzzy number (IFN) as: T = { ( x i , y i , μ i , ν i ) i = 1 , 2 , , N } , where μ i and ν i represent the degrees of membership function and nonmembership functions of x i respectively. Next, the membership and nonmembership functions for each instance are defined as follows.
(1)
Membership Function:
μ ( x i ) = 1 φ ( x i ) c + R + + δ , y i = + 1 , 1 φ ( x i ) c R + δ , y i = 1 ,
where δ > 0 is an adjustable parameter, and the c + ( c ) indicates the center of the positive (negative) class. The R + ( R ) indicates the radius of the positive (negative) class. The φ is a feature mapping that maps the instance space into some feature space. The φ ( x i ) c ± represents the distance from input instance to the corresponding class center. The centers of two class can be written as:
c ± = 1 N ± y i = ± 1 φ ( x i ) ,
where N + and N represent the number of positive and negative instances respectively.
The radius of two classes can be measured by
R ± = max y i = ± 1 φ ( x i ) c ± ,
where
φ ( x i ) c + = φ ( x i ) c + 2
= K ( x i , x i ) + 1 N + y i = + 1 φ ( x i ) 2 2 φ ( x i ) 1 N + y i = + 1 φ ( x i )
= K ( x i , x i ) + 1 N + 2 y m = + 1 y n = + 1 K ( x m , x n ) 2 N + y j = + 1 K ( x i , x j ) .
(2)
Nonmembership Function:
The non-membership function is defined as:
ν ( x i ) = ( 1 μ ( x i ) ) ρ ( x i ) ,
where 0 μ ( x i ) + ν ( x i ) 1 , and ρ ( x i ) is written as:
ρ ( x i ) = | { x j | φ ( x i ) φ ( x j ) β 1 , y j y i } | | { x j | φ ( x i ) φ ( x j ) β 1 } | ,
where β 1 > 0 is a parameter, and the distance between two instances in the inner product space can be expressed as
φ ( x ) φ ( x ) = K ( x , x ) + K ( x , x ) 2 K ( x , x ) .
We define the score function for a given IFN as:
H ( x i ) = μ ( x i ) , ν ( x i ) = 0 , 0 , μ ( x i ) < ν ( x i ) , 1 ν ( x i ) 2 μ ( x i ) ν ( x i ) , o t h e r s .
As we know, the data distribution is generally supposed to obey a Gaussian distribution. The centroid of standardized normal distribution data always approximates the mean of the distribution. Therefore, taking the centroid as a reference can well estimate the importance of each instance. However, the distribution of data in many real-world applications is often complicated, such as the dataset of Figure 1, which follows small disjunctions or manifold distribution. This led to the within-class centroid no longer being a reliable reference, because it could neither describe the manifold structure nor catch the small subclusters well.
In Figure 1a, the instances of group i satisfy x i N ( μ i , σ ) , where μ 1 = ( 0.4 , 0.4 ) , μ 2 = ( 0.15 , 0.15 ) , μ 3 = ( 0.7 , 0.7 ) and σ = [ 0.1 , 0 ; 0 , 0.1 ] . The numbers of these three groups are 800, 50, and 150 respectively. The instances of these three groups are divided into two classes, where the instances of group 1 are set to negative class and the instances of group 2 and group 3 are set to the positive class. So we obtain an imbalanced dataset with an imbalanced ratio of 1:4. In Figure 1b, the centroid of rings is ( 0 , 0 ) , the width of the ring equals 0.2, and the inner radius of negative and positive classes is 0.6, and 0.3 respectively, we obtain an imbalanced dataset with an imbalanced ratio of 1:5.
In Figure 1, we use IF to assign fuzzy values to some instances of Data1 and Data2. From Figure 1a, we can find that the fuzzy values of instances of small clusters of positive instances are generally low because the class center of positive instances is biased toward large clusters. Moreover, the fuzzy values of the instances that are closer to the class center in the small cluster are much smaller than that of the instances that are farther from the class center in the large cluster, and some instances in the small cluster are treated as noise. It can be seen that the fuzzy value obtained by the IF algorithm is unreasonable when the data hold small disjunctions. From Figure 1b, we can find that the fuzzy values of positive instances are generally low because the within-class centroid cannot describe the manifold structure well. Moreover, some important positive instances that are close to negative instances are judged as noise. This is because the instances are surrounded by negative instances, leading to many negative instances in the k-nearest neighbor, so the obtained non-membership degree will be greater than the membership degree, and such instances will be judged as noise.
As we all know, the centroid of data obeying the Gaussian distribution always approximates the mean of the distribution. Therefore, taking the centroid as a reference can well estimate the importance of each instance. However, for the data (especially imbalanced data) obeying non-Gaussian distribution, it is unreasonable to calculate the membership degree based on the class center. Meanwhile, as for the normal SVM, FSVMs are also impacted by class imbalance. Meanwhile, in most cases, to highlight the importance of the minority class, most FSVM series algorithms for class imbalance learning generally set the MVs of the minority class instances to 1, so ignore the distribution characteristics of the minority class itself.
To solve the problem discussed above, our work is based on the following two motivations:
  • To provide a more reliable measure to estimate the importance of each instance.
  • To propose a more preferable fuzzy MF to ensure the fairness of the classification method.

3. Relative Density-Based Intuitionistic Fuzzy Support Vector Machines (RIFSVM)

The relative density-based intuitionistic fuzzy membership approach is proposed first. Then, the relative density-based intuitionistic fuzzy SVM is proposed by using the density-based intuitionistic fuzzy membership for class imbalance learning.

3.1. Relative Density Estimate Based on a k-Nearest-Neighbor Distances

Based on the above analysis, it can be seen that, for some imbalanced datasets with special distribution, it is unreasonable to calculate the degree of membership and non-membership of the instance based on the within-class centroid criterion and calculate the fuzzy value of the instance according to the intuitionistic fuzzy value as well. Therefore, we should take advantage of the prior information of the instances, instead of giving the fuzzy values to instances by assuming that the instance conform to a certain distribution.
It is obvious that it will be easier to discriminate between regular instances and outliers/noise if we can accurately estimate the probability density of each case. However, it is incredibly challenging to quantify the probability density accurately in high dimensional feature space. Instead of precisely measuring the probability density of each instance, previous works [29] have proposed an alternative scheme in which we can determine the proportional relation of the probability densities between any two instances. The information that reflects the proportional relation is called relative density [34]. Here, we estimated the relative density using the K-nearest neighbors-based probability density estimation (KNN-PDE) [30,31] method. The KNN-PDE estimates the probability density distribution of the instances in the multi-dimensional continuous space according to the K-nearest neighbor distance of each instance. The KNN-PDE result can approximately converge to the real probability density distribution when the number of instances reaches infinity [35].
Suppose a dataset includes N + positive instances and N negative instances, and N + + N = N . In this paper, the positive class stands for the minority class and the negative class stands for the majority class. Then for each instance, we find its K-th nearest neighbor and denote the distance between them as d i k . It is not difficult to see that the larger d i k , the sparser the distribution of instance, that is, the smaller the density. As we know, noise and outliers are often in the low-density region, thereby we can estimate the importance of each instance by d i k . We define 1 / d i k as the relative density. Obviously, it assigns greater values to instances of high density and lower values to instances of low density, such as noise and outliers. We can also obtain the proportionality of the relative density of any two instances as follows:
1 / d i k 1 / d j k = d j k d i k .
This proportional relation equals the inverse of the ratio of the k-nearest neighbor distances between these two instances. Choosing an appropriate parameter k may be quite important for the relative density. If the k is too large, some normal instances may be misjudged as noise. If the k is too small, some noise and outliers would not be identified. In this paper, we set k = N by experience, where N is the number of instances.
Next, we introduce the two relative densities: within-class relative density and between-class relative density.
(1)
The within-class relative density refers to the relative density of an instance in its own class. For example, the within-class relative density of positive instance x i is 1 / d i k + , the d i k + is the distance between x i and the k + -th nearest neighbor in the positive instances, where k + = N + . The larger the value of d i k + , the lower the relative density of the instance, then the lower the probability that the instance belongs to this class.
(2)
The between-class relative density refers to the relative density of an instance in another class. The between-class relative density of positive instance x i , for instance, is 1 / d i k , the d i k is the distance between x i and the k -th nearest neighbor in the negative instances, where k = N . The larger the value of d i k , the lower the relative density of the instance, then the farther x i + is from the negative instances, that is the lower the probability that the instance belongs to the negative class.

3.2. Relative Density-Based Intuitionistic MFs for Class Imbalance Learning

Based on the fact above, we proposed a combined formulation of MFs based on relative density-based intuitionistic, which provides the membership values for instances to satisfy two goals:
  • to lessen the impact of class imbalance;
  • to reduce the negative impact of noise and outliers.
(1)
Determination of fuzzy value of majority class
For the instance of the majority class, the MV is calculated by the within-class relative density. Here, the exponential function is used to give the MVs of the instances. Instances with a larger d i k have a smaller relative density, so they should be given lower MVs. The specific calculation process is given as follows.
First, we calculate the MVs of instances according to the following membership degree function:
μ ( x i ) = 2 1 + exp d i k max { d 1 k , d 2 k , d N k } .
Then, we calculate the non-membership values of instances according to the following non-membership degree function:
ν ( x i ) = 1 μ ( x i ) ρ ( x i ) ,
where the ρ ( x i ) is defined as:
ρ ( x i ) = 2 1 + exp d i k + max { d 1 k + , d 2 k + , d N + k + } .
Note that the ρ ( x i ) is calculated by the between-class relative density of instance.
Then, according to the MVs and non-MVs of the instances, we define the score function as:
H ( x i ) = 0 , μ ( x i ) < ν ( x i ) , 1 ν ( x i ) 2 μ ( x i ) ν ( x i ) , o t h e r s .
It can easily distinguish normal instance from noises and outliers [21]. Finally, to lessen the influence of class imbalance, the fuzzy values of negative instances can be defined as:
s i = I R · H ( x i ) ,
where I R is the minority-to-majority class ratio.
(2)
Determination of fuzzy value of minority class
As is well known, there are few instances in the minority class. If the above method is used to calculate the fuzzy value, a large non-membership degree will be obtained, which will make the instance misjudged as noise. It can be seen that it is unreasonable to calculate the fuzzy value of the minority class using the majority class’s fuzzy value. It is also unreasonable to set the fuzzy value of the minority class to 1 directly like FSVM-CIL series algorithms.
Therefore, we directly set the membership degree of minority class instances as the fuzzy value. In this way, we not only give high fuzzy values to minority class instances but also fully utilize the prior information of instances.
μ + ( x i ) = 2 1 + exp d i k + max { d 1 k + , d 2 k + , d N + k + } ,
s i = μ + ( x i ) .
Since the fuzzy values of both majority and minoriy instances are calculated based on relative density, the above algorithm for calculating fuzzy values is denoted as RIF.
As we can see, the fuzzy values of the majority class instances are designed by incorporating the relative density measure into the intuitionistic fuzzy numbers. In this way, with the help of the strong capture ability of the relative density to prior information and the strong recognition ability of the intuitionistic fuzzy score function to outliers and noises, the fuzzy function can be adopted to identify possible noises and outliers existing in the majority class. Additionally, the participation of imbalance ratio reduces the effect of class imbalance. For the minority instances, to lessen disadvantage of being misjudged as noise points because the instances are too sparse, the fuzzy values of minority class instances is directly assigned as the MVs calculated by relative density without the non-MVS.
In order to explore the rationality of RIF, we compute the fuzzy value of Data1 and Data2 using the proposed RIF method and mark the fuzzy values of 20 instances which are the same as those in Figure 1, and the results are displayed in Figure 2. In Figure 2a, it is obvious that the fuzzy values of the small clusters instances are not too small as shown in Figure 1a, so they are not misjudged as noise. Instead, it follows the distribution characteristics of instances, assigning smaller fuzzy values to sparsely distributed points and larger fuzzy values to densely distributed sample points. Instead, it follows the distribution characteristics of instances, assigning smaller fuzzy values to sparsely distributed points and larger fuzzy values to densely distributed instances. From the Figure 2b, we can find that the instance on the edge is not misjudged as noise, and given a reasonable fuzzy value. It can be seen that for such imbalanced datasets with non-normal distribution, our RIF method is relatively reasonable.

3.3. Relative Density-Based Intuitionistic Fuzzy Support Vector Machines

By integrating the relative density-based intuitionistic MVs into FSVM, we propose a novel relative density-based intuitionistic fuzzy support vector machine for imbalanced learning with noise and outliers. The corresponding optimization problem can be expressed as:
min w , b 1 2 w 2 + C i = 1 N s i ξ i s . t . y i ( w T φ ( x i ) + b ) + ξ i 1 , ξ i 0 , i = 1 , 2 , , N ,
where s i = μ + ( x i ) , y i = 1 , I R · H ( x i ) , y i = 1 , and the μ ( x i ) , ν ( x i ) , μ + ( x i ) , and H ( x i ) are shown in (22), (23), (27) and (25) respectively.
At the same time, when C 1 / C + 1 = I R and C + = C , we can also write the objective function of the above optimization problem in the following cost-sensitive form:
f ( w , b , ξ ) = 1 2 w 2 + C + y i = + 1 s i ξ i + C y i = 1 s i ξ i ,
where s i = μ + ( x i ) , y i = 1 , H ( x i ) , y i = 1 .
To solve the optimization problem of RIFSVM in (29), the dual form of it can be written as:
max α i = 1 N α i 1 2 i = 1 N j = 1 N α i α j y i y j K ( x i , x j ) s . t . 0 α i s i C , i = 1 N y i α i = 0 , i = 1 , 2 , , N
where α = [ α 1 , , α N ] T is the Lagrange multiplier. K ( x i , x j ) = φ ( x i ) T φ ( x j ) is the kernel function satisfying Mercer’s theorem.
The optimal values α * can be obtained by solving the optimization (31), and the decision function of a RIFSVM can be expressed as:
h ( x ) = s i g n ( i = 1 N α * i y i K ( x i , x ) + b * ) ,
where w * and b * are shown in (3) and (4).
The above training process is summarized as Algorithm 1.
Algorithm 1 RIFSVM algorithm
Input: 
Training dataset T = { ( x 1 , y 1 ) ( x N , y N ) } , Penalty parameter C, kernel spread parameters h > 0 .
Output: 
The decision function (5).
 1:
The training sample T is divided into T + and T , T + is the set of minority instances, and T is majority instances. Calculate the numbers of minority and majority instances and denote them as N + and N respectively;
 2:
Calculate the nearest neighbor parameter k + = N + , k = N ;
 3:
For the instance x i T , calculate d k and calculate the membership degree μ ( x i ) of the majority class instance according to (22);
 4:
For the instance x i T , calculate d k + in the minority class and calculate the nonmembership degree v ( x i ) of the instance according to (23);
 5:
According to the obtained μ ( x i ) and ν ( x i ) and (25), the final fuzzy value s ( x i ) of the majority instances is calculated by using according to (26);
 6:
For the instance x i T + , calculate d k + and calculate the fuzzy value s ( x i ) of the minority instance according to (28);
 7:
Train NIFSVM model (29) to obtain the decision function.

4. Experiments and Analysis

In this section, we explore the performance and superiority of our proposed algorithms using synthetic and benchmark datasets. The five-fold cross-validation technique [36] is used in this paper to choose all of the parameters for these algorithms. The Gaussian kernel function K ( x 1 , x 2 ) = exp ( h x 1 x 2 2 ) is used for all data sets, kernel spread parameters h are roughly chosen within { 2 5 , 2 4 , 2 4 , 2 5 } . The tradeoff parameters C of our model are selected from the set of { 1 0 5 , 1 0 4 , , 1 0 4 , 1 0 5 } . All the experiments are carried out on a desktop PC with Intel(R) Xeon(R) CPU (3.30 GHz) and 32 GB RAM under the MATLAB 2019a programming environment.

4.1. Evaluation Metrics for Imbalanced Classification

The accuracy-based evaluation metric is usually used to assess the performance of the general classification method. However, it is no longer an appropriate metric for the imbalanced classification method since the effect of minority class on accuracy is smaller than that of majority class. In this paper, we use G-Mean, F-Measure, and AUC to evaluate the performance of imbalanced dataset classification. The G-Mean, a comprehensive measure of minority and majority class, denotes the geometric mean of sensitivity and specificity [37]. A reasonably high value of both True Positive and True Negative ensures a high G-Mean value. The F-Measure denotes the harmonic mean of Precision and Recall A high value of F-Measure means that both sensitivity and Precision are high simultaneously, and the AUC (the area under receiving operator characteristic curve) is another evaluation measure of classification performance in imbalanced problems because the area under the ROC graph is not sensitive to the distribution of two classes.

4.2. Experiments on the Synthetic Imbalanced Datasets

To intuitively show the effectiveness of the proposed model in dealing with class imbalance learning, we classify the two synthetic datasets in Figure 1 using IFSVM and RIFSVM, respectively. The experimental results are shown in Figure 3. In addition, for these two datasets, we trained the model on 80 % of the instances and tested the model on the remaining 20 % . The experiments are performed 10 times, and Table 1 lists the average G-Means, F-Measures, and AUC values. The best results are denoted in boldface.
From Figure 3a and Table 1, we can find that for Data1, the IFSVM has a higher misclassification rate for small cluster instances of minority instances. This is because we can see from Figure 1a that the instances of the small cluster are endowed with lower final fuzzy values, especially some instances of the small cluster are judged as noise, resulting in these important instances will no longer contribute to separating the hyperplane and are likely to be misclassified. However, the classification result obtained by RIFSVM is much better, and the misclassification rate for the instances of the small cluster is also lower. From Figure 3b and Table 1, we can find that the misclassification rate of boundary instances of minority class by IFSVM is higher than that by RIFSVM. This is because the non-membership degree of these instances calculated by IFSVM is higher, which leads to the smaller fuzzy value, and then these instances are misclassified. Similarly, the evaluation measure G-mean, F-Measure, and AUC of RIFSVM are all higher than those of IFSVM. It can be concluded that our method is effective when tackling imbalance datasets.

4.3. Experiments on Benchmark Datasets

In this section, we selected twenty datasets from the keel and UCI repositories in order to assess the performance of the proposed RIFSVM in tackling class imbalance classification. A complete description of each dataset is presented in Table 2. Some multi-classification problems were converted into binary classification problems by a one-versus-others strategy. In addition, we obtain some imbalanced datasets from the Yeast and Block datasets using different class combinations in order to obtain datasets with various imbalance ratios. The positive class and negative class columns in Table 2 contain a complete list of the class combinations for these generated datasets. The IR varies from 1:1.38 to 1:42.72 and the the number of attributes varies from 3 to 90. All attributes are normalized into the interval [ 0 , 1 ] .

4.3.1. Experimental Procedure and Results

We performed experiments on all the datasets in Table 2 to further confirm the effectiveness of the proposed method and compare it to IFSVM and other imbalance learning approaches.
(1)
IFSVM [20]: It uses the membership and non-membership calculated by the sample distribution information to determine the fuzzy value of the instance. The parameter δ = 10 4 , and β 1 = min ( r ) / 5 , where r = ( max ( R n ) , max ( R p ) ) .
(2)
ACFSVM [26]: It assigns the fuzzy value to the majority class instances based on the affinity and class probability and assigns 1 as the fuzzy value of minority class instances to highlight the importance of minority instances. The kernel nearest neighbor parameter k is chosen from the set { 3 , 5 , 7 , 9 , 11 } and the parameter ρ is selected from the set { 0.1 , 0.5 , 1 , 2 , 3 , 5 , 7 , 10 , 20 } .
(3)
GFSVM [25]: It is the supplement and extension of FSVM-CIL, a new distance and a new fuzzy value function, namely Gaussian fuzzy function, is proposed. The parameters μ and σ in the Gaussian function are selected from sets { 0.01 , 0.05 , 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.7 } and { 2 6 , 2 5 , , 2 5 , 2 6 } , respectively.
(4)
EFTWSVM [27]: It uses the information entropy of instances to calculate fuzzy values to minority instances, it fully utilizes the prior information of instances; it assigns 1 as the fuzzy value of minority class instances. Then the fuzzy values obtained are used in the improved TWSVM. The parameters β of EFTWSVM is taken as 0.05, the value of K for K-NN is 10.
(5)
RFLSTSVM [28]: It assigns a more robust fuzzy value to majority class instances and assigns 1 as the fuzzy value of minority class instances, then trains the LSTSVM model on this training set. The parameter k 0 of RFLSTSVM is selected from the set { 0.5 , 1 , 1.5 , 2 , 2.5 } .
(6)
FSVM-WD and FSVM-BD [29]: They are all proposed based on relative density, and the sum of fuzzy values of positive and negative class instances is set as 1 to ensure the robustness of the model. The distinction is that while FSVM-WD is based on information about within-class relative densities, FSVM-BD is based on the between-class relative densities. These two methods use the same strategy to calculate the fuzzy values of instances. The parameters of them are set as λ = 0.01 / N , k = N .
Note that several different algorithms are listed above. They adopt different strategies when calculating the fuzzy values of minority and majority classes. In addition, for some algorithms that set the fuzzy value of minority classes directly to 1, their classification models are also different. The effectiveness of the RIFSVM method is explored by comparing it with different fuzzy value setting methods and different classification model algorithms.
The experiment consisted of two parts. In the first part, the accuracy of minority and majority classes (Se and Sp) of eight algorithms on some datasets are compared. Then, the classification performances of eight algorithms on 20 datasets are compared. All the parameters are chosen by five-fold cross-validation based on G-Means.
First, to compare the accuracies of eight algorithms on positive and negative classes, we performed the experiments on five datasets in Table 2, which contain Wine, Vehicle, Abalone, Ecoli, and Libra. Their imbalance ratios are 1:2.71, 1:3.25, 1:3.86, 1:5.46, and 1:14, respectively. Except for the last Libra dataset, the IR of the other datasets have less difference. The accuracy of positive and negative classes for the seven algorithms on these five datasets is calculated, and the accuracy line chart is drawn in Figure 4a,b. In addition, Yeast1-Yeast5 are selected to explore the relationship between accuracy and imbalance ratio, the difference in imbalance ratio is significantly larger than in the previous five datasets. Their imbalance ratios are 1:8.10, 1:9.07, 1:12.25, 1:32.72, and 1:41.40, respectively. The corresponding experimental results are plotted in Figure 4c,d.
As shown in Figure 4a, the accuracy of these eight algorithms on the minority classes gradually decreases as the imbalance ratio rises, when the imbalance ratio is 1:14, the accuracy of minority classes of the seven algorithms decreases significantly. The accuracy of the proposed RIFSVM is slightly higher than that of other algorithms on the minority classes. When the imbalance ratio becomes large, the accuracy of RIFSVM decreases less than that of other algorithms. The results in Figure 4b show that the accuracy of these seven algorithms on majority classes gradually increases as the imbalance ratio increases. When the imbalance ratio is 1:5.46, the accuracy of majority classes of the seven algorithms is relatively low, which may be related to the dataset. The RIFSVM algorithm still outperforms other algorithms in the accuracy of most classes. From the results shown in Figure 4c, it can be seen that the accuracy of seven algorithms on the minority classes showed a downward trend as the imbalance ratio increased. When the imbalance ratio is 1:9.07, the accuracy of the eight algorithms is relatively high, which may be data related. The accuracy of the our algorithm is slightly better than the other seven algorithms for the minority classes because the algorithm uses relative density to calculate the fuzzy value of the instances and makes full use of the distribution characteristics of the instances. From the results of Figure 4d, we can find that the accuracy of these seven algorithms on majority class increases slowly as the imbalance ratio increases. When the imbalance ratio reaches the maximum, the accuracy of the eight algorithms reaches the highest, indicating that when the imbalance is relatively large, the classifier will still bias toward the majority class. In addition, it can be seen that the accuracy of the FSVM-WD algorithm on the majority class is lower than that of other algorithms. It may be because the algorithm only uses the within-class relative density of instances, which results in inaccurate fuzzy values.
In the second part, we conducted experiments on each of datasets provided in Table 1 in comparison to the other six imbalanced classification algorithms above. Experiments are repeated 10 times, and the G-Means, F-Measures, and AUC are recorded in Table 3. The best values are highlighted in bold.
As shown in Table 3, ACFSVM achieve higher G-Mean and AUC values but obtains lower F-Mean. In fact, a high F-Measure value means high classification accuracy of minority classes. It is evident that the performance of the ACFSVM for minority classification is not high. This may be because the separation hyperplane obtained by the ACFSVM algorithm is still skewed to minority classes, which contributes to the low classification accuracy of minority classes. The performance of the GFSVM algorithm on all datasets is mediocre, but its F-Mean is mostly higher than the ACFSVM. This may be because GFSVM uses the same fuzzy function to calculate the fuzzy value of minority class and majority class, and the difference between majority class and majority class only depends on the imbalance ratio to constrain, which does not result in a significant deviation of the separation hyperplane, so the classification effect will not be greatly improved. EFTWSVM and RFLSTSVM algorithms form a separate hyperplane in each class during training, and their classification performance may be general since the classification accuracy of EFTWSVM and RFLSTSVM is inferior to SVM. The FSVM-BD and FSVM-WD have poor performance in G-Mean and AUC metrics, but their F-Mean is very high. It indicates that the relative density can well reflect the distribution characteristics of minority classes so that the separation hyperplane is no longer biased toward the majority class. In addition, we also find that the classification performance of FSVM-BD, FSVM-WD, and RIFSVM on the Balance dataset is significantly better than that of other methods because these three algorithms all use the relative density of instances when calculating the fuzzy value. It indicates that the relative density can well describe the distribution characteristics of the Balance dataset. On most data sets, the classification performance of the FSVM-BD and FSVM-WD is not good, which may be because the two algorithms use the normalization method to make the sum of the fuzzy values of minority and majority categories equal when calculating the fuzzy values. When the number of instances is large, the fuzzy value of the instances will become small, which leads to poor classification performance.
Finally, the proposed RIFSVM performs well in the three performance indexes of classification on most datasets. This is because RIFSVM not only uses the relative density to give the distribution information of minority and majority class instances but also combines the intuitionistic fuzzy set to give different fuzzy values of positive class sample instances. Thus, the separation hyperplane is no longer skewed to the minority class.

4.3.2. Statistical Comparisons by Friedman Test

The Friedman tests [38] are used to provide a statistical comparison of the proposed RIFSVM with the existing class imbalance learning algorithms. First, we determined average ranks for the G-Mean, F-Measure, and AUC metrics of the seven methods across all datasets. For these datasets, we set the average rank of the algorithm with the best classification performance as 1 while the worst classification performance as 8.
From the results in Figure 5, we can find that the proposed RIFSVM ranks first with an average score of 1.60, 1.70, and 1.85 in terms of G-Mean, F-Measure, and AUC, respectively. It demonstrates that in the mean ranking of all evaluation measures, our proposed RIFSVM performs better than other methods.
Then, the Friedman test is used to judge whether these algorithms all perform equally. Obviously, k(=8) compared algorithms and n(=20) imbalance datasets are considered in this experiments. The r i is the average rank of the j-th algorithms. Under the null hypothesis, which states that all the algorithms are equivalent, and thus their ranks should be equal, the Friedman statistic
Γ χ 2 = 12 N k ( k + 1 ) ( i = 1 k r i 2 k ( k + 1 ) 2 4 )
is distributed according to χ 2 with ( k 1 ) degrees of freedom, when n and k are reasonably large. The Friedman τ χ 2 presents a pessimistic behavior, thus the statistic
τ F = ( n 1 ) Γ χ 2 n ( k 1 ) Γ χ 2
is usually used, which is distributed according to the F-distribution with ( k 1 ) and ( k 1 ) ( n 1 ) degrees of freedom. According to (34), the Friedman statistic τ F of G-Mean, F-Measure, and AUC are 7.50, 6.24, and 6.04 respectively, which at significance level α = 0.05 rejects the null hypothesis of equal performance.
The Nemenyi test is then employed to further identify these eight algorithms. The supposition that “the two algorithms have the same performance” is rejected with corresponding confidence if the difference between the average ranks of the two algorithms exceeds the critical difference (CD). The critical range of average ranks difference C D = q α k ( k + 1 ) 6 n is calculated as 2.3478, where critical values q α are based on the studentized range statistic.
The average rank of each comparing algorithms is indicated along the axis in Figure 6’s CD diagrams of G-Mean, F-Measure, and AUC on the twenty benchmark datasets. The axis has been rotated to place the highest ranks to the right. A red line connects groups of approaches that, according to the Nemenyi test, are not significantly different from one another. In each subfigure, the critical difference is also displayed above the axis.
As can be seen from Figure 6b, our algorithm is optimal in F-Meansure, and it is significantly superior and different from other algorithms. From Figure 6a,c, we can see that our algorithm is optimal in G-Mean and AUC, and is significantly superior different from other algorithms except for ACFSVM. The ACFSVM is only inferior to our method while superior to the other six algorithms in terms of G-Mean and AUC, but it’s not doing very well on the F-Measure. The FSVM-WD and FSVM-BD perform better on F-Measure but are the worst on both G-Mean and AUC. The GFSVM performs generally well in all indicators, and IFSVM, RFLSTSVM and EFTWSVM are mediocre in terms of G-Mean, F-Measure, and AUC. In general, our RIFSVM achieves statistically superior performance on the whole twenty datasets.

4.3.3. Influences of Parameter k on the Performance

In order to evaluate the influence of the parameter k for relative density calculation on the classification performance of our proposed approach. It is selected from the set { N / 3 , N / 2 , N , 2 N , 3 N } . The variation of the G-Means and F-measures with the variation of parameter k on the five chosen datasets is then plotted in Figure 7.
From the results, it can be seen that although there are some fluctuations, the performance increases at the initial stage and decreases after reaching the peak with the increase of k. That is, the performance of the proposed algorithms will deteriorate if k is too high or too low. In fact, if the k is set at a value that is too low, outliers and noise points might be assigned oversize weights. While the k is set at a value that is too high, the instances in the same category might be assigned undersized weights. The results in Figure 7 provide some reference, i.e., the performance of our method could be assured when k is between N / 2 and 2 N . This shows that setting the parameter k as N empirically in our experiment is reasonable. In practical applications, we recommend that the user choose the appropriate parameter k by themselves.

5. Conclusions

In this paper, the FSVM technique for class imbalance learning has been discussed. First, the shortcomings of the settings of existing intuitionistic fuzzy value on some imbalanced datasets with specific distribution and limitations of traditional FSVM-CIL series algorithms in dealing with imbalanced classification task has been analyzed. Then, the importance of digging into the prior information of the instances and merging them into the classification model has been emphasized. Inspired by the work above, a novel relative density-based Intuitionistic FSVM (RIFSVM) has been presented for imbalanced learning with outliers and noise. Specifically, the membership degree and non-membership degree of instances are calculated based on relative density and exponential decay function. According to the above membership degree and non-membership degree, the fuzzy value of majority instances is calculated by the specific score function, and the fuzzy value of minority instances is set as the membership degree. With the help of the strong capture ability of the relative density to prior information and the strong recognition ability of the intuitionistic fuzzy score function to outliers and noises, the proposed RIFSVM not only suppresses the influence of class imbalance but also reduces the impact of noises and outliers. Finally, we demonstrate the effectiveness and superiority of our proposed algorithm on synthetic and benchmark imbalanced datasets. The experimental results on the synthetic datasets show that the proposed RIFSVM can achieve a better classification boundary than IFSVM and be flexible for various types of data distribution. The experimental results on benchmark datasets demonstrate that the proposed algorithm achieves better performance than the other state-of-the-art class imbalance learning algorithms in terms of G-Means, F-Measures, and AUC with good robustness, and a statistical test is applied to verify the performance significance of the proposed method. In addition, after the influences of parameters k on the performance were discussed in this study, we found that when k is between N / 2 and 2 N , the performance of our algorithm could be guaranteed.
In the future, it will be interesting to translate the proposed approach into an efficient approach to multiple classes of problems. In addition, how to combine the relative density with other score functions to obtain reasonable fuzzy values for some other specific classification problems would be investigated in the future, too.

Author Contributions

Conceptualization, C.F. and S.Z.; methodology, C.F.; software, C.F.; validation, C.F., S.Z. and L.C.; writing—original draft preparation, C.F. and D.Z.; writing—review and editing, C.F. and S.Z.; supervision, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61772020, No. 62202433), the Natural Science Foundation of Henan Province (No. 22100002), and the Postdoctoral Research Grant in Henan Province (No. 202103111).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Raghuwanshi, B.S.; Shukla, S. Class-specific extreme learning machine for handling binary class imbalance problem. Neural Netw. 2018, 105, 206–217. [Google Scholar] [CrossRef] [PubMed]
  2. Tao, X.; Li, Q.; Guo, W.; Ren, C.; Li, C.; Liu, R.; Zou, J. Self-adaptive cost weights-based support vector machine cost-sensitive ensemble for imbalanced data classification. Inf. Sci. 2019, 487, 31–56. [Google Scholar] [CrossRef]
  3. Romani, M.; Vigliante, M.; Faedda, N.; Rossetti, S.; Pezzuti, L.; Guidetti, V.; Cardona, F. Face memory and face recognition in children and adolescents with attention deficit hyperactivity disorder: A systematic review. Neurosci. Biobehav. Rev. 2018, 89, 1–12. [Google Scholar] [CrossRef] [PubMed]
  4. Laxmi, S.; Gupta, S.K. Multi-category intuitionistic fuzzy twin support vector machines with an application to plant leaf recognition. Eng. Appl. Artif. Intell. Int. J. Intell.-Real-Time Autom. 2022, 110, 110. [Google Scholar] [CrossRef]
  5. Yadav, A.; Singh, A.; Dutta, M.K.; Travieso, C.M. Machine learning-based classification of cardiac diseases from PCG recorded heart sounds. Neural Comput. Appl. 2019, 32, 17843–17856. [Google Scholar] [CrossRef]
  6. Yang, L.; Xu, Z. Feature extraction by PCA and diagnosis of breast tumors using SVM with DE-based parameter tuning. Int. J. Mach. Learn. Cybern. 2019, 10, 591–601. [Google Scholar] [CrossRef]
  7. Yi, M.; Zhou, C.; Yang, L.; Yang, J.; Tang, T.; Jia, Y.; Yuan, X. Bearing Fault Diagnosis Method Based on RCMFDE-SPLR and Ocean Predator Algorithm Optimizing Support Vector Machine. Entropy 2022, 24, 1696. [Google Scholar] [CrossRef] [PubMed]
  8. Jurgovsky, J.; Granitzer, M.; Ziegler, K.; Calabretto, S.; Portier, P.E.; He-Guelton, L.; Caelen, O. Sequence Classification for Credit-Card Fraud Detection. Expert Syst. Appl. 2018, 100, 234–245. [Google Scholar] [CrossRef]
  9. Carneiro, N.; Figueira, G.; Costa, M. A data mining based system for credit-card fraud detection in e-tail. Decis. Support Syst. 2017, 95, 91–101. [Google Scholar] [CrossRef]
  10. Sebastiani, F. Machine learning in automated text categorization. ACM Comput. Surv. 2002, 34, 1–47. [Google Scholar] [CrossRef]
  11. Tan, S. Neighbor-weighted k-nearest neighbor for unbalanced text corpus. Expert Syst. Appl. 2005, 28, 667–671. [Google Scholar] [CrossRef] [Green Version]
  12. Rodríguez Alvarez, Y.; García Lorenzo, M.M.; Caballero Mota, Y.; Filiberto Cabrera, Y.; García Hilarión, I.M.; Machado Montes de Oca, D.; Bello Pérez, R. Fuzzy prototype selection-based classifiers for imbalanced data. Case study. Pattern Recognit. Lett. 2022, 163, 183–190. [Google Scholar] [CrossRef]
  13. Cherkassky, V. The nature of statistical learning theory. IEEE Trans. Neural Netw. 1997, 8, 1564. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Fu, C.; Zhou, S.; Zhang, J.; Han, B.; Chen, Y.; Ye, F. Risk-Averse support vector classifier machine via moments penalization. Int. J. Mach. Learn. Cybern. 2022, 13, 3341–3358. [Google Scholar] [CrossRef]
  16. Boser, B.E. A Training Algorithm for Optimal Margin Classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Amsterdam, The Netherlands, 7–10 July 2008; Volume 5, pp. 144–152. [Google Scholar]
  17. Zhang, X. Using class-center vectors to build support vector machines. In Neural Networks for Signal Processing IX, Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No. 98TH8468), Madison, WI, USA, 25 August 1999; IEEE: Piscataway, NJ, USA, 1999; pp. 3–11. [Google Scholar]
  18. Lin, C.F.; Wang, S.D. Fuzzy support vector machines. IEEE Trans. Neural Netw. 2002, 13, 464–471. [Google Scholar]
  19. Atanassov, K.T. Intuitionistic fuzzy sets. Fuzzy Sets Syst. 1986, 20, 87–96. [Google Scholar] [CrossRef]
  20. Ming-Hu, H.A.; Huang, S.; Wang, C.; Wang, X.L. Intuitionistic Fuzzy Support Vector Machine. J. Hebei Univ. Sci. Ed. 2011, 31, 226–229. [Google Scholar]
  21. Ha, M.; Wang, C.; Chen, J. The support vector machine based on intuitionistic fuzzy number and kernel function. Soft Comput. 2013, 17, 635–641. [Google Scholar] [CrossRef]
  22. Rezvani, S.; Wang, X.; Pourpanah, F. Intuitionistic Fuzzy Twin Support Vector Machines. IEEE Trans. Fuzzy Syst. 2019, 27, 2140–2151. [Google Scholar] [CrossRef]
  23. Rezvani, S.; Wang, X. Class imbalance learning using fuzzy ART and intuitionistic fuzzy twin support vector machines. Inf. Sci. 2021, 578, 659–682. [Google Scholar] [CrossRef]
  24. Batuwita, R.; Palade, V. FSVM-CIL: Fuzzy support vector machines for class imbalance learning. IEEE Trans. Fuzzy Syst. 2010, 18, 558–571. [Google Scholar] [CrossRef]
  25. Liu, J. Fuzzy support vector machine for imbalanced data with borderline noise. Fuzzy Sets Syst. 2021, 413, 64–73. [Google Scholar] [CrossRef]
  26. Borah, P.; Gupta, D. Affinity and transformed class probability-based fuzzy least squares support vector machines. Fuzzy Sets Syst. 2022, 443, 203–235. [Google Scholar] [CrossRef]
  27. Deepak, G.; Bharat, R.; Parashjyoti, B. A fuzzy twin support vector machine based on information entropy for class imbalance learning. Neural Comput. Appl. 2018, 31, 7153–7164. [Google Scholar]
  28. Richhariya, B.; Tanveer, M. A robust fuzzy least squares twin support vector machine for class imbalance learning. Appl. Soft Comput. 2018, 71, 418–432. [Google Scholar] [CrossRef]
  29. Yu, H.; Changyin, S.; Yang, X.; Zheng, S.; Zou, H. Fuzzy Support Vector Machine With Relative Density Information for Classifying Imbalanced Data. IEEE Trans. Fuzzy Syst. 2019, 27, 2353–2367. [Google Scholar] [CrossRef]
  30. Wang, Q.; Kulkarni, S.R.; Verdu, S. Divergence Estimation for Multidimensional Densities Via k-Nearest-Neighbor Distances. IEEE Trans. Inf. Theory 2009, 55, 2392–2405. [Google Scholar] [CrossRef]
  31. Fukunaga, K.; Hostetler, L. Optimization of k nearest neighbor density estimates. IEEE Trans. Inf. Theory 2003, 19, 320–326. [Google Scholar] [CrossRef]
  32. Lin, C.F.; Wang, S.D. Fuzzy Support Vector Machines with Automatic Membership Setting. In Support Vector Machines: Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2005; Volume 177, pp. 251–253. [Google Scholar]
  33. Fan, Q.; Wang, Z.; Li, D.; Gao, D.; Zha, H. Entropy-based Fuzzy Support Vector Machine for Imbalanced Datasets. Knowl.-Based Syst. 2016, 115, 87–99. [Google Scholar] [CrossRef]
  34. Zheng, S.; Sun, D.; Hualong, Y. Fuzzy weighted extreme learning machine for imbalanced software detect prediction. J. Jiangsu Univ. Sci. Technol. 2019, 33, 7. [Google Scholar]
  35. Biau, G.; Devroye, L. The k-nearest neighbor density estimate. In Lectures on the Nearest Neighbor Method; Springer International Publishing: Cham, Switzerland, 2015; pp. 25–32. [Google Scholar]
  36. Ron, K. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 20–25 August 1995. [Google Scholar]
  37. He, H.; Garcia, E.A. Learning from Imbalanced Data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar]
  38. Demiar, J.; Schuurmans, D. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. The fuzzy membership values of some positive instances of Data1 and Data2 calculated by IF.
Figure 1. The fuzzy membership values of some positive instances of Data1 and Data2 calculated by IF.
Entropy 25 00034 g001
Figure 2. Fuzzy membership values of some positive (minority) class instances calculated by RIF. (a) Fuzzy membership values of Data1; (b) Fuzzy membership values of Data1.
Figure 2. Fuzzy membership values of some positive (minority) class instances calculated by RIF. (a) Fuzzy membership values of Data1; (b) Fuzzy membership values of Data1.
Entropy 25 00034 g002
Figure 3. Plots for comparing the classification boundaries of IFSVM and RIFSVM on the Data1 of Figure 1.
Figure 3. Plots for comparing the classification boundaries of IFSVM and RIFSVM on the Data1 of Figure 1.
Entropy 25 00034 g003
Figure 4. (a,b) are the accuracies of minority and majority class of the eight algorithms on five datasets with different imbalance ratios. (c,d) are the accuracies of minority and majority class of the eight algorithms on the Yeast dataset with different imbalance ratios.
Figure 4. (a,b) are the accuracies of minority and majority class of the eight algorithms on five datasets with different imbalance ratios. (c,d) are the accuracies of minority and majority class of the eight algorithms on the Yeast dataset with different imbalance ratios.
Entropy 25 00034 g004
Figure 5. Mean ranking of all compared imbalanced classification algorithms on test datasets.
Figure 5. Mean ranking of all compared imbalanced classification algorithms on test datasets.
Entropy 25 00034 g005
Figure 6. CD diagrams of the eight comparison models on the twelve benchmark datasets with three performance indexes of classification. It is clear that RIFSVM achieves statistically superior performance on the datasets with different performance indexes of classification.
Figure 6. CD diagrams of the eight comparison models on the twelve benchmark datasets with three performance indexes of classification. It is clear that RIFSVM achieves statistically superior performance on the datasets with different performance indexes of classification.
Entropy 25 00034 g006
Figure 7. Plots of G-Means and F-Measures obtained by RIFSVM with respect to different k of the model on Abalone, Yeast, Balance, Block, and Seed datasets.
Figure 7. Plots of G-Means and F-Measures obtained by RIFSVM with respect to different k of the model on Abalone, Yeast, Balance, Block, and Seed datasets.
Entropy 25 00034 g007
Table 1. Classification results of IFSVM and RIFSVM on two synthetic datasets in Figure 1, the best values are in bold.
Table 1. Classification results of IFSVM and RIFSVM on two synthetic datasets in Figure 1, the best values are in bold.
Results (%)SeSpG-MF-MAUC
Data 1IFSVM86.5099.0092.7590.8192.75
RIFSVM92.50 99.13 95.75 94.39 95.81
Data 2IFSVM98.00 100.00 98.9998.9999.00
RIFSVM 100.00 100.00 100.00 100.00 100.00
Table 2. Details of the imbalanced datasets.
Table 2. Details of the imbalanced datasets.
DatasetFeatureInstancePositiveNegativeMinority ClassMajoritt ClassIR
Liver6345145200class 1class 21:1.38
Seed721070140class 1others1:2.00
Wine1317848130class 1others1:2.71
Haberman330681225class 2class 11:2.78
Glass415050100class 0,1,2,3class 4,5,61:3.19
Vehicle18846199647‘van’others1:3.25
Abalone832667259class 16class 61:3.86
Ecoli733652284‘pp’others1:5.46
Balance462549576‘B’others1:11.76
Libra9036024336class 15others1:14.00
Yeast814844291055class 2others1:2.46
Yeast1814841631321class 4others1:8.10
Yeast2851451463class 5class 11:9.07
Yeast3846435429class 7class 21:12.25
Yeast481484441440class 6others1:32.72
Yeast581484351449class 7others1:41.40
Block1054735604913othersclass 11:8.77
Block11054733295144class 2others1:9.07
Block21051442314913class 3,4,5class 11:21.27
Block31050281154913class 5class 11:42.72
Note: The feature and instance indicate the number of features and instances respectively, and IR is the imbalance ratio.
Table 3. Classification results obtained by eight algorithms on twenty datasets, the best values are in bold.
Table 3. Classification results obtained by eight algorithms on twenty datasets, the best values are in bold.
DatasetResultsIF
SVM
ACF
SVM
GF
SVM
ETW
SVM
RFLST
SVM
FSVM
-BD
FSVM
-WD
RIF
SVM
LiverGm67.8260.1064.4866.5265.8567.4764.33 69.48
F62.08 67.30 58.0061.5560.4161.5458.1864.00
AUC69.9066.2165.9668.4768.8868.8465.0971.34
SeedGm89.2193.6292.4493.6593.3387.0491.05 94.63
F85.7191.2089.4089.4289.5384.6288.89 92.86
AUC89.2993.8992.5093.3993.7587.5091.01 94.64
WineGm88.1996.1793.2491.3588.1994.2887.71 98.06
F87.5093.5591.5888.5087.5094.1275.00 94.71
AUC88.8996.2493.8091.7988.8994.4488.46 98.08
habermanGm57.0162.4361.5763.6567.9665.8362.36 69.13
F42.8647.2446.4348.4552.7253.3350.00 54.12
AUC62.0862.8863.2364.1068.1068.3366.32 69.17
GlassGm79.9677.1779.9569.3069.82 81.65 80.14 81.65
F60.9540.6162.6732.2233.67 80.00 72.30 80.00
AUC81.2877.8281.2870.6471.82 83.33 80.13 83.33
VehicleGm96.6494.9898.1496.6396.6497.94 98.71 98.71
F94.8790.8296.8595.4594.9196.20 98.70 98.70
AUC96.6694.9998.1496.6896.6697.94 98.72 98.72
AbaloneGm94.2594.8395.5990.6492.2492.1593.93 96.00
F84.4984.8485.5578.6279.2782.6281.25 86.67
AUC94.3494.9295.6990.6692.3892.2394.12 96.08
EcoliGm90.2492.2193.4693.4294.1492.5885.53 95.46
F75.0077.0383.6782.8381.1573.6882.22 85.72
AUC90.3692.4593.5093.7094.1692.8686.46 95.50
BalanceGm87.4277.7868.3470.3974.47 100.00 100.00 100.00
F77.7828.4633.5826.8128.29 100.00 100.00 100.00
AUC88.0080.2674.2274.6475.61 100.00 100.00 100.00
LibraGm91.9695.0386.2190.6992.9391.8186.60 98.29
F91.4362.4479.2882.2287.3089.2185.71 93.02
AUC92.5095.1387.0591.9893.3892.3587.50 98.38
YeastGm60.95 70.11 68.4254.8969.7460.4560.8965.51
F49.31 57.43 52.8242.5757.2946.8149.6553.75
AUC65.30 70.14 67.2862.0470.0062.8765.6167.71
Yeast1Gm71.4484.8383.1680.6883.0477.8178.83 84.95
F59.5267.5370.7770.8656.8267.8357.49 72.73
AUC73.8285.0384.0482.1183.1979.9980.82 85.61
Yeast2Gm84.9589.4388.8088.9688.7589.5790.38 93.83
F72.7380.7282.8184.2170.5075.0070.7785.71
AUC85.6189.8189.2889.4688.9589.7390.43 93.91
Yeast3Gm91.9384.0284.0280.6581.0289.4489.56 93.31
F78.5676.9276.9275.2463.81 88.89 88.89 81.82
AUC92.0485.1385.1382.5182.2890.0090.00 93.37
Yeast4Gm83.0181.5481.7683.0682.0978.7878.78 83.30
F66.6763.5569.1758.9659.6966.6766.67 75.00
AUC83.9583.0983.3783.2483.1780.8980.89 84.15
Yeast5Gm49.7368.0765.0464.9260.2761.1453.27 68.76
F33.3334.2144.0040.8039.2341.8836.36 46.15
AUC61.9670.3170.7770.6067.9168.5363.94 70.91
BlockGm90.8694.0693.0893.7694.2389.0387.32 94.73
F83.0782.8783.2679.9681.0280.1880.37 84.96
AUC91.1294.1193.1893.8794.0589.4087.92 94.94
Block1Gm93.2893.78 94.05 93.4792.6689.7392.4193.09
F87.6079.1577.6981.2377.7979.1786.15 87.64
AUC93.4693.88 94.13 93.5992.8790.1292.6493.28
Block2Gm80.4393.4489.5793.33 93.71 87.8285.8093.06
F65.2265.2964.5865.9863.2063.34 80.95 65.37
AUC82.20 93.48 89.9793.4093.7289.5986.7593.09
Block3Gm85.6787.6983.6282.9881.7874.7271.97 89.11
F78.1650.1949.2253.9953.3950.04 57.14 50.06
AUC86.6088.3084.5984.1383.2184.4075.73 89.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, C.; Zhou, S.; Zhang, D.; Chen, L. Relative Density-Based Intuitionistic Fuzzy SVM for Class Imbalance Learning. Entropy 2023, 25, 34. https://doi.org/10.3390/e25010034

AMA Style

Fu C, Zhou S, Zhang D, Chen L. Relative Density-Based Intuitionistic Fuzzy SVM for Class Imbalance Learning. Entropy. 2023; 25(1):34. https://doi.org/10.3390/e25010034

Chicago/Turabian Style

Fu, Cui, Shuisheng Zhou, Dan Zhang, and Li Chen. 2023. "Relative Density-Based Intuitionistic Fuzzy SVM for Class Imbalance Learning" Entropy 25, no. 1: 34. https://doi.org/10.3390/e25010034

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop