Next Article in Journal
Optimization of LFP Pouch Cell Tab Design for Uniform Temperature Distribution
Previous Article in Journal
Lattice Boltzmann Numerical Study on Mesoscopic Seepage Characteristics of Soil–Rock Mixture Considering Size Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Annulus Clustering for Multi-Label Classification

1
School of Computer, Jiangsu University of Science and Technology, Zhenjiang 212100, China
2
Key Laboratory of Oceanographic Big Data Mining and Application of Zhejiang Province, Zhoushan 316022, China
3
School of Science, Jiangsu University of Science and Technology, Zhenjiang 212100, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(8), 1969; https://doi.org/10.3390/math11081969
Submission received: 24 March 2023 / Revised: 13 April 2023 / Accepted: 17 April 2023 / Published: 21 April 2023
(This article belongs to the Section Fuzzy Sets, Systems and Decision Making)

Abstract

:
Label-specific feature learning has become a hot topic as it induces classification models by accounting for the underlying features of each label. Compared with single-label annotations, multi-label annotations can describe samples from more comprehensive perspectives. It is generally believed that the compelling classification features of a data set often exist in the aggregation of label distribution. In this in-depth study of a multi-label data set, we find that the distance between all samples and the sample center is a Gaussian distribution, which means that the label distribution has the tendency to cluster from the center and spread to the surroundings. Accordingly, the double annulus field based on this distribution trend, named DEPT for double annulusfield and label-specific features for multi-label classification, is proposed in this paper. The double annulus field emphasizes that samples of a specific size can reflect some unique features of the data set. Through intra-annulus clustering for each layer of annuluses, the distinctive feature space of these labels is captured and formed. Then, the final classification model is obtained by training the feature space. Contrastive experiments on 10 benchmark multi-label data sets verify the effectiveness of the proposed algorithm.

1. Introduction

In multi-label classification tasks, multiple classification models are derived from a training set for each class label [1]. As a popular paradigm in machine learning, multi-label learning techniques have been widely employed to solve real-world problems, such as image annotation [2], where an image may convey various information; medical diagnosis [3], where the task is to identify the patient’s disease from its symptoms; and sentiment analysis [4,5], where an expression can contain many emotions, etc.
To deal with the classification problem for multi-label data, one of the common approaches is to utilize dependencies between behavioral labels to induce classification models [6,7,8]. Although behavioral labels achieve better results in multi-label classification, this approach might be suboptimal due to ignoring the underlying features for each of its own class labels. For example, altitude-based features are preferred in distinguishing the snow and non-snow labels, while moisture-based features tend to recognize sea and non-sea labels. Therefore, label-specific learning [9,10,11], which extracts the underlying characteristics of each class label, has excellent research significance in multi-label classification.
It needs to be emphasized that the degree of data aggregation intuitively reflects potentially label-specific features. Therefore, some existing methods utilize spherical random clustering, i.e., k-means [9,12], to extract label-specific features by evaluating the similarity between different instances for each label through Euclidean distance. The apparent advantage of the random clusters can be summarized in two points. On the one hand, the iterative process that minimizes squared error between instances and cluster centers is easy to implement and also has low time complexity. On the other hand, favorable adaptability can be obtained, such as rapid convergence and better cluster effect. In particular, it has good scalability for extensive data conforming to the Gaussian distribution. However, there are still some limitations to this method. First, it does not view sufficient discriminative information provided by cluster centers, mainly because random clusters may be trapped in the optimal local solution [13]. Meanwhile, heterogeneous numbers of the cluster centers will further affect the results. Second, it is not suitable for discrete data set. It is also sensitive to abnormal values, i.e., it may generate cluster centers in inappropriate locations due to directly taking into account instances with significant bias.
Motivated by the above-mentioned problems, a novel label-specific feature algorithm based on the double annulus field is proposed in this paper, as shown in Figure 1, which presents the foundation of the annulus model. In other words, due to the distribution characteristics of a multi-label data set, where the majority of samples cluster around the sample center and exhibit a divergent trend towards the surrounding areas, we propose an adaptive annulus model for the distribution trend of a multi-label data set. It should be noted that the double annulus field is an improvement on the single annulus model. The single annulus model is used to construct the label-specific feature space by hierarchically extracting each layer of instances. Although significant biases in the number of positive and negative instances are mitigated through the model, classification ability for more indistinguishable cases is limited, such as the cross-distribution of instances and unbalanced label density. That is, instances within the single annulus are not enough to provide discernable information for both categories simultaneously. In response to this situation, the double annulus field is proposed. The principle of this model is to divide the heterogeneously intersecting instances with high density into different annulus fields, and annulus clustering is then carried out. One of the essential tasks of the double annulus field is to design a mapping process. The mapping process that uses principal component analysis (PCA) [14] can not only bisect each category of instances into two parts but also reduce the classification difficulty in terms of the unbalanced label density. In summary, the contributions of this paper are highlighted as follows:
1.
The concept of the double annulus field for multi-label data is established, which ensures the same number of instances in each layer of annuluses and thus sufficient cluster information to be contained within each annulus.
2.
A cluster strategy within the annulus model is developed, which captures the potential features for each layer of instances and thus effectively prevents information loss. To mitigate the influence of unbalanced label density, each layer of instances can be divided by mapping relationships.
The rest of this paper is organized as follows. In Section 2, several multi-label learning methods are reviewed. In Section 3, the double annulus field for exploring hierarchical label-specific features and intra-annulus clusters are both analyzed. The comparative experimental results and analyses are shown in Section 4. Finally, the conclusions are summarized in Section 5.

2. Related Work

As a pragmatic and popular paradigm in machine learning, multi-label classification has been used intensively in recent research to study aspects such as label correlations [15,16], data streams [17,18], class imbalance [19,20], and feature selection [21,22].
Recent approaches to label correlations can be divided into three techniques by considering the order of label correlations, namely first-order [23,24], second-order [25,26,27], and high-order techniques [28,29,30]. Data streams differentiate themselves from traditional one-time scanning of all data by emphasizing the fact that the concepts contained in the data will change over time and by processing data on time using limited resources, i.e., memory and time. Class imbalance indicates an inherent attribute of multi-label data, where the number of positive training instances is generally much less than its negative counterparts. Feature selection is an effective dimension reduction technique to cope with high-dimensional multi-label data.
Differing from the above methods, label-specific features as an intuitive approach to dealing with multi-label classification problems could explore tailored features and construct feature mappings. In addition, the main idea in such an approach is that the underlying characteristics for each class label are different in their discrimination processes. Therefore, a more effective classification model can be induced through the discriminative features under each label. To achieve this goal, Zhang et al. [9] proposed the LIFT (multi-label learning with label-specific features) algorithm that obtains label-specific features by random clustering and then forms the classification models through the mapping relationships of label-specific features. Xu et al. [12] reduced redundant information on increasing dimensions of multi-label data and performed sample selection through the fuzzy rough set [31]. Zhan et al. [32] appended the ensemble clustering strategy to optimize the unstable random clustering in LIFT. From a deep learning perspective, CLIF (collaborative learning of label semantics and deep label-specific features for multi-label classification) [33] further integrated deep neural networks and label semantics [34] to guide the formation of label-specific features.
Combining label-specific features with the correlation of pairwise labels, Pei et al. [35] proposed JLSE2N (joint label-density-margin space and extreme elastic net for label-specific features), which utilizes the density of multi-label data to calculate the cosine similarity, and then quickly forms the label-specific features based on the elastic net. Lin et al. [36] proposed MULFE (multi-label learning via label-specific feature space ensemble) to obtain maximum margin multi-label classification through ensemble learning and label correlation. Zhang et al. [37] proposed BiLabel (bilabel-specific feature generation and predictive model induction), which emphasized the generation process of label-specific features and generated the label-specific features through heuristic prototype selection. Different from the above-mentioned methods, we focus on stratification in processing multi-label data. Specifically, as shown in Figure 2, the conception of annulus division is constructed based on positive and negative instances to extract hierarchical label-specific features in the design of the double annulus field.

3. The Annulus Feature Space

3.1. Notations

First of all, the following notations are utilized in this paper. Let D = ( x i , y i ) 1 i m represent the multi-label training set with m instances, where x i R d is a d-dimensional feature vector, y i 0 , 1 1 × q is the possible set of q labels associated with x i , and l = l 1 , l 2 , . . . , l q denotes the label space with q labels. Formally, y i k = 1 represents l k y i k , and y i k = 0 represents l k y i k .

3.2. Single Annulus Approach

The single annulus approach aims to construct a label-specific feature space by hierarchically capturing the underlying features of each label. According to each label, the positive and negative samples are separated into two different parts, i.e., P O S k and N E G k , respectively, based on whether it contains the kth label.
To mitigate the class imbalance problem, i.e., | P O S k | | N E G k | , the single annulus centers are selected for a smaller number of samples P O N k :
P O N k = P O S k , | P O S k | | N E G k | ; N E G k , | P O S k | > | N E G k | ,
where | · | denotes the cardinality of the set.
Based on the selected training set P O N k , the center of the annuluses can be determined as C k = mean PON k . To ensure that each annulus contains the same amount of information, the single annulus approach supposes the instances are divided into r annuluses under each class label, and the instances Q k j contained in the jth annulus are also determined accordingly.
For constructing a single annulus clustering feature space, the instances with the shortest sum of distances to others are taken as the cluster center of each layer. Therefore, the positive and negative cluster centers for the jth layer of the annuluses can be respectively expressed as follows:
p c j k = x i x Q k + j , q = 1 α d x i , x q q = 1 α d x , x q , n c j k = x i x Q k j , q = 1 α d x i , x q q = 1 α d x , x q .
Conceptually, we can already obtain the potential features in each layer of annuluses through single annulus clustering. Those features provide prototypes for constructing a single annulus clustering feature space. Here, mapping φ k : D S C F S k from the original training set to a single annulus clustering feature space can be defined as:
φ k x i = [ d x i , p c 1 k , , d x i , p c r k , d x i , n c 1 k , , d x i , n c r k ] ,
in which d ( · , · ) denotes the Euclidean distance between two instances.
Finally, a family of l classification models can be trained by the single annulus feature mapping φ k for the kth class label. Here, for each class label l k Y , a new binary training set G k is created from the original training set T , and the mapping of φ k can be set as follows:
G k = φ k x i , y i k x i D , y i k Y i .
Correspondingly, a classification model V k : S C F S k R for the kth class label can be induced by utilizing any binary learner [9]. Therefore, an unseen example x , which is associated with label set L , can be predicted as: Y = L k V k φ k ( x ) > 0 , 1 k l .

3.3. Double Annulus Field

The sample distribution tendency for each label represents an essential and intuitive connection with feature extraction. In this section, we attempt to construct the double annulus field based on the distribution tendency, where the classification difficulty of heterogeneous samples can be mitigated through the hierarchical division of double annuluses.
Before stratifying positive and negative samples, it is necessary to determine the two centers of the double annulus model. Through a large number of observations on a multi-label data set, it is obvious that most of the samples are clustered in the sample center of positive and negative samples. To preferentially partition the densely-distributed samples, the mean of P O S k is used to indicate the center of the positive samples, i.e., C p k uses the mean of P O S k to indicate the center of the positive samples, while C n k uses the mean of N E G k to indicate the center of the negative samples. In addition, P O S k contains the instances x i + with y i k = 1, and N E G k  contains the instances x i with y i k = 0. The numbers of P O S k and N E G k for the kth label are defined by m k + and m k , respectively. In further measuring the discrete degree of the samples, the distance of the heterogeneous samples from the center point is calculated for the kth label:
G k + = [ d ( x 1 + , C p k ) , d ( x 2 + , C p k ) , , d ( x m k + + , C p k ) ] , G k = [ d ( x 1 , C p k ) , d ( x 2 , C p k ) , , d ( x m k , C p k ) ] ,
where d ( · , · ) denotes the Euclidean distance between two instances and G k + represents the set of Euclidean distances between positive samples and C p k . In addition, S G k + is used to denote an ascending sort of G k + , while S G k is used to denote an ascending sort of G k .
It is well known that the cluster center points generated by random clustering will be affected by densely-distributed samples without obvious discrimination. Therefore, it is essential to ensure that the same amount of information is contained in each layer of the annulus. The extraction of latent features will not be ignored due to the extreme aggregation of samples. α represents the number of instances in each annulus, as long as the number of annuluses r is specified. Then, the maximum amount of information contained in each layer of positive samples can be defined as:
α k + = 1 r · | P O S k | .
Similarly, the maximum amount of information contained in each layer of negative samples can be defined as α k = 1 r · | N E G k | .
In the next step, the positive samples are stratified according to the ascending distance from the center point S G k + , which allows us to obtain instances with α k + stride separately according to the sorted order and assemble them into each layer of the annulus. More specifically, the maximum positive samples contained in the jth annulus for the kth label can be defined as:
R k + j = [ x u , x u + 1 , , x v 1 , x v ] v u + 1 = α k + × j S G k + .
Similarly, we can also obtain the maximum negative samples contained in the jth annulus for the kth label R k j . In other words, R k + j and R k j together form the double annulus field.

3.4. Intra-Annulus Clustering

Observing the distribution of the multi-label data set, we can see that, in addition to the positive and negative samples concentrated in C p k and C n k , there are also more samples at the junction of the positive and negative samples. To further reduce the difficulty of classification, the boundary samples between positive and negative samples are divided into left and right parts relative to the sample centers by PCA, i.e., P k + (relative left division of instances in P O S k ), P k (relative right division of instances in P O S k ), N k + (relative left division of instances in N E G k ), and N k (relative right division of instances in N E G k ). Correspondingly, P k + , P k , N k + , and N k are defined as follows:
P k + = x i π 1 x i < π 1 C p k , P k = x i π 1 x i π 1 C p k , N k + = x i π 1 x i < π 1 C n k , N k = x i π 1 x i π 1 C n k ,
where π n · uses PCA to reduce the features to n-dimensions.
Since the same amount of information is guaranteed to be contained in each annulus, there are potential features that cannot be ignored. Therefore, 2r cluster centers are generated according to the left and right division of relative positions in each layer of the annulus, and the Euclidean distance between each cluster center and other samples is guaranteed to be the smallest. For the convenience of the following expressions, we take “·” as the state parameter representing the two symbols of “±”. The intra-annulus cluster centers within the jth layer of the double annulus field for each label can be defined as:
p c j k · = x i x P k · R k · j , q = 1 p · d x i , x q q = 1 p · d x , x q , n c j k · = x i x N k · R k · j , q = 1 n · d x i , x q q = 1 n · d x , x q ,
where p · = | P k · | | R k · | , n · = | N k · | | R k · | .
In this way, sufficient distinguishable information is provided for forming the label-specific features through the mapping relationship between the instances and the annulus cluster centers. Those features provide prototypes for constructing label-specific feature spaces concerning the double annulus field. Meanwhile, a mapping φ k : D D E P T k from the original training set to label-specific feature space with respect to the double annulus field can be defined as:
φ k x i = d x i , p c 1 k + , , d x i , p c 2 r k , d x i , n c 1 k + , , d x i , n c 2 r k .

3.5. Classification

Therefore, a family of l classification models can be trained by the double annulus field’s label-specific feature space φ k for the kth class label. Here, for each class label l k y , a new binary training set h k is created from the original training set, and the mapping of φ k can be set as follows [9]:
h k = φ k x i , y i k x i D , y i k y i .
Correspondingly, a classification model V k : D E P T k R for the kth class label can be induced by utilizing any binary learner [9]. Therefore, an unseen example x , which is associated with label set L , can be predicted as [9]:
Y = L k V k φ k ( x ) > 0 , 1 k l ;
in other words, from Equations (4) and (12), each classifier h k with reference to the kth class label can be regarded as the composition of V k and φ k . Namely, h k ( x ) = V k φ k ( x ) = V k ( φ k ( x ) ) .

4. Experiments

4.1. Data Set

The details of the data set used in the experiments are shown in Table 1. In addition, we use Card, DL, and Den [1,29] to represent the average of label cardinality, the number of different label sets, and label density, respectively.

4.2. Evaluation Metrics

Compared with traditional single-label criteria, the performance of each evaluation metric is somewhat more complicated, as each instance is related to different labels concurrently. Therefore, Hamming loss (HL), average precision (AP), macro-averaging AUC (AUC), one-error (OE), coverage (CV), and ranking loss (RL) are selected in this paper. Given a test set T = x i , Y i 1 i m t , where Y i Y is the objectively true label subset, P x i Y is a predicted label vector for the ith instance, and μ j i is a confidence score that denotes the degree of x i that belongs to the label; the details of six evaluation metrics are given as follows.
Hamming loss [38] evaluates the number of misclassified instance-label pairs:
HL = 1 m t i = 1 m t | P x i Y i | ,
where ⊕ is the XOR operator.
Macro-averaging AUC [1] evaluates the average AUC value of each label:
AUC = 1 | L | i = 1 | L | | μ 1 , μ 2 μ 1 i μ 2 i , μ 1 , μ 2 Z j × Z ¯ j | | Z j | | Z ¯ j | .
One-error [39] evaluates whether the top-ranked predicted value of the samples is in the objectively true label subset:
OE = 1 m t i = 1 m t [ arg max y Y μ i ] Y i ,
where · is a function for logical judgment.
Ranking loss [40] evaluates the fraction of reversely-ordered label pairs. Namely, ranking and label relevance are negatively correlated:
RL = 1 m t i = 1 m t 1 | Y i | | Y ¯ i | | μ 1 , μ 2 μ 1 i μ 2 i , μ 1 , μ 2 Y i × Y ¯ i | ,
where Y ¯ i is the complementary set of Y i .
Coverage [41] evaluates the average of steps needed to move down the ranked label list that covers all the related labels of the samples:
CV = 1 m t i = 1 m t rank μ i 1 .
Average precision [42] evaluates the average fraction of relevant labels ranked higher than a particular label y Y i :
AP = 1 m t i = 1 m t 1 | Y i | y i Y i | y rank μ 1 i rank μ 2 i , y Y i | rank μ i .

4.3. Comparative Studies

In this section, the proposed DEPT algorithm is compared with the following six algorithms to verify their predictive performances. The only parameter of r is set to 12, according to the subsequent discussion.
  • MLNB [43]: This algorithm adopts the traditional naive Bayes classifiers to deal with multi-label instances, and the parameter is the default value of 0.3.
  • MULFE [36]: It combines ensemble learning with label correlation to construct a multi-label-specific feature space. The parameter of λ is set to 0.1 in this paper.
  • LIFT [9]: It transforms a multi-label into a single-label problem, and then the particular features of each label are used to form a label-specific feature space. In this paper, the ratio parameter ϵ is set to 0.1.
  • FRS-LIFT [12]: Based on label-specific features, this algorithm selects samples and reduces the dimension of the feature space by utilizing the idea of rough sets. The parameter setting is the same as LIFT.
  • ML-KNN [40]: This algorithm is derived from the traditional k-nearest neighbor (k-NN) algorithm, in which the number of nearest neighbors is set to 10, and the smoothing parameter is set to 1.
  • SCFS: Hierarchical label-specific features are extracted from the perspective of a single annulus. The parameter setting is the same as DEPT.
    For a fair comparison, LIBSVM [44] has utilized the linear kernel for training and predicting the induced binary classification model. Moreover, 10-fold cross-validation is used for each compared algorithm. Namely, the data set is divided into 10 groups of the same size; each algorithm is trained repeatedly 10 times on 9 random groups, and one group is used for testing.
Table 2, Table 3 and Table 4 report the detailed experimental results in terms of Hamming loss, average precision, macro-averaging AUC, one-error, ranking loss, and coverage, respectively. Results are compared by mean ± std, and the best performance for each data set is shown in bold. Additionally, “↑” and “↓” represent “the larger, the better” and “the smaller, the better”, respectively.
To analyze the statistical significance of each compared result, the Friedman test [45] with a significance level of 0.05 is employed; this is a well-known statistical test and has been widely utilized for statistically comparative studies over several data sets. Table 5 summarizes the Friedman statistics ( F F ) and the corresponding critical value of each evaluation metric. As shown in Table 5, the null hypothesis that all the compared algorithms have equal performance is rejected for each evaluation metric. In addition, the Bonferroni–Dunn test [46] is used as a post-hoc test to further explore the relative performance of each compared algorithm by treating DEPT as the control algorithm. Accordingly, the significant difference between the average ranks of compared algorithms can be distinguished from others by at least one critical difference (CD):
CD = q α k ( k + 1 ) 6 N ,
where k denotes the number of compared algorithms, and N denotes the number of data sets. For the Bonferroni–Dunn test, q α = 2.638 at significance level α = 0.05 , and thus CD = 2.5486 (k = 7, N = 10).
To intuitively observe the relative performance of DEPT and other comparison algorithms, Figure 3 shows the CD diagrams in terms of each evaluation metric, where the compared algorithm marked on the left axis has lower performance than the right one. Otherwise, any compared algorithm that is not significantly different from DEPT (within one CD) is connected with a black line.
Based on the reported experimental results, the following conclusions can be drawn:
1.
As shown in Figure 3, DEPT significantly outperforms MLNB in terms of each evaluation metric. Furthermore, DEPT differs significantly in performance from other algorithms in more than 38% of cases. That is, in terms of the average precision metric, DEPT is significantly different from MLNB and MLKNN. In terms of the macro-averaging AUC metric, DEPT is significantly different from MLNB and SCFS. In terms of the coverage metric, DEPT is significantly different from MLNB, MULFE, and SCFS. In terms of the Hamming loss metric, DEPT has is statistically significantly different from MLNB, MLKNN, and SCFS. In terms of the one-error metric, DEPT has a statistically significant difference between MLNB and SCFS. Therefore, the DEPT algorithm is statistically superior to others in more than 38% of cases.
2.
As shown in Figure 3, DEPT shows statistically superior or at least comparable performance against SCFS in each evaluation metric. Although the performance of SCFS is similar to MLNB due to the stratified feature space, SCFS uses a single annulus model that can not appropriately solve the cross-distribution of instances and unbalanced label density. Therefore, the double annulus model and further division of the instances are used for DEPT, thus improving the performance.
3.
As shown in Table 2 and Table 3, for the higher-dimensional data sets in this experiment, such as medical, with 1449 dimensions, DEPT ranks first in more than 55% of cases in terms of 6 evaluation metrics, which benefits from the layered strategy. In terms of 7 numeric types of data sets, DEPT ranks first in more than 52% of cases. Furthermore, compared with LIFT, which constructed label-specific features by the k-means algorithm, DEPT performs better on more than 63% of experimental results. DEPT is superior to FRSLIFT in more than 60% of cases, which used sample selection based on LIFT’s label-specific features. Compared with MULFE, which combined label-specific features and ensemble learning, DEPT performs better in more than 78% of cases. In addition, in terms of the average of 6 evaluation metrics, except for the result of one-error, which is close to LIFT, DEPT is superior in terms of macro-averaging AUC, average precision, coverage, ranking loss, and Hamming loss.
4.
Figure Figure 4 shows the spider web diagrams to intuitively illustrate the stability performance of each compared algorithm in terms of each evaluation metric. As shown in Figure 4, DEPT is more stable than the other algorithms. Namely, average precision and macro-averaging AUC tend to be larger, while coverage, ranking loss, Hamming loss, and one-error tend to be smaller.
To sum up, compared with other algorithms, under the condition that the AP is relatively stable, the additional evaluation metrics, such as CV, OE, and HL, have been improved to a certain extent, which finally verifies the effectiveness of the sample stratification and annulus clustering.

4.4. Further Analysis

4.4.1. Parameter Sensitivity

To further investigate the sensitivity of DEPT concerning the only parameter r, Figure 5 demonstrates the variation of each evaluation metric with the specified parameter, and the parameter value is sequentially set from 1 to 14. As shown in Figure 5, it is not difficult to draw the following conclusions:
1.
In most cases, DEPT has relatively poor performance when the parameter is set to 1–2, mainly because the layering effect is hard to achieve on a small number of annuluses.
2.
As the parameters increase sequentially, the performance improves from rapid increase to gradual stability.
3.
The performance reaches its relative optimum as the parameter r increases beyond 12. Therefore, these conclusions justify the parameter setting of DEPT in the experimental parts.

4.4.2. Execution Time

To study the runtime efficiency of SCFS and DEPT, Table 6 records the execution time of 5 algorithms with superior performance in Section 4.3. Combined with the experimental results, conclusions can be drawn as follows:
1.
Since FRSLIFT conducted sample selection when constructing the feature space, a lot of time is sacrificed while improving performance.
2.
Compared with LIFT, SCFS and DEPT can achieve shorter running time. This is mainly because the dimension of the constructed feature space φ d is different, specifically, φ d ( S C F S k ) = 2 r , φ d ( D E P T k ) = 4 r , and φ d ( L I F T k ) = ϵ · m i n ( | P O S | , | N E G | ) . Generally, the relationship can be written as follows:
φ d ( S C F S ) < φ d ( D E P T ) < φ d ( L I F T ) .
3.
The label-specific feature space constructed based on annulus clustering can improve runtime efficiency. The difference is that the hierarchical structure of double annuluses is better than a single annulus in terms of 6 evaluation metrics, and time consumption is also more remarkable than that of the single annulus. In short, these results validate the efficiency of SCFS and DEPT in learning from multi-label data.
Formally, the time complexity of SCFS can be calculated from as follows. Firstly, the cost of calculating the distance between the instances and the center of annuluses is O ( l | P O N k | ) . The time spent dividing the instances in P O N k into annuluses is O ( l r ) . Then, the time spent performing clustering within annuluses is l r ( R k * ) 2 . Therefore, the total time complexity of SCFS is O ( l ( | P O N k | + r + r ( R k * ) 2 ) ) . Analogously, the time complexity of DEPT is O ( l ( t + | P O S k | + | N E G k | + 2 r + 2 r ( R k * ) 2 ) ) , where t denotes the time of dimensional reduction of PCA [14] on P O S k and N E G k . It is obvious that the time complexity of DEPT is higher than that of SCFS.

4.4.3. DEPT vs. SCFS

Both the DEPT and SCFS strategies can achieve the effect of a hierarchical cluster. To explore the difference in performance between the two strategies, we carried out comparative experiments on the two algorithms in terms of different evaluation metrics. Figure 6 reports the detailed experimental results. On the one hand, the two strategies conform to the monotonicity, i.e., either an upward or a downward trend on each evaluation metric. On the other hand, the performance gap between the two algorithms becomes more minor as the number of annuluses increases from 1 to 20.

4.4.4. MCDA

Multi-criteria decision analysis (MCDA) [47] is an analytical method used to make choices among multiple criteria. In this paper, we use MCDA to solve conflicting requirements for speed and reliability of classification. Table 7 shows the calculated average accuracy rate, time consumption, and interaction index of all subsets in the annulus set and Mobius. In the fourth column of Table 7, Mobius corresponding to r 2 , r 3 , r 4 , and r 5 will be used as a measure of information redundancy in the joint contribution of multiple attributes. The fifth column represents the Shapley value, which can be interpreted as a measure of the importance of the corresponding attributes in the feature space, where r 1 and r 2 are complementary. The negative value of the corresponding influence index in this column indicates that the redundancy of the feature space increases with the increase in the number of annuluses. The fifth column can be similarly interpreted. From Table 7, it can be seen that the appropriate number of annuluses to balance time and efficiency is 5.

5. Conclusions

A new method for extracting label-specific feature strategies, DEPT, is proposed in this paper. To form the double annulus domain, the centers of the double annulus field model are shifted to the sample center of the positive and negative samples, and the positive and negative samples are divided into annuluses, respectively. In addition, the principal component analysis (PCA) technique is utilized to divide the samples in each layer of annuluses. Based on this strategy, intra-annulus clustering was performed to further distinguish the two classes of instances close to the center of distribution density, aiming to improve the performance of multi-label classification with class-imbalanced. Comparative experiments with six algorithms on 10 multi-label data sets show the proposed strategy is superior to some others in 6 evaluation metrics. However, the multi-annulus model still faces the following problems and challenges:
1.
Through the sensitivity analysis experiment in Section 4.4.1, we learned that increasing the number of annuluses can improve the overall classification accuracy but may lead to overfitting as the number of annuluses continues to increase.
2.
Increasing the number of annuluses can improve classification accuracy to a certain extent, but it can also significantly impact efficiency because every additional annulus requires an exponential increase in training iterations by a factor of 2.
3.
Based on the experimental results, the performance of the annulus model is not satisfactory for some data sets, which may be due to the single annulus model being unable to universally fit all data sets.
Future work will implement an annulus model that better fits various distribution trends of multi-label data sets and attempt to strike a balance between accuracy and efficiency.

Author Contributions

Conceptualization, Y.L. and C.L.; methodology, J.S.; software, Y.L.; validation, J.S., X.Y. and T.X.; formal analysis, P.W.; investigation, J.S.; resources, X.Y.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, J.S.; visualization, J.S.; supervision, X.Y.; project administration, X.Y.; funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Nos. 61906078, 62076111, 62006099) and the Key Laboratory of Oceanographic Big Data Mining.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, M.; Zhou, Z. A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 2013, 26, 1819–1837. [Google Scholar] [CrossRef]
  2. Rubin, T.; Chambers, A.; Smyth, P.; Steyvers, M. Statistical topic models for multi-label document classification. Mach. Learn. 2012, 88, 157–208. [Google Scholar] [CrossRef]
  3. Bromuri, S.; Zufferey, D.; Hennebert, J.; Schumacher, M. Multi-label classification of chronically ill patients with bag of words and supervised dimensionality reduction algorithms. J. Biomed. Inform. 2014, 51, 165–175. [Google Scholar] [CrossRef] [PubMed]
  4. Trohidis, K.; Tsoumakas, G.; Kalliris, G.; Vlahavas, I. Multi-label classification of music by emotion. EURASIP J. Audio Speech Music Process. 2011, 2011, 4. [Google Scholar] [CrossRef]
  5. Wu, B.; Zhong, E.; Horner, A.; Yang, Q. Music emotion recognition by multi-label multi-layer multi-instance multi-view learning. In Proceedings of the ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 117–126. [Google Scholar]
  6. Neville, J.; Jensen, D. Iterative classification in relational data. In Proceedings of the AAAI-2000 workshop Conference on Learning Statistical Models from Relational Data, Arlington, VA, USA, 12–15 May 2000; pp. 13–20. [Google Scholar]
  7. Wu, T.; Fan, J.; Wang, P. An improved three-way clustering based on ensemble strategy. Mathematics 2022, 10, 1457. [Google Scholar] [CrossRef]
  8. Wang, P.; Yao, Y. Ce3: A three-way clustering method based on mathematical morphology. Knowl. Based Syst. 2018, 155, 54–65. [Google Scholar] [CrossRef]
  9. Zhang, M.; Wu, L. LIFT: Multi-label learning with label-specific features. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 107–120. [Google Scholar] [CrossRef]
  10. Huang, J.; Li, G.; Huang, Q.; Wu, X. Learning label specific features for multi-label classification. In Proceedings of the IEEE International Conference on Data Mining, Atlantic City, NJ, USA, 14–17 November 2015; pp. 181–190. [Google Scholar]
  11. Jia, X.; Zhu, S.; Li, W. Joint label-specific features and correlation information for multi-label learning. J. Comput. Sci. Technol. 2020, 35, 247–258. [Google Scholar] [CrossRef]
  12. Xu, S.; Yang, X.; Yu, H.; Yu, D.; Yang, J.; Tsang, E. Multi-label learning with label-specific feature reduction. Knowl.-Based Syst. 2016, 104, 52–61. [Google Scholar] [CrossRef]
  13. Niknam, T.; Amiri, B.; Olamaei, J.; Arefi, A. An efficient hybrid evolutionary optimization algorithm based on PSO and SA for clustering. J. Zhejiang Univ. Sci. A 2009, 10, 512–519. [Google Scholar] [CrossRef]
  14. Abdi, H.; Williams, L. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  15. Zhang, P.; Gao, W.; Hu, J.; Li, Y. Multi-Label Feature Selection Based on High-Order Label Correlation Assumption. Entropy 2020, 22, 797. [Google Scholar] [CrossRef]
  16. Zhang, Q.; Zhong, G.; Dong, J. A Graph-based Semi-supervised Multi-label Learning Method Based on Label Correlation Consistency. Cogn. Comput. 2021, 13, 1564–1573. [Google Scholar] [CrossRef]
  17. Nguyen, H.; Woon, Y.; Ng, W. A survey on data stream clustering and classification. Knowl. Inf. Syst. 2015, 45, 535–569. [Google Scholar] [CrossRef]
  18. Read, J.; Bifet, A.; Holmes, G.; Pfahringer, B. Scalable and efficient multi-label classification for evolving data streams. Mach. Learn. 2012, 88, 243–272. [Google Scholar] [CrossRef]
  19. Braytee, A.; Liu, W.; Anaissi, A.; Kennedy, P. Correlated multi-label classification with incomplete label space and class imbalance. ACM Trans. Intell. Syst. Technol. 2019, 10, 1–26. [Google Scholar] [CrossRef]
  20. Liu, B.; Tsoumakas, G. Dealing with class imbalance in classifier chains via random undersampling. Knowl. Based Syst. 2020, 192, 105292. [Google Scholar] [CrossRef]
  21. Fan, Y.; Liu, J.; Weng, W.; Chen, B.; Chen, Y.; Wu, S. Multi-label feature selection with local discriminant model and label correlations. Neurocomputing 2021, 442, 98–115. [Google Scholar] [CrossRef]
  22. Liu, J.; Lin, Y.; Wu, S.; Wang, C. Online multi-label group feature selection. Knowl. Based Syst. 2018, 143, 42–57. [Google Scholar] [CrossRef]
  23. Boutell, M.R.; Luo, J.; Shen, X.; Brown, C.M. Learning multi-label scene classification. Pattern Recognit. 2004, 37, 1757–1771. [Google Scholar] [CrossRef]
  24. Zhang, M.; Li, Y.; Liu, X.; Geng, X. Binary relevance for multi-label learning: An overview. Front. Comput. Sci. 2018, 12, 191–202. [Google Scholar] [CrossRef]
  25. Elisseeff, A.; Weston, J. A kernel method for multi-labelled classification. Adv. Neural Inf. Process. Syst. 2001, 14, 681–687. [Google Scholar]
  26. Mencía, E.; Furnkranz, J. Pairwise learning of multilabel classifications with perceptrons. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 2899–2906. [Google Scholar]
  27. Brinker, C.; Mencía, E.; Fürnkranz, J. Graded multilabel classification by pairwise comparisons. In Proceedings of the 2014 IEEE International Conference on Data Mining, Shenzhen, China, 14–17 December 2014; pp. 731–736. [Google Scholar]
  28. Yazici, V.O.; Gonzalez-Garcia, A.; Ramisa, A.; Twardowski, B.; Weijer, J.V.D. Orderless recurrent models for multi-label classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13440–13449. [Google Scholar]
  29. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains for multi-label classification. Mach. Learn. 2011, 85, 333–359. [Google Scholar] [CrossRef]
  30. Tsoumakas, G.; Katakis, I.; Vlahavas, I. Random k-labelsets for multilabel classification. IEEE Trans. Knowl. Data Eng. 2010, 23, 1079–1089. [Google Scholar] [CrossRef]
  31. Song, J.; Tsang, E.; Chen, D.; Yang, X. Minimal decision cost reduct in fuzzy decision-theoretic rough set model. Knowl. Based Syst. 2017, 126, 104–112. [Google Scholar] [CrossRef]
  32. Zhan, W.; Zhang, M. Multi-label learning with label-specific features via clustering ensemble. In Proceedings of the 2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Tokyo, Japan, 19–21 October 2017; pp. 129–136. [Google Scholar]
  33. Hang, J.; Zhang, M. Collaborative Learning of Label Semantics and Deep Label-Specific Features for Multi-Label Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9860–9871. [Google Scholar] [CrossRef]
  34. Che, X.; Chen, D.; Mi, J. A novel approach for learning label correlation with application to feature selection of multi-label data. Inf. Sci. 2020, 512, 795–812. [Google Scholar] [CrossRef]
  35. Pei, G.; Wang, Y.; Cheng, Y.; Zhang, L. Joint label-density-margin space and extreme elastic net for label-specific features. IEEE Access 2019, 7, 112304–112317. [Google Scholar] [CrossRef]
  36. Lin, Y.; Hu, Q.; Liu, J.; Zhu, X.; Wu, X. MULFE: Multi-label learning via label-specific feature space ensemble. ACM Trans. Knowl. Discov. Data 2021, 16, 1–24. [Google Scholar] [CrossRef]
  37. Zhang, M.; Fang, J.; Wang, Y. BiLabel-Specific Features for Multi-Label Classification. ACM Trans. Knowl. Discov. Data 2021, 16, 1–23. [Google Scholar] [CrossRef]
  38. Godbole, S.; Sarawagi, S. Discriminative methods for multi-labeled classification. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 26–28 May 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 22–30. [Google Scholar]
  39. Schapire, R.; Singer, Y. Improved boosting algorithms using confidence-rated predictions. Mach. Learn. 1999, 37, 297–336. [Google Scholar] [CrossRef]
  40. Zhang, M.; Zhou, Z. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  41. Schapire, R.; Singer, Y. BoosTexter: A boosting-based system for text categorization. Mach. Learn. 2000, 39, 135–168. [Google Scholar] [CrossRef]
  42. Salton, G. Developments in automatic text retrieval. Science 1991, 253, 974–980. [Google Scholar] [CrossRef]
  43. Zhang, M.; Peña, J.; Robles, V. Feature selection for multi-label naive Bayes classification. Inf. Sci. 2009, 179, 3218–3229. [Google Scholar] [CrossRef]
  44. Chang, C.; Lin, C. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  45. Friedman, M. A comparison of alternative tests of significance for the problem of m rankings. Ann. Math. Stat. 1940, 11, 86–92. [Google Scholar] [CrossRef]
  46. Dunn, O. Multiple comparisons among means. J. Am. Stat. Assoc. 1961, 56, 52–64. [Google Scholar] [CrossRef]
  47. Montibeller, G.; Franco, A. Multi-criteria decision analysis for strategic decision making. In Handbook of Multicriteria Analysis; Springer: Berlin/Heidelberg, Germany, 2010; pp. 25–48. [Google Scholar]
Figure 1. Illustrative examples of four multi-label data sets in a two-dimensional case. The annuluses are generated in the distribution trend of a multi-label data set.
Figure 1. Illustrative examples of four multi-label data sets in a two-dimensional case. The annuluses are generated in the distribution trend of a multi-label data set.
Mathematics 11 01969 g001
Figure 2. Illustrative example of artificial data in a two-dimensional case. Double annuluses are generated in the centers of positive instances and negative instances.
Figure 2. Illustrative example of artificial data in a two-dimensional case. Double annuluses are generated in the centers of positive instances and negative instances.
Mathematics 11 01969 g002
Figure 3. Comparison of DEPT (control algorithms) against other compared algorithms with the Bonferroni–Dunn test.
Figure 3. Comparison of DEPT (control algorithms) against other compared algorithms with the Bonferroni–Dunn test.
Mathematics 11 01969 g003
Figure 4. Spider web diagrams show the stability performance of compared algorithms with different evaluation metrics.
Figure 4. Spider web diagrams show the stability performance of compared algorithms with different evaluation metrics.
Mathematics 11 01969 g004
Figure 5. Performance of DEPT changes in terms of each evaluation metric as the parameter r increases from 1 to 14 on four regular-scale data set.
Figure 5. Performance of DEPT changes in terms of each evaluation metric as the parameter r increases from 1 to 14 on four regular-scale data set.
Mathematics 11 01969 g005
Figure 6. Comparison of DEPT and SCFS in terms of each evaluation metric as the parameter annulus increases from 1 to 20 on 4 benchmark multi-label data sets.
Figure 6. Comparison of DEPT and SCFS in terms of each evaluation metric as the parameter annulus increases from 1 to 20 on 4 benchmark multi-label data sets.
Mathematics 11 01969 g006
Table 1. Characteristics of the 10 multi-label data sets.
Table 1. Characteristics of the 10 multi-label data sets.
Data SetmdqTypeDomainLabel Features
CardDLDen
Birds64526019bothaudio1.0141330.053
CHD-49555496numericmedicine2.580340.430
Emotions593726numericmusic1.869270.311
Flags194197bothimages3.392540.485
Images20002945numericimages1.236200.247
Medical978144945nominaltext1.245840.028
Reuters-K5006000500103numerictext1.4628110.014
Scene24072946numericimages1.074150.179
WaterQuality10601614numericchemistry5.0738250.362
Yeast241710314numericbiology4.2371980.303
Table 2. Experimental results of each compared algorithm (mean ± std).
Table 2. Experimental results of each compared algorithm (mean ± std).
Comparison
Algorithm
Average Precision ↑
DEPTSCFSLIFTMLKNNMLNBFRSLIFTMULFE
Birds0.750 ± 0.0440.736 ± 0.0410.608 ± 0.0300.587 ± 0.0220.579 ± 0.0590.674 ± 0.0230.733 ± 0.076
CHD-490.814 ± 0.0240.806 ± 0.0260.805 ± 0.0340.773 ± 0.0220.789 ± 0.0350.813 ± 0.0250.813 ± 0.033
Emotions0.819 ± 0.0290.815 ± 0.0210.829 ± 0.0070.814 ± 0.0170.768 ± 0.0260.803 ± 0.0320.812 ± 0.030
Flags0.828 ± 0.0330.822 ± 0.0260.820 ± 0.0210.796 ± 0.0700.802 ± 0.0430.812 ± 0.0300.819 ± 0.039
Images0.805 ± 0.0280.791 ± 0.0180.810 ± 0.0300.779 ± 0.0290.756 ± 0.0300.831 ± 0.0230.823 ± 0.020
Medical0.868 ± 0.0350.825 ± 0.0340.862 ± 0.0050.812 ± 0.0290.777 ± 0.0070.867 ± 0.0180.859 ± 0.044
Reuters-K5000.647 ± 0.0140.617 ± 0.0130.645 ± 0.0020.627 ± 0.0190.592 ± 0.0180.634 ± 0.0190.639 ± 0.016
Scenes0.878 ± 0.0150.865 ± 0.0140.886 ± 0.0050.860 ± 0.0110.839 ± 0.0140.896 ± 0.0160.888 ± 0.017
WaterQuality0.680 ± 0.0160.668 ± 0.0360.673 ± 0.0140.652 ± 0.0090.641 ± 0.0060.679 ± 0.0260.663 ± 0.023
Yeast0.782 ± 0.0190.760 ± 0.0160.766 ± 0.0060.763 ± 0.0210.741 ± 0.0060.778 ± 0.0180.768 ± 0.018
Average0.7870.7700.7700.7460.7290.7790.782
Comparison
algorithm
Hamming loss ↓
DEPTSCFSLIFTMLKNNMLNBFRSLIFTMULFE
Birds0.045 ± 0.0080.049 ± 0.0080.044 ± 0.0070.990 ± 0.0010.971 ± 0.0040.048 ± 0.0090.050 ± 0.005
CHD-490.262 ± 0.0320.300 ± 0.0160.278 ± 0.0230.322 ± 0.0350.289 ± 0.0330.280 ± 0.0330.283 ± 0.020
Emotions0.186 ± 0.0200.194 ± 0.0140.177 ± 0.0130.785 ± 0.0100.834 ± 0.0260.187 ± 0.0200.180 ± 0.020
Flags0.250 ± 0.0490.270 ± 0.0290.280 ± 0.0270.674 ± 0.0410.664 ± 0.0370.261 ± 0.0180.274 ± 0.012
Images0.176 ± 0.0140.197 ± 0.0120.151 ± 0.0130.169 ± 0.0080.199 ± 0.0070.154 ± 0.0160.157 ± 0.010
Medical0.010 ± 0.0010.015 ± 0.0020.012 ± 0.0020.984 ± 0.0010.984 ± 0.0010.013 ± 0.0010.014 ± 0.002
Reuters-K5000.011 ± 0.0010.011 ± 0.0000.011 ± 0.0000.996 ± 0.0000.963 ± 0.0000.022 ± 0.0000.013 ± 0.001
Scenes0.089 ± 0.0080.100 ± 0.0050.076 ± 0.0100.885 ± 0.0070.881 ± 0.0030.076 ± 0.0150.072 ± 0.006
WaterQuality0.296 ± 0.0090.307 ± 0.0110.297 ± 0.0250.839 ± 0.0020.839 ± 0.0050.300 ± 0.0270.307 ± 0.015
Yeast0.191 ± 0.0110.204 ± 0.0090.190 ± 0.0090.190 ± 0.0080.205 ± 0.0080.189 ± 0.0070.196 ± 0.010
Average0.1510.1650.1520.6830.6830.1530.155
Table 3. Experimental results of each comparison algorithm (mean ± std).
Table 3. Experimental results of each comparison algorithm (mean ± std).
Comparison
Algorithm
Macro-Averaging AUC ↑
DEPTSCFSLIFTMLKNNMLNBFRSLIFTMULFE
Birds0.789 ± 0.0450.704 ± 0.0690.807 ± 0.0230.851 ± 0.0480.780 ± 0.0320.708 ± 0.0650.742 ± 0.045
CHD-490.665 ± 0.0690.597 ± 0.0500.640 ± 0.0500.501 ± 0.0610.635 ± 0.0320.591 ± 0.0620.652 ± 0.043
Emotions0.840 ± 0.0290.836 ± 0.0220.851 ± 0.0140.847 ± 0.0190.814 ± 0.0130.849 ± 0.0190.843 ± 0.012
Flags0.722 ± 0.0690.691 ± 0.0660.700 ± 0.0420.686 ± 0.0770.661 ± 0.0770.707 ± 0.0170.721 ± 0.022
Images0.832 ± 0.0190.795 ± 0.0150.831 ± 0.0130.831 ± 0.0170.802 ± 0.0180.829 ± 0.0300.829 ± 0.018
Medical0.905 ± 0.0480.872 ± 0.0380.892 ± 0.0270.850 ± 0.0330.724 ± 0.0470.903 ± 0.0230.894 ± 0.035
Reuters-K5000.809 ± 0.0130.790 ± 0.0160.804 ± 0.0300.777 ± 0.0490.692 ± 0.0270.796 ± 0.0350.792 ± 0.022
Scenes0.935 ± 0.0100.920 ± 0.0050.928 ± 0.0080.924 ± 0.0160.909 ± 0.0130.929 ± 0.0140.921 ± 0.011
WaterQuality0.678 ± 0.0100.662 ± 0.0250.701 ± 0.0230.702 ± 0.0100.664 ± 0.0120.692 ± 0.0310.656 ± 0.014
Yeast0.729 ± 0.0230.638 ± 0.0230.681 ± 0.0320.702 ± 0.0030.675 ± 0.0300.623 ± 0.0250.683 ± 0.015
Average0.7900.7500.7840.7670.7360.7630.773
Comparison
algorithm
One-error ↓
DEPTSCFSLIFTMLKNNMLNBFRSLIFTMULFE
Birds0.668 ± 0.0600.657 ± 0.0480.689 ± 0.0530.675 ± 0.0490.710 ± 0.0380.705 ± 0.0380.694 ± 0.038
CHD-490.207 ± 0.0560.236 ± 0.0530.210 ± 0.0710.219 ± 0.0350.235 ± 0.0540.234 ± 0.0850.232 ± 0.061
Emotions0.204 ± 0.0750.250 ± 0.0460.212 ± 0.0260.254 ± 0.0170.333 ± 0.0600.275 ± 0.0600.232 ± 0.070
Flags0.203 ± 0.0660.231 ± 0.0840.248 ± 0.0740.263 ± 0.1050.234 ± 0.1290.232 ± 0.1200.232 ± 0.139
Images0.316 ± 0.0540.401 ± 0.0290.285 ± 0.0560.343 ± 0.0490.372 ± 0.0450.253 ± 0.0350.264 ± 0.040
Medical0.194 ± 0.0480.212 ± 0.0450.174 ± 0.0110.242 ± 0.0330.435 ± 0.0560.163 ± 0.0270.196 ± 0.042
Reuters-K5000.420 ± 0.0180.454 ± 0.0180.421 ± 0.0100.442 ± 0.0240.474 ± 0.0360.465 ± 0.0270.453 ± 0.045
Scenes0.222 ± 0.0230.255 ± 0.0250.195 ± 0.0250.240 ± 0.0140.261 ± 0.0180.185 ± 0.0290.195 ± 0.027
WaterQuality0.286 ± 0.0440.314 ± 0.0440.327 ± 0.0290.277 ± 0.0240.340 ± 0.0280.318 ± 0.0790.290 ± 0.032
Yeast0.221 ± 0.0300.244 ± 0.0260.228 ± 0.0100.222 ± 0.0380.240 ± 0.0070.238 ± 0.0230.229 ± 0.023
Average0.2940.3250.2990.3180.3630.3070.302
Table 4. Experimental results of each comparison algorithm (mean ± std).
Table 4. Experimental results of each comparison algorithm (mean ± std).
Comparison
Algorithm
Ranking Loss ↓
DEPTSCFSLIFTMLKNNMLNBFRSLIFTMULFE
Birds0.170 ± 0.0500.222 ± 0.0490.173 ± 0.0330.153 ± 0.0390.172 ± 0.0490.206 ± 0.0280.193 ± 0.065
CHD-490.234 ± 0.0520.238 ± 0.0280.211 ± 0.0280.260 ± 0.0660.254 ± 0.0230.223 ± 0.0360.238 ± 0.038
Emotions0.148 ± 0.0310.161 ± 0.0230.149 ± 0.0140.151 ± 0.0210.184 ± 0.0130.140 ± 0.0210.148 ± 0.009
Flags0.206 ± 0.0410.209 ± 0.0310.208 ± 0.0140.222 ± 0.0710.212 ± 0.0220.217 ± 0.0290.220 ± 0.019
Images0.167 ± 0.0220.207 ± 0.0160.140 ± 0.0160.182 ± 0.0240.206 ± 0.0250.142 ± 0.0230.143 ± 0.011
Medical0.021 ± 0.0200.031 ± 0.0130.024 ± 0.0120.037 ± 0.0100.055 ± 0.0060.029 ± 0.0040.032 ± 0.009
Reuters-K5000.040 ± 0.0040.067 ± 0.0030.050 ± 0.0030.061 ± 0.0020.085 ± 0.0040.483 ± 0.0030.103 ± 0.005
Scenes0.074 ± 0.0090.086 ± 0.0090.064 ± 0.0030.079 ± 0.0090.097 ± 0.0130.059 ± 0.0100.063 ± 0.011
WaterQuality0.279 ± 0.0130.295 ± 0.0220.266 ± 0.0170.256 ± 0.0040.294 ± 0.0110.271 ± 0.0250.296 ± 0.025
Yeast0.147 ± 0.0160.175 ± 0.0140.164 ± 0.0120.151 ± 0.0150.178 ± 0.0120.153 ± 0.0140.163 ± 0.007
Average0.1490.1690.1450.1550.1740.1920.160
Comparison
algorithm
Coverage ↓
DEPTSCFSLIFTMLKNNMLNBFRSLIFTMULFE
Birds0.102 ± 0.0450.114 ± 0.0430.123 ± 0.0450.113 ± 0.0180.119 ± 0.0590.139 ± 0.0120.133 ± 0.052
CHD-490.433 ± 0.0260.451 ± 0.0210.450 ± 0.0170.497 ± 0.0300.454 ± 0.0030.444 ± 0.0100.453 ± 0.020
Emotions0.284 ± 0.0230.301 ± 0.0400.290 ± 0.0230.294 ± 0.0290.315 ± 0.0040.289 ± 0.0240.301 ± 0.020
Flags0.513 ± 0.0330.526 ± 0.0450.514 ± 0.0280.536 ± 0.0450.539 ± 0.0370.576 ± 0.0220.552 ± 0.024
Images0.189 ± 0.0140.220 ± 0.0200.164 ± 0.0080.192 ± 0.0180.224 ± 0.0180.172 ± 0.0230.166 ± 0.017
Medical0.040 ± 0.0230.059 ± 0.0160.042 ± 0.0180.060 ± 0.0130.313 ± 0.0340.048 ± 0.0090.055 ± 0.021
Reuters-K5000.076 ± 0.0080.084 ± 0.0050.073 ± 0.0060.090 ± 0.0040.143 ± 0.0050.071 ± 0.0060.095 ± 0.003
Scenes0.075 ± 0.0080.086 ± 0.0080.065 ± 0.0070.080 ± 0.0100.299 ± 0.0160.067 ± 0.0100.084 ± 0.007
WaterQuality0.657 ± 0.0180.675 ± 0.0160.638 ± 0.0340.620 ± 0.0130.666 ± 0.0070.639 ± 0.0390.669 ± 0.023
Yeast0.434 ± 0.0160.443 ± 0.0150.455 ± 0.0230.428 ± 0.0100.460 ± 0.0180.475 ± 0.0100.474 ± 0.016
Average0.2800.2960.2810.2910.3530.2920.298
Table 5. Summary of the Friedman statistics F F (k = 7, N= 10) and the critical value in terms of each evaluation metric (k: number of comparison algorithms; N: number of data sets).
Table 5. Summary of the Friedman statistics F F (k = 7, N= 10) and the critical value in terms of each evaluation metric (k: number of comparison algorithms; N: number of data sets).
Evaluation Metric F F Critical Value ( α = 0.05)
Average precision17.86572.2720
Macro-averaging AUC7.0714
Coverage6.4649
Ranking loss5.2534
Hamming loss13.8468
One-error6.0898
Table 6. Execution time of four compared algorithms (mean ± std) on 6 regular-scale data sets.
Table 6. Execution time of four compared algorithms (mean ± std) on 6 regular-scale data sets.
Execution TimeData SetSCFSDEPTLIFTFRSLIFT
Total time
(in seconds)
Emotions0.227 ± 0.0060.245 ± 0.0100.301 ± 0.01448.090 ± 0.157
CHD490.373 ± 0.0270.429 ± 0.0360.744 ± 0.28553.389 ± 0.781
Images2.672 ± 0.0284.323 ± 0.0576.287 ± 0.0161879.930 ± 57.506
WaterQuality4.125 ± 0.1425.436 ± 0.2166.117 ± 0.106727.960 ± 14.067
Scene5.082 ± 0.1166.036 ± 0.1446.175 ± 0.0611599.054 ± 47.101
Yeast12.839 ± 0.15624.689 ± 0.47046.706 ± 0.73617,198.588 ± 763.569
Table 7. Annuluses of DEPT in Mobius representation and mutual influence index.
Table 7. Annuluses of DEPT in Mobius representation and mutual influence index.
Number of Annuluses Time Accuracy Mobius Shapley Banzhaf
{ r 1 ,r2} 0.21 0.57 –0.67 0.67 0.67
{ r 2 ,r3} 0.34 0.62 0.17 0.32 0.29
{ r 2 ,r3,r4} 0.66 0.66 0 0.11 0.11
{ r 2 ,r3,r4,r5} 1.35 0.75 0.32 0.44 0.5
{ r 2 ,r3,r4,r5,r6} 2.76 0.77 0.67 –0.67 –0.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Liu, C.; Song, J.; Yang, X.; Xu, T.; Wang, P. Multi-Scale Annulus Clustering for Multi-Label Classification. Mathematics 2023, 11, 1969. https://doi.org/10.3390/math11081969

AMA Style

Liu Y, Liu C, Song J, Yang X, Xu T, Wang P. Multi-Scale Annulus Clustering for Multi-Label Classification. Mathematics. 2023; 11(8):1969. https://doi.org/10.3390/math11081969

Chicago/Turabian Style

Liu, Yan, Changshun Liu, Jingjing Song, Xibei Yang, Taihua Xu, and Pingxin Wang. 2023. "Multi-Scale Annulus Clustering for Multi-Label Classification" Mathematics 11, no. 8: 1969. https://doi.org/10.3390/math11081969

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop