Next Article in Journal
Parabolic Hessian Quotient Equation in Exterior Domain
Previous Article in Journal
Global Sensitivity Analysis of Structural Reliability Using Cliff Delta
Previous Article in Special Issue
A Semantic Enhancement Framework for Multimodal Sarcasm Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Updating Correlation-Enhanced Feature Learning for Multi-Label Classification

School of Computer Engineering, Chengdu Technological University, Chengdu 611730, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(13), 2131; https://doi.org/10.3390/math12132131
Submission received: 14 May 2024 / Revised: 30 June 2024 / Accepted: 5 July 2024 / Published: 7 July 2024

Abstract

:
In the domain of multi-label classification, label correlations play a crucial role in enhancing prediction precision. However, traditional methods heavily depend on ground-truth label sets, which can be incompletely tagged due to the diverse backgrounds of annotators and the significant cost associated with procuring extensive labeled datasets. To address these challenges, this paper introduces a novel multi-label classification method called updating Correlation-enhanced Feature Learning (uCeFL), which extracts label correlations directly from the data instances, circumventing the dependency on potentially incomplete label sets. uCeFL initially computes a revised label matrix by multiplying the incomplete label matrix with the label correlations extracted from the data matrix. This revised matrix is then utilized to enrich the original data features, enabling a neural network to learn correlation-enhanced representations that capture intricate relationships between data features, labels, and their interactions. Notably, label correlations are not static; they are dynamically updated during the neural network’s training process. Extensive experiments carried out on various datasets emphasize the effectiveness of the proposed approach. By leveraging label correlations within data instances, along with the hierarchical learning capabilities of neural networks, it offers a significant improvement in multi-label classification, even in scenarios with incomplete labels.

1. Introduction

Given the widespread occurrence of multi-label issues, researchers have introduced various multi-label classification techniques across diverse domains [1,2,3,4]. Unlike traditional single-label classification tasks [5,6], the semantic polysemy associated with multi-label data contributes to the complexity of these challenges. Specifically, as the number of categorical labels proliferates exponentially, the challenge of selecting an apt label set for novel data intensifies significantly [7].
Multi-label sample data consists of a wide array of category labels, many of which exhibit either direct or indirect relationships with each other. For instance, consider two illustrative images: one featuring a pyramid alongside a camel and the other displaying a pyramid situated in a desert. Initially, the camel and the desert in these two images may appear unrelated at a cursory glance. However, upon deeper examination of the contextual relationships inherent within these images, it becomes evident that there is an indirect correlation between the labels “camel” and “desert”. This observation is noteworthy as the terms “camel” and “desert” typically evoke a direct association in real-world scenarios. Research indicates that leveraging these label correlations, especially in scenarios with numerous category labels, can significantly reduce the complexity of classification problems and enhance classifier performance. The investigation of label correlations thus fuels the progress of multi-label classification, constituting a crucial research direction in this field [8,9].
Multi-label classification methods primarily categorize into four key strategies that capitalize on label correlations:
(1) Implicit exploitation of label dependencies, which involves partitioning a multi-label dataset into subsets based on labels and individually learning the data features within each subset. The representative method is label powerset (LP) [10], a simple method that treats each label subset in a training set as a single new label to construct a single-label classification problem. Although LP considers label correlations, it may produce a large number of classes, each with only a few examples. To solve the problem of insufficient training samples of new labels generated by LP, Read proposed pruned problem transformation (PPT) [11], which clips label subsets whose frequencies are much less than a threshold to smaller label subsets with slightly higher frequencies. Random k-Labelsets (RAkEL) [12] is an ensemble algorithm based on LP, which randomly selects m non-repetitive label subset combinations composed of k labels from an original set of K category labels, and trains an LP classifier on the data of each label subset. Multiple LP classifiers vote for the prediction results. Parameters k and m are set artificially. RAkEL degenerates to a BR algorithm when K = 1 and m = K and LP when k = K and m = 1. RAkEL considers correlations between labels while avoiding the drawbacks of LP. However, for the best results, it requires a large number of datasets, as well as cross-testing of input parameters such as subset size and threshold. It is difficult to find the optimal parameters without sufficient training samples.
(2) Incorporating label correlations as prior knowledge during the training of classification models, thereby leveraging pre-existing relationships between labels. Wang et al. [13] proposed a subspace learning algorithm based on label sparse coding for feature extraction. This implies the exploitation of label correlations to annotate different images with the same label and appropriate weights. Zhang and Zhou [7] used label correlations to find a projection matrix that maximizes correlations between data features and labels in the dimensionally reduced feature space. Wang et al. [14] calculated the cosine similarity between category labels and constrained the learning of the objective function with the similarity matrix to obtain a Green’s function enhanced by the label correlations.
(3) Concurrent discovery of label correlations during the training process of the multi-label model, allowing for dynamic exploitation of inter-label relationships. Huang et al. [15] assumed that label correlations are local because sample sets share different label correlations, few of which apply globally. To encode the local effects of label correlations, they derived an LOC code to enhance the feature representation of each sample and combined the global discriminant fit and local correlation sensitivity in a unified framework. Xu et al. [9] captured label correlations by a low-rank structure and introduced a low-rank constraint of the label correlation matrix in the model learning process. Gu et al. [16] used a normal distribution of matrix variables on the vector of classifiers to model label correlations and minimized the label correlation regularized loss of label ranking based on a subset of features. Ji et al. [17] proposed a general framework to extract shared structures, within which a binary classifier was constructed for each label. Unlike traditional binary classifiers, all binary classifiers were defined on both the original data space and low-dimensional shared structures to capture label correlations.
(4) Augmenting the original feature space with a label space, thereby enriching the data with additional information stemming from label correlations. Read et al. introduced classifier chains (CCs) [18], constructing a binary classifier for each label, whose result is used to expand the feature space of the next classifier. However, the order of construction affects classification performance. Montañes et al. proposed dependent binary relevance (DBR) [19], which learns a binary classifier for each label, each employing k − 1 features made up of other labels to expand the original feature space. Learning the label-expanded feature space, its binary classifiers can discover label correlations by themselves. However, the number of classifiers is proportional to the label length.
The aforementioned approaches primarily derive label correlations from the label sets available in the training data. However, these label sets are often incomplete, which is mainly attributed to the diverse backgrounds of grassroots users annotating data features [20] and the high cost associated with extracting a significant number of labeled samples for large-scale tasks [21]. Consequently, the accuracy of label correlations can suffer when labels are missing.
In the context of multi-label learning, a single instance is simultaneously associated with multiple class labels, and each label is determined by unique features specific to that instance. Since the amount of data feature information varies across objects corresponding to different category labels, the importance of labels within the same instance also differs. Nevertheless, the binary value (0 or 1) assigned to a label y i j in a tagged label set merely signifies the presence or absence of the feature represented by label j on data x i . This approach fails to capture the significance of the specific feature corresponding to label j . As a result, relying solely on the given label matrix to derive label correlations may lead to distorted classification results.
Taking into account the aforementioned challenges, a novel multi-label learning architecture known as updating Correlation-enhanced Feature Learning (uCeFL) is introduced in this paper, specifically tailored for multi-label classification tasks. This approach harnesses label information to enrich the features of the original data, echoing the principles outlined in point 4 of the previous discussion. Notably, rather than merely incorporating the given labels directly into the expanded feature space, uCeFL enriches the data feature space by utilizing labels completed through label correlation. This label correlation is extracted from the data matrix, ensuring resilience against missing labels. This innovative approach effectively circumvents the challenges posed by incomplete label matrices, resulting in more precise and robust multi-label classification.
uCeFL employs a covariance matrix to efficiently capture the distinct features of each subset of data affiliated with individual category labels. Through the comparison of covariance matrices within an n × n dimensional space, it accurately quantifies the correlation among these labels. It is universally acknowledged that the covariance between two variables increases as their interaction intensifies, and a covariance matrix comprehensively characterizes the interactions among all variables. The similarity between two covariance matrices suggests that their respective characteristic variables exhibit comparable distributions. Since category labels serve as abstract representations of interrelated data features, it is logically justifiable to express the correlation among labels through the similarity of covariance matrices pertaining to categorical data features. By deriving label relationships directly from the data, uCeFL effectively mitigates any potential interference caused by incomplete labels.
Regarding the intricate new data structure that is expanded with complete labels, this paper refrains from making any assumptions. However, since real-world data often reside in a low-dimensional manifold embedded within a high-dimensional space, the new data structure can be characterized as a nonlinear manifold. Multi-layer neural networks excel at simulating such nonlinear manifold structures without relying on hypotheses. They can acquire high-level data features by learning the hierarchical structure. Therefore, this paper utilizes a neural network to efficiently learn the multi-label data with increased dimensions, obtaining the fundamental variables that explain the enhanced data features. Label correlations are not merely considered as prior knowledge; rather, they are dynamically updated throughout the neural network training process, which ensures that our neural network continuously optimizes the label correlation, ultimately leading to more accurate and refined predictions.
Moreover, to emphasize the effectiveness of uCeFL, this paper introduces a variant named Correlation-enhanced Feature Learning (CeFL). This approach solely treats label correlations as prior knowledge, serving as a static input to the neural network throughout the entire training process. This input remains constant, ensuring consistency and stability.
Distinct from numerous existing methods, uCeFL distinguishes itself as being significantly more intuitive and comprehensible. Rigorous experiments spanning seven diverse databases have unequivocally established its superiority. Moreover, the label correlations inherently extracted from the data effectively capture the varying degrees of importance ascribed to each label, thereby reflecting the intricate relationships and interactions among the distinct features represented by these labels.
In essence, the key contributions of this article are outlined as follows:
  • Our method distinguishes itself from the preponderance of multi-label correlation learning algorithms by directly extracting label correlations from the data itself, thereby enhancing their precision beyond those derived solely from pre-assigned labels.
  • We introduce a revised label matrix, obtained by multiplying the incomplete label matrix with the captured label correlation, which serves to enrich the original data’s feature representation. A multi-layer neural network is then employed to learn the correlation-enhanced features that encapsulate intricate relationships among data features, labels, and their interconnections.
  • Leveraging the high-level semantic features extracted by the multi-layer neural network, our uCeFL approach re-evaluates the label correlation, facilitating continuous updates of these correlations throughout the neural network’s learning trajectory. This approach effectively captures the label propagation effects, ensuring a comprehensive and accurate representation of the label space.
The remainder of this work is structured as follows. Section 2 provides the related work, exploring the existing literature and prior studies pertinent to our investigation. Section 3 delves into the uCeFL approach, offering a comprehensive explanation of the methodologies utilized and the architecture adopted in our research. Additionally, an introduction to CeFL is also provided. Section 4 presents the experimental results and discusses the outcomes obtained through rigorous testing and validation. Finally, Section 5 outlines our conclusions and future work, summarizing the key findings of our study and highlighting potential avenues for further exploration.

2. The Related Work

Among the various multi-label classification algorithms, classifier chains (CC) [18] occupy a significant position as a noteworthy approach. However, a major limitation of CC lies in its sequential nature, where each classifier’s prediction heavily depends on the output of its predecessor. This sequential dependency often leads to error accumulation and propagation, adversely affecting the overall prediction accuracy. To mitigate this limitation, dependent binary relevance (DBR) [19] methods have emerged as a promising alternative. DBR constructs independent binary classifiers for each label, incorporating features relevant to other labels to enrich the input data space. This allows classifiers to independently capture and utilize label dependencies.
However, traditional binary classifiers may not fully realize the potential of DBR when dealing with specific multi-label loss functions. To enhance DBR’s performance, various optimization algorithms have been proposed. Rauber et al. [22] introduced a recursive approach that leverages DBR’s initial predictions as a starting point. This method iteratively improves predictions through feedback, aiming to boost accuracy by continuously modeling label dependencies. Alternatively, Zhang et al. [23] proposed a DBR pruning technique based on Phi coefficients, which evaluates label-task relevance, optimizes the model structure, reduces costs, and demonstrates practical effectiveness. Furthermore, the RDBR algorithm [24], a novel approach within decision tree methods, generates crucial intermediate values for constructing DP matrices. While it offers a unique perspective, ensemble methods still provide superior diversity and robustness.
With the progress of multi-label classification research, researchers have increasingly focused on two key directions: label-specific feature learning and label dependency modeling. These approaches aim to learn the most suitable feature representation for each category and utilize label correlations to improve classification performance.
Label-specific feature learning aims to extract features closely related to the label from the data to improve classification performance. Zhang et al. [25] employed clustering techniques to delve into the inherent patterns of the feature space. However, this clustering approach was executed in a label-specific manner, overlooking the potential interplay between different class labels. ML-DFL, introduced by Zhang et al. [26], relies on a novel spectral clustering algorithm known as Spectral Instance Alignment (SIA). This method excels in capturing the latent patterns between positive and negative instances for each label. In another pursuit, RFS-LIFT [27] endeavors to select a refined set of label-specific features generated by LIFT, leveraging the fuzzy rough set framework. Concurrently, [28] introduced an efficient multi-label feature selection technique grounded in an information-theoretic label selection approach. This method identifies a subset of labels that significantly influence feature importance, enabling an effective output of a tailored feature subset. Nonetheless, a significant oversight among these methodologies is the lack of consideration for label correlations, which could potentially bolster their performance even further.
Label dependency modeling is another crucial research direction in multi-label classification, focusing on capturing and utilizing dependencies between labels. Loza et al. [29] tackle label dependencies in multi-label classification. They propose using rules to represent dependencies and introduce two rule-learning approaches: a bootstrapped stacking method that learns separate rulesets for each label and an algorithm that reincorporates predicted labels for learning subsequent rules. Both methods make dependencies explicit in rules and employ standard techniques to ensure rule usefulness. Li et al. [30] proposed the A-GCN, which leverages popular Graph Convolutional Networks with an adaptive label correlation graph to model label dependencies. Wang et al. [31] leverage Bayesian networks (BNs) to model the intricate correlations among labels and introduce a novel BN-based classifier chain method (BNCC). They utilize conditional entropy to capture the dependencies between labels, transforming the labels into nodes in a Bayesian Network (BN). The strength of these dependencies is represented by the weights of the edges. Furthermore, they devise a novel scoring function to evaluate the quality of the BN structure and employ a heuristic algorithm to refine and optimize the BN. Ultimately, through the application of topological sorting on the optimized BN’s nodes, a label order is derived, which serves as the foundation for constructing the classifier chain model.
Rather than focusing solely on label-specific features or label correlation modeling, some studies have also attempted to combine these two techniques to achieve better classification results. Zhan et al. [32] present LIFTACE, a multi-label learning method that leverages clustering ensemble techniques to generate label-specific features while accounting for label correlations, thus enhancing LIFT’s learning prowess. Zhang et al. [33] integrate partial least squares discriminant analysis (PLS-DA) with an l 1 -norm penalty, effectively capturing the intricate relationships between variables and labels while simultaneously reducing feature dimensionality. In another study, Liu et al. [34] employ partial least squares (PLS) to identify a latent shared space that bridges both data and label variable spaces, laying the foundation for an efficacious learning model. Ma et al. [35] introduce a label-specific dual graph neural network (LDGN) that harnesses categorical insights to derive label-tailored representations from documents. Central to its approach is a dual Graph Convolution Network (GCN) that comprehensively models the intricate and adaptive interactions among these representations, leveraging both statistical label co-occurrences and dynamic reconstruction graphs in a synergistic fashion.
Although numerous methods have been investigated to incorporate label-specific features and label correlations, many algorithms extract these features within the original feature space, which may not be the most effective approach. Therefore, this paper utilizes a neural network framework to map the correlation-augmented feature space into an embedded feature space. This approach captures the intricate interplay between label-specific features, labels, and their interactions while continuously optimizing label correlations. Ultimately, this leads to more accurate and refined predictions.

3. Algorithm Description

A multi-label classification problem can be defined as follows: D = { x i , y i } i = 1 N be a training dataset consisting of N instances, where each instance is associated with K possible category labels ( l 1 , . . . , l K ). Each instance x i is represented by a feature vector x i = x i 1 , x i 2 , . . . , x i m T containing m attribute values. The corresponding label set for x i is denoted by a binary vector y i = y i 1 , y i 2 , . . . , y i K T , where y i j = 1 indicates that the instance x i belongs to label j , and y i j = 0 otherwise.
The training dataset can be further represented as data matrix X = x 1 , x 2 , , x N and label matrix Y = y 1 , y 2 , , y N . The goal of this paper is to leverage the information contained in X and Y to predict the label sets for unlabeled data points.

3.1. Motivation

Category labels serve as concise representations of intrinsic features, and the coexistence of various features within a sample indicates underlying interdependencies among those labels. Therefore, to attain accurate predictions for data samples’ labels, it is crucial to consider label correlations that possess substantial informational worth. Akin to numerous multi-label classification methodologies, such as CC and DBR, our research explicitly utilizes labels as data attributes. However, the distinction lies in our introduction of innovative approaches that harness label correlations inherent in the data itself rather than merely relying solely on ground-truth labels.
This paper first constructs the label-expanded feature vector z i = x i ; y i T , which is utilized instead of the data vector x i to train the model. Obviously, z i contains more distinguishable information than x i , and therefore, training with z i can achieve better classification performance. However, in multi-label problems, sample data with different category labels may share the same labels, suggesting that indirect relationships may exist between labels. This means that solely studying the combination z i may not be sufficient, especially when there are missing labels in y i , as y i , the components of z i , contains only the tagged labels of the corresponding instance; that is, y i can only indicate direct correlations between labels.
To fully leverage labels (including both direct and indirect relationships), this paper introduces a label correlation matrix K . This matrix is used to transform the original labels y i into y i ^ = K × y i , effectively completing label sets with missing labels by utilizing a matrix [36]. Furthermore, by capturing varying degrees of importance among labels, the transformed labels K × y i enable a stronger association between data points with similar characteristics. Conventionally, the label matrix K is defined using the normalized cosine similarity, as described in Equation (1):
K p q = c o s y p , y q = < y p , y q > y p y q ,
where p and q are category labels, with corresponding row vectors y p and y q in label matrix Y . There are many approaches to improve the performance of multi-label classification using a label correlation matrix computed by cosine similarity [37,38]. However, as the amount of data increases and the data types become increasingly diverse, it is difficult to obtain the complete label information of each instance, which results in unreliable label correlations.
Assume that the sample data points x 1 and x 2 are accompanied by their corresponding label vectors y 1 and y 2 , respectively. These label vectors are defined as
y 1 = 1 0 1 a n d y 2 = 0 1 0 ,   respectively ,
which implies that x 1 comprises labels l 1 and l 3 , while x 2 solely contains l 2 . When assessing label correlations via cosine similarity with the original label vectors, it can be deduced that label l 1 is related to label l 3 , whereas label l 2 seems to be completely separate. However, the accuracy of label correlations derived from cosine similarity is compromised in scenarios where, despite x 2 possessing features indicative of label l 1 , this label is absent from the label vector y 2 . Consequently, it becomes imperative to rely on the data matrix, rather than the label matrix, for calculating label correlations whenever missing labels are encountered.

3.2. The Proposed CeFL

Motivated by the desire to enrich the data with supplementary information derived from label correlations and to minimize the influence of missing labels by relying on the data matrix for the computation of these correlations, this paper introduces a novel multi-label learning architecture termed updating Correlation-enhanced Feature Learning (uCeFL). This innovative approach is schematically represented in Figure 1. Subsequently, a detailed explanation of the architecture is provided as follows.
Let X p = { x i | x i , y i D , y i , p = 1 , i = 1 , , N p } denotes a sample subset with label p in the original dataset D, where N p is the number of samples in the subset. The covariance matrix similarity coefficient K c o r r R K × K in Equation (2) is used to obtain a more accurate relationship between labels p and q from the data:
K c o r r C p , C q = t r C p C q C p F C q F ,
t r { . } is the matrix trace, and . F is the Frobenius norm, K c o r r C p , C q = 1 d c o r r C p , C q , where d c o r r is the correlation matrix distance [39], which becomes zero if the covariance matrices are equal up to a scaling factor and 1 if they differ to a maximum extent. C p and C q are the covariance matrices of sets X p and X q , respectively. The covariance matrix is actually a point on the Riemannian manifold, so K c o r r C p , C q can be calculated by the log-Euclidean metric [40], i.e., logarithmically mapping the points from the Riemannian manifold to the Euclidean space and calculating their correlation.
Without prior assumptions on the data distribution, the covariance matrix provides a natural representation for a dataset with any type of features and with an arbitrary number of samples, which can effectively distinguish data subsets with different category labels by encoding the feature-related information specific to each object class. The higher the degree of feature overlap between two subsets, the closer the relationship between the category labels and the greater the value of Equation (2). If the label matrix is used directly to calculate the label correlations, the feature overlap between sample sets cannot be truly represented, while the label correlation may be distorted to some extent when labels are missing. In short, by utilizing a covariance matrix to model the data subset, the issue of label correlation calculation can be resolved, even in datasets with incomplete labeling. Consequently, the label-expanded feature vector z i is updated to z i = x i ; y i ^ T , where y i ^ = K c o r r × y i represents the revised label information.
As the feature space becomes larger, the data structure becomes more complicated, and this paper uses the multi-layer neural network to study the label-enhanced dataset D = { z i , y i } i = 1 N because the neural network can simulate the nonlinear manifold structure of the data with no assumptions. If there are L + 1 layers in the neural network, for feature z i , the output h i l in the l -th layer is calculated as Equation (3):
h i 1 = z i , h i l = f W l 1 h i l 1 + b l 1 , 2 l L , h i L + 1 = P y i , 1 = 1 | x i ; θ P y i , K = 1 | x i ; θ ,
where the first layer is the input layer, and the input is also the output of this layer. The symbol W l denotes the connection parameter between layers l and l + 1, while b l represents the bias term; both of these parameters are learned at the subsequent layer (i.e., layer l + 1). Additionally, f signifies the activation function employed in the hidden layer, which can be a sigmoid function or any other nonlinear function. To address the multi-label problem, softmax regression, trained via batch gradient descent, is utilized. The objective function of this model is described by Equation (4):
J θ = 1 n i = 1 n k = 1 K 1 { y i , k = 1 } l o g P y i , k = 1 | x i ; θ + λ 2 k = 1 K θ k 2 ,
wherer each output item P y i , k = 1 | x i ; θ = e x p θ k T x i j = 1 K e x p θ j T x i in layer L + 1 is the probability that label k 1 , K is assigned to data x i ,  λ is a tradeoff parameter, and θ k is the row vector of matrix W L (i.e., W L = θ 1 T ; θ 2 T ; . . . ; θ K T ).
By utilizing a neural network with concealed layers, the process of acquiring label-expanded data features becomes layered and structured. This hierarchical learning approach ensures that the high-level semantic features of the data are extracted efficiently and accurately. These high-level features represent a fusion of three distinct types of relationships: relationships among data features, relationships connecting labels with data features, and pure relationships between labels. This fusion of relationships results in features that are highly informative and beneficial for subsequent studies of classification models.
As the learning process continues, the augmented features fed into the network are continuously updated with the aid of learned parameters W and b . This updating process ensures that the network adapts to the evolving data distribution and captures the most relevant features for classification. Specifically, after each iteration, a feedforward pass is performed. This pass computes the output h X k L for each subset X k (where k ranges from 1 to K) using the updated parameters W and b learned during that iteration. This output represents the network’s understanding of the data based on the current set of parameters. Subsequently, the covariance matrix similarity coefficient K c o r r is recalculated. This coefficient measures the similarity between different subsets of data based on their covariance matrices. It captures the underlying relationships among the data (or labels) and is crucial for the next iteration, where it will be used to further refine the learning process. By iteratively updating the network parameters, recalculating the covariance matrix similarity coefficient, and performing feedforward passes, the neural network gradually learns to extract meaningful high-level semantic features from the label-expanded data. This process enables the network to adapt to complex datasets and produce accurate classification results.
Based on the preceding analysis, the optimization of the uCeFL can be achieved through gradient descent, and the corresponding pseudo code is outlined in Algorithm 1. Additionally, a variation called CeFL is presented, which is described thoroughly in Algorithm 2. CeFL relies solely on Equation (2) to compute the correlation matrix K c o r r of the original dataset, thus negating the need for frequent updates of the extended features. However, it must be emphasized that if K c o r r remains static, the procedural steps outlined in Algorithm 2 may not fully capture the label propagation effects and the inherent neural network benefits exhibited by uCeFL.
Algorithm 1: uCeFL
Input: Training   data   matrix   X = x 1 , , x n ,   label   matrix   Y = y 1 , , y n ,
             network   layer   number   L + 1 ,   iterative   number   T ,   convergence   error   ε ;
Output: Parameters   W 1 l L l   and   b 1 l < L l .
1. Initialize   W 1 l L l   and   b 1 l < L l with appropriate values;
2. FirstFlag = true;
3. for each label k do
4.       Construct the subset of training data associated with label k as X k = { x i | x i X , y i , k = 1 } ;
5. end for
6. for t = 1, 2, … T do
7.       for each label k do
8.             if FirstFlag = false then
9.                    Z k = X k ; K c o r r × Y k ;
10.                  h X k 1 = Z k ;
11.                  h X k l = f W l 1 h X k l 1 + b l 1 , 1 l L ;
12.           else
13.                  h X k L = X k ;
14.                 FirstFlag = false;
15.           end if
16.           Compute covariance matrix C h X k L of each h X k L ;
17.       end for
18.       Calculate the Covariance Matrix Similarity Coefficient K c o r r by Equation (2);
19.       Train the neural network using Z = X ; K c o r r × Y as input, then update W 1 l L l and b 1 l < L l ;
20.        if t > T or J ( W L ) t J ( W L ) t 1 < ε then
21.             Turn to return;
22.       end if
23.   end for
24. return Parameters W 1 l L l and b 1 l < L l .
Algorithm 2: CeFL
Input: Training data matrix X = x 1 , , x n , label matrix Y = y 1 , , y n ,
            network layer number L + 1 , iterative number T , convergence error ε ;
Output: Parameters W 1 l L l and b 1 l < L l .
1. Initialize W 1 l L l and b 1 l < L l with appropriate values;
2. for each label k do
3.       Construct the subset of training data associated with label k as X k = { x i | x i X , y i , k = 1 } ;
4.       Compute the covariance matrix C X k of matrix X k ;
5. end for
6. Calculate the Covariance Matrix Similarity Coefficient K c o r r by Equation (2);
7. for t = 1,2 , T do
8.       Train the neural network using Z = X ; K c o r r × Y as input, then update W 1 l L l and b 1 l < L l ;
9.       if t > T or J ( W L ) t J ( W L ) t 1 < ε then
10.             Turn to return;
11.     end if
12. end for
13. return Parameters W 1 l L l and b 1 l < L l .

4. Experiments

To validate the efficacy of CeFL and uCeFL, seven diverse multi-label datasets spanning domains such as music, biology, images, and text were carefully selected. Table 1 concisely outlines the essential characteristics of these datasets. Most of these datasets can be conveniently downloaded from Mulan (http://mulan.sourceforge.net/datasets.html (accessed on 15 February 2024)), with the exception of the education dataset, which is obtainable from Yahoo [7] (https://www.lamda.nju.edu.cn/code_MDDM.ashx (accessed on 20 February 2024)). Cardinality indicates the average number of labels per instance, while attributes refer to the fact that all features belong to one of two categories: nominal or numeric. To assess the performance of our methods, they were benchmarked against established approaches such as ML-KNN [41], TRAM [42], LLFS-DL [43], and LCFM [44]. Notably, the use of a label-expanded feature space as the basis for training inputs is a commonality shared by our approaches and LLFS-DL. This common ground facilitated meaningful comparisons and the evaluation of the relative strengths of each method.
ML-KNN was developed from traditional k-nearest neighbor (KNN) problem transformation methods. For each test instance, its k-nearest neighbors are found in the training set. Based on statistical information from these neighbors, such as the number belonging to the same label category, the maximum a posteriori (MAP) principle is employed to determine the label set of the test instance. Although ML-KNN does not consider label correlation, it often outperforms many multi-label approaches.
TRAM examines the intricate composition of label concepts, considering the varying degrees of typicality that individual examples exhibit with respect to specific labels. Initially, TRAM builds a similarity matrix by determining the k -nearest neighbor relationships for all samples, labeled and unlabeled, within the dataset. This step aims to capture the inherent connections and similarities among the data points. Subsequently, TRAM transforms the transductive multi-label learning challenge into an optimization problem, where the focus is on refining the definition and representation of label concept combinations for each multi-label instance through the optimization process. Leveraging the estimated label concept combination details, TRAM then utilizes the label concept combinations of labeled samples, coupled with their corresponding ground-truth labels, to solve an optimization function that admits a closed-form solution. The objective of this optimization is to derive a transformation matrix P , which will serve as a crucial tool in assigning labels to unlabeled samples in subsequent stages.
LLFS-DL utilizes the label matrix Y as additional features to augment the feature space X and supposes that each class label is only determined by a subset of relevant features from a given dataset. Unlike DBR, LLFS-DL does not directly increase the dimension of data features to learn a model. Instead, it solves a variation of the l 1 -norm regularization, incorporating both second-order and high-order label correlations in a sparse stacking manner to improve multi-label classification.
LCFM aims to address label prediction when labels are missing, a goal similar to ours. It combines global and local label features to learn correlations and restore missing labels. The approach utilizes sparse Lasso regression and l 2 regularization for featureselection, optimizing the model through binary cross-entropy loss and the APG algorithm.

4.1. Experimental Criteria

In the experiments, six performance metrics [45] were chosen to evaluate the performance of several multi-label classification approaches. Let T = { x i , y i } i = 1 P be the test dataset, f . be the trained model, and K be the number of labels. The indicators are defined as follows.
  • HammingLoss is an example-based multi-label classification evaluation criterion that indicates the proportion of misclassified instances in two scenarios: when predicted labels do not belong to the instance, and when labels belonging to the instance are not predicted,
    HammingLoss = 1 P i = 1 P y i f x i K ,
    where indicates the symmetric difference between the predicted and actual label vectors for each instance.
  • SubsetAccuracy is a measure of classification accuracy. It considers a classification to be correct when the set of predicted labels exactly matches the set of true labels, and incorrect otherwise. Specifically, SubsetAccuracy calculates the proportion of instances in the test set that are fully and accurately classified,
    SubsetAccuracy = 1 P i = 1 P I f x i = y i ,
    where I . is the indicator function. This metric provides a holistic view of the performance of a classification model with respect to predicting entire sets of labels, rather than just individual labels.
  • Accuracy is the ratio of correctly predicted labels to the total number of predicted labels:
    Accuracy = 1 P i = 1 P y i f x i y i f x i .
  • F1Measure is a comprehensive evaluation index that combines precision and recall, and is also referred to as the comprehensive classification rate. The precision rate measures how many of the predicted labels are actually correct,
    Precision = 1 P i = 1 P y i f x i f x i .
    The recall rate is the ratio of correctly predicted positive labels to the total number of actual positive labels,
    Recall = 1 P i = 1 P y i f x i y i .
    The F1Measure is
    F 1 Measure = 2 × P r e s s i o n × R e c a l l P r e s s i o n + R e c a l l .
    These evaluation metrics are example-based. We also consider label-based evaluation metrics, which are widely used in the literature. Let T P j , T N j , F P j , and F N j represent the number of true positives, true negatives, false positives, and false negatives, respectively, for label l j , j 1 , K .
    Precision j = T P j T P j + F P j ,
    Recall j = T P j T P j + F N j
  • MacroF1Measure is the arithmetic average value of the F1Measure of all labels.
    MacroF 1 Measure = 1 K j = 1 K 2 × Precision j × Recall j Precision j + Recall j .
  • MicroF1Measure sums the Precisionj and Recallj of all labels before calculating the F1measure.
    MicroF 1 Measure = 2 × j = 1 K Precision j × j = 1 K Recall j j = 1 K Precision j + j = 1 K Recall j .
    For all these metrics except HammingLoss, a larger value indicates better classifier performance.

4.2. Implementation Details

The experiment was conducted on a computer that was equipped with an Intel(R) Core(TM) i7-10510U CPU running at 1.80GHz (base clock speed) up to 2.30 GHz (turbo boost speed), and had 16.0 GB of RAM. The MATLAB software version used for the experiment was R2018a, and the experiment was performed in a Microsoft Windows 10 operating system environment.
For the four comparison methods, default settings were applied. The CeFL and uCeFL approaches encounter a similar challenge to the LLSF-DL algorithm, which is the absence of initial label assignments for the test samples, and this obstacle hinders the utilization of a neural network. To align with the input space of z i , certain existing methods were employed to provide initial predictions for the test data x t . Specifically, KNN was adopted to annotate each test instance with the label set identical to that of the nearest training instance, with k set to 1.
When calculating y i ^ = K c o r r × y i , both our algorithms need to input it into a logistic function first before expanding the attribute space. For the five nominal datasets of corel5k, enron, education, genbase, and medical, each item of y i ^ must continue to be reset to 0 or 1. The strategy is to set it to 1 when the item is greater than the average and otherwise to 0.
The neural network in this work is comprised of two hidden layers. The first layer had 300 neurons, while the second layer had 120 neurons. Both hidden layers employed a sigmoid activation function. The network underwent training via the L _ BFGS method, which was implemented with the minFunc (gradient descent toolbox). The regularization parameter λ in Equation (4) was assigned a value of 0.0002.

4.3. Performance Comparison and Discussion

In the experiment, 10-fold cross-validation was performed on each dataset to evaluate the performance of each algorithm. For each experiment, 1/4 of the dataset was randomly selected as test data, and the remaining data was used for training. To assess the effectiveness of uCeFL in scenarios with incomplete labels, the observed labels were randomly removed from the training data at a preset missing rate that ranged from 0 to 0.6, incremented in steps of 0.2. Also, to prevent empty classes or instances without positive labels, at least one instance was retained for each class label, and each instance was ensured to have at least one positive class label.
The average performance of 10 iterations was used to evaluate the performance of each algorithm. Table 2, Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8 display the experimental results (including the standard deviation) of the six evaluation indicators for each algorithm on the seven datasets. The symbol ↓ signifies that the smaller the value of an indicator, the better the performance, whereas ↑ indicates that a larger value signifies better performance. The optimal results within each row are highlighted in bold.
The experimental results indicate that the following conclusions can be drawn:
  • Overall, uCeFL was obviously competitive with the comparison approaches, especially on the emotions, genbase, medical, and corel5k datasets, where uCeFL performed best, regardless of the loss rate.
  • When the label is not missing, TRAM performs relatively well on the three scene, enron, and education databases, and its performance on the other four datasets is also quite competitive. However, it is crucial to note that in the step of deriving the transformation matrix P, there is a significant reliance on the sample ground-truth labels. If labels are missing in the dataset, the accuracy of the obtained transformation matrix P may be compromised, potentially leading to inaccuracies in the labeling of unlabeled samples. Consequently, as the rate of missing labels increases, the overall performance of the system declines accordingly.
  • ML-KNN only performed well on the scene dataset, and its performance was optimal when the missing rate was 0.2. ML-KNN utilizes the label information of the k-nearest neighbors of the test data to estimate the label set, but it does not take into account label correlation. Therefore, poor label annotation performance is possible when some class labels of the training data are missing.
  • Similar to our approaches, the label space Y is augmented in LLSF-DL with the feature space X as additional features. LLSF-DL removes unnecessary dependency relationships by identifying the sparsity coefficient of the label space Y . However, it may disregard correct indirect relationships resulting from labels that are not subject to sparsity coefficient consideration, which could be a reason why its performance is not ideal.
  • Compared to LCFM, which tackles missing labels by integrating both global and local label-specific features, our method leverages neural networks with hidden layers. This facilitates hierarchical learning of label-expanded data features and adeptly extracts high-level semantic features from the data. These features capture three distinct relationships: between data features, between labels and data features, and among labels. Experimental results confirm that these features offer significant benefits for the development of subsequent classification models.
  • At the end of each round of training, uCeFL performs a feedforward process to extract high-level features from the input data. Label correlations are then recalculated to update the input for the next round of training. Therefore, uCeFL not only takes the label correlations as prior knowledge but also continuously improves them during the learning process of the classification model. The experimental results show that uCeFL works better than CeFL, which demonstrates the effectiveness of using neural networks with hidden layers for feature learning.
To demonstrate the efficacy of extracting label correlations directly from the data, a random division of the data into testing and training sets was made, with 1/3 designated for testing and the remaining portion for training. Subsequently, two techniques were employed to compute label correlations: the cosine similarity (CeFL_CS) and the covariance matrix similarity coefficient (uCeFL). Figure 2, Figure 3, Figure 4 and Figure 5 offer a comparative analysis of these techniques across diverse datasets. Notably, the results underscore the potential of extracting label correlations from the data, particularly in scenarios where a substantial proportion of class labels for the training data are missing, resulting in impressive performance.

5. Conclusions and Future Works

In summary, the uCeFL methodology proposed in this paper has incontrovertibly established its superiority in tackling multi-label classification tasks, especially when dealing with missing labels. By directly extracting label correlations from the data instances, uCeFL avoids reliance on potentially incomplete label sets. The method first computes a refined label matrix by multiplying the incomplete label matrix with the label correlations obtained from the data matrix. This refined matrix then augments the original data features, enabling a neural network to learn correlation-enhanced representations that capture the intricate interplay between data features, labels, and their interactions. Importantly, uCeFL does not treat label correlations as static, predefined knowledge. Instead, it iteratively updates the label correlations during the training process, further optimizing its performance. This is achieved through a feedforward mechanism implemented at the end of each training iteration, where high-level features of the input data are extracted, and the label correlations are recalculated to refine the input for the next iteration. This iterative refinement allows uCeFL to adapt to the changing nature of the data and label correlations, ultimately resulting in superior classification outcomes. The experimental results across a range of datasets highlight the competitiveness of uCeFL. Notably, it outperforms other comparative approaches, especially on datasets related to emotions, enron, genbase, and medical, where it achieves optimal performance regardless of the label missing rate.
Future research will consider the following aspects:
  • This paper employed a covariance matrix to represent the dataset comprising category labels and derived label correlations through the calculation of a covariance matrix similarity coefficient. However, computing these correlations becomes challenging when dealing with a substantial data volume. The issue of efficiently acquiring label correlations in a big-data environment demands further exploration.
  • Moreover, the paper only concentrated on one-to-one label correlations. To accurately capture the actual label correlations, further research is needed, encompassing investigations into both local and global correlations.
  • The presence of multi-label data introduces the curse of dimensionality. Future aims can explore the relationship between labels and features to identify a more discriminative subspace.

Author Contributions

Z.Z.: Conceptualization, Methodology, Software, Writing—original draft. X.Z.: Data curation, Writing—review & editing. Y.Y.: Software. X.D.: Validation. S.L.: Writing—review & editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the Qingmiao Plan of Chengdu Technological University (No. QM2023054) and the Doctoral Fundation Program of Chengdu Technological University (No. 2019RC019).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

No author associated with this paper has disclosed any potential or pertinent conflicts of interest that may be perceived as having a bearing on the work presented.

References

  1. Shu, X.; Qiu, J. Speed up kernel dependence maximization for multi-label feature extraction. J. Vis. Commun. Image Represent. 2017, 49, 361–370. [Google Scholar] [CrossRef]
  2. Du, J.; Chen, Q.; Peng, Y.; Xiang, Y.; Tao, C.; Lu, Z. Ml-net: Multi-label classification of biomedical texts with deep neural networks. J. Am. Med. Inform. Assoc. 2019, 26, 1279–1285. [Google Scholar] [CrossRef]
  3. Wang, T.; Liu, L.; Liu, N.; Zhang, H.; Zhang, L.; Feng, S. A multi-label text classification method via dynamic semantic representation model and deep neural network. Appl. Intell. 2020, 50, 2339–2351. [Google Scholar] [CrossRef]
  4. Tang, R.; Yu, Z.; Ma, Y.; Wu, Y.; Chen, Y.-P.P.; Wong, L.; Li, J. Genetic source completeness of hiv-1 circulating recombinant forms (crfs) predicted by multi-label learning. Bioinformatics 2020, 37, 750–758. [Google Scholar] [CrossRef] [PubMed]
  5. Zhou, Z.; Waqas, J. Intrinsic structure based feature transform for image classification. J. Vis. Commun. Image Represent. 2016, 38, 735–744. [Google Scholar] [CrossRef]
  6. Gou, J.; Song, J.; Ou, W.; Zeng, S.; Yuan, Y.; Du, L. Representation-based classification methods with enhanced linear reconstruction measures for face recognition. Comput. Electr. Eng. 2019, 79, 106451. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Zhou, Z.-H. Multilabel dimensionality reduction via dependence maximization. ACM Trans. Knowl. Discov. Data 2010, 4, 1–21. [Google Scholar] [CrossRef]
  8. Li, Y.-K.; Zhang, M.-L. Enhancing binary relevance for multi-label learning with controlled label correlations exploitation. In Pacific Rim International Conference on Artificial Intelligence; Springer: Cham, Switzerland, 2014; pp. 91–103. [Google Scholar] [CrossRef]
  9. Xu, L.; Wang, Z.; Shen, Z.; Wang, Y.; Chen, E. Learning low-rank label correlations for multi-label classification with missing labels. In Proceedings of the 2014 IEEE International Conference on data Mining, Shenzhen, China, 14–17 December 2014; IEEE: New York, NY, USA, 2014; pp. 1067–1072. [Google Scholar] [CrossRef]
  10. Tsoumakas, G.; Katakis, I. Multi-label classification: An overview. Int. J. Data Warehous. Min. 2007, 3, 1–13. [Google Scholar] [CrossRef]
  11. Read, J.; Pfahringer, B.; Holmes, G. Multi-label classification using ensembles of pruned sets. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; IEEE: New York, NY, USA, 2008; pp. 995–1000. [Google Scholar] [CrossRef]
  12. Tsoumakas, G.; Vlahavas, I. Random k-labelsets: An ensemble method for multilabel classification. In European Conference on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2007; pp. 406–417. [Google Scholar] [CrossRef]
  13. Wang, C.; Yan, S.; Zhang, L.; Zhang, H.-J. Multi-label sparse coding for automatic image annotation. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: New York, NY, USA, 2009; pp. 1643–1650. [Google Scholar] [CrossRef]
  14. Wang, H.; Huang, H.; Ding, C. Image annotation using multi-label correlated green’s function. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 27 September–4 October 2009; IEEE: New York, NY, USA, 2009; pp. 2029–2034. [Google Scholar] [CrossRef]
  15. Huang, S.-J.; Zhou, Z.-H. Multi-label learning by exploiting label correlations locally. In Proceedings of the AAAI Conference on Artificial Intelligence, Toronto, ON, USA, 22–26 July 2012; Volume 26. [Google Scholar] [CrossRef]
  16. Gu, Q.; Li, Z.; Han, J. Correlated multi-label feature selection. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management, Glasgow UK, 24–28 October 2011; pp. 1087–1096. [Google Scholar] [CrossRef]
  17. Ji, S.; Tang, L.; Yu, S.; Ye, J. Extracting shared subspace for multi-label classification. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 381–389. [Google Scholar] [CrossRef]
  18. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains for multi-label classification. Mach. Learn. 2011, 85, 333–359. [Google Scholar] [CrossRef]
  19. Montañes, E.; Senge, R.; Barranquero, J.; Quevedo, J.R.; del Coz, J.J.; Hüllermeier, E. Dependent binary relevance models for multi-label classification. Pattern Recognit. 2014, 47, 1494–1508. [Google Scholar] [CrossRef]
  20. Huang, J.; Qin, F.; Zheng, X.; Cheng, Z.; Yuan, Z.; Zhang, W.; Huang, Q. Improving multi-label classification with missing labels by learning label-specific features. Inf. Sci. 2019, 492, 124–146. [Google Scholar] [CrossRef]
  21. Zhu, P.; Xu, Q.; Hu, Q.; Zhang, C.; Zhao, H. Multi-label featureselection with missing labels. Pattern Recognit. 2018, 74, 488–502. [Google Scholar] [CrossRef]
  22. Rauber, T.W.; Mello, L.H.; Rocha, V.F.; Luchi, D.; Varejão, F.M. Recursive dependent binary relevance model for multi-label classification. In Advances in Artificial Intelligence–IBERAMIA 2014: 14th Ibero-American Conference on AI, Santiago de Chile, Chile, 24–27 November 2014; Proceedings 14; Springer: Cham, Switzerland, 2014; pp. 206–217. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Li, Y.; Cai, Z. Correlation-based pruning of dependent binary relevance models for multi-label classification. In Proceedings of the 2015 IEEE 14th International Conference on Cognitive Informatics & Cognitive Computing (ICCICC), Beijing, China, 6–8 July 2015; IEEE: New York, NY, USA, 2015. [Google Scholar] [CrossRef]
  24. Rauber, T.W.; Rocha, V.F.; Mello, L.H.S.; Varejao, F.M. Decision template multi-label classification based on recursive dependent binary relevance. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; IEEE: New York, NY, USA, 2016; pp. 2402–2408. [Google Scholar] [CrossRef]
  25. Zhang, M.L.; Wu, L. Lift: Multi-label learning with label-specific features. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 107–120. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, J.-J.; Fang, M.; Li, X. Multi-label learning with discriminative features for each label. Neurocomputing 2015, 154, 305–316. [Google Scholar] [CrossRef]
  27. Xu, S.; Yang, X.; Yu, H.; Yu, D.-J.; Yang, J.; Tsang, E.C. Multi-label learning with label-specific feature reduction. Knowl. Based Syst. 2016, 104, 52–61. [Google Scholar] [CrossRef]
  28. Lee, J.; Kim, D.-W. Efficient multi-label feature selection using entropy-based label selection. Entropy 2016, 18, 405. [Google Scholar] [CrossRef]
  29. Mencía, E.L.; Janssen, F. Learning rules for multi-label classification: A stacking and a separate-and-conquer approach. Mach. Learn. 2016, 105, 77–126. [Google Scholar] [CrossRef]
  30. Li, Q.; Peng, X.; Qiao, Y.; Peng, Q. Learning label correlations for multi-label image recognition with graph networks. Pattern Recognit. Lett. 2020, 138, 378–384. [Google Scholar] [CrossRef]
  31. RWang, R.; Ye, S.; Li, K.; Kwong, S. Bayesian network based label correlation analysis for multi-label classifier chain. Inf. Sci. 2020, 554, 256–275. [Google Scholar] [CrossRef]
  32. Zhan, W.; Zhang, M.-L. Multi-label learning with label-specific features via clustering ensemble. In Proceedings of the 2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Tokyo, Japan, 19–21 October 2017; IEEE: New York, NY, USA, 2017; pp. 129–136. [Google Scholar] [CrossRef]
  33. Zhang, J.; Li, C.; Cao, D.; Lin, Y.; Su, S.; Dai, L.; Li, S. Multi-label learning with label-specific features by resolving label correlations. Knowl. Based Syst. 2018, 159, 148–157. [Google Scholar] [CrossRef]
  34. Liu, H.; Ma, Z.; Han, J.; Chen, Z.; Zheng, Z. Regularized partial least squares for multi-label learning. Int. J. Mach. Learn. Cybern. 2018, 9, 335–346. [Google Scholar] [CrossRef]
  35. Ma, Q.; Yuan, C.; Zhou, W.; Hu, S. Label-specific dual graph neural network for multi-label text classification. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Bangkok, Thailand, 1–6 August 2021; Volume 1, pp. 3855–3864. [Google Scholar] [CrossRef]
  36. Fu, B.; Xu, G.; Wang, Z.; Cao, L. Leveraging supervised label dependency propagation for multi-label learning. In Proceedings of the 2013 IEEE 13th International Conference on Data Mining, Dallas, TX, USA, 7–10 December 2013; IEEE: New York, NY, USA, 2013; pp. 1061–1066. [Google Scholar] [CrossRef]
  37. Wang, H.; Ding, C.; Huang, H. Multi-label linear discriminant analysis. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2010; pp. 126–139. [Google Scholar] [CrossRef]
  38. Cai, X.; Nie, F.; Cai, W.; Huang, H. New graph structured sparsity model for multi-label image annotations. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 801–808. [Google Scholar] [CrossRef]
  39. Herdin, M.; Czink, N.; Ozcelik, H.; Bonek, E. Correlation matrix distance, a meaningful measure for evaluation of non-stationary mimo channels. In Proceedings of the 2005 IEEE 61st Vehicular Technology Conference, Stockholm, Sweden, 30 May–1 June 2005; IEEE: New York, NY, USA, 2005; Volume 1, pp. 136–140. [Google Scholar] [CrossRef]
  40. Arsigny, V.; Fillard, P.; Pennec, X.; Ayache, N. Geometric means in a novel vector space structure on symmetric positive-definite matrices. SIAM J. Matrix Anal. Appl. 2007, 29, 328–347. [Google Scholar] [CrossRef]
  41. Zhang, M.-L.; Zhou, Z.-H. Ml-knn: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  42. Kong, X.; Ng, M.K.; Zhou, Z.-H. Transductive multilabel learning via label set propagation. IEEE Trans. Knowl. Data Eng. 2011, 25, 704–719. [Google Scholar] [CrossRef]
  43. Huang, J.; Li, G.; Huang, Q.; Wu, X. Learning label-specific features and class-dependent labels for multi-label classification. IEEE Trans. Knowl. Data Eng. 2016, 28, 3309–3323. [Google Scholar] [CrossRef]
  44. Yu, Y.; Zhou, Z.; Zheng, X.; Gou, J.; Ou, W.; Yuan, F. Enhancing label correlations in multi-label classification through global-local label specific feature learning to fill missing labels. Comput. Electr. Eng. 2024, 113, 109037. [Google Scholar] [CrossRef]
  45. Zhang, M.-L.; Zhou, Z.-H. A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 2013, 26, 1819–1837. [Google Scholar] [CrossRef]
Figure 1. The proposed architecture of uCeFL.
Figure 1. The proposed architecture of uCeFL.
Mathematics 12 02131 g001
Figure 2. Performance comparison between two similarity methods on the emotions and scene datasets.
Figure 2. Performance comparison between two similarity methods on the emotions and scene datasets.
Mathematics 12 02131 g002
Figure 3. Performance comparison between two similarity methods on the corel5k and enron datasets.
Figure 3. Performance comparison between two similarity methods on the corel5k and enron datasets.
Mathematics 12 02131 g003
Figure 4. Performance comparison between two similarity methods on the education and genbase datasets.
Figure 4. Performance comparison between two similarity methods on the education and genbase datasets.
Mathematics 12 02131 g004
Figure 5. Performance comparison between two similarity methods on the medical dataset.
Figure 5. Performance comparison between two similarity methods on the medical dataset.
Mathematics 12 02131 g005
Table 1. Characteristics of the datasets used in the experiments.
Table 1. Characteristics of the datasets used in the experiments.
Attributes
DatasetInstancesNominalNumericLabelsCardinality
emotions59307261.869
scene2407029461.074
corel5k500049903743.522
enron170210010533.378
education50005500331.461
genbase66211860271.252
medical97814490451.245
Table 2. Comparison results (mean ± std) of six approaches over the emotions dataset.
Table 2. Comparison results (mean ± std) of six approaches over the emotions dataset.
Emotions
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.2723 ± 0.00490.2986 ± 0.00350.3139 ± 0.00330.4083 ± 0.01010.2617 ± 0.00560.2584 ± 0.0045
SubsetAccuracy (↑)0.1105 ± 0.01980.1467 ± 0.07360.0000 ± 0.00000.0051 ± 0.00580.1847 ± 0.03350.1912 ± 0.0369
Accuracy (↑)0.3016 ± 0.01190.4152 ± 0.01320.0000 ± 0.00000.2735 ± 0.00830.4378 ± 0.01160.4398 ± 0.0178
F1measure (↑)0.3667 ± 0.01120.5112 ± 0.00910.0000 ± 0.00000.3970 ± 0.01190.5231 ± 0.01050.5227 ± 0.0138
MacroF1measure (↑)0.3511 ± 0.00100.4963 ± 0.00580.0000 ± 0.00000.2161 ± 0.00890.5403 ± 0.00780.5599 ± 0.0186
MicroF1Measure (↑)0.4343 ± 0.00390.5265 ± 0.00990.0000 ± 0.00000.3975 ± 0.00820.5575 ± 0.00130.5667 ± 0.0089
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.2835 ± 0.00770.4048 ± 0.00770.3116 ± 0.00110.4101 ± 0.03110.2607 ± 0.00680.2555 ± 0.0107
SubsetAccuracy (↑)0.0838 ± 0.02390.0909 ± 0.00280.0000 ± 0.00000.0051 ± 0.00430.1878 ± 0.01360.1947 ± 0.0037
Accuracy (↑)0.1986 ± 0.00850.2345 ± 0.00880.0000 ± 0.00000.2741 ± 0.01510.3996 ± 0.00680.4116 ± 0.0213
F1measure (↑)0.2415 ± 0.00320.2901 ± 0.00950.0000 ± 0.00000.3966 ± 0.01950.4739 ± 0.01010.4875 ± 0.0318
MacroF1measure (↑)0.2088 ± 0.00670.1646 ± 0.00810.0000 ± 0.00000.2005 ± 0.02410.5218 ± 0.01910.5363 ± 0.0259
MicroF1Measure (↑)0.2910 ± 0.00110.2887 ± 0.01430.0000 ± 0.00000.3952 ± 0.01510.5301 ± 0.01660.5443 ± 0.0211
Missing rate: 0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.2946 ± 0.00380.3897 ± 0.01260.3053 ± 0.00770.3958 ± 0.01950.2684 ± 0.00140.2613 ± 0.0075
SubsetAccuracy (↑)0.0471 ± 0.01310.0502 ± 0.01510.0000 ± 0.00000.0051 ± 0.00520.1308 ± 0.00540.1443 ± 0.0111
Accuracy (↑)0.0944 ± 0.01830.1561 ± 0.02110.0000 ± 0.00000.2798 ± 0.01360.3426 ± 0.00890.3486 ± 0.0304
F1measure (↑)0.1123 ± 0.01870.1932 ± 0.02350.0000 ± 0.00000.4031 ± 0.02120.4158 ± 0.01220.4186 ± 0.0241
MacroF1measure (↑)0.1335 ± 0.01900.0923 ± 0.01770.0000 ± 0.00000.2015 ± 0.02620.4433 ± 0.02120.4575 ± 0.0011
MicroF1Measure (↑)0.1512 ± 0.01210.1943 ± 0.02220.0000 ± 0.00000.4018 ± 0.02040.4687 ± 0.01380.4819 ± 0.0112
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.3093 ± 0.00280.3836 ± 0.00180.3105 ± 0.00090.4055 ± 0.00540.3045 ± 0.02140.2986 ± 0.0178
SubsetAccuracy (↑)0.0067 ± 0.00670.0536 ± 0.00710.0000 ± 0.00000.0153 ± 0.01520.0804 ± 0.02210.0873 ± 0.0204
Accuracy (↑)0.0229 ± 0.02290.1425 ± 0.00690.0000 ± 0.00000.2933 ± 0.02010.2458 ± 0.04280.2621 ± 0.0379
F1measure (↑)0.0285 ± 0.02850.1738 ± 0.00660.0000 ± 0.00000.4151 ± 0.02000.3050 ± 0.04800.3254 ± 0.0424
MacroF1measure (↑)0.0406 ± 0.04060.0685 ± 0.00090.0000 ± 0.00000.2585 ± 0.03360.3330 ± 0.03120.3457 ± 0.0355
MicroF1Measure (↑)0.0403 ± 0.04030.1734 ± 0.00740.0000 ± 0.00000.4187 ± 0.01780.3539 ± 0.045210.3667 ± 0.0454
Table 3. Comparison results (mean ± std) of six approaches over the scene dataset.
Table 3. Comparison results (mean ± std) of six approaches over the scene dataset.
Scene
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0868 ± 0.00210.0946 ± 0.00100.1774 ± 0.00320.4016 ± 0.025740.1013 ± 0.00230.1012 ± 0.0018
SubsetAccuracy (↑)0.6295 ± 0.00760.6918 ± 0.00590.0000 ± 0.00000.0102 ± 0.01310.6685 ± 0.00810.6654 ± 0.0057
Accuracy (↑)0.6743 ± 0.00550.7201 ± 0.00150.0000 ± 0.00000.2773 ± 0.01440.7003 ± 0.00850.7003 ± 0.0077
F1measure (↑)0.6896 ± 0.00470.7295 ± 0.00030.0000 ± 0.00000.3985 ± 0.01220.7113 ± 0.00810.7123 ± 0.0081
MacroF1measure (↑)0.7408 ± 0.00410.7325 ± 0.00210.0000 ± 0.00000.2116 ± 0.00770.7121 ± 0.00880.7158 ± 0.0078
MicroF1Measure (↑)0.7358 ± 0.00480.7256 ± 0.00210.0000 ± 0.00000.3976 ± 0.01380.7071 ± 0.00710.7088 ± 0.0068
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0891 ± 0.00490.2835 ± 0.00680.1803 ± 0.00180.4215 ± 0.03470.1207 ± 0.00750.1194 ± 0.0031
SubsetAccuracy (↑)0.6195 ± 0.03150.1596 ± 0.02480.0000 ± 0.00000.0016 ± 0.00690.5634 ± 0.02670.5665 ± 0.0123
Accuracy (↑)0.6718 ± 0.03340.1741 ± 0.02220.0000 ± 0.00000.2768 ± 0.00640.5945 ± 0.02650.5977 ± 0.0121
F1measure (↑)0.6897 ± 0.03190.1785 ± 0.02240.0000 ± 0.00000.4002 ± 0.01210.6046 ± 0.02610.6081 ± 0.0123
MacroF1measure (↑)0.7410 ± 0.01170.0591 ± 0.00130.0000 ± 0.00000.2151 ± 0.02750.6437 ± 0.02220.6483 ± 0.0075
MicroF1Measure (↑)0.7314 ± 0.02030.1813 ± 0.01890.0000 ± 0.00000.3992 ± 0.01350.6334 ± 0.02380.6373 ± 0.0124
Missing rate: 0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.1199 ± 0.00170.2712 ± 0.00210.1787 ± 0.00340.4042 ± 0.02290.1390 ± 0.00640.1338 ± 0.0054
SubsetAccuracy (↑)0.3928 ± 0.01730.1237 ± 0.00150.0000 ± 0.00000.0052 ± 0.00120.4552 ± 0.02050.4658 ± 0.0015
Accuracy (↑)0.4186 ± 0.01570.1317 ± 0.00660.0000 ± 0.00000.2774 ± 0.00420.4823 ± 0.02670.4948 ± 0.0049
F1measure (↑)0.4273 ± 0.01870.1344 ± 0.00230.0000 ± 0.00000.4011 ± 0.01110.4918 ± 0.02260.5043 ± 0.0101
MacroF1measure (↑)0.5433 ± 0.00640.0458 ± 0.00280.0000 ± 0.00000.2047 ± 0.02330.5567 ± 0.02350.5743 ± 0.0088
MicroF1Measure (↑)0.5542 ± 0.01010.1463 ± 0.00590.0000 ± 0.00000.4015 ± 0.00650.5502 ± 0.02590.5662 ± 0.0056
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.1692 ± 0.00280.2137 ± 0.00170.1797 ± 0.00140.4002 ± 0.02010.1511 ± 0.00580.1493 ± 0.0066
SubsetAccuracy (↑)0.0682 ± 0.01060.0425 ± 0.00550.0000 ± 0.00000.0102 ± 0.00650.4193 ± 0.01750.4229 ± 0.0126
Accuracy (↑)0.0715 ± 0.01010.0458 ± 0.00820.0000 ± 0.00000.2846 ± 0.02290.4498 ± 0.01190.4528 ± 0.0101
F1measure (↑)0.0726 ± 0.00780.0468 ± 0.00870.0000 ± 0.00000.4048 ± 0.03510.4602 ± 0.01010.4632 ± 0.0081
MacroF1measure (↑)0.1263 ± 0.01490.0337 ± 0.00380.0000 ± 0.00000.2552 ± 0.02160.5187 ± 0.01550.5233 ± 0.0141
MicroF1Measure (↑)0.1283 ± 0.01670.0710 ± 0.01320.0000 ± 0.00000.4125 ± 0.03100.5158 ± 0.01360.5203 ± 0.0146
Table 4. Comparison results (mean ± std) of six approaches over the corel5k dataset.
Table 4. Comparison results (mean ± std) of six approaches over the corel5k dataset.
Corel5k
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0093 ± 0.00210.0153 ± 0.00010.0097 ± 0.00110.3380 ± 0.00130.0149 ± 0.00010.0149 ± 0.0011
SubsetAccuracy (↑)0.0041 ± 0.00150.0071 ± 0.00050.0061 ± 0.00250.0000 ± 0.00000.0105 ± 0.00440.0113 ± 0.0001
Accuracy (↑)0.0134 ± 0.00140.1137 ± 0.00190.0161 ± 0.00150.0224 ± 0.00010.1284 ± 0.00230.1313 ± 0.0049
F1measure (↑)0.0172 ± 0.00090.1701 ± 0.00280.0191 ± 0.00050.0436 ± 0.00030.1885 ± 0.00430.1926 ± 0.0027
MacroF1measure (↑)0.0101 ± 0.00530.0281 ± 0.00220.0066 ± 0.00110.0219 ± 0.00010.0476 ± 0.00290.0522 ± 0.0017
MicroF1Measure (↑)0.0292 ± 0.00170.1706 ± 0.00150.0334 ± 0.00160.0434 ± 0.00030.1906 ± 0.00330.1954 ± 0.0014
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0093 ± 0.00110.0154 ± 0.00120.0096 ± 0.00110.3357 ± 0.00090.0133 ± 0.00210.0134 ± 0.0004
SubsetAccuracy (↑)0.0021 ± 0.00020.0004 ± 0.00340.0027 ± 0.00520.0000 ± 0.00000.0055 ± 0.00870.0063 ± 0.0056
Accuracy (↑)0.0128 ± 0.00330.0405 ± 0.00550.0098 ± 0.00390.0220 ± 0.00030.0963 ± 0.00890.0933 ± 0.0075
F1measure (↑)0.0174 ± 0.00550.0656 ± 0.00880.0119 ± 0.00110.0428 ± 0.00060.1427 ± 0.00550.1384 ± 0.0062
MacroF1measure (↑)0.0087 ± 0.00280.0026 ± 0.00230.0042 ± 0.00540.0212 ± 0.00020.0387 ± 0.00000.0372 ± 0.0022
MicroF1Measure (↑)0.0283 ± 0.00310.0696 ± 0.00820.0227 ± 0.00730.0426 ± 0.00060.1558 ± 0.00990.1515 ± 0.0054
Missing rate: 0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0096 ± 0.00120.0143 ± 0.00110.0092 ± 0.00010.3247 ± 0.00210.0143 ± 0.00190.0142 ± 0.0019
SubsetAccuracy (↑)0.0004 ± 0.00110.0000 ± 0.00000.0005 ± 0.00120.0000 ± 0.00000.0043 ± 0.00840.0033 ± 0.0021
Accuracy (↑)0.0112 ± 0.00450.0242 ± 0.00150.0013 ± 0.00110.0226 ± 0.00020.0824 ± 0.00210.0801 ± 0.0056
F1measure (↑)0.0168 ± 0.00650.0389 ± 0.00310.0012 ± 0.00880.0441 ± 0.00040.1226 ± 0.00120.1197 ± 0.0063
MacroF1measure (↑)0.0013 ± 0.00010.0013 ± 0.00120.0003 ± 0.00940.0220 ± 0.00040.0331 ± 0.00210.0327 ± 0.0011
MicroF1Measure (↑)0.0202 ± 0.00890.0436 ± 0.00210.0023 ± 0.00120.0437 ± 0.00040.1296 ± 0.00810.1283 ± 0.0067
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0097 ± 0.00520.0127 ± 0.00120.0091 ± 0.00910.2996 ± 0.00250.0143 ± 0.00040.0143 ± 0.0015
SubsetAccuracy (↑)0.0005 ± 0.00010.0000 ± 0.00000.0000 ± 0.00000.0000 ± 0.00000.0012 ± 0.00140.0025 ± 0.0007
Accuracy (↑)0.0087 ± 0.00590.0072 ± 0.00520.0000 ± 0.00000.0239 ± 0.00030.0634 ± 0.00110.0646 ± 0.0027
F1measure (↑)0.0132 ± 0.01170.0118 ± 0.00870.0000 ± 0.00000.0466 ± 0.00060.0968 ± 0.00550.0981 ± 0.0067
MacroF1measure (↑)0.0012 ± 0.00030.0006 ± 0.00130.0000 ± 0.00000.0227 ± 0.00030.0235 ± 0.00220.0252 ± 0.0001
MicroF1Measure (↑)0.0155 ± 0.01350.0138 ± 0.00210.0000 ± 0.00000.0461 ± 0.00050.1022 ± 0.00170.1035 ± 0.0012
Table 5. Comparison results (mean ± std) of six approaches over the enron dataset.
Table 5. Comparison results (mean ± std) of six approaches over the enron dataset.
Enron
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0523 ± 0.00060.0509 ± 0.00050.0583 ± 0.00060.2778 ± 0.00210.0573 ± 0.00160.0563 ± 0.0014
SubsetAccuracy (↑)0.0353 ± 0.00950.1213 ± 0.00590.0163 ± 0.00460.0000 ± 0.00000.0903 ± 0.02000.1059 ± 0.0071
Accuracy (↑)0.2883 ± 0.03400.4723 ± 0.00050.3470 ± 0.00300.1743 ± 0.00150.3623 ± 0.02350.3819 ± 0.0077
F1measure (↑)0.3813 ± 0.03560.5923 ± 0.00080.4723 ± 0.00690.2880 ± 0.00230.4674 ± 0.02260.4903 ± 0.0100
MacroF1measure (↑)0.0850 ± 0.00700.1873 ± 0.01500.0638 ± 0.00060.1418 ± 0.00500.1754 ± 0.00210.1878 ± 0.0107
MicroF1Measure (↑)0.4548 ± 0.03560.5890 ± 0.00020.5038 ± 0.00290.2883 ± 0.00250.4800 ± 0.01530.4898 ± 0.0101
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0568 ± 0.00060.1083 ± 0.00130.0628 ± 0.00210.2754 ± 0.00460.0553 ± 0.00130.0553 ± 0.0011
SubsetAccuracy (↑)0.0213 ± 0.00700.0073 ± 0.00250.0273 ± 0.00350.0000 ± 0.00000.0973 ± 0.01290.1024 ± 0.0107
Accuracy (↑)0.1953 ± 0.03150.0220 ± 0.00030.0673 ± 0.01060.1743 ± 0.00650.3409 ± 0.00600.3410 ± 0.0030
F1measure (↑)0.2703 ± 0.03690.0298 ± 0.00150.0843 ± 0.01650.2878 ± 0.00900.4384 ± 0.00530.4394 ± 0.0017
MacroF1measure (↑)0.0583 ± 0.00860.0179 ± 0.00250.0203 ± 0.00490.1380 ± 0.00400.1839 ± 0.02700.1724 ± 0.0307
MicroF1Measure (↑)0.3263 ± 0.03950.0347± 0.00210.0903 ± 0.02530.2878 ± 0.00850.4512 ± 0.00500.4513 ± 0.0034
Missing rate:0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0623 ± 0.00060.0973 ± 0.00030.0633 ± 0.00110.2743 ± 0.00790.0603 ± 0.00150.0603 ± 0.0014
SubsetAccuracy (↑)0.0163 ± 0.00950.0024 ± 0.00250.0294 ± 0.00350.0000 ± 0.00000.0729 ± 0.00950.0813 ± 0.0177
Accuracy (↑)0.0633 ± 0.01360.0123 ± 0.00130.0303 ± 0.00330.1754 ± 0.00750.2579 ± 0.01850.2593 ± 0.0240
F1measure (↑)0.0864 ± 0.01600.0193 ± 0.00050.0313 ± 0.00300.2888 ± 0.01050.3383 ± 0.02330.3384 ± 0.0267
MacroF1measure (↑)0.0174 ± 0.00030.0030 ± 0.00000.0039 ± 0.00000.1360 ± 0.00460.1628 ± 0.00000.1528 ± 0.0095
MicroF1Measure (↑)0.1070 ± 0.01030.0210 ± 0.00300.0193 ± 0.00210.2893 ± 0.01000.3473 ± 0.01230.3498 ± 0.0157
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0628 ± 0.00110.0848 ± 0.00050.0633 ± 0.00060.2773 ± 0.00690.0604 ± 0.00070.0603 ± 0.0011
SubsetAccuracy (↑)0.0033 ± 0.00350.0033 ± 0.00130.0200 ± 0.00830.0000 ± 0.00000.0648 ± 0.00590.0673 ± 0.0059
Accuracy (↑)0.0189 ± 0.00050.0099 ± 0.00110.0223 ± 0.00900.1728 ± 0.00750.2213 ± 0.00050.2202 ± 0.0048
F1measure (↑)0.0273 ± 0.00150.0133 ± 0.00060.0233 ± 0.00930.2853 ± 0.01050.2918 ± 0.00210.2924 ± 0.0039
MacroF1measure (↑)0.0084 ± 0.00260.0030 ± 0.00110.0033 ± 0.00060.1309 ± 0.00790.0998 ± 0.00050.1411 ± 0.0018
MicroF1Measure (↑)0.0348 ± 0.00650.0129 ± 0.00050.0153 ± 0.00550.2860 ± 0.01000.2919 ± 0.00450.2923 ± 0.0015
Table 6. Comparison results (mean ± std) of six approaches over the education dataset.
Table 6. Comparison results (mean ± std) of six approaches over the education dataset.
Education
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0378 ± 0.00050.0463 ± 0.00050.0398 ± 0.00000.2813 ± 0.00190.0563 ± 0.00430.0563 ± 0.0049
SubsetAccuracy (↑)0.1968 ± 0.02330.2944 ± 0.00320.1420 ± 0.00450.0000 ± 0.00000.2053 ± 0.03850.2073 ± 0.0497
Accuracy (↑)0.2318 ± 0.02500.4134 ± 0.00180.1704 ± 0.00350.1323 ± 0.00030.2878 ± 0.04390.2873 ± 0.0577
F1measure (↑)0.2443 ± 0.02550.4573 ± 0.00310.1793 ± 0.00300.2273 ± 0.00030.3193 ± 0.04560.3174 ± 0.0601
MacroF1measure (↑)0.1133 ± 0.01290.2173 ± 0.03980.0753 ± 0.00050.1129 ± 0.00650.1433 ± 0.02460.1333 ± 0.0241
MicroF1Measure (↑)0.3364 ± 0.02730.4498 ± 0.00690.2804 ± 0.00110.2304 ± 0.00030.3198 ± 0.04530.3200 ± 0.0544
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0399 ± 0.00210.0683 ± 0.00110.0413 ± 0.00110.2779 ± 0.00060.0508 ± 0.00060.0500 ± 0.0002
SubsetAccuracy (↑)0.1673 ± 0.00660.0780 ± 0.00350.0973 ± 0.00950.0000 ± 0.00000.2353 ± 0.01650.2473 ± 0.0024
Accuracy (↑)0.2069 ± 0.00250.0999 ± 0.00160.1163 ± 0.00600.1328 ± 0.00150.3104 ± 0.01190.3219 ± 0.0014
F1measure (↑)0.2213 ± 0.00130.1083 ± 0.00060.1233 ± 0.00500.2283 ± 0.00250.3390 ± 0.00960.3493 ± 0.0030
MacroF1measure (↑)0.0993 ± 0.00130.0289 ± 0.01550.0350 ± 0.00210.1083 ± 0.00560.1483 ± 0.00480.1404 ± 0.0021
MicroF1Measure (↑)0.2903 ± 0.00560.1048 ± 0.00030.1713 ± 0.00250.2313 ± 0.00250.3353 ± 0.00660.3443 ± 0.0045
Missing rate: 0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0403 ± 0.00050.0668 ± 0.00050.0424 ± 0.00000.2613 ± 0.00250.0484 ± 0.00060.0483 ± 0.0007
SubsetAccuracy (↑)0.1493 ± 0.02750.0753 ± 0.00660.0744 ± 0.00330.0000 ± 0.00000.2493 ± 0.01200.2484 ± 0.0077
Accuracy (↑)0.1809 ± 0.03050.0973 ± 0.00900.0840 ± 0.00590.1370 ± 0.00160.3133 ± 0.00720.3123 ± 0.0057
F1measure (↑)0.1928 ± 0.03160.1054 ± 0.00960.0873 ± 0.00700.2348 ± 0.00250.3363 ± 0.00560.3373 ± 0.0050
MacroF1measure (↑)0.0883 ± 0.00300.0258 ± 0.01600.0219 ± 0.00160.1090 ± 0.00650.1373 ± 0.00230.1433 ± 0.0087
MicroF1Measure (↑)0.2494 ± 0.03190.1028 ± 0.00860.1213 ± 0.01210.2368 ± 0.00230.3338 ± 0.00330.3343 ± 0.0041
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0423 ± 0.00020.0608 ± 0.00050.0433 ± 0.00050.2663 ± 0.00360.0514 ± 0.00210.0513 ± 0.0003
SubsetAccuracy (↑)0.0744 ± 0.00860.0553 ± 0.00130.0524 ± 0.00600.0000 ± 0.00000.1868 ± 0.00200.1933 ± 0.0037
Accuracy (↑)0.0933 ± 0.00760.0708 ± 0.00110.0558 ± 0.00660.1333 ± 0.00130.2528 ± 0.00230.2553 ± 0.0001
F1measure (↑)0.1008 ± 0.00730.0763 ± 0.00050.0573 ± 0.00730.2294 ± 0.00160.2783 ± 0.00400.2794 ± 0.0014
MacroF1measure (↑)0.0388 ± 0.00830.0063 ± 0.00000.0159 ± 0.00190.0989 ± 0.00150.1019 ± 0.01060.1094 ± 0.0050
MicroF1Measure (↑)0.1453 ± 0.00700.0834 ± 0.00000.0780 ± 0.01060.2309 ± 0.00230.2938 ± 0.01250.2929 ± 0.0057
Table 7. Comparison results (mean ± std) of six approaches over the genbase dataset.
Table 7. Comparison results (mean ± std) of six approaches over the genbase dataset.
Genbase
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0043 ± 0.00210.0008 ± 0.00060.0110 ± 0.00230.1723 ± 0.01000.0003 ± 0.00070.0008 ± 0.0008
SubsetAccuracy (↑)0.9158 ± 0.03000.9833 ± 0.01500.8103 ± 0.04530.0453 ± 0.07830.9847± 0.00000.9847± 0.0151
Accuracy (↑)0.9478 ± 0.01950.9820 ± 0.00800.8839 ± 0.03550.2613 ± 0.05930.9913 ± 0.00650.9920 ± 0.0080
F1measure (↑)0.9569 ± 0.01700.9938 ± 0.00630.9093 ± 0.03360.3838 ± 0.04830.9943 ± 0.00430.9950 ± 0.0062
MacroF1measure (↑)0.5610 ± 0.00650.7720 ± 0.00560.4388 ± 0.00150.4593 ± 0.02200.7723 ± 0.00570.7673 ± 0.0107
MicroF1Measure (↑)0.9498 ± 0.01150.9929 ± 0.00700.8809 ± 0.02160.3483 ± 0.02210.9929 ± 0.00590.9938 ± 0.0084
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0133 ± 0.00250.0408 ± 0.00860.0163 ± 0.00130.1483 ± 0.00750.0103 ± 0.00250.0103 ± 0.0021
SubsetAccuracy (↑)0.8013 ± 0.03000.5663 ± 0.12050.7530 ± 0.00600.0000 ± 0.00000.8464 ± 0.01500.8494 ± 0.0120
Accuracy (↑)0.8463 ± 0.03430.5733 ± 0.12150.7789 ± 0.00260.2658 ± 0.01390.9003 ± 0.02050.9029 ± 0.0178
F1measure (↑)0.8653 ± 0.03590.5758 ± 0.12200.7908 ± 0.00110.4043 ± 0.01330.9193 ± 0.02230.9218 ± 0.0201
MacroF1measure (↑)0.3273 ± 0.01430.1838 ± 0.02450.2593 ± 0.00030.4578 ± 0.05650.4463 ± 0.03160.4660 ± 0.0167
MicroF1Measure (↑)0.8363 ± 0.03230.5128 ± 0.10630.7873 ± 0.01460.3803 ± 0.01450.8754 ± 0.02860.8793 ± 0.0247
Missing rate: 0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0443 ± 0.00030.0708 ± 0.00360.0173 ± 0.00390.1179 ± 0.01350.0148 ± 0.00130.0138 ± 0.0021
SubsetAccuracy (↑)0.0453 ± 0.04530.1054 ± 0.02700.7319 ± 0.05730.0468 ± 0.08110.7713 ± 0.06020.7683 ± 0.0572
Accuracy (↑)0.0473 ± 0.04730.1099 ± 0.02550.7470 ± 0.05630.3373 ± 0.04950.8060 ± 0.04690.8053 ± 0.0444
F1measure (↑)0.0483 ± 0.04830.1114 ± 0.02500.7523 ± 0.05560.4773 ± 0.04160.8193 ± 0.04200.8189 ± 0.0400
MacroF1measure (↑)0.0183 ± 0.01850.0163 ± 0.00950.2393 ± 0.00030.4003 ± 0.03060.3623 ± 0.02260.3688 ± 0.0187
MicroF1Measure (↑)0.0739 ± 0.07390.1073 ± 0.02500.7663 ± 0.05300.4298 ± 0.03660.8143 ± 0.01790.8253 ± 0.0254
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0453 ± 0.00050.0600 ± 0.00160.0389 ± 0.00650.0503 ± 0.00560.0123 ± 0.00230.0120 ± 0.0020
SubsetAccuracy (↑)0.0030 ± 0.00300.0783 ± 0.02400.1658 ± 0.16560.5303 ± 0.01950.7923 ± 0.05130.7953 ± 0.0482
Accuracy (↑)0.0043 ± 0.00450.0813 ± 0.02410.1658 ± 0.16560.6973 ± 0.01200.8313 ± 0.04560.8333 ± 0.0440
F1measure (↑)0.0050 ± 0.00500.0823 ± 0.02410.1658 ± 0.16560.7673 ± 0.01330.8464 ± 0.04330.8488 ± 0.0415
MacroF1measure (↑)0.0043 ± 0.00400.0089 ± 0.00150.0363 ± 0.03650.4113 ± 0.04210.4270 ± 0.02350.4338 ± 0.0174
MicroF1Measure (↑)0.0098 ± 0.00960.0933 ± 0.02190.2148 ± 0.21460.6378 ± 0.02960.8413 ± 0.03210.8483 ± 0.0267
Table 8. Comparison results (mean ± std) of six approaches over the medical dataset.
Table 8. Comparison results (mean ± std) of six approaches over the medical dataset.
Medical
Missing rate: 0ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0163 ± 0.00050.0128 ± 0.00210.0213 ± 0.00000.2533 ± 0.00400.0113 ± 0.00210.0113 ± 0.0017
SubsetAccuracy (↑)0.4633 ± 0.03460.6613 ± 0.02850.2673 ± 0.01430.0000 ± 0.00000.6773 ± 0.03660.6858 ± 0.0490
Accuracy (↑)0.5303 ± 0.02250.7378 ± 0.01950.3143 ± 0.00750.1013 ± 0.00300.7623 ± 0.02300.7668 ± 0.0347
F1measure (↑)0.5529 ± 0.01850.7638 ± 0.01630.3293 ± 0.00500.1803 ± 0.00490.7914 ± 0.01860.7943 ± 0.0300
MacroF1measure (↑)0.2033 ± 0.01630.2783 ± 0.00690.0678 ± 0.00890.1043 ± 0.00630.3263 ± 0.00130.3293 ± 0.0254
MicroF1Measure (↑)0.6343 ± 0.01200.7538 ± 0.01730.4773 ± 0.00700.1758 ± 0.00460.7769 ± 0.02030.7803 ± 0.0315
Missing rate: 0.2ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0169 ± 0.00030.0450 ± 0.00060.0218 ± 0.00010.2503 ± 0.00250.0140 ± 0.00130.0138 ± 0.0011
SubsetAccuracy (↑)0.4184 ± 0.02650.0838 ± 0.01030.2388 ± 0.00600.0000 ± 0.00000.6103 ± 0.01030.6184 ± 0.0020
Accuracy (↑)0.4910 ± 0.01210.0959 ± 0.01230.2713 ± 0.00450.1013 ± 0.00160.6940 ± 0.01830.7023 ± 0.0080
F1measure (↑)0.5153 ± 0.00550.1000 ± 0.01290.2828 ± 0.00360.1803 ± 0.00300.7229 ± 0.02060.7308 ± 0.0101
MacroF1measure (↑)0.1793 ± 0.01000.0043 ± 0.00050.0514 ± 0.00630.0947± 0.00630.3343 ± 0.00320.3253 ± 0.0152
MicroF1Measure (↑)0.6143 ± 0.00200.0963 ± 0.01260.4153 ± 0.00290.1763 ± 0.00330.7170 ± 0.02250.7243 ± 0.0151
Missing rate: 0.4ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0210 ± 0.00030.0443 ± 0.00210.0223 ± 0.00210.2300 ± 0.00350.0160 ± 0.00030.0159 ± 0.0005
SubsetAccuracy (↑)0.2878 ± 0.00600.0714 ± 0.01030.2043 ± 0.00830.0000 ± 0.00000.5043 ± 0.00200.5063 ± 0.0041
Accuracy (↑)0.3373 ± 0.01460.0793 ± 0.00650.2429 ± 0.01230.1099 ± 0.00300.5913 ± 0.00830.5933 ± 0.0078
F1measure (↑)0.3543 ± 0.01830.0820 ± 0.00500.2558 ± 0.01350.1940 ± 0.00460.6213 ± 0.01110.6234 ± 0.0098
MacroF1measure (↑)0.1299 ± 0.02030.0038 ± 0.00000.0323 ± 0.00210.0934 ± 0.00430.2573 ± 0.01350.2583 ± 0.0092
MicroF1Measure (↑)0.4608 ± 0.00860.0810 ± 0.00330.3610 ± 0.02750.1873 ± 0.00430.6553 ± 0.00230.6573 ± 0.0015
Missing rate: 0.6ML-KNNTRAMLLFS-DLLCFMCeFLuCeFL
HammingLoss (↓)0.0268 ± 0.00210.0350 ± 0.00030.0273 ± 0.00130.2200 ± 0.00250.0193 ± 0.00060.0193 ± 0.0002
SubsetAccuracy (↑)0.0573 ± 0.00400.0510 ± 0.00600.0020 ± 0.00200.0000 ± 0.00000.4510 ± 0.00600.4533 ± 0.0287
Accuracy (↑)0.0663 ± 0.00300.0553 ± 0.00600.0020 ± 0.0180.1133 ± 0.00330.5463 ± 0.02350.5463 ± 0.0361
F1measure (↑)0.0694 ± 0.00550.0563 ± 0.00600.0020 ± 0.00210.1999 ± 0.00450.5809 ± 0.02900.5804 ± 0.0389
MacroF1measure (↑)0.0494 ± 0.01130.0048 ± 0.00110.0003 ± 0.00030.0893 ± 0.00230.2608 ± 0.02430.2853 ± 0.0222
MicroF1Measure (↑)0.1103 ± 0.00960.0699 ± 0.00730.0034 ± 0.00350.1929 ± 0.00430.6088 ± 0.02350.6108 ± 0.0225
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Z.; Zheng, X.; Yu, Y.; Dong, X.; Li, S. Updating Correlation-Enhanced Feature Learning for Multi-Label Classification. Mathematics 2024, 12, 2131. https://doi.org/10.3390/math12132131

AMA Style

Zhou Z, Zheng X, Yu Y, Dong X, Li S. Updating Correlation-Enhanced Feature Learning for Multi-Label Classification. Mathematics. 2024; 12(13):2131. https://doi.org/10.3390/math12132131

Chicago/Turabian Style

Zhou, Zhengjuan, Xianju Zheng, Yue Yu, Xin Dong, and Shaolong Li. 2024. "Updating Correlation-Enhanced Feature Learning for Multi-Label Classification" Mathematics 12, no. 13: 2131. https://doi.org/10.3390/math12132131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop