Next Article in Journal
Entropy Profiles for Li-Ion Batteries—Effects of Chemistries and Degradation
Previous Article in Journal
Vector Flows That Compute the Capacity of Discrete Memoryless Channels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

OMAL: A Multi-Label Active Learning Approach from Data Streams

1
School of Computer, Jiangsu University of Science and Technology, Zhenjiang 212100, China
2
School of Information Technology, Yancheng Institute of Technology, Yancheng 224051, China
*
Author to whom correspondence should be addressed.
The authors contribute equally to this work.
Entropy 2025, 27(4), 363; https://doi.org/10.3390/e27040363
Submission received: 16 March 2025 / Revised: 26 March 2025 / Accepted: 26 March 2025 / Published: 29 March 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
With the rapid growth of digital computing, communication, and storage devices applied in various real-world scenarios, more and more data have been collected and stored to drive the development of machine learning techniques. It is also noted that the data that emerge in real-world applications tend to become more complex. In this study, we regard a complex data type, i.e., multi-label data, acquired with a time constraint in a dynamic online scenario. Under such conditions, constructing a learning model has to face two challenges: it requires dynamically adapting the variances in label correlations and imbalanced data distributions and it requires more labeling consumptions. To solve these two issues, we propose a novel online multi-label active learning (OMAL) algorithm that considers simultaneously adopting uncertainty (using the average entropy of prediction probabilities) and diversity (using the average cosine distance between feature vectors) as an active query strategy. Specifically, to focus on label correlations, we use a classifier chain (CC) as the multi-label learning model and design a label co-occurrence ranking strategy to arrange label sequence in CC. To adapt the naturally imbalanced distribution of the multi-label data, we select weight extreme learning machine (WELM) as the basic binary-class classifier in CC. The experimental results on ten benchmark multi-label datasets that were transformed into streams show that our proposed method is superior to several popular static multi-label active learning algorithms in terms of both the Macro-F1 and Micro-F1 metrics, indicating its specifical adaptions in the dynamic data stream environment.

1. Introduction

In recent years, active learning has been developed to be a significant machine learning paradigm for specific application in a scenario where it is easy to collect an amount of instances, but labeling them requires unacceptable time, human, and economic costs [1,2]. In such a scenario, active learning can best improve the performance of a learning model by iteratively querying a few significant instances from human annotators (oracles). Some previous studies [3,4,5] have indicated that by active learning, a learning model can acquire comparative or better performance to a model trained with labeling all instances.
We note that most existing active learning studies focused on single-label data [3,4,5,6,7], i.e., each instance is only associated with a specific label. However, in many real-world applications, data may simultaneously associate with multiple labels, e.g., an image may contain desert, camels, clouds, and the Sun (see Figure 1); a news article may cover several different topics, such as war, economics, and politics; and a movie may relate to more than one topic, such as comedy, cartoon, and action. In general, we refer to this type of data as multi-label data, which widely exist in various real-world applications, including image recognition [8], recommender system [9,10], text classification [11], disease diagnosis [12], and sensor data analysis [13,14]. For multi-label data, learning models are required to simultaneously label all labels for an instance; thus, it is obviously more difficult to evaluate the significance level of an unlabeled multi-label instance than a single-label one in active learning [15].
Several studies have focused on multi-label active learning issues where the core concern lies in how to design a query strategy for determining which unlabeled instances are important for significantly improving the performance of learning models. In previous studies, uncertainty [16] and diversity [17] have been the two most widely used criteria for designing query strategies. Specifically, the uncertainty criterion always searches instances that are close to the classification boundary of the current learning model, while the diversity criterion focuses on the deviation level between unlabeled instances and the instances in the labeled set. Representative multi-label active learning methods include example-based active learning (EMAL) [18], label cardinality inconsistency (LCI) [19], and category vector inconsistency and ranking of scores (CVIRS) [20]. EMAL only adopts the uncertainty criterion to select significant unlabeled instances. Specifically, it first independently measures the uncertainty level in each label by the traditional single-label uncertainty level detection strategy, and then it uses the sum of these uncertainty levels to evaluate the uncertainty level in the corresponding multi-label instance. LCI regards the unlabeled instance whose number of pseudo-labels predicted by the current learning model has the most significant difference with the average label cardinality of all labeled instances as the most uncertain one. CVIRS focuses on both uncertainty and diversity, where the uncertainty level is measured by ranking the distances between all unlabeled instances and the current classification boundary, and the diversity level is quantized by calculating the average difference between the predicted label vector of an unlabeled instance and the real label vectors of all labeled instances.
Although there are some multi-label active learning methods, most of them only focus on static scenarios, i.e., all unlabeled instances are previously available. However, in many real-world applications, data can only be collected with a time constraint, i.e., emerging in the form of a data stream [21,22,23]. In such a scenario, considering the potential unlimitedness of the data stream and the finiteness of the data store, it is inevitable that the learning model must abide by the one pass rule [24]. This means that when a new unlabeled data chunk is received, the active learning model is required to immediately decide which instances are important and should be labeled and stored and then remove all other instances, which cannot be revisited. The active learning algorithms based on such a constraint are expected to simultaneously improve the quality of the learning model and adapt to potential concept drift [25,26,27]. To our best knowledge, only one previous work partially focuses on such a learning scenario and proposes an ensemble learning solution [28]. However, it has two significant drawbacks, as follows: (1) it independently trains a KNN classifier for each label, thus failing to focus on label correlations [29,30] that have been proven to be important for training multi-label learning models, and (2) it neglects the naturally imbalanced issue existing in multi-label data [31,32]. Therefore, it is necessary to design some effective and efficient multi-label active learning algorithms for application in the data stream environment.
For the purpose mentioned above, a novel online multi-label active learning algorithm, called OMAL, is proposed in this study. Similar to CVIRS [20], OMAL simultaneously focuses on both the uncertainty and diversity criteria within its query strategy. Specifically, the uncertainty level of an unlabeled instance is measured by calculating the average entropy of the prediction probability provided by the learning model across all labels. As we know, the entropy of the prediction probability can reflect how close an instance is the classification boundary. As for the diversity criterion, we do not consider the strategy of CVIRS but adopt the average cosine distance between the feature vector of an unlabeled instance and all labeled instances to calculate it. In contrast to calculating the difference between label vectors, such a strategy can more directly discover those diverse unlabeled instances. In the OMAL query strategy, the significance level of an unlabeled instance is integrated by weighting its uncertainty and diversity levels. In addition, considering that label dependency and a naturally imbalanced data distribution always exist in multi-label data, we adopt a classifier chain (CC) [33,34] as the classification model, in which a label co-occurrence ranking strategy is embedded to arrange the label sequence and guarantee the performance of classification models. We also use weighted extreme learning machine (WELM) [35] as the basic binary classification model for adapting the imbalanced data distribution. The experimental results on ten benchmark multi-label datasets that were transformed to be streams show that our proposed method is superior to several popular static multi-label active learning algorithms in terms of both the Macro-F1 and Micro-F1 metrics, indicating its effectiveness and superiority.
Specifically, the novelties of this study are listed as follows:
  • A novel instance significance active query strategy, which simultaneously considers uncertainty and diversity, is proposed;
  • To adapt both requirements of leveraging label correlations and treating class imbalance in multi-label learning, CC and WELM learning models are integrated into an online active learning framework;
  • A novel multi-label online active learning algorithm, called OMAL, is proposed in this study, and to our best knowledge, it is the first algorithm that totally considers all requirements in this specific learning scenario.
The rest of this paper is organized as follows. Section 2 first simply describes the basic framework of stream-based active learning; then, it introduces the query strategy and classification models in detail and, finally, presents the flow path of the proposed OMAL algorithm and discusses its time complexity. In Section 3, the experimental results and analysis are provided in detail. Finally, Section 4 discusses the findings and contributions of this study and further indicates the future work.

2. Methods

2.1. The Basic Framework of Stream-Based Active Learning

In a stream-based active learning framework, there are generally five basic components, as follows: a labeled set Φ L , a most recently received unlabeled set U t , a classification model S, a pre-designed query strategy Q, and an oracle H, which may be either a human annotator or others, e.g., ChatGPT. During active learning, when a new unlabeled data chunk U t is received, the query strategy Q and classification model S are both activated to select some significant instances from U t to be submitted to the oracle H for labeling. Next, this new labeled set U t , where U t U t , is added into the labeled set Φ L , i.e., Φ L = Φ L U t , to be used for training a new classifier to replace S. The aforementioned procedure is iteratively conducted while continually receiving new unlabeled data. The basic framework is described in Figure 2. Specifically, when data simultaneously associate with multiple labels, both the classification model S and the query strategy Q are required to adapt to the multi-label scenario. The differences between pool-based active learning and stream-based active learning are reflected in the two following aspects: (1) in static active learning, Q can directly visit the whole unlabeled set U, while online active learning can only visit the current unlabeled chunk U t , and (2) static active learning reserves all unlabeled instances U = U \ U t after each iteration, while online active learning abandons all remaining unlabeled instances U t = U t \ U t in the current data chunk to hinder revisiting forever.

2.2. Query Strategy of OMAL

Suppose that Φ L = { x i , Y i | x i R m , i 1,2 , , n , Y i = y i 1 , y i 2 , , y i L } is a labeled multi-label set and u t = { x j | x j R m , j 1,2 , , k } is a new received unlabeled block, where x i and Y i respectively denote a feature vector and a label vector corresponding to the ith instance, n and k respectively represent the number of instances in the labeled set and new received unlabeled block, y i k indicates the kth class label of the ith instance, m denotes the number of features, and |L| represents the number of labels. Active learning is required to extract k′ (k′ < k) instances, according to a pre-designed query strategy Q, from u t for submission to the oracle H for labeling, which are then used to extend Φ L . For an active learning algorithm, its query strategy is most important as it directly relates to the quality of the algorithm.
In this study, we combine both the uncertainty and diversity criteria to design a multi-label query strategy. As indicated in Section 1, the uncertainty criterion aims to search those instances that are most close to the current classification boundary. These instances often emerge in the class overlapping area, and thus, their label confidences are much lower than those of other instances. We use the average entropy of the prediction probability across all |L| labels to search close-to-boundary instances. The difficulty lies in identifying how to calculate the prediction probability for a non-Bayes classification model. Fortunately, for our selected WELM classifier, there is a way to transform the soft outputs into prediction probabilities [36]. Suppose that for an unlabeled instance x j , its output in the trained WELM corresponding to the ith label is f i ( x j ) ; then, it can be transformed as a prediction probability P ( y i | f i ( x j ) ) by the sigmoid function as follows:
P y i f i x j = 1 1 + e x p ( f i ( x j ) )
It has been shown that the transformation strategy can create an approximately accurate mapping between the outputs of extreme learning machine and posterior probabilities of the naïve Bayes classifier in theory [36]. Based on this transformation, the average entropy of x j can be calculated as follows:
E N x j = 1 | L | i = 1 | l |   P y i f i x j log P y i f i x j + 1 P y i f i x j log 1 P y i f i x j
where E N x j denotes the average entropy of the instance x j across all labels. It is clear that if an instance has a larger EN, it will be more uncertain.
As for the diversity criterion, we consider that if two instances have a more approximated distribution in the feature space, then it might mean that their labels are more similar. In the same fashion, when an instance has a large difference from all other instances, it may represent a novel pattern. Based on this basic hypothesis, we suggest adopting the average cosine distance between an unlabeled instance x j and all n labeled instances in Φ L to calculate the diversity of x j . The reason for adopting the cosine distance but not the Euclidean distance is that the cosine distance focuses more on the direction difference between two instances, and thus, it is more robust than the Euclidean distance, especially when the data are highly dimensional. For an unlabeled instance x j and any one labeled instance x i , their cosine distance is calculated as follows:
cos x i , x j = x i · x j x i   | | x j | |
where x i · x j represents their dot product and x i denotes the norm of the vector x i . However, we note that when the cosine distance between two instances is larger, their difference is smaller, which is inconsistent with the requirement of diversity evaluation. Thus, we define the diversity measure as follows:
d x j = 1 1 n i = 1 n cos x i , x j
where d x j denotes the diversity level of x j corresponding to all n labeled instances. Selecting those unlabeled instances with a large diversity to the labeling instances helps to cover diverse instance patterns, further improving the robustness of the classification models.
In our proposed query strategy, the significance level of an unlabeled instance can be calculated by weighting both the uncertainty level and diversity level as follows:
s x j = λ × E N x j + ( 1 - λ ) × d x j
where s x j denotes the significance level of the unlabeled instance x j and λ ( λ [ 0,1 ] ) is the weighing factor that is used to regulate the relative significance of uncertainty and diversity. In this study, we empirically designate λ as 0.5, i.e., uncertainty and diversity contribute equally to the significance level.

2.3. Classification Models Used in OMAL

2.3.1. Classifier Chains

It is well known that the reason why it is more difficult to construct a multi-label learning model than a single-label one lies in the fact that some label correlations exist in multi-label data [29,30], e.g., if an image has the sandbeach label, then it has a higher probability of simultaneously owning the sea label but a lower probability of holding the waterfall label. Therefore, it is necessary to make multi-label classification models leverage label correlation information. CC [33,34] belongs to this kind of learning model, and it has been widely used in many real-world multi-label applications. CC first arranges all |L| labels into a chain as follows: l 1 l 2 l | L | . It then sequentially trains the |L| binary-class classifier to predict each label. Specifically, for the ith learning model, the feature space of training instances is extended to include both original features and the first i − 1 labels. Suppose the original feature space is represented by X; then, the first binary classifier can be represented as C L ( X ) l 1 , and the second binary classifier can be represented as C L ( X l 1 ) l 2 . In a similar fashion, the final binary classifier can be represented as C L ( X l 1 l 2 l | L 1 | ) l | L | , where C L denotes a binary classifier. By this way, class correlations are partially embedded into the learning model S, helping to improve its prediction performance.
However, for CC, some previous studies have found that the order of labels in sequence can directly influence the performance of the learning model [37,38]. If some difficult labels are first predicted, then the error accumulation phenomenon would emerge to misguide the prediction of subsequent labels. Aiming to address this issue, we used the idea of label dependency drift detector (LD3) [39] to design a label significance ranking strategy. Specifically, LD3 was originally proposed to detect label matching level between two data chunks existing in a multi-label data stream. In our algorithm, we only take advantage of the label correlation significance ranking strategy adopted by LD3. For the labeled set Φ L , it first calculates the co-occurrence matrix, which is obtained by counting the number of times each class label occurs as “1” alongside other labels. The generated matrix is then ranked within each row, which is called local ranking, by creating a ranking for each label based on their co-occurrence frequencies. After acquiring the local ranking, the ranks can be further aggregated as the global ranking by
r i = 1 j = 1 , j i | L | 1 r i j
where r i denotes the ranking score of the label l i and r i j represents the ranking of the label l i in the jth row of the local ranking matrix. Next, we obtain the global ranking position sequence R by ranking all r i in ascending ranking. Obviously, the labels with more correlations with other labels will have a higher ranking in R, and predicting these labels first could provide more useful information for improving the accuracy of other labels.

2.3.2. Weighted Extreme Learning Machine

As we know, multi-label data always suffer from the class imbalance problem on most or all of the labels [31,32]; thus, we should not ignore this issue when selecting the basic binary classification model. For the binary-class imbalanced issue, there are abundant solutions, including sampling [40], cost-sensitive learning [35], threshold moving [41], and ensemble learning [42,43]. Considering the time-complexity requirement of the online environment, we decided to adopt an efficient, cost-sensitive learning method, that is, weighted extreme learning machine (WELM), which is both fast and robust [35].
WELM is a variant of ELM [44,45], aiming to alleviate the impact of imbalanced data distributions. ELM is a fast learning algorithm used to train single-hidden-layer feedforward neural networks (SLFNs). For a binary-class problem, let us suppose there are n training instances ( x i ,   t i ), where x i R m and t i { 1,1 } . If an SLFNs with L hidden nodes can approximate these n instances with zero error, then it implies that there exists a i , b i , and β i , such that:
f L x j = i = 1 K β i G ( a i , b i , x j ) = t j ,   j = 1,2 , , n
where a i and b i respectively denote the weight and bias of the ith hidden node that is randomly generated, K represents the number of hidden nodes, G denotes the activation function, and β i indicates the weight vector connecting the ith hidden node to the output nodes. Then, we can write Equation (7) compactly as follows:
H β = T
where
H = h ( x 1 ) h ( x n ) = G ( a 1 , b 1 , x 1 ) G ( a K , b K , x 1 ) G ( a 1 , b 1 , x n ) G ( a K , b K , x n ) β = β 1 β K ,   T = t 1 t K
Since H and T are both known, β can be directly calculated as follows:
β ^ = H T
where H denotes the Moore–Penrose generalized inverse of the hidden layer output matrix H . ELM can be also trained in the viewpoint of optimization, i.e., simultaneously minimizing H β T 2 and β 2 . Then, the issue can be described as follows:
M i n i m i z e : 1 2 β 2 + 1 2 C i = 1 n ξ i 2 S u b j e c t t o : h x i β = t i ξ i ,   i = 1 ,   2 , , n  
where ξ i denotes the training error of the training instance x i and C represents the penalty factor, which is the tradeoff between training errors and the generalization ability of the learning model. Then, β can be solved as:
β = H T ( I C + H H T ) 1 T , when   n K ( I C + H H T ) 1 H T T , when   n > K
However, ELM treats the training error of each instance equally, which can cause the learning model to be partial to the majority class when the data distribution is biased. WELM addresses this issue by introducing a weighted matrix W into the optimization formula described in Equation (11), which is rewritten as follows:
M i n i m i z e : 1 2 β 2 + 1 2 C W i = 1 n ξ i 2 S u b j e c t   t o : h x i β = t i ξ i , i = 1 ,   2 , , n  
where W is a n × n diagonal matrix in which each value on its diagonal represents the corresponding regulation weight of the penalty factor C. In [35], W was suggested to be set as:
W i i = 1 j = 1 n I N ( t j = 1 ) , i f   t i = 1   1 j = 1 n I N ( t j = 1 ) , i f   t i = 1
where I N is the indicator function, which returns 1 if the corresponding condition holds; otherwise, it returns 0. Then, the solution of β is represented as:
β = H T ( I C + W H H T ) 1 W T , when   n K ( I C + H W H T ) 1 H T W T , when   n > K
By means of this weighting strategy, the training errors of the minority instances are exerted larger penalties than those of the majority class, further providing an impartial training result.

2.4. Description of the OMAL Algorithm

Next, the procedure of the OMAL algorithm (Algorithm 1) is described as follows:
Algorithm 1: OMAL
Input: a null labeled set Φ L , an unlabeled multi-label data block stream u 1 , u 2 , ,   u t , , the significance level of query strategy regulation weight λ , the querying rate θ , the activation function G, the number of hidden nodes K, and the penalty factor C.
Output: the current classification model S.
Procedure:
  • Label u 1 by the oracle when we have received it;
  • Put the labeled u 1 into Φ L ;
  • Remove u 1 ;
  • On Φ L , for each label i = 1, 2,…,|L|, calculate r i by Equation (6);
  • Acquire the label ranking sequence R by arranging all r i in descending order;
  • For each label in R, train the corresponding binary classifiers C L 1 ,   C L 2 ,   ,   C L | L | by calling WELM;
  • Acquire the initial CC classification model S = [ C L 1 ,   C L 2 ,   ,   C L | L | ];
  • When a new unlabeled data chunk u t is received:
  •   Transform the output of each instance in u t on S to be posterior probabilities by Equation (1);
  •   Calculate the average entropy of each instance in u t by Equation (2);
  •   Calculate the diversity level of each instance in u t by Equations (3) and (4);
  •   Calculate the significance level of each instance in u t by Equation (5);
  •   Rank the significance level of all instances in u t in descending order and further select the first θ instances from them for querying;
  •   Label these selected instances by the oracle;
  •   Add the new labeled instances to Φ L ;
  •   Remove u t ;
  •   On Φ L , for each label i = 1, 2,…,|L|, calculate r i by Equation (6);
  •   Acquire the label ranking sequence R by arranging all r i in descending order;
  •   For each label in R, train the corresponding binary classifiers C L 1 ,   C L 2 ,   ,   C L | L | by calling WELM;
  •   Update the CC classification model S by the newly trained |L| binary classifiers, i.e., S = [ C L 1 ,   C L 2 ,   ,   C L | L | ];
  •   Output the current classifier S;
  •   Return to step 8.
Specifically, the querying rate θ [ 0 % , 100 % ] is a user-specified parameter that is used to decide how many instances should be selected from an unlabeled block for querying. An undersize θ tends to miss some significant instances, which may help to improve classification performance, while an oversize θ tends to introduce some useless information while increasing the labeling burden and training time. In next section, the impact of θ will be presented by describing the experimental results in detail.
In addition, it can be observed that OMAL fails to consider concept drift [46], which may emerge in multi-label data streams. In such a scenario, it is suggested that LD3 [39] can be first used to detect whether concept drift has happened, and if it has, its strength should be further measured. Then, the labeled instances with time stamps should be refined based on the estimated concept drift strength. This means that some older labeled instances should be removed from the labeled set Φ L to avoid their participation in training new classification models. Such manipulation can be seen as an alternative of the forgetting mechanism.

2.5. Time Complexity of the OMAL Algorithm

Finally, we try to analyze the time complexity of the proposed OMAL algorithm running on a round. First, both the transformation of posterior probabilities and the calculation of average entropy consume O(k) time. The time complexity of calculating the average diversity is O(nk). Then, calculating the significance levels of all unlabeled instances in the current block u t consumes O(k) time. Next, generating the co-occurrence matrix costs O(n|L|) time, calculating the global rankings consumes O(|L|2) time, and ranking the label sequence takes O(|L|log|L|) time. As we know, the time complexity of training a WELM classifier is O(n3), and thus, training |L| WELMs in CC costs O(n3|L|) in total. In general, k << n and |L| << n3, and thus, the time complexity of OMAL is O(n3|L|). In other words, training binary classifiers dominates the time consumption of OMAL.
Therefore, there exist two potential risks that hinder the application of OMAL in some scenarios that have high demands for running time. The first one is that with a continuous increase in labeled instances, the time burden of updating classifiers will increase in exponential fashion. While the second one is that if facing a dataset with tremendous labels, the time consumption may become unacceptable. In such scenarios, we suggest using incremental multi-label learning models, which do not require revisiting old labeled instances, and, at the same time, can focus on label correlations to some extent and dynamically adapt to the variance in class imbalance, e.g., LW-ELM [47], to replace the combination of CC and WELM. Although such a replacement may have some negative impacts on the classification performance of learning models, but it can still be regarded as an effective alternative to maintain the tradeoff between model performance and running time.

3. Results

3.1. Datasets Used in This Study

In this study, all experiments were conducted on ten multi-label datasets acquired from the Multi-Label Classification Dataset Repository, which is available at http://www.uco.es/kdis/mllresources/ (accessed on 17 December 2024). These datasets contain 194~10,810 instances, 16~1449 features, 5~45 labels, and 1.074~5.073 label cardinality (LC), which indicates the average number of labels on each instance, and cover various fields, including image, medicine, music, chemistry, biology, and text. Specifically, these datasets were randomly transformed to be data streams with different block sizes, and for each dataset, 30% of the instances were reserved as testing instances. A detailed description about these datasets is presented in Table 1.

3.2. Experimental Settings

All experiments were conducted in the environment of Python 3.8 on AMD Ryzen 9 7945HX with Radeon Graphics 2.50 GHz and 32 GB RAM. We compared the proposed OMAL algorithm with a baseline query algorithm, i.e., Random, and three previous algorithms, including EMAL [18], LCI [19], and CVIRS [20]. Specifically, Random extracts unlabeled instances randomly for submission to the oracle for labeling. To guarantee the impartiality of the comparative experiments, all algorithms used CC [29,30] as the basic multi-label classifier, where WELM [35] was adopted as the basic binary classifier. In addition, the grid search strategy was used to determine the best combination of parameters K and C by an internal five-fold cross validation, in which K { 100,200 , . , 2000 } and C { 10 3 , 10 2 , , 10 5 } , respectively.
As for the performance evaluation metrics, we used the two most popular metrics for evaluating the quality of a multi-label learning algorithm considering imbalanced data distributions, which are the Macro-F1 and Micro-F1 metrics [48]. Both Macro-F1 and Micro-F1 evaluate the harmonic means of precision and recall, and they can be defined as follows:
Macro F 1 = 2 × P macro   × R macro   P macro   + R macro    
Micro F 1 = 2 × P micro   × R micro   P micro   + R micro  
where P macro   and R macro   represent, respectively, macroscopic precision and recall, and they are defined as follows:
P macro   = 1 | L | i = 1 | L |   T P i T P i + F P i
R macro   = 1 | L | i = 1 | L |   T P i T P i + F N i
where T P i , F P i , and F N i denote the number of true positive, false positive, and false negative instances on the ith label predicted by the classification model, respectively. P micro   and R micro   respectively denote the microscopic precision and recall, which can be calculated as follows:
P micro   = i = 1 | L |   T P i i = 1 | L |   T P i + F P i
R micro   = i = 1 | L |   T P i i = 1 | L |   T P i + F N i
Furthermore, considering the randomness of data streams, each experiment was randomly conducted twenty times, and then, the average performance was presented in the corresponding learning curve. The area under learning curve (ALC) was finally calculated to compare the quality of the various active learning algorithms.

3.3. Experimental Results

First, we varied the querying rate θ from 10% to 100% with an increment of 10% to observe the impact of this parameter on the learning effect. Figure 3 presents the trends of the various active learning algorithms with the variance of θ on each dataset.
The results in Figure 3 show the following:
(1)
When increasing the value of θ , i.e., actively querying more unlabeled instances, the various active learning algorithms tend to yield better classification performance;
(2)
When θ is designated as a small value, increasing it can provide a more significant performance improvement;
(3)
After designating a medium value for θ , continually increasing the value of θ can not significantly improve classification performance, and even on some datasets, the performance presents a declining trend.
It is not difficult to understand the first phenomenon, where, although active learning can provide the most informative instances, if they are extremely scare, it still tends to yield underfitting results for classification models. The two other phenomena tell us that when a sufficient number of informative instances have been learned, it is not necessary to continually learn the remaining instances; otherwise, more noise will be added, and more labeling burdens will be required. According to the feedback from the results in Figure 3, we empirically set θ = 60 % .
Next, we present the learning curves of several comparable algorithms on each data stream in Figure 4. Furthermore, the ALC values and average rankings of various algorithms based on the Macro-F1 and Micro-F1 metrics are shown in Table 2 and Table 3, respectively.
From the results in Figure 4 and Table 2 and Table 3, we can draw several conclusions, as follows:
(1)
In comparison to Random, several of the active learning algorithms except EMAL yielded higher ALC values in terms of both the Macro-F1 and Micro-F1 metrics, indicating that these algorithms can select more significant unlabeled instances to query and further help to improve the quality of a multi-label learning model to a large extent. As the worst querying strategy, EMAL only focuses those instances that are closest to the classification boundary, but it neglects the diversity of the data distribution, which tends to make the classification boundary converge to a local optimum one. This explains why EMAL performs worse than Random.
(2)
Although LCI acquired a higher average ranking than Random on both metrics, its superiority is not significant enough. In essence, LCI can be seen as an unconventional query strategy with partial exploration ability, as it always queries those unlabeled instances with a significant difference in terms of label cardinality with labeled instances. Therefore, we cannot prevent it from converging to a local optimum boundary rapidly, but its convergence speed is obviously slower than EMAL.
(3)
Both CVIRS and the proposed OMAL significantly outperform several of the other algorithms. It is difficult to understand this result since both these algorithms simultaneously focus on the significance of unlabeled instances in terms of both uncertainty and diversity. Therefore, we can say that CVIRS and OMAL both own exploitation and exploration abilities. More informative instances can be selected for querying by them.
(4)
In comparison to CVIRS, OMAL adopts a direct way of exploring diversity in the feature space, which helps to rapidly adapt the potential variance in the feature space in the data stream. In contrast, the exploration of diversity by observing the variance in the label vectors adopted by CVIRS may be not robust enough. In addition, the time consumption of CVIRS is always higher than that of OMAL, owing to the sophisticated ranking aggregation strategy adopted by CVIRS.
(5)
OMAL yielded the best ALC value based on the Macro-F1 metric on seven datasets and the best ALC value based on the Micro-F1 metric on eight datasets. In addition, OMAL acquired the lowest average ranking of 1.4 on both metrics. These results show its effectiveness and superiority in dealing with multi-label data stream active learning scenarios. We believe that on a static unlabeled set, where all instances are initially available, the continuously strong space exploration ability of OMAL may be not necessary. While on a dynamic data stream, it will contribute more to the performance improvement of the classification model.

3.4. Significance Analysis

Next, the Nemenyi test, a post hoc test for the Friedman test [49,50], was adopted to observe whether there exists a significant difference between the proposed OMAL algorithm and any one of the compared algorithms. Specifically, if the average ranking of OMAL and that of a compared algorithm differ by at least one critical difference (CD) unit, then we considered that their performances are significantly different. Specifically, the CD was calculated as follows:
C D = q α J ( J + 1 ) 6 D
where q α denotes the significance level, J represents the number of compared algorithms, and D indicates the number of datasets. In our experiments, J = 5 and D = 10, and thus, with the significance level α = 0.05 , we calculate C D = 1.9288 by Equation (22). The CD diagrams are presented in Figure 5.
In Figure 5, we observe that on both metrics, our proposed OMAL algorithm significantly outperforms both the EMAL and Random algorithms. While in comparison with the LCI and CVIRS algorithms, although the superiority of OMAL is not significant, it still acquires an obviously lower average ranking than them.

3.5. Ablation Experiments

Next, we conducted three groups of ablation experiments to estimate whether the adoption of some components or strategies are necessary.
In the first group of ablation experiments, we focused on the query strategy. Specifically, we compared the performance of only adopting the uncertainty query criterion, only adopting the diversity query criterion, and integrating both. The results in Table 4 show that integrating both uncertainty and diversity is necessary as it yielded the best results on seven and eight datasets in terms of the Macro-F1 and Micro-F1 metrics, respectively. This means that the local search ability and the global exploration ability are both very important for active learning. In addition, we note that using only uncertainty as the querying strategy seems to easily yield a better performance than using only diversity. However, this does not mean that actively querying diversity instances is useless.
In the second group of ablation experiments, we compared ELM and WELM further to judge whether the class imbalance issue should be considered in multi-label active learning. The results in Table 5 illustrate that WELM yielded the best results on nine and eight datasets, which is significantly superior to ELM, indicating it is necessary to use a classifier that is specifically designed for dealing with the class imbalance issue.
The third group of ablation experiments refers to the label sequence in CC. We compared the performance of CC based on a random label sequence and ranking labels by the significance estimation of LD3. The results in Table 6 indicate that it is necessary to rank the order of labels according to their correlation significance because, on most datasets, adopting the random label sequence yielded a poorer performance than that based on LD3. This means that putting high correlation labels in front of the label sequence of CC can provide more information for improving the classification performance of sequent binary classifiers.

3.6. Results of Running Time

Finally, we also compared the average running time of several active learning algorithms throughout all rounds. The results in Table 7 indicate that OMAL is a time-saving algorithm as it always costs an approximately equal running time as the Random, EMAL, and LCI algorithms, but it costs significantly less running time than CVIRS. Considering in our experimental framework, training classifiers dominate the time complexity of several of the algorithms except CVIRS, and thus, the differences in the running time among several active learning algorithms can be ignored. For OMAL, the difference in the number of training instances n and the number of labels |L| on different datasets result in the difference in the running time.

4. Conclusions

This study proposed an active learning algorithm called OMAL, aiming to improve the performance of multi-label learning models in a data stream scenario. Specifically, OMAL adopts a novel integrated query strategy, where the average entropy of prediction probabilities is used to estimate the uncertainty level of unlabeled instances, and the average cosine distance in feature space is used to estimate its diversity level. To focus the label correlations, CC is used as the basic multi-label classifier. Also, to adapt the naturally imbalanced data distribution existing in multi-label data, WELM is used as the basic binary classifier. The experimental results on ten transformed multi-label data streams illustrated that OMAL outperforms random querying and several traditional static active querying algorithms, indicating its effectiveness and superiority. In addition, our ablation experiments verified the significance of each component embedded in OMAL.
Specifically, the contributions of this study can be summarized as follows:
  • It proposed a novel multi-label active query strategy that simultaneously satisfies the requirements of exploration and exploitation in online environments;
  • It designed an effective multi-label online solution to simultaneously leverage label correlation information and adapt the class imbalance distribution;
  • It introduced LD3 label correlation information into a CC model to avoid error accumulation;
  • It presented a subtle multi-label online active learning algorithm that can produce excellent performance and, meanwhile, is relatively time-saving.
In future work, we expect to develop more efficient active learning algorithms for adapting high-speed multi-label data streams. In addition, methods to improve OMAL for tracking and adapting drifting multi-label data streams will also be investigated.

Author Contributions

Conceptualization, Q.F. and H.Y.; methodology, Q.F., C.X. and J.D.; software, C.X. and B.S.; validation, Q.F., C.X. and C.S.; formal analysis, S.X. and H.Y.; investigation, Q.F. and X.Y.; resources, C.S. and H.Y.; data curation, Q.F. and C.X.; writing—original draft preparation, Q.F., C.X. and B.S.; writing—review and editing, J.D., C.S., X.Y. and H.Y.; visualization, Q.F.; supervision, S.X., X.Y. and H.Y.; project administration, H.Y.; funding acquisition, S.X., X.Y. and H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Natural Science Foundation of China under grants No. 62176107, No. 62076111, and No. 62076215.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All multi-label datasets used in the experiments are from a multi-label classification dataset repository that is available at http://www.uco.es/kdis/mllresources/ (accessed on 17 December 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ALCarea under learning curve
CCclassifier chain
CDcritical difference
CVIRScategory vector inconsistency and ranking of scores
ELMextreme learning machine
EMALexample-based active learning
KNNK-nearest neighbors
LClabel cardinality
LCIlabel cardinality inconsistency
LD3label dependency drift detector
OMALonline multi-label active learning
WELMweighted extreme learning machine

References

  1. Amin, S.U.; Hussain, A.; Kim, B.; Seo, S. Deep learning based active learning technique for data annotation and improve the overall performance of classification models. Expert Syst. Appl. 2023, 228, 120391. [Google Scholar]
  2. Zhang, L.; Chen, C.; Bu, J.; Cai, D.; He, X.; Huang, T.S. Active learning based on locally linear reconstruction. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2026–2038. [Google Scholar] [CrossRef]
  3. Guo, H.; Wang, W. An active learning-based SVM multi-class classification model. Pattern Recognit. 2015, 48, 1577–1597. [Google Scholar]
  4. Goudjil, M.; Koudil, M.; Bedda, M.; Ghoggali, N. A novel active learning method using SVM for text classification. Int. J. Autom. Comput. 2018, 15, 290–298. [Google Scholar]
  5. Cebron, N.; Berthold, M.R. Active learning for object classification: From exploration to exploitation. Data Min. Knowl. Discov. 2009, 18, 283–299. [Google Scholar]
  6. Sourati, J.; Akcakaya, M.; Dy, J.G.; Leen, T.K.; Erdogmus, D. Classification active learning based on mutual information. Entropy 2016, 18, 51. [Google Scholar] [CrossRef]
  7. Karaca, A.C.; Bilgin, G. An end-to-end active learning framework for limited labelled hyperspectral image classification. Int. J. Remote Sens. 2025, 1–28. [Google Scholar] [CrossRef]
  8. Shen, X.; Dong, G.; Zheng, Y.; Lan, L.; Tsang, I.W.; Sun, Q.S. Deep co-image-label hashing for multi-label image retrieval. IEEE Trans. Multimed. 2021, 24, 1116–1126. [Google Scholar] [CrossRef]
  9. Bogaert, M.; Lootens, J.; Van den Poel, D.; Ballings, M. Evaluating multi-label classifiers and recommender systems in the financial service sector. Eur. J. Oper. Res. 2019, 279, 620–634. [Google Scholar]
  10. Laghmari, K.; Marsala, C.; Ramdani, M. An adapted incremental graded multi-label classification model for recommendation systems. Prog. Artif. Intell. 2018, 7, 15–29. [Google Scholar]
  11. Wang, R.; Ridley, R.; Qu, W.; Dai, X. A novel reasoning mechanism for multi-label text classification. Inf. Process. Manag. 2021, 58, 102441. [Google Scholar] [CrossRef]
  12. Zhou, L.; Zheng, X.; Yang, D.; Wang, Y.; Bai, X.; Ye, X. Application of multi-label classification models for the diagnosis of diabetic complications. BMC Med. Inform. Decis. Mak. 2021, 21, 182. [Google Scholar] [CrossRef] [PubMed]
  13. Sonkul, N.A.; Dhage, G.S.; Vyas, N.S. Single and multi-label fault classification in rotors from unprocessed multi-sensor data through deep and parallel CNN architectures. Expert Syst. Appl. 2021, 185, 115565. [Google Scholar] [CrossRef]
  14. Jethanandani, M.; Sharma, A.; Perumal, T.; Chang, J.R. Multi-label classification based ensemble learning for human activity recognition in smart home. Internet Things 2020, 12, 100324. [Google Scholar] [CrossRef]
  15. Wu, J.; Sheng, V.S.; Zhang, J.; Li, H.; Dadakova, T.; Swisher, C.L.; Cui, Z.; Zhao, P. Multi-label active learning algorithms for image classification: Overview and future promise. ACM Comput. Surv. (CSUR) 2020, 53, 1–35. [Google Scholar] [CrossRef]
  16. Nguyen, V.L.; Shaker, M.H.; Hüllermeier, E. How to measure uncertainty in uncertainty sampling for active learning. Mach. Learn. 2022, 111, 89–122. [Google Scholar] [CrossRef]
  17. Wang, R.; Wang, X.Z.; Kwong, S.; Xu, C. Incorporating diversity and informativeness in multiple-instance active learning. IEEE Trans. Fuzzy Syst. 2017, 25, 1460–1475. [Google Scholar] [CrossRef]
  18. Wu, J.; Sheng, V.S.; Zhang, J.; Zhao, P.; Cui, Z. Multi-label active learning for image classification. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 5227–5231. [Google Scholar]
  19. Li, X.; Guo, Y. Active Learning with Multi-Label SVM Classification. Int. Jt. Conf. Artif. Intell. (IJCAI) 2013, 13, 1479–1485. [Google Scholar]
  20. Reyes, O.; Morell, C.; Ventura, S. Effective active learning strategy for multi-label learning. Neurocomputing 2018, 273, 494–508. [Google Scholar] [CrossRef]
  21. Song, Y.; Lu, J.; Lu, H.; Zhang, G. Learning data streams with changing distributions and temporal dependency. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 3952–3965. [Google Scholar] [CrossRef]
  22. Gaudreault, J.G.; Branco, P. A systematic literature review of novelty detection in data streams: Challenges and opportunities. ACM Comput. Surv. 2024, 56, 1–37. [Google Scholar] [CrossRef]
  23. Buddhika, T.; Pallickara, S.L.; Pallickara, S. Pebbles: Leveraging sketches for processing voluminous, high velocity data streams. IEEE Trans. Parallel Distrib. Syst. 2021, 32, 2005–2020. [Google Scholar] [CrossRef]
  24. Bahri, M.; Bifet, A.; Gama, J.; Gomes, H.M.; Maniu, S. Data stream analysis: Foundations, major tasks and tools. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2021, 11, e1405. [Google Scholar] [CrossRef]
  25. Lu, J.; Liu, A.; Dong, F.; Gu, F.; Gama, J.; Zhang, G. Learning under concept drift: A review. IEEE Trans. Knowl. Data Eng. 2018, 31, 2346–2363. [Google Scholar] [CrossRef]
  26. Gama, J.; Žliobaitė, I.; Bifet, A.; Pechenizkiy, M.; Bouchachia, A. A survey on concept drift adaptation. ACM Comput. Surv. (CSUR) 2014, 46, 1–37. [Google Scholar] [CrossRef]
  27. Webb, G.I.; Hyde, R.; Cao, H.; Nguyen, H.L.; Petitjean, F. Characterizing concept drift. Data Min. Knowl. Discov. 2016, 30, 964–994. [Google Scholar] [CrossRef]
  28. Wang, P.; Zhang, P.; Guo, L. Mining multi-label data streams using ensemble-based active learning. In Proceedings of the 2012 SIAM International Conference on Data Mining, Anaheim, CA, USA, 26–28 April 2012; pp. 1131–1140. [Google Scholar]
  29. Zhu, Y.; Kwok, J.T.; Zhou, Z.H. Multi-label learning with global and local label correlation. IEEE Trans. Knowl. Data Eng. 2017, 30, 1081–1094. [Google Scholar] [CrossRef]
  30. Duan, J.; Yang, X.; Gao, S.; Yu, H. A partition-based problem transformation algorithm for classifying imbalanced multi-label data. Eng. Appl. Artif. Intell. 2024, 128, 107506. [Google Scholar] [CrossRef]
  31. Duan, J.; Gu, Y.; Yu, H.; Yang, X.; Gao, S. ECC++: An algorithm family based on ensemble of classifier chains for classifying imbalanced multi-label data. Expert Syst. Appl. 2024, 236, 121366. [Google Scholar] [CrossRef]
  32. Zhang, M.L.; Li, Y.K.; Yang, H.; Liu, X.Y. Towards class-imbalance aware multi-label learning. IEEE Trans. Cybern. 2020, 52, 4459–4471. [Google Scholar] [CrossRef]
  33. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains for multi-label classification. Mach. Learn. 2011, 85, 333–359. [Google Scholar]
  34. Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains: A review and perspectives. J. Artif. Intell. Res. 2021, 70, 683–718. [Google Scholar]
  35. Zong, W.; Huang, G.B.; Chen, Y. Weighted extreme learning machine for imbalance learning. Neurocomputing 2013, 101, 229–242. [Google Scholar]
  36. Yu, H.; Sun, C.; Yang, W.; Yang, X.; Zuo, X. AL-ELM: One uncertainty-based active learning algorithm using extreme learning machine. Neurocomputing 2015, 166, 140–150. [Google Scholar]
  37. Gonçalves, E.C.; Plastino, A.; Freitas, A.A. A genetic algorithm for optimizing the label ordering in multi-label classifier chains. In Proceedings of the 2013 IEEE 25th International Conference on Tools with Artificial Intelligence, Herndon, VA, USA, 4–6 November 2013; pp. 469–476. [Google Scholar]
  38. Mishra, N.K.; Singh, P.K. Linear ordering problem based classifier chain using genetic algorithm for multi-label classification. Appl. Soft Comput. 2022, 117, 108395. [Google Scholar]
  39. Gulcan, E.B.; Can, F. Unsupervised concept drift detection for multi-label data streams. Artif. Intell. Rev. 2023, 56, 2401–2434. [Google Scholar]
  40. Soltanzadeh, P.; Hashemzadeh, M. RCSMOTE: Range-Controlled synthetic minority over-sampling technique for handling the class imbalance problem. Inf. Sci. 2021, 542, 92–111. [Google Scholar]
  41. Yu, H.; Sun, C.; Yang, X.; Yang, W.; Shen, J.; Qi, Y. ODOC-ELM: Optimal decision outputs compensation-based extreme learning machine for classifying imbalanced data. Knowl.-Based Syst. 2016, 92, 55–70. [Google Scholar]
  42. Khan, A.A.; Chaudhari, O.; Chandra, R. A review of ensemble learning and data augmentation models for class imbalanced problems: Combination, implementation and evaluation. Expert Syst. Appl. 2024, 244, 122778. [Google Scholar]
  43. Abedin, M.Z.; Guotai, C.; Hajek, P.; Zhang, T. Combining weighted SMOTE with ensemble learning for the class-imbalanced prediction of small business credit risk. Complex Intell. Syst. 2023, 9, 3559–3579. [Google Scholar]
  44. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  45. Huang, G.; Huang, G.B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef]
  46. Feng, B.; Gu, Y.; Yu, H.; Yang, X.; Gao, S. DME: An adaptive and just-in-time weighted ensemble learning method for classifying block-based concept drift steam. IEEE Access 2022, 10, 120578–120591. [Google Scholar] [CrossRef]
  47. Yu, H.; Sun, C.; Yang, X.; Zheng, S.; Wang, Q.; Xi, X. LW-ELM: A fast and flexible cost-sensitive learning framework for classifying imbalanced data. IEEE Access 2018, 6, 28488–28500. [Google Scholar] [CrossRef]
  48. Cheng, K.; Gao, S.; Dong, W.; Yang, X.; Wang, Q.; Yu, H. Boosting label weighted extreme learning machine for classifying multi-label imbalanced data. Neurocomputing 2020, 403, 360–370. [Google Scholar] [CrossRef]
  49. García, S.; Fernández, A.; Luengo, J.; Herrera, F. Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Inf. Sci. 2010, 180, 2044–2064. [Google Scholar] [CrossRef]
  50. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. An example of a multi-label image.
Figure 1. An example of a multi-label image.
Entropy 27 00363 g001
Figure 2. The basic framework of stream-based active learning.
Figure 2. The basic framework of stream-based active learning.
Entropy 27 00363 g002
Figure 3. Best performance variance of the various algorithms with the variance in the parameter θ on the various datasets.
Figure 3. Best performance variance of the various algorithms with the variance in the parameter θ on the various datasets.
Entropy 27 00363 g003aEntropy 27 00363 g003b
Figure 4. Learning curve of each active learning algorithm on each dataset.
Figure 4. Learning curve of each active learning algorithm on each dataset.
Entropy 27 00363 g004aEntropy 27 00363 g004b
Figure 5. CD diagrams at the standard level of significance α = 0.05 in terms of both the Macro-F1 and Micro-F1 metrics.
Figure 5. CD diagrams at the standard level of significance α = 0.05 in terms of both the Macro-F1 and Micro-F1 metrics.
Entropy 27 00363 g005
Table 1. Datasets used in this study.
Table 1. Datasets used in this study.
DatasetDomain#Instances#Features#LabelsLCBlock Size
FlagsImage1941973.39220
CHD_49Medicine5554962.58050
EmotionsMusic5937261.86850
MedicalText9781449451.245100
Water qualityChemistry106016145.073100
ImageImage200029451.236200
SceneImage240729461.074200
YeastBiology2417103144.237200
EukaryotePseAACBiology7766440221.146500
YelpText10,81067151.638800
#Instances, #Features, and #Labels denote the number of instances, features, and labels included in the corresponding dataset.
Table 2. ALC values and average ranking of several comparable algorithms based on Macro-F1 metric.
Table 2. ALC values and average ranking of several comparable algorithms based on Macro-F1 metric.
DatasetRandomEMALLCICVIRSOMAL
Flags0.5751 ± 0.01590.5762 ± 0.00640.5856 ± 0.00610.5663 ± 0.00890.5774 ± 0.0120
CHD_490.4346 ± 0.00670.4176 ± 0.00920.4447 ± 0.00650.4349 ± 0.00660.4503 ± 0.0040
Emotions0.5732 ± 0.01000.5729 ± 0.00610.5642 ± 0.00700.5771 ± 0.01060.5841 ± 0.0067
Medical0.2661 ± 0.00760.2449 ± 0.00540.2315 ± 0.00780.2875 ± 0.00300.2865 ± 0.0031
Water-quality0.5497 ± 0.00230.5365 ± 0.00150.5426 ± 0.00260.5408 ± 0.00240.5504 ± 0.0025
Image0.5636 ± 0.00590.5649 ± 0.00700.5648 ± 0.00420.5678 ± 0.00510.5708 ± 0.0032
Scene0.6832 ± 0.00370.6738 ± 0.00330.6782 ± 0.00580.6876 ± 0.00310.6940 ± 0.0034
Yeast0.4318 ± 0.00510.4247 ± 0.00330.4309 ± 0.00430.4321 ± 0.00240.4397 ± 0.0027
EukaryotePseAAC0.1189 ± 0.00250.1143 ± 0.00080.1153 ± 0.00050.1167 ± 0.00180.1193 ± 0.0013
Yelp0.4716 ± 0.00250.4722 ± 0.00410.4796 ± 0.00210.4771 ± 0.00210.4726 ± 0.0013
Average ranking3.44.33.32.61.4
All results are presented in the form of mean ± standard deviation, and the best one for each dataset is highlighted in bold.
Table 3. ALC values and average rankings of several comparable algorithms based on the Micro-F1 metric.
Table 3. ALC values and average rankings of several comparable algorithms based on the Micro-F1 metric.
Dataset Random EMAL LCI CVIRS OMAL
Flags0.6414 ± 0.01030.6413 ± 0.00580.6507 ± 0.00540.6341 ± 0.00710.6561 ± 0.0041
CHD_490.5493 ± 0.00760.5337 ± 0.01100.5630 ± 0.00690.5453 ± 0.00780.5704 ± 0.0041
Emotions0.5752 ± 0.01040.5759 ± 0.00570.5718 ± 0.00720.5819 ± 0.01050.5866 ± 0.0061
Medical0.7501 ± 0.00610.7336 ± 0.00340.7230 ± 0.00770.7587 ± 0.00410.7602 ± 0.0022
Water quality0.5666 ± 0.00210.5527 ± 0.00120.5577 ± 0.00210.5551 ± 0.00190.5727 ± 0.0022
Image0.5625 ± 0.00580.5635 ± 0.00650.5633 ± 0.00460.5683 ± 0.00460.5695 ± 0.0033
Scene0.6644 ± 0.00380.6586 ± 0.00320.6649 ± 0.00600.6684 ± 0.00330.6735 ± 0.0036
Yeast0.5577 ± 0.00530.5541 ± 0.00230.5683 ± 0.00220.5588 ± 0.00260.5619 ± 0.0031
EukaryotePseAAC0.3442 ± 0.00870.3464 ± 0.00140.3471 ± 0.00420.3631 ± 0.00340.3663 ± 0.0029
Yelp0.4861 ± 0.00260.4896 ± 0.00360.4955 ± 0.00210.4929 ± 0.00130.4881 ± 0.0010
Average ranking3.84.12.92.81.4
All results are presented in the form of mean ± standard deviation, and the best one for each dataset is highlighted in bold.
Table 4. Ablation experiment on the querying strategy, in which the best results have been highlighted in bold.
Table 4. Ablation experiment on the querying strategy, in which the best results have been highlighted in bold.
DatasetMacro-F1Micro-F1
UncertaintyDiversityBothUncertaintyDiversityBoth
Flags0.5519 ± 0.01130.5762 ± 0.00470.5774 ± 0.00120.6397 ± 0.00600.6445 ± 0.00230.6561 ± 0.0041
CHD_490.4494 ± 0.00440.4448 ± 0.00400.4503 ± 0.00400.5678 ± 0.00610.5601 ± 0.00390.5704 ± 0.0041
Emotions0.5888 ± 0.00730.5808 ± 0.00560.5841 ± 0.00670.5927 ± 0.00740.5834 ± 0.00530.5866 ± 0.0061
Medical0.2852 ± 0.00450.2777 ± 0.00240.2865 ± 0.00310.7591 ± 0.00430.7511 ± 0.00290.7602 ± 0.0022
Water quality0.5557 ± 0.00160.5399 ± 0.00150.5504 ± 0.00250.5772 ± 0.00110.5581 ± 0.0010.5727 ± 0.0022
Image0.5642 ± 0.00460.5656 ± 0.00450.5708 ± 0.00320.5629 ± 0.00410.5643 ± 0.00460.5695 ± 0.0033
Scene0.6922 ± 0.00310.6786 ± 0.00270.6940 ± 0.00340.6731 ± 0.00290.6582 ± 0.0030.6735 ± 0.0036
Yeast0.4351 ± 0.00290.4350 ± 0.00130.4397 ± 0.00270.5585 ± 0.00210.5497 ± 0.00090.5619 ± 0.0031
EukaryotePseAAC0.1276 ± 0.00140.1124 ± 0.00080.1193 ± 0.00130.3461 ± 0.00210.3642 ± 0.00120.3663 ± 0.0029
Yelp0.4678 ± 0.00130.4718 ± 0.00140.4726 ± 0.00130.4838 ± 0.00130.4866 ± 0.00200.4881 ± 0.0010
Table 5. Ablation experiment on considering the class imbalance issue, in which the best results have been highlighted in bold.
Table 5. Ablation experiment on considering the class imbalance issue, in which the best results have been highlighted in bold.
DatasetMacro-F1Micro-F1
ELMWELMELMWELM
Flags0.5766 ± 0.01590.5774 ± 0.00120.6682 ± 0.01180.6561 ± 0.0041
CHD_490.4454 ± 0.00580.4503 ± 0.00400.5855 ± 0.00740.5704 ± 0.0041
Emotions0.5742 ± 0.00790.5841 ± 0.00670.5765 ± 0.00810.5866 ± 0.0061
Medical0.2805 ± 0.00520.2865 ± 0.00310.7194 ± 0.00550.7602 ± 0.0022
Water quality0.4822 ± 0.00970.5504 ± 0.00250.5415 ± 0.00570.5727 ± 0.0022
Image0.4461 ± 0.00600.5708 ± 0.00320.4498 ± 0.00590.5695 ± 0.0033
Scene0.5965 ± 0.00360.6940 ± 0.00340.5769 ± 0.00410.6735 ± 0.0036
Yeast0.4023 ± 0.00310.4397 ± 0.00270.5601 ± 0.00400.5619 ± 0.0031
EukaryotePseAAC0.0877 ± 0.00310.1193 ± 0.00130.2735 ± 0.00160.3663 ± 0.0029
Yelp0.4771 ± 0.00370.4726 ± 0.00130.4871 ± 0.00340.4881 ± 0.0010
Table 6. Ablation experiment on the label ranking strategy in CC, in which the best results have been highlighted in bold..
Table 6. Ablation experiment on the label ranking strategy in CC, in which the best results have been highlighted in bold..
DatasetMacro-F1Micro-F1
RandomLD3RandomLD3
Flags0.5764 ± 0.00770.5774 ± 0.00120.6553 ± 0.00710.6561 ± 0.0041
CHD_490.4526 ± 0.00460.4503 ± 0.00400.5692 ± 0.00850.5704 ± 0.0041
Emotions0.5791 ± 0.00930.5841 ± 0.00670.5802 ± 0.01210.5866 ± 0.0061
Medical0.2809 ± 0.00270.2865 ± 0.00310.7547 ± 0.00410.7602 ± 0.0022
Water quality0.5487 ± 0.00770.5504 ± 0.00250.5782 ± 0.00540.5727 ± 0.0022
Image0.5671 ± 0.00600.5708 ± 0.00320.5662 ± 0.00530.5695 ± 0.0033
Scene0.6916 ± 0.00330.6940 ± 0.00340.6748 ± 0.00390.6735 ± 0.0036
Yeast0.4386 ± 0.00600.4397 ± 0.00270.5562 ± 0.01790.5619 ± 0.0031
EukaryotePseAAC0.1132 ± 0.00510.1193 ± 0.00130.2855 ± 0.03190.3663 ± 0.0029
Yelp0.4712 ± 0.00290.4726 ± 0.00130.4841 ± 0.00420.4881 ± 0.0010
Table 7. Average running time of each active learning algorithm throughout all rounds (seconds).
Table 7. Average running time of each active learning algorithm throughout all rounds (seconds).
DatasetRandomEMALLCICVIRSOMAL
Flags0.04090.04640.04830.31520.0443
CHD_490.16460.17580.16652.29450.1687
Emotions0.40590.42600.42712.79050.4117
Medical8.06108.35678.183614.71878.4605
Water quality0.60950.63420.70398.24120.6541
Image1.02621.06041.113426.64371.1383
Scene3.26783.53513.544240.35673.5444
Yeast8.46418.85648.734746.96479.1096
EukaryotePseAAC109.7874110.9367111.7936483.9314109.8256
Yelp34.095734.772434.1290780.740835.3902
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fang, Q.; Xiang, C.; Duan, J.; Soufiyan, B.; Shao, C.; Yang, X.; Xu, S.; Yu, H. OMAL: A Multi-Label Active Learning Approach from Data Streams. Entropy 2025, 27, 363. https://doi.org/10.3390/e27040363

AMA Style

Fang Q, Xiang C, Duan J, Soufiyan B, Shao C, Yang X, Xu S, Yu H. OMAL: A Multi-Label Active Learning Approach from Data Streams. Entropy. 2025; 27(4):363. https://doi.org/10.3390/e27040363

Chicago/Turabian Style

Fang, Qiao, Chen Xiang, Jicong Duan, Benallal Soufiyan, Changbin Shao, Xibei Yang, Sen Xu, and Hualong Yu. 2025. "OMAL: A Multi-Label Active Learning Approach from Data Streams" Entropy 27, no. 4: 363. https://doi.org/10.3390/e27040363

APA Style

Fang, Q., Xiang, C., Duan, J., Soufiyan, B., Shao, C., Yang, X., Xu, S., & Yu, H. (2025). OMAL: A Multi-Label Active Learning Approach from Data Streams. Entropy, 27(4), 363. https://doi.org/10.3390/e27040363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop