Next Article in Journal
A No-Reference Quality Assessment Method for Screen Content Images Based on Human Visual Perception Characteristics
Previous Article in Journal
G-Band Broad Bandwidth Extended Interaction Klystron with Traveling-Wave Output Structure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Self-Paced Learning Based on Adaptive Mixture Weighting

1
Xi’an Precision Machinery Research Institute, Xi’an 710077, China
2
School of Marine Science and Technologyg, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(19), 3154; https://doi.org/10.3390/electronics11193154
Submission received: 8 September 2022 / Revised: 20 September 2022 / Accepted: 23 September 2022 / Published: 1 October 2022
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
Self-paced learning (SPL) is a learning mechanism inspired by human and animal learning processes that gives variable weights to samples, gradually introducing simple to more complicated samples into the learning set as the “age” parameter increases. To regulate the learning process, a self-paced weighting regularization term with an “age” parameter is introduced to the learning function. Several self-paced weighting methods have been proposed, and different regularization terms might result in varied learning performance. However, on the one hand, it is difficult to select a suitable weighting method for SPL. On the other hand, it is challenging to determine the “age” parameter, and it is easy for SPL to obtain poor results as the “age” of the model increases. To solve the aforementioned difficulties, an ensemble SPL approach with an adaptive mixture weighting mechanism is proposed in this study. First, as the “age” parameter increases, a set of base classifiers is collected to produce a new data set, which is used to learn the second-level classifier. Then, the ensemble model is used to generate the final output to avoid the selection of the optimal “age” parameter. An adaptive mixture weighting method is designed to reduce the dependence of parameters on human experience. The previous methods find it difficult to determine the “age” parameters or self-paced parameters. In this paper, these parameters can be adjusted adaptively during the learning process. In comparison with the previous SPL techniques, the proposed method achieves the best results in 27 of the 32 datasets in the experiments with the adaptive parameters. The statistical tests are carried out to show that the proposed method is superior to other state-of-the-art algorithms.

1. Introduction

Self-paced learning (SPL) is a machine learning mechanism that is based on human and animal learning processes [1]. SPL can be traced back to curriculum learning [2,3,4]. Curriculum learning and SPL both use the method of gradually introducing simple to more complex examples into the training set for learning. In curriculum learning, curriculum knowledge must be specified in advance as priori information, and then, the information remains intact. However, based on what the learners have already learned, the SPL mechanism can dynamically generate curriculum information [5,6]. Self-paced learning has been widely employed in pattern recognition and computer vision since its inception, such as multimedia event detection [7,8], matrix factorization [9,10], long-term tracking [11,12], image segmentation and classification [13,14,15], co-saliency detection [16], and so on. Many studies have found that SPL can help models avoid local minima and improve their generalization ability [17,18]. SPL can also lessen the impact of noise and anomalous data in the data set, resulting in more reliable learning.
In recent years, self-paced learning has been widely developed [5,8,19,20,21,22]. Kumar et al. [1] gave the basic form of SPL and solved it in an iterative manner. Then, SPL was used to for ranking problems in the field of zero-example multimedia search [7]. In [19], a self-paced learning method with diversity (SPLD) is proposed to consider sample diversity. This method improves the diversity of the current training set by selecting easy samples from different categories, thereby avoiding the imbalance of sample categories caused by different attributes. Li et al. [8] proposed a multi-objective self-paced learning method (MOSPL), by introducing multi-objective optimization technology to optimize the loss function and the self-paced regular term at the same time. This method obtains the complete solution path of the self-paced learning problem when the self-paced parameters have no prior knowledge and are difficult to obtain. In [9], matrix factorization problems were solved by self-paced learning, which achieves good results in the related applications, such as background subtraction and structure from motion problems. In order to solve the problem of prior knowledge, Jiang et al. [5] combined curriculum learning and self-paced learning, and they proposed self-paced curriculum learning (SPCL), which also considered the prior knowledge before training and the curriculum information during the learning process. Liang et al. [20] proposed active self-paced learning (ASPL), which combines active learning and self-paced learning to make the classifier have better classification in the case of less labeled sample data. Following [20], a new Deep Self-paced Active Learning (DSAL) was proposed in [6] by combining ASPL with deep learning techniques. ASPL realizes incremental learning under weak manual labeling on face recognition problems. In [22], to deal with spatial prior information and make the suggested technique noise-resistant, the spatial information of the samples is taken into account in the self-paced learning process. Furthermore, some intelligent optimization algorithms were used to solve self-paced problem to find the optimal solution path [10]. The design of SPL regularization terms has become a key research direction in recent years [1,23,24]. SPL reflects the difficulty of the sample by assigning a weight to each sample, and the SPL regularization term determines the selection of samples and the calculation of weights. Finally, the curriculum information and the model are jointly learned by iteratively updating the weight variables and model parameters. Kumar et al. [1] control the optimization process of SPL by introducing the l 1 -norm of the weight variable to the objective function as the regularization term. The weight variable only uses two value states “0” or “1” to indicate whether the sample is selected, which is the hard weighting form of SPL. In order to measure the importance of samples on the basis of selected samples, Jiang et al. [19] proposed several SPL regularization terms, such as linear soft weighting, logarithmic soft weighting, and mixture weighting, which are used to solve learning problems in different data and problem contexts. When the sample loss value is small, compared with other weighting forms, the mixture weighting method can assign a large weight to the sample. Therefore, the mixture weighting method favors the sample with a small loss value to a certain extent, and it is widely used in the real-world applications [19,22].
As indicated in [23,25], the current SPL methods increase the “age” parameter to gradually involve more samples into training. However, it is difficult to select a suitable “age” parameter to achieve the best results. This paper aims to learn an ensemble classifier to avoid the selection of the “age” parameter, which is able to improve classification performance by combining multiple base classifiers [26,27,28]. The current ensemble learning methods can be roughly divided into three categories: bagging, boosting, and stacking. Bagging employs numerous versions of a training set generated via a random draw with the substitution of n samples, where n is the size of the original training set [29,30]. A new model is trained using each of these data sets, and voting is used to aggregate the model outputs into a single output. Boosting is an ensemble approach for turning a collection of weak classifiers into a strong classifier, which allows the predictors to be learned sequentially [31,32]. Stacking is divided into two levels: basic models (first level) and stacking model (second level) [33,34]. To learn from a dataset, the base learner employs a variety of models to learn from the current data set. To produce a new data set, the outputs of each model are collected. The stacking model then uses that data set to generate the final output.
In the current SPL regimes, on the one hand, it is a difficult task to determine the optimal “age” parameter. On the other hand, the form of the soft part of the mixture weighting is too simple, and only one form of soft weighting is considered. It is an urgent task to extend the mixture weighting method to adaptively penalize the objective function. In this paper, in order to address the above issues, we propose an ensemble self-paced learning method based on an adaptive mixture weighting strategy (ESPL). As the “age” parameter increases, a set of weak classifiers is constructed to formulate a new training set. Then, these classifiers are stacked to acquire the second-level classifier. Next, an adaptive mixture weighting regularization term is proposed to penalize the SPL objective function. Experiments on classification problems have demonstrated that the proposed ESPL is robust to the selection of “age” parameter and can achieve the best results in terms of statistical tests.
This paper proposes an ensemble self-paced learning (ESPL) method based on an adaptive mixture weighting scheme. On the one hand, this paper aims to avoid the selection of “age” parameters by stacking multiple base self-paced learning models with different “age” parameters. On the other hand, this paper aims to propose a new adaptive mixture weighting strategy to penalize the SPL objective function adaptively. The remainder of this paper is structured as follows: The background knowledge of self-paced learning is described in Section 2. Section 3 introduces the proposed ensemble SPL method in detail. Experimental studies are shown in Section 4. Finally, concluding remarks are given in Section 5.

2. Background

In this section, the background of self-paced learning is described. Some well-known self-paced weighing regularizers are shown in this section. Finally, the background of ensemble stacking learning is stated.

2.1. Self-Paced Learning

In this paper, D = { ( x 1 , y 1 ) , , ( x n , y n ) } denotes the training data, where x i R m represents the ith observed sample, and y i is its corresponding label. To indicate the ease of a sample, SPL uses the weight variable s . The latent weight variable s and the model parameters w are jointly learned by minimizing:
min w , s E ( w , s ; ϕ ) = i = 1 n s i L ( y i , g ( x i , w ) ) + f ( s ; ϕ ) ,
where s [ 0 , 1 ] n and f ( s ; ϕ ) is the SPL regularization term. ϕ is called the “age” parameter. In general, an increasing pace sequence is given in advance to define this “age” parameter. The cost between the ground truth label y i and the estimated label g ( x i , w ) is calculated using the loss function L ( y i , g ( x i , w ) ) . The algorithm of SPL is shown in Algorithm 1. Alternative convex search is commonly used to solve the aforementioned equation. The variables are divided into two discontinuous blocks in this biconvex optimization iterative approach. Each iteration optimizes one set of variables while leaving the others alone.
When w is fixed, as shown in Algorithm 1, we can derive closed-form optimum solutions using various SPL regularization terms. The weighted training loss and the negative l 1 -norm regularizer s 1 = i = 1 n s i are minimized with an increasing pace parameter in the original SPL implementation [1], where s i is a binary variable. Then, some more efficient SPL regularization terms have been proposed as follows [7]:
Linear soft weighting regularizer: The losses are penalized linearly in this method, which can be expressed as the following function:
f ( s ; ϕ ) = ϕ 1 2 s 2 2 i = 1 n s i ,
where ϕ > 0 . The SPL model’s closed-form solutions under the linear soft weighting scheme are as follows:
s i * = 1 L i ϕ L i < ϕ 0 L i ϕ .
Logarithmic soft weighting regularizer: This approach can be represented as the following function to logarithmically discriminate samples with respect to losses:
f ( s ; ϕ ) = i = 1 n ζ s i ζ s i log ζ ,
where ζ = 1 ϕ and 0 < ϕ < 1 . Then, using the logarithmic soft weighting method, the closed-form optimal solution is expressed as:
s i * = 1 log ζ log ( L i + ζ ) L i < ϕ 0 L i ϕ .
Mixture weighting regularizer: The mixture approach is a combination of the “soft” and the “hard” schemes, as represented by the function:
f ( s ; ϕ ) = ζ i = 1 n log s i + 1 ϕ 1 ζ ,
where ζ = ϕ 1 ϕ 2 ϕ 1 ϕ 2 and ϕ 1 > ϕ 2 > 0 . The following is the closed-form optimal solution:
s i * = 1 L i ϕ 2 0 L i ϕ 1 ζ L i ζ ϕ 1 otherwise .
When s is set, the available off-the-shelf learning methods can be used to find the best w * . The “age” of the SPL model to control the learning pace is the parameter ϕ . We gradually raise the “age” parameter during the evolution of SPL to learn new samples.
Algorithm 1: Algorithm of SPL
Input: The training dataset D .
Output: Model parameter w .
Electronics 11 03154 i001

2.2. Ensemble Stacking Learning

Several classifiers are used in ensemble approaches. Each of these classifiers creates a model based on the data. Finally, learning, fusion, or voting among various classifiers is completed for classification, and the class with the highest votes is the final class. Stacking is a very powerful ensemble learning strategy, in addition to the bagging and boosting methods discussed above. The basic idea behind stacking is to improve the generalization performance by enhancing the heterogeneity of the basis classifiers and combining the outputs predicted by the base classifiers using the second-level classifier.
Figure 1 illustrates the general stacking method, which consists of three major steps. First, the first-level classifiers are created using the original training data set. There are several strategies for learning base classifiers. For example, we can use the bootstrap sampling technique to learn independent classifiers. To build various base classifiers, we can tune parameters in a learning algorithm. Furthermore, we can use a variety of classification and/or sampling approaches to generate base classifiers. Second, a new data set is established from the output of the base classifiers. The labels of the first-level classifiers are considered new features in this case, but the original class labels are retained as labels in the new generated data set. Finally, using the newly produced data set, we create a second-level classifier. The second-level classifier could be learned using any learning approach.

3. Methodology

In this section, we describe the proposed ensemble self-paced learning (ESPL) method in detail. First, the framework of ESPL is introduced. Next, a new adaptive mixture weighting method is designed. Finally, the Majorization Minimization algorithm is used to optimize the objective function of the proposed method.

3.1. Ensemble Self-Paced Learning

Ensemble learning brings together several learners to address the same problem. Traditional machine learning algorithms learn a decision function from training data and use it to generate predictions on the test set, whereas ensemble learning methods attempt to construct a set of decision functions and combine them. As described above, during the evolution of SPL, a set of classifiers is generated based on the increasing “age” parameter. The previous studies have indicated that it is difficult to determine this “age” parameter, since SPL can easily obtain bad results as the “age” parameter increases. In order to avoid the above issues, this paper proposes an ensemble self-paced learning method to ensemble multiple SPL models generated in the evolution of SPL.
To train independent base learners, bagging uses bootstrap sampling and takes the majority as the resulting prediction [28]. Every round, boosting adjusts the weight distribution, learns the appropriate base classifiers, and then merges them based on their accuracy. Stacking, unlike the above two methods, trains a high-level classifier on top of the underlying classifiers. In this paper, stacking is considered to learn a second-level classifier, which is used to combine multiple SPL models generated in the evolution of SPL. As shown in Figure 2, as the “age” parameter increases, more samples are included into the training process during the evolution of SPL. At a certain “age” parameter, the samples with weight s > 0 constitute the current training set to establish a first-level classifier. Then, a new training set is constructed based on the output of first-level classifiers. Finally, a second-level classifier is created using the newly formed data set. The algorithm of the proposed ESPL method is shown in Algorithm 2.
Algorithm 2: Algorithm of Ensemble SPL
Input: The training dataset D = { x i , y i } i = 1 n .
Output: An ensemble clssifier H.
Electronics 11 03154 i002

3.2. Adaptive Mixture Weighting

The previous mixture weighting method used a combination of hard weights and logarithmic soft weights, and the form was relatively simple. This paper aims to extend this hybrid weight to a more general form and to be able to adaptively select hard and soft weight combination forms. In order to solve the above problems, this paper proposes an adaptive mixture weighting method, as shown in the following formula:
f ( s ; { α 1 , α 2 , β } ) = i = 1 n α 1 β s i β ( α 1 + α 2 ) s i ,
where α 1 > 0 , α 2 > 0 , and β > 1 .
According to the above formula, the objective function of SPL based on the adaptive mixture weighting method can be obtained as:
min w , s E ( w , s ) = i = 1 n s i L y i , g ( x i , w ) + i = 1 n α 1 β s i β ( α 1 + α 2 ) s i .
The Hessian matrix H of f ( s ; { α 1 , α 2 , β } ) with respect to s can be represented as:
α 1 β 1 s 1 β 2 0 0 0 α 1 β 1 s 2 β 2 0 0 0 α 1 β 1 s n β 2
where α 1 > 0 and β > 1 .
s i ( 0 , 1 ] is the weight of the ith sample for the sampled involved in training. It can be observed that α 1 β 1 s i β 2 is greater than zero for every i = 1 , 2 , , n . Therefore, H is a symmetric diagonal matrix. The diagonal elements are the eigenvalues of a diagonal matrix. The main diagonal entries are obviously greater than zero. Then, the Hessian of f is positive definite. Therefore, the proposed regularizer is convex.

3.3. Optimization and Analysis

The Majorization Minimization (MM) algorithm can be used to optimize the objective function of the proposed method. By iterating the two phases of identifying the optimization function and minimizing it, the technique seeks to reduce a complex optimization issue to a simple one. Assume that s * ( α ; L ) is the optimal weights and the integrative function of s * ( α ; L ) is
F α ( L ) = 0 L s * ( α ; L ) d L .
Q α ( w | w * ) denotes a tractable surrogate for F α ( L ( w ) ) . In [17], Meng et al. have indicated that
Q α ( i ) ( w | w * ) = F α ( L i ( w * ) ) + s i * ( α ; L i ( w * ) ) ( L i ( w ) L i ( w * ) ) ,
and then we have:
i = 1 n F α ( L i ( w ) ) i = 1 n Q α ( i ) ( w | w * ) .
In this paper, w t is used to represent the model parameters in the tth iteration of the MM method.

3.3.1. Majorization Step

To acquire each Q ϕ ( i ) ( w | w t ) , we can compute s i * ( ϕ ; L i ( w t ) ) by solving the following problem:
s i * ( ϕ ; L i ( w t ) ) = arg min s i [ 0 , 1 ] s i L i ( w t ) + f ( s i ; ϕ ) .
where the form of f ( s i ; ϕ ) has been given in the Equation (8). As described above, the proposed f ( s i ; ϕ ) in (8) is convex. Then, we have
E s i = L i + α 1 s i β 1 ( α 1 + α 2 ) = 0 ,
and
s i β 1 = α 1 + α 2 L i α 1 .
Finally, the closed-form optimal solution is
v n * = 1 , L i α 2 0 , L i α 1 + α 2 α 1 + α 2 L i α 1 1 β 1 , otherwise .

3.3.2. Minimization Step

In this minimization step, we have
w t + 1 = arg min w i = 1 n v i L i ( w t ) .
Here, support vector machine (SVM) is used as the learning model for self-paced learning. The proposed method combined with support vector machine can be expressed as the following form:
min w , s 1 2 w 2 2 + C i = 1 n s i L i + f ( s ; λ ) s . t . i , y i w T ϕ x i + b 1 L i , L i 0 y { 1 , + 1 } n , s [ 0 , 1 ] n ,
where L i is the standard hinge loss:
L i = max 0 , 1 y i · w T ϕ x i + b .
ϕ x i is a feature mapping function. The regularization parameter C ( C > 0 ) is a trade-off between the hinge loss and the margin.

3.3.3. Parameter Analysis

The parameters in self-paced learning are usually set based on human experience. The above-mentioned mixture weighting method includes three types of parameters: α 1 , α 2 , and β . As shown in Equation (17), the parameter α 2 controls the number of samples with v = 1 . Therefore, we let α 2 = 0.2 · α 1 to reduce the number of parameters. Let L s represent the loss L in ascending order and N t be the number of samples involved in training. Then, we have
α 1 + α 2 = 1.2 α 1 = L s N t .
For the polynomial parameter β , researchers can assign specific values to obtain a SPL regularization term suitable for the specific problem. This paper expects that it can dynamically adjust according to the loss of samples in each iteration. Therefore, we have
β = tan 1 N t 2 n + 1 · π 2 .
In the early stage of SPL, simple samples are selected into the data set, and they are given larger weights. Therefore, this stage β should have large values. In the iterative process, more complex samples are selected into the training set; then, β should have a small value.

3.4. Analysis

As shown in Figure 3, the flowchart of the proposed method is given. The results of each iteration process constitute a set of first-level classifiers. Then, the prediction results of the first-level classifiers on the training set and the test set constitute the training matrix and the test matrix, and the training matrix and the test matrix in each iteration process together to learn a second-level classifier. In the previous studies on SPL weighting methods, some soft weighting methods have been proposed to penalize the loss. It has been demonstrated that the mixture weighting method is able to receive better generalization ability than the hard and sort weighting methods. However, the form of the soft part of the mixture weighting is too simple, and only one form of soft weighting is considered. In formula (8), in comparison with the original mixture weighting method, an additional parameter β is involved in the objective function. As shown in formula (17), the parameter β is able to control the form of the soft part of the proposed mixture weighting method. Researchers can assign specific values to obtain an SPL regularization term suitable for the specific problem. Furthermore, this paper expects that it can dynamically adjust according to the loss of samples in each iteration, which has been shown in formula (22).

4. Experimental Studies

In the experiments, a set of data sets from the KEEL repository are employed [35], which are widely used for comparative studies in several literatures [36,37]. Using the k-fold cross-validation technique, these datasets accessed on 8 September 2022 (https://sci2s.ugr.es/keel/imbalanced.php) were divided into five training/test subsets. A subset of the data was utilized as a test set, while the rest was used as a training set in cross-validation. As a result, the cross-validation procedure is carried out five times. The five-fold results can then be averaged to provide a single estimate. The data sets utilized in this study are listed in Table 1. We show the number of instances and attributes for each data collection separately.
For comparing several algorithms across multiple data sets, we use the Friedman and Nemenyi tests [38]. The Friedman statistic value is computed as follows
F F = ( J 1 ) X F 2 J ( K 1 ) X F 2 ,
where
X F 2 = 12 J K ( K + 1 ) j R j 2 K ( K + 1 ) 2 4 .
K is the number of algorithms being compared, J is the number of datasets and R j is the average rank of the jth algorithm. The null hypothesis is that there is no difference in the average ranks. If the Friedman statistic is larger than the critical value of F distribution with K 1 and ( K 1 ) × ( J 1 ) degrees of freedom for 0.05, we can reject the null hypothesis and perform Nemenyi tests to further analyze the average ranks. The Nemenyi statistic is obtained as follows
C D = q 0.05 K ( K + 1 ) 6 J ,
where q 0.05 is the Studentized range statistic divided by 2 . Let d i j denote the difference between the average ranks of the ith algorithm and jth algorithm. If d i j > C D , the difference between the algorithms is significant.

4.1. Test of the Proposed Ensemble Strategy

Since the growth of the “age” parameter, the present SPL approaches may generate unsatisfactory results, as discussed in the preceding sections. In most cases, the results obtained by self-paced learning will get better first and then worse. As the “age” parameter increases, some hard samples involved in the training will cause the model performance to degrade. However, due to a lack of prior information, determining the “age” parameter is problematic. The proposed ESPL method considers the models generated in the SPL implementation as the first-level classifiers to avoid the section of the “age” parameter.
Figure 4 shows the convergence curves obtained by SPL and the proposed ESPL methods. It can be observed that the index values obtained by SPL and ESPL gradually increase in the first few iterations. After several iterations, the results obtained by the original SPL will gradually deteriorate. As shown in Figure 4, the blue curves obtained by SPL first rise and then fall. However, the proposed ESPL method ensembles the previous models and is robust to the “age” parameter.

4.2. Test of the Proposed Adaptive Mixture Weighting Method

As described in Section 2, the hard weighting scheme, the linear soft weighting scheme, the logarithmic soft weighting scheme, and the mixture weighting scheme have been widely used in the current SPL-based applications. The previous mixture weighting method used a simple form, which is a combination of hard weights and logarithmic soft weights. The proposed adaptive mixture weighting method broadens the scope of this hybrid weight and enables the adaptive selection of hard and soft weight combinations.
The values in terms of Accuracy, F-measure, and Gmean obtained by different weighting schemes are reported in Figure 5, Figure 6 and Figure 7. Note that all these algorithms with different weighting schemes utilize the proposed ensemble strategy and only ESPL employs the proposed adaptive mixture weighting method. It is obvious that the proposed ESPL with adaptive mixture weighting achieves the best results in these dataset.
We conduct a statistical test to validate statistical differences between the results generated by different weighting methods within the proposed ESPL framework in order to further evaluate the performance of the proposed adaptive mixture weighting method. To examine the results in terms of Accuracy, F-measure, and Gmean, we apply the Friedman and Nemenyi tests [38]. The comparison involves five ( K = 5 ) algorithms with 32 ( J = 32 ) data sets. The p values of Friedman statistic in terms of Accuracy, F-measure, and Gmean in the experiments are 5.611 × 10 10 , 1.43 × 10 15 , and 1.005 × 10 27 , respectively. As a result, with a 95% confidence level, we find that the evaluated algorithms are significantly different in the three criteria. The results of the Nemenyi test are shown in Figure 8. In terms of Accuracy and Gmean, it can be seen that the proposed adaptive mixture weighting scheme and other weighting algorithms are significantly different. Although there are no significant differences in terms of F-measure between the adaptive mixture weighting scheme and the hard weighting scheme, the proposed adaptive mixture weighting system has the lowest average rank and has been proved to be superior in experiments.

4.3. Experimental Results

We compare ESPL to the Decision Tree classifier [39], linear discriminant analysis (LDA) [40], and AdaBoost [41] algorithms to validate its performance. One of the most extensively used and practical approaches for inductive inference is decision tree learning. First, the decision tree with C4.5 was selected as the competing algorithm in the experiments. It is a method for approximating discrete-valued functions that can learn disjunctive expressions and is robust to noisy data [39]. In comparison to the spread within each class, LDA finds the subspace in which the data between various classes are most spread out, which makes LDA suitable for classification problems. Furthermore, AdaBoost is also chosen as the competing algorithm in the following experiments. AdaBoost is an ensemble method that has been successfully applied to a variety of applications while avoiding the concerns of overfitting. In the next iteration, patterns that are incorrectly identified are given a higher weight. Patterns around the decision boundary are typically more difficult to categorize and, as a result, receive high weights after several iterations.
The results on KEEL datasets obtained by four methods with respect to Accuracy are shown in Table 2. It is obvious that the proposed ESPL algorithm achieves the best results in 23 of the 32 datasets. The decision tree algorithm achieves the best results on certain data such as glass0, ecoli2, glass-0-4_vs_5, vowel0, and page_blocks_1_3_vs_4. The LDA method achieves the best results in eight of the 32 datasets. On some data, such as new_thyroid1, ecoli-0-3-4-6_vs_5, ecoli-0-2-6-7_vs_3-5, ecoli-0-1-4-7_vs_5-6, shuttle_c2_vs_c4, and dermatology-6-5-5tst, the AdaBoost algorithm produces the best results. The results on KEEL datasets obtained by four methods with respect to F-measure are shown in Table 3. The decision tree algorithm achieves the best results on some data sets such as glass0, ecoli2, yeast-0-3-5-9_vs_7-8, glass-0-4_vs_5, vowel0, and page_blocks_1_3_vs_4. The LDA algorithm achieves the best results on yeast_2_vs_4, ecoli-0-1_vs_5, and led7digit-0-2-4-5-6-7-8-9_vs_1. For new_thyroid1, ecoli-0-3-4-6_vs_5, ecoli-0-2-6-7_vs_3-5, shuttle_c2_vs_c4, and dermatology-6-5-5tst, AdaBoost obtains the best results. The results on KEEL datasets obtained by four methods with respect to Gmean are shown in Table 4. It is obvious that the proposed ESPL algorithm achieves the best results in 27 of the 32 datasets. From Table 2, Table 3 and Table 4, it can be observed that the proposed ESPL algorithm achieves satisfactory results in terms of Accuracy, F-measure, and Gmean.
We conduct a statistical test to validate statistical differences between the results provided by Decision Tree, LDA, AdaBoost, and ESPL in order to further evaluate the performance of the proposed ESPL technique. In terms of Accuracy, F-measure, and Gmean, the Friedman and Nemenyi tests are used to examine the data. The comparison involves five ( K = 4 ) algorithms with 32 ( J = 32 ) datasets. The p values of Friedman statistic in terms of Accuracy, F-measure, and Gmean in the experiments are 3.616 × 10 8 , 1.633 × 10 10 , and 7.082 × 10 14 , respectively. As a result, with a 95% confidence level, we find that the evaluated algorithms are significantly different in the three indexes. The results of the Nemenyi test are show in Figure 9. It can be observed that from a statistical point of view, the performance of our method is superior to the comparison algorithms.

5. Concluding Remarks

The mixture weighting schemes have the characteristics of both hard and soft weights, so they are widely used in many self-paced learning problems. This paper has proposed an ensemble self-paced learning method and designed an adaptive mixture weighting strategy. As the “age” parameter increases, a group of first-level classifiers is generated for each “age” parameter, which is then utilized to establish the new data set and train the second-level classifier. To avoid selecting the best “age” parameter, the ensemble model is utilized to generate the final output. Furthermore, an adaptive mixture weighting approach has been proposed, which is incorporated into the learning objective of the SVM algorithm. Then, the proposed method is solved by the MM algorithm, and the closed-form solutions of the proposed optimization problem are derived. Experimental results on classification challenges illustrate the effectiveness of the proposed ensemble SPL method. Statistical tests have been conducted to demonstrate that the suggested method outperforms other state-of-the-art algorithms. In the future, we hope to investigate the deep learning-based algorithms under the proposed ensemble self-paced learning framework and explore more case studies to discuss the proposed method.

Author Contributions

Conceptualization, L.L. and Z.W.; methodology, L.L.; software, L.L.; validation, J.B., X.Y. and Y.Y.; formal analysis, J.Z.; investigation, J.Z.; resources, J.B.; data curation, Z.W.; writing—original draft preparation, L.L.; writing—review and editing, Z.W.; visualization, J.B.; supervision, J.Z.; project administration, Y.Y.; funding acquisition, L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Guangdong Province grant number [2214050001305], the Science Foundation of Zhanjiang City grant number [211207157080994], the National Natural Science Foundation of China grant number [11904290] and the sixth Youth Talent Promotion Project of China Association for Science and Technology grant number [2020QNRC002].

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kumar, M.P.; Packer, B.; Koller, D. Self-paced learning for latent variable models. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 6–9 December 2010; pp. 1189–1197. [Google Scholar]
  2. Bengio, Y.; Louradour, J.; Collobert, R.; Weston, J. Curriculum learning. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 41–48. [Google Scholar]
  3. Bae, J.; Kim, T.; Lee, W.; Shim, I. Curriculum learning for vehicle lateral stability estimations. IEEE Access 2021, 9, 89249–89262. [Google Scholar] [CrossRef]
  4. Wang, X.; Chen, Y.; Zhu, W. A survey on curriculum learning. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4555–4576. [Google Scholar] [CrossRef] [PubMed]
  5. Jiang, L.; Meng, D.; Zhao, Q.; Shan, S.; Hauptmann, A.G. Self-paced curriculum learning. In Proceedings of the 29th AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 2694–2700. [Google Scholar]
  6. Wang, W.; Feng, R.; Chen, J.; Lu, Y.; Chen, T.; Yu, H.; Chen, D.Z.; Wu, J. Nodule-plus R-CNN and deep self-paced active learning for 3D instance segmentation of pulmonary nodules. IEEE Access 2019, 7, 128796–128805. [Google Scholar] [CrossRef]
  7. Jiang, L.; Meng, D.; Mitamura, T.; Hauptmann, A.G. Easy samples first: Self-paced reranking for zero-example multimedia search. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 547–556. [Google Scholar]
  8. Li, H.; Gong, M.; Meng, D.; Miao, Q. Multi-objective self-paced learning. In Proceedings of the 30th AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1802–1808. [Google Scholar]
  9. Zhao, Q.; Meng, D.; Jiang, L.; Xie, Q.; Xu, Z.; Hauptmann, A.G. Self-paced learning for matrix factorization. In Proceedings of the 29th AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 3196–3202. [Google Scholar]
  10. Li, H.; Gong, M.; Wang, C.; Miao, Q. Pareto self-paced learning based on differential evolution. IEEE Trans. Cybern. 2021, 51, 4187–4200. [Google Scholar] [CrossRef]
  11. Supancic, J.S.; Ramanan, D. Self-paced learning for long-term tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2379–2386. [Google Scholar]
  12. Ge, D.; Song, J.; Qi, Y.; Wang, C.; Miao, Q. Self-paced dense connectivity learning for visual tracking. IEEE Access 2019, 7, 37181–37191. [Google Scholar] [CrossRef]
  13. Peng, J.; Wang, P.; Desrosiers, C.; Pedersoli, M. Self-paced contrastive learning for semi-supervised medical image segmentation with meta-labels. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 7–10 December 2021; Volume 34. [Google Scholar]
  14. Li, H.; Li, J.; Zhao, Y.; Gong, M.; Zhang, Y.; Liu, T. Cost-sensitive self-paced learning with adaptive regularization for classification of image time series. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2021, 14, 11713–11727. [Google Scholar] [CrossRef]
  15. Pan, X.; Wei, D.; Zhao, Y.; Ma, M.; Wang, H. Self-paced learning with diversity for medical image segmentation by using the query-by-committee and dynamic clustering techniques. IEEE Access 2020, 9, 9834–9844. [Google Scholar] [CrossRef]
  16. Zhang, D.; Meng, D.; Li, C.; Jiang, L.; Zhao, Q.; Han, J. A self-paced multiple-instance learning framework for co-saliency detection. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 594–602. [Google Scholar]
  17. Meng, D.; Zhao, Q.; Jiang, L. A theoretical understanding of self-paced learning. Inform. Sci. 2017, 414, 319–328. [Google Scholar] [CrossRef]
  18. Klink, P.; Abdulsamad, H.; Belousov, B.; D’Eramo, C.; Peters, J.; Pajarinen, J. A probabilistic interpretation of self-paced learning with applications to reinforcement learning. J. Mach. Learn. Res. 2021, 22, 1–52. [Google Scholar]
  19. Jiang, L.; Meng, D.; Yu, S.-I.; Lan, Z.; Shan, S.; Hauptmann, A. Self-paced learning with diversity. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 27, pp. 2078–2086. [Google Scholar]
  20. Lin, L.; Wang, K.; Meng, D.; Zuo, W.; Zhang, L. Active self-paced learning for cost-effective and progressive face identification. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 7–19. [Google Scholar] [CrossRef] [Green Version]
  21. Sun, L.; Zhou, Y. FSPMTL: Flexible self-paced multi-task learning. IEEE Access 2020, 8, 132012–132020. [Google Scholar] [CrossRef]
  22. Li, H.; Gong, M.; Zhang, M.; Wu, Y. Spatially self-paced convolutional networks for change detection in heterogeneous images. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2021, 14, 4966–4979. [Google Scholar] [CrossRef]
  23. Gong, M.; Li, H.; Meng, D.; Miao, Q.; Liu, J. Decomposition-based evolutionary multi-objective optimization to self-paced learning. IEEE Trans. Evol. Comput. 2019, 23, 288–302. [Google Scholar] [CrossRef]
  24. Huang, Z.; Ren, Y.; Liu, W.; Pu, X. Self-paced multi-view clustering via a novel soft weighted regularizer. IEEE Access 2019, 7, 168629–168636. [Google Scholar] [CrossRef]
  25. Gu, B.; Zhai, Z.; Li, X.; Huang, H. Finding age path of self-paced learning. In Proceedings of the 2021 IEEE International Conference on Data Mining, Auckland, New Zealand, 7–10 December 2021; pp. 151–160. [Google Scholar]
  26. Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
  27. Galar, M.; Fernandez, A.; Barrenechea, E.; Bustince, H.; Herrera, F. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Trans. Syst. Man Cybern. C Appl. Rev. 2011, 42, 463–484. [Google Scholar] [CrossRef]
  28. Khoshgoftaar, T.M.; Van Hulse, J.; Napolitano, A. Comparing boosting and bagging techniques with noisy and imbalanced data. IEEE Trans. Syst. Man Cybern. A Syst. Humans 2010, 41, 552–568. [Google Scholar] [CrossRef]
  29. Martínez-Muñoz, G.; Suárez, A. Out-of-bag estimation of the optimal sample size in bagging. Pattern Recogn. 2010, 43, 143–152. [Google Scholar] [CrossRef] [Green Version]
  30. Mordelet, F.; Vert, J.-P. A bagging SVM to learn from positive and unlabeled examples. Pattern Recogn. Lett. 2014, 37, 201–209. [Google Scholar] [CrossRef] [Green Version]
  31. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 2–18 December 2018; Volume 31, pp. 6638–6648. [Google Scholar]
  32. Shen, C.; Li, H. On the dual formulation of boosting algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2216–2231. [Google Scholar] [CrossRef] [Green Version]
  33. Akyol, K. Stacking ensemble based deep neural networks modeling for effective epileptic seizure detection. Expert Syst. Appl. 2020, 148, 113239. [Google Scholar] [CrossRef]
  34. Jiang, W.; Chen, Z.; Xiang, Y.; Shao, D.; Ma, L.; Zhang, J. SSEM: A novel self-adaptive stacking ensemble model for classification. IEEE Access 2019, 7, 120337–120349. [Google Scholar] [CrossRef]
  35. Alcalá-Fdez, J.; Fernández, A.; Luengo, J.; Derrac, J.; García, S.; Sánchez, L.; Herrera, F. KEEL data-mining software tool: Data set repository, integration of algorithms and experimental analysis framework. J. Mult. Valued Log. Soft Comput. 2011, 17, 255–287. [Google Scholar]
  36. Ramírez-Gallego, S.; García, S.; Benítez, J.M.; Herrera, F. Multivariate discretization based on evolutionary cut points selection for classification. IEEE Trans. Cybern. 2016, 46, 595–608. [Google Scholar] [CrossRef]
  37. Rosales-Pérez, A.; García, S.; Gonzalez, J.A.; Coello, C.A.C.; Herrera, F. An evolutionary multiobjective model and instance selection for support vector machines with Pareto-based ensembles. IEEE Trans. Evol. Comput. 2017, 21, 863–877. [Google Scholar] [CrossRef]
  38. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  39. Polat, K.; Güneş, S. A novel hybrid intelligent method based on C4. 5 decision tree classifier and one-against-all approach for multi-class classification problems. Expert Syst. Appl. 2009, 36, 1587–1592. [Google Scholar] [CrossRef]
  40. Ioffe, S. Probabilistic linear discriminant analysis. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 531–542. [Google Scholar]
  41. Rätsch, G.; Onoda, T.; Müller, K.-R. Soft margins for adaboost. Mach. Learn. 2001, 42, 287–320. [Google Scholar] [CrossRef]
Figure 1. Ensemble learning.
Figure 1. Ensemble learning.
Electronics 11 03154 g001
Figure 2. Illustration of ensemble self-paced learning.
Figure 2. Illustration of ensemble self-paced learning.
Electronics 11 03154 g002
Figure 3. Flowchart of the proposed ESPL method.
Figure 3. Flowchart of the proposed ESPL method.
Electronics 11 03154 g003
Figure 4. The convergence curves obtained by SPL and the proposed ESPL methods. The three rows shows the results on the glass0, vehicle3, and glass-0-4_vs_5 data sets, respectively.
Figure 4. The convergence curves obtained by SPL and the proposed ESPL methods. The three rows shows the results on the glass0, vehicle3, and glass-0-4_vs_5 data sets, respectively.
Electronics 11 03154 g004
Figure 5. The results in terms of Accuracy obtained by different weighting schemes.
Figure 5. The results in terms of Accuracy obtained by different weighting schemes.
Electronics 11 03154 g005
Figure 6. The results in terms of F-measure obtained by different weighting schemes.
Figure 6. The results in terms of F-measure obtained by different weighting schemes.
Electronics 11 03154 g006
Figure 7. The results in terms of Gmean obtained by different weighting schemes.
Figure 7. The results in terms of Gmean obtained by different weighting schemes.
Electronics 11 03154 g007
Figure 8. The graphs of the results of Nemenyi test obtained by different weighting schemes in terms of Accuracy, F-measure, and Gmean.
Figure 8. The graphs of the results of Nemenyi test obtained by different weighting schemes in terms of Accuracy, F-measure, and Gmean.
Electronics 11 03154 g008aElectronics 11 03154 g008b
Figure 9. The graphs of the results of Nemenyi test obtained by Decision Tree, LDA, AdaBoost, and ESPL in terms of Accuracy, F-measure, and Gmean.
Figure 9. The graphs of the results of Nemenyi test obtained by Decision Tree, LDA, AdaBoost, and ESPL in terms of Accuracy, F-measure, and Gmean.
Electronics 11 03154 g009
Table 1. Description of the data sets used in this paper. For each data set, we present the number of attributes and the number of instances.
Table 1. Description of the data sets used in this paper. For each data set, we present the number of attributes and the number of instances.
DatasetInstancesAttributes
wisconsin68310
pima7689
glass021410
haberman3064
vehicle284619
vehicle184619
vehicle384619
vehicle084619
ecoli13368
new_thyroid12156
new_thyroid22156
ecoli23368
glass621410
ecoli-0-3-4_vs_52008
ecoli-0-2-3-4_vs_52028
yeast-0-3-5-9_vs_7-85069
ecoli-0-4-6_vs_52037
ecoli-0-3-4-6_vs_52058
ecoli-0-3-4-7_vs_5-62578
ecoli-0-1_vs_2-3-52448
yeast_2_vs_45149
glass-0-4_vs_59210
ecoli-0-2-6-7_vs_3-52248
vowel098814
ecoli-0-6-7_vs_52207
ecoli-0-1-4-7_vs_2-3-5-63368
ecoli-0-1_vs_52407
led7digit-0-2-4-5-6-7-8-9_vs_14438
ecoli-0-1-4-7_vs_5-63327
page_blocks_1_3_vs_447211
shuttle_c2_vs_c412910
dermatology-6-5-5tst35835
Table 2. Results on KEEL datasets obtained by four algorithms in terms of Accuracy. The best results are marked in bold.
Table 2. Results on KEEL datasets obtained by four algorithms in terms of Accuracy. The best results are marked in bold.
DatasetDecisionLDAADAESPL
TreeBoost
wisconsin0.94590.95900.96780.9693
pima0.72270.76820.76430.7786
glass00.81750.76170.76660.7897
haberman0.66010.74180.73830.7548
vehicle20.95270.96810.96340.9693
vehicle10.74590.80020.80610.8144
vehicle30.73640.79550.79080.8014
vehicle00.94560.95150.96690.9728
ecoli10.89300.88410.87490.8958
new_thyroid10.94420.94880.99070.9907
new_thyroid20.96280.93950.99070.9953
ecoli20.91670.91370.90180.9167
glass60.93010.96280.95790.9627
ecoli-0-3-4_vs_50.93000.94000.95000.9550
ecoli-0-2-3-40.92560.95540.96040.9654
_vs_5
yeast-0-3-5-90.88930.91700.75460.9170
_vs_7-8
ecoli-0-4-6_vs_50.94600.96570.96560.9656
ecoli-0-3-4-60.95120.95120.96590.9659
_vs_5
ecoli-0-3-4-70.93780.94560.96890.9728
_vs_5-6
ecoli-0-1_vs0.94250.95890.95480.9589
_2-3-5
yeast_2_vs_40.94360.95720.94170.9533
glass-0-4_vs_50.98890.95610.96730.9784
ecoli-0-2-6-70.93760.95980.96870.9687
_vs_3-5
vowel00.98580.94940.97270.9727
ecoli-0-6-7_vs_50.95000.95910.95910.9682
ecoli-0-1-4-70.94940.95830.96430.9673
_vs_2-3-5-6
ecoli-0-1_vs_50.95830.96670.95000.9542
led7digit-0-2-40.95940.96170.95940.9616
-5-6-7-8-9_vs_1
ecoli-0-1-4-70.94880.96100.96690.9699
_vs_5-6
page_blocks0.99150.96400.95340.9746
_1_3_vs_4
shuttle_c2_vs_c40.99230.97661.00001.0000
dermatology0.99440.99720.99720.9972
-6-5-5tst
Table 3. Results on KEEL datasets obtained by four algorithms in terms of F-measure. The best results are marked in bold.
Table 3. Results on KEEL datasets obtained by four algorithms in terms of F-measure. The best results are marked in bold.
DatasetDecisionLDAADAESPL
TreeBoost
wisconsin0.92270.93990.95440.9565
pima0.58520.62770.63990.6712
glass00.70370.58020.62910.6758
haberman0.28730.23030.48360.5319
vehicle20.90900.93790.92970.9407
vehicle10.49200.57210.59050.6567
vehicle30.47280.54750.57240.6318
vehicle00.88660.89710.93210.9436
ecoli10.76810.75940.74300.7694
new_thyroid10.83640.80300.97130.9713
new_thyroid20.88570.75470.97130.9867
ecoli20.70000.71390.68370.7200
glass60.73450.85760.85080.8718
ecoli-0-3-4_vs_50.62540.70210.75240.7835
ecoli-0-2-3-40.61760.77540.78570.8143
_vs_5
yeast-0-3-5-90.42980.33600.28660.3612
_vs_7-8
ecoli-0-4-6_vs_50.68810.82140.80630.8254
ecoli-0-3-4-60.73810.75290.80400.8040
_vs_5
ecoli-0-3-4-70.62640.68640.83110.8529
_vs_5-6
ecoli-0-1_vs0.66330.73960.70640.7521
_2-3-5
yeast_2_vs_40.71290.75180.63420.7292
glass-0-4_vs_50.93330.76670.80760.8933
ecoli-0-2-6-70.64940.76000.79440.7944
_vs_3-5
vowel00.91780.72650.83830.8425
ecoli-0-6-7_vs_50.75430.76110.73480.8262
ecoli-0-1-4-70.67980.71880.78000.7921
_vs_2-3-5-6
ecoli-0-1_vs_50.72440.79440.70240.7310
led7digit-0-2-40.76140.78890.76140.7766
-5-6-7-8-9_vs_1
ecoli-0-1-4-70.63400.68100.77210.7988
_vs_5-6
page_blocks0.93810.59580.66830.8426
_1_3_vs_4
shuttle_c2_vs_c40.93330.80001.00001.0000
dermatology0.94920.97140.97780.9778
-6-5-5tst
Table 4. Results on KEEL datasets obtained by four algorithms in terms of Gmean. The best results are marked in bold.
Table 4. Results on KEEL datasets obtained by four algorithms in terms of Gmean. The best results are marked in bold.
DatasetDecisionLDAADAESPL
TreeBoost
wisconsin0.94160.94860.96650.9686
pima0.67410.70210.71370.7452
glass00.77310.66810.71290.7523
haberman0.45390.36730.61840.6718
vehicle20.94050.95720.95800.9626
vehicle10.63340.68160.72020.7931
vehicle30.62200.66490.69970.7662
vehicle00.93040.93250.95900.9663
ecoli10.83810.84280.82630.8885
new_thyroid10.90420.82260.98240.9824
new_thyroid20.93030.78340.98240.9972
ecoli20.78820.81200.85220.8834
glass60.83420.90220.91760.9495
ecoli-0-3-4_vs_50.76520.83920.85570.9143
ecoli-0-2-3-40.75500.87580.85430.8811
_vs_5
yeast-0-3-5-90.62580.45950.47430.6092
_vs_7-8
ecoli-0-4-6_vs_50.78430.88160.85650.9221
ecoli-0-3-4-60.82060.84930.85600.9047
_vs_5
ecoli-0-3-4-70.73310.78490.88600.9069
_vs_5-6
ecoli-0-1_vs0.77740.82520.81220.8470
_2-3-5
yeast_2_vs_40.83320.80980.79470.8801
glass-0-4_vs_50.99400.87240.94100.9743
ecoli-0-2-6-70.76750.82840.83850.8947
_vs_3-5
vowel00.93930.84770.92160.9599
ecoli-0-6-7_vs_50.64420.85200.81310.8863
ecoli-0-1-4-70.79580.78440.84570.8803
_vs_2-3-5-6
ecoli-0-1_vs_50.83620.87960.83560.8785
led7digit-0-2-40.87040.89930.87040.8917
-5-6-7-8-9_vs_1
ecoli-0-1-4-70.75470.75530.86070.9021
_vs_5-6
page_blocks0.99550.68590.92350.9863
_1_3_vs_4
shuttle_c2_vs_c40.94140.93321.00001.0000
dermatology0.97170.97320.99850.9985
-6-5-5tst
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, L.; Wang, Z.; Bai, J.; Yang, X.; Yang, Y.; Zhou, J. Ensemble Self-Paced Learning Based on Adaptive Mixture Weighting. Electronics 2022, 11, 3154. https://doi.org/10.3390/electronics11193154

AMA Style

Liu L, Wang Z, Bai J, Yang X, Yang Y, Zhou J. Ensemble Self-Paced Learning Based on Adaptive Mixture Weighting. Electronics. 2022; 11(19):3154. https://doi.org/10.3390/electronics11193154

Chicago/Turabian Style

Liu, Liwen, Zhong Wang, Jianbin Bai, Xiangfeng Yang, Yunchuan Yang, and Jianbo Zhou. 2022. "Ensemble Self-Paced Learning Based on Adaptive Mixture Weighting" Electronics 11, no. 19: 3154. https://doi.org/10.3390/electronics11193154

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop