Next Article in Journal
Cost-Effective Data Acquisition Systems for Advanced Structural Health Monitoring
Previous Article in Journal
A Retinal Vessel Segmentation Method Based on the Sharpness-Aware Minimization Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Broad Learning System under Label Noise: A Novel Reweighting Framework with Logarithm Kernel and Mixture Autoencoder

by
Jiuru Shen
,
Huimin Zhao
* and
Wu Deng
College of Electronic Information and Automation, Civil Aviation University of China, Tianjin 300300, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(13), 4268; https://doi.org/10.3390/s24134268
Submission received: 7 June 2024 / Revised: 27 June 2024 / Accepted: 29 June 2024 / Published: 30 June 2024

Abstract

:
The Broad Learning System (BLS) has demonstrated strong performance across a variety of problems. However, BLS based on the Minimum Mean Square Error (MMSE) criterion is highly sensitive to label noise. To enhance the robustness of BLS in environments with label noise, a function called Logarithm Kernel (LK) is designed to reweight the samples for outputting weights during the training of BLS in order to construct a Logarithm Kernel-based BLS (L-BLS) in this paper. Additionally, for image databases with numerous features, a Mixture Autoencoder (MAE) is designed to construct more representative feature nodes of BLS in complex label noise environments. For the MAE, two corresponding versions of BLS, MAEBLS, and L-MAEBLS were also developed. The extensive experiments validate the robustness and effectiveness of the proposed L-BLS, and MAE can provide more representative feature nodes for the corresponding version of BLS.

1. Introduction

In recent years, deep learning, as a research hotspot in the field of artificial intelligence, has achieved significant breakthroughs and widespread applications in various domains [1,2]. Although deep learning boasts powerful learning capabilities, its training process is extremely time-consuming due to its complex model structure and the iterative adjustment of numerous hyperparameters [3,4].
Against this backdrop, Chen et al. [5] proposed a neural network learning framework called the Broad Learning System (BLS) in 2018. It originated from the Random Vector Functional Link Neural Network [6]. BLS extends the width of neural networks to fit the data and has advantages such as a simple network structure and fewer model parameters [7,8,9]. BLS does not require iterative adjustment and has an extremely fast learning speed. Since its proposal, BLS has attracted widespread attention and has rapidly developed in both theoretical and applied research fields.
Researchers have further explored the excellent performance of BLS, leading to further developments in many challenging areas [10,11,12,13]. Chen et al. [14] developed a cascade structure, a recurrent structure, and a wide and deep combination structure based on BLS. Ye et al. [15] implemented image denoising by using cascaded BLS that connects feature mapping node groups and enhancement node groups in sequence. For chaotic time series prediction.
Furthermore, BLS demonstrates excellent performance in semi-supervised learning. For instance, Zhao et al. [16] developed a semi-supervised BLS (SS-BLS) that utilizes manifold regularization to obtain pseudo-labels for unknown data, thereby expanding the training set. Huang et al. [17] designed a generalized model with manifold regularization sparse features (BLS-MS), which utilizes latent information hidden in unlabeled samples for representation learning. Bhosle et al. [18] designed a deep learning CNN model for the recognition of the Devanagari digit. Deng et al. [19] developed a cluster clustering-based modular with a deep neural network for predicting flight arrival time.
The aforementioned BLS methods and their variants have shown good generalization performance and practical effectiveness, but these results are typically obtained under conditions where training samples are free from label noise. In recent decades, the rapid advancement of sensor technology has greatly raised the bar for sensor complexity, accuracy, and efficiency. The operating status of a sensor is critical to its health and reliability. However, during the data annotation process, factors such as sensor aging, human annotation errors, and environmental issues can degrade data quality, leading to incorrect labels, posing significant challenges to sensor operation. Addressing label noise is critical to improving the accuracy and reliability of sensor prediction and health management systems. The broad learning system employs ridge regression for output weights; although BLS uses one-hot encoding for sample labels, it is still highly sensitive to labeling noisy data. This has been elaborated on in detail [20]. Therefore, in the label noise environment, we urgently need a more robust BLS method.
In response to the above problems, in previous studies, Lu et al. [21] proposed a robust weighted least squares support vector machine. Jin et al. [22] combined the L1 and L2 norms and effectively optimized the BLS model using the augmented Lagrange multiplier method to enhance model robustness. Chu et al. [23] introduced several weighted penalty factors to enhance model performance, resulting in the design of a weighted BLS (WBLS). Liu et al. [24] utilized modal regression instead of least squares measurement to train a generalized network, which generated a diagonal weight matrix through optimization strategies for stronger noise penalties. A notable work in improving the robustness of BLS is the robust manifold BLS (RM-BLS) [25]. By introducing manifold embedding and random perturbation approximation, it aims to achieve robust mapping characteristics in certain specialized applications, such as predicting chaotic time series with noise. Zheng et al. [26] employed the maximum correlation entropy criterion to train network connection coefficients, achieving outstanding regression and classification performance in noisy environments. The graph regularized BLS introduced in [27] incorporates a target function based on maximum likelihood estimation, assigning appropriate weights to each sample for classifying data with label noise. In addition, other methods have also been proposed in recent years [28,29,30,31,32,33,34,35,36].
The above methods have achieved certain results for BLS, but there is still room for improvement. In this paper, the main contributions are listed as follows:
(1)
Based on the Gaussian kernel, a novel function called Logarithm Kernel (LK) is constructed to effectively enhance the robustness of BLS under label noise conditions.
(2)
A new robust broad learning reweighting framework (L-BLS) is designed by adding Logarithm Kernel (LK) to BLS for training output weights in order to significantly mitigate or eliminate the impact of label noise on BLS.
(3)
A Mixture Autoencoder (MAE) is constructed to create more representative feature nodes in BLS for image databases with numerous features.
(4)
MAEBLS and L-MAEBLS were developed to improve the expressiveness of the enhanced feature nodes of the corresponding BLS version and reduce their sensitivity to label noise.
The rest of this paper is organized as follows: Section 2 provides a brief review of BLS. In Section 3, we elaborate on the proposed methods. Section 4 presents the robustness analysis of the Logarithm Kernel (LK). Section 5 offers extensive experimental results and discussions. Finally, Section 6 concludes the paper.

2. Review of the Broad Learning System

To maintain consistency in this text, variables are represented in italics, vectors are represented in bold lowercase letters (e.g., x), and matrices are represented in bold uppercase letters (e.g., X). More generally, for any matrix A R m × n , a i R 1 × n , a j R m × 1 , a i , j represents the element of matrix A at the ith row and jth column. The capital letter T indicates the transpose operator. Let us denote X = x 1 ; x 2 ; ; x N R N × M as the training sample matrix, where N represents the number of samples, M represents the feature dimension, and Y = [ y 1 ; y 2 ; ; y N ] R N × C as the training label matrix, where C means the number of classes.
The introduction of the Broad Learning System (BLS) provides an effective and efficient learning framework for classification and regression problems. The main advantage of BLS is its ability to map input data into a series of random feature spaces and determine output weights through optimized least squares. When new nodes or inputs arise, the training process can be extended to an incremental learning mode.
Here, a set of labeled training samples is provided to the BLS, for example, {X,Y}. Assume that BLS has n groups of feature mapping nodes; each group contains k feature mapping nodes. The feature node can be represented by Z i .
Z i = ϕ i X W e i + β e i R N × k , i = 1 , 2 , 3 , , n
where ϕ i ( · ) represents the activation function.   W e i R M × k   a n d   β e i R N × k both represent randomly generated weight matrices and biases.
Connecting the n groups of feature mapping nodes together forms the feature mapping layer. Here, the feature mapping layer is represented by Z n .
Z n = Z 1 ; Z 2 ; Z 3 ; Z n R N × n k
The feature mapping layer Z n is passed to enhancement nodes to construct the feature enhancement layer. The enhancement node is represented by H j .
H j = ξ j Z n W h j + β h j R N × b , i = 1 , 2 , 3 , , n
where ξ j i ( · ) represents the activation function. W h j R n k × b   a n d   β h j R N × b both represent randomly generated weight matrices and biases.
The enhancement nodes are connected together to form the enhancement layer, resulting in H m . The enhancement layer has m groups of enhancement nodes; each group contains b nodes.
H m = H 1 ; H 2 ; H 3 ; ; H m R N × m b
Connecting the feature mapping layer and the enhancement layer together, the complete state matrix A can be represented as follows:
A = Z n H m
This optimization problem can be formulated as finding the regularized least squares solution of Y = AW. Hence, the BLS can be trained as follows:
min W Y A W 2 2 + λ W 2 2
W = A + Y
where λ is the regularization parameter. W R ( n k + m b ) × C , A + = ( A T A + λ I ) 1 A T is the Moore–Penrose pseudoinverse, and more details can be found in [5].

3. The Proposed Method

To enhance the robustness of the BLS framework, this section delves into a comprehensive exploration of the proposed Logarithm Kernel (LK) function. We elucidate its integration into the width learning system for training output weights, thereby yielding the reweighting framework L-BLS. At the end of this section, we will introduce a Mixture Autoencoder (MAE) that helps BLS build more representative feature nodes in a noisy label environment and two BLS versions of MAEBLS and L-MAEBLS for image data sets.

3.1. Logarithm Kernel (LK)

The correlation entropy between two random variables X and Y serves as a correlation measure in kernel space, as elucidated in [37,38,39,40].
V X , Y = E Φ X , Φ Y H = Φ X , Φ Y H d F X Y X , Y
where E · represents the expectation operator, F X Y ( x , y ) represents the joint distribution function, Φ ( x ) , Φ ( y ) H = κ ( x , y ) , κ ( x , y ) is the Mercer kernel [21] controlled by the kernel size σ . We can clearly obtain the following equation:
V X , Y = E κ X , Y
In the field of machine learning, E ( κ ( X , Y ) ) is commonly used to estimate the degree of correlation between the true values and the predicted values. In complex, noisy environments, it is important to accurately quantify the correlation between the true values and the predicted values and enhance the robustness of parameter estimation.
Non-second-order statistical measures can be elegantly defined as second-order measures in kernel space. Non-second-order statistical measures can be elegantly defined as second-order measures in kernel space. According to Property 3 provided in [39], the correlation entropy has the potential to capture the second-order and higher order statistical characteristics of the error when using a Gaussian kernel. With an appropriate kernel size setting, the second-order statistical characteristics of the error can dominate, which makes entropy-based optimization criteria a suitable choice in label noise environments as well. This paper presents a new function called the Logarithm Kernel (LK).
ζ X , Y = l o g 1 + k e r n e l X , Y # 10
where k e r n e l ( X , Y ) is equal to κ ( X , Y ) . So unless specified, this paper employs the Gaussian kernel as the function for the Logarithm Kernel. So Equation (10) can be deduced to Equation (11).
ζ y i , y i = l o g 1 + k e r n e l G y i , y i
where k e r n e l G ( y i , y i )   i s   e q u a l   t o   e x p ( | | y i y i | | 2 2 σ 2 ) and σ > 0 , y i represents the predicted label, and y i represents the real label.
Clearly, we can obtain V ( y i , y i ) = E ( k e r n e l G ( y i , y i ) ) . By using the Taylor series expansion to V ( y i , y i ) , we have
V y i , y i = n = 1 1 n 2 n n ! E ( y i , y i ) 2 n σ 2 n
V ( y i , y i ) can be regarded as the weighted sum of all the even moments of y i and y i , with the weights of the second and higher order moments controlled by the kernel size σ. The kernel size determines the weight of the individual, even at moments when calculating the weighted sum. When σ increases, it means that we consider the difference between y i and y i more smoothly, and the weight of the higher order even moments decreases. This property allows the correlated entropy to better adapt to different data situations and improves its robustness in practical applications. For finite sample data, V ( y i , y i ) can be approximated as follows:
V ~ y i , y i = 1 N n = 1 N k e r n e l G y i , y i
Similarly, each sample is introduced into ζ ( X , Y ) = l o g ( 1 + k e r n e l ( X , Y ) ) and changes the following equation:
ζ y i , y i = 1 N n = 1 N ( 1 + l o g k e r n e l G y i , y i

3.2. The Proposed L-BLS

Based on the above motivations and to address the worse performance of ridge regression in complex label noise environments., we transform LK to a BLS-based reweighting framework, L-BLS, for training output weights.
Similar to BLS, the state matrix A can be constructed to build the feature mapping via (1)–(5). Therefore, to enhance the robustness of the BLS, incorporating the LK proposed in Section A and sample reweighting techniques into the optimization model of the BLS can be represented as follows:
argmax W i = 1 N log 1 + exp y i a i W 2 2 σ 2 λ 2 W 2 2
where a i R L represents the feature of the i-th sample among N data samples. According to our experience, if we need to find W satisfying the condition in Equation (15), we should first calculate the gradient of this equation. To simplify the calculation, let us set ϕ ( W ) = i = 1 N l o g ( 1 + e x p ( | | y i a i W | | 2 2 σ 2 ) ) λ 2 | | W | | 2 2 . By taking the derivative of the function ϕ ( W ) with respect to W, we obtain the following equation:
ϕ W W = i N   exp y i a i W 2 2 σ 2 a i T a i W y i λ W σ 2 1 + exp y i a i W 2 2 σ 2
To simplify subsequent calculations, ϕ ( W ) W can be represented as follows:
ϕ W W = 1 σ 2 A T U A W Y λ W
where U is equal to
  exp y 1 a 1 W 2 2 σ 2 1 + exp y 1 a 1 W 2 2 σ 2   exp y N a N W 2 2 σ 2 1 + exp y N a N W 2 2 σ 2
Letting the partial derivative ϕ ( W ) W be zero, W can be expressed as follows:
ϕ W W = 1 σ 2 A T U A W Y λ W = 0
W = A T U A + λ σ 2 E 1 A T U Y
Observing W and U, we can see that the right-hand side can be viewed as a function of W, so we can further rewrite them as follows:
W = f W
F W = A T U A + λ σ 2 E 1 A T U Y
A more intuitive L-BLS framework and detailed algorithm are shown in Figure 1 and Algorithm 1. In Algorithm 1, the weight of each sample is continuously updated by iteratively optimizing U and W. Reasonable weights are assigned to samples with the correct labels. The effectiveness of the sample weighting framework can be more intuitively presented in Section 5.
Algorithm 1 Proposed L-BLS Sample Reweighting Framework
Input: Training samples X with corrupted label Y
O u t p u t : O u t p u t   w e i g h t   m a t r i x   W
1: Initialization: parameters, regularization parameter λ , kernel   size   σ , termination tolerance η , and maximum iteration number T;
2: for i = 1, 2, …, n do
3:     Random   W e i , β e i ;
4:     Calculate   Z i = [ ϕ i ( X W e i + β e i ) ] ;
5: end for
6: Set   the   feature   mapping   group   Z n = [ Z 1 ; Z 2 ; Z 3 ; Z n ] ;
7: for j = 1, 2, …, m do
8:      Random   W h j , β h j ;
9:      Calculate   H j = [ ξ j ( Z n W h j + β h j ) ] ;
10: end for
11: Set   the   enhancement   node   group   H m = [ H 1 ; H 2 ; H 3 ; ; H m ] ;
12: Set the state matrix A according to Equation (5);
13: for t = 1, 2, …, T do
14:    Compute U according to Equation (18);
15:    Update W according to Equation (20);
16:     if   | | W t W t 1 | | 2 2 < η then
17:        break;
18:    end if
19: end for

3.3. Mixture Autoencoder

In this section, for image databases with numerous features, a novel Mixture Autoencoder (MAE) is constructed by utilizing convolutional autoencoder techniques [41] and the advantages of variational autoencoders [42]. The purpose of MAE is to help BLS and its variants create more representative feature nodes under label noise conditions [43,44].
The encoder network consists of convolutional layers and nonlinear activation functions, which are used to extract features from the input images. The convolutional layers employ multiple convolutional kernels to perform convolutional operations on the input images, capturing the spatial structure and features of the images. Nonlinear activation functions are then applied to introduce nonlinear transformations, enhancing the expressive power and robustness of the features. The structure of the Mixture Autoencoder is shown in Figure 2.
To achieve efficient latent representation learning, a reparameterization of E n c o d e r O u t p u t is performed, and reparameterization factors a and b are introduced, ensuring a + b = 1, as in Equation (23), to allocate reasonable weights. This enhances the model’s robustness while ensuring feature integrity.
D e c o d e r I n p u t = a E + b V a r i a n c e E × E + M e a n E
where DecoderInput represents the input to the decoder, E represents EncoderOutput, V a r i a n c e ( x ) denotes the standard deviation of x, and M e a n ( x ) denotes the mean of the vector x.
The decoder network is responsible for mapping the learned low-dimensional latent representation back to the original image space to verify whether the encoder part provides representative feature nodes. To simplify the model in the decoding part of MAE, a fully connected approach is adopted. Nonlinear activation functions are applied layer by layer to decode the features, reconstructing the input image through inverse transformations to recover the information of the input image to the maximum extent.
The number of convolutional layers and the size of each convolution kernel determine the receptive field and feature extraction capability of the model. More convolutional layers and larger convolution kernels can not only capture more complex and high-order features but also increase computational complexity. So, in MAE, we designed three convolutional layers with 5 × 5 convolution kernels.
On the other hand, we encode by retaining 40%–50% of the original data features in order to provide the most representative features of the image to BLS during MAE training. The entire decoder design gradually maps low-dimensional features back to high-dimensional features through linear layers and ReLU activation functions. The original input data can be efficiently reconstructed by gradually increasing the dimensionality and complexity of the data. In the process of reconstructing the original image, if the decoder’s reconstruction step size is too small, it will increase the uncertainty and complexity of the decoding process. In this paper, we gradually reconstruct the original data through two stages of similar step size. This can improve the stability of the decoder while reducing the uncertainty and complexity of the decoder. In addition, the encoding and decoding methods of the convolutional layer and the fully connected layer determine the choice of subsequent parameters to a certain extent.
Here are the detailed parameter settings, as shown in Table 1 and Table 2. We use C o n v i and F u l l y C o n n e c t e d i to represent the i-th convolutional layer and the i-th fully connected layer.
This process enhances the feature extraction capability of BLS, strengthens BLS’s understanding and utilization of data, and provides a more reliable foundation and support for subsequent applications. In any version of BLS, MAE can be used to achieve high-quality extraction of image features.

3.4. The Proposed MAEBLS and L-MAEBLS

On the image database, based on BLS, LK, and MAE, we developed MAEBLS and L-MAEBLS. As can be seen from Figure 3, we embed MAE into the feature layer of BLS and use MAE to encode complex image data to help BLS build more representative feature nodes to obtain MAEBLS. At the same time, the decoder is used to verify the effectiveness of the encoder. Based on MAEBLS, transform LK to a MAEBLS-based reweighting framework for training output weights to obtain L-MAEBLS. This enables BLS to focus more on valuable features, thereby alleviating the performance degradation caused by insufficient feature extraction capabilities when label noise exists. In Section 5, the experiment results will validate our statement.
E n c o d e r O u t p u t = M A E d a t a

4. Proof of Robustness

The method proposed in this paper demonstrates impressive robustness; thus, its inherent mechanism for robustness to label noise will be proved in this section. On one hand, we can approximate ζ ( y i , y i ) as ζ (e), where e = | | y i y i | | 2 . We can observe that ζ (e) is bounded, smooth, and non-convex loss. Therefore, ζ (e) exhibits robustness under noisy conditions. On the other hand, we will proceed to demonstrate the robustness mechanism of the method proposed in this paper.
Theorem 1. 
Through L-BLS, normal samples are assigned larger weights, while noisy samples are assigned smaller weights. Therefore, L-BLS can be more robust than BLS.
Proof. 
The error term of the robust width learning model L-BLS proposed in this paper can be regarded as E = { l o g ( 1 + e x p ( e 1 2 ) ) , l o g ( 1 + e x p ( e 2 2 ) ) , , l o g ( 1 + e x p ( e N 2 ) ) . Let the boundary for determining whether a sample is corrupted be θ . If l o g ( 1 + e x p ( e i 2 ) ) < θ , the i-th sample is considered an intact sample. Otherwise, it is considered a corrupted (containing label noise) sample. We can set the weights of the k-th sample in BLS, and L-BLS are assigned as follows:
W B L S k = e k 2 i = 1 N e i 2
W L B L S k = L k   i = 1 N L i  
where, for ease of representation and understanding, let us set L k   = l o g ( 1 + e x p ( e k 2 ) ) . We define δ k = W L B L S k W B L S k . Substituting the above equation into Equation (30), we get
δ k = W B L S k W L B L S k = L k   i = 1 N e i 2 e k 2 i = 1 N L i  
Furthermore, by normalizing the Cauchy–Schwarz inequality [30] to the above equation, we obtain the following equation:
i = 1 N ( 1 N ) 1 2 L k   e i 2 L i   e k 2 δ k
i = 1 N ( 1 N L k 2 e i 4 L i 2 e k 4 ) 1 2 δ k
An obvious fact is that the error of noisy data is much larger than that of intact data. Therefore, when e k belongs to normal training samples, we obtain the following equation:
i = 1 N L k 2 e i 4 L i 2 e k 4 > N
Further derivation leads to the following equation:
δ k = W L B L S k     W B L S k > 1
This implies that in L-BLS, compared to the base width learning system, normal data can be assigned larger weights. Thus, all training data can be concluded as i = 1 N W B L S i = i = 1 N W L B L S i = 1 . Thus, we can derive the following equation:
| e i   | < θ W L B L S i > | e i   | < θ   W B L S i
| e i   | > θ W L B L S i < | e i   | > θ   W B L S i
Thus, this proof is finished. □

5. Experimental Results

In this section, the performance of the proposed L-BLS, L-MAEBLS, and MAEBLS for classifying data with label noise was evaluated through extensive experiments. Accuracy (ACC) was chosen as the evaluation metric.
A C C = 1 N n = 1 N ψ y i , y i
where ψ ( a , b ) is a function that computes the number of correctly classified samples. Unless otherwise stated, all experiments were conducted using Python 3.10 on a computer equipped with an Intel i7 2.5-GHz CPU and 16-GB RAM.

5.1. Dataset Selection and Parameter Settings

Our experiments utilized six datasets from the UC Irvine (UCI) Machine Learning Repository [45] and three image classification datasets: Coil20 [46], ORL [47], and UMIST. Their features and partitions are shown in Table 3.
In the UCI datasets, we selected BLS [5] and four robust BLS models, including WBLS [23], C-BLS [26], ENBLS [22], and GRBLS [27], as comparison methods. On the image datasets, we constructed MAEBLS and L-MAEBLS based on the above methods and compared them with their original versions to demonstrate the ability of MAE in feature extraction under label noise [48,49,50]. To ensure fairness, we conducted grid searches within the same range to search for common parameters of the comparison methods in order to obtain the best performance.
Commonly used parameters include the number of feature mapping groups N ω , the number of feature mapping nodes per group N f , the number of enhancement nodes N e , and the L2 regularization parameter λ .
More specifically, for each UCI dataset, the search range for N f and N ω is [1,15], with a step size of 2. The search range for N e is [10,50], with a step size of 5. Search for the L2 regularization parameter within the range [ 2 30 , 2 25 , , 2 0 ] . For each image dataset, the search range for N f and N ω is [10, 50], with a step size of 5. Besides, the search range for N e is [1000, 5000], with a step size of 1000. The search range for the L2 regularization parameter λ is set to be the same as the corresponding range for the UCI datasets.
In the comparison methods, refer to the corresponding methodological papers, in the Huber–WBLS model, the positive adjustable parameter is set to 1.345. For the C-BLS model, the kernel size; for the ENBLS model, the L1 regularization parameter λ 1 and the L2 regularization parameter λ 2 ; for the GRBLS model, the regularization factor for the manifold term; and for the L-BLS model, the kernel width are all searched within the range of [ 2 30 , 2 25 , 2 0 ] .
Additionally, to eliminate the scale effect, we normalize the attributes of the datasets to [–1, 1]. For the UCI datasets, their attributes are normalized individually. For the three 8-bit grayscale image datasets, all attributes are divided by 127.5 and then subtracted by 1. Each dataset undergoes 50 repeated processes with all comparison methods, using the corresponding fixed optimal parameters to ensure the stability and reliability of the results.

5.2. Noise Modeling

People generally believe that labels always have a greater impact on the modeling process. In the model learning process, since the importance of each feature to model learning varies, label noise tends to have a more profound and detrimental effect than feature noise.
Ghosh et al. [51] proposed that label noise can be represented as follows:
y ~ =   y   ,                               with   probability     1 η y j , j ϵ N , y j y , with   probability     η j = η N 1
where η represents the noise ratio, and satisfies η = η j . N represents the total number of classes. When η is a constant, η j can be represented as η j = η N 1 . In this case, the type of label noise is symmetric or uniform noise. Otherwise, the noise type is asymmetric noise, where the true labels are randomly flipped to another class.
To be more realistic, in our experiments, we chose uniform label noise to simulate common noise situations. As for the process of mislabeling, it is completely random, with equal probabilities for all other classes.

5.3. Performance Evaluation on Data with Label Noise

In this section, the comparison results of the above methods on different UCI datasets under different pollution rates ( η = 0 %, η = 1 0%, …, η = 5 0%) are compared. The results are shown in Table 4 and Table 5 as the average values ± STD (%), with the best results highlighted in bold.
After thorough experimentation, we have confirmed the superior performance of L-BLS. Analyzing Table 4 and Table 5, we can make the following observations: First of all, across most UCI datasets, L-BLS outperforms competitors at various contamination levels, especially at higher noise levels like 40% and 50%. For instance, on the Wine dataset with 40% label noise, L-BLS achieves an impressive average accuracy of 94.25% with a minimal standard deviation (STD) of 0.96, while other methods struggle to reach 90% accuracy. This underscores L-BLS’s robustness to label noise. Figure 4 and Figure 5 provide a visual comparison of method trends on the Iris and Ecoil datasets. L-BLS outperforms other methods under any contamination rate conditions. In the Iris dataset, the accuracy of the L-BLS method is always above 96%, which is difficult to achieve with other methods. In particular, in the Ecoil dataset, when other methods are seriously affected by label noise, the L-BLS method can still remain stable and be far ahead of other methods. We can conclude that as the contamination rate increases, L-BLS maintains an acceptable accuracy drop, while the accuracy of other methods drops significantly. Second, L-BLS surpasses all comparison methods on COIL20, ORL, and UMIST datasets, except for L-MAEBLS, even maintaining acceptable performance degradation under severe contamination rates. This shows that L-BLS can still demonstrate strong robustness on image datasets. Third, on most clean datasets, L-BLS demonstrates superior performance compared to standard BLS and other methods, suggesting its capability to reweight and detect samples, further enhancing its performance. Overall, L-BLS consistently outperforms other methods across different scenarios, with acceptable performance even under high contamination rates. These comprehensive results indicate promising applications for L-BLS.
Notably, L-BLS consistently exhibits superior performance and accuracy across different contamination rates, especially under high noise levels, showcasing its robustness in label noise environments for effective classification tasks.

5.4. The Performance of L-BLS Combined with MAE

In the image dataset, in order to verify the MAE’s ability to enhance feature extraction and robustness under label noise conditions in the wide learning system, we added MAEBLS and L-MAEBLS to the experimental part and compared them with their corresponding BLS versions. In the COIL20, ORL, and UMIST datasets, MAEBLS shows more powerful feature extraction capabilities than BLS, and L-MAEBLS shows more powerful feature extraction capabilities than L-BLS. Notably, L-MAEBLS exhibited even greater robustness. Key findings from Table 5 are as follows:
First of all, for the COIL20, ORL, and UMIST datasets, the proposed MAEBLS and L-MAEBLS are superior to BLS and L-BLS at most pollution rates. From Figure 6 and Figure 7, it can be seen intuitively that MAE can effectively help BLS construct feature nodes. In particular, as shown in Figure 8, L-MAEBLS shows a significant improvement over BLS in accuracy. When η = 5 0%, L-MAEBLS can improve the accuracy by 5.91% compared with L-BLS, and other noise conditions are also significantly improved. Second, in the above three image datasets, L-MAEBLS outperforms all other comparison methods in most cases. From Figure 9, we can intuitively see the excellent performance of L-MAEBLS on the ORL database, especially when there is 50% label noise. The average accuracy of L-MAEBLS is 73.33%, with a standard deviation of only 0.65, while other methods struggle to achieve an accuracy of 70%.

5.5. The Effectiveness of L-BLS

In order to further verify the effectiveness of the reweighting framework L-BLS, in this section, we take the COIL20 dataset under the condition of symmetric label noise with a contamination rate of η = 3 0% as an example. We plot the sum of squares of residuals for each sample to visually evaluate the effectiveness of the proposed method. Since one-hot encoding is used to encode the labels of the samples, it is necessary to calculate the residuals for each element of each sample and then sum the squares of the residuals for each element.
S U M Y Y 2
where SUM(A) represents the sum of rows in matrix A.
The residuals and their squares for each sample in the initial iteration and at the optimal iteration are shown in Figure 10a.
From the results depicted in Figure 10a, it can be inferred that there exists an overlapping region, making it challenging to differentiate between clean and noisy samples. Hence, employing an iterative sample identification approach is more reasonable. Initially, only a few samples are distinctly identified as clean or noisy, with appropriate element weights, while the remaining samples are considered unidentified samples with moderate weights, effectively avoiding the negative impact of misjudgment. As the iterations progress, the squared errors of clean samples gradually decrease, while those of noisy samples gradually increase, facilitating more accurate sample identification.
Observing the results from Figure 10b, it can be noted that in the optimal iteration process, almost all clean samples are correctly identified, while the adverse effects of some noisy samples are effectively suppressed.

5.6. Statistical Analysis

In this section, we provide an analysis using the Friedman test to evaluate the statistical significance of the differences between the specified methods on the UCI datasets and image datasets.
First, we use the Friedman test with a confidence level of α = 0.1 to test the overall performance of different algorithms. As shown in Table 6, we can see that on the UCI dataset, our proposed L-BLS ranks the highest. C-BLS is second, while Huber-WBLS, ENBLS, and GRBLS perform better than the standard BLS. In the image dataset, as shown in Table 7, our proposed L-MAEBLS and L-BLS rank first and second, respectively, and our proposed MAEBLS also shows an improvement in ranking compared to BLS.
Second, as noise level is a critical factor affecting classifier performance, we employed the Friedman test at a confidence level of α = 0.1 to test the statistical differences of the complete set of algorithms under various noise levels. Table 8 and Table 9 display the results of the Friedman tests for the UCI datasets and the image datasets, respectively. In Table 8, the participants included BLS, ENBLS, C-BLS, GRBLS, Huber-WBLS, and L-BLS. Building on these methods, MAEBLS and L-MAEBLS also took part in the Friedman test, as shown in Table 9.
This paper conducts comparative experiments on the current advanced, robust BLS family methods under a more rigorous grid search scope. In the vast majority of cases, the p-values are less than 0.1, which implies that the statistical differences among the mentioned methods on the UCI database and image databases are significant under various noise levels.

6. Conclusions

In order to enhance the robustness and feature extraction ability of BLS in label noisy environments, we proposed in this paper, three versions of BLS: L-BLS, L-MAEBLS, and MAEBLS for modeling label noisy data. Based on the Gaussian kernel LK function, a new sample reweighting framework, L-BLS, is derived. Then the loss function and regularization term of the L-BLS classification problem are comprehensively studied. And for the image dataset, a mixture autoencoder MAE is constructed based on the convolutional autoencoder technology and the variational autoencoder technology, and the corresponding L-MAEBLS and MAEBLS versions are developed. Then the robustness and effectiveness of L-BLS are proven. The proposed L-BLS and L-MAEBLS can accurately simulate the loss distribution and are insensitive to label noise. MAE can provide more representative feature nodes for the corresponding BLS version. The experimental results show that the proposed L-BLS does achieve better performance in terms of robustness and effectiveness, and MAE does improve the feature extraction ability of the corresponding BLS version under label noise conditions.
This paper mainly focuses on the classification problem of BLS without considering the robust regression problem, which will be our future work.

Author Contributions

Conceptualization, J.S.; methodology, J.S.; software, J.S.; validation, J.S. and H.Z.; formal analysis, J.S.; investigation, W.D.; resources, H.Z.; data curation, J.S.; writing—original draft preparation, J.S.; writing—review and editing, W.D. and H.Z.; visualization, J.S.; supervision, H.Z.; project administration, H.Z.; funding acquisition, W.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (grant number 52372436), and the Research and Innovation Funding Project for Postgraduates of Civil Aviation University of China (grant No. 2023YJSKC02011).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank all the reviewers for their constructive comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Du, J.; Zhou, Y.; Liu, P.; Vong, C.-M.; Wang, T. Parameter-free loss for class-imbalanced deep learning in image classification. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 3234–3240. [Google Scholar] [CrossRef] [PubMed]
  2. Johnson, F.T.; Adebukola, O.S.; Ojo, O.E.; Alaba, A.; Victor, O. A Task Performance and Fitness Predictive Model Based on Neuro-Fuzzy Modeling. Artif. Intell. Appl. 2023, 2, 66–72. [Google Scholar] [CrossRef]
  3. Chen, H.; Long, H.; Chen, T.; Song, Y.; Chen, H.; Zhou, X.; Deng, W. M3FuNet: An Unsupervised Multivariate Feature Fusion Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5513015. [Google Scholar]
  4. Deng, W.; Cai, X.; Wu, D.; Song, Y.; Chen, H.; Ran, X.; Zhou, X.; Zhao, H. MOQEA/D: Multi-objective QEA with decomposition mechanism and excellent global search and its application. IEEE Trans. Intell. Transp. Syst. 2024. [Google Scholar] [CrossRef]
  5. Chen, C.L.P.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 10–24. [Google Scholar] [CrossRef]
  6. Pao, Y.-H.; Park, G.-H.; Sobajic, D.J. Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 1994, 6, 163–180. [Google Scholar] [CrossRef]
  7. Zhao, H.; Wu, Y.; Deng, W. An Interpretable Dynamic Inference System Based on Fuzzy Broad Learning. IEEE Trans. Instrum. Meas. 2023, 72, 2527412. [Google Scholar] [CrossRef]
  8. Li, M.; Wang, Y.; Yang, C.; Lu, Z.; Chen, J. Automatic diagnosis of depression based on facial expression information and deep convolutional neural network. IEEE Trans. Comput. Soc. Syst. 2024. [Google Scholar] [CrossRef]
  9. Deng, W.; Chen, X.; Li, X.; Zhao, H. Adaptive federated learning with negative inner product aggregation. IEEE Internet Things J. 2023, 11, 6570–6581. [Google Scholar] [CrossRef]
  10. Chu, F.; Liang, T.; Chen, C.L.P.; Wang, X.; Ma, X. Compact Broad Learning System Based on Fused Lasso and Smooth Lasso. IEEE Trans. Cybern. 2024, 54, 435–448. [Google Scholar] [CrossRef]
  11. Yang, K.; Yu, Z.; Chen, C.L.P.; Cao, W.; You, J.; Wong, H.-S. Incremental weighted ensemble broad learning system for imbalanced data. IEEE Trans. Knowl. Data Eng. 2022, 34, 5809–5824. [Google Scholar] [CrossRef]
  12. Shi, Y.; Yang, K.; Yu, Z.; Chen, C.L.P.; Zeng, H. Adaptive Ensemble Clustering with Boosting BLS-Based Autoencoder. IEEE Trans. Knowl. 2023, 35, 12369–12383. [Google Scholar] [CrossRef]
  13. Chen, C.L.P.; Liu, Z.; Feng, S. Universal approximation capability of broad learning system and its structural variations. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 1191–1204. [Google Scholar] [CrossRef] [PubMed]
  14. Ye, H.; Li, H.; Chen, C.L.P. Adaptive deep cascade broad learning system and its application in image denoising. IEEE Trans. Cybern. 2021, 51, 4450–4463. [Google Scholar] [CrossRef]
  15. Yi, J.; Huang, J.; Zhou, W.; Chen, G.; Zhao, M. Intergroup cascade broad learning system with optimized parameters for chaotic timeseries prediction. IEEE Trans. Artif. Intell. 2022, 3, 709–721. [Google Scholar] [CrossRef]
  16. Zhao, H.; Zheng, J.; Deng, W.; Song, Y. Semi-supervised broad learning system based on manifold regularization and broad network. IEEE Trans. Circuits Syst. I Reg. Pap. 2020, 67, 983–994. [Google Scholar] [CrossRef]
  17. Huang, S.; Liu, Z.; Jin, W.; Mu, Y. Broad learning system with manifold regularized sparse features for semi-supervised classification. Neurocomputing 2021, 463, 133–143. [Google Scholar] [CrossRef]
  18. Bhosle, K.; Musande, V. Evaluation of Deep Learning CNN Model for Recognition of Devanagari Digit. Artif. Intell. Appl. 2023, 1, 114–118. [Google Scholar] [CrossRef]
  19. Deng, W.; Li, K.; Zhao, H. A flight arrival time prediction method based on cluster clustering-based modular with deep neural network. IEEE Trans. Intell. Transp. Syst. 2023, 25, 6238–6247. [Google Scholar] [CrossRef]
  20. Liu, L.; Cai, L.; Liu, T.; Chen, C.L.P.; Tang, X. Cauchy regularized broad learning system for noisy data regression. Inf. Sci. 2022, 603, 210–221. [Google Scholar] [CrossRef]
  21. Lu, X.; Liu, W.; Zhou, C.; Huang, M. Robust Least-Squares Support Vector Machine with Minimization of Mean and Variance of Modeling Error. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2909–2920. [Google Scholar] [CrossRef]
  22. Jin, J.W.; Chen, C.L.P. Regularized robust broad learning system for uncertain data modeling. Neurocomputing 2018, 322, 58–69. [Google Scholar] [CrossRef]
  23. Chu, F.; Liang, T.; Chen, C.L.P.; Wang, X.; Ma, X. Weighted broad learning system and its application in nonlinear industrial process modeling. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 3017–3031. [Google Scholar] [CrossRef] [PubMed]
  24. Liu, L.; Liu, T.; Chen, C.L.P.; Wang, Y. Modal-regression-based broad learning system for robust regression and classification. IEEE Trans. Neural Netw. Learn. Syst. 2023. [Google Scholar] [CrossRef] [PubMed]
  25. Feng, S.; Ren, W.; Han, M.; Chen, Y.W. Robust manifold broad learning system for large-scale noisy chaotic time series prediction:A perturbation perspective. Neural Netw. 2019, 117, 179–190. [Google Scholar] [CrossRef]
  26. Zheng, Y.; Chen, B.; Wang, S.; Wang, W. Broad learning system based on maximum correntropy criterion. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3083–3097. [Google Scholar] [CrossRef]
  27. Jin, J.; Li, Y.; Chen, C.L.P. Pattern classification with corruptedlabeling via robust broad learning system. IEEE Trans. Knowl. Data Eng. 2022, 34, 4959–4971. [Google Scholar] [CrossRef]
  28. Liu, W.; Pokharel, P.P.; Príncipe, J.C. Correntropy: Properties and applications in non-Gaussian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  29. Zhao, X.; Fu, X.; Zeng, X.; Zhang, N. An Integrated Fishery Meteorological Information Service Terminal Based on End-Side Deep Learning Technology. Artif. Intell. Appl. 2024. [Google Scholar] [CrossRef]
  30. Shao, H.; Zhou, X.; Lin, J.; Liu, B. Few-shot cross-domain fault diagnosis of bearing driven by Task-supervised ANIL. IEEE Internet Things J. 2024, 11, 22892–22902. [Google Scholar] [CrossRef]
  31. Dong, W.; Zhou, L.; Ding, S.; Wang, A.; Cai, J. Two-Staged Method for Ice Channel Identification Based on Image Seg- mentation and Corner Point Regression. China Ocean. Eng. 2024, 38, 313–325. [Google Scholar] [CrossRef]
  32. Xiao, Y.; Shao, H.; Wang, J.; Yan, S.; Liu, B. Bayesian variational transformer: A generalizable model for rotating machinery fault diagnosis. Mech. Syst. Signal Process. 2024, 207, 110936. [Google Scholar] [CrossRef]
  33. Xie, C.; Zhou, L.; Ding, S.; Lu, M.; Zhou, X. Research on self-propulsion simulation of a polar ship in a brash ice channel based on body force model. Int. J. Nav. Archit. Ocean. Eng. 2023, 15, 100557. [Google Scholar] [CrossRef]
  34. Luo, J.; Shao, H.; Lin, J.; Liu, B. Meta-learning with elastic prototypical network for fault transfer diagnosis of bearings under unstable speeds. Reliab. Eng. Syst. Saf. 2024, 245, 110001. [Google Scholar] [CrossRef]
  35. Zhou, L.; Sun, Q.; Ding, S.; Han, S.; Wang, A. A Machine-Learning-Based Method for Ship Propulsion Power Prediction in Ice. J. Mar. Sci. Eng. 2023, 11, 1381. [Google Scholar] [CrossRef]
  36. Aronszajn, N. “Theory of reproducing kernels. Trans. Amer. Math. Soc. 1950, 68, 337–404. [Google Scholar] [CrossRef]
  37. Wu, Y.; Liu, Y. Robust truncated hinge loss support vector machines. J. Amer. Statist. Assoc. 2007, 102, 974–983. [Google Scholar] [CrossRef]
  38. Frenay, B.; Verleysen, M. Classification in the presence of label noise: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 845–869. [Google Scholar] [CrossRef] [PubMed]
  39. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements Statistical Learning: Data Mining, Inference, Predict, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  40. Miao, Q.; Cao, Y.; Xia, G.; Gong, M.; Liu, J.; Song, J. RBoost: Label noise-robust boosting algorithm based on a nonconvex loss function and the numerically stable base learners. IEEE Trans. Neural Netw. Learn. Syst. 2015, 27, 2216–2228. [Google Scholar] [CrossRef]
  41. Masci, J.; Meier, U.; Cireşan, D.; Schmidhuber, J. Stacked convolutional auto-encoders for hierarchical feature extraction. In International Conference on Artificial Neural Network; Springer: Berlin/Heidelberg, Germany, 2011; pp. 52–59. [Google Scholar]
  42. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. International Conference on Learning Representations. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  43. Yang, X.W.; Tan, L.J.; He, L.F. A robust least squares support vector machine for regression and classification with noise. Neurocomputing 2014, 140, 41–52. [Google Scholar] [CrossRef]
  44. Aldaz, J.M.; Barza, S.; Fujii, M.; Moslehian, M.S. Advances in operator Cauchy–Schwarz inequalities and their reverses. Ann. Funct. Anal. 2015, 6, 275–295. [Google Scholar] [CrossRef]
  45. Asuncion, A.; Newman, D. UCI Machine Learning Repository; School of Information and Computer Sciences, University California: Irvine, CA, USA, 2007. [Google Scholar]
  46. Murase, H.; Nayar, S.K. Visual learning and recognition of 3-D objects from appearance. Int. J. Comput. Vis. 1995, 14, 5–24. [Google Scholar] [CrossRef]
  47. Samaria, F.; Harter, A. ORL Face Database. 1994. Available online: https://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html (accessed on 14 August 2023).
  48. Wang, Z.; Wang, Q.; Liu, Z.; Wu, T. A deep learning interpretable model for river dissolved oxygen multi-step and interval prediction based on multi-source data fusion. J. Hydrol. 2024, 629, 130637. [Google Scholar] [CrossRef]
  49. Rui, S.; Dai, L.; Zhang, X.; He, M.; Xu, F.; Wu, W.; Armstrong, D.G.; You, Y.; Xiao, X.; Ma, Y.; et al. Exosomal miRNA-26b-5p from PRP suppresses NETs by targeting MMP-8 to promote diabetic wound healing. J. Control. Release 2024, 372, 221–233. [Google Scholar] [CrossRef]
  50. Yao, Z.; Wang, Z.; Wang, D.; Wu, J.; Chen, L. An ensemble CNN-LSTM and GRU adaptive weighting model based improved sparrow search algorithm for predicting runoff using historical meteorological and runoff data as input. J. Hydrol. 2023, 625, 129977. [Google Scholar] [CrossRef]
  51. Ghosh, A.; Kumar, H.; Sastry, P. Robust Loss Functions under Label Noise for Deep Neural Networks. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 1919–1925. [Google Scholar]
Figure 1. Architecture of the L-BLS.
Figure 1. Architecture of the L-BLS.
Sensors 24 04268 g001
Figure 2. Architecture of a Mixture Autoencoder.
Figure 2. Architecture of a Mixture Autoencoder.
Sensors 24 04268 g002
Figure 3. Architecture of the L-MAEBLS.
Figure 3. Architecture of the L-MAEBLS.
Sensors 24 04268 g003
Figure 4. Classification performance trends of different algorithms on the Iris dataset corrupted by noise of diverse levels.
Figure 4. Classification performance trends of different algorithms on the Iris dataset corrupted by noise of diverse levels.
Sensors 24 04268 g004
Figure 5. Classification performance trends of different algorithms on the Ecoil dataset corrupted by noise of diverse levels.
Figure 5. Classification performance trends of different algorithms on the Ecoil dataset corrupted by noise of diverse levels.
Sensors 24 04268 g005
Figure 6. Classification performance of MAEBLS corrupted by noise of diverse levels in three image datasets.
Figure 6. Classification performance of MAEBLS corrupted by noise of diverse levels in three image datasets.
Sensors 24 04268 g006
Figure 7. Classification performance of MAEBLS corrupted by noise of diverse levels in three image datasets.
Figure 7. Classification performance of MAEBLS corrupted by noise of diverse levels in three image datasets.
Sensors 24 04268 g007
Figure 8. Performance of L-MAEBLS on the UMIST database.
Figure 8. Performance of L-MAEBLS on the UMIST database.
Sensors 24 04268 g008
Figure 9. Classification performance trends of different algorithms on the ORL database corrupted by noise of diverse levels.
Figure 9. Classification performance trends of different algorithms on the ORL database corrupted by noise of diverse levels.
Sensors 24 04268 g009
Figure 10. Residual of each training sample in the COIL20 database at contamination rate η = 30%. (a) First iteration. (b) Optimal iteration.
Figure 10. Residual of each training sample in the COIL20 database at contamination rate η = 30%. (a) First iteration. (b) Optimal iteration.
Sensors 24 04268 g010
Table 1. Convolutional layer parameter settings.
Table 1. Convolutional layer parameter settings.
Name of LayerInput ChannelOutput ChannelKernel Size
C o n v 1 125 × 5
C o n v 2 245 × 5
C o n v 3 415 × 5
Table 2. Fully connected layer parameter settings.
Table 2. Fully connected layer parameter settings.
Name of LayerInput FeaturesOutput Features
F u l l y C o n n e c t e d 1 2020
F u l l y C o n n e c t e d 2 400700
F u l l y C o n n e c t e d 3 7001024
Table 3. Characteristics of the selected datasets.
Table 3. Characteristics of the selected datasets.
DatasetsAttributesClassesNo. of Instances
TrainingTesting
Statlog (Heart)132162108
Iris439060
Wine13310771
Balance scale43375250
Zoo1676140
Ecoil88202134
Coil201024201008432
ORL102440280120
UMIST102420345230
Table 4. The classification results of different methods on different test sets of UCI databases.
Table 4. The classification results of different methods on different test sets of UCI databases.
DatasetsAlgorithms η = 0% η = 10% η = 20% η = 30% η = 40% η = 50%
Aver ± STD (%)
Statlog (Heart)BLS80.74 ± 2.0677.78 ± 1.1474.72 ± 3.2771.43 ± 2.2466.64 ± 3.1660.92 ± 3.07
Huber-WBLS82.23 ± 0.8982.04 ± 0.8879.69 ± 0.8477.90 ± 0.9467.53 ± 0.9763.27 ± 0.66
ENBLS83.14 ± 1.0882.22 ± 0.6980.19 ± 0.4573.19 ± 2.4467.24 ± 1.5054.44 ± 0.37
C-BLS82.96 ± 0.9481.39 ± 0.4981.02 ± 0.6280.09 ± 0.7576.85 ± 1.0166.29 ± 1.31
GRBLS81.12 ± 6.1377.99 + 9.1272.37 ± 9.2671.72 ± 9.2068.24 ± 8.2467.81 ± 8.99
L-BLS85.93 ± 1.3683.24 ± 0.4882.96 ± 0.9481.67 ± 0.3777.59 ± 1.3673.33 ± 0.37
IrisBLS96.50 ± 0.5093.49 ± 1.7193.13 ± 1.3291.10 ± 1.3190.76 ± 1.7786.59 ± 3.07
Huber-WBLS96.77 ± 1.4796.36 ± 0.6794.33 ± 1.1193.17 ± 0.8992.93 ± 1.9287.93 ± 5.09
ENBLS95.80 ± 0.9695.66 ± 0.8295.33 ± 0.1195.33 ± 1.4595.30 ± 0.8289.33 ± 2.81
C-BLS96.77 ± 1.4796.23 ± 0.7396.40 ± 0.6195.80 ± 1.0695.23 ± 1.2594.53 ± 2.11
GRBLS96.39 ± 6.6795.13 ± 7.7494.70 ± 6.2394.47 ± 8.2292.53 ± 5.8582.27 ± 8.36
L-BLS98.33 ± 1.1196.56 ± 0.3996.43 ± 0.5796.23 ± 0.9896.20 ± 0.7496.13 ± 2.65
WineBLS95.27 ± 1.6194.08 ± 0.7493.09 ± 1.1686.14 ± 0.7177.15 ± 7.2960.98 ± 4.28
Huber-WBLS98.09 ± 0.6797.25 ± 0.8496.90 ± 0.8593.66 ± 0.9488.17 ± 1.3177.82 ± 1.72
ENBLS96.19 ± 1.2795.77 ± 0.8594.36 ± 1.2891.41 ± 1.8580.98 ± 1.8469.98 ± 7.92
C-BLS97.54 ± 0.6196.95 ± 1.1395.85 ± 0.9992.92 ± 1.0389.84 ± 1.0474.87 ± 1.76
GRBLS97.01 ± 1.1896.25 ± 9.3795.18 ± 9.6894.85 ± 3.7093.94 ± 3.1475.52 ± 10.16
L-BLS98.73 ± 0.4298.62 ± 1.1998.62 ± 0.5395.74 ± 2.0994.25 ± 1.9694.39 ± 9.69
Balance scaleBLS92.44 ± 0.2290.99 ± 0.6090.60 ± 1.3988.35 ± 0.8787.29 ± 1.3978.51 ± 1.72
Huber-WBLS94.40 ± 0.3194.38 ± 0.0993.82 ± 0.5391.22 ± 0.7390.86 ± 1.9786.76 ± 0.63
ENBLS93.08 ± 1.7192.02 ± 1.3591.12 ± 2.4090.64 ± 2.6589.76 ± 2.8585.07 ± 2.45
C-BLS93.72 ± 0.1891.60 ± 0.3691.44 ± 0.5190.60 ± 0.4588.12 ± 0.1887.04 ± 1.22
GRBLS91.44 ± 1.2291.34 ± 3.1391.12 ± 1.8589.65 ± 1.5088.96 ± 1.9385.45 ± 2.36
L-BLS94.88 ± 0.6494.68 ± 0.1894.56 ± 0.4194.19 ± 0.2793.08 ± 0.8189.40 ± 0.76
ZooBLS96.10 ± 2.0190.64 ± 2.7387.45 ± 6.3584.15 ± 2.6381.50 ± 1.2266.90 ± 4.08
Huber-WBLS97.56 ± 4.2292.20 ± 1.4689.51 ± 2.6888.54 ± 2.2085.61 ± 3.1773.90 ± 2.45
ENBLS93.90 ± 3.7592.15 ± 4.0190.41 ± 4.0491.50 ± 4.1583.05 ± 2.3677.05 ± 5.58
C-BLS98.71 ± 1.2597.30 ± 2.1795.60 ± 1.7892.39 ± 2.4987.40 ± 0.7075.05 ± 5.39
GRBLS97.05 ± 1.6396.80 ± 1.4296.15 ± 1.6793.85 ± 3.0985.50 ± 2.9283.65 ± 2.13
L-BLS98.95 ± 1.2397.45 ± 0.3597.25 ± 0.7596.25 ± 2.1490.75 ± 3.0185.10 ± 1.93
EcoilBLS88.82 ± 1.1187.81 ± 0.5483.64 ± 1.5983.45 ± 1.1582.37 ± 1.7669.70 ± 3.08
Huber-WBLS90.30 ± 1.0989.17 ± 0.6988.72 ± 1.3286.09 ± 1.6485.68 ± 1.6981.47 ± 5.11
ENBLS90.15 ± 1.1587.61 ± 1.3485.07 ± 2.4583.28 ± 1.1778.21 ± 3.6473.66 ± 2.94
C-BLS90.44 ± 0.5689.55 ± 0.7586.26 ± 0.8985.82 ± 1.4985.44 ± 1.0782.08 ± 1.29
GRBLS87.78 ± 0.6887.21 ± 1.1486.72 ± 0.6885.91 ± 1.5785.53 ± 1.0783.48 ± 2.27
L-BLS92.87 ± 1.7990.29 ± 1.2490.02 ± 1.0189.78 ± 1.0587.53 ± 1.0685.84 ± 1.11
Table 5. The classification results of different methods on different test sets of Images databases.
Table 5. The classification results of different methods on different test sets of Images databases.
DatasetsAlgorithms η = 0% η = 10% η = 20% η = 30% η = 40% η = 50%
Aver ± STD (%)
COIL20BLS99.95 ± 0.0395.52 ± 0.7690.21 ± 1.0485.91 ± 0.7379.01 ± 1.2859.75 ± 1.34
Huber-WBLS100 ± 0.0099.54 ± 0.3296.81 ± 0.6396.03 ± 0.5794.89 ± 0.6292.52 ± 0.84
ENBLS99.19 ± 0.4698.89 ± 0.3295.54 ± 0.5894.62 ± 1.0490.06 ± 0.9789.98 ± 1.28
C-BLS100 ± 0.0095.61 ± 0.2395.33 ± 0.0992.10 ± 0.2391.82 ± 0.3190.02 ± 0.31
GRBLS97.85 ± 0.2197.21 ± 0.3497.16 ± 0.2194.73 ± 0.7694.08 ± 0.5888.22 ± 0.62
L-BLS100 ± 0.0099.45 ± 0.1598.87 ± 0.2796.49 ± 0.6795.10 ± 0.3294.25 ± 0.97
MAEBLS100 ± 0.0096.54 ± 0.4693.99 ± 0.3986.93 ± 0.5081.76 ± 0.3976.17 ± 0.66
L-MAEBLS100 ± 0.0099.97 ± 0.1999.52 ± 0.1598.86 ± 0.2396.59 ± 0.3294.41 ± 0.31
ORLBLS92.67 ± 1.7878.83 ± 0.6672.33 ± 1.7860.50 ± 2.3358.52 ± 2.2046.00 ± 1.61
Huber-WBLS94.92 ± 1.1589.58 ± 0.8185.17 ± 0.9574.21 ± 0.9873.79 ± 2.1259.75 ± 1.57
ENBLS94.39 ± 1.7889.67 ± 2.7278.50 ± 2.2073.16 ± 2.0766.17 ± 1.7256.14 ± 1.13
C-BLS92.83 ± 0.4192.34 ± 0.3388.00 ± 0.6774.83 ± 1.1066.79 ± 1.0556.67 ± 1.17
GRBLS94.67 ± 0.6791.42 ± 0.8484.63 ± 0.7181.92 ± 1.1073.92 ± 1.1862.58 ± 1.37
L-BLS94.68 ± 0.6692.83 ± 1.1388.83 ± 0.9185.50 ± 1.6773.25 ± 1.2267.92 ± 2.30
MAEBLS93.36 ± 1.0584.00 ± 1.4373.98 ± 1.2264.00 ± 1.8662.17 ± 2.6147.66 ± 1.33
L-MAEBLS95.34 ± 0.8593.83 ± 0.6789.02 ± 0.8287.00 ± 1.3077.35 ± 0.9773.33 ± 0.65
UMISTBLS99.26 ± 0.2095.98 ± 0.7592.01 ± 1.2585.53 ± 1.3081.67 ± 1.5169.27 ± 1.65
Huber-WBLS97.81 ± 0.3697.22 ± 0.4394.89 ± 0.9389.95 ± 0.9482.68 ± 1.0874.30 ± 1.25
ENBLS96.40 ± 0.2996.37 ± 0.4193.90 ± 0.4092.01 ± 0.3896.13 ± 0.9275.65 ± 4.76
C-BLS99.12 ± 0.8597.89 ± 0.2996.52 ± 0.6793.04 ± 0.9282.61 ± 0.8276.09 ± 0.91
GRBLS96.61 ± 0.6495.32 ± 0.5793.74 ± 0.6390.13 ± 0.6286.09 ± 0.6775.69 ± 0.99
L-BLS99.13 ± 0.1198.09 ± 0.2896.85 ± 0.5093.74 ± 0.8986.70 ± 0.5976.26 ± 1.09
MAEBLS99.80 ± 0.1896.23 ± 0.4192.75 ± 1.0887.39 ± 0.6183.48 ± 1.8571.16 ± 1.82
L-MAEBLS99.28 ± 0.2198.99 ± 0.2497.61 ± 0.2294.96 ± 0.2988.17 ± 1.3982.17 ± 0.72
Table 6. Average rankings of different algorithms in classification accuracy in UCI databases.
Table 6. Average rankings of different algorithms in classification accuracy in UCI databases.
BLSHuber-WBLSENBLSC-BLSGRBLSL-BLS
Accuracy ranks5.6667 3.1528 4.2639 2.9583 3.9583 1
Chi-sq 125.19
p-value 2.49293   ×   10 25
Table 7. Average rankings of different algorithms in classification accuracy in image databases.
Table 7. Average rankings of different algorithms in classification accuracy in image databases.
BLSHuber-WBLSENBLSC-BLSGRBLSL-BLSMAEBLSL-MAEBLS
Accuracy Ranks7.5556 4.1111 5.22224.55564.72222.44446.16671.2222
Chi-sq84.54
p-value 1.62493   ×   10 15
Table 8. Statistical testing of classification accuracy on the UCI databases.
Table 8. Statistical testing of classification accuracy on the UCI databases.
Noise LevelChi-sqp-Value
r = 0%23.56 2.63 × 10 4
r = 10%23.14 3.17 × 10 3
r = 20%21.84 5.61 × 10 4
r = 30%21.08 7.93 × 10 4
r = 40%21.62 6.18 × 10 4
r = 50%21.14 7.61 × 10 4
Table 9. Statistical testing of classification accuracy on the image database.
Table 9. Statistical testing of classification accuracy on the image database.
Noise LevelChi-sqp-Value
r = 0%10.50 0.1619
r = 10%16.00 0.0251
r = 20%19.11 0.0784
r = 30%18.78 0.0089
r = 40%15.44 0.0307
r = 50%19.01 0.0081
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, J.; Zhao, H.; Deng, W. Broad Learning System under Label Noise: A Novel Reweighting Framework with Logarithm Kernel and Mixture Autoencoder. Sensors 2024, 24, 4268. https://doi.org/10.3390/s24134268

AMA Style

Shen J, Zhao H, Deng W. Broad Learning System under Label Noise: A Novel Reweighting Framework with Logarithm Kernel and Mixture Autoencoder. Sensors. 2024; 24(13):4268. https://doi.org/10.3390/s24134268

Chicago/Turabian Style

Shen, Jiuru, Huimin Zhao, and Wu Deng. 2024. "Broad Learning System under Label Noise: A Novel Reweighting Framework with Logarithm Kernel and Mixture Autoencoder" Sensors 24, no. 13: 4268. https://doi.org/10.3390/s24134268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop