Next Article in Journal
Analytical Thermal Analysis of Radially Functionally Graded Circular Plates with Coating or Undercoating under Transverse and Radial Temperature Distributions
Previous Article in Journal
Study on the Classification and Identification Methods of Surrounding Rock Excavatability Based on the Rock-Breaking Performance of Tunnel Boring Machines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interactive Guiding Sparse Auto-Encoder with Wasserstein Regularization for Efficient Classification

Department of Computer Engineering, Inha University, Inha-ro 100, Michuhol-gu, Incheon 22212, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(12), 7055; https://doi.org/10.3390/app13127055
Submission received: 26 April 2023 / Revised: 6 June 2023 / Accepted: 9 June 2023 / Published: 12 June 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
In the era of big data, feature engineering has proved its efficiency and importance in dimensionality reduction and useful information extraction from original features. Feature engineering can be expressed as dimensionality reduction and is divided into two types of methods, namely, feature selection and feature extraction. Each method has its pros and cons. There are a lot of studies that combine these methods. The sparse autoencoder (SAE) is a representative deep feature learning method that combines feature selection with feature extraction. However, existing SAEs do not consider feature importance during training. It causes extracting irrelevant information. In this paper, we propose an interactive guiding sparse autoencoder (IGSAE) to guide the information by two interactive guiding layers and sparsity constraints. The interactive guiding layers keep the main distribution using Wasserstein distance, which is a metric of distribution difference, and it suppresses the leverage of guiding features to prevent overfitting. We perform our experiments using four datasets that have different dimensionalities and numbers of samples. The proposed IGSAE method produces a better classification performance compared to other dimensionality reduction methods.

1. Introduction

Efficient feature learning algorithms have become an essential element and the hot research topic in the era of big data [1]. Massive and high-dimensional data generally have a lot of redundancies and noise. This brings huge challenges to extracting important features, processing them, and identifying meaningful patterns. Many traditional methods are also fine in some cases. For example, many methods are based on linear relationships. This is intuitive and easily understandable. However, there are also non-linear data structures. Therefore, a machine learning-based model can be a solution to explain the non-linear structure [2,3,4,5]. An autoencoder [6,7,8,9] is a representative machine learning technique for finding manifolds that represent large-scale data and is employed extensively in the field of feature extraction and dimensionality reduction. It consists of two parts: an encoder and a decoder. The encoder transforms the input into a latent vector that contains the main information. The decoder reconstructs the latent vector to new features that are the same as the input form. It learns to minimize the reconstruction error. This method has become one of the more useful feature engineering methods.
Feature engineering is a process of extracting important information from raw data and can be generally divided into two types: feature selection [10,11,12] and feature extraction [13,14]. Feature selection can be explained as a method of finding a subset of features with higher feature importance for classification. Therefore, it is possible to effectively remove less relevant features. These methods generally have a high generalization performance and computational efficiency. For example, there is a regularization [15,16], information gain [17], chi-square test [18], etc. However, these methods have the disadvantage of removing potential non-noise factors of the original data. On the other hand, feature extraction methods are usually based on the reconstruction of features and projection from high-dimensional data into low-dimensional subspace. This kind of algorithm normally preserves the variance of features, so it has high information retention. Since these algorithms consider almost all features in the transformation process, the dimensionality reduction effect is usually depleted. Principal component analysis (PCA) [19,20], canonical correlation analysis (CCA) [21], linear discriminant analysis (LDA) [22,23], and autoencoder using machine learning models for feature extraction are representative methods.
However, an autoencoder has overfitting problems and highly complex structural issues. Improved structures have been proposed in recent years to solve these problems. Peicheng et al. [24] applied a stacked autoencoder (SAE) to hyperspectral image classification to generate high-quality hyperspectral image data to maximize the performance of the task. Meidi et al. [25] applied a stacked autoencoder to fault diagnosis of rolling bearings and obtained promising results. They increase the size of the network to obtain a better result. However, the autoencoder-based simple method is still a field of feature extraction, so it is difficult to modify and remove features with less relevance, which is the advantage of feature selection. To improve this shortcoming, new techniques applying regularization have been studied. Notably, the stacked sparse autoencoder [26,27,28] is one of the representative deep autoencoder methods that have feature selection ideas. The stacked sparse autoencoder (SSAE) improved the autoencoder by a combination of two advanced autoencoders: a sparse autoencoder and a stacked autoencoder. A sparse autoencoder adds sparse constraints to the autoencoder. It restricts the useless node to reduce the dimension efficiently. A stacked autoencoder improves the training performance by stacking autoencoders. It mediates to deeply learn the features in comparison with the simple learning process of traditional autoencoders. Anush et al. [29] proposed the structure of a group sparse autoencoder (GAE) with regularization applied to solve the overfitting, which is a chronic problem of autoencoders.
SSAE’s regularization-based methods achieved successful results in various applications. However, it is difficult to apply the dynamic and efficient feature selection method due to its structural issues. In many cases, there are vanishing gradient problems and over-generalizing information (e.g., on sensitive applications). Zhilei et al. [30] proposed a label and sparse regularization autoencoder (LSRAE). The main idea of LSRAE is to learn by dividing the output layer of the autoencoder into two parts, the actual label and output data. LSRAE extracts features that involve the relationship between features and targets, and prevents overfitting via sparse regularization. Most regularization-based autoencoders optimize the loss function without considering the structure. However, LSRAE has structural changes and the advantages of supervised learning. This is an important point in the feature engineering field. Hongteng et al. [31] proposed a relational regularized autoencoder (RAE) that flexibly learns a structured prior distribution using Fused Gromov–Wasserstein (FGW) distance [32]. The main idea is to limit the difference by regularizing the structural distance between the latent prior and the corresponding posterior using a metric determining the distribution similarity. To obtain high-quality and low-dimensional features, the structure of SSAE needs to be improved and reformed.
In this paper, we propose an interactive guiding sparse autoencoder (IGSAE), which is an advanced autoencoder model. It is hard to prevent overfitting with existing autoencoders because the weights are updated to reduce reconstruction loss. To avoid overfitting, we modify the autoencoder by inserting two guiding layers before and after the reconstruction layer to adjust the leverage of feature importance. These guiding layers learn the distribution of selected labels and relieve overfitting caused by the reconstruction loss. Next, we design a regularization module to prevent the reconstruction layer and guiding layer from having excessive influence. Therefore, our method is a new manifold learning system that can improve the quality of deep features.

Main Contributions

The main contributions of our method are summarized as follows:
  • We reduce overfitting of the autoencoder to propose a new structure that can utilize both label information and reconstruction information, similar to semi-supervised learning by proposing a guiding layer before and after the reconstruction layer of the autoencoder.
  • We design an architecture that consists of two different autoencoders and show outstanding classification performance compared to six representative feature extraction methods.
The rest of this paper is organized as follows. Section 2 introduces the related work, including the basic theory and advanced theory of the autoencoder that is related to the proposed method. Section 3 describes the details of the proposed IGSAE method and shows its framework. Section 4 illustrates the experiment environment, evaluation metrics, information on datasets, and experimental results. Finally, Section 5 concludes the paper.

2. Background and Related Works

In this section, we describe the autoencoder-based models.

2.1. Sparse Autoencoder (SAE)

An autoencoder consists of two parts: an encoder and a decoder. In the encoder step, the autoencoder transforms the original feature into a latent representation. It is model-interpretable information. In the decoder step, the autoencoder reconstructs the latent representation into new meaningful features.
An encoding function f maps the input data sample x m to a low-dimensional latent feature space n . It is formulated as:
h = f ( x ) = ϕ ( W 1 T x + b 1 )
where W 1 T is the transformed weight matrix ( W 1 ) between the input layer and hidden layer, ϕ is the activation function, b 1 is a bias, and h is a latent vector of the hidden layer. A decoding function g maps h to reconstructed feature x . It can be expressed as:
x = g ( x ) = ϕ ( W 2 T h + b 2 )
where W 2 T is the weight matrix ( W 2 ) between the hidden layer and output layer and b 2 is a bias. The purpose of the autoencoder is to find a structure that minimizes the reconstruction error, and the objective function is defined as follows:
min θ   J A E ( θ ) = 1 2 n i = 1 n x i x i 2
where J A E is the reconstruction loss of the autoencoder and n is the number of input samples. SAE’s structure is based on an autoencoder. However, in a basic autoencoder, when reconstructing the input, the distribution of reconstructed features has a lot of overlap due to problems such as overfitting [33], and simply copying the input may not represent key information well. To solve these problems, we add a sparsity constraint as a penalty term that can suppress the output of neurons when the number of neurons in the hidden layer is large and formulated as follows:
min θ   J S A E ( θ ) = 1 n i = 1 n x i x i 2 + ε j = 1 s K L ( ρ | | ρ ^ j )
where J S A E is the objective of the sparse autoencoder (SAE), s is the number of neurons in the hidden layer, ρ ^ j is the activation probability, and ρ is the sparse parameter. KL divergence is briefly a measure of how differences between two probability distributions [34].

2.2. Label and Sparse Regularization Autoencoder (LSRAE)

Zhilei et al. proposed LSRAE, a structure that combines supervised learning methods with the autoencoder, which is unsupervised learning, to improve the performance of classification tasks. They thought that the overfitting and the irregular information problem of classification, which are chronic problems of autoencoders, occur because the relation between the target data and the input data is not considered at all in the reconstruction process. Therefore, they solved the problem by introducing two main methods: the first is adding a label regularization term to the objective function to reduce the difference between the actual label and the desired label, and the second is using the extreme learning machine (ELM) [35] for fast optimization. Label regularization term ( J l a b e l ) of the LSRAE is formulated as follows:
J l a b e l = 1 2 n i = 1 n L T 2 = 1 2 n i = 1 n H β ^ T 2
where L is the actual label as the output of the autoencoder, T is the desired label, and n is the number of input samples. Figure 1 shows a way to utilize the ELM on the autoencoder. The blue lines connect the hidden layer and the output layer. The red lines connect the hidden layer and label layer. SoftMax classifier is adopted for the classifier. In the training process, they compute the reconstruction loss, the weight decay term, and the sparse regularization term, in order. Then, they use the concept of ELM to calculate the real label. The real label is used to obtain label regularization. The LSRAE learns through backpropagation by obtaining the squared error of the expression as Equation (5). It can restrict overfitting of the autoencoder that focuses only on reducing reconstruction loss. It improves the classification accuracies, and they also use the sparse constraint to restrict useless nodes. The simple objective function of LSRAE is calculated as follows:
J s p a r s e = j = 1 s K L ( ρ | | ρ ^ j )
J L S R A E = J A E + λ J l a b e l + β J s p a r s e
where J A E is the loss function of the basic autoencoder, and λ and β are coefficients.

3. Interactive Guiding Sparse Autoencoder

Autoencoders have traditionally been used as one of the dimensionality reduction methods, such as PCA and LDA. However, when overfitting occurs in the learning process, the following two major problems occur. First, even if unrelated data are used as input, the autoencoder transforms the data into a similar distribution with a training set. Second, while learning the information of features with low importance, a noise can be recognized as important information, which affects classification accuracy. To solve the problem, we developed an interactive guiding sparse autoencoder (IGSAE) that learns the manifold of selected features and guides it to the rest of the features.
As shown in Figure 2, IGSAE has a structure with two added guiding layers to the basic autoencoder so that the two layers are interactive. From left to right, it is composed in the order of input layer, hidden layer (with guiding layer), reconstruction layer, and guiding layer (output layer). The whole process is explained in small five processes. In the first process of IGSAE, H and G 1 are the output of hidden units and the output of the first guiding units, respectively. The step to calculate H and G 1 can be represented as follows:
H = ϕ ( W 1 X + b 1 )
G 1 = ϕ ( W 1 X + b 1 )
where X = { x 1 ,   x 2 , ,   x n } is input data, ϕ is activation function (RELU), and   W 1 and b 1 denote the weight and bias between the input and hidden layer, respectively. W 1 and b 1 denote the weight and bias between the input and guiding layer, respectively. X denotes a guided feature and X g denotes a guiding feature. The learning process of IGSAE can be divided into two parts: (a) to learn the guiding feature as a target from the guided feature to involve the relationship between the two features, and (b) to reconstruct the guided feature. A pretrained model from guided feature X to guiding feature X g is required to involve the relationship between the features. The corresponding part of IGSAE maps the guided feature of the input layer X to the guiding feature of the first guiding layer G 1 and is pretrained partially. The cost function J f g of the partial pretrained model is formulated as follows:
J f g = 1 2 n   X g G 1 2 = 1 2 n   X g ϕ ( W 1 X + b 1 ) 2
where X g = { x n + 1 ,   x n + 2 , ,   x m } is input data of the guiding layer. The cost function used for pretraining is also used as a regularization term for the objective function in Equation (18).
In the second process, as the number of units in the hidden layer increases, there may be too much sparse information. As a sparse constraint, the average activity of the hidden unit is calculated. If it is an important unit, it is activated, and if not, it is deactivated. The sparsity criterion for determining the importance of a unit uses the Kullback–Leibler (KL) divergence. KL divergence is the relative entropy to measure the similarity between a sparsity parameter ρ and the average activation probability ρ ^ j . s is the number of the hidden units. The average activation ρ ^ j is represented as:
ρ ^ j = 1 n i = 1 n ϕ ( x i )
and J s p a r s e (KL divergence) is represented as:
J s p a r s e = j = 1 s K L ( ρ | | ρ ^ j ) = j = 1 s ρ log ρ ρ ^ j + ( 1 ρ ) log 1 ρ 1 ρ ^ j
In the third process, there is a reconstruction loss J A E , and the loss can be calculated by the output vector of the reconstruction layer X . The output of the reconstruction layer is calculated as follows:
X = ϕ ( W 2 H + b 2 ) + ϕ ( W 2 G + b 2 ) = ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) + ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 )
where ϕ is the activation function, and W 2 and b 2 denote the weight and bias between the hidden and output layers, respectively. W 2 and b 2 denote the weight and bias between the guiding and output layers, respectively. The J A E is a typical loss used in the existing autoencoder and is the main loss of IGSAE. In general, the mean squared deviation is used for linear cases, and cross-entropy is used for non-linear cases. In this section, we compute using mean squared deviation. J A E is computed as follows:
J A E = 1 2 n X X 2 = 1 2 n X ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) + ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) 2
In the fourth process, IGSAE has a second guiding part. Its operation step is almost the same to obtain the first guiding layer, and the two guide layers operate complementarily. The second guiding regularization term ( J s g ) is calculated as follows:
G 2 = ϕ ( W 3 X + b 3 ) = ϕ ( W 3 ( ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) + ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) ) + b 3 )
J s g = 1 2 n   X g G 2 2 = 1 2 n   X g ϕ ( W 3 ( ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) + ϕ ( W 2 ϕ ( W 1 X + b 1 ) + b 2 ) ) + b 3 ) 2
where W 3 and b 3 are the weight and bias between the output layer and the second guiding layer, respectively. The guiding term ( J g u i d e ) is computed by two guiding terms Equations (10) and (16), and is formulated as follows:
J g u i d e = J f g + J s g
J g u i d e is the loss of the two guiding layers that learn the distribution of the real label. If the distribution of the label can be influenced in the learning process of the autoencoder, then overfitting, which learns only to reduce the reconstruction loss [36], can be restricted. Therefore, the objective function of IGSAE using the above Equations (12), (14) and (17), and the Wasserstein regularization term ( J w a s ) is defined as:
min w W   J I G S A E = min w W   J A E + ε J s p a r s e + λ J g u i d e + μ J w a s = min w W   1 2 n X X 2 + ε j = 1 s K L ( ρ | | ρ ^ j ) + λ 1 2 n   X g G 2 2 + 1 2 n   X g G 1 2 + μ inf r ( P θ 1 ,   P θ 2 ) E ( X ,   H ) ~ γ [ d ( G 1 ,   G 2 ) ]
where ε , λ , and μ are coefficients to balance the J I G S A E . W is the weight of IGSAE. θ 1 is the parameter of the first guiding layer and θ 2 is the parameter of the second guiding layer. P θ 1 is the pretrained prior distribution, which we assume is real data distribution. P θ 2 is the latent variable distribution. ( P θ 1 ,   P θ 2 ) denotes the set of all joint distributions d ( G 1 ,   G 2 ) whose marginals are P θ 1 and P θ 2 , respectively.
When learning IGSAE, the influence of the first guiding layer becomes too strong, which can increase the reconstruction loss. If reconstruction performance decreases, then classification performance can be also degraded [37,38]. Therefore, the Wasserstein guiding term is proposed to limit the influence of the first guiding layer. The Wasserstein regularization term ( J w a s ) is explained in the next part and shown as:
J w a s = inf r ( P θ 1 ,   P θ 2 ) E ( X ,   H ) ~ γ [ d ( G 1 ,   G 2 ) ]
where J w a s is the lowest estimate of the expected value of ( P θ 1 ,   P θ 2 ) . However, the infimum in Equation (19) is too intractable and complex. Kantorovich–Rubinstein duality suggests that the solution to replace the problem is with a low-complexity duality problem. It can be formulated as:
min w W   J I G S A E = min w W   J A E + ε J s p a r s e + λ J g u i d e + μ   sup   f L E x ~ P θ 1 [ f ( x ) ] E x ~ P θ 2 [ f ( x ) ]
where L means all ranges that satisfy the 1-Lipschitz function and f is a function about θ 1 that satisfies the 1-Lipschitz function. By adding a weight clipping parameter to the f function to fit the duality condition, it can be expressed as follows:
min w W   J I G S A E = min w W   J A E + ε J s p a r s e + λ J g u i d e + μ max   w W E x ~ P θ 1 [ f w ( x ) ] E x ~ P θ 2 [ f w ( g θ 2 ( x ) ) ]
IGSAE assumes that we already have good performance f w throught pretraining. Ultimately our remaining goal is to find the optimum to update W 1 , W 2 , and W 2 through the Wasserstein regularization. That is, it should find optimal g θ 2 . Furthermore, we could consider and compute differentiating the changed J w a s through the back-propagation method in Equation (21) via obtaining max   w W E x ~ P θ 1 [ f w ( x ) ] . According to the above methods, using backpropagation of the entire objective function is possible by the gradient descent algorithm. In the partial derivative formula to optimize J I G S A E , bias is omitted to simplify the expression.
The weights W is formulated in the order as follows:
J I G S A E W 1 = J A E W 1 + ε J s p a r s e W 1 + λ J g u i d e W 1 + μ J w a s W 1                       = J A E W 1 + ε J s p a r s e W 1 + λ J s g W 1 + μ J w a s W 1
Since W 1 is not affected by J f g , it is removed from J g u i d e and only J s g remains after differentiation. W 2 is computed as follows:
J I G S A E W 2 = J A E W 2 + ε J s p a r s e W 2 + λ J g u i d e W 2 + μ J w a s W 2 = J A E W 2 + λ J s g W 2 + μ J w a s W 2
Since W 2 is the weight between the first guiding layer and the reconstruction layer, it is not affected by J s p a r s e . Therefore, J s p a r s e is removed. W 1 is computed as follows:
J I G S A E W 1 = J I G A E W 3 = [ J A E W 1 + ε J s p a r s e W 1 + λ J g u i d e W 1 + μ J w a s W 1 ]                               + [ J A E W 3 + ε J s p a r s e W 3 + λ J g u i d e W 3 + μ J w a s W 3 ] = λ J g u i d e W 1 + μ J w a s W 3
Since W 1 and W 3 must have the adjacent gradient value and are updated in interactive, the sum of partial derivative of J g u i d e with respect to W 1 and partial derivative of J w a s with respect to W 3 is updated. W 2 is computed as follows:
J I G S A E W 2 = J A E W 2 + ε J s p a r s e W 2 + λ J g u i d e W 2 + μ J w a s W 2 = J A E W 2 + λ J g u i d e W 2 + μ J w a s W 2
The main operating process of IGSAE is shown in the Algorithm 1 and Figure 3 as follows:
Algorithm 1: IGSAE
Input: Guided feature X (input data), Guiding feature G
1: Setting hyper parameters: ε ,     μ ,     and     λ . Number of hidden layers and units.
2: Pretraining about G 1 of the first guiding layer using Equation (10)
3: Training start
4:      Compute S and     G 1 of the first guiding layer and the J s p a r s e and     J f g terms.
5:      Compute X of the reconstruction layer and J A E term.
6:      Compute   G 2 of the second guiding layer and the J g u i d e and     J w a s terms.
7:      Compute partial derivative J A E W 1 ,     J A E W 2 ,     J A E W 1 ,     J A E W 2   , and J l a b e l W 3 to compute the gradients by Equations (22)–(25)
8:     Update the weights using gradient descent.
9: Repeat until iterations.
Output: Reconstructed guided feature X

4. Experimental Result of IGSAE

4.1. Datasets and Experimental Setups

In this paper, experiments are conducted using six public datasets with various dimensionalities: Heart Disease (Heart), Wine quality (Wine), U.S. Postal Service (USPS), Fashion-MNIST, Yale face (Yale), and Columbia Object Image Library (COIL-100) datasets. The features of the dataset are divided into low-dimensional features, medium-dimensional features, and high-dimensional features. Brief information of the dataset is summarized in Table 1. We carry out the experiments on a personal server with the following configurations: two CPUs (AMD 7742 64-core), eight GPUs (NVIDIA DGX A100), memory (1 TB), and NVMe Storage (15 TB).
The proposed IGSAE has three regularization terms: an interactive guiding term ( J g u i d e ), a sparse regularization term ( J s p a r s e ), and a Wasserstein regularization term ( J w a s ) from Equation (18). We find the optimal value of ε ,     λ , and μ in the equation for different datasets. In particular, λ and μ affect the leverage of the guiding feature. Thus, the classification performance depends on the sensitively tuned hyperparameters. The parameters of IGSAE that contain guiding layers and the coefficients are summarized in Table 2.

4.2. Authentication of Main Innovation

The main innovation of this study is to solve the overfitting problem of the autoencoder model. Overfitting of the autoencoder is mainly divided into two categories: learning the distribution of the identity function and transforming unrelated data such as training data [39,40,41,42]. To solve the two overfitting categories, experiments were conducted in two phases. First, we designed an experiment to prove that IGSAE is not learning the distribution of identity functions. Second, the experiment was conducted to test whether the deep features extracted from IGSAE could potentially express the main information. In this section, an ablation study was applied to prove the performance of the basic structure of IGSAE and the effects of sparse regularization and Wasserstein regularization. Table 3 shows the results of using SVM as a classifier. The symbols in Table 3 are explained as follows. RF: raw features; DF: deep features trained by IGSAE without Wasserstein regularization and sparse regularization; DF & WR: deep features with Wasserstein regularization; and DF & WR & SR: deep features with Wasserstein regularization and sparse regularization. As shown in Table 1 and Table 3, our methods have low-dimensional features compared to conventional RF in most datasets. However, the results show no significant performance degradation. DF & WR (DW) and DF & WR & SR (DWS) showed excellent results as follows: DWS shows 0.17% higher on the Heart; DWS shows 0.59% lower on the Wine; DW shows 1.42% lower on the USPS; DWS shows 1.73% lower on the Fashion-MNIST; DWS shows 3.21% lower on the Yale; and DWS shows 0.86% lower on the COIL-100. In most cases, the performance slightly decreases, but the training efficiency increases a lot as follows: the efficiency increases about 31% from 13 features to 9 on the Heart; the efficiency increases about 33% from 12 to 8 on the Wine; the efficiency increases about 29% from 256 to 180 on the USPS; the efficiency increases about 29% from 784 to 549 on the Fashion-MNIST; the efficiency increases about 30% increase from 1024 to 717 on the Yale; and the efficiency increases about 30% increase from 1024 to 717 on the COIL-100. Most of the results show a high learning efficiency of about 30%. Therefore, this strategy has been proven to be effective and can improve the performance and stability of the existing autoencoder model.

4.3. Comparison of the Benchmark Datasets

The efficiency of the proposed IGSAE is proven in Section 4.2, and in this section, we experiment to prove that IGSAE shows better performance than existing dimensionality reduction algorithms and does not learn the distribution of identity function. When the autoencoder model learns the distribution of the identity function to minimize the reconstruction error, it is difficult to expect improvement in classification because there are few transformations of the feature. Therefore, if the classification accuracy is improved compared to the result of using only the raw features by the hybrid feature of the deep features extracted from IGSAE and the raw features, it shows that our IGSAE is not overfitted without learning the distribution of the identity function. In Table 4, Hybrid IGSAE showed improvement in accuracy of 0.75%, 1.09%, 0.45%, 0.48%, 2.02%, and 0.57% for 6 datasets, respectively. This means that IGSAE does not learn the distribution of identity functions. In addition, Table 4 illustrates how much performance improvement of IGSAE was achieved compared to the existing feature extraction methods. The proposed IGSAE is compared with representative autoencoder models and dimensionality reduction methods such as Kernel PCA, LDA, stacked autoencoder (SAE), stacked sparse autoencoder (SSAE), uniform manifold approximation and projection (UMAP), and t-Stochastic neighbor embedding (t-SNE) [43,44,45]. As shown in the table, the proposed method shows superior achievement in most cases compared to other methods on the six datasets. The IGSAE on the datasets Heart, USPS, Fashion-MNIST, and COIL-100 shows the best outcome. Moreover, IGSAE shows slightly lower performance by 0.62% and 0.13% on the Wine and Yale compared to the SSAE and Kernal PCA, respectively. In addition, the accuracy of Hybrid IGSAE on the dataset Heart is higher by 0.75% and 0.58% than RF. Similarly, the performance increases by 1.09% and 1.06% on the Wine, 0.45% and 1.87% on the USPS, 0.48% and 2.21% on the Fashion-MNIST, 2.02% and 5.23% on the Yale, and 0.57% and 1.43% on the COIL-100, respectively, to RF and the best algorithm. This means that the proposed method can avoid overfitting the autoencoder and display superior results.

4.4. Prevention of Transforming Irrelevant Data Such as Training Data

In Section 4.3, we proved that IGSAE does not learn the distribution of identity functions, but basically, the autoencoder performance relies on reconstruction quality, which is the most important component in the encoding–decoding process. The reconstruction error can still be an insufficient indicator due to the other type of overfitting. Thus, additional experiments are necessary. We compare the reconstruction error with the actual image simultaneously and show that the proposed IGSAE can solve the overfitting that transforms data unrelated to the training set into the distribution of the training set. In Figure 4, the experimental setup of the IGSAE on Fashion-MNIST and COIL-100 is illustrated in Table 2. In addition, Gaussian noise was used as an extra test set to confirm the transformation of the data distribution. In Figure 4, the IGSAE shows better restoration ability than the basic autoencoder model for detailed information. However, in Table 5, the reconstruction error of the IGSAE on Fashion-MNIST and COIL-100 is higher than the basic autoencoder. This means that the overfitting occurred because the basic autoencoder, which has a lower reconstruction error, shows lower performance than the IGSAE. In the figure, the result of using Gaussian noise as an input value shows that most of the noise information is vaguely expressed in the case of a basic autoencoder. If there are many similar data in the training set, the reconstruction error can be reduced by transforming the information. However, the main information can be lost. IGSAE shows that this overfitting is improved through the guiding layer, Wasserstein regularization, and sparse regularization.

4.5. Effect of the Coefficients and Classifiers

4.5.1. Coefficients Optimization on the Six Datasets

In this experiment, two optimization techniques are used to find the optimum of the three coefficients ε ,     λ , and μ . The first one is a grid search algorithm that searches all values within a specific range and the second is a Bayesian optimization algorithm that probabilistically finds the optimal parameter composed of a surrogate model (Gaussian process) and an acquisition function [46,47,48]. Figure 5 shows the experimental results for coefficients ε ,     λ , and μ using grid search, and the ranges are { 0.005 × e | e = 1 ,   2 ,   3 , ,   10 } ,   { 0.01 × e | e = 1 ,   2 ,   3 , ,   10 } and { 0.001 × e | e = 1 ,   2 ,   3 , ,   10 } . Since λ is relatively more influential than ε and μ , we reduce the range of the sensitive coefficient by the results of a grid search and, additionally, test the coefficient using the Bayesian optimization algorithm to obtain a suitable value efficiently. The six datasets are split into two subsets: one accounting for about one-fifth of all samples as the test data, and the rest as the training data. We ran tests five times with randomly shuffled samples. As a classifier, SVM was chosen to optimize all coefficients. Figure 5a shows the results of the parameter experiment on the Heart dataset. As the Wasserstein regularization parameter increases, the classification accuracy decreases. It shows superior outputs for values between about 0.003 and 0.004. λ is the most important coefficient for extracting core features, and the optimal value should be obtained through experiments since the performance varies greatly depending on the data. If λ is too low, it causes a huge loss of the guiding feature and low accuracy of classification. If λ is too high, the weights are updated to improve the performance of guiding layers only. It brings a high reconstruction error of the IGSAE. The μ is the role to prevent overfitting. If μ is too low, the first guiding and second guiding layers are not interactive. In this case, the imbalanced guiding layers are updated only to reduce the reconstruction error, which causes overfitting easily. If μ is too high, the first guiding layer and the second guiding layer become almost the same. This is too stiff to be optimal and leads to high reconstruction error. The ε filters sparse nodes out. It may increase or decrease the performance depending on the data and hidden unit. In Figure 4c, IGSAE shows better results when sparse regularization is not applied. Therefore, the parameter optimization process to obtain useful features is imperative in IGSAE. Figure 5b–f are the optimization results of Wine, USPS, Fashion-MNIST, Yale, and COIL-100, respectively.

4.5.2. Effects of Classifier

The classifier has a significant impact on the performance of IGSAE. We conducted an experiment using the following three representative classifiers: support vector machine (SVM), K-nearest neighbor (KNN), and random forest [49]. The experimental results are shown in Table 6. As a result, KNN shows poor performance. The random forest achieves higher accuracies than KNN and has the highest accuracy for Yale. However, it does not show optimal performance because all hyperparameters are set as default. Since SVM can be applied stably to the proposed algorithm and shows excellent accuracies in most cases, it was chosen as a final classifier for IGSAE. Since the proposed IGSAE is in the field of dimensionality reduction and feature extraction domain, further experiments on different classifiers were not carried out. However, if more suitable and optimized classifiers are used in future work, performance can be improved, and it can be extended to other tasks.

5. Conclusions and Future Work

In this paper, we introduced a new method called IGSAE that works with interactive guiding layers. It is a dimensionality reduction method that combines the ideas of feature selection and feature extraction. There are two key points. One is using guiding layers to flexibly update the weights. This offers the advantages of feature selection and manages the effects of each feature to reduce classification performance loss. The other is a constraint method by Wasserstein distance and KL divergence. Wasserstein distance prevents overfitting in guiding layers, and KL divergence restricts the useless nodes of the network. These two main ideas interactively perform to extract the main information of the origin feature and reduce the dimensionality. In the experiment, we compared IGSAE with other traditional dimensionality reduction methods and show better performance. The proposed method outperformed other methods on average and showed the highest classification accuracies because the coefficients of J g u i d e and J w a s are efficiently optimized through the experimental tests.
As a future work, we need to expand our methodology to vision tasks by applying convolutional layers to ensure compatibility for image data and high-dimensional data representation [50]. Dimensionality reduction methods are used not only for efficiency and performance but also for visualization to show a representative data structure.

Author Contributions

Conceptualization, S.K.; methodology, L.H.; software, H.L. and C.H.; validation, H.L. and C.H.; formal analysis, H.L. and B.I.; investigations, H.L.; resources, H.L.; data curation, H.L.; writing—original draft preparation, H.L. and S.K.; writing—review and editing, H.L., C.H., B.I. and S.K.; visualization, H.L. and B.I.; supervision, H.L. and C.H.; project administration, S.K.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Inha University Grant.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

All authors declare that they have no conflict of interest.

References

  1. Storcheus, D.; Rostamizadeh, A.; Kumar, S. A survey of modern questions and challenges in feature extraction. In Proceedings of the 1st International Workshop on Feature Extraction: Modern Questions and Challenges at NIPS 2015, PMLR2015, Montreal, QC, Canada, 11 December 2015; pp. 1–18. [Google Scholar]
  2. Zhou, F.; Fan, H.; Liu, Y.; Zhang, H.; Ji, R. Hybrid Model of Machine Learning Method and Empirical Method for Rate of Penetration Prediction Based on Data Similarity. Appl. Sci. 2023, 13, 5870. [Google Scholar] [CrossRef]
  3. Janiesch, C.; Zschech, P.; Heinrich, K. Machine learning and deep learning. Electron Mark. 2021, 31, 685–695. [Google Scholar] [CrossRef]
  4. Chen, X.; Ding, M.; Wang, X.; Xin, Y.; Mo, S.; Wang, Y.; Han, S.; Luo, P.; Zeng, G.; Wang, J. Context autoencoder for self-supervised representation learning. arXiv 2022, arXiv:2202.03026. [Google Scholar]
  5. Aguilar, D.L.; Medina-Pérez, M.A.; Loyola-Gonzalez, O.; Choo KK, R.; Bucheli-Susarrey, E. Towards an interpretable autoencoder: A decision-tree-based autoencoder and its application in anomaly detection. IEEE Trans. Dependable Secur. Comput. 2022, 20, 1048–1059. [Google Scholar] [CrossRef]
  6. Wang, Y.; Yao, H.; Zhao, S. Auto-encoder based dimensionality reduction. Neurocomputing 2016, 184, 232–242. [Google Scholar] [CrossRef]
  7. Liou, C.-Y.; Cheng, W.-C.; Liou, J.-W.; Liou, D.-R. Autoencoder for words. Neurocomputing 2014, 139, 84–96. [Google Scholar] [CrossRef]
  8. Li, J.; Luong, M.-T.; Jurafsky, D. A hierarchical neural autoencoder for paragraphs and documents. arXiv 2015, arXiv:1506.01057. [Google Scholar]
  9. Tschannen, M.; Bachem, O.; Lucic, M. Recent advances in autoencoder-based representation learning. arXiv 2018, arXiv:1812.05069. [Google Scholar]
  10. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 1–45. [Google Scholar] [CrossRef] [Green Version]
  11. Jović, A.; Brkić, K.; Bogunović, N. A review of feature selection methods with applications. In Proceedings of the 2015 38th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 25–29 May 2015; pp. 1200–1205. [Google Scholar]
  12. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  13. Osia, S.A.; Taheri, A.; Shamsabadi, A.S.; Katevas, K.; Haddadi, H.; Rabiee, H.R. Deep private-feature extraction. IEEE Trans. Knowl. Data Eng. 2018, 32, 54–66. [Google Scholar] [CrossRef] [Green Version]
  14. Ghojogh, B.; Samad, M.N.; Mashhadi, S.A.; Kapoor, T.; Ali, W.; Karray, F.; Crowley, M. Feature selection and feature extraction in pattern analysis: A literature review. arXiv 2019, arXiv:1905.02845. [Google Scholar]
  15. Schmidt, M.; Fung, G.; Rosales, R. Fast optimization methods for L1 regularization: A comparative study and two new approaches. In Proceedings of the 18th European Conference on Machine Learning, Warsaw, Poland, 17–21 September 2007; pp. 286–297. [Google Scholar]
  16. Van Laarhoven, T. L2 regularization versus batch and weight normalization. arXiv 2017, arXiv:1706.05350. [Google Scholar]
  17. Azhagusundari, B.; Thanamani, A.S. Feature selection based on information gain. Int. J. Innov. Technol. Explor. Eng. 2013, 2, 18–21. [Google Scholar]
  18. Bryant, F.B.; Satorra, A. Principles and practice of scaled difference chi-square testing. Struct. Equ. Model. A Multidiscip. J. 2012, 19, 372–398. [Google Scholar] [CrossRef]
  19. Mika, S.; Schölkopf, B.; Smola, A.J.; Müller, K.-R.; Scholz, M.; Rätsch, G. Kernel PCA and De-noising in feature spaces. Adv. Neural Inf. Process. Syst. 1998, 11, 536–542. [Google Scholar]
  20. Ding, C.; Zhou, D.; He, X.; Zha, H. R 1-pca: Rotational invariant l 1-norm principal component analysis for robust subspace factorization. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 281–288. [Google Scholar]
  21. Andrew, G.; Arora, R.; Bilmes, J. Livescu, Deep canonical correlation analysis. In Proceedings of the International Conference on Machine Learning, PMLR2013, Atlanta, GA, USA,, 17–19 June 2013; pp. 1247–1255. [Google Scholar]
  22. Yu, H.; Yang, J. A direct LDA algorithm for high-dimensional data—With application to face recognition. Pattern Recognit. 2001, 34, 2067–2070. [Google Scholar] [CrossRef] [Green Version]
  23. Martinez, A.M.; Kak, A.C. Pca versus lda. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 228–233. [Google Scholar] [CrossRef] [Green Version]
  24. Zhou, P.; Han, J.; Cheng, G.; Zhang, B. Learning compact and discriminative stacked autoencoder for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4823–4833. [Google Scholar] [CrossRef]
  25. Sun, M.; Wang, H.; Liu, P.; Huang, S.; Fan, P. A sparse stacked denoising autoencoder with optimized transfer learning applied to the fault diagnosis of rolling bearings. Measurement 2019, 146, 305–314. [Google Scholar] [CrossRef]
  26. Coutinho, M.G.; Torquato, M.F.; Fernandes, M.A. Deep neural network hardware implementation based on stacked sparse autoencoder. IEEE Access 2019, 7, 40674–40694. [Google Scholar] [CrossRef]
  27. Shi, Y.; Lei, J.; Yin, Y.; Cao, K.; Li, Y.; Chang, C.-I. Discriminative feature learning with distance constrained stacked sparse autoencoder for hyperspectral target detection. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1462–1466. [Google Scholar] [CrossRef]
  28. Xiao, Y.; Wu, J.; Lin, Z.; Zhao, X. A semi-supervised deep learning method based on stacked sparse auto-encoder for cancer prediction using RNA-seq data. Comput. Methods Programs Biomed. 2018, 166, 99–105. [Google Scholar] [CrossRef] [PubMed]
  29. Sankaran, A.; Vatsa, M.; Singh, R.; Majumdar, A. Group sparse autoencoder. Image Vis. Comput. 2017, 60, 64–74. [Google Scholar] [CrossRef]
  30. Chai, Z.; Song, W.; Wang, H.; Liu, F. A semi-supervised auto-encoder using label and sparse regularizations for classification. Appl. Soft Comput. 2019, 77, 205–217. [Google Scholar] [CrossRef]
  31. Xu, H.; Luo, D.; Henao, R.; Shah, S.; Carin, L. Learning autoencoders with relational regularization. In Proceedings of the International Conference on Machine Learning, PMLR2020, Virtual Event, 13–18 July 2020; pp. 10576–10586. [Google Scholar]
  32. Vayer, T.; Chapel, L.; Flamary, R.; Tavenard, R.; Courty, N. Fused Gromov-Wasserstein distance for structured objects. Algorithms 2020, 13, 212. [Google Scholar] [CrossRef]
  33. Liang, J.; Liu, R. Stacked denoising autoencoder and dropout together to prevent overfitting in deep neural network. In Proceedings of the 2015 8th International Congress on Image and Signal Processing (CISP), Shenyang, China, 14–16 October 2015; pp. 697–701. [Google Scholar]
  34. Goldberger, J.; Gordon, S.; Greenspan, H. An Efficient Image Similarity Measure Based on Approximations of KL-Divergence Between Two Gaussian Mixtures. In Proceedings of the Ninth IEEE International Conference on Computer Vision, ICCV2003, Nice, France, 13–16 October 2003; pp. 487–493. [Google Scholar]
  35. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar]
  36. Yang, Z.; Xu, B.; Luo, W.; Chen, F. Autoencoder-based representation learning and its application in intelligent fault diagnosis: A review. Measurement 2022, 189, 110460. [Google Scholar] [CrossRef]
  37. Zheng, Q.; Zhao, P.; Zhang, D.; Wang, H. MR-DCAE: Manifold regularization-based deep convolutional autoencoder for unauthorized broadcasting identification. Int. J. Intell. Syst. 2021, 36, 7204–7238. [Google Scholar] [CrossRef]
  38. Li, Y.; Lei, Y.; Wang, P.; Jiang, M.; Liu, Y. Embedded stacked group sparse autoencoder ensemble with L1 regularization and manifold reduction. Appl. Soft Comput. 2021, 101, 107003. [Google Scholar] [CrossRef]
  39. Steck, H. Autoencoders that don′t overfit towards the Identity. Adv. Neural Inf. Process. Syst. 2020, 33, 19598–19608. [Google Scholar]
  40. Probst, M.; Rothlauf, F. Harmless Overfitting: Using Denoising Autoencoders in Estimation of Distribution Algorithms. J. Mach. Learn. Res. 2020, 21, 2992–3022. [Google Scholar]
  41. Kunin, D.; Bloom, J.; Goeva, A.; Seed, C. Loss landscapes of regularized linear autoencoders. In Proceedings of the International Conference on Machine Learning, PMLR2019, Long Beach, CA, USA, 9–15 June 2019; pp. 3560–3569. [Google Scholar]
  42. Pretorius, A.; Kroon, S.; Kamper, H. Learning dynamics of linear denoising autoencoders. In Proceedings of the International Conference on Machine Learning, PMLR2018, Stockholm, Sweden, 10–15 July 2018; pp. 4141–4150. [Google Scholar]
  43. Bunte, K.; Haase, S.; Biehl, M.; Villmann, T. Stochastic neighbor embedding (SNE) for dimension reduction and visualization using arbitrary divergences. Neurocomputing 2012, 90, 23–45. [Google Scholar] [CrossRef] [Green Version]
  44. McInnes, L.; Healy, J.; Melville, J. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  45. Becht, E.; McInnes, L.; Healy, J.; Dutertre, C.A.; Kwok, I.W.; Ng, L.G.; Newell, E.W. Dimensionality reduction for visualizing single-cell data using UMAP. Nat. Biotechnol. 2018, 37, 38–44. [Google Scholar] [CrossRef]
  46. Wang, H.; van Stein, B.; Emmerich, M.; Back, T. A new acquisition function for Bayesian optimization based on the moment-generating function. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5–8 October 2017; pp. 507–512. [Google Scholar]
  47. Snoek, J.; Larochelle, H.; Adams, R.P. Practical bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar]
  48. Audet, C.; Denni, J.; Moore, D.; Booker, A.; Frank, P. A surrogate-model-based method for constrained optimization. In Proceedings of the 8th Symposium on Multidisciplinary Analysis and Optimization, Long Beach, CA, USA, 6–8 September 2000; p. 4891. [Google Scholar]
  49. Lin, W.; Wu, Z.; Lin, L.; Wen, A.; Li, J. An ensemble random forest algorithm for insurance big data analysis. IEEE Access 2017, 5, 16568–16575. [Google Scholar] [CrossRef]
  50. Nikoloulopoulou, N.; Perikos, I.; Daramouskas, I.; Makris, C.; Treigys, P.; Hatzilygeroudis, I. A Convolutional Autoencoder Approach for Boosting the Specificity of Retinal Blood Vessels Segmentation. Appl. Sci. 2023, 13, 3255. [Google Scholar] [CrossRef]
Figure 1. The virtual structure of the autoencoder using ELM.
Figure 1. The virtual structure of the autoencoder using ELM.
Applsci 13 07055 g001
Figure 2. The architecture of IGSAE. It consists of first guiding layer for the reconstruction of guided features and second guiding layer to reflect the distribution of guiding features.
Figure 2. The architecture of IGSAE. It consists of first guiding layer for the reconstruction of guided features and second guiding layer to reflect the distribution of guiding features.
Applsci 13 07055 g002
Figure 3. Step-by-step flowchart of the proposed method.
Figure 3. Step-by-step flowchart of the proposed method.
Applsci 13 07055 g003
Figure 4. Reconstructed image of Overfitted autoencoder model.
Figure 4. Reconstructed image of Overfitted autoencoder model.
Applsci 13 07055 g004
Figure 5. Effects of the coefficients for classification accuracies on the six datasets: (a) Heart; (b) Wine; (c) USPS; (d) Fashion-MNIST; (e) Yale; (f) COIL-100.
Figure 5. Effects of the coefficients for classification accuracies on the six datasets: (a) Heart; (b) Wine; (c) USPS; (d) Fashion-MNIST; (e) Yale; (f) COIL-100.
Applsci 13 07055 g005aApplsci 13 07055 g005b
Table 1. Summary of six public datasets.
Table 1. Summary of six public datasets.
DatasetClassesFeaturesTrainingTestingDimensionality after Feature Extraction
Low-dimensionHeart213240639
Wine212487316248
Medium-dimensionUSPS1025672912007180
Fashion-MNIST1078460,00010,000549
High-dimensionYale15102413035717
COIL-10010010241080360717
Table 2. Parameters of IGSAE for six datasets.
Table 2. Parameters of IGSAE for six datasets.
DatasetGuiding Units of IGSAE ε λ μ
Heart64 – 32 – 640.0400.1190.004
Wine64 – 32 – 640.0350.1660.003
USPS256 – 128 – 25600.0250.001
Fashion-MNIST256 – 128 – 2560.0300.1150.003
Yale512 – 256 – 5120.0300.1710.002
COIL-100512 – 256 – 5120.0350.1720.003
Table 3. Ablation study on key components of IGSAE.
Table 3. Ablation study on key components of IGSAE.
DatasetRF (%)DF (%)DF & WR (%)DF & WR & SR (%)
Heart82.47 ± 2.181.24 ± 1.382.44 ± 1.882.64 ± 2.0
Wine96.84 ± 0.594.62 ± 2.096.12 ± 0.696.25 ± 1.1
USPS94.92 ± 3.493.15 ± 1.693.50 ± 2.493.24 ± 2.3
Fashion-MNIST90.60 ± 0.486.76 ± 2.388.46 ± 2.188.87 ± 1.8
Yale74.56 ± 2.170.51 ± 5.771.08 ± 4.271.35 ± 5.5
COIL-10096.19 ± 0.293.87 ± 1.194.56 ± 1.295.33 ± 0.5
Table 4. Classification accuracies of three different classifiers by seven dimensionality reduction methods.
Table 4. Classification accuracies of three different classifiers by seven dimensionality reduction methods.
DatasetRF
(%)
Kernel PCA
(%)
LDA
(%)
SAE
(%)
SSAE
(%)
UMAP
(%)
t-SNE
(%)
IGSAE
(%)
Hybrid IGSAE
(%)
Heart82.47 ± 2.181.82 ± 1.176.86 ± 4.582.23 ± 2.881.82 ± 2.774.38 ± 5.260.33 ± 9.282.64 ± 2.083.22 ± 1.7
Wine96.84 ± 0.596.00 ± 1.293.15 ± 1.495.98 ± 1.596.87 ± 1.195.54 ± 2.295.52 ± 2.996.25 ± 1.197.93 ± 0.8
USPS94.92 ± 3.491.46 ± 5.692.78 ± 5.793.47 ± 3.993.23 ± 4.589.23 ± 9.986.36 ± 8.493.50 ± 2.495.37 ± 3.1
Fashion-MNIST90.60 ± 0.487.07 ± 3.587.27 ± 3.287.98 ± 2.788.65 ± 1.881.28 ± 6.481.38 ± 4.788.87 ± 1.891.08 ± 1.0
Yale74.56 ± 2.171.48 ± 6.868.26 ± 3.667.44 ± 7.470.78 ± 4.169.33 ± 5.561.45 ± 8.571.35 ± 5.576.58 ± 4.2
COIL-10096.19 ± 0.294.55 ± 1.392.95 ± 1.595.01 ± 0.995.12 ± 1.190.03 ± 4.689.85 ± 4.295.33 ± 0.596.76 ± 0.8
Table 5. Reconstruction loss of the autoencoder models.
Table 5. Reconstruction loss of the autoencoder models.
DatasetBasic AE
(%)
SAE
(%)
SSAE
(%)
IGSAE
(%)
Heart0.35210.32130.34940.3132
Wine0.45600.46510.44760.4242
USPS0.17570.16340.16540.1537
Fashion-MNIST0.24630.25780.25630.2487
Yale0.87540.87430.88640.8435
COIL-1000.47490.48360.47980.4803
Table 6. Classification accuracies of the IGSAE under different classifiers.
Table 6. Classification accuracies of the IGSAE under different classifiers.
HeartWineUSPSFashion-MNISTYaleCOIL-100
SVM (%)82.64 ± 2.096.25 ± 1.193.50 ± 2.488.87 ± 1.871.35 ± 5.595.33 ± 0.5
KNN (%)77.48 ± 6.392.78 ± 3.289.75 ± 4.783.79 ± 2.969.95 ± 4.186.46 ± 4.6
Random Forest (%)81.72 ± 1.894.55 ± 2.190.56 ± 2.286.78 ± 2.772.17 ± 4.894.8 ± 11.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, H.; Hur, C.; Ibrokhimov, B.; Kang, S. Interactive Guiding Sparse Auto-Encoder with Wasserstein Regularization for Efficient Classification. Appl. Sci. 2023, 13, 7055. https://doi.org/10.3390/app13127055

AMA Style

Lee H, Hur C, Ibrokhimov B, Kang S. Interactive Guiding Sparse Auto-Encoder with Wasserstein Regularization for Efficient Classification. Applied Sciences. 2023; 13(12):7055. https://doi.org/10.3390/app13127055

Chicago/Turabian Style

Lee, Haneum, Cheonghwan Hur, Bunyodbek Ibrokhimov, and Sanggil Kang. 2023. "Interactive Guiding Sparse Auto-Encoder with Wasserstein Regularization for Efficient Classification" Applied Sciences 13, no. 12: 7055. https://doi.org/10.3390/app13127055

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop