Next Article in Journal
A Novel Walking Activity Recognition Model for Rotation Time Series Collected by a Wearable Sensor in a Free-Living Environment
Previous Article in Journal
Prediction Techniques on FPGA for Latency Reduction on Tactile Internet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Holistic Strategy for Classification of Sleep Stages with EEG

1
Department of Artificial Intelligence Convergence, Hallym University, Chuncheon 24252, Korea
2
Department of ECE, Bannari Amman Institute of Technology, Sathyamangalam 638401, India
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(9), 3557; https://doi.org/10.3390/s22093557
Submission received: 11 March 2022 / Revised: 3 May 2022 / Accepted: 5 May 2022 / Published: 7 May 2022

Abstract

:
Manual sleep stage scoring is usually implemented with the help of sleep specialists by means of visual inspection of the neurophysiological signals of the patient. As it is a very hectic task to perform, automated sleep stage classification systems were developed in the past, and advancements are being made consistently by researchers. The various stages of sleep are identified by these automated sleep stage classification systems, and it is quite an important step to assist doctors for the diagnosis of sleep-related disorders. In this work, a holistic strategy named as clustering and dimensionality reduction with feature extraction cum selection for classification along with deep learning (CDFCD) is proposed for the classification of sleep stages with EEG signals. Though the methodology follows a similar structural flow as proposed in the past works, many advanced and novel techniques are proposed under each category in this work flow. Initially, clustering is applied with the help of hierarchical clustering, spectral clustering, and the proposed principal component analysis (PCA)-based subspace clustering. Then the dimensionality of it is reduced with the help of the proposed singular value decomposition (SVD)-based spectral algorithm and the standard variational Bayesian matrix factorization (VBMF) technique. Then the features are extracted and selected with the two novel proposed techniques, such as the sparse group lasso technique with dual-level implementation (SGL-DLI) and the ridge regression technique with limiting weight scheme (RR-LWS). Finally, the classification happens with the less explored multiclass Gaussian process classification (MGC), the proposed random arbitrary collective classification (RACC), and the deep learning technique using long short-term memory (LSTM) along with other conventional machine learning techniques. This methodology is validated on the sleep EDF database, and the results obtained with this methodology have surpassed the results of the previous studies in terms of the obtained classification accuracy reporting a high accuracy of 93.51% even for the six-classes classification problem.

1. Introduction

Sleep is one of the most important functions of the brain, and it plays a vital role in a person’s life, which includes factors such as learning ability, concentration, and memory [1]. A partial or full unconsciousness is rendered by sleep to an individual, thereby making the brain a less complicated network. Conditions such as insomnia and obstructive sleep apnea are quite frequent, and they greatly affect the physical health [2]. Sleep issues cause depression, fatigue, lack of interest in academics/office, headache, frequent colds, and joint problems, and can sometimes lead even to death. A lot of road traffic accidents and fatalities are caused by drowsiness [3]. Therefore, automatic detection and analysis of sleep patterns are quite important to trace sleep-related conditions, including fatigue, drowsiness, apnea, insomnia, and so forth [3]. For the analysis of human sleep, sleep stage scoring is the gold standard, and it helps to identify the sleep stages that are important in treating sleep disorders. Based on the polysomnographic (PSG) recordings obtained from the patients during sleep time, the scoring of sleep stages is usually considered [4]. The overnight PSG recordings include electroencephalography (EEG), electrooculography (EOG), electromyography (EMG), and electrocardiography (ECG) recordings, and the visual scoring of them is performed by experts per the guidelines modelled by Rechtschaffen and Kales (R and K) [5]. The division of the PSG recording is made into a 20 or 30 s epoch, which is further classified into wakefulness (W) stage, rapid eye movement (REM) sleep, and non-REM (NREM) stages. NREM is further divided into four stages—S1, S2, S3, and S4—based on the guidelines of R and K. Multiple signal channels are included in the PSG recordings, and therefore, the visual examination of it by an expert is highly time-consuming, expensive, and prone to a lot of human errors [6]. Moreover, when the recording of the signals is performed, the sleep efficiency of the patient can be severely disturbed as the patient sleeps in an unfamiliar environment for the full night with so many adhesive electrodes and wires attached to the patient. Therefore, when analyzing these challenges, an automated sleep stage classification system is necessary as it would mitigate the time demand of the clinicians and can easily improve the diagnosis of sleep disorders. For the analysis of sleep stages, the EEG signal plays a vital role, and single/multiple EEG channels have been utilized in the past [7]. An EEG is a very efficient modality helping in acquiring the brain signals that correspond to different states from the area of scalp surface [8]. The EEG, apart from sleep analysis, helps in many other important applications, such as motor imagery classification [9], visual feedback classification [10], subject independent brain–computer interface classification [11], schizophrenia classification [12], and epilepsy classification [13]. In this work, the design of the proposed methodology is implemented only for sleep stage classification. A lot of methods and algorithms have been proposed in the past for automated sleep stage classification systems, and few important related works are discussed as follows.
A lot of analysis has been performed in many sleep-related datasets, such as sleep-EDF dataset, expanded sleep-EDF dataset, Montreal Archive of Sleep Studies (MASS) dataset, Sleep Heart Health Study (SHHS), and Massachusetts Institute of Technology–Beth Israel Hospital (MIT–BIH) dataset, ISRUC dataset, Massachusetts General Hospital (MGH), University College Dublin Sleep Apnea Database (UCD) dataset, and CAP dataset [14]. In this work, only the related works conducted on sleep-EDF are discussed as the present work was implemented only on this dataset in a very exhaustive manner. A recent survey paper published in 2020 highlighted all the previous works in the past in the field of automated sleep stage classification, which includes all the methodologies involved, algorithms implemented, classification techniques used, and so forth, thereby easing the work of other authors not to repeat the past literature over and over again [15]. However, the most recent and relevant works in the automated sleep stage classification published in recent years on sleep-EDF dataset are provided in a short manner for the readers’ elaborate understanding. It has become a fashion in recent years to use deep learning for almost all the applications in every domain, and so papers published in the past 2 to 3 years in automated sleep stage classification utilizing deep learning are discussed as follows. A convolutional neural network (CNN) design was implemented in [16,17,18,19] for automated sleep stage classification, and they reported classification accuracies of 92.5%, 84.5%, 83.6%, and 81.3%, respectively. Attention CNN produced a classification accuracy of 93.7% in [20], and a one-dimensional 1D-CNN was implemented in [21], producing a classification accuracy of 90.8% for EEG signals, 89.8% for EOG signals, and 91.2% for EEG and EOG combined signals, respectively. Elman recurrent neural network (RNN) analysis was used in [22], reporting a classification accuracy of 87.2%. A multitask CNN was utilized in [23], reporting a classification accuracy of 82.3% for EEG and EOG combined signals and 81.9% for EEG signals, respectively. A deep neural network (DNN) was implemented, reporting a classification accuracy of 86.1% in [24]. Hybrid deep learning models for automated sleep stage classification utilized CNN with bidirectional LSTM (CNN-BiLSTM) [25], CNN with bidirectional RNN (CNN-BiRNN) [26], RNN-LSTM [27], and CRNN [28] reporting classification accuracies of 82.0%, 84.3%, 86.7%, and 83.9%, respectively. Generally, the features of the signals extracted with or without dimensionality reduction and later classified by a classification procedure is the standard protocol followed. For feature extraction techniques, time domain methods, frequency domain methods, nonlinear complex methods, and so forth are utilized widely [29]. The common time domain methods used in the past for feature extraction are standard statistical methods, such as mean, variance, standard deviation, skewness, kurtosis, threshold percentile, median, Shannon entropy, Renyi entropy, zero crossing, Hjorth parameters, detrended fluctuation analysis, mutual information, and Tsallis entropy. The common frequency domain methods utilized in the past include nonparametric analysis, parametric analysis, higher-order spectra (HOS), median frequency, harmonic parameters, coherence analysis, Itakura distance, spectral entropy, and so forth [29]. The time–frequency domain methods include wavelet transform, signal decomposition, short-time Fourier transform, empirical mode decomposition, energy distribution, and Choi–Williams technique. Other nonlinear parameters involved in the past for feature extraction included correlation dimension, Lempel–Ziv complexity, Lyapunov exponent, fractal dimension, approximate entropy, sample entropy, autoregressive coefficients, phase space components, Hurst exponent, energy operators, permutation operator, and multiscale entropy [29]. Feature selection techniques involved in the past for sleep stage classification includes fuzzy C-means clustering, minimum redundancy maximum relevance, sequential techniques, artificial immune clustering, large margin neural network, fast correlation-based filter, fisher score, t-test, recursive feature elimination, principal component analysis, ReliefF method, and metaheuristic algorithms such as differential evolution, genetic algorithm, and particle swarm optimization (PSO) feature selection methods. Many machine learning techniques, such as linear discriminant analysis (LDA), support vector machine (SVM), artificial neural networks (ANN), naïve Bayesian classifier (NBC), quadratic discriminant analysis (QDA), K-nearest neighbor (KNN), decision trees (DT), Adaboost, K-means classifier, Gaussian mixture model (GMM), hidden Markov model (HMM), and bagging, have been utilized for sleep stage classification in the past [30]. A lot of concerns always exist with the automatic sleep stage classification system, such as deviation in the ranges of classification accuracy with sensitivity and specificity measures, careful selection of versatile feature extraction and selection methods, and execution of advanced classification techniques. Many other considerations, such as strong mathematical modelling, good computational time, high generalization and stability, and prospects for good hardware implementation in real-time situations, are also considered while developing automated sleep stage classification systems. On analyzing all the past literature, in this paper something novel was implemented, and the major contributions of the work are as follows:
(a)
Initially, the clustering was implemented to EEG signals, and the clustering methodology incorporates hierarchical clustering, spectral clustering, and the proposed PCA-based subspace clustering techniques, which is the first of its kind to implement all the three techniques for EEG signal processing utilized for automated sleep stage classification.
(b)
The dimensionality of the signals was then reduced with the help of the proposed SVD-based spectral algorithm and the standard VBMF. Though VBMF is already existing in the literature, very few works have been reported on its application for reducing the dimensionality of EEG, and so it is considered in this work along with the proposed SVD-based spectral technique.
(c)
The features were extracted and selected with the help of techniques, such as the proposed sparse group lasso technique with dual-level implementation (SGL-DLI) and the proposed ridge regression technique with limiting weight scheme (RR-LWS). Both these two developed novel techniques have been successfully utilized in our work.
(d)
Finally, classification happens with the less explored multiclass Gaussian process classification (MGC), the proposed RACC method, and the deep learning technique using LSTM, and the performance is compared with the other conventional machine learning techniques too.
A very good mathematical modelling was provided for all the proposed techniques, and the interesting factor is the novel convergence of all the proposed techniques, which makes the whole paper in general very interesting and easy to perform the experiment and provide better results than the previous works. The workflow of the methodology is shown in Figure 1.
The organization of the work is as follows: In Section 2, the clustering and dimensionality reduction techniques are discussed, followed by the usage of feature extraction and selection techniques in Section 3. Section 4 discusses the usage of classifiers, followed by the results and discussion in Section 5 and ended with the conclusion in Section 6.

2. Clustering and Dimensionality Reduction Techniques

2.1. Clustering Techniques

The primary task of assimilating a group of objects in such a manner in which the objects in the same cluster/group are quite similar to each other compared with those present in the other cluster/group is called clustering. In this work, hierarchical clustering, spectral clustering, and the proposed PCA-based subspace clustering are utilized for the clustering, and they are applied to the signals once the preprocessing is done with the help of independent component analysis (ICA).

2.1.1. Hierarchical Clustering

One of the famous and strong manifestations of the curse of dimensionality problem is that the points considered from high dimensional distributions are quite far from their nearest neighbors, and to address the noise and outliers associated with it becomes a huge challenge [31]. In order to model the low-dimensional structure, various assumptions are imposed on the data such that the clusters should be drawn from the affine subspaces. Spectral clustering usually takes place when the cluster shape is unknown or when it deviates severely from the linear structure. With respect to the geometry of the clusters considering the noise and outliers, this clustering seems to be a very popular and effective approach. An initial distance or a similarity measure is required by the spectral clustering as the operation of it is performed on a graph constructed between the neighbors assessed and the weights dependent on such distances. In the procedure of assessing the groupings within the data and based on these groupings, the assigning of labels to the data points without supervision is performed, and the procedure is termed clustering. In some circumstances, it can perform well, but in some other circumstances, it can never perform well as we have learned with K-means clustering, fuzzy C-means clustering, and so forth [31]. Statistical assumptions are usually placed on the data so that a good performance assessment is provided. The most famous clustering technique is K-means along its variants and is utilized with many feature extraction techniques for EEG signal processing. However, in this paper, a different attempt to utilize other kinds of clustering is implemented, and the notations utilized for the clustering concept are as follows: The data points to the clusters are denoted by X = { x i } i = 1 n d , the intrinsic dimension of cluster sets is expressed by d , the number of clusters is denoted by K , and the discrete data clustering is denoted by { X l } l = 1 K . The discrete noise data are represented as X , and the denoised data are represented as X N . The number of points that remains in the cluster is denoted as n min , and W represents the weight matrix. The arbitrary value is represented by ρ . A family of clusters is built at distinct hierarchical levels by the hierarchical clustering algorithm [31]. The initiation of the individual points is performed as their own clusters, and then the merging of it is performed in an iterative manner unless it reaches a stopping criterion, and therefore, the algorithms are agglomerative. The merging of the clusters at a certain iteration is determined by the agglomerative techniques by utilizing a clustering dissimilarity metric ρ c . For two clusters, C i , C j , ρ c ( C i , C j ) implies that the clusters are strong candidates for merging purposes. For every data point in X , let the metric be expressed as ρ X , and along with the standard ρ c , the corresponding clustering techniques include:
ρ S L ( C i , C j ) = min x i C i , x j C j ρ X ( x i , x j ) ,   single   linkage   clustering
ρ C L ( C i , C j ) = max x i C i , x j C j ρ X ( x i , x j ) ,   complete   linkage   clustering

2.1.2. Spectral Clustering

To define an embedding of the data, a spectral decomposition of a Laplacian matrix is utilized, and then the embedded data are clustered using a standard algorithm called K-means. On the data, a weighted graph is constructed that can specify the local relationships. For the points that are far apart from each other, the graph has very low edge weights. For the points that are very close to each other, the graph has very high edge weights. Then the partitioning of the graph is performed into clusters so that small edge weights are present between each cluster and large edge weights are within each cluster. A kernel function is denoted here as f σ : [ 0 , 1 ] with a specific scale parameter σ . Assume that W i j = f σ ( ρ ( x i , x j ) ) is the respective weight matrix for a given metric ρ : D × D [ 0 , ) and some discrete set X = { x i } i = 1 n D . The degree of point x i is expressed as d i = j = 1 n W i j , and the diagonal degree matrix D i 1 = d i , D i j = 0 is also defined for i j . By expressing L = D W , the graph Laplacian is defined and is normalized to get the symmetric Laplacian L S Y M = I D 1 2 W D 1 2 . It can also be normalized to obtain a random walk Laplacian L R W = I D 1 W . To define an embedding, utilizing the eigenvectors of L leads to un-normalized spectral clustering, while the eigenvector of L S Y M leads to normalized spectral clustering. The within-cluster similarity is always maximized by normalized spectral clustering and is generally utilized in practice.
The spectral clustering with L S Y M is considered, and then the spectral embedding is constructed [32]. In order to indicate the matrix L S Y M computed on the data set X , L S Y M ( X , ρ , f σ ) is utilized effectively with a metric ρ and kernel f σ . The eigenvalues of L S Y M are specified by λ 1 λ n , and the respective eigenvectors are denoted by ϕ 1 , , ϕ n . The data have to be clustered into K groups, and initially, a n × K matrix Φ is formed where columns are expressed by { ϕ i } i = 1 k , where the K eigenvectors are called as the K principal eigenvectors. To obtain matrix V , the normalization of the rows of Φ is performed and is expressed as:
V i j = Φ i j | ( Σ j Φ i j 2 ) 1 2
The rows of V are specified by { v i } i = 1 n K . If g : D K is implemented to specify the spectral embedding, then v i = g ( x i ) . At the end, the clustering of { v i } i = 1 n is performed into K groups by applying K-means where the partition of data points { x i } i = 1 n is expressed. Similarly, L R W can be utilized. A vital aspect of spectral clustering is to choose K . In order to estimate the total number of clusters as the largest empirical eigenmap, the eigenvalues of L S Y M are often used and are represented as K ^ = arg max i λ i + 1 λ i . It should also be observed that λ k ^ + 1 λ k ^ is maximal and also λ i should be close to zero for i k ^ . The spectral clustering algorithm utilized in this work is given in Algorithm 1. When utilizing a sparse Laplacian especially, where the sparse nearest neighbor graph is defined by W , this algorithm can be utilized.
Algorithm 1: Spectral clustering process with a metric ρ .
  • Input: (data) { x i } i = 1 n , (kernel function) f σ , and (scaling parameter) σ > 0
  • Output: Y (labels) with clustered values
    •  1. The weight matrix W n × n is computed with W i j = f σ ( ρ ( x i , x j ) ) .
    •  2. The diagonal degree matrix D n × n is computed with D i i = j = 1 n W i j .
    •  3. The symmetric normalized Laplacian L S Y M = I D 1 2 W D 1 2 is formed.
    •  4. The eigendecomposition { ( ϕ k , λ k ) } k = 1 n is computed and sorted so that 0 = λ 1 λ 2 λ n .
    •  5. The number of clusters K is estimated as: K ^ = arg max k λ k + 1 λ k .
    •  6. The row normalized spectral embedding is defined by v i = ( ϕ 1 ( x i ) , ϕ 2 ( x i ) , , ϕ k ^ ( x i ) ) / ϕ 1 ( x i ) , ϕ 2 ( x i ) , , ϕ k ^ ( x i ) 2 for 1 i n .
    •  7. By implementing K-means on the data { v i } i = 1 n , the labels Y are computed by utilizing K ^ as the total number of clusters, thereby implementing the concept of clustering successfully.

2.1.3. Proposed PCA-Based Subspace Clustering

Subspace clustering is an important technique, and the mainstream approach has two important phases, such as calculation of the affinity matrix, followed by the application of spectral clustering [33]. To enhance the scalability of sparse subspace clustering, many techniques have been proposed in the literature. A random subset of the whole dataset in clustering is considered, and then it utilizes these clusters to group the output of sample data points. When the random subset is small or large, this technique can be scaled well. From the raw dataset X , which has N observations of various parameters as input, the clustering assignment for each point is considered an output in the dataset. In the sample clustering stage, a subset X ˜ is drawn from n N points. The ( d max + 1 ) nearest neighbor points in X is found out for every point x ˜ i X ˜ . The index set of these points is denoted by C i . Therefore, the subclustering of x ˜ i is called X c i . The affirmative matrix D is computed by each element [ D ] i j and is nothing but the similarity computed between X c i and X c j . By eliminating the spurious connections with the implementation of principal component analysis (PCA), the affinity matrix is sparsified [34]. With the sparsified affinity matrix, the spectral clustering on X ˜ is conducted. To the clustered points in X ˜ , a classifier is fit, and the points may be classified, but in our case, the dimensionality of it is still reduced, and the best features are extracted and selected so that a very good classification accuracy can be obtained at a later stage. A total of n subclusters are formulated with the help of the sampled dataset. These n subclusters are divided and grouped into k clusters. The linear model of subspaces is central to the concept of clustering. Around every sampled point, the neighborhood of points is computed by applying a thresholding based on the similarity score of inner products. There is huge dependence on the self-representative property of linear subspaces here, and therefore, the concept of distance in between the subclusters is developed to build an affinity matrix. The proposed PCA-based subspace clustering is expressed in Algorithm 2 as follows.
Algorithm 2: Proposed PCA-based subspace clustering.
  • Input: Data X , number of subspaces k , sampling size n , regularization parameter λ 1 and λ 2 , neighborhood threshold d max , residual minimization parameter m , affinity threshold t max .
  • Output: The label vector l of all points in X with clustered values.
    •  1. The uniform sampling of n points X ˜ from X is performed.
    •  2. The subclusters are constructed.
    •  3. Implement PCA on the subclusters.
    •  4. An affinity matrix is constructed.
    •  5. The adjacency matrix is sparsified.
      •   For j = 1 t o n d o
      •    w : = [ D ] j
      •   For i = 1 t o n d o
      •   If [ D ] i j w ( n d max ) , then
      •     | D | i j : = 0
      •    End
      •   End
    •  6. Cluster X ˜ : s e t D : = D + D T .
    •  7. Sample points in X ˜ are clustered by implementing spectral clustering on D .
    •  8. Indicate the labels of X ˜ by l i n .
    •  9. The label of the entire dataset X is obtained by combining l i n and l o u t so that the entire l can be obtained and the clustering is performed successfully.

2.2. Dimensionality Reduction Techniques

To reduce the overall dimension, dimensionality reduction techniques are highly useful, and techniques incorporated here are the proposed SVD-based spectral algorithm and the standard variational Bayesian matrix factorization technique. Once the clustering of the signals is done using the above three techniques, the dimensionality of it is reduced so that the aim of achieving a high classification accuracy is achieved later.

2.2.1. SVD-Based Spectral Algorithm

Here in this approach, an undirected graph G = ( [ n ] , E ) is initially assembled along with an unknown vector r n , where the score related with node i is expressed as r i for the obtained clustered values. The assumption of G is considered G ( n , p ) , where the edges between the vertices have a huge independence with a probability p . For each i , the assumption is that r i is handled uniformly, r i [ 0 , M ] . Therefore, r i r j [ M , M ] for all i , j . M is not considered to be known to the algorithm. A noisy and independent measurement R i j is obtained for every { i , j } E as
R i j = { r i r j ; w . p η U [ M , M ] ; w . p ( 1 η )
In order to control the noise level, the parameter η [ 0 , 1 ] is used, and the indication of the noise level in an explicit manner is performed by γ = 1 η . The parameters η and p are not considered to be known to the algorithm.
The measurement matrix H n × n is formed by initializing the following conditions:
H i j = 0 , i = 1 , , n
H i j = R i j   and   H j i = R i j ,   if   ( i , j ) E
H i j = 0 ,   if   ( i , j ) E
where R i j denotes the independent measurements.
The main intention is to recover the score vector r and also to recover the ranking π , which is induced by r . The complete graph G along with the noise-free measurement conditions makes H = r e T e r T , which is nothing but a rank 2 skew-symmetric matrix. By specifying α = r T e n , it can be understood that the two nonzero left singular vectors are u 1 = e / n , u 2 = r α e r α e 2 with equal nonzero singular vector σ 1 = σ 2 = r α e 2 n [35].
Training a vector orthonormal to e / n is important for any orthonormal basis for span { u 1 , u 2 } so that the candidate solutions ± r α e r α e 2 are obtained. In order to recover the scale information of r , these candidates are multiplied by σ 1 / n . By selecting the best candidate that has high consistency among the measurements, the resolving of the sign ambiguity is performed easily. Therefore, ranking and synchronization is implemented here as it is a famous spectral technique to recover the ranks and scores of items. The application of SVD for ranking and synchronization is expressed in Algorithm 3 as follows.
Algorithm 3: SVD for ranking and synchronization.
  • Input: Measurement graph G = ( [ n ] , E ) and pairwise measurement R i j for { i , j } E assigned for the clustered values.
  • Output: Rank estimates: π ^ and score estimates r ^ n considered as the dimensionally reduced values.
    •  1. Measurement matrix formation H n × n by utilizing R i j .
    •  2. Trace the top 2 left singular vectors of H , namely, u ^ 1 , u ^ 2 .
    •  3. As an orthogonal projection of u 1 = e / n onto space { u ^ 1 , u ^ 2 } ,obtain vector u ¯ 1 .
    •  4. Unit vector u ˜ 2 s p a n { u ^ 1 , u ^ 2 } is obtained.
    •  5. Rank recovery: induced by u ˜ 2 , the ranking π ˜ is obtained.
    •  6. Minimize the number of upsets and reconcile its global sign.
    •  7. The ranking estimate π ^ is found out.
    •  8. Score recovery: To recover the scale τ , u ˜ 2 , H is utilized and the output is expressed, giving the dimensionally reduced values as:
      r ^ = r u ˜ 2 e T ( T u ˜ 2 ) n e

2.2.2. Variational Bayesian Matrix Factorization

In order to uncover a low-rank latent structure of data, matrix factorization is utilized, where a product of two factor matrices is obtained by approximating the data matrix [36]. For the purposes of collaborative prediction, the most famous technique utilized is matrix factorization, where the user and item factor matrices are used to predict the unknown ratings; therefore, the approximation of a user-item matrix as their respective product can be analyzed well. Assuming that Z P × Q indicates a user-item rating matrix, Z p q of the ( p , q ) entry indicates the user rating p on item q . The factor matrices U = [ u 1 , , u P ] K × P and V = [ v 1 , , v Q ] K × Q are determined by the matrix factorization so that the rating matrix Z is approximated by U T V .
Z U T V
Here, the rank of the factor matrices is denoted by K .
The regularized squared error loss is minimized and expressed as:
( p , q ) Ω [ ( Z p q u p T v q ) 2 + λ ( u p 2 + v q 2 ) ]
where a collection of indices of the observed entries in Z is represented by Ω and the regularization parameter is represented by λ . By alternating the stochastic gradient descent techniques or the least squares, the solving of the problem (7) can be performed in an efficient manner. The metaparameters, such as learning rate, regularization parameters, and the total number of iterations, should be carefully tuned so that the overfitting on the training data is avoided. All the model parameters are integrated so that the overfitting problem is alleviated successfully by means of the implementation of the Bayesian concept on matrix factorization. Thus, without the need for more parameter tuning, the learning of the complex models can be conducted easily. The side information can be easily incorporated by the Bayesian matrix factorization by implementing the Gaussian priors on user and item factor matrices. To the respective side information, each prior can be repressed so that the time and space complexity is reduced. With respect to the rank K, a cubic time and quadratic space complexity is obtained by VBMF as the variational distributions are considered to be matrixwise independent. An additional cubic time and quadratic space complexity is necessary if the incorporation of the side information is performed to the VBMF depending on the feature vector size obtained by the side information. Thus, with the prohibition of the usage of rich side information, high-dimensional feature vector is achieved, and the dimensionally reduced values are obtained. In order to satisfy the element-wise independence, the full factorization of the variational distribution is performed.

3. Feature Extraction and Selection Techniques

Once the dimensionality is reduced, the features have to be extracted and selected, and therefore, the two techniques proposed here are the sparse group lasso technique with dual-level implementation and a ridge regression technique with limiting optimal weight scheme.

3.1. Proposed Sparse Group Lasso Technique with Dual-Level Implementation

To identify the significant groups and features in a simultaneous manner, the most powerful regression technique utilized is sparse group lasso (SGL). The lasso and group lasso are combined by the SGL so the sparsity can be yielded at both the individual and group feature levels [37]. SGL has been implemented in machine learning, bioinformatics, signal processing, and so forth. A two-layer feature screening technique called dual layer features is proposed here. The inactive groups and features are quickly identified by this method, and ultimately, zero coefficients are guaranteed in the solution. To deal efficiently with multiple sparsity-inducing regularities, a dual-level technique is widely used. Through the framework of Fenchel duality, the dual feasible solution of SGL is developed [38]. The upper bounds should be estimated so that an efficient dual-level technique is developed.
Assume that 1 , , is indicated as the l 1 , l 2 and l norms, respectively. The unit l 1 , l 2 and l norm balls in n are denoted by B 1 n , B n and B n , respectively. For set C, assume that int C is its interior value. Assume that Γ 0 ( n ) is the class of proper close convex function on n . The domain of f is the set dom f : = { w : f ( n ) < } . Assume [ w ] i as the i t h component for w n . G { 1 , 2 , , n } is considered an index set, and the corresponding subvector of w is denoted by [ w ] G | G | , where the number of elements in G is denoted by | G | .
Assume that y N is the response vector and X P N × q is the matrix of features. The SGL problem is expressed here with the group information available and represented as:
min β q 1 2 y g = 1 G X g β g 2 + λ 1 + g = 1 G n g β g + λ 2 β 1
where the number of features in the g t h group is represented as n g . The predictors in the group with the respective coefficient vector β g are expressed as X g N × n g .
The positive regularization parameters are represented as λ 1 , λ 2 . Without loss of generality, assume λ 1 = α λ and λ 2 = λ with α > 0 . Equation (8), therefore, can be written as follows:
min β q 1 2 y g = 1 G X g β g 2 + λ ( α g = 1 G n g β g + λ 2 β 1 )
The dual problem of SGL can be obtained as follows using the Lagrangian techniques as:
s u p θ 1 2 y 2 1 2 y λ θ 2
such that:
X g T θ D g α : = α n g β + β , g = 1 , , G
The intersection of closed half spaces enables the dual feasible set of lasso. Using Fenchel’s duality theorem, the dual feasible set of SGL is analyzed well. For every X g T θ D g α , Fenchel’s duality leads to an explicit decomposition X g T θ = b 1 + b 2 , where one belongs to α n g β and the other belongs to B . The procedure for developing dual-level feature extraction and selection is expressed in Algorithm 4 as follows.
Algorithm 4: Procedure for developing dual-level feature extraction and selection.
  • Estimate a region θ that has dual optimum θ * ( λ , α ) of Equations (10) and (11) for a given pair of parameter values ( λ , α ) .
  • The two optimization problems are solved as follows:
    s g * = sup ξ g { S 1 ( ξ g ) : ξ g Ξ g X g T Θ } , where X g T Θ = { X g T θ : θ Θ } ,
    t g k * = sup θ { | x g k T θ | : θ Θ } , where x g k is the k t h column of X g
    .
  • The dual feature screening ensures the form as:
    s g * < α n g β g * ( λ , α ) = 0
    t g k * 1 [ β g * ( λ , α ) ] k = 0 ,
    where the optimal solution of SGL in (9) is expressed as β * ( λ , α ) , giving the best extracted and selected features.

3.2. Proposed Ridge Regression Technique with Limiting Optimal Weight Scheme

For the dimensionally reduced values, a subset of samples is considered initially, and then ridge regression is trained on these local data. The local dataset is arranged into a feature matrix X i , where every row has a sample or data point along with an outcome vector Y i , where each entry is an outcome. The local ridge regression estimates [39] are computed as follows:
β ^ i = ( X i T X i + λ i I p ) 1 X i T Y i
where the regularization parameter is termed as λ i . By using a weighted combination, the aggregation of them is performed so that a single-shot distributed ridge estimator is constructed as:
β ^ d i s t = i = 1 q w i β ^ i
where q represents the number of sites. By a finite sample analysis of estimation error in linear models, the distributed ridge regression can be studied well. The standard linear model is considered here as Y = X β + ε . For n independent samples, the n-dimensional continuous outcome vector is represented as Y n . X is the n × p design matrix having the values of p features for each sample. The p-dimensional vector of unknown regression coefficients is expressed as β = ( β 1 , , β p ) T T . In order to predict the outcome variable of future samples and to firmly estimate the respective coefficients, this technique is used. Random noise can greatly affect the outcome vector ε = ( ε 1 , , ε n ) T n . The coordinates of ε are assumed to be independent random variables with zero mean and variance σ 2 . For estimation and prediction purposes in linear models, the ridge regression estimation is the most widely used. The ridge estimator of β is recalled as follows [39]:
β ^ ( λ ) = ( X T X + n λ I p ) 1 X T Y
where λ denotes a tuning parameter. Many justifications are present in their estimation. An improved estimation can be performed as the coefficient of the ordinary least squares estimators are shrunk. Supposing that the distribution of the samples is performed across q different sites or machines, the partitioning is performed and expressed as follows:
X = [ X 1 : : X q ] , Y = [ Y 1 : : Y q ] ,
Therefore, for the sake of approximation in a distributed setting, ridge regression estimation is widely used. When the ridge regression is performed locally on every subset of the data, a one-shot weighting technique is given more focus, and finally, the regression coefficients are aggregated by a weighted sum. The weighting technique is utilized as it serves as a useful method of initialization to iterative techniques. Moreover, a variety of new phenomena about one-shot weights can be discovered easily. Therefore, for every dataset X i , Y i , the local ridge estimators are defined with a regularization parameter λ i and are expressed as follows:
β ^ i ( λ i ) = ( X i T X i + n i λ i I p ) 1 X i T Y i
By using a weighted one-shot distributed estimation summation, the local ridge estimators are combined and expressed as:
β ^ d i s tan c e ( w ) = i = 1 q w i β ^ i
The local ridge estimators are well defined, and they are not like ordinary least squares (OLS). As the ridge estimators are biased, it is not necessary to consider whether any constraints should be added on the weights or not. The proposed algorithm works well for designs X with arbitrary covariance structures Σ . Assuming the samples distributed to be n , it is considered that there is an equal distribution of the samples. A local ridge estimator β ^ 1 is computed along with the local estimators σ ^ i 2 and α ^ i 2 of the SNR and the noise level. The qualities necessary to find the optimal weights are m , m and λ . The procedure of Ridge regression with limiting optimal weights is expressed in Algorithm 5 as follows.
Algorithm 5: Ridge regression with limiting optimal weights.
  • Input: Data matrices ( n i × p ) and outcomes ( n i × 1 ) , ( X i , Y i ) distributed across q sites.
  • Output: Distributed ridge estimator β ^ d i s t of regression coefficients β indicating the best features extracted and selected.
    •  1. For i 1 to q do.
         Calculate the MLE θ ^ i = ( σ ^ i 2 , α ^ i 2 ) locally.
         Progress θ ^ i to the global data center
         End
    •  2. Get a global estimator θ ^ = ( σ ^ 2 , α ^ 2 ) = q 1 i = 1 q θ ^ i .
    •  3. Tuning parameters S is chosen around the initial guess λ 0 = q p / ( n α ^ 2 ) .
    •  4. For λ S do.
         For i 1 t o q d o .
         Compute the local ridge estimator β ^ i ( λ ) = ( X i T X i + n i λ I p ) 1 X i T Y .
         The weight w i is computed for the i t h local estimator as w i ( λ ) = σ ^ 2 α ^ 2 ( 1 λ m ) F + q G .
         Progress β ^ i ( λ ) and w i ( λ ) to the global data center.
         End
         Terminate the performance of the distributed ridge estimator.
         End
    •  5. Select the best tuning parameter λ * .
    •  6. Output the respective distributed ridge estimator
      β ^ d i s t ( λ * ) = i = 1 q w i ( λ * ) β ^ i ( λ * )
The tuning parameter λ is chosen by the grid search process. Therefore, the limiting optimal weights too are estimated successfully by this algorithm, and the best features are extracted and selected through this technique.

4. Classification Techniques

The features extracted and selected are then fed to the classification stage. The classification techniques proposed in this work are the multiclass Gaussian process classification (MGC), the proposed random arbitrary collective classification (RACC), deep learning methods, and other standard conventional techniques. A standard 10-fold cross-validation technique was utilized for all the implemented pattern recognition and machine learning techniques.

4.1. Multiclass Gaussian Process Classification

The multiclass classification problems can be well addressed by Gaussian processes [40], and it is as follows: a dataset comprising N instances with X = ( x 1 , , x N ) T as the observed explaining attributes and y = ( y 1 , , y N ) T as the target class labels, where y i { 1 , , C } and C > 2 are the number of classes. Making predictions about the label y * of a new instance x * is the primary task of interest, given the observed data X and y . Every class label has been obtained by utilizing the labelling rule for the multiclass classification with the Gaussian process as follows:
y i = arg max c f c ( x i )
where f c ( . ) , for c = 1 , , C are the various latent functions, and each of them communicates with various class labels. By analyzing the latent function with the highest value at the data point y i , the class label can be obtained. Assume f i = ( f ( x i ) , , f C ( x i ) ) T . The likelihood of the values of every latent function at a training point under this labelling rule is expressed by:
p ( y i | f i ) = c y i Θ ( f y i ( x i ) f c ( x i ) )
where the Heaviside step function is denoted as Θ ( ) . By analyzing and marginalizing noise present around the latent functions f c ( ) , other likelihood functions, such as the softmax likelihood, are considered. Around each f c , the Gaussian noise is considered. The actual class label y i , which is related to x i , is considered to account for the labelling errors, and it could have been replaced with a particular probability ε so that some other class labels are reached. Therefore, the likelihood becomes as follows:
p ( y i | f i ) = ( 1 ε ) c y i Θ ( f ( y i ) ( x i ) f c ( x i ) ) + ε C 1 [ 1 c y i Θ ( f y i ( x i ) f c ( x i ) ) ]
For every latent function f c ( ) , the assumptions of a GP prior are performed so that a multiclass classification with GPs is addressed. Therefore, it is represented as p ( f c ) G P ( 0 , k θ ( , ) ) , where k θ c ( , ) represents a covariance function with hyperparameter θ c . A famous example of covariance function includes the squared exponential covariance function and is represented as follows:
k θ c ( x , x ) = σ 2 exp { 1 2 j = 1 d ( x j x j ) 2 l j } + I [ x = x ] σ 0 2
where the indicator function is represented as I [ ] and θ c = { σ 2 , σ 0 , { l j } j = 1 d } , which are the hyperparameters. The length scale is represented by l j , the amplitude parameter is represented by σ 2 , and the level of additive Gaussian noise and f c is represented by σ 0 2 . For each latent function f c ( ) , the hyperparameter will be quite different from each other. The posterior distribution of f = { f i } i = 1 y is computed so that the predictions about the potential class label of a new data point x * is made. The latent function values that are pretty compatible with the observed data are summarized by this distribution. Using Baye’s rule, the computation of the posterior distribution is performed as follows:
p ( f | y ) = p ( y | f ) p ( f ) p ( y ) = [ i = 1 N p ( y i | f i ) ] [ c = 1 C p ( f c ) ] p ( y )
where f c = ( f c ( x 1 ) , , f c ( x N ) ) T and p ( f c ) = N ( f c | 0 , K c ) are a multivariate Gaussian distribution with zero mean and covariance matrix K c with K i , j c = k θ c ( x i , x j ) . The normalization constant is represented as p ( y ) = p ( y | f ) p ( f ) d f and is known as marginal likelihood. In order to get good values for the model hyperparameters θ c , it can be maximized well. Computing the marginal likelihood is slightly difficult, and so to approximate the posterior, inference methods are utilized. A famously used inference technique is variational inference technique. The main advantage of using variational inference is that it can transform the approximate inference issue into a goal optimization problem, and it can be solved easily using stochastic optimization techniques.

4.2. Proposed Random Arbitrary Collective Classification

For classification tasks, a very famous framework is ensemble classification, which is nothing but a combination of the consequences of a lot of weak learners to get the ultimate classification [41]. The stability and accuracy of weak classifiers are improved greatly, always leading to a higher performance than the individual weak classifier. Here, a random arbitrary collective classification (RACC) is proposed that can be hybrid with any base classifier. The base classifier used here is SVM. RACC is thus a very flexible ensemble classification framework. Assume that the observation pair ( x , y ) considers values from X × { 0 , 1 } , where X denotes an open subset of q , q represents a positive integer, and the class label is represented by y . A total of n observation pairs { ( x i , y i ) , i = 1 , , n } are assumed in the training set. To indicate the prediction result of the classifier, C n S T ( x ) { 0 , 1 } is utilized. B 2 random subspaces { S j k } k = 1 B 2 are generated from the j t h ( j { 1 , , B 1 } ) weak learner. The optimal one S j * is chosen based on some criterion to be mentioned. By utilizing only a portion of training samples in this subspace S j * , the training of the weak learner is performed. To form the decision function, the aggregation of the B 1 weak classifiers C n S 1 * T , , C n S B 1 * T is performed as:
C n R A C C ( x ) = 1 ( 1 B 1 j = 1 B 1 C n S j * T ( x ) > α )
where α indicates a threshold, which has to be determined. A flexible framework can be admitted here, where any selected classification techniques can act as the base classifiers, such as LDA, KNN, SVM, QDA, and DT. The ranking on the significance of variables in the B 1 subspaces { b j * } j = 1 B 1 is explained by this ensemble process. The minimal discriminative set for every learner can be covered easily with this procedure, and the methodology is as follows.
Assuming that n pairs of observations are present { ( x i , y i ) , i = 1 , n } ~ ( x , y ) X × { 0 , 1 } , where X denotes an open subset of q , q indicates a positive integer and y = { 0 , 1 } is the class label. To specify the whole feature set, S F U L L = { 1 , , q } is utilized. For classes 0 (y = 0) and 1 (y = 1), the marginal densities of x are assumed and expressed as f ( 0 ) and f ( 1 ) . The respective probability estimates they influence are indicated as p ( 0 ) and p ( 1 ) . Using the following mixture model, the joint distribution of ( x , y ) can be expressed as follows:
x | y = y 0 ~ ( 1 y 0 ) f ( 0 ) + y 0 f ( 1 ) , y 0 = 0 , 1
where the Bernoulli variable is represented as y with a success probability π 1 = 1 π 0 ( 0 , 1 ) .
To express the cardinality, | S | is used for any subspace S . The probability estimate observed by the marginal distribution of x is indicated as Q x , which is expressed as π 0 Q ( 0 ) + π 1 Q ( 1 ) . For classes 0 and 1, the respective marginal densities are expressed as f s ( 0 ) and f s ( 1 ) when they are restricted to the feature subspace S . The generation of the B 2 independent arbitrary subspaces is performed as S j 1 , , S j B 2 so that each weak learner can be trained. Then the selection of the optimal subspace S j * is performed based on some important criterion, and only in S j * , the weak learners are trained, and therefore, the B 1 weak classifiers { C n S j * T } j = 1 B are obtained. The final decision function is obtained by means of aggregation of the outputs of { C n S j * T } j = 1 B by computing a simple average. Algorithm 6 expresses the whole procedure in detail.
Algorithm 6: RACC.
  • Input: Training data { ( x i , y i ) } i = 1 n , new data x , type of base classifiers T , subspace distribution D , integers B 1 and B 2 , criterion C .
  • Output: Predicted label C n R A C C ( x ) , the chosen proposition of every feature η .
    •  1. Generate random subspaces independently, S j k ~ D , 1 j B 1 , 1 k B 2 .
    •  2. For j 1 to B 1 do.
         Choosing of optimal subspace S j * is performed from { S j k } k = 1 B 2 based on C and T .
         End
    •  3. Develop the collective decision function as an ensembled one, and represent it as:
      v n ( x ) = 1 B 1 j = 1 B 1 C n S j * T ( x )
      .
    •  4. Based on Equation (2), the threshold is set.
    •  5. The predicted label C n R A C C ( x ) = 1 ( v n ( x ) > α ^ ) is given as output, which is the chosen proposition of every feature η = ( n 1 , , n q ) T .
Hierarchical uniform distribution is used to choose the subspace distribution D. From the uniform distribution over { 1 , , D } , the generation of the subspace size ‘d’ is performed. The adjustment of the subspace distribution could be performed if sufficient details are present with respect to the data structure.

4.3. LSTM Recurrent Network

One of the famous time recurrent neural networks is LSTM [42]. For predicting the time series of important events, LSTM is highly useful. The historical information can be easily retained by this neural network, and therefore, the learning of long-term dependence information is easily realized. An input gate, a forget gate, and an output gate are the most common gates contained in an LSTM network. In order to update and retain the historical information, a cell unit is utilized. Figure 2 shows the structure of an LSTM block.
By utilizing a simple single neuron, it helps to control the forget gate f t in the LSTM memory block. To enable the historical information storage, it helps to assess which information must be retained or discarded. The input gate i t is a part where neurons and the previous memory unit effects are used to create an LSTM block. To assess the historical information of the LSTM block, it is activated widely. Using a tanh neuron, the calculation of the candidate update content c i n is performed. By utilizing the current candidate cell c i n , input gate information i t , forget gate information f t , and the previous time state c t 1 , the current time memory cell state value c t is computed. The generation of o t for the LSTM block in the current time is performed at the output gate. The amount of information about the current cell state is determined by a t , and it is the output. The calculation of the activation of every gate along with the updation of the current cell state is performed as follows:
i t = s i g m o i d ( W i . [ a t 1 , x t , c t 1 ] + b i )
f t = s i g m o i d ( W f . [ a t 1 , x t , c t 1 ] + b f )
o t = s i g m o i d ( W o . [ a t 1 , x t , c t 1 ] + b o )
c t = f t c t + i t c i n
a t = o t . tanh ( c t )
c i n = tanh ( W c [ a t 1 , x t , c t 1 ] + b c )
For each position, the hidden vector is computed, and the last hidden vector is considered as the EEG signal representation. It is fed to a linear layer, and finally, a softmax output layer is utilized to classify the EEG. A four-layer LSTM architecture was used in this paper, which includes an input layer, an LSTM layer, and two fully connected (FC) layers. The illustration of the proposed LSTM for the EEG signal feature extraction and classification is shown in Figure 3.

Focal Loss

To deal with imbalanced datasets, one of the most effective ways is focal loss [43]. By transforming the cross-entropy (CE) loss function, it is obtained. The computation of CE is performed as follows:
C E ( y ^ ) = log ( y ^ )
A dynamically scaled CE is focal loss, where the confidence of the classification increases when the scaling factor decays to zero. The contribution of EEG examples can be automatically downweighed by this scaling factor when the model training focuses on the hard examples. The computation of FL is performed as follows:
F L ( y ^ ) = ( 1 y ^ ) γ . log ( y ^ ) , γ 0 ,
where the modulating factor is denoted by ( 1 y ^ ) γ , and the focusing parameter is expressed by γ . When the misclassification of EEG is performed and the value of y ^ is very small, then the value of the modulation factor is close to 1, and in such cases, the loss is barely affected. For the network parameters, optimization is important. Many kinds of gradient descent optimization algorithms are present, such as Adam, Nadam, Adagrad, and Adadelta. Here in this work, Adam is utilized.

5. Results and Discussion

5.1. Dataset Description

The sleep-EDF database contains raw physiological data having 61 data recordings considered from 42 Caucasian subjects [44,45]. The initial 39 recordings are considered from 20 healthy volunteers (SC-PSG.edf files), and they do not have any sleep-related disease. There were 10 males and 10 females, and at the time of recordings, the demographic range was between 25 to 34 years. The rest, 22 data records, were obtained from 22 participants (ST-PSG.edf files), and there were 7 males and 15 females within the demographic range of 18–79. These 22 subjects had the problem of falling asleep. Dual-channel EEG from FPz-Cz and Pz-Oz is considered in this database with a sampling rate of 100 Hz. Many other physiological signals, such as EMG, EOG, and oronasal respiration, are present in it. To understand the automatic sleep staging, dual-channel EEG data are utilized in this work as they are effective for sleep stage classification. Based on the R and K standards, the manual scoring of the 30 s epoch was performed, and the primary annotations are named as AWA, REM, S1, S2, S3, S4, ‘Movement Time’ and ‘Unscored’. Based on the R and K criteria, the total number of samples is expressed in Table 1. The total number of samples is 127,658 after movement time, and unscored categories are ignored.
Once the clustering is done, a total of 90,000 samples are obtained, and once when dimensionality reduction is obtained, a total of 30,000 samples are obtained as the dimensionality is reduced by threefold time. When the feature extraction and selection techniques are implemented, a total of 2000 samples are selected, and finally they are fed to classification implementing a 10-fold cross-validation method. For the deep learning application model, once the clustering is performed, all the 90,000 samples are provided to it, and the classification results are obtained. The hyperparameter set for the LSTM deep learning is as follows: The number of LSTM cells is set at 64, the network layers are 4, the optimizer chosen is Adam, the dropout rate is set to 0.1 (after several trial-and-error experiments), the batch size is 128, the cost function is focal loss, and the value of focusing on parameter γ is set to 2 finally again after several trial-and-error experimentations.
Table 2 shows the results of the hierarchical clustering with SVD-based spectral algorithm dimensionality reduction technique and its performance analysis with SGL-DLI and RR-LWS for the different classifiers. When MGC is utilized, a high classification accuracy of 97.73% is obtained for two classes, 94.43% for three classes, 93.73% for four classes, 92.73% for five classes, and 92.16% for six classes under SGL-DLI technique. Similarly, when RACC is utilized, a high classification accuracy of 97.96% is obtained for two classes, 94.56% for three classes, 92.99% for four classes, 92.96% for five classes, and 92.72% for six classes under SGL-DLI technique. Similarly, when MGC is utilized, a high classification accuracy of 96.68% is obtained for two classes, 92.84% for three classes, 92.31% for four classes, 92.67% for five classes, and 91.45% for six classes under RR-LWS technique. Similarly, when RACC is utilized, a high classification accuracy of 97.55% is obtained for two classes, 91.78% for three classes, 93.56% for four classes, 91.34% for five classes, and 92.12% for six classes under RR-LWS technique. All the present results surpassed the previous results to a great extent.
Table 3 shows the results of the spectral clustering with SVD-based spectral algorithm dimensionality reduction technique and its performance analysis with SGL-DLI and RR-LWS for the different classifiers. When MGC is utilized, a high classification accuracy of 95.87% is obtained for two classes, 92.56% for three classes, 93.81% for four classes, 93.07% for five classes, and 93.51% for six classes under SGL-DLI technique. Similarly, when RACC is utilized, a high classification accuracy of 94.34% is obtained for two classes, 91.34% for three classes, 93.24% for four classes, 90.19% for five classes, and 91.20% for six classes under SGL-DLI technique. Similarly, when MGC is utilized, a high classification accuracy of 95.74% is obtained for two classes, 90.80% for three classes, 90.01% for four classes, 89.12% for five classes, and 88.75% for six classes under RR-LWS technique. Similarly, when RACC is utilized, a high classification accuracy of 95.68% is obtained for two classes, 91.01% for three classes, 90.35% for four classes, 88.38% for five classes, and 86.49% for six classes under RR-LWS technique.
Table 4 shows the results of the subspace clustering with SVD-based spectral algorithm dimensionality reduction technique and its performance analysis with SGL-DLI and RR-LWS for the different classifiers. When MGC is utilized, a high classification accuracy of 97.57% is obtained for two classes, 96.64% for three classes, 96.61% for four classes, 91.76% for five classes, and 90.49% for six classes under SGL-DLI technique. Similarly, when RACC is utilized, a high classification accuracy of 98.41% is obtained for two classes, 97.56% for three classes, 97.21% for four classes, 94.78% for five classes, and 93.12% for six classes under SGL-DLI technique. Similarly, when MGC is utilized, a high classification accuracy of 92.26% is obtained for two classes, 92.25% for three classes, 91.47% for four classes, 91.07% for five classes, and 89.23% for six classes under RR-LWS technique. Similarly, when RACC is utilized, a high classification accuracy of 96.78% is obtained for two classes, 95.97% for three classes, 94.32% for four classes, 93.89% for five classes, and 89.11% for six classes under RR-LWS technique.
Table 5 shows the results of the hierarchical clustering with VBMF dimensionality reduction technique and its performance analysis with SGL-DLI and RR-LWS for the different classifiers. When MGC is utilized, a high classification accuracy of 95.23% is obtained for two classes, 91.35% for three classes, 91.45% for four classes, 89.31% for five classes, and 90.34% for six classes under SGL-DLI technique. Similarly, when RACC is utilized, a high classification accuracy of 95.11% is obtained for two classes, 92.22% for three classes, 90.89% for four classes, 89.21% for five classes, and 92.24% for six classes under SGL-DLI technique. Similarly, when MGC is utilized, a high classification accuracy of 95.35% is obtained for two classes, 90.29% for three classes, 90.01% for four classes, 90.03 for five classes, and 90.01% for six classes under RR-LWS technique. Similarly, when RACC is utilized, a high classification accuracy of 96.11% is obtained for two classes, 90.83% for three classes, 91.09% for four classes, 89.01% for five classes, and 90.12% for six classes under RR-LWS technique.
Table 6 shows the results of the spectral clustering with VBMF dimensionality reduction technique and its performance analysis with SGL-DLI and RR-LWS for the different classifiers. When MGC is utilized, a high classification accuracy of 93.66% is obtained for two classes, 90.23% for three classes, 90.15% for four classes, 90.34% for five classes, and 89.65% for six classes under SGL-DLI technique. Similarly, when RACC is utilized, a high classification accuracy of 92.54% is obtained for two classes, 89.11% for three classes, 88.77% for four classes, 88.02% for five classes, and 87.18% for six classes under SGL-DLI technique. Similarly, when MGC is utilized, a high classification accuracy of 92.98% is obtained for two classes, 90.55% for three classes, 90.02% for four classes, 88.98% for five classes, and 87.67% for six classes under RR-LWS technique. Similarly, when RACC is utilized, a high classification accuracy of 93.09% is obtained for two classes, 89.67% for three classes, 88.11% for four classes, 86.71% for five classes, and 85.45% for six classes under RR-LWS technique.
Table 7 shows the results of the subspace clustering with VBMF dimensionality reduction technique and its performance analysis with SGL-DLI and RR-LWS for the different classifiers. When MGC is utilized, a high classification accuracy of 96.45% is obtained for two classes, 95.26% for three classes, 94.08% for four classes, 90.45% for five classes, and 90.08% for six classes under SGL-DLI technique. Similarly, when RACC is utilized, a high classification accuracy of 97.33% is obtained for two classes, 96.98% for three classes, 95.85% for four classes, 93.84% for five classes, and 92.03% for six classes under SGL-DLI technique. Similarly, when MGC is utilized, a high classification accuracy of 92.89% is obtained for two classes, 91.02% for three classes, 90.74% for four classes, 90.56% for five classes, and 88.42% for six classes under RR-LWS technique. Similarly, when RACC is utilized, a high classification accuracy of 95.01% is obtained for two classes, 94.05% for three classes, 93.13% for four classes, 92.43% for five classes, and 89.85% for six classes under RR-LWS technique.
Table 8 shows the results of the clustering methodology with deep learning LSTM method. For the two-classes classification, the classification accuracies produced are 96.35% for hierarchical clustering, 97.85% for spectral clustering, and 97.38% for subspace clustering. For the three-classes classification, the classification accuracies produced are 95.11% for hierarchical clustering, 95.99% for spectral clustering, and 96.78% for subspace clustering. For the four-classes classification, the classification accuracies produced are 94.71% for hierarchical clustering, 95.37% for spectral clustering, and 96.42% for subspace clustering. For the five-classes classification, the classification accuracies produced are 90.65% for hierarchical clustering, 93.35% for spectral clustering, and 93.22% for subspace clustering. For the six-classes classification, the classification accuracies produced are 90.42% for hierarchical clustering, 92.31% for spectral clustering, and 92.47% for subspace clustering.

5.2. Performance Comparison with Previous Works

The results obtained in this work are compared with the previous works and expressed in Table 9.
In the literature, there was only one recently published paper reporting results from two channels (Pz-Oz and Fpz-Cz), and so the present results were compared with it. Most of the papers have concentrated only on a single-channel EEG or EOG, and some have clubbed both EEG and EOG as reported in [21], and therefore, the current results obtained cannot be compared with them. Moreover, many results have utilized the extended version of sleep-EDF database, which has about 197 recordings released in 2018. It is a very huge database, and it is rarely analyzed as a whole dataset, and most of the reports have analyzed only a small portion or subset of it. Therefore, the current results cannot be compared with those results too as the database itself was completely different and is an extended version of the currently used database. Considering these points in mind, the hierarchical clustering with SVD-based spectral algorithm methodology with suitable classifiers produced a classification accuracy of 97.96% for two classes, 94.56% for three classes, 93.73% for four classes, 92.96% for five classes, and 92.72% for six classes. The spectral clustering with SVD-based spectral algorithm methodology with suitable classifiers produced a classification accuracy of 95.87% for two classes, 92.56% for three classes, 93.81% for four classes, 93.07% for five classes, and 93.51% for six classes, respectively. The subspace clustering with SVD-based spectral algorithm methodology with suitable classifiers produced a classification accuracy of 98.41% for two classes, 97.56% for three classes, 96.61% for four classes, 94.78% for five classes, and 93.12% for six classes. The hierarchical clustering with VBMF produced a classification accuracy of 96.11% for two classes, 92.22% for three classes, 91.45% for four classes, 90.03% for five classes, and 92.24% for six classes. The spectral clustering with VBMF produced a classification accuracy of 93.66% for two classes, 90.55% for three classes, 90.15% for four classes, 90.34% for five classes, and 89.65% for six classes. The subspace clustering with VBMF produced a classification accuracy of 97.33% for two classes, 96.98% for three classes, 95.85% for four classes, 93.84% for five classes, and 92.03% for six classes. The clustering methodology with deep learning LSTM produced a classification accuracy of 97.85% for two classes, 96.78% for three classes, 96.42% for four classes, 93.35% for five classes, and 92.47% for six classes. All the results obtained surpassed the previous results, and this shows that the present work is quite a versatile methodology. The statistical significance of the results too was analyzed. Cohen’s kappa coefficient was computed for the extracted and selected features, and the values ranged in the category of 0.6 to 1, proving that the values reached good agreement and sometimes very good agreement. The Friedman test analysis too was conducted for the process, and distinct values were obtained, proving the uniqueness in the selected features. The standard two-sided Wilcoxon test too was conducted, and the obtained ρ value was less than 0.05 in our experiment, thereby proving that a higher confidence level is achieved.

6. Conclusions and Future Work

Sleep disorder is a very common symptom of many neurological disorders that affects the quality of life to a great extent. Some of the common problems created due to sleep disorders are insomnia, narcolepsy, sleep-related breathing disorders, and sleep-related movement disorders. The PSG recordings of subjects are the physiological signals that are obtained during an entire night of sleep. The signal recordings, such as EEG, ECG, EOG, and EMG, are found here as PSG is a multivariate system. Once the recording is done, the scoring of sleep stages is performed on the PSG recordings by sleep experts who evaluate and grade the sleep stages. The manual determination of sleep stages is very complex and costly by means of visual inspection of PSG signals. Detecting the EEG signal variations is hard as it has a random and chaotic nature. As a result, automated sleep detection systems are developed so that the experts can be assisted well. The widely used PSG signals for the purpose of sleep stage classification are the EEG data or one or more channels. EEG is widely preferred as it is obtained using wearable technologies, and it consists of more important information. In the EEG signal processing phase, factors such as dimensionality reduction, feature extraction, and feature selection techniques are quite important, and based on that, a novel attempt to implement an interesting flow of methodology is proposed in this paper. Initially, three clustering techniques, followed by two dimensionality reduction techniques and two feature extraction cum selection techniques, were utilized and classified with around 10 classifiers to conduct an exhaustive performance analysis. Among all the results, the best results were obtained when subspace clustering with SVD-based spectral algorithm with suitable classification was performed for a two-class classification problem reporting a classification accuracy of 98.41%. Future works aim to work with many other modified versions of clustering algorithms, modified versions of dimensionality mitigation schemes, and feature extraction techniques along with plenty of other deep learning techniques to obtain a higher classification accuracy and a faster execution time with much easier applicability.

Author Contributions

Data curation, S.K.P.; Formal analysis, S.K.P.; Investigation, H.R.; Methodology, H.R.; Resources, S.R.; Software, S.R.; Supervision, I.c.J.; Validation, I.c.J.; Writing—original draft, D.-O.W.; Writing—review & editing, D.-O.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Hallym University Research Fund, 2021 (HRF-202111-010).

Data Availability Statement

All the programming codes developed can be obtained upon request to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Andlauer, O.; Moore, H.; Jouhier, L.; Drake, C.; Peppard, P.E.; Han, F.; Hong, S.-C.; Poli, F.; Plazzi, G.; O’Hara, R.; et al. Nocturnal Rapid Eye Movement Sleep Latency for Identifying Patients with Narcolepsy/Hypocretin Deficiency. JAMA Neurol. 2013, 70, 891–902. [Google Scholar] [CrossRef] [PubMed]
  2. Sharma, M.; Goyal, D.; Achuth, P.; Acharya, U.R. An accurate sleep stages classification system using a new class of optimally time-frequency localized three-band wavelet filter bank. Comput. Biol. Med. 2018, 98, 58–75. [Google Scholar] [CrossRef] [PubMed]
  3. Acharya, U.R.; Bhat, S.; Faust, O.; Adeli, H.; Chua, E.C.-P.; Lim, W.J.E.; Koh, J.E.W. Nonlinear Dynamics Measures for Automated EEG-Based Sleep Stage Detection. Eur. Neurol. 2015, 74, 268–287. [Google Scholar] [CrossRef] [PubMed]
  4. Drake, C.L.; Roehrs, T.; Roth, T. Insomnia causes, consequences, and therapeutics: An overview. Depress. Anxiety 2003, 18, 163–176. [Google Scholar] [CrossRef]
  5. Liang, S.-F.; Kuo, C.-E.; Hu, Y.-H.; Pan, Y.-H.; Wang, Y.-H. Automatic Stage Scoring of Single-Channel Sleep EEG by Using Multiscale Entropy and Autoregressive Models. IEEE Trans. Instrum. Meas. 2012, 61, 1649–1657. [Google Scholar] [CrossRef]
  6. Sharma, R.; Pachori, R.B.; Upadhyay, A. Automatic sleep stages classification based on iterative filtering of electroencephalogram signals. Neural Comput. Appl. 2017, 28, 2959–2978. [Google Scholar] [CrossRef]
  7. Prabhakar, S.K.; Rajaguru, H.; Kim, S.-H. Schizophrenia EEG Signal Classification based on Swarm Intelligence Computing. Comput. Intell. Neurosci. 2020, 2020, 8853835. [Google Scholar] [CrossRef]
  8. Prabhakar, S.K.; Rajaguru, H. Alcoholic EEG Signal Classification with Correlation Dimension Based Distance Metrics Approach and Modified Adaboost Classification. Heliyon 2020, 6, e05689. [Google Scholar] [CrossRef]
  9. Ahn, M.; Jun, S.C. Performance variation in motor imagery brain—Computer interface: A brief review. J. Neurosci. Methods 2015, 243, 103–110. [Google Scholar] [CrossRef]
  10. Trincado-Alonso, F.; Lopez-Larraz, E.; Resquín, F.; Ardanza, A.; Pérez-Nombela, S.; Pons, J.L.; Montesano, L.; Gil-Agudo, Á. A pilot study of brain-triggered electrical stimulation with visual feedback in patients with incomplete spinal cord injury. J. Med. Biol. Eng. 2017, 38, 790–803. [Google Scholar] [CrossRef]
  11. Cao, L.; Li, J.; Ji, H.; Jiang, C. A hybrid brain computer interface system based on the neurophysiological protocol and brain-actuated switch for wheelchair control. J. Neurosci. Methods 2014, 229, 33–43. [Google Scholar] [CrossRef] [PubMed]
  12. Prabhakar, S.K.; Rajaguru, H.; Lee, S.-W. A Framework for Schizophrenia EEG Signal Classification With Nature Inspired Optimization Algorithms. IEEE Access 2020, 8, 39875–39897. [Google Scholar] [CrossRef]
  13. Übeyli, E.D. Wavelet/mixture of experts network structure for EEG signals classification. Expert Syst. Appl. 2008, 34, 1954–1962. [Google Scholar] [CrossRef]
  14. Aboalayon, K.A.I.; Faezipour, M.; Almuhammadi, W.S.; Moslehpour, S. Sleep Stage Classification Using EEG Signal Analysis: A Comprehensive Survey and New Investigation. Entropy 2016, 18, 272. [Google Scholar] [CrossRef]
  15. Loh, H.W.; Ooi, C.P.; Vicnesh, J.; Oh, S.L.; Faust, O.; Gertych, A.; Acharya, U.R. Automated Detection of Sleep Stages Using Deep Learning Techniques: A Systematic Review of the Last Decade (2010–2020). Appl. Sci. 2020, 10, 8963. [Google Scholar] [CrossRef]
  16. Qureshi, S.; Karilla, S.; Vanichayobon, S. GACNN SleepTuneNet: A genetic algorithm designing the convolutional neural network architecture for optimal classification of sleep stages from a single EEG channel. Turk. J. Electr. Eng. Comput. Sci. 2019, 27, 4203–4219. [Google Scholar] [CrossRef] [Green Version]
  17. Wei, L.; Lin, Y.; Wang, J.; Ma, Y. Time-Frequency Convolutional Neural Network for Automatic Sleep Stage Classification Based on Single-Channel EEG. In Proceedings of the 2017 IEEE 29th International Conference on Tools with Artificial Intelligence, Boston, MA, USA, 6–8 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 88–95. [Google Scholar]
  18. Zhang, X.; Xu, M.; Li, Y.; Su, M.; Xu, Z.; Wang, C.; Kang, D.; Li, H.; Mu, X.; Ding, X.; et al. Automated multi-model deep neural network for sleep stage scoring with unfiltered clinical data. Sleep Breath. 2020, 24, 581–590. [Google Scholar] [CrossRef] [Green Version]
  19. Vilamala, A.; Madsen, K.H.; Hansen, L.K. Deep convolutional neural networks for interpretable analysis of EEG sleep stage scoring. In Proceedings of the 2017 IEEE 27th International Workshop on Machine Learning for Signal Processing (MLSP), Tokyo, Japan, 25–28 September 2017; pp. 1–6. [Google Scholar]
  20. Zhu, T.; Luo, W.; Yu, F. Convolution- and Attention-Based Neural Network for Automated Sleep Stage Classification. Int. J. Environ. Res. Public Health 2020, 17, 4152. [Google Scholar] [CrossRef]
  21. Yildirim, O.; Baloglu, U.B.; Acharya, U.R. A Deep Learning Model for Automated Sleep Stages Classification Using PSG Signals. Int. J. Environ. Res. Public Health 2019, 16, 599. [Google Scholar]
  22. Hsu, Y.-L.; Yang, Y.-T.; Wang, J.-S.; Hsu, C.-Y. Automatic sleep stage recurrent neural classifier using energy features of EEG signals. Neurocomputing 2013, 104, 105–114. [Google Scholar] [CrossRef]
  23. Phan, H.; Andreotti, F.; Cooray, N.; Chén, O.Y.; Vos, M.D. Joint classification and prediction CNN framework for automatic sleep stage classification. IEEE Trans. Biomed. Eng. 2019, 66, 1285–1296. [Google Scholar] [CrossRef] [PubMed]
  24. Xu, M.; Wang, X.; Zhang, X.; Bin, G.; Jia, Z.; Chen, K. Computation-Efficient Multi-Model Deep Neural Network for Sleep Stage Classification. In Proceedings of the ASSE’ 20: 2020 Asia Service Sciences and Software Engineering Conference, Nagoya, Japan, 13–15 May 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–8. [Google Scholar]
  25. Supratak, A.; Dong, H.; Wu, C.; Guo, Y. DeepSleepNet: A Model for Automatic Sleep Stage Scoring Based on Raw Single-Channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1998–2008. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Mousavi, S.; Afghah, F.; Acharya, U.R. SleepEEGNet: Automated sleep stage scoring with sequence to sequence deep learning approach. PLoS ONE 2019, 14, e0216456. [Google Scholar] [CrossRef] [Green Version]
  27. Michielli, N.; Acharya, U.R.; Molinari, F. Cascaded LSTM recurrent neural network for automated sleep stage classification using single-channel EEG signals. Comput. Biol. Med. 2019, 106, 71–81. [Google Scholar] [CrossRef] [PubMed]
  28. Seo, H.; Back, S.; Lee, S.; Park, D.; Kim, T.; Lee, K. Intra- and inter-epoch temporal context network (IITNet) using sub-epoch features for automatic sleep scoring on raw single-channel EEG. Biomed. Signal Process. Control 2020, 61, 102037. [Google Scholar] [CrossRef]
  29. Fraiwan, L.; Lweesy, K.; Khasawneh, N.; Wenz, H.; Dickhaus, H. Automated sleep stage identification system based on time–frequency analysis of a single EEG channel and random forest classifier. Comput. Methods Programs Biomed. 2011, 108, 10–19. [Google Scholar] [CrossRef] [PubMed]
  30. Wu, H.-T.; Talmon, R.; Lo, Y.-L. Assess Sleep Stage by Modern Signal Processing Techniques. IEEE Trans. Biomed. Eng. 2014, 62, 1159–1168. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Maqbool, O.; Babri, H. Hierarchical Clustering for Software Architecture Recovery. IEEE Trans. Softw. Eng. 2007, 33, 759–780. [Google Scholar] [CrossRef]
  32. Wang, L.; Ding, S.; Jia, H. An Improvement of Spectral Clustering via Message Passing and Density Sensitive Similarity. IEEE Access 2019, 7, 101054–101062. [Google Scholar] [CrossRef]
  33. Pham, D.-S.; Budhaditya, S.; Phung, D.; Venkatesh, S. Improved subspace clustering via exploitation of spatial constraints. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’ 12), Providence, RI, USA, 16–21 June 2012; pp. 550–557. [Google Scholar]
  34. Jackson, J.E. A User’s Guide to Principal Components; John Wiley & Sons: New York, NY, USA, 2004. [Google Scholar]
  35. Cong, F.; Chen, J.; Dong, G.; Zhao, F. Short-time matrix series based singular value decomposition for rolling bearing fault diagnosis. Mech. Syst. Signal Process. 2013, 34, 218–230. [Google Scholar] [CrossRef]
  36. Schmidt, M.N.; Laurberg, H. Nonnegative Matrix Factorization with Gaussian Process Priors. Comput. Intell. Neurosci. 2008, 2008, 361705. [Google Scholar] [CrossRef] [PubMed]
  37. Liu, X.; Cao, P.; Zhao, D.; Banerjee, A. Multi-task spare group lasso for characterizing Alzheimers disease. In Proceedings of the 5th Workshop on Data Mining for Medicine and Healthcare, Miami, FL, USA, 5–7 May 2016; p. 49. [Google Scholar]
  38. Boţ, R.I.; Grad, S.-M.; Wanka, G. Fenchel’s Duality Theorem for Nearly Convex Functions. J. Optim. Theory Appl. 2007, 132, 509–515. [Google Scholar] [CrossRef]
  39. Ogutu, J.O.; Schulz-Streeck, T.; Piepho, H.P. Genomic selection using regularized linear regression models: Ridge regression, lasso, elastic net and their extensions. BMC Proc. 2012, 6 (Suppl. 2), S10. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Calvo, C.V.; Zaldivar, B.; Merchan, E.C.G.; Lobato, D.H. Multi-class Gaussian Process Classification with Noisy Inputs. J. Mach. Learn. Res. 2020, 21, 1–52. [Google Scholar]
  41. Partalas, I.; Tsoumakas, G.; Vlahavas, I. A Study on Greedy Algorithms for Ensemble Pruning; Technical Report; TR-LPIS-360-12; Department of Informatics, Aristotle University of Thessaloniki: Thessaloniki, Greece, 2012. [Google Scholar]
  42. Gers, F.A.; Schraudolph, N.N.; Schmidhuber, J. Learning precise timing with LSTM recurrent networks. J. Mach. Learn. Res. 2003, 3, 115–143. [Google Scholar]
  43. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  44. Kemp, B.; Zwinderman, A.H.; Tuk, B.; Kamphuisen, H.A.; Oberye, J.J. Analysis of a Sleep-dependent Neuronal Feedback Loop: The Slow-wave Microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 2000, 47, 1185–1194. [Google Scholar] [CrossRef]
  45. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. Physiobank, physiotoolkit, and physionet. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [Green Version]
  46. Delimayanti, M.K.; Purnama, B.; Nguyen, N.G.; Faisal, M.R.; Mahmudah, K.R.; Indriani, F.; Kubo, M.; Satou, K. Classification of Brainwaves for Sleep Stages by High-Dimensional FFT Features from EEG Signals. Appl. Sci. 2020, 10, 1797. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Pictorial representation of the work.
Figure 1. Pictorial representation of the work.
Sensors 22 03557 g001
Figure 2. An LSTM representation.
Figure 2. An LSTM representation.
Sensors 22 03557 g002
Figure 3. Illustration of the LSTM implementation.
Figure 3. Illustration of the LSTM implementation.
Sensors 22 03557 g003
Table 1. Total number of samples in sleep-EDF dataset (R and K criteria).
Table 1. Total number of samples in sleep-EDF dataset (R and K criteria).
Number of ClassesAWAREMS1S2S3S4
674,82711,848484827,29250703773
574,82711,848484827,2928843
474,82711,84832,1408843
374,82711,84840,983
274,82752,831
Table 2. Hierarchical clustering with SVD-based spectral algorithm.
Table 2. Hierarchical clustering with SVD-based spectral algorithm.
SGL-DLIRR-LWS
2 Classes3 Classes4 Classes5 Classes6 Classes2 Classes3 Classes4 Classes5 Classes6 Classes
LDA93.2492.3490.6288.2486.3492.6793.6792.4589.6788.35
KNN91.3291.2690.6789.3288.3692.3290.8391.7891.1386.16
NBC89.9288.1187.1385.9284.6588.9887.5689.4387.8486.56
DT89.2788.6987.5786.2783.6888.2788.3188.2187.7981.26
RF87.5785.7384.2982.5781.8286.7684.6985.4484.5484.68
Adaboost88.3285.2583.4183.3282.7889.1586.4584.6783.6383.16
SVM96.5793.2292.6291.5790.9195.9394.3294.8692.8590.83
MGC97.7394.4393.7392.7392.1696.6892.8492.3192.6791.45
RACC97.9694.5692.9992.9692.7297.5591.7893.5691.3492.12
Table 3. Spectral clustering with SVD-based spectral algorithm.
Table 3. Spectral clustering with SVD-based spectral algorithm.
SGL-DLIRR-LWS
2 Classes3 Classes4 Classes5 Classes6 Classes2 Classes3 Classes4 Classes5 Classes6 Classes
LDA92.4690.2188.6789.4585.0994.0892.2291.0187.1186.22
KNN93.7889.3291.5391.2189.8693.8788.8189.9286.5884.79
NBC87.4587.9185.5987.4887.8287.6186.3988.6185.3183.53
DT87.3289.7289.8788.988.3488.2587.4187.7888.6980.16
RF86.1286.7887.5184.3284.5787.6885.3786.9282.8380.81
Adaboost87.3585.1685.3485.5783.8188.9387.2883.5381.4781.21
SVM95.6994.8391.6992.8989.2493.2691.6192.8090.8489.34
MGC95.8792.5693.8193.0793.5195.7490.8090.0189.1288.75
RACC94.3491.3493.2490.1991.2095.6891.0190.3588.3886.49
Table 4. Subspace clustering with SVD-based spectral algorithm.
Table 4. Subspace clustering with SVD-based spectral algorithm.
SGL-DLIRR-LWS
2 Classes3 Classes4 Classes5 Classes6 Classes2 Classes3 Classes4 Classes5 Classes6 Classes
LDA93.3491.0190.1189.0988.2195.1194.2192.3491.1189.89
KNN95.5792.4391.2490.0187.6596.8795.5892.5291.3688.36
NBC92.9791.9991.6788.8186.7993.6292.9791.6790.8789.72
DT91.4290.5190.8489.6188.8194.4194.4193.9492.4286.38
RF92.5692.1291.5788.5488.2794.6493.4692.2892.1988.67
Adaboost94.8993.9592.8289.3488.7593.9893.8792.4191.8289.93
SVM96.4295.5895.4392.7290.4394.7392.5491.8691.3488.67
MGC97.5796.6496.6191.7690.4992.2692.2591.4791.0789.23
RACC98.4197.5697.2194.7893.1296.7895.9794.3293.8989.11
Table 5. Hierarchical clustering with VBMF.
Table 5. Hierarchical clustering with VBMF.
SGL-DLIRR-LWS
2 Classes3 Classes4 Classes5 Classes6 Classes2 Classes3 Classes4 Classes5 Classes6 Classes
LDA92.0190.0488.6682.4681.4490.5289.4490.1787.1286.48
KNN90.9790.2989.6787.3283.5991.1590.2490.4388.4684.23
NBC87.5388.1186.4285.8983.8687.0586.5688.9285.8785.73
DT87.7386.7586.3183.2683.3186.5385.1487.3685.6380.56
RF85.1584.5882.5784.5184.1384.7883.8484.8282.2282.98
Adaboost89.6886.3281.8681.8785.5688.1585.1482.5781.5980.78
SVM94.7592.8790.3289.5688.9894.8492.6292.2390.9889.56
MGC95.2391.3591.4589.3190.3495.3590.2990.0190.0390.01
RACC95.1192.2290.8989.2192.2496.1190.8391.0989.0190.12
Table 6. Spectral clustering with VBMF.
Table 6. Spectral clustering with VBMF.
SGL-DLIRR-LWS
2 Classes3 Classes4 Classes5 Classes6 Classes2 Classes3 Classes4 Classes5 Classes6 Classes
LDA90.0988.8786.1187.0283.2292.2591.0491.0185.1281.47
KNN91.0387.3188.3689.4987.4591.0489.5689.2485.4582.32
NBC86.2385.6584.8983.2385.6685.0985.78843183.7881.65
DT85.5686.9885.0385.9883.7885.3084.9284.8981.7479.87
RF84.7483.2386.0382.2882.9884.1283.3483.6780.5280.47
Adaboost84.1381.8780.5683.5181.9285.5184.7883.0580.1480.23
SVM94.8791.6590.8790.7886.3491.6790.9490.3687.6785.11
MGC93.6690.2390.1590.3489.6592.9890.5590.0288.9887.67
RACC92.5489.1188.7788.0287.1893.0989.6788.1186.7185.45
Table 7. Subspace clustering with VBMF.
Table 7. Subspace clustering with VBMF.
SGL-DLIRR-LWS
2 Classes3 Classes4 Classes5 Classes6 Classes2 Classes3 Classes4 Classes5 Classes6 Classes
LDA92.4690.2390.1488.1187.5794.4693.5491.9890.6788.11
KNN94.8691.6790.7889.2487.1495.9894.8191.7690.3387.23
NBC91.4590.0990.0588.6787.5292.7291.7990.2490.2187.56
DT90.0289.3588.4188.8986.8993.1493.4692.5791.9885.78
RF91.6491.78905987.4386.5793.6992.9391.8991.5687.33
Adaboost93.0592.0690.2388.6087.2392.0392.5791.3590.1288.24
SVM95.8994.8193.0190.0289.9593.2691.3490.2590.7587.78
MGC96.4595.2694.0890.4590.0892.8991.0290.7490.5688.42
RACC97.3396.9895.8593.8492.0395.0194.0593.1392.4389.85
Table 8. Results of the clustering methodology with deep learning LSTM.
Table 8. Results of the clustering methodology with deep learning LSTM.
2 Classes3 Classes4 Classes5 Classes6 Classes
Hierarchical clustering96.3595.1194.7190.6590.42
Spectral clustering97.8595.9995.3793.3592.31
Subspace clustering97.3896.7896.4293.2292.47
Table 9. Comparison with previous works for two channels (Pz-Oz and Fpz-Cz).
Table 9. Comparison with previous works for two channels (Pz-Oz and Fpz-Cz).
ReferenceMethodologyNumber of ClassesAccuracy (%)
[46]
(Pz-Oz and Fpz-Cz)
High-dimensional FFT features with SVM classifier297.88
394.41
492.82
591.73
690.77
Proposed method
(Pz-Oz and Fpz-Cz)
Hierarchical clustering with SVD-based spectral algorithm297.96
394.56
493.73
592.96
692.72
Proposed method
(Pz-Oz and Fpz-Cz)
Spectral clustering with SVD-based spectral algorithm295.87
392.56
493.81
593.07
693.51
Proposed method
(Pz-Oz and Fpz-Cz)
Subspace clustering with SVD-based spectral algorithm298.41
397.56
496.61
594.78
693.12
Proposed method
(Pz-Oz and Fpz-Cz)
Hierarchical clustering with VBMF296.11
392.22
491.45
590.03
692.24
Proposed method
(Pz-Oz and Fpz-Cz)
Spectral clustering with VBMF293.66
390.55
490.15
590.34
689.65
Proposed method
(Pz-Oz and Fpz-Cz)
Subspace clustering with VBMF297.33
396.98
495.85
593.84
692.03
Proposed method
(Pz-Oz and Fpz-Cz)
Clustering methodology with deep learning LSTM297.85
396.78
496.42
593.35
692.47
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Prabhakar, S.K.; Rajaguru, H.; Ryu, S.; Jeong, I.c.; Won, D.-O. A Holistic Strategy for Classification of Sleep Stages with EEG. Sensors 2022, 22, 3557. https://doi.org/10.3390/s22093557

AMA Style

Prabhakar SK, Rajaguru H, Ryu S, Jeong Ic, Won D-O. A Holistic Strategy for Classification of Sleep Stages with EEG. Sensors. 2022; 22(9):3557. https://doi.org/10.3390/s22093557

Chicago/Turabian Style

Prabhakar, Sunil Kumar, Harikumar Rajaguru, Semin Ryu, In cheol Jeong, and Dong-Ok Won. 2022. "A Holistic Strategy for Classification of Sleep Stages with EEG" Sensors 22, no. 9: 3557. https://doi.org/10.3390/s22093557

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop