Next Article in Journal
Investigating the Driving Factors of Public Participation in Public-Private Partnership (PPP) Projects—A Case Study of China
Previous Article in Journal
Disparities in COVID-19 Information Sources and Knowledge in South Korea
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CAttSleepNet: Automatic End-to-End Sleep Staging Using Attention-Based Deep Neural Networks on Single-Channel EEG

1
School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
2
School of Computer and Communication Engineering, Shanghai Polytechnic University, Shanghai 201209, China
3
School of Computer Science and Technology, Kashi University, Kashi 844008, China
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(9), 5199; https://doi.org/10.3390/ijerph19095199
Submission received: 12 March 2022 / Revised: 20 April 2022 / Accepted: 22 April 2022 / Published: 25 April 2022
(This article belongs to the Topic Artificial Intelligence in Healthcare)

Abstract

:
Accurate sleep staging results can be used to measure sleep quality, providing a reliable basis for the prevention and diagnosis of sleep-related diseases. The key to sleep staging is the feature representation of EEG signals. Existing approaches rarely consider local features in feature extraction, and fail to distinguish the importance of critical and non-critical local features. We propose an innovative model for automatic sleep staging with single-channel EEG, named CAttSleepNet. We add an attention module to the convolutional neural network (CNN) that can learn the weights of local sequences of EEG signals by exploiting intra-epoch contextual information. Then, a two-layer bidirectional-Long Short-Term Memory (Bi-LSTM) is used to encode the global correlations of successive epochs. Therefore, the feature representations of EEG signals are enhanced by both local and global context correlation. Experimental results achieved on two real-world sleep datasets indicate that the CAttSleepNet model outperforms existing models. Moreover, ablation experiments demonstrate the validity of our proposed attention module.

1. Introduction

As an important physiological activity, high-quality sleep can effectively restore people’s physical and mental strength, while long-term sleep deprivation or disorder can seriously affect physical and emotional health. It has been shown that certain diseases, such as Parkinson’s disease and Alzheimer’s disease, are strongly associated with sleep disorders or abnormalities [1,2]. Therefore, it is important to improve sleep quality and prevent diseases caused by sleep disorders through a detailed scoring of sleep stages. In the process of sleep staging, sleep experts divide the polysomnography (PSG) into 30 s (30-s) epochs and mark the corresponding sleep stages of each epoch according to the Rechtschaffen and Kales (R&K) [3] and American Academy of Sleep Medicine (AASM) [4] guidelines. Sleep specialists usually label an epoch by analyzing contextual information to find important sleep-related events, such as LAMF and k-complex. However, artificial sleep staging is time-consuming and complex, and the sleep staging results produced by different sleep experts sometimes vary.
Recently, a growing number of researchers have tried to apply artificial intelligence techniques, such as machine learning and deep learning, to solve the issue of sleep staging. Machine learning-based methods usually choose appropriate features from physiological signals (i.e., EEG, EOG, and EMG) [5,6,7,8]. Then, the feature selection algorithm is used to select more representative signal features. Finally, the classifier categorizes sleep stages according to the selected features. Although these approaches have led to some achievements, they still demonstrate some problems. For instance, selecting the most discriminative sleep features requires researchers to have professional experience in the area of sleep medicine, which can be very challenging to obtain. For another, traditional machine learning algorithms are poor at modeling temporal physiological signals.
Deep learning has achieved prominence in various fields (i.e., medical imaging [9,10], emotion recognition [11,12], and bio-signal processing [13]) and provides a novel approach to sleep staging. Some researchers have used convolutional neural networks (CNNs) to construct sleep staging models [14,15,16]. These methods can improve sleep staging performance to some extent. However, traditional CNNs only focus on information within the current receptive field, ignoring context-related information. Therefore, they can easily interfere with non-key features and cannot accurately obtain key local features. Some researchers have gradually come to focus on recurrent neural networks (RNNs) [17,18,19,20,21]. Most of the literature on this subject only considers the contextual association between sleep stages from the global sequence, and only a few works have focused on the local sequence as well. Seo et al. [19] proposed a model called IITNet, which performs feature learning by considering contextual information within and between epochs. Although IITNet has achieved excellent performance, it fails to discriminate the weight of local features.
Feature extraction often plays a decisive role in the results of sleep staging. To obtain more typical feature representations for high-accuracy automatic sleep staging, we propose a novel model named CAttSleepNet. First, local sequences within a 30-s epoch are fed into the conventional CNN module to learn the local sequence features. Second, we feed longer local sequences within a 30-s epoch that are centered on the sequence input to the CNN to the designed attention module. In this way, our proposed attention module can calculate the weights of local sequence features obtained by the CNN network by mining the contextual associations of local sequences. Finally, a two-layer Bi-LSTM is used to encode the global representation of different epochs. Accordingly, our model can more comprehensively capture temporal contextual information and learn the importance of local sequence features. Our main contributions can be summarized as follows.
  • To obtain more discriminative feature representation, our model makes full use of temporal contextual correlation at both local and global levels to achieve high-precision automatic sleep staging on single-channel EEG.
  • To solve the problem whereby traditional CNNs cannot distinguish feature importance due to their limited receptive fields, we add an attention module, which learns the weights of local features by mining the contextual relations of local sequences.
  • Our proposed model is evaluated on the public data sets sleep-edfx-2013 and sleep-edfx-2018. The experimental results show that CAttSleepNet outperforms the existing state-of-the-art methods.
The remainder of the paper is organized as follows. In Section 2, related works on sleep staging are introduced. Section 3 describes CAttSleepNet in detail. Section 4 experimentally evaluates CAttSleepNet. Section 5 concludes the paper and points out directions for future work.

2. Related Work

2.1. Machine Learning-Based Sleep Staging Methods

Many methods have been proposed for sleep staging by machine learning. For example, Li et al. [5] used 30 EEG signal features, including temporal, frequency, and nonlinear features, to train a random forest model for this task. In [6], the Naive Bayes classifier was applied to classify sleep stages on single-channel EEG. Zhu et al. [7] extracted the graph domain features of EEG and then input these features into a support vector machine model. Seifpour et al. [8] fed extracted EEG time-domain features into a support vector machine for sleep staging. Lajnef et al. [22] proposed a multi-class support vector machine (SVM) classification model based on decision trees. The model used hierarchical clustering techniques and extensive time and frequency domain feature extraction to obtain a decision tree or dendrogram. Hassan et al. [23] utilized bootstrap aggregation (Bagging) and complete ensemble empirical pattern decomposition with adaptive noise (CEEMDAN) to perform a sleep staging task. These sleep staging methods usually require specialized sleep medicine knowledge. In addition, machine learning models do not excel at processing temporal signals.

2.2. Deep Learning-Based Sleep Staging Methods

Many studies have applied deep learning to sleep staging. The authors of [14,15,16] used convolutional neural networks for sleep staging. The model shown in reference [14] combined the convolution, max-pooling, and dropout layers. This model used one fully connected layer and the softmax classifier to divide the sleep process into five categories. For this purpose, the classification method proposed in Reference [15] used seven convolution layers, one dropout layer, and one fully connected layer. Traditional CNNs perform well in feature extraction, but fail to exploit the temporal context of sleep stages due to their limited receptive fields.
In recent years, an increasing number of specialists have started to use RNNs in their classification models. Michielli et al. [17] designed a new model based on a cascaded RNN structure with long short-term memory (LSTM) blocks to process single-channel EEG signals. Sun et al. [18] proposed a two-stage sleep staging method—namely, feature learning and sleep pattern recognition. The features in this method were fused manually and trained using CNNs. The fused features were then fed to the RNN module. Supratak et al. [20] combined CNNs with Bi-LSTM, using CNNs to extract representative features from each 30-s epoch and Bi-LSTM to consider contextual correlations for consecutive epochs. The experimental results showed that considering the contextual correlation of signals is an effective technique in automatic sleep staging. Similarly, the model proposed in [21] utilized a three-scale CNN architecture to extract features for each 30-s epoch and fused hand-crafted features with extracted features. Finally, the fusion results were fed into the Bi-LSTM network to learn the transition rules between consecutive epochs. However, most of these models do not consider the contextual correlations of local sequences within a 30-s epoch and ignore the differences in the importance of local features.

3. CAttSleepNet

Figure 1 shows the overall architecture of the CAttSleepNet model. Firstly, EEG signals, including Fpz-Cz and Pz-Oz channels, are obtained from PSGs. The extracted EEG signals are divided into many 30-s epochs. Next, consecutive epochs are fed into CAttSleepNet to extract their features. After that, we input feature representations of all epochs to a two-layer Bi-LSTM to capture the time dependence of different epochs. The final sleep staging results can be obtained by adding a softmax layer after the two-layer Bi-LSTM.

3.1. The Attention-Based CNN for Local Sequence Feature Extraction

According to the AASM manual [4], sleep specialists focus on certain key features when labeling sleep stages, such as low-amplitude mixed frequency (LAMF) and vertex sharp waves in the N1 stage. The variability in the importance of temporal signal features is difficult for standard CNNs to capture, leading models to incorrectly assess key features or even directly ignore them due to their less frequent occurrence. If the CNN is given the ability to pay more attention to high-importance features, the effective extraction of local sequence features and the optimization of the input information of the Bi-LSTM layer can be achieved more easily. Therefore, we designed an attention module to solve this problem.
Attention mechanisms can be categorized as hard attention and soft attention [24]. Hard attention mechanisms filter out the regions of interest as the input, which can help the model to focus on the target object in image processing. However, this method, which directly limits the input content, is not applicable in time series signal classification and prediction. In addition, hard attention mechanisms are more difficult to train and less versatile. In contrast, soft attention mechanisms achieve the purpose of focusing on specific spatial regions or channels by obtaining weights through training learning and then weighting input features on spaces or channels. At the same time, this approach is differentiable in reverse computation, so an end-to-end method can be used to learn the attention network. Based on the above principles, we introduced a soft attention mechanism into the CNN to weight all local sequence features, focus on specific spaces and channels, and achieve the extraction of significant fine-grained features of time series.
In this work, the local feature extraction consisted of two branches: the standard CNN branch and the attention branch. The standard CNN branch was used for extracting local features. The attention branch was used to compute the corresponding attention scores. Then, the attention score was used to reweight the feature map by element-wise multiplication. This fusion method has been verified to be feasible and effective in various works in the literature [25,26]. Specifically, to obtain richer short sequence features, different scales of inputs were used for the CNN and attention branches. Furthermore, when performing the convolution operation, we set the filter size of the attention branch to be larger than that of the CNN branch. This approach was motivated by previous research [20,27]. Additionally, since the frequency range of sleep stages differs for different time steps [28], we employed different filter kernel sizes to capture the sleep-related frequency band features for different time steps.
When given N 30-s EEG epoch { X 1 ,   X 2 ,   ,   X N } input to our model, each epoch was input to the CNN and attention module with lengths of 2 s and 4 s. The input sequence of the attention module was centered on the input sequence of the CNN module. In order to prevent the loss of information, two modules slide forward in steps of 1 s. The process is presented in Figure 2. In other words, the 30-s epoch X i can be divided into 29 subsequences of 2 s or 29 subsequences of 4 s, which can be expressed as X i = { x 1 c ,   x 2 c , ,   x n c } = { x 1 a ,   x 2 a , ,   x n a } . x i c represents the i-th short sequence input to the CNN module, and x i a represents the i-th short sequence input to the attention module. This division approach is inspired by the study [29]. We used two modules to extract features from the i-th EEG epoch X i , as shown below.
h i c = CNN ( x i c )
h i a = Attention ( x i a )
a i   = h i c     h i a
where CNN ( x i c ) denotes the operation of the standard CNN module, which can convert the input subsequence x i c into the feature vector h i c ; Attention ( x i a ) denotes the operation of the attention module, which can convert the input subsequence x i a into the feature vector h i a , and denotes element-wise multiplication. After the above operations, epoch X i can be represented as a feature vector A I   = { a 1 ,   a 2 , ,   a n } . Thereby, the feature vector { A 1 ,   A 2 , …, A N } for N epochs { X 1 ,   X 2 ,   ,   X N } can be obtained.

3.2. The Two-Layer Bi-LSTM for Global Sequence Modeling

Due to the individual variability of recorded sleep signals, we needed to focus not only on intra-epoch variations, but also on inter-epoch variations to minimize the impact of this variability on the sleep staging task. The simple multiclassification task ignores temporal context information obtained between consecutive sleep epochs. Therefore, we considered the sleep staging problem as a sequential multiclassification task. Specifically, for N EEG epochs { X 1 ,   X 2 , ,   X N } , CAttSleepNet calculated the output sequence { Y 1 ,   Y 2 , ,   Y N } to maximize the conditional probability p ( X 1 ,   X 2 , ,   X N | Y 1 ,   Y 2 , ,   Y N ) .
In this work, we used a two-layer Bi-LSTM to capture global context information between consecutive sleep epochs. Bi-LSTM was composed of a forward LSTM and a backward LSTM. Therefore, compared to LSTM, Bi-LSTM could utilize the information in both the forward and backward directions. In addition, Bi-LSTM was used to extract coarse-grained features from the significant fine-grained features extracted by the attention-based CNN network. Meanwhile, it prevented the memory loss and gradient dispersion problems caused by the use of excessively long steps. In conclusion, a two-layer Bi-LSTM could capture the temporal dependence of consecutive epochs to achieve coarse and fine-grained feature fusion and fully characterize time-series data.
For the feature representation { A 1 , A 2 , , A N } of N 30-s EEG epoch { X 1 ,   X 2 , ,   X N } , where A i = { a 1 , a 2 , , a n } , we modeled the global sequence between 30-s epochs, as follows.
H t   = LSTM ( A t ,   H t 1 ,   C t 1 )
H t   = LSTM ( A t ,   H t + 1 ,   C t + 1 )
O t = H t     H t
where LSTM ( · ) denotes the operation of a two-layer LSTM, which can model the feature vector A t from the front and back directions; C and H are the vectors of cells and hidden states; and ‖ is a concatenation operation. Finally, sleep staging results can be obtained by adding a softmax layer after a two-layer Bi-LSTM.

3.3. Model Training and Parameter Optimization

Figure 3 shows the specific structure of CAttSleepNet. The top branch is the CNN module, and the middle is the attention module. Table 1 displays the specific parameters of CAttSleepNet. The CNN branch consisted of seven one-dimensional convolution layers, one max-pooling layer, and two dropout layers. The attention branch contained nine one-dimensional convolutional layers, one max-pooling layers, and three dropout layers. To solve the problem of overfitting during training, we used a dropout layer with the parameter set to 0.5. In particular, the batch normalization and application of corrected linear unit (ReLU) activation were performed for each convolutional layer in two branches. The attention branch restricted feature values extracted through a series of convolution and pooling operations between 0 and 1, through a sigmoid function. The output features of the CNN branch were multiplied element-wise with the output weights of the corresponding attention branch. Specifically, the more important the feature of the CNN branch was, the closer the output weight of the corresponding attention branch was to 1. On the contrary, the less important the feature of the CNN branch was, the closer the output weight of the corresponding attention branch was to 0.
We input five minutes EEG segments—i.e., ten 30-s EEG epochs—into the model without preprocessing for end-to-end automatic sleep staging. The CNN and the corresponding attention branches were slid forward simultaneously in one-second steps. The sliding window size of the CNN block was 2 s, and the sliding window size of the attention block was 4 s. Since the sampling rate of EEG was 100 Hz, per second EEG signal contained 100 data points. After the attention-based CNN module, the vector of shape (1, 256) was obtained. Fused results were flattened and fed to a two-layer Bi-LSTM. The number of hidden units in the Bi-LSTM layer was 64. Eventually, the softmax classifier outputs the most likely sleep stage.
We utilized the Adam optimizer with a learning rate of 0.001 to optimize the model parameters. CAttSleepNet could be trained end-to-end using the back-propagation algorithm. The cross-entropy loss function was employed. The loss function was defined as follows.
loss = 1 S k = 0 K 1 i = 0 S 1 y i , k log ( p i , k )
where K denotes the number of classes; S denotes the total number of samples; and y i , k and p i , k denote the actual label and predicted probability of the i-th sample for class k, respectively. The maximum training epoch was set to 300. Moreover, we performed our experiments on a device with two GPUs (NVIDIA GeForce GTX 1080 Ti) using Python 3.6 and Tensorflow 1.10. For the sleep-edfx-2013 dataset, each training epoch took about 1 min, while the sleep-edfx-2018 dataset needed roughly 4 min.

4. Experimental Evaluation

4.1. Experiment Datasets and Evaluation Metrics

4.1.1. Experiment Datasets

In this experiment, we used the public Physionet sleep-edf expanded (sleep-edfx) dataset [30,31], including version 1 from 2013 and version 2 from 2018, to evaluate the performance of CAttSleepNet. The sleep-edf dataset was first released in 2002 (version 0), and only contained a small amount of data; thus, it was not used in our study. It was expanded in 2013 and 2018, with sleep-edfx-2013 containing 61 whole-night PSGs, while sleep-edfx-2018 reached 197 whole-night PSGs. Additionally, these sleep signals were derived from two different studies—namely, studies on the impact of age on healthy adults (SC) and the effect of temazepam medication on sleep (ST). We only utilized data from * SC files for this task. Subjects’ EEG (Fpz-Cz and Pz-Oz channels), EOG, EMG, and event markers were recorded in each PSG. The sampling rates of EEG, EOG, and EMG were 100 Hz, 100 Hz, and 1 Hz, respectively. Some PSGs also recorded respiration and body temperature. These PSGs were manually labeled by professionals following the R&K manual [3]. Specifically, each 30-s epoch was labeled with one of the following classes: wakefulness (W), rapid eye movement (REM), MOVEMENT, UNKNOWN, and non-rapid eye movement (NREM, which was further divided into N1, N2, N3, and N4 stages). We combined N3 and N4 into N3 according to the AASM manual [4] and removed the MOVEMENT and UNKNOWN stages. Therefore, sleep stages were classified into five categories: W, REM, N1, N2, and N3. Furthermore, we eliminated some W stages and retained only the 30-min waking periods before and after sleep. This approach was consistent with the study [20]. As shown in Table 2, we performed sleep staging using EEG signals from Fpz-Cz and Pz-Oz channels.

4.1.2. Evaluation Metrics

K-fold cross-validation can improve the learning ability of deep learning models and make them more robust. In this experiment, we applied the k-fold cross-validation method to evaluate the performance of the CAttSleepNet model. To ensure a fairer comparison of the experimental results, the k values in this paper were set to be the same as those studies [20,27,29,32,33,34,35,36,37]—i.e., the k values were taken as 20 and 10 on the sleep-edfx-2013 and sleep-edfx-2018 datasets, respectively. The detailed process was as follows:
  • The sleep-edfx-2013 and sleep-edfx-2018 datasets are shuffled into k equal parts. K was set to 20 and 10, correspondingly.
  • One of the k equal parts was taken as a test set and the rest as a training set.
  • We trained the model and calculated the accuracy on the test set.
Steps 2 and 3 were repeated k times, and then the average of k test results was calculated as the final result.
To more comprehensively assess the behavior of our model in the sleep staging task, we considered the model evaluation both overall and per-category separately. On the one hand, due to the different number of samples for each sleep stage, we used the overall accuracy (ACC), Macro-F1 score (MF1), and Cohen’s Kappa coefficient (K) [38,39] to obtain a more intuitive and realistic portrayal of the overall classification. On the other hand, we calculated precision, recall, and F1-score for each class separately. The calculation formula used for the above indicators was as follows.
precision i = TP i TP i + FP i
recall i = TP i TP i + FN i
F 1 i = 2   ×   precision i   ×   recall i precision i + recall i
ACC = i = 1 C TP i S
MF 1 = 1 C i = 1 C F 1 i
K = p o   p e 1     p e
where FP i , FN i , and TP i are false positive, false negative, and true positive for the i-th class, respectively; S is the total amount of samples; and C is the number of categories. In this experiment, C is set to 5. p o is the actual agreement rate, while p e is the theoretical agreement rate.

4.2. Experimental Results of CAttSleepNet

On the sleep-edfx-2013 and sleep-edfx-2018 datasets, we obtained four k-fold cross-validation confusion matrices and corresponding receiver operating characteristic (ROC) curves, as shown in Figure 4, Figure 5, Figure 6, and Figure 7. In the confusion matrices, diagonal positions represent the proportions of correct classification, and other positions indicate proportions misclassified as other classes. The darker the color is, the higher the percentage is. For the sleep-edfx-2013 dataset, except for the N1 stage, the other four classes achieved a high accuracy. The highest classification accuracy was obtained for the W and N2 stages. The classification accuracy of the N1 stage on two channels was 41% and 42%, respectively. N1 stages were mainly misclassified as N2 and REM stages. Due to the small number of samples and few features learned during training, the N1 stage was the most indistinguishable among the five classes. The confusion matrix distribution of the sleep-edfx-2018 dataset was similar to that of the sleep-edfx-2013 dataset. On the other hand, the same ROC curve was obtained, since our model obtained similar classification results for both channels on the sleep-edfx-2013 dataset. The ROC curves on the two channels of sleep-edfx-2018 were vastly different. The CAttSleepNet model had the highest area under curve (AUC) on the sleep-edfx-2013 dataset and the lowest AUC on the Pz-Oz channel of sleep-edfx-2018. This may be due to the poor performance of the CAttSleepNet model on the Pz-Oz channel of sleep-edfx-2018 classification for each sleep stage, especially for the N1 stage.
Moreover, Table 3 also displays the ACC, MF1, and K of the overall classification, precision, recall, and F1-score for each category. It is obvious that our model outperformed the Pz-Oz channel on the Fpz-Cz channel, both overall and per-class. On the sleep-edfx-2013 dataset, the ACC, MF1, and K of the Fpz-Cz channel were 1.56%, 2.12%, and 1.51% higher than those of the Pz-Oz channel, respectively. On the sleep-edfx-2018 dataset, they were 2.8%, 4.06%, and 4.48%, correspondingly. On the other hand, the above indicator values for the sleep-edfx-2013 dataset were significantly higher than those for the sleep-edfx-2018 dataset.

4.3. Comparison with State-of-the-Art Methods

In this section, we compare the performance of our model with that of existing models [20,27,29,32,33,34,35,36,37]. Reference [32] used convolutional neural networks to automatically score sleep stages on the Fpz-Cz channel of EEG without using prior knowledge. Reference [33] designed a sleep staging model with stacked sparse autoencoders. Supratak et al. [20] proposed the DeepsleepNet model for sleep staging. The literature [29] used an approach combining attention mechanisms and bidirectional recurrent neural networks. Meanwhile, the literature [34] utilized a 1-max pooling CNN and time-frequency image features for automatic sleep staging. Additionally, a CNN framework for joint classification and prediction was proposed in the study [35]. Zhu et al. [36] proposed an automatic sleep staging method based on the attention mechanism and convolutional neural networks. Yang et al. [27] designed the 1D-CNN-HMM model, which combines the hidden Markov model (HMM) and one-dimensional convolutional neural network (1D-CNN). A new model for automatic sleep staging called SleepEEGNet, was proposed in the work of [37].
The results of our method compared with those of other methods are shown in Table 4. The validation of our experiments was consistent with these methods—i.e., using 20-fold cross-validation and 10-fold cross-validation on the sleep-edfx-2013 and sleep-edfx-2018 datasets, separately. On the sleep-edfx-2013 dataset, our model achieved the best results in terms of the overall metrics (ACC, MF1, and K) for both channels. That is, the ACC, MF1, and K on the Fpz-Cz channel were 84.1%, 78.2%, and 78%, respectively, while the ACC, MF1, and K on the Pz-Oz channel were 82.58%, 76.69%, and 76%, respectively. Second, the CAttSleepNet model achieved the highest F1-score for each class on the Pz-Oz channel, and the highest F1-score for the N1 and N2 stages on the Fpz-Cz channel. It is more difficult to achieve satisfactory classification results in N1 stages because of the smaller sample and more similar modality to the N2 stages, as is verified by experimental results recorded in the literature [20,27,29,32,33,34,35,36,37]. Without dealing with class imbalance, our model still outperformed these methods in the N1 stage. There have been few studies on the sleep-edfx-2018 dataset so far, and we only compared our results with those in the literature [37]. Our model outperformed that in the study of [37] in terms of overall classification metrics, performing satisfactorily in each class. In addition, we can see that the models of [36,37] achieved better or similar classification results compared to our model in the W, N1, and REM stages; this may be due to the different contributions of different classifiers to the different categories. Moreover, in the literature [36,37], there was a greater number of samples in the W stages, which may also have affected the actual classification results. In summary, CAttSleepNet achieved a highly competitive performance compared to that of other well-established sleep staging models.

4.4. Ablation Experiment

To verify the validity of the attention approach proposed in this paper, we conducted ablation experiments. Without changing the model parameters, we compare the performance of CAttSleepNet with and without attention in the Fpz-Cz channel of the sleep-edfx-2013 dataset. Additionally, both models used the same optimizer, loss function, and experimental environment. We calculated the ACC, MF1, and K, and per-class F1-score for the two models separately. Table 5 shows the experimental results obtained. The model with the attention branch outperformed the model without the attention branch. Specifically, ACC improved by 2.19%, K increased by 3.59%, and MF1 was enhanced by 2.94%. The visualization comparison results are shown in Figure 8. It can be seen from the figure that with the attention branch, CattSleepNet achieved better or similar classification results for each sleep stage. With the attention branch, the performance of CAttSleepNet was more consistent with the actual classification results. Therefore, the values of K and MF1 also increased significantly.

5. Conclusions

In this paper, we proposed a deep learning model named CAttSleepNet for automatic end-to-end sleep staging based on raw single-channel EEG. First, the attention-based CNN architecture could calculate the weights of local features by mining contextual associations; thus, it could differentiate the importance of key and non-key local features. Second, a two-layer Bi-LSTM was applied to globally model consecutive epochs, enabling end-to-end automatic sleep staging by exploiting the transition rules between sleep stages. The experimental results obtained on the sleep-edfx-2013 and sleep-edfx-2018 datasets demonstrate that our model achieved a better performance than the existing models. Furthermore, ablation experiments proved the effectiveness of our proposed attention module for use in sleep staging. Although our model achieved a promising performance, it still had some shortcomings. In the future, we hope to fuse information from multiple modalities, such as EOG and EMG, to enhance the performance of CAttSleepNet.

Author Contributions

Formal analysis, T.L., S.H. and Z.X.; Methodology, T.L., B.Z. and Z.X.; Software, T.L. and S.H.; Writing—original draft, T.L., B.Z., H.L., S.H. and Z.X.; Writing—review and editing, T.L., B.Z., H.L., S.H., Z.X. and Y.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key R & D Program of China grant number 3532017YFC0907505.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

EEG signals used in our experiments are derived from the public Physionet sleep-edf expanded dataset.

Conflicts of Interest

All authors declare that they have no conflict of interest with this work.

References

  1. Bjornara, K.A.; Dietrichs, E.; Toft, M. Longitudinal assessment of probable rapid eye movement sleep behaviour disorder in Parkinson’s disease. Eur. J. Neurol. 2015, 22, 1242–1244. [Google Scholar] [CrossRef] [PubMed]
  2. Zhong, G.; Naismith, S.; Rogers, N.; Lewis, S. Sleep–wake disturbances in common neurodegenerative diseases: A closer look at selected aspects of the neural circuitry. J. Neurol. Sci. 2011, 307, 9–14. [Google Scholar] [CrossRef] [PubMed]
  3. Wolpert, E.A. A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects. Arch. Gen. Psychiatry 1969, 20, 246. [Google Scholar] [CrossRef]
  4. Iber, C.; Ancoli-Israel, S.; Chesson, A.; Quan, S.F. The AASM Manual for the Scoring of Sleep and Associated Events: Rules, Terminology and Technical Specifications. 2007. Available online: https://www.sleep.pitt.edu/wp-content/uploads/2020/03/The-AASM-Manual-for-Scoring-of-Sleep-and-Associated-Events-2007-.pdf (accessed on 11 March 2022).
  5. Li, X.; Cui, L.; Tao, S.; Chen, J.; Zhang, X.; Zhang, G.-Q. HyCLASSS: A Hybrid Classifier for Automatic Sleep Stage Scoring. IEEE J. Biomed. Health Inform. 2017, 22, 375–385. [Google Scholar] [CrossRef]
  6. Dimitriadis, S.I.; Salis, C.; Linden, D. A novel, fast and efficient single-sensor automatic sleep-stage classification based on complementary cross-frequency coupling estimates. Clin. Neurophysiol. 2018, 129, 815–828. [Google Scholar] [CrossRef] [PubMed]
  7. Zhu, G.; Li, Y.; Wen, P.P. Analysis and Classification of Sleep Stages Based on Difference Visibility Graphs from a Single-Channel EEG Signal. IEEE J. Biomed. Health Inform. 2014, 18, 1813–1821. [Google Scholar] [CrossRef] [PubMed]
  8. Seifpour, S.; Niknazar, H.; Mikaeili, M.; Nasrabadi, A.M. A new automatic sleep staging system based on statistical behavior of local extrema using single channel EEG signal. Expert Syst. Appl. 2018, 104, 277–293. [Google Scholar] [CrossRef]
  9. Adegun, A.A.; Viriri, S.; Ogundokun, R.O. Deep Learning Approach for Medical Image Analysis. Comput. Intell. Neurosci. 2021, 2021, 6215281. [Google Scholar] [CrossRef]
  10. Astley, J.R.; Wild, J.M.; Tahir, B.A. Deep learning in structural and functional lung image analysis. Br. J. Radiol. 2022, 95, 20201107. [Google Scholar] [CrossRef]
  11. Said, Y.; Barr, M. Human emotion recognition based on facial expressions via deep learning on high-resolution images. Multimedia Tools Appl. 2021, 80, 25241–25253. [Google Scholar] [CrossRef]
  12. Aspandi, D.; Sukno, F.; Schuller, B.W.; Binefa, X. Audio-Visual Gated-Sequenced Neural Networks for Affect Recognition. IEEE Trans. Affect. Comput. 2022. [Google Scholar] [CrossRef]
  13. Comas, J.; Aspandi, D.; Binefa, X. End-to-end Facial and Physiological Model for Affective Computing and Applications. In Proceedings of the 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 16–20 November 2020; pp. 93–100. [Google Scholar]
  14. Yildirim, O.; Baloglu, U.B.; Acharya, U.R. A Deep Learning Model for Automated Sleep Stages Classification Using PSG Signals. Int. J. Environ. Res. Public Health 2019, 16, 599. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Fernandez-Blanco, E.; Rivero, D.; Pazos, A. Convolutional neural networks for sleep stage scoring on a two-channel EEG signal. Soft Comput. 2020, 24, 4067–4079. [Google Scholar] [CrossRef]
  16. Sors, A.; Bonnet, S.; Mirek, S.; Vercueil, L.; Payen, J.-F. A convolutional neural network for sleep stage scoring from raw single-channel EEG. Biomed. Signal Process. Control 2018, 42, 107–114. [Google Scholar] [CrossRef]
  17. Michielli, N.; Acharya, U.R.; Molinari, F. Cascaded LSTM recurrent neural network for automated sleep stage classification using single-channel EEG signals. Comput. Biol. Med. 2019, 106, 71–81. [Google Scholar] [CrossRef]
  18. Sun, C.; Fan, J.; Chen, C.; Li, W.; Chen, W. A Two-Stage Neural Network for Sleep Stage Classification Based on Feature Learning, Sequence Learning, and Data Augmentation. IEEE Access 2019, 7, 109386–109397. [Google Scholar] [CrossRef]
  19. Seo, H.; Back, S.; Lee, S.; Park, D.; Kim, T.; Lee, K. Intra- and inter-epoch temporal context network (IITNet) using sub-epoch features for automatic sleep scoring on raw single-channel EEG. Biomed. Signal Process. Control 2020, 61, 102037. [Google Scholar] [CrossRef]
  20. Supratak, A.; Dong, H.; Wu, C.; Guo, Y. DeepSleepNet: A Model for Automatic Sleep Stage Scoring Based on Raw Single-Channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1998–2008. [Google Scholar] [CrossRef] [Green Version]
  21. Sun, C.; Chen, C.; Li, W.; Fan, J.; Chen, W. A Hierarchical Neural Network for Sleep Stage Classification Based on Comprehensive Feature Learning and Multi-Flow Sequence Learning. IEEE J. Biomed. Health Inform. 2019, 24, 1351–1366. [Google Scholar] [CrossRef]
  22. Lajnef, T.; Chaibi, S.; Ruby, P.; Aguera, P.E.; Eichenlaub, J.-B.; Samet, M.; Kachouri, A.; Jerbi, K. Learning machines and sleeping brains: Automatic sleep stage classification using decision-tree multi-class support vector machines. J. Neurosci. Methods 2015, 250, 94–105. [Google Scholar] [CrossRef]
  23. Hassan, A.R.; Bhuiyan, M.I.H. Computer-aided sleep staging using Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and bootstrap aggregating. Biomed. Signal Process. Control 2016, 24, 1–10. [Google Scholar] [CrossRef]
  24. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; Bengio, Y. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 2048–2057. [Google Scholar]
  25. Li, C.; Hou, Y.; Wang, P.; Li, W. Joint Distance Maps Based Action Recognition with Convolutional Neural Networks. IEEE Signal Process. Lett. 2017, 24, 624–628. [Google Scholar] [CrossRef] [Green Version]
  26. Wang, P.; Li, W.; Gao, Z.; Zhang, Y.; Tang, C.; Ogunbona, P. Scene Flow to Action Map: A New Representation for RGB-D Based Action Recognition with Convolutional Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 595–604. [Google Scholar]
  27. Yang, B.; Zhu, X.; Liu, Y.; Liu, H. A single-channel EEG based automatic sleep stage classification method leveraging deep one-dimensional convolutional neural network and hidden Markov model. Biomed. Signal Process. Control 2021, 68, 102581. [Google Scholar] [CrossRef]
  28. Memar, P.; Faradji, F. A Novel Multi-Class EEG-Based Sleep Stage Classification System. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 26, 84–95. [Google Scholar] [CrossRef] [PubMed]
  29. Phan, H.; Andreotti, F.; Cooray, N.; Chen, O.Y.; De Vos, M. Automatic Sleep Stage Classification Using Single-Channel EEG: Learning Sequential Features with Attention-Based Recurrent Neural Networks. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 1452–1455. [Google Scholar]
  30. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [Green Version]
  31. Kemp, B.; Zwinderman, A.; Tuk, B.; Kamphuisen, H.; Oberye, J. Analysis of a sleep-dependent neuronal feedback loop: The slow-wave microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 2000, 47, 1185–1194. [Google Scholar] [CrossRef]
  32. Tsinalis, O.; Matthews, P.M.; Guo, Y.; Zafeiriou, S. Automatic Sleep Stage Scoring with Single-Channel EEG Using Convolutional Neural Networks. arXiv 2016, arXiv:1610.01683. [Google Scholar]
  33. Tsinalis, O.; Matthews, P.M.; Guo, Y. Automatic Sleep Stage Scoring Using Time-Frequency Analysis and Stacked Sparse Autoencoders. Ann. Biomed. Eng. 2016, 44, 1587–1597. [Google Scholar] [CrossRef] [Green Version]
  34. Phan, H.; Andreotti, F.; Cooray, N.; Chen, O.Y.; De Vos, M. DNN Filter Bank Improves 1-Max Pooling CNN for Single-Channel EEG Automatic Sleep Stage Classification. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 453–456. [Google Scholar]
  35. Phan, H.; Andreotti, F.; Cooray, N.; Chen, O.Y.; De Vos, M. Joint Classification and Prediction CNN Framework for Automatic Sleep Stage Classification. IEEE Trans. Biomed. Eng. 2018, 66, 1285–1296. [Google Scholar] [CrossRef]
  36. Zhu, T.; Luo, W.; Yu, F. Convolution- and Attention-Based Neural Network for Automated Sleep Stage Classification. Int. J. Environ. Res. Public Health 2020, 17, 4152. [Google Scholar] [CrossRef]
  37. Mousavi, S.; Afghah, F.; Acharya, U.R. SleepEEGNet: Automated sleep stage scoring with sequence to sequence deep learning approach. PLoS ONE 2019, 14, e0216456. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Cohen, J. A Coefficient of Agreement for Nominal Scales. Educ. Psychol. Meas. 1960, 20, 37–46. [Google Scholar] [CrossRef]
  39. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
Figure 1. General architecture of CattSleepNet.
Figure 1. General architecture of CattSleepNet.
Ijerph 19 05199 g001
Figure 2. The CNN and attention branches perform feature extraction on the 30-s EEG epoch. The red solid line indicates the input sequence length of attention branch. The black solid line indicates the input sequence length of CNN branch. The dotted line indicates where the two branches will slide forward next.
Figure 2. The CNN and attention branches perform feature extraction on the 30-s EEG epoch. The red solid line indicates the input sequence length of attention branch. The black solid line indicates the input sequence length of CNN branch. The dotted line indicates where the two branches will slide forward next.
Ijerph 19 05199 g002
Figure 3. Detailed structure of CAttSleepNet model.
Figure 3. Detailed structure of CAttSleepNet model.
Ijerph 19 05199 g003
Figure 4. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Fpz-Cz channel of the sleep-edfx-2013 dataset. (a) Confusion matrix; (b) ROC curve.
Figure 4. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Fpz-Cz channel of the sleep-edfx-2013 dataset. (a) Confusion matrix; (b) ROC curve.
Ijerph 19 05199 g004
Figure 5. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Pz-Oz channel of the sleep-edfx-2013 dataset. (a) Confusion matrix; (b) ROC curve.
Figure 5. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Pz-Oz channel of the sleep-edfx-2013 dataset. (a) Confusion matrix; (b) ROC curve.
Ijerph 19 05199 g005
Figure 6. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Fpz-Cz channel of the sleep-edfx-2018 dataset. (a) Confusion matrix; (b) ROC curve.
Figure 6. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Fpz-Cz channel of the sleep-edfx-2018 dataset. (a) Confusion matrix; (b) ROC curve.
Ijerph 19 05199 g006
Figure 7. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Pz-Oz channel of the sleep-edfx-2018 dataset. (a) Confusion matrix; (b) ROC curve.
Figure 7. The CAttSleepNet model’s confusion matrix and ROC curve are obtained from the Pz-Oz channel of the sleep-edfx-2018 dataset. (a) Confusion matrix; (b) ROC curve.
Ijerph 19 05199 g007
Figure 8. Visual comparison of the two models. (a) Comparison of overall indicators; (b) comparison of per-class F1-score.
Figure 8. Visual comparison of the two models. (a) Comparison of overall indicators; (b) comparison of per-class F1-score.
Ijerph 19 05199 g008
Table 1. Detailed parameters of CAttSleepNet model.
Table 1. Detailed parameters of CAttSleepNet model.
BranchLayer TypeNumber of FiltersKernel SizeRegion SizeStrideOutput Shape
Input-----(200, 1)
The CNNConv1641 × 5-3(67, 64)
Conv2641 × 5-3(23, 64)
Conv3641 × 5-3(8, 64)
Conv41281 × 3-2(4, 128)
Conv51281 × 3-2(2, 128)
Conv61281 × 3-1(2, 128)
Conv72561 × 3-1(2, 256)
Max-pooling--1 × 21(1, 256)
Input-----(400, 1)
The AttentionConv1641 × 7-3(134, 64)
Conv2641 × 7-3(45, 64)
Conv3641 × 7-3(15, 64)
Max-pooling--1 × 22(8, 64)
Conv41281 × 5-2(4, 128)
Conv51281 × 5-2(2, 128)
Conv61281 × 5-2(1, 128)
Conv72561 × 3-1(1, 256)
Conv82561 × 3-1(1, 256)
Conv92561 × 3-1(1, 256)
Table 2. Detailed distribution of sleep stages in sleep-edfx-2013 and sleep-edfx-2018 datasets.
Table 2. Detailed distribution of sleep stages in sleep-edfx-2013 and sleep-edfx-2018 datasets.
StageSleep-Edfx-2013Sleep-Edfx-2018
Training SetTest SetTotalTraining SetTest SetTotal
W7734292802655,69710,02365,720
N12666138280419,207231521,522
N216,80599417,79989,789634396,132
N35449254570311,879116013,039
REM7295422771723,452238325,835
Total39,949210042,049200,02422,224222,248
Table 3. Evaluation indicators for the overall and each category are obtained from two datasets.
Table 3. Evaluation indicators for the overall and each category are obtained from two datasets.
Sleep-Edfx-2013Sleep-Edfx-2018
EEG Fpz-Cz (%)EEG Pz-Oz (%)EEG Fpz-Cz (%)EEG Pz-Oz (%)
PreRecF1PreRecF1PreRecF1PreRecF1
W88.8690.2889.5689.5689.4489.5092.5691.6792.1289.2490.6789.95
N155.5940.8747.1151.9841.9346.4246.1041.6143.7445.7934.3339.24
N286.2488.2187.2284.6387.1085.8481.7184.9083.2878.5283.7981.07
N386.5783.4784.5082.5579.3880.9477.2576.6676.9671.1865.2068.06
REM80.0584.3082.1279.4582.0980.7576.5176.8976.7071.0173.1672.07
Overall IndicatorsACCKMF1ACCKMF1ACCKMF1ACCKMF1
84.1478.0978.2082.5875.9776.6980.8173.5174.5678.0169.4570.08
Note: Pre = precision, Rec = recall, F1 = F1-score.
Table 4. Comparison among CAttSleepNet and other models.
Table 4. Comparison among CAttSleepNet and other models.
ApproachOverall Performance (%)Per-Class F1-Score (%)
ACCMF1KWN1N2N3REM
Dataset: Sleep-Edfx-2013 EEG Channel: Fpz-Cz
Tsinalis et al. [32]74.869.8-65.443.780.684.974.5
Tsinalis et al. [33]78.973.7-71.647.084.684.081.4
Supratak et al. [20]82.076.90.7684.746.685.984.882.4
Phan et al. [29]79.169.80.7075.527.386.085.674.8
Phan et al. [34]79.872.00.7277.033.386.886.376.4
Phan et al. [35]81.973.80.74-----
Zhu et al. [36]82.877.8-90.347.186.082.183.2
Yang et al. [27]82.1373.50.7587.823.086.290.981.8
CAttSleepNet84.178.20.7889.647.187.285.082.1
Dataset: Sleep-Edfx-2013 EEG Channel: Pz-Oz
Supratak et al. [20]79.873.10.7288.13782.777.380.3
Yang et al. [27]80.5468.70.7285.317.585.078.275.8
CAttSleepNet82.5876.690.7689.546.485.880.980.8
Dataset: Sleep-Edfx-2018 EEG Channel: Fpz-Cz
Mousavi et al. [37]80.0373.550.7391.7244.0582.4973.4576.06
CAttSleepNet80.8174.560.7492.1243.7483.2876.9676.70
Dataset: Sleep-Edfx-2018 EEG Channel: Pz-Oz
Mousavi et al. [37]77.5670.000.69-----
CAttSleepNet78.0170.080.6989.9539.2481.0768.0672.07
Note: The highest performance metrics are highlighted in bold. Except for the K indicator, values of other indicators are all percentiles.
Table 5. Ablation experiments on Fpz-CZ channel of the sleep-edfx-2013 dataset.
Table 5. Ablation experiments on Fpz-CZ channel of the sleep-edfx-2013 dataset.
CAttSleepNet (%)CAttSleepNet without Attention (%)
PreRecF1PreRecF1
W88.8690.2889.5683.1383.1383.14
N155.5940.8747.1148.6340.0943.95
N286.2488.2187.2290.0588.2089.57
N386.5783.4784.5078.0778.9278.49
REM80.0584.3082.1275.5887.6481.87
Overall IndicatorsACCKMF1ACCKMF1
84.1478.0978.2081.9574.5075.26
Note: Pre = precision, Rec = recall, F1 = F1-score.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, T.; Zhang, B.; Lv, H.; Hu, S.; Xu, Z.; Tuergong, Y. CAttSleepNet: Automatic End-to-End Sleep Staging Using Attention-Based Deep Neural Networks on Single-Channel EEG. Int. J. Environ. Res. Public Health 2022, 19, 5199. https://doi.org/10.3390/ijerph19095199

AMA Style

Li T, Zhang B, Lv H, Hu S, Xu Z, Tuergong Y. CAttSleepNet: Automatic End-to-End Sleep Staging Using Attention-Based Deep Neural Networks on Single-Channel EEG. International Journal of Environmental Research and Public Health. 2022; 19(9):5199. https://doi.org/10.3390/ijerph19095199

Chicago/Turabian Style

Li, Tingting, Bofeng Zhang, Hehe Lv, Shengxiang Hu, Zhikang Xu, and Yierxiati Tuergong. 2022. "CAttSleepNet: Automatic End-to-End Sleep Staging Using Attention-Based Deep Neural Networks on Single-Channel EEG" International Journal of Environmental Research and Public Health 19, no. 9: 5199. https://doi.org/10.3390/ijerph19095199

APA Style

Li, T., Zhang, B., Lv, H., Hu, S., Xu, Z., & Tuergong, Y. (2022). CAttSleepNet: Automatic End-to-End Sleep Staging Using Attention-Based Deep Neural Networks on Single-Channel EEG. International Journal of Environmental Research and Public Health, 19(9), 5199. https://doi.org/10.3390/ijerph19095199

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop