Next Article in Journal
Thermal Stability of Oxidation-Resistant Ta-Zr-Si-B-N and Ta-Zr-Si-B-C Coatings under In Situ TEM Heating and Vacuum Annealing
Previous Article in Journal
A Study on Maize Leaf Pest and Disease Detection Model Based on Attention and Multi-Scale Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Sleep Stage Classification Using a Taguchi-Based Multiscale Convolutional Compensatory Fuzzy Neural Network

Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung 411, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(18), 10442; https://doi.org/10.3390/app131810442
Submission received: 24 August 2023 / Revised: 14 September 2023 / Accepted: 17 September 2023 / Published: 18 September 2023
(This article belongs to the Topic Electronic Communications, IOT and Big Data)

Abstract

:
Current methods for sleep stage detection rely on sensors to collect physiological data. These methods are inaccurate and take up considerable medical resources. Thus, in this study, we propose a Taguchi-based multiscale convolutional compensatory fuzzy neural network (T-MCCFNN) model to automatically detect and classify sleep stages. In the proposed T-MCCFNN model, multiscale convolution kernels extract features of the input electroencephalogram signal and a compensatory fuzzy neural network is used in place of a traditional fully connected network as a classifier to improve the convergence rate during learning and to reduce the number of model parameters required. Due to the complexity of general deep learning networks, trial and error methods are often used to determine their parameters. However, this method is very time-consuming. Therefore, this study uses the Taguchi method instead, where the optimal parameter combination is identified over a minimal number of experiments. We use the Sleep-EDF database to evaluate the proposed model. The results indicate that the proposed T-MCCFNN sleep stage classification accuracy is 85.3%, which is superior to methods proposed by other scholars.

1. Introduction

According to Maslow’s hierarchy of needs, human beings can only fulfill higher-order needs if basic ones are fulfilled [1], and sleep is a key physiological need. In general, good sleep contributes to physical and mental health and emotional stability [2]. Conversely, a lack of sleep can cause health problems, such as cardiovascular disease [3], hypertension [4], and stroke [5].
Researchers have used several methods to improve the automated detection of sleep quality and stage, including polysomnography (PSG), where the individual’s physiological data are recorded. PSG records biological signals through several mechanisms, including electrocardiography, electromyography (EMG), electrooculography (EOG), and electroencephalogram (EEG), with EEG being the most widely used method in the field. Experts require complete and well-defined criteria to conduct sleep stage research. Rechtschaffen and Kales (R&K) proposed criteria for reporting sleep data [6], and although these criteria have not been rigorously validated [7], the American Academy of Sleep Medicine (AASM) improved upon the foundational work of R&K [8], making R&K’s criteria a standard in the medical community. AASM classifies EEG sleep stages through 30-s interval frequency sample recordings. These sleep stages are divided into Wake (W), rapid eye movement (REM), and non-rapid eye movement (NREM). NREM is subdivided into NREM1, NREM2, NREM3, and NREM4 stages that correspond to the shallowest to deepest stages of sleep, respectively. NREM3 and NREM4 have no obvious differences, and they are thus often collectively referred to as slow-wave sleep or deep sleep.
Although scholars have developed a relatively complete sleep stage classification, PSG records are rich with data, and experts have different approaches to signal evaluation [9]. Manual sleep stage classification by experts from PSG records is time-consuming, labor-intensive, and inconsistent, and thus, not a feasible solution. Therefore, automatic sleep stage classification is required.
Many scholars have developed automatic sleep stage classification systems based on machine learning. Machine learning typically involves feature extraction and classification. Phan et al. [10] extracted features from single-channel EEG signals and used the k nearest neighbor algorithm with the Mahalanobis distance for classification. The method performs well in addressing four types of sleep stage problems. Liang et al. [11] extracted 13 features from EEG, EOG, and EMG readings and used fuzzy vector quantization and a genetic algorithm to improve the resolution of a hidden Markov model (HMM). Fraiwan et al. [12] used Hilbert–Huang transform to extract features from the input EEG signal, random forest for classification, Choi-Williams distribution, and continuous wavelet transform. Khalighi et al. [13] used six EEG channels, two EOG channels, and maximal overlap discrete wavelet transformation for feature extraction. They also used a support vector machine to classify the selected features.
Although machine learning methods can solve sleep stage classification problems, feature extraction is performed independently from classification, and thus, is reliant on experts. Deep learning technology, which has been widely used in the medical field for the diagnosis of health problems such as breast cancer [14], epilepsy [15], or arrhythmia [16], can solve several machine learning problems. Deep learning techniques eliminate the need for manual feature extraction. During model training, features are automatically generated according to signals. Khan et al. [17] employed a dilated convolutional neural network and multi-headed attention mechanism for feature extraction. Ishaq et al. [18] extracted features using time series-based Temporal Convolutional Neural (TCN). Other scholars have employed deep learning-based models for sleep stage detection as well. Kanwal et al. [19] used a Fourier transform to convert the original one-dimensional EEG signal into the input data of a convolutional neural network (CNN) in the form of a two-dimensional image. Huang et al. [20] developed a neural network architecture that combined a one-dimensional CNN and HMM. The neural network architecture performed well on Fpz-Cz and Pz-Oz EEG channels. Toma and Choi [21] used the Fpz-Cz channel’s EEG and EOG signals as input for the CNN and a recurrent neural network hybrid model. They evaluated the performance of the model in the Sleep-EDF dataset [22].
Although scholars have used deep learning to solve sleep stage classification problems, their experiments have typically involved the use of a fully connected network (FCN) as a model classifier. FCN, an artificial neural network, relies on non-physically interpretable weight updates through backpropagation. To address this problem, many scholars have introduced fuzzy theory into neural networks (i.e., ANFIS [23] and IT2FNN [24]). FNNs combine neural network architecture with fuzzy logic. They have the neural networks’ parameter learning ability and fuzzy theory’s capability for human-like thinking. FNN can be divided into Mamdani-type FNN [25] and Takagi-Sugeno-Kang (TSK)-type FNN [26]. TSK-type FNNs have better performance than the Mamdani-type FNN [26]. Zhang and Kandel [27] proposed compensatory fuzzy sets to improve the convergence speed and performance of FNNs.
Because deep learning model architectural parameters are obtained through training results (which may rely on academic experience or trial and error), obtaining them can be a time-consuming and costly process. In the field of engineering, Taguchi’s method [28] is often used to improve parameter optimization: it requires minimal experimentation to produce optimal parameters and improve system performance.
In this study, we proposed a Taguchi-based multiscale convolutional compensatory FNN (T-MCCFNN) for the automatic detection and classification of sleep stages, which assists doctors in obtaining patient’s sleep information and identifying sleep problems, such as insomnia. In the proposed T-MCCFNN model, multiscale convolution kernels were used to extract features of the input EEG signal and a compensatory FNN was used as a classifier. Since the parameters of the general deep learning network are not easy to determine, the Taguchi method is used to find the optimal parameter combination with a small number of experiments.
The rest of this paper is organized as follows: Section 2 introduces the proposed T-MCCFNN architecture and the Taguchi method’s experimental procedure. Section 3 describes the results of the T-MCCFNN architecture using the Sleep-EDF dataset, and Section 4 concludes the paper.

2. Materials and Methods

This section presents the proposed sleep stage detection method. Section 2.1 introduces the proposed T-MCCFNN model. The T-MCCFNN model was employed to classify the EEG signals recorded from the Fpz-Cz channel into five distinct sleep stages: W, REM, N1, N2, and N3. Section 2.2 introduces our use of the Taguchi method to optimize the T-MCCFNN parameters through a minimal number of experiments.

2.1. Proposed T-MCCFNN Model

The proposed T-MCCFNN model includes two parts: a convolutional neural network and a TSK-type compensatory fuzzy neural network. The convolutional neural network uses convolution kernels with different ratios to obtain feature signals of different scales from the input EEG signal. A TSK-type compensatory fuzzy neural network adopts TSK-type fuzzy rules, and increases the convergence speed of network training by compensatory fuzzy sets. The architecture of the T-MCCFNN model is shown in Figure 1, where C and T represent the compensatory operation and the TSK-type fuzzy rule, respectively. The T-MCCFNN model includes a convolutional layer, a max pool layer, a concatenate layer, a global maximum pooling layer, a fuzzified layer, a compensatory rule layer, and an output layer. In the T-MCCFNN model, the operations of each layer are as follows:

2.1.1. Convolutional Layer

The convolutional layer performs convolution operations. It extracts features from the EEG signal by sliding the convolution kernel. The operation of the convolutional layer is as follows:
y f , i c o n v = j = 0 k s 1 x i × s c o n v + j w f , j c o n v
where y c o n v is the output of the convolutional layer, x is the input signal, is the convolution operator, w c o n v is the weight of the convolution kernel, k s is the size of the convolution kernel, and s c o n v is the movement pace of the convolution kernel.

2.1.2. Max Pooling Layer

Max pooling layers reduce the output feature maps of convolutional layers, thereby helping the model learn abstract features to improve generalizability. The operation of the maximum pooling layer is as follows:
y f , i m p = m a x 0 j p s 1 y f , i × s m p + j c o n v
where f is the index value of the feature map, i is the index of the value in the feature map, mp is the max pooling, y m p is the output of the maximum pooling layer,   m a x ( · ) is the maximum value operator, and p s is the pooling kernel size.

2.1.3. Concatenate Layer

The concatenate layer concatenates output feature maps from differing sizes along the channel axis of the input features as follows:
y f , i D = { y 1,1 m p , , y 1 , N m p , , y f , 1 m p , , y f , N m p }
where y D is the output of the concatenate layer, { · } is the concatenation operator, and N is the size of the input feature map.

2.1.4. Global Max Pool Layer

To reduce feature maps from the concatenate layers, we used a global max pooling layer to perform the maximum operation on each feature map. The operation of the global max pooling layer is as follows:
y f G = max 0 i N y f , i D
where y G is the output of the local maximum pooling layer.

2.1.5. Fuzzified Layer

In the fuzzified layer, the input data is fuzzified. Fuzzy if–then rules with compensation are used and are expressed as follows:
R u l e : ( I F   y 0 G   i s   A 1 , j   a n d a n d   y N G   i s   A N , j ) 1 r j + r j R T H E N   y = w 0 , j + w 1 , j y 1 G + + w N , j y N G
where A is a fuzzy set, r is the degree of compensation r [ 0 ,   1 ] , R is the number of rules, and w is the TSK-type fuzzy rule weight. In this study’s method, the Gaussian membership function is used as a fuzzy set A , which is defined as follows:
y i , j F = e ( y i G m i , j σ i , j ) 2
where y F is the output of the fuzzy layer,  e ( · ) is the exponentiation operation, and m and σ represent the mean and standard deviation, respectively, of the fuzzy set A .

2.1.6. Compensatory Rule Layer

The compensatory rule layer performs a fuzzy AND operation on the degree of membership obtained from the previous layer operation. We used the product operation to obtain the firing strength fuzzy rule. Additionally, we incorporated a compensatory operation into the rule layer. The compensatory rule layer is governed by the following equation.
y j C = ( i = 0 N y i , j F ) 1 r j + r j R

2.1.7. Output Layer

The output layer uses the center of gravity method to obtain the T-MCCFNN model output. The output layer is operated according to the following equation:
y = j = 0 R y j C × ( w 0 , j + i = 1 N w i , j y i G ) j = 0 R y j C

2.2. Taguchi Method

The Taguchi method is a widely used method in engineering. This method generates an orthogonal array (OA) using control factor selection, level selection, and statistical principles. OA is a high-efficiency technology that improves the effectiveness of experiments and thus reduces the need for extensive experimentation. After the OA experiment was completed, the results were converted to a signal-to-noise (S/N) ratio and used to determine the optimal control factors for improving system stability. In this experiment, we used the Taguchi method to solve the problem of hyperparameters in the neural network.
We found, through previous research experience, that features extracted by the convolutional layer exert a considerable influence on classification results. Therefore, we chose Conv1_filters, Conv1_kernel_size, Conv1_strides, Conv2_filters, Conv2_kernel_size, and Conv2_strides as control factors in our experiments. To improve the diversity of convolutional layers in feature extraction, we used multiscale convolution branches for feature extraction. When more branches are used as control factors, the required number of experiments increases. Therefore, this study used a scale factor to adjust the convolution kernel size of the remaining two branches. Table 1 lists each control factor and its level.
This experiment used an OA of L27( 3 7 ) (as shown in Table 2) based on the seven control factors and three levels (shown in Table 1). The contribution of each control factor and level were calculated after the experiment was completed. An S/N ratio (where larger ratios are preferable) was adopted for evaluation. The S/N ratio is calculated as follows:
S N = 10 log 10 1 n ( i = 0 n 1 y i 2 )
where n is the number of experiments, y i represents the accuracy rate obtained in the i-th time, and S N represents the S/N ratio and its unit in db.

3. Experimental Results

We evaluated the proposed T-MCCFNN model on the Sleep-EDF database. We also discussed the Taguchi experiment’s proposed T-MCCFNN model results for comparison and analysis. Finally, we compared the results with other state-of-the-art methods.

3.1. Dataset

This study used the publicly available Sleep-EDF database [22], which was labeled and verified by experts. The database analyzed the sleep data of 20 individuals (10 men and 10 women) between the ages of 25 and 34 years old. The PSG signals in the data set include a horizontal EOG signal, a chin EMG signal, and two EEG signals: the Fpz-Cz channel and the Pz-Cz channel. This experiment adopted the Fpz-Cz location channel (Figure 2). These signals were sampled at a rate of 100 Hz and segmented into 30-s intervals for a total of 3000 sampling points. In addition, the signals in the data set are not scaled or preprocessed. They were then classified into six categories according to R&K’s classification standards: W, N1, N2, N3, N4, and REM. They represent different levels of the patient’s depth of sleep. The amount of data for W, N1, N2, N3, and REM are 8285, 2804, 17,799, 5703, and 7717, respectively. To meet the classification standards of AASM, this experiment combined the two sleep stages of N3 and N4. These recordings included a total of 42,308 signals divided into training data and testing data at a ratio of 7:3.

3.2. Taguchi Experimental Results

The proposed T-MCCFNN model included a CNN and TSK-type compensatory FNN. The CNN, which used the scale factor to adjust the convolution kernel size of the two remaining convolutional branches, comprised three convolutional layers and two maximum pooling layers. The TSK-type compensatory FNN comprised a fuzzy layer, a compensatory rule layer, and an output layer. The initial parameters of the proposed T-MCCFNN model are depicted in Table 3. The model architecture included many tunable hyperparameters. The Taguchi method was used to find the optimal parameter combination through the use of L27(37), as shown in Table 2. The subsequently generated OA reduced the number of experiments required. In our Taguchi experiment, we selected Adam as our optimizer, a learning rate of 0.001, a batch size of 256, and 50 training epochs. To improve the stability of the Taguchi experiment, we conducted ten experiments for each set of parameter combinations and calculated the average. The results of OA are shown in Table 4.
The S/N ratio of each control factor was calculated and used to analyze the contributions of each parameter to model performance. As shown in Table 5, the optimal parameter combinations of Conv1_filters (A), Conv1_kernel_size (B), Conv1_strides(C), Scale (D), Conv2_filters (E), Conv2_kernel_size (F), and Conv2_strides (G) values were 64, 50, 4, 4, 256, 9, and 1, respectively. The control factors with the highest contribution to the model performance were Conv2_strides (G), Conv2_filters (E), and Connv1_strides (C). Thus, changes in stride size tended to affect the performance of the T-MCCFNN model. Figure 3 depicts the main effects on the S/N ratio. In particular, a smaller stride size was associated with a higher S/N ratio for both Conv1_strides (C) and Conv2_strides (G).

3.3. Model Evaluation

This study used performance metrics of accuracy, precision, recall, and F1-score, which are defined as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F P
P r e c i s i o n = T P T P + F N
F 1 - s c o r e = 2 × R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n
where T P , T N , F P , and F N are the numbers of true positives, false positives, true negatives, and false negatives, respectively.
Figure 4a shows the MCCFNN confusion matrix. Although the model performed poorly in predicting N1, it performed well in predicting W, N2, N3, and REM in terms of accuracy. Figure 4b indicates that the use of T-MCCFNN parameter optimization in the Taguchi experiment yielded improvements in the model’s predictions of W, N3, and REM by 3.02%, 2.98%, and 8.54%, respectively, in terms of accuracy.
The comparison results of the proposed T-MCCFNN with other FNNs are shown in Table 6. In Table 6, MCCNN uses a multi-scale convolution kernel and a fully connected network. A Mamdani-type MCFNN and TSK-type MCFNN use a multi-scale convolution kernel and Mamdani-type and TSK-type FNNs. In Table 6, the accuracy of MCCFNN with TSK-type is 84.4%, which is better than 84.1% of Mamdani-type MCCFNN. After parameter optimization through the Taguchi experiment, T-MCCFNN can increase the accuracy by 0.9% compared to TSK-type MCCFNN, and it is better than other FNNs and fully connected network models. In Table 6, the standard deviation of the accuracy of the proposed T-MCCFNN is 0.0321, which means that the proposed T-MCCFNN model has good robustness and universality in terms of accuracy.
In order to verify that the compensatory FNN has a better convergence speed, this study uses MCCFNN to compare it with MCFNN without the compensatory fuzzy set. It can be observed from Figure 5 that MCCFNN tends to converge after nearly 15 epochs. However, MCFNN slowly converges at 25 epochs. The experimental results indicate that the MCCFNN with the compensatory fuzzy set has a better convergence speed than the MCFNN without the compensatory fuzzy set.
In this paper, we use cross-entropy as the loss function, which is defined as follows:
C E = i y i l o g   y ^ i
where C E represents cross-entropy; y and y ^ represent our model’s prediction and ground truth.
In addition, this study compares the performance of the multi-scale convolution kernel (T-MCCFNN) and the single-scale convolution kernel (T-SSCFNN), as shown in Table 7. In Table 7, the accuracy, recall, precision, and F1-score of T-MCCFNN with multi-scale convolution kernels have higher performance than T-SSCFNN with single-scale convolution kernels. In particular, there is a 1.1% increase in the accuracy of T-MCCFNN.
Table 8 illustrates the comparison of T-MCCFNN with other deep learning methods. Supratak et al. [29] proposed the DeepSleepNet network architecture. This architecture uses a two-branched convolutional layer to extract features from the raw EEG signal, and a two-layer bi-directional long short-term memory (Bi-LSTM) as a classifier. Fiorillo et al. [30] improved the architecture of DeepSleepNet and proposed a DeepSleepNet-lite network architecture. This architecture discards Bi-LSTM, but uses a general fully connected network as the output. Phan et al. [31] proposed that the XSleepNet network architecture is different from [29] and [30]. This architecture uses the original signal and the time-frequency signal as dual inputs to obtain the characteristics of the time-domain signal and the frequency-domain signal. Eldele et al. [32] proposed the multi-resolution convolutional neural network (MRCNN). The architecture of MRCNN introduces an attention mechanism, enabling it to adaptively focus on important information in the signal. Zhou et al. [33] proposed a lightweight LightSleepNet network architecture. This architecture uses a spectrogram as the signal input, and proposes two modules, the concatenation of max-pooling2D and average-pooling2D module and the multi-convolution block module, to extract various features from the spectrogram.
Compared with various deep learning methods, the accuracy and F1-score of the proposed T-MCCFNN are 85.3% and 85.3%, respectively, and both are superior to other deep learning network methods. In Table 8, although the N1 and REM of the per-class F1-score in Phan et al. [31] are 52.9% and 85.4%, respectively, which are better than the proposed method; the W, N2, and N3 of the per-class F1-score in the proposed T-MCCFNN are 92.7%, 88.5%, and 91.2%, respectively, which are better than the method of Phan et al. [31]. Overall, the accuracy and F1-score of the proposed T-MCCFNN are superior to other deep learning network methods.

4. Conclusions

In this study, a T-MCCFNN was proposed for the automatic classification of sleep stage detection. In this study, the Sleep-EDF database was used to verify the performance of the proposed model. The experimental results indicate that the accuracy and F1-score of the proposed T-MCCFNN in sleep stage classification are 85.3% and 85.3%, respectively, which is better than the methods proposed by other methods. In addition, we have also shown that the compensatory FNN has a better convergence speed than the general FNN, and the model with multi-scale convolution kernels has better accuracy, recall, precision, and F1-score than the model with single-scale convolution kernels. The major contributions of this study were as follows:
  • Multiscale convolution kernels in the T-MCCFNN model were used to extract input EEG signal features.
  • A compensatory FNN was used in place of the traditional FCN classifier to improve the learning convergence speed and reduce model parameters.
  • The Taguchi method was used to find the optimal parameter combination through minimal experimentation.
  • The experimental results indicate that the accuracy, recall, precision, and F1-score of the proposed T-MCCFNN in sleep stage classification were 85.3%, 83.3%, 87.3%, and 85.3%, respectively, which is superior to the scores of the methods proposed by other scholars.
To detect the patient’s sleep quality in real time, the proposed T-MCCFNN model will be deployed into the Field-Programmable Gate Array (FPGA) in future work. Because FPGA is known for its excellent performance and low power consumption, it is very suitable to be embedded in wearable devices to assist doctors in instantly assessing patients’ sleep quality.

Author Contributions

Conceptualization, C.-J.L. (Cheng-Jian Lin) and X.-Q.L.; Methodology, C.-J.L. (Chun-Jung Lin) and C.-J.L. (Cheng-Jian Lin); Software, C.-J.L. (Chun-Jung Lin) and X.-Q.L.; Validation, C.-J.L. (Chun-Jung Lin); Data curation, X.-Q.L.; Writing—original draft, C.-J.L. (Cheng-Jian Lin). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of the Republic of China, grant number NSTC 112-2221-E-167-026.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors confirm that the data supporting the findings of this study are available within the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maslow, A. A Theory of Human Motivation. Psychol. Rev. 1943, 50, 370–396. [Google Scholar] [CrossRef]
  2. Pilcher, J.; Ginter, D.; Sadowsky, B. Sleep Quality Versus Sleep Quantity: Relationships Between Sleep and Measures of Health Well-Being and Sleepiness in College Students. J. Psychosom. Res. 1997, 6, 583–596. [Google Scholar] [CrossRef] [PubMed]
  3. Schwartz, S.; Anderson, W.M.; Cole, S.R.; Cornoni-Huntley, J.; Hays, J.C.; Blazer, D. Insomnia and Heart Disease: A Review of Epidemiologic Studies. J. Psychosom. Res. 1999, 4, 313–333. [Google Scholar] [CrossRef] [PubMed]
  4. Bruno, R.M.; Palagini, L.; Gemignani, A.; Virdis, A.; Giulio, A.D.; Ghiadoni, L.; Riemann, D.; Taddei, S. Poor Sleep Quality and Resistant Hypertension. Sleep Med. 2013, 11, 1157–1163. [Google Scholar] [CrossRef]
  5. Wallace, D.M.; Ramos, A.R.; Rundek, T. Sleep Disorders and Stroke. Int. J. Stroke 2012, 3, 231–242. [Google Scholar] [CrossRef]
  6. Rechtschaffen, A.; Kales, A. A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects. Clin. Neurophysiol. 1969, 2, 246–247. [Google Scholar]
  7. Himanena, S.L.; Hasan, J. Limitations of Rechtschaffen and Kales. Sleep Med. Rev. 2004, 2, 149–167. [Google Scholar] [CrossRef]
  8. Iber, C. The AASM Manual for the Scoring of Sleep and Associated Events: Rules Terminology and Technical Specifications; American Academy of Sleep Medicine: Westchester, IL, USA, 2007. [Google Scholar]
  9. Alickovic, E.; Subasi, A. Ensemble SVM Method for Automatic Sleep Stage Classification. IEEE Instrum. Meas. 2018, 6, 1258–1265. [Google Scholar] [CrossRef]
  10. Phan, H.; Do, Q.; Do, T.L.; Vu, D.L. Metric Learning for Automatic Sleep Stage Classification. In Proceedings of the 35th Annual International Conference of the IEEE EMBS, Osaka, Japan, 3–7 July 2013; pp. 5025–5028. [Google Scholar]
  11. Liang, S.F.; Chen, C.F.; Zeng, J.H.; Pan, S.T. Application of Genetic Algorithm and Fuzzy Vector Quantization on EEG-Based Automatic Sleep Staging by Using Hidden Markov Model. In Proceedings of the 2014 International Conference on Machine Learning and Cybernetics, Lanzhou, China, 13–16 July 2014; pp. 567–572. [Google Scholar]
  12. Fraiwan, L.; Lweesy, K.; Khasawneh, N.; Wenz, H.; Dickhaus, H. Automated Sleep Stage Identification System Based on Time-Frequency Analysis of A Single EEG Channel and Random Forest Classifier. Comput. Methods Programs Biomed. 2012, 1, 10–19. [Google Scholar] [CrossRef]
  13. Khalighi, S.; Sousa, T.; Oliveira, D.; Pires, G.; Nunes, U. Efficient Feature Selection for Sleep Staging Based on Maximal Overlap Discrete Wavelet Transform and SVM. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 3306–3309. [Google Scholar]
  14. Wang, Z.; Li, M.; Wang, H.; Jiang, H.; Yao, Y.; Zhang, H.; Xin, J. Breast Cancer Detection Using Extreme Learning Machine Based on Feature Fusion with CNN Deep Features. IEEE Access 2019, 7, 105146–105158. [Google Scholar] [CrossRef]
  15. Sun, M.; Wang, F.; Min, T.; Zang, T.; Wang, Y. Prediction for High Risk Clinical Symptoms of Epilepsy Based on Deep Learning Algorithm. IEEE Access 2018, 6, 77596–77605. [Google Scholar] [CrossRef]
  16. Hammad, M.; Iliyasu, A.M.; Subasi, A.; Ho, E.S.L.; El-Latif, A.A.A. A Multitier Deep Learning Model for Arrhythmia Detection. IEEE Trans. Instrum. Meas. 2021, 70, 3033072. [Google Scholar] [CrossRef]
  17. Khan, M.; Saddik, A.E.; Alotaibi, F.S.; Pham, N.T. AAD-Net: Advanced end-to-end signal processing system for human emotion detection & recognition using attention-based deep echo state network. Knowl.-Based Syst. 2023, 270, 110525. [Google Scholar]
  18. Ishaq, M.; Khan, M.; Kwon, S. TC-Net: A Modest & Lightweight Emotion Recognition System Using Temporal Convolution Network. Comput. Syst. Sci. Eng. 2023, 46, 3355–3369. [Google Scholar]
  19. Kanwal, S.; Uzair, M.; Ullah, H.; Khan, S.D.; Ullah, M.; Cheikh, F.A. An image based prediction model for sleep stage identification. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1366–1370. [Google Scholar]
  20. Huang, J.; Ren, L.; Zhou, X.; Yan, K. An Improved Neural Network Based on SENet for Sleep Stage Classification. IEEE J. Biomed. Health Informat. 2022, 10, 4948–4956. [Google Scholar] [CrossRef]
  21. Toma, T.I.; Choi, S. An End-to-End Convolutional Recurrent Neural Network with Multi-Source Data Fusion for Sleep Stage Classification. In Proceedings of the 2023 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Bali, Indonesia, 20–23 February 2023; pp. 564–569. [Google Scholar]
  22. Kemp, B.; Zwinderman, A.H.; Tuk, B.; Kamphuisen, H.A.C.; Oberye, J.J.L. Analysis of A Sleep-Dependent Neuronal Feedback Loop: The Slow-Wave Microcontinuity of The EEG. IEEE Trans. Biomed. Eng. 2000, 9, 1185–1194. [Google Scholar] [CrossRef]
  23. Jang, J.S.R. ANFIS: Adaptive-Network-Based Fuzzy Inference System. IEEE Trans. Syst. Man Cybern. 1993, 3, 665–685. [Google Scholar] [CrossRef]
  24. Limei, W.; Zongxue, Z.; Xiaoying, L. The IT2FNN Synchronous Control for H-Type Gantry Stage Driven by Dual Linear Motors. In Proceedings of the 2017 29th Chinese Control And Decision Conference (CCDC), Chongqing, China, 28–30 May 2017; pp. 4716–4721. [Google Scholar]
  25. Mamdani, E.H.; Assilian, S. An Experiment in Linguistic Synthesis with A Fuzzy Logic Controller. Int. J. Man-Mach. Stud. 1975, 1, 1–13. [Google Scholar] [CrossRef]
  26. Takagi, T.; Sugeno, M. Fuzzy Identification of Systems and Its Applications to Modeling and Control. IEEE Trans. Syst. Man Cybern. 1985, 1, 116–132. [Google Scholar] [CrossRef]
  27. Zhang, Y.Q.; Kandel, A. Compensatory Neurofuzzy Systems with Fast Learning Algorithms. IEEE Trans. Neural Netw. 1998, 1, 83–105. [Google Scholar] [CrossRef]
  28. Karna, S.K.; Sahai, R. An Overview on Taguchi Method. Int. J. Eng. Math. Sci. 2012, 1, 1–7. [Google Scholar]
  29. Supratak, A.; Dong, H.; Wu, C.; Guo, Y. DeepSleepNet: A Model for Automatic Sleep Stage Scoring Based on Raw Single-Channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 11, 1998–2008. [Google Scholar] [CrossRef] [PubMed]
  30. Fiorillo, L.; Favaro, P.; Faraci, F.D. DeepSleepNet-Lite: A Simplified Automatic Sleep Stage Scoring Model with Uncertainty Estimates. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 2076–2085. [Google Scholar] [CrossRef] [PubMed]
  31. Phan, H.; Chén, O.Y.; Tran, M.C.; Koch, P.; Mertins, A.; Vos, M.D. XSleepNet: Multi-View Sequential Model for Automatic Sleep Staging. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 9, 5903–5915. [Google Scholar] [CrossRef] [PubMed]
  32. Eldele, E.; Chen, Z.; Liu, C.; Wu, M.; Kwoh, C.-K.; Li, X.; Guan, C. An Attention-Based Deep Learning Approach for Sleep Stage Classification with Single-Channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 809–818. [Google Scholar] [CrossRef] [PubMed]
  33. Zhou, D.; Xu, Q.; Wang, J.; Zhang, J.; Hu, G.; Kettunen, L.; Chang, Z.; Cong, F. LightSleepNet: A Lightweight Deep Model for Rapid Sleep Stage Classification with Spectrograms. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Mexico City, Mexico, 1–5 November 2021; pp. 43–46. [Google Scholar]
Figure 1. Architecture of the proposed T-MCCFNN model.
Figure 1. Architecture of the proposed T-MCCFNN model.
Applsci 13 10442 g001
Figure 2. Fpz-Cz channel location.
Figure 2. Fpz-Cz channel location.
Applsci 13 10442 g002
Figure 3. S/N ratio response graph.
Figure 3. S/N ratio response graph.
Applsci 13 10442 g003
Figure 4. Confusion matrix of (a) MCCFNN and (b) T-MCCFNN.
Figure 4. Confusion matrix of (a) MCCFNN and (b) T-MCCFNN.
Applsci 13 10442 g004
Figure 5. Comparison of learning curves of MCCFNN and MCFNN. (a) Accuracy (b) Cross-entropy.
Figure 5. Comparison of learning curves of MCCFNN and MCFNN. (a) Accuracy (b) Cross-entropy.
Applsci 13 10442 g005
Table 1. Each control factor and its level.
Table 1. Each control factor and its level.
No.Control FactorsLevel 1Level 2Level 3
AConv1_filters3264128
BConv1_kernel_size50100150
CConv1_strides468
DScale248
EConv2_filters64128256
FConv2_kernel_size789
GConv2_strides123
Table 2. L27( 3 7 ) orthogonal array.
Table 2. L27( 3 7 ) orthogonal array.
No.ABCDEFG
13250426471
232504212882
332504225693
432100646471
5321006412882
6321006425693
732150886471
8321508812882
9321508825693
106450686483
1164506812891
1264506825672
1364100826483
14641008212891
15641008225672
1664150446483
17641504412891
18641504425672
1912850846492
20128508412873
21128508425681
22128100486492
231281004812873
241281004825681
25128150626492
261281506212873
271281506225681
Table 3. Initial parameters of the proposed T-MCCFNN model.
Table 3. Initial parameters of the proposed T-MCCFNN model.
LayerChannelKernel SizeStridesOther Parameters
Conv layer 1641006-
MaxPool layer 1-88-
Conv layer 2-112881-
Conv layer 2-212881-
MaxPool layer 2-44-
Scale---Scale = 4
Concatenate layer---Mode = Depthwise
GlobalMaxPool layer----
Compensatory fuzzy layer---Rule_node = 64
Output layer5---
Table 4. Experimental results of OA.
Table 4. Experimental results of OA.
No.ABCDEFGAvg_acc (%)
1325042647184.40
23250421288283.63
33250422569382.89
43210064647184.11
532100641288283.43
632100642569382.60
73215088647183.25
832150881288283.13
932150882569382.89
10645068648382.18
116450681289184.19
126450682567284.11
136410082648382.71
1464100821289184.16
1564100822567283.79
166415044648382.62
1764150441289184.56
1864150442567283.68
191285084649283.30
2012850841287382.66
2112850842568184.72
2212810048649283.49
23128100481287383.04
24128100482568184.60
2512815062649283.41
26128150621287381.71
27128150622568184.37
Table 5. S/N ratios of varying parameter combinations.
Table 5. S/N ratios of varying parameter combinations.
LevelABCDEFG
1−1.580−1.560−1.550−1.571−1.590−1.575−1.487
2−1.560−1.561−1.583−1.5564−1.578−1.567−1.561
3−1.569−1.588−1.576−1.573−1.541−1.556−1.661
Delta0.0190.0290.0330.0090.0480.0090.174
Rank5436271
Best parameter64504425691
Table 6. Comparison of results of the proposed T-MCCFNN with other fuzzy neural networks.
Table 6. Comparison of results of the proposed T-MCCFNN with other fuzzy neural networks.
MethodsAccuracyRecallPrecisionF1-ScoreStandard Deviation of Accuracy
MCCNN84.2%81.2%87.5%84.2%0.0559
Mamdani-type MCFNN84.3%82.1%86.7%84.3%0.0298
TSK-type MCFNN84.2%80.8%87.4%84.0%0.0282
Mamdani-type MCCFNN84.1%82.5%86.0%84.2%0.0966
MCCFNN84.4%82.4%86.8%84.5%0.0397
T-MCCFNN85.3%83.3%87.3%85.3%0.0321
Table 7. Performance comparison of the multi-scale and the single-scale convolution kernel.
Table 7. Performance comparison of the multi-scale and the single-scale convolution kernel.
MethodsAccuracyRecallPrecisionF1-Score
T-SSCFNN84.2%81.4%87.3%84.2%
T-MCCFNN85.3%83.3%87.3%85.3%
Table 8. Comparison results of various deep learning methods.
Table 8. Comparison results of various deep learning methods.
MethodsAccuracyF1-ScorePer-Class F1-Score
WN1N2N3REM
Supratak et al. [29]82.0%76.9%84.7%46.6%85.9%84.8%82.4%
Fiorillo et al. [30]82.6%76.3%81.6%42.4%87.4%87.9%82.1%
Phan et al. [31]83.9%78.7%81.6%52.9%88.1%85.3%85.4%
Eldele et al. [32]84.4%78.1%89.7%42.6%88.8%90.2%79.0%
Zhou et al. [33]83.8%75.3%90.0%31.0%88.0%89.0%78.0%
MCCFNN84.4%84.5%89.6%39.2%92.1%88.2%76.3%
T-MCCFNN85.3%85.3%92.7%35.9%88.5%91.2%84.8%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, C.-J.; Lin, C.-J.; Lin, X.-Q. Automatic Sleep Stage Classification Using a Taguchi-Based Multiscale Convolutional Compensatory Fuzzy Neural Network. Appl. Sci. 2023, 13, 10442. https://doi.org/10.3390/app131810442

AMA Style

Lin C-J, Lin C-J, Lin X-Q. Automatic Sleep Stage Classification Using a Taguchi-Based Multiscale Convolutional Compensatory Fuzzy Neural Network. Applied Sciences. 2023; 13(18):10442. https://doi.org/10.3390/app131810442

Chicago/Turabian Style

Lin, Chun-Jung, Cheng-Jian Lin, and Xue-Qian Lin. 2023. "Automatic Sleep Stage Classification Using a Taguchi-Based Multiscale Convolutional Compensatory Fuzzy Neural Network" Applied Sciences 13, no. 18: 10442. https://doi.org/10.3390/app131810442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop