Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (16)

Search Parameters:
Keywords = electroencephalogram spectrogram

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 930 KB  
Article
Investigation of the MobileNetV2 Optimal Feature Extraction Layer for EEG-Based Dementia Severity Classification: A Comparative Study
by Noor Kamal Al-Qazzaz, Sawal Hamid Bin Mohd Ali and Siti Anom Ahmad
Algorithms 2025, 18(10), 620; https://doi.org/10.3390/a18100620 - 1 Oct 2025
Abstract
Diagnosing dementia and recognizing substantial cognitive decline are challenging tasks. Thus, the objective of this study was to classify electroencephalograms (EEGs) recorded during a working memory task in 15 patients with mild cognitive impairment (MCogImp), 5 patients with vascular dementia (VasD), and 15 [...] Read more.
Diagnosing dementia and recognizing substantial cognitive decline are challenging tasks. Thus, the objective of this study was to classify electroencephalograms (EEGs) recorded during a working memory task in 15 patients with mild cognitive impairment (MCogImp), 5 patients with vascular dementia (VasD), and 15 healthy controls (NC). Before creating spectrogram pictures from the EEG dataset, the data were subjected to preprocessing, which included preprocessing using conventional filters and the discrete wavelet transformation. The convolutional neural network (CNN) MobileNetV2 was employed in our investigation to identify features and assess the severity of dementia. The features were extracted from five layers of the MobileNetV2 CNN architecture—convolutional layers (‘Conv-1’), batch normalization (‘Conv-1-bn’), clipped ReLU (‘out-relu’), 2D Global Average Pooling (‘global-average-pooling2d1’), and fully connected (‘Logits’) layers. This was carried out to find the efficient features layer for dementia severity from EEGs. Feature extraction from MobileNetV2’s five layers was carried out using a decision tree (DT) and k-nearest neighbor (KNN) machine learning (ML) classifier, in conjunction with a MobileNetV2 deep learning (DL) network. The study’s findings show that the DT classifier performed best using features derived from MobileNetV2 with the 2D Global Average Pooling (global-average-pooling2d-1) layer, achieving an accuracy score of 95.9%. Second place went to the characteristics of the fully connected (Logits) layer, which achieved a score of 95.3%. The findings of this study endorse the utilization of deep processing algorithms, offering a viable approach for improving early dementia identification with high precision, hence facilitating the differentiation among NC individuals, VasD patients, and MCogImp patients. Full article
(This article belongs to the Special Issue Machine Learning in Medical Signal and Image Processing (3rd Edition))
Show Figures

Figure 1

23 pages, 2640 KB  
Article
DenseNet-Based Classification of EEG Abnormalities Using Spectrograms
by Lan Wei and Catherine Mooney
Algorithms 2025, 18(8), 486; https://doi.org/10.3390/a18080486 - 5 Aug 2025
Viewed by 651
Abstract
Electroencephalogram (EEG) analysis is essential for diagnosing neurological disorders but typically requires expert interpretation and significant time. Purpose: This study aims to automate the classification of normal and abnormal EEG recordings to support clinical diagnosis and reduce manual workload. Automating the initial screening [...] Read more.
Electroencephalogram (EEG) analysis is essential for diagnosing neurological disorders but typically requires expert interpretation and significant time. Purpose: This study aims to automate the classification of normal and abnormal EEG recordings to support clinical diagnosis and reduce manual workload. Automating the initial screening of EEGs can help clinicians quickly identify potential neurological abnormalities, enabling timely intervention and guiding further diagnostic and treatment strategies. Methodology: We utilized the Temple University Hospital EEG dataset to develop a DenseNet-based deep learning model. To enable a fair comparison of different EEG representations, we used three input types: signal images, spectrograms, and scalograms. To reduce dimensionality and simplify computation, we focused on two channels: T5 and O1. For interpretability, we applied Local Interpretable Model-agnostic Explanations (LIME) and Gradient-weighted Class Activation Mapping (Grad-CAM) to visualize the EEG regions influencing the model’s predictions. Key Findings: Among the input types, spectrogram-based representations achieved the highest classification accuracy, indicating that time-frequency features are especially effective for this task. The model demonstrated strong performance overall, and the integration of LIME and Grad-CAM provided transparent explanations of its decisions, enhancing interpretability. This approach offers a practical and interpretable solution for automated EEG screening, contributing to more efficient clinical workflows and better understanding of complex neurological conditions. Full article
(This article belongs to the Special Issue AI-Assisted Medical Diagnostics)
Show Figures

Figure 1

25 pages, 6137 KB  
Article
EEG-Based Seizure Detection Using Dual-Branch CNN-ViT Network Integrating Phase and Power Spectrograms
by Zhuohan Wang, Yaoqi Hu, Qingyue Xin, Guanghao Jin, Yazhou Zhao, Weidong Zhou and Guoyang Liu
Brain Sci. 2025, 15(5), 509; https://doi.org/10.3390/brainsci15050509 - 16 May 2025
Cited by 1 | Viewed by 1585
Abstract
Background/Objectives: Epilepsy is a common neurological disorder with pathological mechanisms closely associated with the spatiotemporal dynamic characteristics of electroencephalogram (EEG) signals. Although significant progress has been made in epileptic seizure detection methods using time–frequency analysis, current research still faces challenges in terms of [...] Read more.
Background/Objectives: Epilepsy is a common neurological disorder with pathological mechanisms closely associated with the spatiotemporal dynamic characteristics of electroencephalogram (EEG) signals. Although significant progress has been made in epileptic seizure detection methods using time–frequency analysis, current research still faces challenges in terms of an insufficient utilization of phase information. Methods: In this study, we propose an effective epileptic seizure detection framework based on continuous wavelet transform (CWT) and a hybrid network consisting of convolutional neural network (CNN) and vision transformer (ViT). First, the raw EEG signals are processed by the CWT. Then, the phase spectrogram and power spectrogram of the EEG are generated, and they are sent into the designed CNN and ViT branches of the network to extract more discriminative EEG features. Finally, the features output from the two branches are fused and fed into the classification network to obtain the detection results. Results: Experimental results on the CHB-MIT public dataset and our SH-SDU clinical dataset show that the proposed framework achieves sensitivities of 98.09% and 89.02%, specificities of 98.21% and 95.46%, and average accuracies of 98.45% and 94.66%, respectively. Furthermore, we compared the spectral characteristics of CWT with other time–frequency transforms within the hybrid architecture, demonstrating the advantages of the CWT-based CNN-ViT architecture. Conclusions: These results highlight the outstanding epileptic seizure detection performance of the proposed framework and its significant clinical feasibility. Full article
Show Figures

Figure 1

24 pages, 18899 KB  
Review
Utility of Quantitative EEG in Neurological Emergencies and ICU Clinical Practice
by Misericordia Veciana de las Heras, Jacint Sala-Padro, Jordi Pedro-Perez, Beliu García-Parra, Guillermo Hernández-Pérez and Merce Falip
Brain Sci. 2024, 14(9), 939; https://doi.org/10.3390/brainsci14090939 - 20 Sep 2024
Cited by 2 | Viewed by 3089
Abstract
The electroencephalogram (EEG) is a cornerstone tool for the diagnosis, management, and prognosis of selected patient populations. EEGs offer significant advantages such as high temporal resolution, real-time cortical function assessment, and bedside usability. The quantitative EEG (qEEG) added the possibility of long recordings [...] Read more.
The electroencephalogram (EEG) is a cornerstone tool for the diagnosis, management, and prognosis of selected patient populations. EEGs offer significant advantages such as high temporal resolution, real-time cortical function assessment, and bedside usability. The quantitative EEG (qEEG) added the possibility of long recordings being processed in a compressive manner, making EEG revision more efficient for experienced users, and more friendly for new ones. Recent advancements in commercially available software, such as Persyst, have significantly expanded and facilitated the use of qEEGs, marking the beginning of a new era in its application. As a result, there has been a notable increase in the practical, real-world utilization of qEEGs in recent years. This paper aims to provide an overview of the current applications of qEEGs in daily neurological emergencies and ICU practice, and some elementary principles of qEEGs using Persyst software in clinical settings. This article illustrates basic qEEG patterns encountered in critical care and adopts the new terminology proposed for spectrogram reporting. Full article
Show Figures

Figure 1

20 pages, 6326 KB  
Article
A Novel Prognostic Model Using Chaotic CNN with Hybridized Spoofing for Enhancing Diagnostic Accuracy in Epileptic Seizure Prediction
by Preethi Palanisamy, Shabana Urooj, Rajesh Arunachalam and Aime Lay-Ekuakille
Diagnostics 2023, 13(21), 3382; https://doi.org/10.3390/diagnostics13213382 - 3 Nov 2023
Cited by 22 | Viewed by 1788
Abstract
Epileptic seizure detection has undergone progressive advancements since its conception in the 1970s. From proof-of-concept experiments in the latter part of that decade, it has now become a vibrant area of clinical and laboratory research. In an effort to bring this technology closer [...] Read more.
Epileptic seizure detection has undergone progressive advancements since its conception in the 1970s. From proof-of-concept experiments in the latter part of that decade, it has now become a vibrant area of clinical and laboratory research. In an effort to bring this technology closer to practical application in human patients, this study introduces a customized approach to selecting electroencephalogram (EEG) features and electrode positions for seizure prediction. The focus is on identifying precursors that occur within 10 min of the onset of abnormal electrical activity during a seizure. However, there are security concerns related to safeguarding patient EEG recordings against unauthorized access and network-based attacks. Therefore, there is an urgent need for an efficient prediction and classification method for encrypted EEG data. This paper presents an effective system for analyzing and recognizing encrypted EEG information using Arnold transform algorithms, chaotic mapping, and convolutional neural networks (CNNs). In this system, the EEG time series from each channel is converted into a 2D spectrogram image, which is then encrypted using chaotic algorithms. The encrypted data is subsequently processed by CNNs coupled with transfer learning (TL) frameworks. To optimize the fusion parameters of the ensemble learning classifiers, a hybridized spoofing optimization method is developed by combining the characteristics of corvid and gregarious-seeking agents. The evaluation of the model’s effectiveness yielded the following results: 98.9 ± 0.3% accuracy, 98.2 ± 0.7% sensitivity, 98.6 ± 0.6% specificity, 98.6 ± 0.6% precision, and an F1 measure of 98.9 ± 0.6%. When compared with other state-of-the-art techniques applied to the same dataset, this novel strategy demonstrated one of the most effective seizure detection systems, as evidenced by these results. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

14 pages, 2883 KB  
Article
A New Hybrid Approach Based on Time Frequency Images and Deep Learning Methods for Diagnosis of Migraine Disease and Investigation of Stimulus Effect
by Fırat Orhanbulucu, Fatma Latifoğlu and Recep Baydemir
Diagnostics 2023, 13(11), 1887; https://doi.org/10.3390/diagnostics13111887 - 28 May 2023
Cited by 8 | Viewed by 2406
Abstract
Migraine is a neurological disorder that is associated with severe headaches and seriously affects the lives of patients. Diagnosing Migraine Disease (MD) can be laborious and time-consuming for specialists. For this reason, systems that can assist specialists in the early diagnosis of MD [...] Read more.
Migraine is a neurological disorder that is associated with severe headaches and seriously affects the lives of patients. Diagnosing Migraine Disease (MD) can be laborious and time-consuming for specialists. For this reason, systems that can assist specialists in the early diagnosis of MD are important. Although migraine is one of the most common neurological diseases, there are very few studies on the diagnosis of MD, especially electroencephalogram (EEG)-and deep learning (DL)-based studies. For this reason, in this study, a new system has been proposed for the early diagnosis of EEG- and DL-based MD. In the proposed study, EEG signals obtained from the resting state (R), visual stimulus (V), and auditory stimulus (A) from 18 migraine patients and 21 healthy control (HC) groups were used. By applying continuous wavelet transform (CWT) and short-time Fourier transform (STFT) methods to these EEG signals, scalogram-spectrogram images were obtained in the time-frequency (T-F) plane. Then, these images were applied as inputs in three different convolutional neural networks (CNN) architectures (AlexNet, ResNet50, SqueezeNet) that proposed deep convolutional neural network (DCNN) models and classification was performed. The results of the classification process were evaluated, taking into account accuracy (acc.), sensitivity (sens.), specificity (spec.), and performance criteria, and the performances of the preferred methods and models in this study were compared. In this way, the situation, method, and model that showed the most successful performance for the early diagnosis of MD were determined. Although the classification results are close to each other, the resting state, CWT method, and AlexNet classifier showed the most successful performance (Acc: 99.74%, Sens: 99.9%, Spec: 99.52%). We think that the results obtained in this study are promising for the early diagnosis of MD and can be of help to experts. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

17 pages, 1908 KB  
Article
An AI-Inspired Spatio-Temporal Neural Network for EEG-Based Emotional Status
by Fahad Mazaed Alotaibi and Fawad
Sensors 2023, 23(1), 498; https://doi.org/10.3390/s23010498 - 2 Jan 2023
Cited by 18 | Viewed by 3504
Abstract
The accurate identification of the human emotional status is crucial for an efficient human–robot interaction (HRI). As such, we have witnessed extensive research efforts made in developing robust and accurate brain–computer interfacing models based on diverse biosignals. In particular, previous research has shown [...] Read more.
The accurate identification of the human emotional status is crucial for an efficient human–robot interaction (HRI). As such, we have witnessed extensive research efforts made in developing robust and accurate brain–computer interfacing models based on diverse biosignals. In particular, previous research has shown that an Electroencephalogram (EEG) can provide deep insight into the state of emotion. Recently, various handcrafted and deep neural network (DNN) models were proposed by researchers for extracting emotion-relevant features, which offer limited robustness to noise that leads to reduced precision and increased computational complexity. The DNN models developed to date were shown to be efficient in extracting robust features relevant to emotion classification; however, their massive feature dimensionality problem leads to a high computational load. In this paper, we propose a bag-of-hybrid-deep-features (BoHDF) extraction model for classifying EEG signals into their respective emotion class. The invariance and robustness of the BoHDF is further enhanced by transforming EEG signals into 2D spectrograms before the feature extraction stage. Such a time-frequency representation fits well with the time-varying behavior of EEG patterns. Here, we propose to combine the deep features from the GoogLeNet fully connected layer (one of the simplest DNN models) together with the OMTLBP_SMC texture-based features, which we recently developed, followed by a K-nearest neighbor (KNN) clustering algorithm. The proposed model, when evaluated on the DEAP and SEED databases, achieves a 93.83 and 96.95% recognition accuracy, respectively. The experimental results using the proposed BoHDF-based algorithm show an improved performance in comparison to previously reported works with similar setups. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

16 pages, 3039 KB  
Article
An Ensemble Learning Method for Emotion Charting Using Multimodal Physiological Signals
by Amna Waheed Awan, Syed Muhammad Usman, Shehzad Khalid, Aamir Anwar, Roobaea Alroobaea, Saddam Hussain, Jasem Almotiri, Syed Sajid Ullah and Muhammad Usman Akram
Sensors 2022, 22(23), 9480; https://doi.org/10.3390/s22239480 - 4 Dec 2022
Cited by 16 | Viewed by 4200
Abstract
Emotion charting using multimodal signals has gained great demand for stroke-affected patients, for psychiatrists while examining patients, and for neuromarketing applications. Multimodal signals for emotion charting include electrocardiogram (ECG) signals, electroencephalogram (EEG) signals, and galvanic skin response (GSR) signals. EEG, ECG, and GSR [...] Read more.
Emotion charting using multimodal signals has gained great demand for stroke-affected patients, for psychiatrists while examining patients, and for neuromarketing applications. Multimodal signals for emotion charting include electrocardiogram (ECG) signals, electroencephalogram (EEG) signals, and galvanic skin response (GSR) signals. EEG, ECG, and GSR are also known as physiological signals, which can be used for identification of human emotions. Due to the unbiased nature of physiological signals, this field has become a great motivation in recent research as physiological signals are generated autonomously from human central nervous system. Researchers have developed multiple methods for the classification of these signals for emotion detection. However, due to the non-linear nature of these signals and the inclusion of noise, while recording, accurate classification of physiological signals is a challenge for emotion charting. Valence and arousal are two important states for emotion detection; therefore, this paper presents a novel ensemble learning method based on deep learning for the classification of four different emotional states including high valence and high arousal (HVHA), low valence and low arousal (LVLA), high valence and low arousal (HVLA) and low valence high arousal (LVHA). In the proposed method, multimodal signals (EEG, ECG, and GSR) are preprocessed using bandpass filtering and independent components analysis (ICA) for noise removal in EEG signals followed by discrete wavelet transform for time domain to frequency domain conversion. Discrete wavelet transform results in spectrograms of the physiological signal and then features are extracted using stacked autoencoders from those spectrograms. A feature vector is obtained from the bottleneck layer of the autoencoder and is fed to three classifiers SVM (support vector machine), RF (random forest), and LSTM (long short-term memory) followed by majority voting as ensemble classification. The proposed system is trained and tested on the AMIGOS dataset with k-fold cross-validation. The proposed system obtained the highest accuracy of 94.5% and shows improved results of the proposed method compared with other state-of-the-art methods. Full article
(This article belongs to the Special Issue Advances in IoMT for Healthcare Systems)
Show Figures

Figure 1

17 pages, 3033 KB  
Article
A Deep Learning Method Approach for Sleep Stage Classification with EEG Spectrogram
by Chengfan Li, Yueyu Qi, Xuehai Ding, Junjuan Zhao, Tian Sang and Matthew Lee
Int. J. Environ. Res. Public Health 2022, 19(10), 6322; https://doi.org/10.3390/ijerph19106322 - 23 May 2022
Cited by 67 | Viewed by 8978
Abstract
The classification of sleep stages is an important process. However, this process is time-consuming, subjective, and error-prone. Many automated classification methods use electroencephalogram (EEG) signals for classification. These methods do not classify well enough and perform poorly in the N1 due to unbalanced [...] Read more.
The classification of sleep stages is an important process. However, this process is time-consuming, subjective, and error-prone. Many automated classification methods use electroencephalogram (EEG) signals for classification. These methods do not classify well enough and perform poorly in the N1 due to unbalanced data. In this paper, we propose a sleep stage classification method using EEG spectrogram. We have designed a deep learning model called EEGSNet based on multi-layer convolutional neural networks (CNNs) to extract time and frequency features from the EEG spectrogram, and two-layer bi-directional long short-term memory networks (Bi-LSTMs) to learn the transition rules between features from adjacent epochs and to perform the classification of sleep stages. In addition, to improve the generalization ability of the model, we have used Gaussian error linear units (GELUs) as the activation function of CNN. The proposed method was evaluated by four public databases, the Sleep-EDFX-8, Sleep-EDFX-20, Sleep-EDFX-78, and SHHS. The accuracy of the method is 94.17%, 86.82%, 83.02% and 85.12%, respectively, for the four datasets, the MF1 is 87.78%, 81.57%, 77.26% and 78.54%, respectively, and the Kappa is 0.91, 0.82, 0.77 and 0.79, respectively. In addition, our proposed method achieved better classification results on N1, with an F1-score of 70.16%, 52.41%, 50.03% and 47.26% for the four datasets. Full article
Show Figures

Figure 1

15 pages, 887 KB  
Article
Representation Learning for EEG-Based Biometrics Using Hilbert–Huang Transform
by Mikhail Svetlakov, Ilya Kovalev, Anton Konev, Evgeny Kostyuchenko and Artur Mitsel
Computers 2022, 11(3), 47; https://doi.org/10.3390/computers11030047 - 20 Mar 2022
Cited by 13 | Viewed by 3782
Abstract
A promising approach to overcome the various shortcomings of password systems is the use of biometric authentication, in particular the use of electroencephalogram (EEG) data. In this paper, we propose a subject-independent learning method for EEG-based biometrics using Hilbert spectrograms of the data. [...] Read more.
A promising approach to overcome the various shortcomings of password systems is the use of biometric authentication, in particular the use of electroencephalogram (EEG) data. In this paper, we propose a subject-independent learning method for EEG-based biometrics using Hilbert spectrograms of the data. The proposed neural network architecture treats the spectrogram as a collection of one-dimensional series and applies one-dimensional dilated convolutions over them, and a multi-similarity loss was used as the loss function for subject-independent learning. The architecture was tested on the publicly available PhysioNet EEG Motor Movement/Imagery Dataset (PEEGMIMDB) with a 14.63% Equal Error Rate (EER) achieved. The proposed approach’s main advantages are subject independence and suitability for interpretation via created spectrograms and the integrated gradients method. Full article
(This article belongs to the Special Issue Explainable Artificial Intelligence for Biometrics 2021)
Show Figures

Figure 1

17 pages, 4136 KB  
Article
The Choice of Time–Frequency Representations of Non-Stationary Signals Affects Machine Learning Model Accuracy: A Case Study on Earthquake Detection from LEN-DB Data
by Marko Njirjak, Erik Otović, Dario Jozinović, Jonatan Lerga, Goran Mauša, Alberto Michelini and Ivan Štajduhar
Mathematics 2022, 10(6), 965; https://doi.org/10.3390/math10060965 - 17 Mar 2022
Cited by 6 | Viewed by 3724
Abstract
Non-stationary signals are often analyzed using raw waveform data or spectrograms of those data; however, the possibility of alternative time–frequency representations being more informative than the original data or spectrograms is yet to be investigated. This paper tested whether alternative time–frequency representations could [...] Read more.
Non-stationary signals are often analyzed using raw waveform data or spectrograms of those data; however, the possibility of alternative time–frequency representations being more informative than the original data or spectrograms is yet to be investigated. This paper tested whether alternative time–frequency representations could be more informative for machine learning classification of seismological data. The mentioned hypothesis was evaluated by training three well-established convolutional neural networks using nine time–frequency representations. The results were compared to the base model, which was trained on the raw waveform data. The signals that were used in the experiment are three-component seismogram instances from the Local Earthquakes and Noise DataBase (LEN-DB). The results demonstrate that Pseudo Wigner–Ville and Wigner–Ville time–frequency representations yield significantly better results than the base model, while spectrogram and Margenau–Hill perform significantly worse (p < 0.01). Interestingly, the spectrogram, which is often used in signal analysis, had inferior performance when compared to the base model. The findings presented in this research could have notable impacts in the fields of geophysics and seismology as the phenomena that were previously hidden in the seismic noise are now more easily identified. Furthermore, the results indicate that applying Pseudo Wigner–Ville or Wigner–Ville time–frequency representations could result in a large increase in earthquakes in the catalogs and lessen the need to add new stations with an overall reduction in the costs. Finally, the proposed approach of extracting valuable information through time–frequency representations could be applied in other domains as well, such as electroencephalogram and electrocardiogram signal analysis, speech recognition, gravitational waves investigation, and so on. Full article
Show Figures

Graphical abstract

15 pages, 3842 KB  
Article
GaborPDNet: Gabor Transformation and Deep Neural Network for Parkinson’s Disease Detection Using EEG Signals
by Hui Wen Loh, Chui Ping Ooi, Elizabeth Palmer, Prabal Datta Barua, Sengul Dogan, Turker Tuncer, Mehmet Baygin and U. Rajendra Acharya
Electronics 2021, 10(14), 1740; https://doi.org/10.3390/electronics10141740 - 20 Jul 2021
Cited by 110 | Viewed by 7930
Abstract
Parkinson’s disease (PD) is globally the most common neurodegenerative movement disorder. It is characterized by a loss of dopaminergic neurons in the substantia nigra of the brain. However, current methods to diagnose PD on the basis of clinical features of Parkinsonism may lead [...] Read more.
Parkinson’s disease (PD) is globally the most common neurodegenerative movement disorder. It is characterized by a loss of dopaminergic neurons in the substantia nigra of the brain. However, current methods to diagnose PD on the basis of clinical features of Parkinsonism may lead to misdiagnoses. Hence, noninvasive methods such as electroencephalographic (EEG) recordings of PD patients can be an alternative biomarker. In this study, a deep-learning model is proposed for automated PD diagnosis. EEG recordings of 16 healthy controls and 15 PD patients were used for analysis. Using Gabor transform, EEG recordings were converted into spectrograms, which were used to train the proposed two-dimensional convolutional neural network (2D-CNN) model. As a result, the proposed model achieved high classification accuracy of 99.46% (±0.73) for 3-class classification (healthy controls, and PD patients with and without medication) using tenfold cross-validation. This indicates the potential of proposed model to simultaneously automatically detect PD patients and their medication status. The proposed model is ready to be validated with a larger database before implementation as a computer-aided diagnostic (CAD) tool for clinical-decision support. Full article
Show Figures

Figure 1

25 pages, 3512 KB  
Article
A Statistical Method for Exploratory Data Analysis Based on 2D and 3D Area under Curve Diagrams: Parkinson’s Disease Investigation
by Olga Sergeevna Sushkova, Alexei Alexandrovich Morozov, Alexandra Vasilievna Gabova, Alexei Vyacheslavovich Karabanov and Sergey Nikolaevich Illarioshkin
Sensors 2021, 21(14), 4700; https://doi.org/10.3390/s21144700 - 9 Jul 2021
Cited by 19 | Viewed by 7301
Abstract
A statistical method for exploratory data analysis based on 2D and 3D area under curve (AUC) diagrams was developed. The method was designed to analyze electroencephalogram (EEG), electromyogram (EMG), and tremorogram data collected from patients with Parkinson’s disease. The idea of the method [...] Read more.
A statistical method for exploratory data analysis based on 2D and 3D area under curve (AUC) diagrams was developed. The method was designed to analyze electroencephalogram (EEG), electromyogram (EMG), and tremorogram data collected from patients with Parkinson’s disease. The idea of the method of wave train electrical activity analysis is that we consider the biomedical signal as a combination of the wave trains. The wave train is the increase in the power spectral density of the signal localized in time, frequency, and space. We detect the wave trains as the local maxima in the wavelet spectrograms. We do not consider wave trains as a special kind of signal. The wave train analysis method is different from standard signal analysis methods such as Fourier analysis and wavelet analysis in the following way. Existing methods for analyzing EEG, EMG, and tremor signals, such as wavelet analysis, focus on local time–frequency changes in the signal and therefore do not reveal the generalized properties of the signal. Other methods such as standard Fourier analysis ignore the local time–frequency changes in the characteristics of the signal and, consequently, lose a large amount of information that existed in the signal. The method of wave train electrical activity analysis resolves the contradiction between these two approaches because it addresses the generalized characteristics of the biomedical signal based on local time–frequency changes in the signal. We investigate the following wave train parameters: wave train central frequency, wave train maximal power spectral density, wave train duration in periods, and wave train bandwidth. We have developed special graphical diagrams, named AUC diagrams, to determine what wave trains are characteristic of neurodegenerative diseases. In this paper, we consider the following types of AUC diagrams: 2D and 3D diagrams. The technique of working with AUC diagrams is illustrated by examples of analysis of EMG in patients with Parkinson’s disease and healthy volunteers. It is demonstrated that new regularities useful for the high-accuracy diagnosis of Parkinson’s disease can be revealed using the method of analyzing the wave train electrical activity and AUC diagrams. Full article
(This article belongs to the Special Issue On the Applications of EMG Sensors and Signals)
Show Figures

Figure 1

20 pages, 6522 KB  
Article
Data Augmentation for Motor Imagery Signal Classification Based on a Hybrid Neural Network
by Kai Zhang, Guanghua Xu, Zezhen Han, Kaiquan Ma, Xiaowei Zheng, Longting Chen, Nan Duan and Sicong Zhang
Sensors 2020, 20(16), 4485; https://doi.org/10.3390/s20164485 - 11 Aug 2020
Cited by 106 | Viewed by 9023
Abstract
As an important paradigm of spontaneous brain-computer interfaces (BCIs), motor imagery (MI) has been widely used in the fields of neurological rehabilitation and robot control. Recently, researchers have proposed various methods for feature extraction and classification based on MI signals. The decoding model [...] Read more.
As an important paradigm of spontaneous brain-computer interfaces (BCIs), motor imagery (MI) has been widely used in the fields of neurological rehabilitation and robot control. Recently, researchers have proposed various methods for feature extraction and classification based on MI signals. The decoding model based on deep neural networks (DNNs) has attracted significant attention in the field of MI signal processing. Due to the strict requirements for subjects and experimental environments, it is difficult to collect large-scale and high-quality electroencephalogram (EEG) data. However, the performance of a deep learning model depends directly on the size of the datasets. Therefore, the decoding of MI-EEG signals based on a DNN has proven highly challenging in practice. Based on this, we investigated the performance of different data augmentation (DA) methods for the classification of MI data using a DNN. First, we transformed the time series signals into spectrogram images using a short-time Fourier transform (STFT). Then, we evaluated and compared the performance of different DA methods for this spectrogram data. Next, we developed a convolutional neural network (CNN) to classify the MI signals and compared the classification performance of after DA. The Fréchet inception distance (FID) was used to evaluate the quality of the generated data (GD) and the classification accuracy, and mean kappa values were used to explore the best CNN-DA method. In addition, analysis of variance (ANOVA) and paired t-tests were used to assess the significance of the results. The results showed that the deep convolutional generative adversarial network (DCGAN) provided better augmentation performance than traditional DA methods: geometric transformation (GT), autoencoder (AE), and variational autoencoder (VAE) (p < 0.01). Public datasets of the BCI competition IV (datasets 1 and 2b) were used to verify the classification performance. Improvements in the classification accuracies of 17% and 21% (p < 0.01) were observed after DA for the two datasets. In addition, the hybrid network CNN-DCGAN outperformed the other classification methods, with average kappa values of 0.564 and 0.677 for the two datasets. Full article
Show Figures

Figure 1

16 pages, 1810 KB  
Article
EEG-Based Multi-Modal Emotion Recognition using Bag of Deep Features: An Optimal Feature Selection Approach
by Muhammad Adeel Asghar, Muhammad Jamil Khan, Fawad, Yasar Amin, Muhammad Rizwan, MuhibUr Rahman, Salman Badnava and Seyed Sajad Mirjavadi
Sensors 2019, 19(23), 5218; https://doi.org/10.3390/s19235218 - 28 Nov 2019
Cited by 88 | Viewed by 8087
Abstract
Much attention has been paid to the recognition of human emotions with the help of electroencephalogram (EEG) signals based on machine learning technology. Recognizing emotions is a challenging task due to the non-linear property of the EEG signal. This paper presents an advanced [...] Read more.
Much attention has been paid to the recognition of human emotions with the help of electroencephalogram (EEG) signals based on machine learning technology. Recognizing emotions is a challenging task due to the non-linear property of the EEG signal. This paper presents an advanced signal processing method using the deep neural network (DNN) for emotion recognition based on EEG signals. The spectral and temporal components of the raw EEG signal are first retained in the 2D Spectrogram before the extraction of features. The pre-trained AlexNet model is used to extract the raw features from the 2D Spectrogram for each channel. To reduce the feature dimensionality, spatial, and temporal based, bag of deep features (BoDF) model is proposed. A series of vocabularies consisting of 10 cluster centers of each class is calculated using the k-means cluster algorithm. Lastly, the emotion of each subject is represented using the histogram of the vocabulary set collected from the raw-feature of a single channel. Features extracted from the proposed BoDF model have considerably smaller dimensions. The proposed model achieves better classification accuracy compared to the recently reported work when validated on SJTU SEED and DEAP data sets. For optimal classification performance, we use a support vector machine (SVM) and k-nearest neighbor (k-NN) to classify the extracted features for the different emotional states of the two data sets. The BoDF model achieves 93.8% accuracy in the SEED data set and 77.4% accuracy in the DEAP data set, which is more accurate compared to other state-of-the-art methods of human emotion recognition. Full article
Show Figures

Figure 1

Back to TopTop