Next Article in Journal
Correction Compensation and Adaptive Cost Aggregation for Deep Laparoscopic Stereo Matching
Next Article in Special Issue
A Systematic Literature Review of Modalities, Trends, and Limitations in Emotion Recognition, Affective Computing, and Sentiment Analysis
Previous Article in Journal
Multicriteria Decision Model for Port Evaluation and Ranking: An Analysis of Container Terminals in Latin America and the Caribbean Using PCA-TOPSIS Methodologies
Previous Article in Special Issue
Real-Time Analysis of Facial Expressions for Mood Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Accuracy Classification of Multiple Distinct Human Emotions Using EEG Differential Entropy Features and ResNet18

1
School of Electronic and Information Engineering, Harbin Institute of Technology at Shenzhen, Shenzhen 518055, China
2
School of Computer Science and Engineering, Huizhou University, Huizhou 516007, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(14), 6175; https://doi.org/10.3390/app14146175
Submission received: 12 June 2024 / Revised: 2 July 2024 / Accepted: 11 July 2024 / Published: 16 July 2024
(This article belongs to the Special Issue Application of Affective Computing)

Abstract

:
The high-accuracy detection of multiple distinct human emotions is crucial for advancing affective computing, mental health diagnostics, and human–computer interaction. The integration of deep learning networks with entropy measures holds significant potential in neuroscience and medicine, especially for analyzing EEG-based emotion states. This study proposes a method combining ResNet18 with differential entropy to identify five types of human emotions (happiness, sadness, fear, disgust, and neutral) from EEG signals. Our approach first calculates the differential entropy of EEG signals to capture the complexity and variability of the emotional states. Then, the ResNet18 network is employed to learn feature representations from the differential entropy measures, which effectively captures the intricate spatiotemporal dynamics inherent in emotional EEG patterns using residual connections. To validate the efficacy of our method, we conducted experiments on the SEED-V dataset, achieving an average accuracy of 95.61%. Our findings demonstrate that the combination of ResNet18 with differential entropy is highly effective in classifying multiple distinct human emotions from EEG signals. This method shows robust generalization and broad applicability, indicating its potential for extension to various pattern recognition tasks across different domains.

1. Introduction

Recognizing human emotional states through EEG signals is of great significance, and has wide applications in neuroscience, psychology, and brain–computer interface fields [1]. These signals contribute to a deeper comprehension of emotional mechanisms and cognitive processes within the brain [2]. By analyzing EEG signal changes, researchers can uncover patterns and characteristics of brain activity associated with different emotional states, thereby advancing our knowledge of emotional regulation, disorders, and related phenomena. The essence of emotional EEG patterns lies in elucidating the neural underpinnings of human emotions and cognitive functions, alongside their extensive utility in diverse fields like medicine, psychology, and engineering, providing valuable insights and support for enhancing human lives and well-being [3].
Deep learning models play a pivotal role in learning sophisticated feature representations from EEG signals, thereby enhancing performance in tasks such as emotion recognition and brain–computer interfaces [4]. These models autonomously learn abstract features embedded within EEG signals, effectively capturing crucial information. Given the complex nonlinear nature of EEG signals, traditional emotion recognition methods relying on manual feature extraction and classification may struggle to capture intricate patterns and details [5]. Deep learning models, on the other hand, autonomously extract features from raw data and represent them at higher abstraction levels, thereby boosting accuracy and efficiency in emotion recognition tasks [6].
Moreover, deep learning models can optimize model structures and algorithms, leading to higher running speeds and real-time performance, thus making emotion recognition systems more practical for real-world applications [7]. For instance, the integration of deep learning with EEG technology represents a promising avenue for advancing emotion recognition capabilities and improving their feasibility in practical settings. For example, Chen et al. introduced a Graph Convolutional Network (GCN) based on Ordinary Differential Equations (ODE), which significantly enhances performance in handling the relationships among EEG channels [8]. In another instance, Xu et al. presented a method leveraging spiking neural networks to achieve high accuracy in recognizing emotions within EEG datasets [9].
EEG signals are known for their inherent complexity and nonlinearity, making them challenging to analyze directly. Entropy serves as a valuable metric to quantify this complexity, offering a quantitative measure of the irregularity within the signal [10]. A higher entropy value signifies increased unpredictability and complexity in the signal, while a lower entropy value indicates a more structured and ordered signal. Leveraging entropy as a feature extraction tool facilitates the conversion of raw EEG signals into numerical representations, thereby enabling various applications such as emotion classification and cognitive state recognition [11]. For instance, Lu et al. proposed a novel EEG-based emotion recognition approach that utilizes dynamic entropy measures, demonstrating robust generality in emotion recognition tasks [12]. Additionally, in Patel’s research, a novel classification method was investigated, leveraging Tsallis entropy in conjunction with the K-Nearest Neighbors (KNN) classifier [10].
Examining entropy provides insights into the dynamic characteristics of the nervous system. Differential entropy (DE) offers a means to capture the temporal variations present in signals. Given that EEG signals exhibit dynamic changes over time, employing DE proves particularly advantageous, as it offers a more precise depiction of this evolving process. Moreover, DE surpasses standard entropy by accounting for signal variations over time, thereby enhancing feature discrimination and proving especially beneficial for complex emotional and cognitive state recognition tasks [13].
In our study, we propose a novel approach to train and classify EEG emotion states by integrating ResNet18 network architecture with DE. ResNet18 is adept at capturing both temporal and spatial characteristics of signals, with its cross-layer connections designed to capture spatiotemporal features across different levels. By incorporating DE alongside ResNet18, we aim to better model the temporal dynamics inherent in EEG signals, thereby enhancing the network’s ability to perceive temporal features within the signals. This synergistic approach allows for a more comprehensive understanding of signal dynamics, ultimately improving the network’s performance in analyzing EEG data for emotion classification and cognitive state recognition. Combining ResNet18 with DE for training and classifying EEG signals enables us to harness the strengths of deep learning and information theory synergistically.
Our approach merges the ResNet18 network and the DE measures to enhance the classification performance of human emotion states significantly. In essence, our contributions are as follows:
(1)
A framework for classifying multiple distinct human emotions from EEG signals is proposed, capable of identifying five distinct human emotions (happiness, sadness, fear, disgust, and neutrality).
(2)
A high-accuracy classification performance for the different emotional states of happiness, sadness, fear, disgust, and neutrality is achieved, by exploiting EEG differential entropy measures and ResNet18.
The remaining sections of our paper are structured as follows. The second part details the method of the DE combing with ResNet18 network. The third part presents the outcomes of data analysis and EEG-based emotion classification experiments. The fourth part provides a comparative analysis of our method with existing related research, showcasing experimental results. The final part concludes our findings and contributions.

2. Method

2.1. Proposed Framework for Classifying Multiple Distinct Human Emotions from EEG Signals

As depicted in Figure 1, the experimental subjects watched emotion-inducing videos corresponding to different emotions, allowing them to reach corresponding emotional states and collect corresponding 62 channel EEG signals in real time. The collected EEG signals undergo initial preprocessing steps, including downsampling, bandpass filtering, and rectifying artifacts. Subsequently, DE is extracted from the preprocessed signals. Subsequently, this extracted measure of DE serves as feature input to the ResNet18 network, where the classification process is trained and executed to obtain the final five types of human emotions (happiness, sadness, fear, disgust, and neutral).

2.2. Experimental Data

The SEED-V database delved into an examination of five distinct emotions—happiness, sadness, fear, disgust, and neutral. A curated selection of 15 film clips, with 3 clips representing each emotion, served as stimuli. As shown in Figure 2, five specific emotional stimulation videos correspond to five different emotions, namely fear, happiness, disgust, sadness, and neutral. Each of the 16 subjects, comprising 6 males and 10 females, partook in the experiment 3 times, with each session spanning 50 min. The ages of the 16 subjects were 21, 21, 23, 21, 20, 23, 23, 21, 23, 24, 24, 20, 22, 22, 19, and 19, respectively. The average age is 21.63, with a standard deviation of 1.63. Throughout the experiment, a meticulously designed protocol was adhered to, as follows: a 15 s initial pause preceded each stimulus, followed by 2 to 4 min of stimulus exposure, with intervals lasting either 15 or 30 s, depending on the emotion portrayed [14].
EEG signals were captured using a sophisticated 62-channel EEG device with a sampling frequency set at 1000 Hz, later downsampled to a 200 Hz rate to facilitate processing. Preprocessing of EEG data involved the application of a band-pass filter spanning 1 to 75 Hz, aimed at eliminating noise and artifacts. Subsequent to the experiments, participants engaged in a self-assessment segment, where they were prompted to rate the impact of the stimuli on a scale ranging from 0 to 5. A score of 5 indicated the most favorable response, while 0 denoted the least favorable.

2.3. Feature Extraction of Differential Entropy Measures

Given the non-Gaussian distribution of EEG signals, utilizing DE emerges as a potent technique for extracting features from such signals. This method aptly captures the complexity and uncertainty inherent in EEG signals during feature extraction. Lower entropy values signify a higher degree of signal orderliness and predictability, while higher entropy values indicate greater signal complexity and unpredictability.
As a method for feature extraction, entropy measurement offers a means to characterize the diversity and intricacy of information embedded within EEG signals, facilitating the extraction of pertinent features crucial for subsequent analysis and classification. Specifically, the employment of DE enables us to quantify the complexity and stochasticity inherent in EEG signals. Through the computation of DE, we gain insights into the temporal or spatial uncertainty present in the signals, alongside a quantitative assessment of the informational content they carry. Moreover, by examining the dynamic attributes of the signal, we can derive valuable features conducive to further analysis and classification endeavors.
As illustrated in Figure 3, the initial EEG signal collected comprises data from 62 channels, which underwent preprocessing. Initially, the EEG data were downsampled to 200 Hz and then subjected to a bandpass filter ranging from 0 to 75 Hz to eliminate any signal noise. We identified and rectified artifacts in the EEG signals stemming from electrode displacement and external interference.
After preprocessing the EEG signals from 62 channels, we utilized a 256-point Short-Time Fourier Transform technique along with a 1 s non-overlapping Hanning window to extract five distinct frequency bands from the following preprocessed EEG signals: delta (1–4 Hz), theta (4–8 Hz), alpha (8–14 Hz), beta (14–31 Hz), and gamma (31–50 Hz). From each frequency band, the DE feature is computed, culminating in a total of 310 DE features (62 channels × 5 frequency bands). The calculation formula for DE is outlined as follows:
D E = P ( x ) ln ( P ( x ) ) d x
We assume that the EEG signals follow a Gaussian distribution: x ~ N ( μ , σ 2 ) . Consequently, the calculation of the DE features can be simplified:
D E = 1 2 π σ 2 exp ( x μ ) 2 2 σ 2 log ( 1 2 π σ 2 exp ( x μ ) 2 2 σ 2 ) d x = 1 2 log ( 2 π e σ 2 )

2.4. Pattern Classifier Based on ResNet18

To establish an end-to-end model for deep feature extraction and classification of EEG data, we employed the ResNet18 network to process the extracted features of DE. Initially, the DE feature, sized at 310, undergoes reshaping to (62,5), integrating data from 62 channels scanned using ESI. As illustrated in Figure 4a, within the overall model architecture, the input size is structured as 5 features with 62 channels, wherein conv represents the convolutional layer, Max pool signifies the maximum pooling layer, AVG pool denotes the average pooling layer, and FC denotes the fully connected layer. The multi-layer convolutional structure amalgamates DE representations from diverse frequency bands and channels to learn multi-level features in EEG data. Concurrently, residual connections facilitate the direct propagation of shallow information into deeper networks, thereby retaining crucial EEG data. Subsequently, the deeply extracted DE features are employed for the classification of 5 distinct emotions through a fully connected layer. Table 1 shows the details of the emotion classification method based on the ResNet18. The model computes the output probability for each emotion type post multi-layer convolution. Throughout the network’s training process, cross-entropy loss serves as the optimization criterion, facilitating the refinement of weights and other parameters in the model.
The primary network utilized in this study comprises multiple analogous residual blocks linked sequentially, as illustrated in Figure 4b. Each residual block encompasses two paths: F(x), and x. The F(x) path, serving to adjust residuals, is termed the residual path, while the x path represents an identity mapping, referred to as the shortcut path. The residual path comprises two convolutional layers with a kernel size of 3, followed by a ReLU activation layer. The shortcut path can be categorized into two scenarios: one involves directly outputting the input x, while the other necessitates a 1 × 1 convolution for upsampling or downsampling, ensuring consistency in the shape of the shortcut output with that of the F(x) path output. Upon combining the residual and shortcut paths, they traverse through the ReLU activation layer and function as input for the subsequent residual block.
In this study, we employed the Adam (Adaptive Moment Estimation) optimizer to update the parameters of the ResNet network. In our experiment, we configured the parameters of the Adam optimizer as follows: β 1 = 0.9 , β 2 = 0.999 , learning rate = 1 × 10−4, weight decay = 1 × 10−2. The Adam optimizer [15] is an optimization algorithm that amalgamates momentum and adaptive learning rate techniques, facilitating the discovery of superior solutions within the parameter space. Its update rules involve computing the moving average of the first and second moments of the gradient, thereby correcting their deviations to enable adaptive parameter updates.

3. Results

In our experiment, we commenced by computing the confusion matrix using EEG data from 16 participants in the SEED-V dataset, thereby elucidating the precise classification scenario. Subsequently, our investigation encompassed the classification of five classes of human emotions (happiness, sadness, fear, disgust, and neutral) based on EEG signals from the aforementioned 16 individuals, followed by an in-depth analysis of the classification outcomes. To visually represent the classification of EEG emotion states, we undertook a two-dimensional mapping of the classification results and portrayed them via graphical imagery. Additionally, we computed and showcased ROC curves for the five categories of emotional states, thereby facilitating a comprehensive assessment of the efficacy of the proposed network classification model.
We utilized the proposed network to classify the five EEG emotional states of 16 participants in the dataset and conducted a comparative analysis of the results. Figure 5 illustrates the classification results for the emotional states of the 16 subjects, with accuracy values spanning from 87.64% to 99.21%. Notably, the 5th individual attained the lowest classification accuracy, while the remaining 15 individuals achieved accuracies exceeding 90%, with 11 individuals surpassing the 95% mark.
In addition to demonstrating recognition accuracy, we present the confusion matrix of the proposed method applied to the SEED-V dataset in Figure 6. This matrix provides insights into the recognition accuracy of the ResNet18 network across different emotional states. Within the confusion matrix, varying color depths indicate different values, transitioning from blue to red with increasing magnitude. The deeper the red color, the higher the corresponding value.
As illustrated in Figure 6, the ResNet18 network achieves recognition accuracies of 86.16%, 90.19%, 94.61%, 96.78%, and 96.13% for the following five emotional states: fear, sadness, neutral, happiness, and disgust, respectively. It is evident that the network exhibits the highest accuracy in recognizing a happy emotional state, while its performance is comparatively lower for fearful emotional states. The accuracy rate for detecting fear emotions is 86.16%, with misclassifications as follows: 3.46% as sad, 1.73% as neutral, 1.8% as happy, and 6.85% as doubt. Notably, fear often tends to be misclassified as doubt due to the similarity in their EEG signals, posing a challenge for accurate differentiation. Moreover, the graphical representation reveals that 2.5% and 0.72% of happy emotional states are misclassified as neutral and disgust emotional states, respectively.
To further underscore the efficacy of the ResNet18 network in EEG emotion classification, we employed a two-dimensional visualization of the classification outcomes. The 512 dimensional features output by the average pooling layer of ResNet18 are mapped to a two-dimensional image using the t-SNE algorithm [14]. Different colors represent different emotions, and different emotions are completely separated without overlapping areas, indicating excellent classification results for the five emotions. Illustrated in Figure 7a–d, we depict the classification results for subjects 2, 5, 6, and 9 utilizing the ResNet18 network across five distinct EEG emotional states.
Initially, emotions in EEG recordings appear intertwined and challenging to discern. However, post-training and classification using the ResNet18 network, Figure 6 reveals a distinct clustering of emotions of the same type. This visual representation demonstrates that EEG signals corresponding to different emotions are more visually distinguishable, thereby corroborating the effectiveness of the ResNet18 network model in EEG emotion classification.
To assess the ResNet18 network’s capability in deciphering EEG emotion states, we conducted an analysis utilizing the ROC curve and AUC (Area Under the ROC Curve) for Subject 4. Each of the five emotion categories was treated as a positive example, while all other categories were deemed negative examples. By calculating the True Positive Rate (TPR) and False Positive Rate (FPR) for each category and plotting the curve, we obtained a comprehensive evaluation model, showcasing its performance across diverse categories.
The multi-class ROC curve provides a holistic overview of performance indicators across all categories through the following two perspectives: micro average, and macro average. The micro average ROC curve computes the overall TPR and FPR by aggregating the number of true and false positive cases across all categories, thus treating each category with equal importance and consolidating their performance into a singular metric. Conversely, the macro average ROC curve evaluates the model’s performance by calculating the TPR and FPR for each category separately and then averaging the values, assigning equal weight to each category.
The significance of the multi-class ROC curve lies in its ability to offer a comprehensive evaluation perspective on the performance of multi-class classification models. By scrutinizing the TPR and FPR of different categories, one can gauge the model’s recognition ability across each category. Both micro average and macro average ROC curves facilitate the comparison of overall performance among different models in multi-class classification tasks.
As illustrated in Figure 8, we chose Subject 4 as the focal point of our experimentation. Across five distinct emotions—disgust, fear, sadness, neutrality, and happiness—the corresponding AUC values were 0.9871, 0.9886, 0.8976, 0.9330, and 0.9718, respectively. A higher AUC value indicates stronger discrimination capability within the network for different emotions.

4. Discussion

In our research, we integrated the ResNet18 network with DE to classify EEG-based human emotion states. DE serves as a valuable tool for quantifying the uncertainty and information content within EEG signals, aiding in the discernment of distinct signal types. ResNet18, renowned for its deep convolutional architecture, boasts formidable feature extraction capabilities. When EEG signals are fed into the ResNet18 network, it adeptly extracts high-level abstract feature representations, enabling the acquisition of intricate insights into the brain activity’s spatiotemporal dynamics. Our findings from emotion recognition using EEG signals underscore the synergistic benefits of this combination, as it enhances both the accuracy and robustness of signal classification through sophisticated EEG signal-processing techniques.
To demonstrate the superior performance of the ResNet18 network in conjunction with the DE method for distinguishing emotion states, we conducted a comparative analysis with recent studies utilizing the same SEED-V dataset. Our evaluation encompassed studies by Zhu, Zhou, Jin, Aslan, Kamble, and Conejo.
In Zhu’s research, an emotion recognition method termed Joint Distributed Instances Represent Transfer (JD-IRT) was employed. This method integrates two core components: Joint Distribution Deep Adaptation (JDDA), and Instance-Representation Transfer (I-RT). Zhou’s study utilized a progressive graph convolution network (PGCN) to learn discriminative EEG features. Jin proposed a Graph Adaptive Semi-supervised Discriminative Subspace Learning (GASDSL) model for EEG-based emotion recognition. Aslan’s study utilized Discrete Wavelet Transform (DWT) and Statistical Moments (SM) methods to extract features from EEG signals, and applied Convolutional Neural Networks (CNN) for classification. Liu’s study employed the Deep Canonical Correlation Analysis (DCCA) method for emotion recognition from EEG signals. Conejo’s study proposed two dimension reduction algorithms based on Local Binary Pattern (LBP) derived from machine learning models, which included electroencephalogram channel selection and conflict learning.
Comparing the results of EEG-based emotion classification within the SEED-V dataset, as depicted in Table 2, the accuracies of EEG emotion classification for Zhu’s, Zhou’s, Jin’s, Aslan’s, Liu’s, and Conejo’s studies were 65.91%, 71.40%, 82.03%, 82.06%, 85.30%, and 93.85%, respectively. In contrast, our proposed method, which combines the ResNet18 network with DE, achieves an accuracy of 95.61% in classifying EEG-based emotion states, significantly enhancing the resolution performance. Through this comparison of classification accuracy, it becomes evident that our method retains more pertinent information while eliminating redundant data, thus leading to a more precise and efficient EEG emotion classification outcome.
However, our research still faces several limitations. Schubling et al. [20] segmented the beta band into low and high beta bands and explored their relationship with emotional processing. Wijk et al. [21] suggested that different frequency ranges within the beta band significantly influence motor control. Borzelli et al. [22] identified increased synchronization among motor units involved in stiffness modulation, supported by cross-muscle coherence analysis, revealing peaks in the β-band, typically linked with cortical activity. These studies are about the impact of frequency band separation on brain activity, and we have not yet considered this factor in our experiments. We plan to incorporate frequency band separation considerations in future experiments. Additionally, Duclos et al. [23] proposed that posture affects emotional experiences similarly to facial expressions, while Carney et al. [24] suggested that body posture can induce physiological and behavioral changes, eliciting various emotions. Buisine et al. [25] explored emotional expression detection through different body positions, emphasizing the close connection between the motor system and emotions. Both facial expressions and body posture significantly influence emotional recognition. Our study utilized the SEED-V open-source dataset without considering facial expressions and body posture. Future experiments will investigate the interplay between emotion and motor control and incorporate these factors. Furthermore, Leonardi et al. [26] proposed brain oscillatory rhythms as a potential measure for stroke recovery, indicating their consistent patterns as a promising tool for predicting rehabilitation outcomes. However, our study did not include patients with neurological conditions, limiting its scope. Future research will involve neurological patients to explore EEG-based emotion recognition.

5. Conclusions

In our study, we propose a novel approach that combines the ResNet18 network with the DE method to classify human emotion states effectively from EEG signals. The DE method, serving as an information measurement technique, offers the ability to precisely quantify the uncertainty and information content inherent in EEG signals. Utilized as input to the network, DE enriches the model with a more comprehensive dataset, thereby enhancing its capacity to understand and interpret EEG signals. Leveraging the ResNet18 network, renowned for its deep convolutional architecture and multi-level feature extraction capabilities, further amplifies our approach’s analytical prowess.
By integrating DE with the ResNet18 network, we are able to unveil the spatiotemporal dynamics within EEG signals, extracting intricate feature representations across various network levels. This amalgamation enables us to more effectively capture the nuanced complexity and abstraction present in EEG data, thereby elevating the accuracy and robustness of our classification system. Our experimental findings unequivocally demonstrate the superior performance achieved with the ResNet18 network in tandem with DE, showcasing its remarkable classification proficiency in human emotional states.
Moreover, our method underwent rigorous comparison against existing approaches using identical datasets, affirming its superior efficacy in emotion recognition from EEG signals. This substantiates the validity and reliability of our proposed methodology. Ultimately, our approach not only represents a highly effective means of classifying EEG-based emotion states, but also holds promising implications for advancing research within the broader domain of EEG pattern analysis.

Author Contributions

Conceptualization, L.Y.; methodology, L.Y. and Y.L.; software, L.Y., Y.Q. and C.H.; validation, L.Y., Y.Q. and C.H.; formal analysis, L.Y.; writing—original draft preparation, L.Y. and Y.L.; supervision, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China, under grants No. 62276076 and No. 62176102, by the Guangdong Basic and Applied Basic Research Foundation (No. 2023A1515140109), the Professorial and Doctoral Scientific Research Foundation of Huizhou University under Grant 2020JB058, the Planning Project of Enhanced Independent Innovation Ability of Huizhou University under Grant HZU202018.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: [https://bcmi.sjtu.edu.cn/~seed/seed-v. 10 June 2024].

Acknowledgments

The authors gratefully acknowledge the use of the SJTU Emotion EEG Dataset (SEED-V) provided by Wei-Long Zheng, Shanghai Jiao Tong University, 800 Dong Chuan Road, Shanghai 200240, China.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zou, Y.; Donner, R.V.; Marwan, N.; Donges, J.F.; Kurths, J. Complex network approaches to nonlinear time series analysis. Phys. Rep. 2018, 787, 1–97. [Google Scholar] [CrossRef]
  2. Yao, L.; Wang, M.; Lu, Y.; Li, H.; Zhang, X. EEG-Based Emotion Recognition by Exploiting Fused Network Entropy Measures of Complex Networks across Subjects. Entropy 2021, 23, 984. [Google Scholar] [CrossRef] [PubMed]
  3. Aslan, M.; Baykara, M.; Alakuş, T.B. Analysis of brain areas in emotion recognition from EEG signals with deep learning methods. Multimed. Tools Appl. 2024, 83, 32423–32452. [Google Scholar] [CrossRef]
  4. Pan, J.; Liang, R.; He, Z.; Li, J.; Liang, Y.; Zhou, X.; He, Y.; Li, Y. ST-SCGNN: A Spatio-Temporal Self-Constructing Graph Neural Network for Cross-Subject EEG-Based Emotion Recognition and Consciousness Detection. IEEE J. Biomed. HealthInform. 2024, 28, 777–788. [Google Scholar] [CrossRef] [PubMed]
  5. Zhong, X.; Gu, Y.; Luo, Y.; Zeng, X.; Liu, G. Bi-hemisphere asymmetric attention network: Recognizing emotion from EEG signals based on the transformer. Appl. Intell. 2023, 53, 15278–15294. [Google Scholar] [CrossRef]
  6. Zhong, M.-Y.; Yang, Q.-Y.; Liu, Y.; Zhen, B.-Y.; Zhao, F.-D.; Xie, B.-B. EEG emotion recognition based on TQWT-features and hybrid convolutional recurrent neural network. Biomed. Signal Process. Control. 2023, 79, 104211. [Google Scholar] [CrossRef]
  7. Alotaibi, F.M.; Fawad. An AI-Inspired Spatio-Temporal Neural Network for EEG-Based Emotional Status. Sensors 2023, 23, 498. [Google Scholar] [CrossRef] [PubMed]
  8. Chen, Y.; Xu, X.; Bian, X.; Qin, X. EEG emotion recognition based on Ordinary Differential Equation Graph Convolutional Networks and Dynamic Time Wrapping. Appl. Soft Comput. 2024, 152, 111181. [Google Scholar] [CrossRef]
  9. Xu, F.; Pan, D.; Zheng, H.; Ouyang, Y.; Jia, Z.; Zeng, H. EESCN: A novel spiking neural network method for EEG-based emotion recognition. Comput. Methods Programs Biomed. 2024, 243, 107927. [Google Scholar] [CrossRef]
  10. Patel, P.; Balasubramanian, S.; Annavarapu, R.N. Cross subject emotion identification from multichannel EEG sub-bands using Tsallis entropy feature and KNN classifier. Brain Inf. 2024, 11, 7. [Google Scholar] [CrossRef]
  11. Wang, T.; Huang, X.; Xiao, Z.; Cai, W.; Tai, Y. EEG emotion recognition based on differential entropy feature matrix through 2D-CNN-LSTM network. EURASIP J. Adv. Signal Process. 2024, 2024, 49. [Google Scholar] [CrossRef]
  12. Lu, Y.; Wang, M.; Wu, W.; Han, Y.; Zhang, Q.; Chen, S. Dynamic entropy-based pattern learning to identify emotions from EEG signals across individuals. Measurement 2020, 150, 107003. [Google Scholar] [CrossRef]
  13. Zheng, W.-L.; Zhu, J.-Y.; Lu, B.-L. Identifying Stable Patterns over Time for Emotion Recognition from EEG. IEEE Trans. Affect. Comput. 2019, 10, 417–429. [Google Scholar] [CrossRef]
  14. Liu, W.; Qiu, J.-L.; Zheng, W.-L.; Lu, B.-L. Comparing Recognition Performance and Robustness of Multimodal Deep Learning Models for Multimodal Emotion Recognition. IEEE Trans. Cogn. Dev. Syst. 2022, 14, 715–729. [Google Scholar] [CrossRef]
  15. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. Comput. Sci. 2014. [Google Scholar] [CrossRef]
  16. Zhu, L.; Yu, F.; Huang, A.; Ying, N.; Zhang, J. Instance-representation transfer method based on joint distribution and deep adaptation for EEG emotion recognition. Med Biol. Eng. Comput. 2024, 62, 479–493. [Google Scholar] [CrossRef] [PubMed]
  17. Zhou, Y.; Li, F.; Li, Y.; Ji, Y.; Shi, G.; Zheng, W.; Zhang, L.; Chen, Y.; Cheng, R. Progressive graph convolution network for EEG emotion recognition. Neurocomputing 2023, 544, 126262. [Google Scholar] [CrossRef]
  18. Jin, F.; Peng, Y.; Qin, F.; Li, J.; Kong, W. Graph adaptive semi-supervised discriminative subspace learning for EEG emotion recognition. J. King Saud Univ. Comput. Inf. Sci. 2023, 35, 101648. [Google Scholar] [CrossRef]
  19. Almanza-Conejo, O.; Avina-Cervantes, J.G.; Garcia-Perez, A.; Ibarra-Manzano, M.A. A channel selection method to find the role of the amygdala in emotion recognition avoiding conflict learning in EEG signals. Eng. Appl. Artif. Intell. 2023, 126, 106971. [Google Scholar] [CrossRef]
  20. Schubring, D.; Schupp, H.T. Emotion and Brain Oscillations: High Arousal is Associated with Decreases in Alpha- and Lower Beta-Band Power. Cereb. Cortex 2020, 31, 1597–1608. [Google Scholar] [CrossRef]
  21. Van Wijk, B.C.; Neumann, W.J.; Schneider, G.H.; Sander, T.H.; Litvak, V.; Kühn, A.A. Low-beta cortico-pallidal coherence decreases during movement and correlates with overall reaction time. NeuroImage 2017, 159, 1–8. [Google Scholar] [CrossRef] [PubMed]
  22. Borzelli, D.; Vieira, T.M.; Botter, A.; Gazzoni, M.; Lacquaniti, F.; d’Avella, A. Synaptic inputs to motor neurons underlying muscle coactivation for functionally different tasks have different spectral characteristics. J. Neurophysiol. 2024, 131, 1126–1142. [Google Scholar] [CrossRef] [PubMed]
  23. Duclos, S.E.; Laird, J.D.; Schneider, E.; Sexter, M.; Stern, L.; Van Lighten, O. Emotion-Specific Effects of Facial Expressions and Postures on Emotional Experience. J. Pers. Soc. Psychol. 1989, 57, 100–108. [Google Scholar] [CrossRef]
  24. Carney, D.R.; Cuddy, A.J.; Yap, A.J. Power posing: Brief nonverbal displays affect neuroendocrine levels and risk tolerance. Psychol. Sci. 2010, 21, 1363–1368. [Google Scholar] [CrossRef] [PubMed]
  25. Buisine, S.; Courgeon, M.; Charles, A.; Clavel, C.; Martin, J.-C.; Tan, N.; Grynszpan, O. The Role of Body Postures in the Recognition of Emotions in Contextually Rich Scenarios. Int. J. Hum.–Comput. Interact. 2013, 30, 52–62. [Google Scholar] [CrossRef]
  26. Leonardi, G.; Ciurleo, R.; Cucinotta, F.; Fonti, B.; Borzelli, D.; Costa, L.; Tisano, A.; Portaro, S.; Alito, A. The role of brain oscillations in post-stroke motor recovery: An overview. Front. Syst. Neurosci. 2022, 16, 947421. [Google Scholar] [CrossRef]
Figure 1. Framework for EEG-based emotion classification using ResNet18 combined with the DE method.
Figure 1. Framework for EEG-based emotion classification using ResNet18 combined with the DE method.
Applsci 14 06175 g001
Figure 2. Five emotion-related stimulation videos: fear, happy, disgust, sad, and neutral.
Figure 2. Five emotion-related stimulation videos: fear, happy, disgust, sad, and neutral.
Applsci 14 06175 g002
Figure 3. Framework of extracting DE features.
Figure 3. Framework of extracting DE features.
Applsci 14 06175 g003
Figure 4. ResNet18 network combined with DE for emotion recognition from EEG signals. (a) ResNet 18 network; (b) diagram of residual block.
Figure 4. ResNet18 network combined with DE for emotion recognition from EEG signals. (a) ResNet 18 network; (b) diagram of residual block.
Applsci 14 06175 g004
Figure 5. Classification results for the emotional states of the 16 subjects.
Figure 5. Classification results for the emotional states of the 16 subjects.
Applsci 14 06175 g005
Figure 6. Confusion matrix of different emotion states on the SEED-V dataset.
Figure 6. Confusion matrix of different emotion states on the SEED-V dataset.
Applsci 14 06175 g006
Figure 7. Feature distribution visualization by the t-SNE algorithm. The different colors stand for different emotions. Two-dimensional display of classification results from four subjects. (a) Subject 2, (b) Subject 5, (c) Subject 6, (d) Subject 9.
Figure 7. Feature distribution visualization by the t-SNE algorithm. The different colors stand for different emotions. Two-dimensional display of classification results from four subjects. (a) Subject 2, (b) Subject 5, (c) Subject 6, (d) Subject 9.
Applsci 14 06175 g007
Figure 8. ROC curves of each of the five emotions in Subject 4 compared to other emotions.
Figure 8. ROC curves of each of the five emotions in Subject 4 compared to other emotions.
Applsci 14 06175 g008
Table 1. The details of structure of ResNet18-based emotion recognition method.
Table 1. The details of structure of ResNet18-based emotion recognition method.
Module LevelModule/Layer TypeConfigurations
0Input featuresChannel = 62
Frequency bands = 5
Feature size = 62 × 5
1Input conv layerIn_channel = 62
Out_channel = 64
Kernel = 7
2MaxpoolKernel = 3
3Conv layers #1In_channel = 64
Out_channel = 64
Kernel = 3
4Conv layers #2In_channel = 64
Out_channel = 128
Kernel = 3
5Conv layers #3In_channel = 128
Out_channel = 256
Kernel = 3
6Conv layers #4In_channelg = 256
Out_channel = 512
Kernel = 3
7Avgpooloutput_size = 1
8FCIn_feature = 512
Out_feature = 5
Table 2. The results of classification accuracy from Zhu’s, Zhou’s, Jin’s, Aslan’s, Liu’s, Conejo’s, and our studies.
Table 2. The results of classification accuracy from Zhu’s, Zhou’s, Jin’s, Aslan’s, Liu’s, Conejo’s, and our studies.
TitleDatasetMethodologyMean AccuracySTDYear
Zhu’s study [16]SEED-VJD-IRT65.91%12.31%2023
Zhou’s study [17]SEED-VPGCN71.40%9.43%2023
Jin’s study [18]SEED-VGASDSL82.03%-2023
Aslan’s study [3]SEED-VDWT and SM82.06%-2023
Liu’s study [14]SEED-VDCCA85.30%5.6%2022
Conejo’s study [19]SEED-VLBP93.85%5.64%2023
Our workSEED-VResNet18-DE95.61%4.02%2024
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, L.; Lu, Y.; Qian, Y.; He, C.; Wang, M. High-Accuracy Classification of Multiple Distinct Human Emotions Using EEG Differential Entropy Features and ResNet18. Appl. Sci. 2024, 14, 6175. https://doi.org/10.3390/app14146175

AMA Style

Yao L, Lu Y, Qian Y, He C, Wang M. High-Accuracy Classification of Multiple Distinct Human Emotions Using EEG Differential Entropy Features and ResNet18. Applied Sciences. 2024; 14(14):6175. https://doi.org/10.3390/app14146175

Chicago/Turabian Style

Yao, Longxin, Yun Lu, Yukun Qian, Changjun He, and Mingjiang Wang. 2024. "High-Accuracy Classification of Multiple Distinct Human Emotions Using EEG Differential Entropy Features and ResNet18" Applied Sciences 14, no. 14: 6175. https://doi.org/10.3390/app14146175

APA Style

Yao, L., Lu, Y., Qian, Y., He, C., & Wang, M. (2024). High-Accuracy Classification of Multiple Distinct Human Emotions Using EEG Differential Entropy Features and ResNet18. Applied Sciences, 14(14), 6175. https://doi.org/10.3390/app14146175

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop