Next Article in Journal
Associations of Depression Score with Dialkyl Phosphate Metabolites in Urine: A Cross-Sectional Study
Next Article in Special Issue
Complex Spiking Neural Network Evaluated by Injury Resistance Under Stochastic Attacks
Previous Article in Journal
Age- and Sex-Based Developmental Biomarkers in Eye Movements
Previous Article in Special Issue
Sensitivity of Spiking Neural Networks Due to Input Perturbation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Emotion Recognition Model of EEG Signals Based on Double Attention Mechanism

1
Xi’an Key Laboratory of High Pricision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China
2
Affiliation College of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
3
School of Mathematics and Statistics, Zhengzhou University, Zhengzhou 710003, China
4
West China Institute of Children’s Brain and Cognition, Chongqing University of Education, Chongqing 400065, China
*
Authors to whom correspondence should be addressed.
Brain Sci. 2024, 14(12), 1289; https://doi.org/10.3390/brainsci14121289
Submission received: 13 November 2024 / Revised: 14 December 2024 / Accepted: 19 December 2024 / Published: 21 December 2024

Abstract

:
Background: Emotions play a crucial role in people’s lives, profoundly affecting their cognition, decision-making, and interpersonal communication. Emotion recognition based on brain signals has become a significant challenge in the fields of affective computing and human-computer interaction. Methods: Addressing the issue of inaccurate feature extraction and low accuracy of existing deep learning models in emotion recognition, this paper proposes a multi-channel automatic classification model for emotion EEG signals named DACB, which is based on dual attention mechanisms, convolutional neural networks, and bidirectional long short-term memory networks. DACB extracts features in both temporal and spatial dimensions, incorporating not only convolutional neural networks but also SE attention mechanism modules for learning the importance of different channel features, thereby enhancing the network’s performance. DACB also introduces dot product attention mechanisms to learn the importance of spatial and temporal features, effectively improving the model’s accuracy. Results: The accuracy of this method in single-shot validation tests on the SEED-IV and DREAMER (Valence-Arousal-Dominance three-classification) datasets is 99.96% and 87.52%, 90.06%, and 89.05%, respectively. In 10-fold cross-validation tests, the accuracy is 99.73% and 84.26%, 85.40%, and 85.02%, outperforming other models. Conclusions: This demonstrates that the DACB model achieves high accuracy in emotion classification tasks, demonstrating outstanding performance and generalization ability and providing new directions for future research in EEG signal recognition.

1. Introduction

1.1. Motivation

Emotions are fundamental to human behavior. Positive emotions help improve daily work efficiency, while negative emotions can lead to erroneous decisions. Emotions play a crucial role in people’s lives, profoundly impacting cognition, decision-making, and interpersonal interactions. Positive emotion (PE) affects cognitive processes, including executive function (i.e., inhibition, working memory, cognitive flexibility) [1]. For example, positive emotions often increase motivation and focus, making people more willing to engage in tasks and better able to focus. Conversely, negative emotions such as anxiety or depression may be distracting and reduce productivity. At the same time, moderate positive emotions can improve working memory performance and may promote more creative and flexible thinking.
In recent years, the application of artificial intelligence technology has indeed become increasingly widespread across various fields, and emotion recognition technology has emerged as a key technology in areas such as human-computer interaction, disease assessment, emotion detection, and healthcare. Specifically, it can be used to assist mental health professionals to more accurately assess and diagnose mood disorders such as depression and anxiety. In terms of user experience and human-computer interaction, product design can be optimized by analyzing users’ emotional responses. In the field of education, students’ emotional fluctuations in the learning process can be understood to help teachers adjust teaching methods and improve learning efficiency.
Over the past decades, there are two main approaches used to analyze and classify emotions: the locationist approach and the constructive approach [2]. The first approach assumes that each emotion has a specific brain structure and pattern. Studies of the constructionist approach use affective valence models or motivational direction models to deal with emotions. Affective valence means dividing emotions into positive and negative. Positive emotions include happiness, joy, excitement, and interest, while negative emotions include sadness, anger, disappointment, and anxiety. On the contrary, the motivational direction category has different characteristics. On the other hand, the constructionist model consists of valence-arousal-dominance (VAD) and activity-temperature-weight (VTT) models, where the VAD model is commonly used. Valence (positiveness-negativeness/pleasure-displeasure) refers to the emotional tendency conveyed by the vocabulary, with positive values indicating positive emotions, negative values indicating negative emotions, and zero indicating neutral emotions. Arousal (active-passive) represents the emotional intensity elicited by the vocabulary, with high values indicating high excitement or arousal, and low values indicating low excitement or calmness. Dominance (dominant-submissive) refers to the authority and control conveyed by vocabulary, where high values indicate high authority or control, while low values indicate low authority or control [3]. It can be considered that there are three primary colors in color theory, combining different numerical values to create a color. In VAD, valence, arousal, and dominance are three dimensions that assign a value to represent an emotion.
Emotion is a neurophysiological state that includes thoughts, feelings, excitement, pleasure, and personal behavioral responses [4]. Generally, human emotions can be detected in two ways: non-physiological signals and physiological signals [5]. Non-physiological signals include facial expressions, voice, language, gestures, etc. [6,7,8]. However, using non-physiological signals in emotion detection and classification may suffer from inaccurate recognition results because humans can produce false verbal intonations and facial expressions [9,10], and these technologies themselves are also fragile. Compared to non-physiological signals, physiological signals are harder to hide or disguise, making them more reliable. Physiological signals include electroencephalogram (EEG), electrooculography (EOG), electrocardiogram (ECG), electromyogram (EMG), galvanic skin response (GSR), skin temperature (ST), blood volume pulse (BVP), etc. EOG, ECG, EMG, and other physiological signals are usually indirect reactions caused by emotions, with lower recognition accuracy [11]. EEG, on the other hand, is generated by the central nervous system and directly reflects brain activity, making it possible to use EEG signals to reflect the working state of the human brain. Most researchers believe that EEG signals are more objective and accurate for studying human emotion recognition [12,13]. It has been confirmed by various psychophysiological studies that there are correlations between EEG signals and human emotions. Of course, it should also be noted that some signals are better for arousal (e.g., gsr), while others can be used for valence (EEG).

1.2. Related Work

Research on emotion recognition based on EEG signals mainly includes two parts: feature extraction and emotion classification models. EEG signals, as a kind of physiological signal with a low signal-to-noise ratio that is non-stationary and asymmetric, pose significant challenges for robust feature extraction and selection of emotional EEG signals [14]. Traditional methods of emotion recognition based on EEG signals mainly extract features in the time domain or frequency domain. For instance, Zhu et al. [15] used differential entropy (DE) as features and a support vector machine (SVM) for classification. Wagh et al. [16] utilized discrete wavelet transform (DWT) to decompose EEG signals in different frequency bands and employed SVM, k-nearest neighbors (KNN), and decision trees to classify emotional states on the SEED dataset. Liu et al. [17] adopted a method combining multi-channel EEG in the time domain with text features to recognize different human emotions, incorporating six time-domain statistical features into a single feature vector for emotion classification. Finally, the performance was evaluated on the DEAP dataset for physiological signal emotion analysis. Gannounid et al. [18]. proposed ZTWBES algorithm to identify time points, to select electrodes that successfully identify time points in advance, and to determine the relevant electrodes for each frequency band for each emotional state. Changes in brain activity compared with neutral emotional states were extracted as features.
However, the limitations of traditional electroencephalogram (EEG) analysis methods for emotion recognition have become increasingly apparent. These traditional methods have limited generalization ability and require extensive prior knowledge to obtain deep features of EEG signals. This results in significant time and energy consumption during the manual feature extraction process, greatly affecting the accuracy of emotion recognition [19].
In recent years, with the continuous development of artificial intelligence, deep learning has gradually gained popularity in various fields, attracting widespread attention from computer vision, speech recognition, natural language processing, biomedical signal processing, and other fields. Many researchers have successfully introduced deep learning methods into the field of EEG-based emotion recognition. Tripathi et al. [20] extracted features such as skewness and kurtosis from EEG signals in the DEAP dataset, which represents the benchmark for emotion classification research. Then, they used deep neural networks (DNNs) and convolutional neural networks (CNNs) to carry out binary and ternary classification for valence and arousal features, and the accuracy of this model has been greatly improved compared with previous studies.
Kim et al. [21] proposed a long short-term memory network considering the changes in emotions over time and applied an attention mechanism to allocate weights to the emotional states that appear at specific moments based on the peak–end rule in psychology. They used 32-channel EEG data from the DEAP database. Two-level (low and high) and three-level (low, middle, and high) classification experiments were performed on the valence and arousal emotion models. Wang et al. [22] proposed the concept of an electrode-frequency distribution map (EFDM) based on the short-time Fourier transform (STFT). They employed a deep convolutional neural network with residual blocks for automatic feature extraction and classification on the DEAP dataset. Cui et al. [23] designed an end-to-end regional asymmetric convolutional neural network (RACNN) for emotion recognition considering spatial information from neighboring and symmetric channels. Li et al. [24] proposed an emotion recognition method based on EEG combining multi-scale residual networks (MSRN) and meta-transfer learning (MTL) strategies and evaluated its performance using the DEAP and SEED datasets. Rajpoot et al. [25] proposed a subject-independent emotion recognition method, which uses an unsupervised Long Short-Term Memory (LSTM) with channel-wise attention autoencoder.

1.3. Contribution

Although many studies in the literature have proposed various deep learning methods and models for EEG-based emotion recognition and classification, there are still many unresolved issues, such as the following: (1) The original EEG signals already contain a lot of information, but due to the nonlinear nature of these signals, it is difficult to directly extract valuable features. (2) How to design the structure or framework of the EEG emotion decoding model that can fully explore high-level abstract features while maintaining high accuracy and robustness. (3) EEG data are composed of multiple channels; how to find the best channels for feature extraction without manual selection. To tackle the aforementioned challenges, this study introduces a hybrid deep learning model named DACB, which integrates a dual attention mechanism, a convolutional neural network, and a bidirectional long short-term memory network for effective emotion classification.
The primary contributions of this paper are outlined as follows: (1) The proposed model in this study utilizes only the original EEG signals, eliminating the need for any time-frequency domain transformation for feature extraction. This simplifies its portability and integration into brain-computer interface devices. (2) While prior studies predominantly emphasized spatial feature extraction, temporal sequence features play a pivotal role in EEG signals. Hence, this paper introduces a bidirectional long short-term memory network to enhance the model’s capability in capturing temporal features. (3) A one-dimensional Squeeze-and-Excitation (SE) network attention mechanism alongside a dot product attention mechanism was introduced. The former enables the model to learn the significance of various channel features, while the latter teaches it the importance of spatial and temporal features. (4) By applying the model to two-dimensional datasets, this paper evaluates its performance across both the discrete emotion model and the three dimensions of the multidimensional emotion space model (valence, arousal, and dominance). This comprehensive testing further validates the model’s generalization capabilities by using EEG signals.
The rest of the study is structured as follows: Section 2 details the methods presented in this paper and provides a description of the dataset and experimental setup. Section 3 describes the basic process of the experiment and compares the experimental results with other methods. Finally, Section 4 summarizes the main conclusions of this study.

2. Materials and Methods

2.1. The Structure of the DACB Model

Figure 1 illustrates the basic structure of the DACB model. As shown in Figure 1, the DACB model consists of four components: EEG input for emotion recognition, feature extraction, feature selection, and classification. Firstly, the raw one-dimensional EEG signals are normalized and then simultaneously fed into the feature extraction module. The feature extraction has two parallel channels, including spatial features and temporal features.
As shown in Figure 1, the DACB structure includes a spatial feature extraction module, which consists of a convolutional layer with a ReLU activation function, a max pooling layer, an SE attention mechanism module, and a flatten layer. The convolutional layer primarily captures the interaction information between spatial physiological signals and feature channels. The ReLU activation function is simple and efficient, effectively alleviating the problems of gradient disappearance and explosion, and has been widely used in recent years. The SE attention mechanism learns the importance of each channel in the feature mapping from the channel dimension data using a two-layer neural network and then multiplies it by the original channel data. The temporal feature extraction module is composed of two layers: Bidirectional Long Short Term Memory network (Bi-LSTM) and Flatten. By considering both forward and backward time series information simultaneously along the temporal dimension, Bi-LSTM can achieve more comprehensive and accurate predictions.
The EEG signals are processed through a spatial feature extraction module, where the data’s spatial features are obtained using CNN. Then, the SE attention mechanism is applied to further derive the weight distribution across various channels. Meanwhile, the Bi-LSTM extracts temporal features from the EEG time series signals. A flatten layer concatenates the features of two different dimensions, and, finally, a fully connected layer integrates these features to output classification results. This dual-channel model allows for a more in-depth analysis of emotional features from both temporal and spatial perspectives, thereby enhancing the accuracy of the prediction results. The DACB architecture parameters are shown in Table 1.

2.2. Convolutional Neural Network

In recent years, Convolutional Neural Networks (CNNs) have achieved significant performance in various challenging tasks such as computer vision, speech recognition, and natural language processing (NLP). For instance, CNNs can outperform humans in some difficult visual tasks [26]. Moreover, they are not limited to specific types of input data but excel in various fields, including but not limited to signal processing for brain-computer interfaces (BCI) and biomedical signal processing [27,28].
CNNs have emerged as the most prevalent deep-learning-based network model. Contrary to traditional machine learning algorithms, CNNs eliminate the need for manual feature design. Instead, they leverages local receptive fields created by convolutional kernels to automatically extract abstract features from raw data for classification, thereby preventing the loss of valuable information. Unlike the traditional framework that separates feature learning and classification into distinct steps, CNNs are capable of learning features and performing classification simultaneously through multiple layers of neural networks. A standard CNN comprises five layers: an input layer, a convolutional layer, an activation layer, a pooling layer, and a fully connected layer. Typically, a comprehensive CNN architecture consists of one or more such blocks, culminating in a classifier layer. Given a 2D or 3D array as input data X and a CNN with n computational blocks, the output of the convolutional layer in the L-th block can be computed using the following formula.
f L ( x ) = i = 1 L ( X i w i + b i )
Here, Xi, wi, and bi represent the input, weight, and bias, respectively, with ∗ denoting the convolution operation. Once the activation and pooling layers have been computed and processed accordingly, the output of the convolutional layer serves as the input for the subsequent block. The output of the final convolutional layer is then fed consecutively into the following fully connected layer and classifier layer. The entire network undergoes training through the backpropagation of a supervised loss function.

2.3. Bidirectional Long Short-Term Memory Network

There exists a specific temporal dependency among the sequential signals within the emotional electroencephalogram. To better capture the temporal dependency features of the time series, this paper incorporates Bidirectional Long Short-Term Memory (Bi-LSTM). Bi-LSTM not only processes the contextual information but also incorporates the content of future contexts, thereby enhancing the model’s accuracy. Bi-LSTM consists of two LSTMS. The LSTM network comprises “gates” with distinct functions and operations. Figure 2 illustrates the LSTM cell unit and its internal structure.
The forget gate determines which information should be discarded or retained, as expressed in the following formula.
f t = σ ( w f × [ h t 1 , x t ] + b f )
The aim of updating the input gate and output gate is to ascertain whether to substitute the memory cell with a candidate value and produce the activation segment of the current time step. This process can be further represented as follows:
i t = σ ( w i × [ h t 1 , x t ] + b i )
o t = σ ( w o × [ h t 1 , x t ] + b o )
The behavior of the LSTM unit is determined by the following formula.
C ~ t = tanh ( w c × [ h t 1 , x t ] + b c )
C t = f t × C t 1 + i t × C ~ t
h t = o t × tanh ( C t )
In the formula, σ denotes the sigmoid activation function, while tanh represents the hyperbolic tangent activation function. wf, wi, and wo stand for weight matrices, xt represents the input vector, ht1 signifies the previous hidden state, and bi, bo, and bc represent the bias vector.
In Bi-LSTM, the feature data obtained at time t contain both past and future information. Compared to a single LSTM structure, Bi-LSTM is more effective in extracting EEG signal features. Bi-LSTM can utilize both early and late sequence information, aiding in uncovering deep brain information from long EEG signal sequences, with its performance surpassing that of a unidirectional LSTM [29].
h t = σ ( W h × [ h t , h t ] + b h )
h t represents the current state in the forward direction, and h t represents the current state in the backward direction. Wh and bh represent the weight matrix and bias vector.

2.4. 1D SE-Block

The 2D SE attention mechanism, proposed by Hu et al. [30], secured the top spot in the ILSVRC 2017 classification task, demonstrating that the integration of SE blocks can bolster network performance. The SE attention mechanism, which operates on a channel-based approach, elevates attention along the channel dimension, emphasizing squeeze and excitation. This mechanism primarily identifies the attention level of each channel within the feature map, assigning an attention weight to each feature channel accordingly. This allows the convolutional network to prioritize these feature channels, thereby emphasizing the channels in the feature map that are pertinent to the current task while suppressing those that are less relevant.
Hence, this paper designs an SE block for one-dimensional data, as depicted in Figure 3. As shown in Figure 3, this one-dimensional SE module can deliberately amplify or attenuate specific channels by leveraging global information, thus enhancing the network’s performance. When compared to the straightforward usage of a single fully connected layer in the excitation operation, the benefit of employing two fully connected layers lies in their increased nonlinearity, enabling a superior fit for the intricate correlations among channels. Consequently, the one-dimensional SE block is well-suited for applications in electroencephalogram signal classification.
The one-dimensional SE block comprises two integral components: squeeze and excitation. Initially, the squeeze operation quantifies each feature channel into a corresponding real number (Zc) by employing a one-dimensional global average pooling operation. The formula for computing Zc is as follows [30]:
Z c = F s q ( u c ) = 1 W i = 1 W u c ( i )
W is the data length of a channel, and uc is the feature channel, ZcRc. Excitation generates the weight values Sc [30] for each feature channel, calculated as follows:
S c = F e x ( z c , W ) = σ ( g ( z c , W ) ) = σ ( W 2 δ ( W 1 z C ) )
σ and δ are the Sigmoid and ELU activation functions. W1 is the first fully connected layer (dimensionality reduction layer), W1   R c r × c , controlled by the reduction ratio (R). W2 is the second fully connected layer (dimensionality increase layer), W2 R c r × c .
Finally, multiply the weight values sc of the feature channel by the original channel uc to obtain the output of the one-dimensional SE block [31], as shown below.
F s c a l e ( u c , s c ) = u c s c

2.5. Dot Multiplier Attention Mechanism

Common attention mechanisms include dot product attention, scaled attention, spatial attention, and others. In this article, apart from utilizing the SE attention mechanism, we also incorporate dot product attention, which significantly improves the model’s accuracy, reduces computation time, and simplifies algorithm complexity. Within the dot product attention mechanism, attention scores are derived by computing the dot product between query and key vectors. Specifically, for a designated query vector Q and key vector K, the formula for calculating dot product attention is as follows: attention score = Q·K.
The dot product attention mechanism stands out for its efficient and straightforward calculation. It generates attention scores by directly computing the dot product between query and key vectors, eliminating the need for extra parameters or complex computational steps. Take the dot product attention mechanism as an example; its role is to assign weights to the expressions of hidden layer vectors outputted by the model. By integrating the attention mechanism into the feature extraction model, features that have a more significant impact on the output variables can be prioritized, thereby improving the accuracy of the method. Figure 4 illustrates the attention mechanism for point products.
The dot product attention comprises three integral components: the key matrix K, the value matrix V, and the query vector q [32].
First, Equation (12) defines the method for calculating the key matrix.
K = tanh ( V W a )
In this formula, Wa is a randomly initialized weight matrix, V is the original input vector, and K is the computed key vector. Using the hyperbolic tangent function tanh ensures that the values of the key vector are within the range of −1 to 1.
d = s o f t max ( q K T )
Next, Equation (13) describes how to calculate the attention weights (also known as the normalized probability vector d). Here, q is the query vector, and KT is the transpose of the key vector. The soft max function converts the computed attention scores into a probability distribution format, ensuring that the sum of all scores equals 1.
a = d V
Finally, Equation (14) defines how to apply the attention weights to obtain the attention vector a. Here, d is the previously calculated attention weight vector, and V is the original value vector. By multiplying the attention weights with the value vector, a new vector can be obtained that reflects the importance of various positions in the original vector.

2.6. Dataset

Our experiments were carried out on two widely used datasets: the Database for Emotion Analysis using Physiological Signals (SEED-IV) [33,34,35] (https://bcmi.sjtu.edu.cn/~seed/index.html, accessed on 18 December 2024) and the Database for Emotion Recognition utilizing EEG and ECG signals (DREAMER) [36,37] (https://zenodo.org/records/546113, accessed on 18 December 2024).
The movie clips used for both datasets are shown in Table 2 and Table 3.
The Shanghai JiaoTong University Emotional EEG Dataset (SEED-IV) is generously provided by Professor Lu from the BCMI Lab at Shanghai Jiao Tong University. SEED-IV comprises EEG recordings from 15 participants, with each participant undergoing three separate experiments conducted on different days. Each experiment encompassed 24 movie clips. During each experiment, participants watched a movie clip while their EEG signals and eye movements were meticulously captured from 62 channels, utilizing the ESI neuroimaging system alongside SMI eye-tracking glasses. The objective of these experiments was to evoke feelings of happiness, sadness, fear, or neutrality. The selection criteria for movie clips were as follows: (1) the duration of each experiment was kept brief to prevent participant fatigue, (2) the videos were designed to be self-explanatory, and (3) the videos were intended to evoke a singular target emotion.
DREAMER is a multimodal physiological dataset for emotional computing, collected by Katsigiannis et al. The experimental data comprise multi-channel EEG signals gathered from 23 healthy subjects (9 females and 14 males). In this database, emotions were elicited in each subject by presenting 18 movie clips. Each movie clip encompassed nine emotion categories: anger, fear, calm, amusement, sadness, surprise, disgust, happiness, and excitement. The duration of each movie clip varied from 65 to 393 s, averaging 199 s. EEG signal acquisition utilized 14 channels (adhering to the standard 10–20 electrode system) with a sampling rate of 128 Hz, while ECG signal acquisition employed a sampling rate of 256 Hz. Following the viewing of each movie clip, subjects self-assessed their valence, arousal, and dominance using a two-dimensional emotion space model, scoring from 1 to 5 as the outcome classification labels. Prior to the commencement of each clip, subjects were instructed to watch a neutral video to assist in adjusting their emotions to a neutral state.
To test the stability and portability of the model in emotion recognition, experiments were conducted on the SEED-IV and DREAMER datasets. The parameter of the original signal is channel × time, the number of channels in the SEED-IV dataset is 62, and the number of channels in DREAMER is 14. Due to the large amount of data, 1000 consecutive datasets were extracted from each person in the experiment in the middle of each video segment, resulting in 360,000 for SEED and SEED-IV (15 people × 24 videos × 1000 EEG data) and 1,242,000 for 23 people × 18 videos × 1000 EEG data × three labels (valence, arousal, and dominance).

2.7. Experimental Setup

In this experiment, due to the significant temporal correlation of each individual’s EEG data, using the same person’s EEG data as both training and testing sets may lead to biased and unreasonable results. Therefore, to ensure more reasonable and accurate results, this study divides the experiments of the SEED-IV dataset into approximately 80% (training set of 12 individuals) and approximately 20% (testing set of 3 individuals) and similarly divides the experiments of the DREAMER dataset into approximately 80% (training set of 18 individuals) and approximately 20% (testing set of 5 individuals). The experiment was conducted with the Adam optimizer for 150 training epochs, with a batch size of 1024.
To further validate the performance of the model, this paper also conducted a 10-fold cross-validation experiment. Ten-fold cross-validation involves dividing the dataset into 10 parts, with 9 parts of EEG signals used for the training set and the remaining 1 part as the test set, and so on, repeating this process 10 times. Next, the experimental results of the test set are saved, and the average value is used as the model evaluation criterion. This strategy can promote the model to learn samples from multiple aspects, avoiding the model becoming stuck in local extreme values. Ensuring the same data distribution between the training and test sets, all pre-training data are set with the same random seed, randomly shuffled, and passed to the network model. This paper implemented and modeled DACB and other comparative network models using the same parameter settings on GeForce RTX 4090 (Colorful, Xi’an, China). The training parameters of the DACB model are shown in Table 4.

2.8. Evaluation Index

The evaluation criteria used in the paper include accuracy, precision, recall rate, F1-score, and Matthew’s correlation coefficient to assess the model’s performance. Accuracy measures the proportion of correctly predicted samples among all samples. Precision is the proportion of true positive samples among all positive samples predicted by the classifier. Recall rate measures the proportion of correctly predicted positive samples among all true positive samples. F1-score is the harmonic mean of precision and recall. MCC is calculated based on the confusion matrix, considering true positive, true negative, false positive, false negative, etc., providing a comprehensive evaluation of the model’s performance. Specifically, the range of MCC values is [−1, 1], where 1 indicates perfect predictive ability, 0 indicates random prediction or no predictive ability, and −1 indicates completely opposite predictive ability. When MCC is a positive number, it represents a higher likelihood of the model’s predicted results matching the actual situation; when MCC is a negative number, it represents a higher likelihood of the model’s predicted results being opposite to the actual situation. The definitions of accuracy, precision, recall rate, F1-score, and Matthew’s correlation coefficient are as follows:
a c c u r a c y = T P + T N T P + T N + F P + F N
p r e c i s i o n = T P T P + F P
r e c a l l = T P T P + F N
F 1 s c o r e = 2 1 p r e c i s i o n + 1 r e c a l l
M C C = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
Among them, TP represents true positive samples; TN represents true negative samples; FP represents false positive samples, that is, incorrectly predicted as positive samples; and FN represents false negative samples, that is, incorrectly predicted as negative samples positive samples.

3. Results

3.1. Single-Test Result of DACB Model

To validate the proposed DACB model’s classification performance in EEG-based emotion detection, this study compares the model with Convolutional Neural Network-Recurrent Neural Network (CNN-RNN), Convolutional Neural Network-Long Short-Term Memory (CNN-LSTM), Convolutional Neural Network-Bidirectional Long Short-Term Memory (CNN-Bi-LSTM), 1D Convolutional Autoencoder (1D CAE), and 1D InceptionV1. 1D CAE consists of two convolutional layers replacing fully connected layers, with input symbols downsampled to provide smaller-dimension latent representations. On the other hand, 1D InceptionV1 replaces the 2D convolutional kernels of InceptionV1 with 1D convolutional kernels. In addition to comparing this model with other deep learning models, we also compare it with three popular traditional machine learning types: Adaboost, Bayes, and XGBoost. Traditional machine learning methods have also found widespread applications in various computer fields. Traditional machine learning typically requires manual feature engineering, extracting meaningful features from raw data. These methods are simple, effective, and interpretable for particularly straightforward tasks. In contrast, deep learning can automatically learn features through multi-layer neural networks, obtaining more sophisticated representations from raw data.
The DACB model was validated on the SEED-IV and DREAMER datasets, as presented in Table 5, Table 6, Table 7 and Table 8. The SEED-IV dataset comprises four distinct classes of labels, happiness, sadness, fear, and neutrality, representing various emotional states. On the other hand, the DREAMER dataset employs a two-dimensional emotional space model for self-assessment of valence, arousal, and dominance, with result classification labels ranging from 1 to 5. The DABC model is trained and tested separately in valence, arousal, and dominance.
In the SEED-IV dataset, the DACB model proposed in this paper performed the best in the four-classification task, with an accuracy of 99.96%, precision of 99.96%, recall of 99.96%, F1-score of 99.96%, and MCC of 99.95%. The CNN-Bi-LSTM model followed closely behind DACB, with an accuracy of 99.96%, precision of 99.96%, recall of 99.96%, F1-score of 99.96%, and MCC of 99.95%. Bayes showed the worst results, with an accuracy of 26.10%, precision of 30.44%, recall of 26.10%, F1-score of 17.39%, and MCC of 2.46%.
In the valence classification task on the DREAMER dataset, the DACB model achieved an accuracy of 87.52%, precision of 87.74%, recall of 87.41%, F1-score of 87.52%, and MCC of 84.28%. In the arousal classification task, the DACB model achieved an accuracy of 90.06%, precision of 90.15%, recall of 89.40%, F1-score of 89.76%, and MCC of 86.75%. In the dominance classification task, the DACB model achieved an accuracy of 89.05%, precision of 89.93%, recall of 88.69%, F1-score of 89.28%, and MCC of 85.24%. The DACB model performed the best in all three classification tasks on the DREAMER dataset, while the Bayesian method performed the worst. The results indicate that the proposed DACB model outperforms other models because its parallel structure can extract spatial-temporal features of input signals simultaneously, thereby improving model accuracy.

3.2. Ten-Fold Cross Validation Result of DACB Model

A single-model test verification may have randomness and uncertainty, so this paper also verified the performance of each model through 10-fold cross-validation. The results of the DACB model verified by 10-fold cross-validation are shown in Figure 5, Figure 6, Figure 7 and Figure 8. Table 9, Table 10, Table 11 and Table 12 summarize the average results of 10-fold cross-validation.
In the SEED-IV dataset, as shown in Table 9, the DACB model proposed in this paper still performs the best in the four-classification task, with an accuracy of 99.73%, precision of 99.72%, recall of 99.72%, F1-score of 99.72%, and MCC of 99.64%. Similarly, in 10-fold cross-validation, the CNN-Bi-LSTM model is only second to DACB, with an accuracy of 92.87%, precision of 92.89%, recall of 92.87%, F1-score of 92.87%, and MCC of 90.50%.
In the valence classification task on the DREAMER dataset, as shown in Table 10, the accuracy of the DACB model is 84.26%, precision is 84.39%, recall is 84.03%, F1-score is 84.18%, and MCC is 80.12%. The accuracy of the XGBoost model is 75.90%, precision is 77.08%, recall is 75.15%, F1-score is 75.89%, and MCC is 69.48%. In the arousal classification task, as shown in Table 11, the accuracy of the DACB model is 85.40%, precision is 85.54%, recall is 84.63%, F1-score is 85.04%, and MCC is 80.49%. The accuracy of the XGBoost model is 76.96%, precision is 81.75%, recall is 73.75%, F1-score is 76.89%, and MCC is 68.97%. In the dominance classification task, as shown in Table 12, the accuracy of the DACB model is 77.56%, precision is 81.58%, recall is 84.47%, F1-score is 85.10%, and MCC is 79.85%. The accuracy of the XGBoost model is 77.56%, precision is 81.58%, recall is 76.03%, F1-score is 78.35%, and MCC is 69.79%.
In two datasets, among the traditional three machine learning models, XGBoost exhibits better performance, with its overall performance second only to the DCBA model proposed in this paper. This is likely because the XGBoost algorithm is an ensemble learning method that enhances model accuracy and stability by constructing multiple decision trees. Among the five deep learning models, the CNN-Bi-LSTM model performs well on the SEED-IV dataset but shows a decline in performance on the DREAMER dataset, indicating poor generalization ability of the model. From the results of single validation and 10-fold cross-validation, it can be seen that the model in this paper performs the best among all comparison models, indicating the robustness of the DACB model in EEG-based emotion classification.

3.3. Ablation Experiment

To demonstrate and validate the benefits of concurrently extracting spatiotemporal information within a multi-channel model, this study also conducted ablation experiments on the SEED-IV and DREAMER datasets. Block1 eliminated the temporal feature module, while Block2 removed the spatial feature module, with all other modules remaining unchanged. The results of the ablation experiments are presented in Table 13 and Table 14, revealing that the single-channel spatial feature module performs inferiorly to the DACB model. In the SEED-IV dataset, Block2 outperforms Block1, indicating that temporal features are relatively important. It is noteworthy that, in the dreamer dataset, Block1 and Block2 show a performance difference of approximately 50% compared to our model, highlighting the superiority of the DACB model in combining spatial and temporal features for emotion recognition. This also demonstrates that relying solely on spatial or temporal features does not yield satisfactory experimental results, further validating the effectiveness of the proposed method.

3.4. Feature Visualization

This article adopts the SE (Squeeze-and-Excitation) attention mechanism, which enhances the model’s expressiveness by assigning different weights to different channels. To improve the model’s interpretability, the channel weights are visualized. Figure 9 presents the visualization results of electrode channel contributions: (a) corresponds to the analysis results of the SEED-IV dataset, while (b), (c), and (d) show the analysis results for the three emotional dimensions of valence, arousal, and dominance in the DREAMER dataset. The visual attention score is not calculated by the softmax function, so there is a negative number. Emotion is a complex phenomenon involving various cognitive processing mechanisms, and these processes are thought to result from the interaction of bottom-up and top-down information processing pathways [38,39]. These cognitive mechanisms exhibit significant neural activity in the parietal, temporal-parietal, and occipital regions of the brain [40,41,42]. The distribution of electrode channel feature weights extracted from EEG (electroencephalogram) signals using the SE attention mechanism aligns with these known brain cognitive functional areas, further validating the interpretability and effectiveness of the model proposed in this study.

4. Conclusions

This study introduces a multi-channel automatic emotion electroencephalogram (EEG) signal classification method, the DACB model. This proposed model incorporates various strategies to enhance accuracy. The proposed DACB model, a multi-channel spatiotemporal feature fusion network, achieves impressive classification accuracies of 99.73%, 84.26%, 85.40%, and 85.02% across four classification tasks in the SEED-IV and DREAMER datasets. The DACB model effectively captures spatial features through a one-dimensional SE module, extracts temporal features using a bidirectional long short-term memory network tailored for EEG as a time series signal, and selects these critical features via a dot product attention mechanism. Testing results on two distinct open-source EEG emotion datasets demonstrate the remarkable capabilities of the proposed deep learning model, which can learn features and classify stages without relying on manual feature extraction. This model holds promise for designing efficient and accurate real-time brain-computer interface frameworks. In future work, we aim to further improve the DACB model, such as exploring how DACB methods can enhance artifact removal or detect specific pathological patterns in EEG, so as to improve the accuracy of EEG recognition and classification and reduce the computing overhead of the model to facilitate deployment in the usability testing framework of emotion detection based on EEG signals [43,44]. In addition, we plan to incorporate multi-modal data, including EMG, ECG, or MRI, to further improve the generalization of the model.

Author Contributions

Conceptualization, Data curation, Investigation, Methodology, Validation, Visualization, Writing—original draft, Y.M. and Z.H.; Data curation, Project administration, Resources, Supervision, Y.M.; Supervision, Writing—review and editing, Y.M. and Z.H.; Investigation, Y.Y., S.Z., Q.D., R.W., and L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the General Projects of Shaanxi Science and Technology Plan (No. 2023-JC-YB-504), the National Natural Science Foundation of China (No. 62172338), the Shaanxi province innovation capacity support program (No. 2018KJXX-095), the Xijing University special talent research fund (No. XJ17T03), and the Natural Science Foundation of Chongqing CSTC (No. CSTB2022NSCQ-MSX1581).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from [BCMI laboratory, zenodo] and are available [from the authors/at https://bcmi.sjtu.edu.cn/~seed/index.html and https://zenodo.org/records/546113] with the permission of BCMI laboratory, zenodo].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lautenbach, F. Effects of positive affect and positive emotions on executive functions: A systematic review and meta-analysis. Cogn. Emot. 2024, 38, 1–22. [Google Scholar] [CrossRef]
  2. Gannouni, S.; Aledaily, A.; Belwafi, K.; Aboalsamh, H. Adaptive emotion detection using the valence-arousal-dominance model and EEG brain rhythmic activity changes in relevant brain lobes. IEEE Access 2020, 8, 67444–67455. [Google Scholar] [CrossRef]
  3. Mohammad, S. Obtaining reliable human ratings of valence, arousal, and dominance for 20,000 English words. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 174–184. [Google Scholar]
  4. Yıldırım, Ö.; Baloglu, U.B.; Acharya, U.R. A deep convolutional neural network model for automated identification of abnormal EEG signals. Neural Comput. Appl. 2020, 32, 15857–15868. [Google Scholar] [CrossRef]
  5. Liu, J.; Wu, G.; Luo, Y.; Qiu, S.; Yang, S.; Li, W.; Bi, Y. EEG-based emotion classification using a deep neural network and sparse autoencoder. Front. Syst. Neurosci. 2020, 14, 43. [Google Scholar] [CrossRef] [PubMed]
  6. Chen, L.; Wu, M.; Zhou, M.; Liu, Z.-T.; She, J.; Hirota, K. Dynamic emotion understanding in human–robot interaction based on two-layer fuzzy SVR-TS model. IEEE Trans. Syst. Man Cybern. Syst. 2017, 50, 490–501. [Google Scholar] [CrossRef]
  7. Busso, C.; Deng, Z.; Yildirim, S.; Bulut, M.; Lee, C.M.; Kazemzadeh, A.; Lee, S.; Neumann, U.; Narayanan, S. Analysis of emotion recognition using facial expressions, speech and multimodal information. In Proceedings of the 6th International Conference on Multimodal Interfaces, State College, PA, USA, 13–15 October 2004; pp. 205–211. [Google Scholar]
  8. Emerich, S.; Lupu, E.; Apatean, A. Emotions recognition by speechand facial expressions analysis. In Proceedings of the 2009 17th European Signal Processing Conference, Glasgow, UK, 24–28 August 2009; IEEE: Piscateaway, NJ, USA, 2009; pp. 1617–1621. [Google Scholar]
  9. Gunes, H.; Pantic, M. Automatic, dimensional and continuous emotion recognition. Int. J. Synth. Emot. IJSE 2010, 1, 68–99. [Google Scholar] [CrossRef]
  10. Liu, J.; Meng, H.; Nandi, A.; Li, M. Emotion detection from EEG recordings. In Proceedings of the 2016 12th International Conference on Natural Computation, Fuzzy Systems And Knowledge Discovery (ICNC-FSKD), Changsha, China, 13–15 August 2016; IEEE: Piscateaway, NJ, USA, 2016; pp. 1722–1727. [Google Scholar]
  11. Miao, M.; Zheng, L.; Xu, B.; Yang, Z.; Hu, W. A multiple frequency bands parallel spatial–temporal 3D deep residual learning framework for EEG-based emotion recognition. Biomed. Signal Process. Control. 2023, 79, 104141. [Google Scholar] [CrossRef]
  12. Ahern, G.L.; Schwartz, G.E. Differential lateralization for positive and negative emotion in the human brain: EEG spectral analysis. Neuropsychologia 1985, 23, 745–755. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Cheng, C.; Zhang, Y.D. Multimodal emotion recognition based on manifold learning and convolution neural network. Multimed. Tools Appl. 2022, 81, 33253–33268. [Google Scholar] [CrossRef]
  14. Rahman, A.; Rahman, H.; Biswas, P.; Hossain, S.; Islam, R.; Hannan, A.; Uddin, J.; Rhim, H. Potential Therapeutic Role of Phytochemicals to Mitigate Mitochondrial Dysfunctions in Alzheimer’s Disease. Antioxidants 2020, 10, 23. [Google Scholar] [CrossRef] [PubMed]
  15. Zhu, J.Y.; Zheng, W.L.; Lu, B.L. Cross-subject and cross-gender emotion classification from EEG. In Proceedings of the World Congress on Medical Physics and Biomedical Engineering, Toronto, ON, Canada, 7–12 June 2015; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 1188–1191. [Google Scholar]
  16. Wagh, K.P.; Vasanth, K. Performance evaluation of multi-channel electroencephalogram signal (EEG) based time frequency analysis for human emotion recognition. Biomed. Signal Process. Control. 2022, 78, 103966. [Google Scholar] [CrossRef]
  17. Liu, Y.; Fu, G. Emotion recognition by deeply learned multi-channel textual and EEG features. Future Gener. Comput. Syst. 2021, 119, 1–6. [Google Scholar] [CrossRef]
  18. Gannouni, S.; Aledaily, A.; Belwafi, K.; Aboalsamh, H. Emotion detection using electroencephalography signals and a zero-time windowing-based epoch estimation and relevant electrode identification. Sci. Rep. 2021, 11, 7071. [Google Scholar] [CrossRef] [PubMed]
  19. Naser, D.S.; Saha, G. Influence of music liking on EEG based emotion recognition. Biomed. Signal Process. Control. 2021, 64, 102251. [Google Scholar] [CrossRef]
  20. Tripathi, S.; Acharya, S.; Sharma, R.; Mittal, S.; Bhattacharya, S. Using deep and convolutional neural networks for accurate emotion classification on DEAP data. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31, pp. 4746–4752. [Google Scholar]
  21. Kim, Y.; Choi, A. EEG-based emotion classification using long short-term memory network with attention mechanism. Sensors 2020, 20, 6727. [Google Scholar] [CrossRef]
  22. Wang, F.; Wu, S.; Zhang, W.; Xu, Z.; Zhang, Y.; Wu, C.; Coleman, S. Emotion recognition with convolutional neural network and EEG-based EFDMs. Neuropsychologia 2020, 146, 107506. [Google Scholar] [CrossRef]
  23. Cui, H.; Liu, A.; Zhang, X.; Chen, X.; Wang, K.; Chen, X. EEG-based emotion recognition using an end-to-end regional-asymmetric convolutional neural network. Knowl. Based Syst. 2020, 205, 106243. [Google Scholar] [CrossRef]
  24. Li, J.; Hua, H.; Xu, Z.; Shu, L.; Xu, X.; Kuang, F.; Wu, S. Cross-subject EEG emotion recognition combined with connectivity features and meta-transfer learning. Comput. Biol. Med. 2022, 145, 105519. [Google Scholar] [CrossRef]
  25. Rajpoot, A.S.; Panicker, M.R. Subject independent emotion recognition using EEG signals employing attention driven neural networks. Biomed. Signal Process. Control. 2022, 75, 103547. [Google Scholar]
  26. Géron, A. Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques To Build Intelligent Systems; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2019. [Google Scholar]
  27. Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Mapp. 2017, 38, 5391–5420. [Google Scholar] [CrossRef]
  28. Hou, Y.; Zhou, L.; Jia, S.; Lun, X. A novel approach of decoding EEG four-class motor imagery tasks via scout ESI and CNN. J. Neural Eng. 2020, 17, 016048. [Google Scholar] [CrossRef] [PubMed]
  29. Chen, J.; Han, Z.; Qiao, H.; Li, C.; Peng, H. EEG-based sleep staging via self-attention based capsule network with Bi-LSTM model. Biomed. Signal Process. Control. 2023, 86, 105351. [Google Scholar] [CrossRef]
  30. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  31. Liu, X.; Xiong, S.; Wang, X.; Liang, T.; Wang, H.; Liu, X. A compact multi-branch 1D convolutional neural network for EEG-based motor imagery classification. Biomed. Signal Process. Control. 2023, 81, 104456. [Google Scholar] [CrossRef]
  32. Liu, C.; Liu, Y.; Yan, Y.; Wang, J. An intrusion detection model with hierarchical attention mechanism. IEEE Access 2020, 8, 67542–67554. [Google Scholar] [CrossRef]
  33. Zheng, W.L.; Lu, B.L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  34. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; IEEE: Piscateaway, NJ, USA, 2013; pp. 81–84. [Google Scholar]
  35. Zheng, W.-L.; Liu, W.; Lu, Y.; Lu, B.-L. Emotionmeter: A multimodal framework for recognizing human emotions. IEEE Trans. Cybern. 2018, 49, 1110–1122. [Google Scholar] [CrossRef]
  36. Gabert-Quillen, C.A.; Bartolini, E.E.; Abravanel, B.T.; Sanislow, C.A. Ratings for emotion film clips. Behav. Res. Methods 2015, 47, 773–787. [Google Scholar] [CrossRef] [PubMed]
  37. Katsigiannis, S.; Ramzan, N. DREAMER: A database for emotion recognition through EEG and ECG signals from wireless low-cost off-the-shelf devices. IEEE J. Biomed. Health Inform. 2017, 22, 98–107. [Google Scholar] [CrossRef] [PubMed]
  38. McRae, K.; Misra, S.; Prasad, A.K.; Pereira, S.C.; Gross, J.J. Bottom-up and top-down emotion generation: Implications for emotion regulation. Soc. Cogn. Affect. Neurosci. 2011, 7, 253–262. [Google Scholar] [CrossRef]
  39. Ochsner, K.N.; Ray, R.R.; Hughes, B.; McRae, K.; Cooper, J.C.; Weber, J.; Gabrieli, J.D.; Gross, J.J. Bottom-up and top-down processes in emotion generation: Common and distinct neural mechanisms. Psychol. Sci. 2009, 20, 1322–1331. [Google Scholar] [CrossRef]
  40. Comte, M.; Schön, D.; Coull, J.T.; McRae, K.; Cooper, J.C.; Weber, J.; Gabrieli, J.D.; Gross, J.J. Dissociating bottom-up and top-down mechanisms in the cortico-limbic system during emotion processing. Cereb. Cortex 2016, 26, 144–155. [Google Scholar] [CrossRef] [PubMed]
  41. Desimone, R.; Duncan, J. Neural mechanisms of selective visual attention. Annu. Rev. Neurosci. 1995, 18, 193–222. [Google Scholar] [CrossRef] [PubMed]
  42. Kajal, D.S.; Fioravanti, C.; Elshahabi, A.; Ruiz, S.; Sitaram, R.; Braun, C. Involvement of top-down networks in the perception of facial emotions: A magnetoencephalographic investigation. NeuroImage 2020, 222, 117075. [Google Scholar] [CrossRef] [PubMed]
  43. Gannouni, S.; Belwafi, K.; Aledaily, A.; Aboalsamh, H.; Belghith, A. Software Usability Testing Using EEG-Based Emotion Detection and Deep Learning. Sensors 2023, 23, 5147. [Google Scholar] [CrossRef]
  44. Gannouni, S.; Aledaily, A.; Belwafi, K.; Aboalsamh, H. Electroencephalography based emotion detection using ensemble classification and asymmetric brain activity. J. Affect. Disord. 2022, 319, 416–427. [Google Scholar] [CrossRef] [PubMed]
Figure 1. DACB structure.
Figure 1. DACB structure.
Brainsci 14 01289 g001
Figure 2. LSTM cell.
Figure 2. LSTM cell.
Brainsci 14 01289 g002
Figure 3. 1D SE-block.
Figure 3. 1D SE-block.
Brainsci 14 01289 g003
Figure 4. Dot multiplier attention mechanism.
Figure 4. Dot multiplier attention mechanism.
Brainsci 14 01289 g004
Figure 5. Accuracy of the DACB model on the SEED-IV dataset based on 10-fold cross-validation.
Figure 5. Accuracy of the DACB model on the SEED-IV dataset based on 10-fold cross-validation.
Brainsci 14 01289 g005
Figure 6. Accuracy of the DACB model on the DREAMER-Valence dataset based on 10-fold cross-validation.
Figure 6. Accuracy of the DACB model on the DREAMER-Valence dataset based on 10-fold cross-validation.
Brainsci 14 01289 g006
Figure 7. Accuracy of the DACB model based on 10-fold cross-validation on the DREAMER-Arousal dataset.
Figure 7. Accuracy of the DACB model based on 10-fold cross-validation on the DREAMER-Arousal dataset.
Brainsci 14 01289 g007
Figure 8. The accuracy of the DACB model on the DREAMER-Dominance dataset based on 10-fold cross-validation.
Figure 8. The accuracy of the DACB model on the DREAMER-Dominance dataset based on 10-fold cross-validation.
Brainsci 14 01289 g008
Figure 9. Visualization of electrode channel contribution.
Figure 9. Visualization of electrode channel contribution.
Brainsci 14 01289 g009
Table 1. Parameters of the DACB architecture.
Table 1. Parameters of the DACB architecture.
ModuleTypeNumber
Block1Conv1d64
Flatten1
Block2Bi-LSTM32
Flatten1
FC1Dense64
FC2Dense32
FC3Softmax1
Table 2. SEED-IV dataset movie snippets.
Table 2. SEED-IV dataset movie snippets.
Serial No.Film Clips’ SourcesEmotion LabelStart Time PointEnd Time Point
01Black Keyssad42:3245:41
02The Eye 3fear49:25:0051:00:00
03Rob-B-Hoodhappy41:07:0045:06
04A Bite of Chinaneutral30:2932:48
05The Child’s Eyefear41:0042:37
06A Bite of Chinaneutral5:198:05
07A Bite of Chinaneutral24:4226:41
08Very Happysad17:0921:13
09A Bite of Chinaneutral31:1833:44
10A Wedding Invitationsad1:34:041:38:50
11Bunshinsaba IIfear42:2443:33
12Dearestsad1:31:081:33:29
13Aftershocksad20:1324:14
14Foster Fathersad24:2927:10
15Bunshinsaba IIIfear1:04:521:09:49
16Promo for applying the Olympic Winter Gameshappy0:002:54
17Hungry Ghost Ritualfear45:0746:48
18Hungry Ghost Ritualfear1:10:211:13:33
19Very Happyhappy34:3037:15
20You are my life more completehappy39:3240:44
21A Bite of China neutral18:5920:56
22Hear Mehappy1:33:2796:10
23A Bite of Chinaneutral16:2819:24
24Very Happyhappy12:4815:31
Table 3. DREAMER dataset movie snippets.
Table 3. DREAMER dataset movie snippets.
Serial No.Film Clips’ SourcesValenceArousalDominance
01Searching for Bobby Fischer3.17 ± 0.722.26 ± 0.752.09 ± 0.73
02D.O.A.3.04 ± 0.883.00 ± 1.002.70 ± 0.88
03The Hangover4.57 ± 0.733.83 ± 0.833.83 ± 0.72
04The Ring2.04 ± 1.024.26 ± 0.694.13 ±0.87
053003.22 ± 1.173.70 ± 0.703.52 ± 0.95
06National Lampoon’s VanWilder2.70 ±1.553.83 ± 0.834.04 ± 0.98
07Wall-E4.52 ± 0.593.17 ± 0.983.57 ± 0.99
08Crash1.35 ± 0.653.96 ± 0.774.35 ± 0.65
09My Girl1.39 ± 0.663.00 ± 1.093.48 ± 0.95
10The Fly2.17 ± 1.153.30 ± 1.023.61 ± 0.89
11Pride and Prejudice3.96 ± 0.641.96 ± 0.822.61 ± 0.89
12Modern Times3.96 ± 0.562.61 ± 0.892.70 ± 0.82
13Remember the Titans4.39 ± 0.663.70 ± 0.973.74 ± 0.96
14Gentleman’s Agreement2.35 ± 0.652.22 ± 0.852.39 ± 0.72
15Psycho2.48 ± 0.853.09 ± 1.003.22 ± 0.9
16The Bourne Identity3.65 ± 0.653.35 ± 1.073.26 ± 1.14
17The Shawshank Redemption1.52 ± 0.593.00 ± 0.743.96 ± 0.77
18The Departed2.65 ± 0.783.91 ± 0.853.57 ± 1.04
Table 4. DACB model training parameters.
Table 4. DACB model training parameters.
ParametersNumber
epoch number150
learning rate0.001
batch size1024
optimizerAdam
loss functioncategorical_crossentropy
Random seed42
Table 5. The performance of the DACB model and other models on SEED-IV of task test sets.
Table 5. The performance of the DACB model and other models on SEED-IV of task test sets.
MethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
CNN-RNN61.5362.9461.5360.7249.50
CNN-Bi-LSTM96.1596.1696.1596.1594.88
1D CAE90.2890.2890.2890.2887.05
1D InceptionV179.2879.7379.2879.3372.49
Adaboost37.4937.5237.4937.4116.69
Bayes26.1030.4426.1017.392.46
XGBoost87.2587.3587.2487.2683.02
DACB99.9699.9699.9699.9699.95
Table 6. The performance of the DACB and other models on DREAMER-Valence of task test sets.
Table 6. The performance of the DACB and other models on DREAMER-Valence of task test sets.
MethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
CNN-RNN58.2559.1358.2558.4647.52
CNN-Bi-LSTM75.6876.1675.2975.6569.31
1D CAE67.3867.4567.2467.2458.89
1D InceptionV134.8239.3032.2931.9316.82
Adaboost34.6634.6332.7032.6416.76
Bayes22.5323.9422.2815.963.96
XGBoost81.1682.0880.8681.3576.24
DACB87.5287.7487.4187.5284.28
Table 7. The performance of DACB and other models on DREAMER-Arousal of task test sets.
Table 7. The performance of DACB and other models on DREAMER-Arousal of task test sets.
MethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
CNN-RNN63.3866.6458.7161.4450.56
CNN-Bi-LSTM76.4777.8974.5076.0068.49
1D CAE71.3874.1568.3470.6861.56
1D InceptionV140.5949.4131.4232.2017.85
Adaboost38.6937.2030.6231.2215.72
Bayes15.2625.4525.2915.544.78
XGBoost81.9585.2980.5682.6275.81
DACB90.0690.1589.4089.7686.75
Table 8. The performance of DACB model and other models on DREAMER-Dominance of task test sets.
Table 8. The performance of DACB model and other models on DREAMER-Dominance of task test sets.
MethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
CNN-RNN65.2067.1162.1764.0652.73
CNN-Bi-LSTM73.2176.7670.6873.0163.67
1D CAE70.4671.9171.0470.9460.56
1D InceptionV140.2147.5629.7829.6615.74
Adaboost37.3033.6829.2729.6912.86
Bayes21.3625.7526.1918.096.37
XGBoost82.3385.8882.2783.8576.10
DACB89.0589.9388.6989.2885.24
Table 9. The performance of the DACB model based on 10-fold cross-validation (SEED-IV dataset).
Table 9. The performance of the DACB model based on 10-fold cross-validation (SEED-IV dataset).
MethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
CNN-RNN62.3162.5362.3062.1949.87
CNN-Bi-LSTM92.8792.8992.8792.8790.50
1D CAE85.2185.2885.2185.2280.30
1D InceptionV177.6077.6677.6077.5970.15
Adaboost35.9336.0035.9335.8214.61
Bayes25.7728.8425.7717.341.66
XGBoost80.5780.8180.5780.6174.15
DACB99.7399.7299.7299.7299.64
Table 10. The performance of the DACB model based on 10-fold cross-validation (DREAMER-Valence).
Table 10. The performance of the DACB model based on 10-fold cross-validation (DREAMER-Valence).
MethodsAccuracy
(%)
Precision
(%)
Recall
(%)
F1-Score
(%)
MCC
(%)
CNN-RNN26.1824.9820.098.501.76
CNN-Bi-LSTM33.0937.7530.2528.7714.60
1D CAE50.6852.9149.3149.9137.53
1D InceptionV128.1337.6724.1919.117.76
Adaboost32.9532.7229.7328.8713.72
Bayes21.8821.1821.1113.772.24
XGBoost75.9077.0875.1575.8969.48
DACB84.2684.3984.0384.1880.12
Table 11. The performance of the DACB model based on 10-fold cross-validation (DREAMER-Arousal).
Table 11. The performance of the DACB model based on 10-fold cross-validation (DREAMER-Arousal).
MethodsAccuracy
(%)
Precision
(%)
Recall
(%)
F1-Score
(%)
MCC
(%)
CNN-RNN32.1119.2118.5610.672.17
CNN-Bi-LSTM40.5344.6930.5729.9717.06
1D CAE51.9554.0145.3846.4733.74
1D InceptionV134.4547.9924.2420.679.38
Adaboost38.0035.3129.0029.0913.91
Bayes12.4224.1124.7912.924.36
XGBoost76.9681.7573.7576.8968.97
DACB85.4085.5484.6385.0480.49
Table 12. The performance of the DACB model based on 10-fold cross-validation (DREAMER-Dominance).
Table 12. The performance of the DACB model based on 10-fold cross-validation (DREAMER-Dominance).
MethodsAccuracy
(%)
Precision
(%)
Recall
(%)
F1-Score
(%)
MCC
(%)
CNN-RNN32.4119.1718.269.941.23
CNN-Bi-LSTM37.4540.2126.9624.4612.73
1D CAE53.1658.6846.7648.7635.76
1D InceptionV134.0345.2822.5116.787.00
Adaboost36.5532.0826.6625.9711.13
Bayes10.4429.4222.6211.364.30
XGBoost77.5681.5876.0378.3569.79
DACB85.0285.7484.4785.1079.85
Table 13. Ablation experiments of the DACB model based on 10-fold cross-validation (SEED-IV dataset).
Table 13. Ablation experiments of the DACB model based on 10-fold cross-validation (SEED-IV dataset).
DataSetMethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
SEED-IVBlock180.8885.3480.9179.7975.80
SEED-IVBlock290.6899.6890.6899.6899.58
SEED-IVDACB99.7399.7299.7299.7299.64
Table 14. Ablation experiments of the DACB model based on 10-fold cross-validation (DREAMER dataset).
Table 14. Ablation experiments of the DACB model based on 10-fold cross-validation (DREAMER dataset).
DataSetMethodsAccuracy (%)Precision (%)Recall (%)F1-Score (%)MCC (%)
DREAMER
(valence)
Block126.087.2120.008.270.11
DREAMER
(valence)
Block232.165.2120.008.270
DREAMER
(valence)
DACB84.2684.3984.0384.1880.12
DREAMER
(arousal)
Block131.8810.3720.009.670.31
DREAMER
(arousal)
Block231.8712.1120.009.660
DREAMER
(arousal)
DACB85.4085.5484.6385.0480.49
DREAMER
(dominance)
Block132.3821.8020.019.810.84
DREAMER
(dominance)
Block232.366.4620.009.770
DREAMER
(dominance)
DACB85.0285.7484.4785.1079.85
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, Y.; Huang, Z.; Yang, Y.; Zhang, S.; Dong, Q.; Wang, R.; Hu, L. Emotion Recognition Model of EEG Signals Based on Double Attention Mechanism. Brain Sci. 2024, 14, 1289. https://doi.org/10.3390/brainsci14121289

AMA Style

Ma Y, Huang Z, Yang Y, Zhang S, Dong Q, Wang R, Hu L. Emotion Recognition Model of EEG Signals Based on Double Attention Mechanism. Brain Sciences. 2024; 14(12):1289. https://doi.org/10.3390/brainsci14121289

Chicago/Turabian Style

Ma, Yahong, Zhentao Huang, Yuyao Yang, Shanwen Zhang, Qi Dong, Rongrong Wang, and Liangliang Hu. 2024. "Emotion Recognition Model of EEG Signals Based on Double Attention Mechanism" Brain Sciences 14, no. 12: 1289. https://doi.org/10.3390/brainsci14121289

APA Style

Ma, Y., Huang, Z., Yang, Y., Zhang, S., Dong, Q., Wang, R., & Hu, L. (2024). Emotion Recognition Model of EEG Signals Based on Double Attention Mechanism. Brain Sciences, 14(12), 1289. https://doi.org/10.3390/brainsci14121289

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop