Next Article in Journal
Learning Causes of Functional Dynamic Targets: Screening and Local Methods
Previous Article in Journal
Causalized Convergent Cross Mapping and Its Implementation in Causality Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GC-STCL: A Granger Causality-Based Spatial–Temporal Contrastive Learning Framework for EEG Emotion Recognition

1
School of Software Technology, Dalian University of Technology, Dalian 116024, China
2
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
3
School of Innovation and Entrepreneurship, Dalian University of Technology, Dalian 116024, China
4
School of Computer Science and Technology, Dalian University of Technology, Dalian 116024, China
*
Authors to whom correspondence should be addressed.
Entropy 2024, 26(7), 540; https://doi.org/10.3390/e26070540
Submission received: 12 April 2024 / Revised: 6 June 2024 / Accepted: 17 June 2024 / Published: 24 June 2024
(This article belongs to the Section Complexity)

Abstract

:
EEG signals capture information through multi-channel electrodes and hold promising prospects for human emotion recognition. However, the presence of high levels of noise and the diverse nature of EEG signals pose significant challenges, leading to potential overfitting issues that further complicate the extraction of meaningful information. To address this issue, we propose a Granger causal-based spatial–temporal contrastive learning framework, which significantly enhances the ability to capture EEG signal information by modeling rich spatial–temporal relationships. Specifically, in the spatial dimension, we employ a sampling strategy to select positive sample pairs from individuals watching the same video. Subsequently, a Granger causality test is utilized to enhance graph data and construct potential causality for each channel. Finally, a residual graph convolutional neural network is employed to extract features from EEG signals and compute spatial contrast loss. In the temporal dimension, we first apply a frequency domain noise reduction module for data enhancement on each time series. Then, we introduce the Granger–Former model to capture time domain representation and calculate the time contrast loss. We conduct extensive experiments on two publicly available sentiment recognition datasets (DEAP and SEED), achieving 1.65% improvement of the DEAP dataset and 1.55% improvement of the SEED dataset compared to state-of-the-art unsupervised models. Our method outperforms benchmark methods in terms of prediction accuracy as well as interpretability.

1. Introduction

As a crucial component of affective computing, emotion recognition has garnered increasing attention from scholars in recent years and has emerged as a significant research topic at the intersection of neuroscience, psychology, computer science, and artificial intelligence [1]. Broadly speaking, emotion recognition methods can be categorized into two groups: one based on non-physiological signals such as speech, text, facial expressions, etc. [2,3], and the other based on physiological signals like electroencephalogram (EEG), electrocardiogram (ECG), EMG, etc. [4,5,6]. Physiological signals directly reflect the brain’s state under different emotions and are considered to be more objective. With advancements in EEG acquisition equipment and technology, EEG has become the preferred method for studying the brain’s response to emotional stimuli.
Firstly, the mining of complex spatial topological relationships among EEG channels poses a significant challenge. The EEG signals are captured through multiple electrode channels that collectively form the spatial structure of these signals. Existing studies often determine channel spatial topological relationships solely based on the physical distance between electrodes or channel correlation, as exemplified by GCNN, P-GCNN, and GCNs-Net [7,8,9]. However, previous research has demonstrated that different brain electrical channels frequently exhibit intricate causal relationships. Therefore, simple undirected graph modeling struggles to accurately depict the complex information transmission between these channels [10,11]. Consequently, we argue that directed causal graph modeling can effectively capture the causal relationship between channels.
Secondly, high levels of noise interference constitute the primary factor contributing to model overfitting in emotion recognition tasks. During the process of EEG data collection, it is often subject to interference from internal emotions and external devices, resulting in a low signal-to-noise ratio characteristic of EEG data. Furthermore, the high temporal resolution of EEG signals necessitates a time-consuming and labor-intensive sample labeling process. Most supervised or semi-supervised methods are ineffective when dealing with limited labeled training data.Previous studies, such as SGMC and CLISA [12,13], employed video alignment to obtain similar samples for constructing positive sample pairs. These studies utilized an unsupervised contrast learning framework to extract intrinsic features of specific emotional EEG signals, thereby enhancing the model’s generalization capability. However, these methods still lack robustness against noise interference. Therefore, an effective strategy for constructing positive samples pairs temporally can further augment the model’s generalization ability.
To address the aforementioned challenges, this paper proposes a Granger causality based spatial–temporal contrast learning framework called GC-STCL. In the spatial dimension, we initially consider the EEG channel as a node and utilize the physical distance between sensors as edges to construct the original EEG graph structure. Subsequently, we introduced a graph augmentation method at the structural level by leveraging Granger causality. In contrast to approaches based on channel distance or correlation, our proposed method constructs a directed graph through rigorous Granger causality testing. This not only captures the causal relationships between brain electrical channels but also eliminates unnecessary connections, thereby enhancing the accuracy of emotion recognition. Furthermore, we employ a sampling strategy to select positive sample pairs from individuals who watched the same video. Finally, spatial feature extraction of EEG signals is achieved through multi-layer graph convolution connected by residual edges, followed by calculation of spatial contrast loss based on these extracted features. In terms of time dimension, we employ a frequency domain noise reduction data augmentation strategy to construct positive sample pairs. Differing from traditional time series data augmentation methods such as translation, masking and amplification, our proposed approach exhibits stronger resistance against noise interference. Furthermore, inspired by Granger causality tests, we propose using Granger–Former to capture time representations. Lastly, time comparison loss is calculated based on temporal characteristics.
To summarize, the primary contributions of this paper are as follows:
  • We propose a spatial–temporal contrast learning framework (Granger-STCL) based on the Granger causality test, which effectively enhances the emotion recognition capability of EEG signals.
  • We validate the significance of directed causal graph and temporal causal modeling. Furthermore, we introduce spatial–temporal positive sample pair construction strategies for enhancing causal graph and reducing frequency domain noise respectively, successfully improving the generalization ability of contrast learning.
  • Extensive experimental results on two publicly available datasets demonstrate that GC-STCL outperforms the benchmark method in terms of prediction accuracy and interpretability.

2. Related Works

In this section, we comprehensively examine EEG-based methods for emotion recognition, with a primary focus on deep learning and contrastive learning approaches.

2.1. EEG-Based Emotion Recognition

As a result of the success of deep neural networks, deep learning-based emotion recognition has received more attention. Various deep learning methods have been utilized for this purpose. Zhang et al. [14] proposed a model which consists of two RNN layers to extract both temporal and spatial features from EEG for emotion recognition. Li et al. [15] introduced a self-attention network that captures spatial domain and frequency domain features from EEG by utilizing a parallel convolutional neural network (PCNN) layer. Chen et al. [16] presented layered bidirectional gated cycle unit (GRU) networks to mitigate the impact of long-term non-stationarity in EEG.
In order to investigate the relationship between EEG channels and extract spatial features, the researchers employed graph convolutional networks (GCN) to capture the interrelationships among brain electrical channels by constructing EEG networks based on either physical proximity or functional connectivity between channels. Yin et al. [17] constructed the graph network using a Gaussian kernel function and extracted both spatial and temporal features through GCN combined with LSTM. Du et al. [18] integrated Cartesian distance between channels and prior knowledge to establish a graph network, from which they extracted EEG features utilizing GCN. Lin et al. [19] determined the relationships among EEG channels using a phase lag index (PLI), then extracted intra-channel and inter-channel features through 1D convolutional layers and Graph Attention Networks (GAT). Feng et al. [20] identified brain electrical channel connections via a Pearson correlation coefficient and achieved emotion recognition by employing GCN, Bi-LSTM, as well as an attention mechanism.
Despite the favorable outcomes achieved by these approaches, a majority of them construct EEG data using undirected graphs, thereby overlooking the directional information flow between EEG signals. However, it has been demonstrated in previous studies that there is a causal relationship among brain regions [21], which significantly contributes to extracting EEG features and enhancing accuracy in emotion recognition.

2.2. Contrastive Learning

Contrastive learning, a self-supervised algorithm that leverages unlabeled data to acquire representations, has found extensive applications in computer vision (CV) [22], natural language processing (NLP) [23], and bioinformatics [24], etc. In the domain of EEG emotion recognition, there have also been studies employing contrastive learning techniques. Mohsenvand et al. [25] proposed a contrastive learning framework, SeqCLR, based on SimCLR. The model augments the original data through various techniques such as masking, linear scaling, and Gaussian noise addition, etc. By employing both pre-training and fine-tuning stages, impressive emotion recognition results are achieved. Li et al. [26] proposed an innovative approach for emotion identification through EEG by employing a convolutional neural network and incorporating supervised contrastive learning (ECNN-C). The proposed model not only mitigates computational complexity, but also enhances the accuracy of emotion recognition. Guo et al. [27] proposed a contrastive learning approach that integrates multi-head attention mechanisms for unsupervised representation learning and feature fusion, aiming to enhance the accuracy of emotion recognition.
Currently, there is a paucity of research on EEG based emotion recognition utilizing the contrastive learning framework. Furthermore, the current research often overlooks the interdependence between EEG channels during feature extraction, and there is a dearth of studies on noise data processing. Consequently, there exists significant potential for further development of contrast learning in the domain of emotion recognition based on EEG signals.

3. Methodology

3.1. Problem Definition

Given a time series X R L × V , where L represents the length of the input sequence and V represents the number of variables, our objective is to learn a nonlinear function f θ that maps the input sequence X to the representation vector Z R F e , where F e denotes the dimensionality of the representation. To achieve this goal, we propose two contrast learning frameworks for extracting EEG temporal and spatial features, respectively. In the temporal dimension, we employ a frequency noise reduction method to eliminate noise data from X , resulting in X T . We consider both original time series data X and filtered data X T as positive sample pairs, encode them using Granger–Former, calculate contrast loss, and obtain the time domain representation of X , which is denoted as Z T . For spatial dimension analysis, we collect EEG samples from two subjects watching identical video clips as positive samples. By constructing and enhancing graph data through the Granger causality test, encoding with ResGCN model, calculating contrast loss, we derive a spatial representation of X , which is referred to as Z S . Finally, we combine these representations by summing them up as follows: Z = Z T + Z S . The overall structure of the model is presented in Figure 1.

3.2. Spatial Contrast Learning

3.2.1. Graph Construction and Augmentation

In the spatial dimension, utilizing a graph structure to model EEG signals enables more effective utilization of information between EEG channels. Given a graph G = Λ , E , Λ represents a set of nodes and E represents a set of edges. For EEG signals, Λ c denotes the channels of the signals. Let A R C × C represent the adjacency matrix of G, which can depict the topology structure of EEG signals. The EEG graph can be defined as G = Λ , E , A . According to previous studies [28], the correlation between EEG channels exhibits a negative relationship with the square of their Euclidean distance d i j .
d i s t i , j = θ d i j 2
The adjacency matrix can be represented as follows:
A i j = 1 , d i s t i , j δ 0 , o t h e r w i s e
where θ is the calibration constant and δ is the hyperparameter, which is used to remove the less correlated channels.
During the construction of the aforementioned adjacency matrix, the consideration of directionality in information transmission was omitted. Hence, this paper proposes a data enhancement method based on the Granger causality test to construct an asymmetric adjacency matrix that accurately reflects the directional relationships between EEG channels and improves emotion recognition accuracy [21,29]. For two EEG channels i and j, with their respective time series data denoted as X i = { x i 1 , x i 2 , , x i L } and X j = { x j 1 , x j 2 , , x j L } , the fundamental concept behind the Granger causality test is that if combining the value of X i with that of X j yields better results than considering only the value of X j , it indicates that X i has a Granger causal influence on X j . The autoregressive model representing this relationship can be expressed as follows:
x i t = p = 1 P α 1 , p x i t p + ε 1 t x j t = p = 1 P α 1 , p x j t p + ε 2 t
The bivariate regressive model can be formulated as follows:
x i t = p = 1 P α 11 , p x i t p + p = 1 P α 12 , p x j t p + ε 3 t x j t = p = 1 P α 21 , p x i t p + p = 1 P α 22 , p x j t p + ε 4 t
The lag order of the model is denoted as P, the regression coefficients α 1 , p and α 2 , p represent the parameters for the univariate model, while α 11 , p , α 12 , p , α 21 , p , α 22 , p are the binary regression coefficients for the bivariate model. The prediction errors of the univariate autoregressive model, denoted as ε 1 t and ε 2 t , have variances var ε 1 and var ε 2 respectively. Similarly, the prediction errors of the binary regression model, denoted as ε 3 t and ε 4 t , have variances var ε 3 and var ε 4 respectively. In this context, Granger causality between the two channels can be defined.
G C Λ j Λ i = ln var ε 1 var ε 3
G C Λ i Λ j = ln var ε 2 var ε 4
After conducting the Granger causality test to enhance the aforementioned adjacency matrix, we obtain
A i j G C = 1 , d i s t i , j τ a n d G C Λ j Λ i > 0 0 , o t h e r w i s e

3.2.2. Spatial Encoder

In order to extract EEG features for contrast learning, an encoder based on ResGCN and ResNet18 is designed in this paper [30,31]. It consists of nine residual blocks, each containing two GCN layers, all of which are connected by a residual. Given an input graph G with its representation at layer l denoted as G l , the encoder first generates the residual graph representation G l + 1 r e s at the next layer using a residual map F and then adds vertices to obtain the output G l + 1 . Let H represent the underlying map of the encoder and W l denote the parameters to be learned at layer l. We obtain the following:
G l + 1 = H ( G l , W l ) = F ( G l , W l ) = G l + 1 r e s + G l
The features extracted from each GCN layer are subsequently combined through a 1 × 1 convolutional layer and a maximum pooling layer, resulting in the acquisition of a comprehensive representation that integrates both global and local features.
G ˜ = h G
The structure of the spatial encoder is illustrated in Figure 2.

3.2.3. Projector

According to GraphCL [32], a nonlinear projection head is introduced between the encoder and the final contrast loss in order to project the encoded EEG signal learned by the encoder into a latent space for computing the contrast loss. In this study, we employ a Feed-forward Network (FFN) as the projection head. It comprises three 1 × 1 convolutional layers with Batch Normalization applied to each layer. Ultimately, this yields the feature representation.
Z S = F F N S G ˜

3.2.4. Spatial Contrastive Loss Function

The sampling strategy is employed to select samples X i that have watched the same video clips as positive samples, corresponding to the representations Z S , i and Z S , i , respectively, for sample X i . The negative sample set D i S is randomly drawn from samples that have watched other video clips. We employ a modified version of contrast learning inspired by MoCo to establish the spatial contrastive loss [33].
L S = 1 N i = 1 N log exp ( s i m ( Z S , i , Z S , i ) / τ ) exp ( s i m ( Z S , i , Z S , i ) / τ ) + Z S , j D i S exp ( s i m ( Z S , i , Z S , j ) / τ )
s i m ( Z S , i , Z S , i ) = Z S , i · Z S , i Z S , i Z S , i
where N denotes the number of samples and τ is the temperature parameter.

3.3. Temporal Contrast Learning

3.3.1. Frequency Noise Reduction

The traditional methods for enhancing time series data typically employ time shift, masking, and amplitude scale, etc. [25]. However, these aforementioned techniques are not suitable for EEG signal prediction tasks. Firstly, EEG signals exhibit high complexity and individual specificity. Traditional data enhancement methods may disrupt the subtle patterns and correlations present in EEG signals, thereby significantly impacting the accuracy of prediction tasks. Secondly, EEG signals possess strong temporal dependencies. Enhancement techniques such as time shift or amplitude scale can alter the temporal relationships within the signal, distorting its inherent time dependence. Lastly, each component of an EEG signal potentially carries crucial physiological information regarding brain states. Methods like adding masks or introducing random noise may obscure or modify this critical physiological information, resulting in predictive models that fail to accurately interpret or identify specific brain states. Consequently, we propose a frequency domain noise reduction data enhancement strategy that is better suited for processing EEG signals.
Specifically, for a given input signal X T , it is initially transformed into the frequency domain using the fast Fourier transform ( F F T ). Subsequently, noise reduction is accomplished by filtering out frequencies corresponding to the Top-K amplitudes, thereby obtaining the respective positive sample X T .
X T = F 1 ( S e l e c t F ( T o p K ( A ( F ( X T ) ) ) ) )
where F denotes the FFT and F 1 is its inverse, F ( X T ) R E , where E = L / 2 + 1 , S e l e c t F ( ) denotes the frequency that corresponds to the selected amplitude, A ( F ) = F r 2 + F i 2 , F r and F i represent the corresponding real and imaginary parts, respectively.

3.3.2. Granger–Former

To enhance the capture of temporal relationships in time series data, we propose extracting temporal EEG representations based on Granger causality and employ the Granger causality test instead of traditional self-attention mechanisms to depict causal connections among the data.
Transformer exhibits significant potential in time series prediction tasks [34]. However, conventional self-attention mechanisms can only capture the relationships between time points in time series modeling and fail to account for more intricate causal associations preceding or succeeding specific segments of time. Hence, we adopt the foundational framework of Transformer and substitute self-attention mechanism with Granger-attention to achieve temporal causality modeling. Specifically, for a given time series X R L × V , we first capture the time period after each time point with a length of P through a sliding window. If the length after the time point is less than P, it is padded with zeros. This results in obtaining X P R L × P × V . Subsequently, the queries Q R L × P × F e , keys K R L × P × F e , and values V R L × P × F e are mapped through three linear layers. Finally, the attention relevance score is ultimately computed via a Granger causality test, thereby facilitating the comprehensive aggregation of values.
Z T ˜ = s o f t m a x ( R e L U ( G r a n g e r ( Q , K ) ) ) · V
The calculation details of G r a n g e r ( ) are presented in Section 3.2.1. Associations below zero are rectified to zero using the R e L U ( ) activation function.The structure of Granger–Former is presented in Figure 3.
We also use the MLP as the temporal projector, facilitating the acquisition of a refined representation of the input signal in the time domain.
Z T = M L P T Z T ˜

3.3.3. Temporal Contrastive Loss Function

Similar to the calculation of contrast loss in the spatial dimension, we select sample X T , i after noise reduction as a positive pair for sample X T , i . These samples are then subjected to Granger causality analysis to allow us to determine their corresponding weights, resulting in Z T , i and Z T , i . From all samples, we randomly draw the negative sample set D i T . Finally, the temporal contrastive loss can be expressed as follows:
L T = 1 N i = 1 N log exp ( s i m ( Z T , i , Z T , i ) / τ ) exp ( s i m ( Z T , i , Z T , i ) / τ ) + Z T , j D i T exp ( s i m ( Z T , i , Z T , j ) / τ )
The total loss of GC-STCL is formulated as follows:
L = L S + L T

4. Experiments

To validate the model, we conducted extensive experiments on two widely used emotion recognition datasets, namely DEAP and SEED, and compared them with state-of-the-art methods. Additionally, we designed ablation experiments to verify the rationality of each component of the model and further discussed the role of frequency noise reduction and the Granger causality test in enhancing its performance.

4.1. Settings

4.1.1. Datasets

DEAP [35]: The DEAP dataset comprises 32 channels of EEG signals and 8 channels of other physiological signals, which were recorded from 32 subjects during their observation of a series of 40 one-minute music videos. Each segment of the EEG data consists of a baseline period lasting for 3 s followed by a test signal lasting for 60 s. All the data were downsampled to a frequency of 128 Hz and processed using a bandpass filter with cutoff frequencies ranging from 4 Hz to 45 Hz. Subsequently, participants provided ratings on a scale from 1 to 9 regarding their emotional state after watching the videos, encompassing dimensions such as arousal, valence, liking, and dominance. In this paper, we employed a 1 s sliding window to partition the EEG signal into distinct segments totaling up to 63 segments. Subsequently, the average of the preceding 3 s baseline signal was subtracted from each 1 s emotion-related signal, serving as the input data. For each subject, we have access to a dataset comprising 2400 signals (40 videos, with each video containing 60 1 s signals), resulting in a total of 76,800 input samples. Consistent with previous research in this field, our specific focus lies in investigating arousal and valence for emotion recognition purposes. More specifically, scores equal to or above 5 indicate high arousal or high valence levels, while scores below 5 indicate low arousal or low valence levels.
SEED [36]: The SEED dataset consists of EEG signals recorded from 15 subjects while they observed 15 videos, each with an approximate duration of 4 min. Prior to the experiments, the videos were pre-classified as negative, neutral, or positive. The participants underwent three separate sessions with a one-week interval between them. Within a timeframe of 45 s after watching each video, the subjects rated their emotional responses. The collected EEG signals comprise a total of 62 channels and were downsampled to a frequency of 200 Hz before undergoing processing using a bandpass filter ranging from 0 Hz to 75 Hz. We implemented a data processing methodology similar to that employed in the DEAP dataset, resulting in 3394 1 s signals for each subject and yielding a cumulative total of 152,730 input samples (obtained from 15 subjects during 3 sessions).

4.1.2. Baselines

For the DEAP dataset, we compared our model with state-of-the-art emotion recognition models in valence, arousal dimensions and valence and arousal classification. These models include LResNet [37], an emotion recognition method based on a deep residual network; MMResLSTM [38], a multimodal residual LSTM network for emotion recognition; ERDL [17], a GCN and LSTM network for emotion recognition; ACRNN [39], an attention-based convolutional RNN model; GLFANet [40], a global and local feature aggregation net. We also compared it with two self-supervised emotion recognition frameworks: GANSER, which is based on Generative Adversarial Network [41], and SGMC, which is based on based on group meiosis contrastive learning [12] for EEG-based emotion recognition.
Similar to the DEAP dataset, we selected the five most advanced supervised models and two self-supervised models for comparison on the SEED dataset. These include BiHDM [42], a bi-hemispheric discrepancy model; ResNet18-1D [31], a residual network consisting of 17 convolutional layers and 1 fully connected layer; DGCNN [43], a GCN and LSTM network; RGNN [44], a regularized graph neural network, and CLISA [13], a contrastive Learning method for learning Inter-Subject Alignment. Additionally, GLFANet [40] and SGMC [12] models, which serve as baseline models in the DEAP dataset, are also incorporated.

4.1.3. Implementation Details

The experiment was conducted on the Geforce RTX 3090Ti GPU using the pytorch [45] framework. For optimization, we selected the Adam optimizer [46] with a learning rate of 0.001.The experimental setup we adopted bears a resemblance to that of SGMC [12]. During pre-training, we performed 3000 training sessions on two datasets with a batch size of 32. In the spatial contrastive learning framework, based on prior research experience [40], the calibration constant θ was set to 6 and was used to normalize the reciprocal distance between EEG channels to a range of (0,1). The hyperparameter δ was set to 0.1 in order to exclude channels exhibiting lower correlation. In the temporal contrastive learning framework, based on empirical evidence, we set the value of top-K to 20, indicating that we selected the frequencies corresponding to the top 20 largest amplitudes as representative samples after applying noise reduction. The hyperparameter τ for calculating contrast loss was set to 0.1. For fine-tuning, we used a batch size of 256 and trained for a total of 100 epochs. We employed a five-fold cross-validation methodology and calculated the average performance across all five validation iterations as our final experimental result.

4.2. Performance Comparison

We employed classification accuracy to evaluate the model’s performance, and the corresponding results for both datasets are presented in Table 1 and Table 2. Additionally, Figure 4 depicts the confusion matrices of the model. The experimental results demonstrated that the proposed model achieved recognition accuracies of 96.79% and 96.81% for the valence and arousal dimensions, respectively, on the DEAP dataset, surpassing the performance of previous state-of-the-art models by 1.55% and 1.06%. The accuracy of the valence and arousal classifications reached 94.96%, demonstrating an increase of 1.65%. Moreover, for the SEED dataset, our model achieved an accuracy improvement of 1.55%, reaching an accuracy rate of 96.43%. The primary reason is that the model effectively captures both the temporal and spatial features of EEG. When extracting temporal features, signal noise is efficiently eliminated through frequency domain noise reduction, and Granger causality between time segments is extracted for each signal, which has not been considered in previous studies. The experimental results also validate the efficacy of this approach. In terms of extracting spatial features, the Granger causality test is employed to extract directed causality between brain electrical channels, thereby reflecting the information flow within the EEG signal in line with brain mechanisms. Simultaneously, leveraging a contrast learning framework enables better extraction of inherent characteristics specific to various emotions within the EEG signal and ultimately achieves the optimal emotion recognition performance.
Based on the confusion matrix results, our model has demonstrated excellent performance in accurately recognizing various emotions across both datasets, indicating the effectiveness of the contrast learning framework in extracting EEG features associated with diverse emotional states.

4.3. Ablation Study

In order to investigate the rationality of the model, a series of ablation experiments were conducted. Firstly, we individually eliminated the space feature extraction module and the time feature extraction module to validate their respective functions in this paper’s model. Additionally, to further examine the impact of each component in the model, we also removed the frequency noise reduction, Granger–Former, and Granger graph augmentation modules. When the frequency noise reduction module was omitted, we employed the data mask method to enhance the temporal dimension data; when excluding Granger–Former, we utilized a conventional transformer as the encoder; and when disregarding Granger graph augmentation, we applied a random edge dropping method to augment spatial dimension data, as depicted in Table 3. The experiment was conducted on two datasets, and the corresponding results are presented in Table 4.
The removal of the temporal feature extraction module led to decreases in the performance on the DEAP and SEED datasets of approximately 6.46% and 8.24%, respectively, while the removal of the spatial feature extraction module resulted in decreases of about 7.83% and 8.24%. These findings provide evidence of the crucial role played by simultaneous extraction of both temporal and spatial features in enhancing the emotion recognition efficacy of EEG signals.
When employing data masking instead of a frequency noise reduction module, the accuracy experiences reductions of approximately 4.51% and 4.44%, substantiating the efficacy of the proposed method in eliminating noise from EEG signals and enhancing emotion recognition accuracy, thus establishing it as an effective approach for data augmentation. The utilization of a transformer instead of Granger–Former results in a reduction in accuracy by approximately 4.64% and 4.42%, thereby substantiating the efficacy of Granger–Former in effectively extracting pertinent features between time segments of EEG signals, a facet that has been overlooked in prior studies on emotion recognition. When random edge deletion was employed instead of Granger causal graph augmentation, the accuracy exhibited decreases of approximately 3.82% and 3.44%. This decline may be attributed to the effectiveness of the Granger causality test in capturing the directional flow of information in EEG signals during spatial information extraction, aligning with brain mechanisms and facilitating more accurate extraction of spatial features between brain electrical channels, thereby enhancing recognition accuracy.

5. Discussion

5.1. Granger Causality Analysis

The Granger causality test is employed during the process of spatial feature extraction to augment the initial symmetric adjacency matrix into an asymmetric matrix, thereby capturing the directional information flow within EEG signals. Taking the DEAP dataset as an example, a subject was randomly selected, and the EEG signals of two emotions (high arousal and high valence, low arousal, and low valence) were recorded after enhancement. The adjacency matrix heat map is presented in Figure 5. The adjacency matrix of high-arousal and high-valence emotions exhibits a greater number of connections compared to that of low-arousal and low-valence emotions, implying a higher level of complexity in information transmission within the former.
We also compute the disparity between the out and in degrees of each channel in these two emotions. When the positive difference of the channel is observed, it indicates that the channel serves as an input for information transmission, and vice versa. The top five channels for both information input and output are illustrated in Figure 6.
The color orange represents the outbound channel of information flow, while green signifies the inbound channel for information transmission. As depicted, the frontal and temporal lobes are closely related to emotions, and the flow of information is directional in different emotions, which aligns with previous research findings [47,48,49].

5.2. Noise Disturbance Experiment

In order to further validate the efficacy of the proposed frequency domain noise reduction module, we conducted a noise disturbance experiment on the DEAP dataset. Specifically, we applied noise perturbations of 1%, 5%, 10%, and 20% to the input signal during the time contrast learning stage, respectively. Subsequently, experiments were conducted on the DEAP and SEED datasets, and the corresponding results are illustrated in Figure 7.
The results demonstrate that the inclusion of a 20% noise disturbance in the data results in a reduction of merely 3.47% and 4.13% on the two datasets, thereby indicating the efficacy of frequency domain noise reduction approach in effectively eliminating EEG signal interference and consequently enhancing emotion recognition accuracy.

6. Conclusions

In this paper, we proposed a Granger causal-based data augmentation graph contrastive learning framework for emotion recognition. This framework constructs the graph structure as the input to use the relationship between the channels of the EEG signals. To reduce the influence of noise on emotion recognition while maintaining the graph structure, we chose the Granger causality test to enhance the input data, and then we trained a residual graph convolutional network to extract the features of the EEG signals.
To verify the validity of the proposed method, we conducted experimental tests on the DEAP and SEED datasets. The accuracies of the DEAP dataset for valence and arousal are 96.79% and 96.81% and the accuracy of SEED dataset is 96.43%. The experimental results show that the framework achieved state-of-the-art results for both datasets. We also performed ablation experiments to verify the validity of the various components of the framework.
Despite the commendable outcomes achieved using both datasets, the proposed method still exhibits certain noteworthy limitations. Firstly, it is important to acknowledge that the current public datasets are relatively small in size, encompassing a maximum of 32 subjects (DEAP). Consequently, this constraint may impede further advancements in the performance of this study. Furthermore, within the proposed contrastive learning framework in this paper, Granger–Former exhibits relatively high computational complexity, while spatial contrastive learning requires a large number of model iterations to ensure comprehensive sample training, resulting in increased computational costs. Finally, given the substantial inter-individual differences observed in EEG signals, it is imperative to validate the model’s performance across different subjects. These challenges collectively pose obstacles to practical implementation of the model.
In future research, it is imperative to construct a comprehensive dataset encompassing a larger number and diverse range of subjects in order to analyze the EEG signal characteristics across different emotional states among various populations. Furthermore, optimizing the model to reduce computational complexity, enhancing its generalization ability, and conducting subject-independent verification are crucial directions for future investigations.

Author Contributions

Conceptualization, L.W.; methodology, L.W. and S.W.; software, L.W. and S.W.; validation, L.W.; formal analysis, L.W. and S.W.; investigation, B.J. and X.W.; data curation, L.W.; writing—original draft preparation, L.W.; writing—review and editing, L.W. and S.W.; visualization, L.W.; supervision, B.J. and X.W.; project administration, B.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Natural Science Foundation of China (No. 62172074), Program of Introducing Talents of Discipline to Universities (Plan 111) (No. B20070).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Two publicly available datasets were analyzed in this study. These data can be found at the following address: http://www.eecs.qmul.ac.uk/mmv/datasets/deap/index.html (accessed on 20 July 2022), and https://bcmi.sjtu.edu.cn/home/seed/ (accessed on 16 March 2021).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Prabowo, D.W.; Nugroho, H.A.; Setiawan, N.A.; Debayle, J. A systematic literature review of emotion recognition using EEG signals. Cogn. Syst. Res. 2023, 82, 101152. [Google Scholar] [CrossRef]
  2. Zhao, J.; Mao, X.; Chen, L. Speech emotion recognition using deep 1D & 2D CNN LSTM networks. Biomed. Signal Process. Control 2019, 47, 312–323. [Google Scholar]
  3. Huang, X.; Wang, S.J.; Liu, X.; Zhao, G.; Feng, X.; Pietikäinen, M. Discriminative spatiotemporal local binary pattern with revisited integral projection for spontaneous facial micro-expression recognition. IEEE Trans. Affect. Comput. 2017, 10, 32–47. [Google Scholar] [CrossRef]
  4. Rahman, M.M.; Sarkar, A.K.; Hossain, M.A.; Hossain, M.S.; Islam, M.R.; Hossain, M.B.; Quinn, J.M.; Moni, M.A. Recognition of human emotions using EEG signals: A review. Comput. Biol. Med. 2021, 136, 104696. [Google Scholar] [CrossRef] [PubMed]
  5. Nita, S.; Bitam, S.; Heidet, M.; Mellouk, A. A new data augmentation convolutional neural network for human emotion recognition based on ECG signals. Biomed. Signal Process. Control 2022, 75, 103580. [Google Scholar] [CrossRef]
  6. Perdiz, J.; Pires, G.; Nunes, U.J. Emotional state detection based on EMG and EOG biosignals: A short survey. In Proceedings of the 2017 IEEE 5th Portuguese Meeting on Bioengineering (ENBENG), Coimbra, Portugal, 16–18 February 2017; IEEE: New York, NY, USA, 2017; pp. 1–4. [Google Scholar]
  7. Jang, S.; Moon, S.E.; Lee, J.S. EEG-based video identification using graph signal modeling and graph convolutional neural network. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; IEEE: New York, NY, USA, 2018; pp. 3066–3070. [Google Scholar]
  8. Wang, Z.; Tong, Y.; Heng, X. Phase-locking value based graph convolutional neural networks for emotion recognition. IEEE Access 2019, 7, 93711–93722. [Google Scholar] [CrossRef]
  9. Hou, Y.; Jia, S.; Lun, X.; Hao, Z.; Shi, Y.; Li, Y.; Zeng, R.; Lv, J. GCNs-net: A graph convolutional neural network approach for decoding time-resolved eeg motor imagery signals. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 7312–7323. [Google Scholar] [CrossRef] [PubMed]
  10. Reid, A.T.; Headley, D.B.; Mill, R.D.; Sanchez-Romero, R.; Uddin, L.Q.; Marinazzo, D.; Lurie, D.J.; Valdés-Sosa, P.A.; Hanson, S.J.; Biswal, B.B.; et al. Advancing functional connectivity research from association to causation. Nat. Neurosci. 2019, 22, 1751–1760. [Google Scholar] [CrossRef] [PubMed]
  11. Herrmann, C.S.; Strüber, D.; Helfrich, R.F.; Engel, A.K. EEG oscillations: From correlation to causality. Int. J. Psychophysiol. 2016, 103, 12–21. [Google Scholar] [CrossRef] [PubMed]
  12. Kan, H.; Yu, J.; Huang, J.; Liu, Z.; Wang, H.; Zhou, H. Self-supervised group meiosis contrastive learning for eeg-based emotion recognition. Appl. Intell. 2023, 53, 27207–27225. [Google Scholar] [CrossRef]
  13. Shen, X.; Liu, X.; Hu, X.; Zhang, D.; Song, S. Contrastive learning of subject-invariant EEG representations for cross-subject emotion recognition. IEEE Trans. Affect. Comput. 2022, 14, 2496–2511. [Google Scholar] [CrossRef]
  14. Zhang, T.; Zheng, W.; Cui, Z.; Zong, Y.; Li, Y. Spatial–temporal recurrent neural network for emotion recognition. IEEE Trans. Cybern. 2018, 49, 839–847. [Google Scholar] [CrossRef] [PubMed]
  15. Li, D.; Xie, L.; Chai, B.; Wang, Z.; Yang, H. Spatial-frequency convolutional self-attention network for EEG emotion recognition. Appl. Soft Comput. 2022, 122, 108740. [Google Scholar] [CrossRef]
  16. Chen, J.; Jiang, D.; Zhang, Y. A hierarchical bidirectional GRU model with attention for EEG-based emotion classification. IEEE Access 2019, 7, 118530–118540. [Google Scholar] [CrossRef]
  17. Yin, Y.; Zheng, X.; Hu, B.; Zhang, Y.; Cui, X. EEG emotion recognition using fusion model of graph convolutional neural networks and LSTM. Appl. Soft Comput. 2021, 100, 106954. [Google Scholar] [CrossRef]
  18. Du, G.; Su, J.; Zhang, L.; Su, K.; Wang, X.; Teng, S.; Liu, P.X. A multi-dimensional graph convolution network for EEG emotion recognition. IEEE Trans. Instrum. Meas. 2022, 71, 2518311. [Google Scholar] [CrossRef]
  19. Lin, X.; Chen, J.; Ma, W.; Tang, W.; Wang, Y. EEG emotion recognition using improved graph neural network with channel selection. Comput. Methods Programs Biomed. 2023, 231, 107380. [Google Scholar] [CrossRef] [PubMed]
  20. Feng, L.; Cheng, C.; Zhao, M.; Deng, H.; Zhang, Y. EEG-based emotion recognition using spatial-temporal graph convolutional LSTM with attention mechanism. IEEE J. Biomed. Health Inform. 2022, 26, 5406–5417. [Google Scholar] [CrossRef] [PubMed]
  21. Hesse, W.; Möller, E.; Arnold, M.; Schack, B. The use of time-variant EEG Granger causality for inspecting directed interdependencies of neural assemblies. J. Neurosci. Methods 2003, 124, 27–44. [Google Scholar] [CrossRef] [PubMed]
  22. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; PMLR: New York, NY, USA, 2020; pp. 1597–1607. [Google Scholar]
  23. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  24. Liu, X.; Luo, Y.; Li, P.; Song, S.; Peng, J. Deep geometric representations for modeling effects of mutations on protein-protein binding affinity. PLoS Comput. Biol. 2021, 17, e1009284. [Google Scholar] [CrossRef] [PubMed]
  25. Mohsenvand, M.N.; Izadi, M.R.; Maes, P. Contrastive representation learning for electroencephalogram classification. In Proceedings of the Machine Learning for Health, Virtual, 11 December 2020; PMLR: New York, NY, USA, 2020; pp. 238–253. [Google Scholar]
  26. Li, C.; Lin, X.; Liu, Y.; Song, R.; Cheng, J.; Chen, X. EEG-based emotion recognition via efficient convolutional neural network and contrastive learning. IEEE Sens. J. 2022, 22, 19608–19619. [Google Scholar] [CrossRef]
  27. Guo, Y.; Zhang, T.; Huang, W. Emotion Recognition based on Physiological Signals Multi-head Attention Contrastive Learning. In Proceedings of the 2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Istanbul, Turkiye, 5–8 December 2023; IEEE: New York, NY, USA, 2023; pp. 1929–1934. [Google Scholar]
  28. Salvador, R.; Suckling, J.; Coleman, M.R.; Pickard, J.D.; Menon, D.; Bullmore, E. Neurophysiological architecture of functional magnetic resonance images of human brain. Cereb. Cortex 2005, 15, 1332–1342. [Google Scholar] [CrossRef] [PubMed]
  29. Granger, C.W. Investigating causal relations by econometric models and cross-spectral methods. Econom. J. Econom. Soc. 1969, 37, 424–438. [Google Scholar] [CrossRef]
  30. Li, G.; Muller, M.; Thabet, A.; Ghanem, B. Deepgcns: Can gcns go as deep as cnns? In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9267–9276. [Google Scholar]
  31. Cheah, K.H.; Nisar, H.; Yap, V.V.; Lee, C.Y.; Sinha, G. Optimizing residual networks and vgg for classification of eeg signals: Identifying ideal channels for emotion recognition. J. Healthc. Eng. 2021, 2021, 5599615. [Google Scholar] [CrossRef] [PubMed]
  32. You, Y.; Chen, T.; Sui, Y.; Chen, T.; Wang, Z.; Shen, Y. Graph contrastive learning with augmentations. Adv. Neural Inf. Process. Syst. 2020, 33, 5812–5823. [Google Scholar]
  33. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9729–9738. [Google Scholar]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  35. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. Deap: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef]
  36. Zheng, W.L.; Lu, B.L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  37. Liu, N.; Fang, Y.; Li, L.; Hou, L.; Yang, F.; Guo, Y. Multiple feature fusion for automatic emotion recognition using EEG signals. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; IEEE: New York, NY, USA, 2018; pp. 896–900. [Google Scholar]
  38. Ma, J.; Tang, H.; Zheng, W.L.; Lu, B.L. Emotion recognition using multimodal residual LSTM network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 176–183. [Google Scholar]
  39. Tao, W.; Li, C.; Song, R.; Cheng, J.; Liu, Y.; Wan, F.; Chen, X. EEG-based emotion recognition via channel-wise attention and self attention. IEEE Trans. Affect. Comput. 2020, 14, 382–393. [Google Scholar] [CrossRef]
  40. Liu, S.; Zhao, Y.; An, Y.; Zhao, J.; Wang, S.H.; Yan, J. GLFANet: A global to local feature aggregation network for EEG emotion recognition. Biomed. Signal Process. Control 2023, 85, 104799. [Google Scholar] [CrossRef]
  41. Zhang, Z.; Zhong, S.h.; Liu, Y. GANSER: A self-supervised data augmentation framework for EEG-based emotion recognition. IEEE Trans. Affect. Comput. 2022, 14, 2048–2063. [Google Scholar] [CrossRef]
  42. Li, Y.; Wang, L.; Zheng, W.; Zong, Y.; Qi, L.; Cui, Z.; Zhang, T.; Song, T. A novel bi-hemispheric discrepancy model for EEG emotion recognition. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 354–367. [Google Scholar] [CrossRef]
  43. Song, T.; Zheng, W.; Song, P.; Cui, Z. EEG emotion recognition using dynamical graph convolutional neural networks. IEEE Trans. Affect. Comput. 2018, 11, 532–541. [Google Scholar] [CrossRef]
  44. Zhong, P.; Wang, D.; Miao, C. EEG-based emotion recognition using regularized graph neural networks. IEEE Trans. Affect. Comput. 2020, 13, 1290–1301. [Google Scholar] [CrossRef]
  45. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  46. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  47. Zheng, W.L.; Zhu, J.Y.; Lu, B.L. Identifying stable patterns over time for emotion recognition from EEG. IEEE Trans. Affect. Comput. 2017, 10, 417–429. [Google Scholar] [CrossRef]
  48. Bogert, B.; Numminen-Kontti, T.; Gold, B.; Sams, M.; Numminen, J.; Burunat, I.; Lampinen, J.; Brattico, E. Hidden sources of joy, fear, and sadness: Explicit versus implicit neural processing of musical emotions. Neuropsychologia 2016, 89, 393–402. [Google Scholar] [CrossRef] [PubMed]
  49. Wang, H.Y.; Xu, G.Q.; Ni, M.F.; Zhang, C.H.; Sun, X.P.; Chang, Y.; Zhang, B.W. Neural mechanisms of implicit cognitive reappraisal: Preceding descriptions alter emotional response to unpleasant images. Neuroscience 2017, 347, 65–75. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overall structure of GC-STCL. The model comprises a pre-training process and a fine-tuning process. During the pre-training process, we employ the time contrast learning framework and the space contrast learning framework to extract temporal and spatial features from a small batch of input data, ultimately obtaining the feature representation of the input sample. In the fine-tuning process, classifiers are utilized to refine the pre-trained time and space encoders while conducting emotion classification training.
Figure 1. The overall structure of GC-STCL. The model comprises a pre-training process and a fine-tuning process. During the pre-training process, we employ the time contrast learning framework and the space contrast learning framework to extract temporal and spatial features from a small batch of input data, ultimately obtaining the feature representation of the input sample. In the fine-tuning process, classifiers are utilized to refine the pre-trained time and space encoders while conducting emotion classification training.
Entropy 26 00540 g001
Figure 2. The architectures of spatial encoder.
Figure 2. The architectures of spatial encoder.
Entropy 26 00540 g002
Figure 3. The architectures of Granger attention.
Figure 3. The architectures of Granger attention.
Entropy 26 00540 g003
Figure 4. The confusion matrix of classification on DEAP and SEED. (a) Valence on DEAP; (b) arousal on DEAP; (c) valence and arousal classifications on DEAP; (d) SEED.
Figure 4. The confusion matrix of classification on DEAP and SEED. (a) Valence on DEAP; (b) arousal on DEAP; (c) valence and arousal classifications on DEAP; (d) SEED.
Entropy 26 00540 g004
Figure 5. Heat map of the adjacency matrix before and after Granger causality augmentation. (a) Adjacency matrix without Granger causal augmentation; (b) augmented adjacency matrices incorporating Granger causality for emotions characterized by high arousal and high valence; (c) augmented adjacency matrices incorporating Granger causality for emotions characterized by low arousal and low valence.
Figure 5. Heat map of the adjacency matrix before and after Granger causality augmentation. (a) Adjacency matrix without Granger causal augmentation; (b) augmented adjacency matrices incorporating Granger causality for emotions characterized by high arousal and high valence; (c) augmented adjacency matrices incorporating Granger causality for emotions characterized by low arousal and low valence.
Entropy 26 00540 g005
Figure 6. The channels exhibiting higher levels of activity in both emotional responses. (a) High arousal and high valence; (b) low arousal and low valence.
Figure 6. The channels exhibiting higher levels of activity in both emotional responses. (a) High arousal and high valence; (b) low arousal and low valence.
Entropy 26 00540 g006
Figure 7. The results of the noise disturbance experiment.
Figure 7. The results of the noise disturbance experiment.
Entropy 26 00540 g007
Table 1. Performances on DEAP.
Table 1. Performances on DEAP.
MethodAccuracy (%)
ValenceArousalValence & Arousal
LResNet (2018)90.3989.06-
MMResLstm (2019)92.8792.30-
ERDL (2021)90.4590.60-
ACRNN (2020)93.7293.38-
GLFANet (2023)94.5394.9192.92
GANSER (2022)93.5294.2189.74
SGMC (2023)95.3195.7993.42
Proposed96.7996.8194.96
Table 2. Performances on SEED.
Table 2. Performances on SEED.
MethodAccuracy (%)
BiHDM (2020)93.12
ResNet18-1D (2021)93.42
DGCNN (2018)90.40
GLFANet (2023)93.19
RGNN (2020)94.24
CLISA (2022)86.4
SGMC (2023)94.96
Proposed96.43
Table 3. The design of the ablation study.
Table 3. The design of the ablation study.
MethodTemporalSpatialNoise ReductionGranger–FormerGraph Augmentation
GC-STCL-temporal×××
GC-STCL-spatial××
GC-STCL-noise×
GC-STCL-gf×
GC-STCL-ga×
Table 4. The results of the ablation study.
Table 4. The results of the ablation study.
MethodValence and Arousal Classifications Accuracy (%) on DEAPAccuracy (%) on SEED
GC-STCL-temporal88.83/4.1889.76/4.57
GC-STCL-spatial87.52/3.4588.48/4.10
GC-STCL-noise90.68/3.3892.15/4.37
GC-STCL-gf90.55/3.2992.16/3.39
GC-STCL-ga91.33/3.2393.11/3.30
GC-STCL94.96/2.7096.43/2.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Wang, S.; Jin, B.; Wei, X. GC-STCL: A Granger Causality-Based Spatial–Temporal Contrastive Learning Framework for EEG Emotion Recognition. Entropy 2024, 26, 540. https://doi.org/10.3390/e26070540

AMA Style

Wang L, Wang S, Jin B, Wei X. GC-STCL: A Granger Causality-Based Spatial–Temporal Contrastive Learning Framework for EEG Emotion Recognition. Entropy. 2024; 26(7):540. https://doi.org/10.3390/e26070540

Chicago/Turabian Style

Wang, Lei, Siming Wang, Bo Jin, and Xiaopeng Wei. 2024. "GC-STCL: A Granger Causality-Based Spatial–Temporal Contrastive Learning Framework for EEG Emotion Recognition" Entropy 26, no. 7: 540. https://doi.org/10.3390/e26070540

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop