You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

23 October 2025

STCCA: Spatial–Temporal Coupled Cross-Attention Through Hierarchical Network for EEG-Based Speech Recognition

,
,
and
1
School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
Beijing BigData Center, Beijing 101100, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue EEG Signal Processing Techniques and Applications—3rd Edition

Abstract

Speech recognition based on Electroencephalogram (EEG) has attracted considerable attention due to its potential in communication and rehabilitation. Existing methods typically process spatial and temporal features with sequential, parallel, or constrained feature fusion strategies. However, the intricate cross-relationships between spatial and temporal features remain underexplored. To address these limitations, we propose a spatial–temporal coupled cross-attention mechanism through a hierarchical network, named STCCA. The proposed STCCA consists of three key components: local feature extraction module (LFEM), coupled cross-attention (CCA) fusion module, and global feature extraction module (GFEM). The LFEM employs CNNs to extract local temporal and spatial features, while the CCA fusion module leverages a dual-directional attention mechanism to establish deep interactions between temporal and spatial features. The GFEM uses multi-head self-attention layers to model long-range dependencies and extract global features comprehensively. STCCA is validated on three EEG-based speech datasets, achieving accuracies of 45.45%, 25.91%, and 29.07%, corresponding to improvements of 1.95%, 3.98%, and 1.98% over the comparison models.

1. Introduction

Language impairment caused by neurological damage is a severe disability that limits both work and social life [,]. The recognition of speech neural signals based on Electroencephalogram (EEG) offers an efficient communication and rehabilitation method for patients with language impairments [,,]. As a widely used non-invasive technique for monitoring brain activity, EEG is characterized by its high temporal resolution and rich spatial–temporal features [,,]. EEG-based speech recognition, encompassing inner speech, silent speech, and imagined speech recognition [,], holds significant potential for language impairments by decoding the intricate cross-relationships between spatial–temporal features [,].
Recently, various EEG-based speech recognition methods have been proposed, with deep learning (DL) methods taking a leading role, starting with convolutional neural networks (CNNs) [,]. CNNs have demonstrated exceptional capability in extracting local features [], leading to their widespread adoption in EEG analysis. One of the most representative architectures is EEGNet [,], which proposes multiple convolution layer architectures to sequentially extract temporal and spatial representations from EEG signals. EEGNet has demonstrated notable performance across multiple EEG decoding tasks and remains a widely adopted baseline in the field. In addition to EEGNet, other representative models have emerged to enhance spatial–temporal representation learning. For instance, TSception [] applies multi-scale temporal convolutions to capture temporal features, followed by asymmetric spatial convolutions to extract spatial representations. Despite these advancements, existing methods often treat parallel temporal and spatial features as independent entities, neglecting their inherent interconnections. To establish joint spatial–temporal relationships, several studies, including those by Chang et al. [] and Chen et al. [], have typically employed CNNs, dynamic connectogram to extract spatial and temporal features separately, and combined the parallel strategies such as feature summation, concatenation, or weighted averaging to fuse spatial–temporal features. In practice, the spatial locations of electrodes and the temporal evolution of EEG signals are tightly coupled due to the dynamic nature of cognitive and language-related brain activity. These universal approaches are inherently limited in capturing the intricate interactions and deep dependencies between spatial and temporal dimensions. In summary, existing approaches remain challenging due to the prevalent adoption of sequential, parallel, or constrained feature fusion strategies for processing temporal and spatial characteristics. Moreover, CNNs inherently operate with fixed-size receptive fields, which restrict their capacity to capture global context and long-range dependencies in EEG signals [,,].
To address the aforementioned challenges, we propose a spatial–temporal coupled cross-attention mechanism within a hierarchical network, named STCCA, which is designed to extract and fuse local spatial–temporal features, and subsequently model long-range dependencies. The framework consists of three key components. First, the local feature extraction module (LFEM) leverages temporal and spatial convolutional layers to independently extract fine-grained temporal and spatial features from EEG signals, enabling effective representation of both temporal dynamics and spatial distributions. Second, the extracted features are fed into the coupled cross-attention (CCA) fusion module, where a bidirectional attention mechanism is employed to establish intricate joint relationships between temporal and spatial dimensions. The CCA facilitates the capture of joint spatial–temporal dependencies, addressing the limitations of traditional feature fusion approaches. Finally, the fused features are passed to the global feature extraction module (GFEM), which utilizes multi-head self-attention layers to model long-range dependencies, facilitating the representation of global spatial–temporal patterns. The hierarchically processed features are passed to the fully connected layers to output the recognition results. The main contributions of this letter can be summarized as follows:
  • We propose a novel CCA module, which applies attention mechanisms bidirectionally to two input features. To the best of our knowledge, this is the first CCA fusion method designed to fuse temporal and spatial features in EEG-based speech recognition.
  • We propose an innovative STCCA network that hierarchically captures local and global characteristics of EEG signals. The LFEM is employed to extract local temporal and spatial features, the CCA module fuses these features by capturing their coupled interactions, and the GFEM extracts global dependencies.
  • STCCA achieves accuracy improvements of 1.95%, 3.98%, and 1.98% on three EEG-based speech datasets with 22 subjects. Ablation experiments further validate the superior performance of the CCA compared to other commonly used fusion methods, and highlight the essential roles of LFEM and GFEM in enhancing overall performance.
The remainder of this article is organized as follows. Section 2 describes the proposed method, including LFEM, CCA, and GFEM. Section 3 presents the dataset preprocess, experiment settings, results and analyses, and ablation studies. Finally, Section 4 presents the conclusion and future work directions.

2. Method

2.1. Problem Definition

The research goal of EEG-based speech recognition is to train a brain-to-speech decoding network f : x p , where the output p R m represents the coded representation of EEG signal across m target speech categories.
A set of EEG signal acquisition nodes V = v 1 , v 2 , , v c h is spatially distributed over the scalp, where c h = V denotes the number of electrode channels. Each node samples EEG signals at a frequency f s . The collected EEG signals are represented as x R c h × t , where t denotes the number of sampling time points. Each EEG segment is annotated with a ground-truth label y 1 , 2 , , m , corresponding to a specific speech class. Fianlly, the extracted features are calculated into logits, resulting in predictions y ^ 1 , 2 , , m .

2.2. Overview of STCCA

The overall architecture of the proposed STCCA framework is illustrated in Figure 1, and its complete processing pipeline is outlined in Algorithm 1 using pseudo-code. The hierarchical design aims to first extract and fuse local spatial–temporal features, followed by modeling long-range dependencies for enhanced representation. The proposed method consists of three key components:
Figure 1. Architecture of the proposed STCCA.
(1)
LFEM. LFEM extracts local fine-grained temporal and spatial features from EEG signals through convolutional layers along the time and channel dimensions, enabling effective representation of both temporal dynamics and spatial distributions.
(2)
CCA. CCA leverages a bidirectional attention mechanism to dynamically assign weights, enabling the modeling of complex interactions between temporal and spatial features. By capturing joint spatial–temporal dependencies, CCA effectively overcomes the limitations of traditional feature fusion approaches.
(3)
GFEM. GFEM employs multi-head self-attention layers to model long-range dependencies, thereby enhancing the representation of global spatial–temporal patterns. The final classification results are then obtained through the fully connected layers.
Algorithm 1: Pseudo-code of STCCA
Sensors 25 06541 i001

2.3. Local Feature Extraction Module

In view of the fact that neural activity is continuous, extracting local features from EEG signals plays a crucial role in improving EEG decoding performance. In the LFEM, two separate convolutional pathways are designed to capture both temporal and spatial local characteristics of the input EEG signals, as shown in Figure 1. Each convolutional pathway consists of convolutional layers (Conv), average pooling layers (AG), batch normalization layers (BN), and ELU activation functions. In the temporal convolutional pathway, temporal–spatial joint features ( f t ) are calculated by Equation (1) along the time and channel dimensions. Similarly, in the spatial convolutional pathway, spatial–temporal joint features ( f s ) are extracted based on Equation (2) along the channel and time dimensions. The parameters of LFEM are shown in Table 1.
f t = AG ϕ BN Conv 2 Conv 1 ( x ) ,
f s = AG ϕ BN Conv 4 Conv 3 ( x ) ,
where x is the input EEG signal, and ϕ is the activation function (ELU).
Table 1. Parameters of the LFEM.

2.4. Coupled Cross-Attention Fusion Module

The f t and f s obtained from the LFEM are fed into the CCA fusion module. This module utilizes a bidirectional attention mechanism to explore the intrinsic relationships between the temporal and spatial dimensions, dynamically allocating weights to the spatial–temporal features, thereby achieving a comprehensive fusion of spatial–temporal complementary features. Specifically, CCA enables temporal query to focus on spatial patterns and spatial query to capture temporal information, enhancing the learning of context-aware interactions between spatial and temporal representations. The entire process is illustrated in Figure 1.
In the first direction of attention, the f s serves as the key and value, respectively, and the f t is designated as the query. The basic Scaled Dot-Product Attention [] is applied, defined as
Atten ( Q , K , V ) = softmax ( Q K T d k ) V ,
where Q, K, and V represent the query, key, and value, respectively, each with a shape of R N b × 512 × k . Here, N b is the batch size, 512 represents the sequence length, and k is the feature dimension after linear projection, which is equal to the output dimension of the convolutional layer in Table 1. d k is the size of the query. In this direction, temporal features f t act as query to focus on spatial patterns f s , enabling the model to selectively emphasize spatial components that are most relevant at each time step. The formula is defined as follows:
f t s = Atten ( f t , f s , f s ) ,
In the second direction of attention, f s acts as the query, while f t is used as both the key and value, and the Equation (3) operation applies. This reverse attention allows spatial features to capture the temporal dynamics that is most informative for each spatial patterns, facilitating better temporal–contextual representation learning. The formula is defined as follows:
f s t = Atten ( f s , f t , f t ) ) ,
Finally, the f t s , f s t results from both directions are aggregated to produce the fused feature with learnable weight parameters, and layer normalization (LN) is applied to normalize the fused feature. The formula is defined as follows:
f fusion = LN ( α f t s + β f s t ) ,
where α and β are learnable unconstrained scalar weights for f t s and f s t , respectively.

2.5. Global Feature Extraction Module

Due to the fixed receptive field size, convolutional layers can only capture local temporal and spatial features, limiting their ability to extract global features from EEG signals. In the GFEM, we use multi-head self-attention (MHA) mechanism to learn global dependencies of EEG features, complementing the limited receptive field in the LFEM, as shown in Figure 1. The f fusion are divided into h segments. Each segment is linearly transformed into global query ( G Q ) , global key ( G K ), and global value ( G V ). The linear projection dimension is k. This process can be formulated as
f g = MHA ( G Q , G K , G V ) = [ h e a d 0 , , h e a d h 1 ] , h e a d l = Atten ( G Q l , G K l , G V l ) ,
where G Q l , G K l , G V l denote different G Q , G K , G V in the l t h head, respectively. The G Q l , G K l , G V l are processed by Equation (3), and all head outputs are concatenated.
y ^ = argmax ( FC ( f g ) ) ,
where y ^ is the predicted labels.

3. Experiment and Results

3.1. EEG Data and Preprocessing

In this study, three datasets are applied to evaluate the effectiveness of the proposed STCCA. The key characteristics of these datasets are summarized in Table 2. Appendix A.1 reports the number of trials per class per subject for each dataset to illustrate the data distribution in detail. A detailed description of each dataset is presented in the following paragraphs.
Table 2. The information about three datasets.
Dataset I: Nieto et al. [] developed an inner speech EEG dataset consisting of 10 native Spanish-speaking participants. EEG signals were collected using the 10–20 system with 128 EEG channels at a sampling rate of 1024 Hz. Participants were instructed to silently imagine four Spanish words: “arriba” (up), “abajo” (down), “derecha” (right), and “izquierda” (left). During preprocessing, the data were re-referenced by earlobe channels and subjected to band-pass filtering (0.5–100 Hz), along with a 50 Hz notch filter. The sampling rate was downsampled to 256 Hz. Independent Component Analysis (ICA) was applied to eliminate artifacts and ensure the integrity of the signal.
Dataset II: Zhao et al. [] collected EEG signals, facial expressions, and voice patterns while subjects silently imagined four words (“pat,” “pot,” “knew,” and “gnaw”) and seven syllables (“iy,” “uw,” “piy,” “tiy,” “diy,” “m,” and “n”), forming the Kara One dataset. The EEG signals were recorded at a sampling frequency of 1000 Hz using the 10–20 system with 62 EEG channels. In this study, EEG data from six subjects were randomly selected for analysis. No further filtering or artifact correction was applied to the publicly available dataset provided by the authors.
Dataset III: To complement the publicly available dataset, we collected a private inner speech EEG dataset with 32 EEG channels at a sampling rate of 1000 Hz. Subjects were instructed to silently imagine the ten most frequently used Chinese words, which were “我” (I), “你” (you), “他” (he), “上” (up), “下” (down), “左” (left), “右” (right), “是” (are), “水” (water), and “饭” (food). Each subject completed five sessions, with each session comprising 100 trials. As illustrated in Figure 2, each trial consists of a 2-s concentrate period, a 2-s Chinese words cue presentation, a 4-s silent action imagination period, and a final 2-s relaxation period. During preprocessing, the data were re-referenced using CPz channel and filtered with a band-pass filter ranging from 0.5 to 100 Hz, along with a notch filter at 50 Hz. The sampling rate was then downsampled to 250 Hz. ICA was employed to remove artifacts, ensuring signal quality. This study complied with the Declaration of Helsinki and was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of the Beijing University of Posts and Telecommunications (Ethic approval code: 202302003).
Figure 2. Trial workflow in each session. Chinese words “我”, “休息” correspond to “down”, “relax”, respectively.

3.2. Experimental Setting

The three datasets, consisting of Dataset I with four classes, Dataset II with eleven classes, and Dataset III with ten classes, are evaluated using five-fold cross-validation []. Specifically, all EEG trials of a subject are pooled and then randomly divided into five folds. The training-to-testing ratio in each iteration is 4:1, as shown in Figure 3. To ensure class balance within each fold, StratifiedKFold is applied. Accuracy is used as the evaluation metric, and the final accuracy is expressed as the mean ( μ ) ± standard deviation ( σ ), calculated as follows:
Accuracy = μ ± σ
μ = 1 5 i = 1 5 Accuracy i
σ = 1 5 i = 1 5 ( Accuracy i μ ) 2
where Accuracy i denotes the accuracy obtained in the i th iteration. The Wilcoxon Signed Rank Test is applyed to analyze final accuracy statistical significance.
Figure 3. Schematic diagram of five-fold cross-validation.
The implementation uses the PyTorch library (1.8.1) and runs in parallel on six NVIDIA 1080Ti GPUs. The input 2D data (channels × time samples) are standardized using Z-score normalization. Both ICA and Z-score normalization are estimated exclusively on the training data and then applied to the test data to avoid data leakage. In the proposed model, the output dimension of the convolutional layer is set to k = 32 , and the number of attention heads is set to h = 4 , which provides a balance between model complexity and representation capacity. The model is optimized using the Adam optimizer for 200 epochs, with a StepLR scheduler that reduces the learning rate by a factor of 0.5 every 50 epochs. A dropout rate of 0.3 is applied during the training process to prevent overfitting, and cross-entropy loss function is applied, defined as
L = 1 N b i = 1 N b c = 1 m y log ( y ^ ) ,
where m is the number of EEG categories, y and y ^ are the ground-truth and predicted labels, and N b is the batch size.

3.3. Experimental Results and Analyses

To evaluate the effectiveness of STCCA, we conducted extensive experiments on three benchmark datasets. The performance of STCCA was compared with ten baseline methods: ShallowConvNet [,], DeepConvNet [,], EEGNet [,], RACNN [], EEG-ChannelNet [], Conformer [], LMDA-Net [], AISR(SPWVD+CNN) [], EEG-Deformer [], and D-FaST []. As illustrated in Figure 4, the feature distributions become increasingly compact and separable with the progression of training epochs. The test results for each dataset, summarized in Table 3, Table 4 and Table 5, demonstrate the superior performance of the proposed STCCA model compared to the baseline methods. Our proposed STCCA achieves the highest average accuracies of 45.45 ± 6.48 % , 25.91 ± 4.84 % , and 29.07 ± 5.44 % on Dataset I, Dataset II, and Dataset III. The results demonstrate that STCCA significantly outperforms the baseline methods, with accuracy improvements of 1.95% (p < 0.05), 3.98% (p < 0.05), and 1.98% (p < 0.05), respectively. These baseline methods can be broadly categorized into CNN-based and attention-based methods. CNN-based methods excel at capturing local spatial or temporal patterns, whereas attention-based methods are more effective at modeling global dependencies. However, these methods generally fail to explicitly capture the coupling relationships between temporal and spatial features. For example, methods such as EEGNet treat temporal and spatial features independently, extracting them sequentially without considering their mutual interactions. Similarly, methods like D-Fast employ constrained fusion strategies—such as direct concatenation of features—that are limited in their ability to capture the intrinsic interactions and deep dependencies between temporal and spatial features. In contrast, STCCA integrates the complementary strengths of both CNNs and attention mechanisms within a hierarchical network. It not only extracts local and global features effectively but also models the complex and complementary relationships between temporal and spatial features. As a result, STCCA achieves the highest average accuracy across all datasets (p < 0.05). Figure 5 presents the confusion matrix for the classification of two randomly selected subjects. The confusion matrices of all ten subjects are provided in Appendix A.2. The diagonal elements represent the probabilities of accurate classification for each respective category. These results demonstrate that STCCA achieves superior recognition performance compared to other models in three EEG-based speech recognition for 18 subjects on three datasets.
Figure 4. T-SNE visualizations of feature distributions at different training epochs on Dataset I: (a) 3 epochs; (b) 50 epochs. Different colors represent different feature categories.
Table 3. Accuracy (%) across different methods on Dataset I. Bold values indicate the best performance in each subject.
Table 4. Accuracy (%) across different methods on Dataset II. Bold values indicate the best performance in each subject.
Table 5. Accuracy (%) across different methods on Dataset III. Bold values indicate the best performance in each subject.
Figure 5. Confusion matrices of (a) S09 and (b) S10 on Dataset I.
Furthermore, the F1_scores of STCCA are computed, as shown in Table 6, demonstrating stable classification performance across subjects. To further evaluate the sensitivity of STCCA, we conduct channel dropout experiments, where 5% and 10% of EEG channels are randomly removed. The results show a decline in accuracy under channel loss conditions, which can be attributed to the spatial–temporal fusion mechanism of STCCA, where dropping channels inevitably leads to the loss of spatial information. In addition, Table 7 compares the parameters, FLOPs, and inference time among several representative baseline models. Although STCCA introduces a moderate number of parameters (7.155 M) and computational cost (2581.882 MFLOPs), it achieves the fastest inference time (4.597 ms). Collectively, these results demonstrate that STCCA maintains competitive performance while remaining computationally efficient, making it suitable for EEG-based speech recognition.
Table 6. F1_score and accuracy(%) under channel dropout conditions (5% and 10%) on Dataset I.
Table 7. Summary of parameters, FLOPs, and inference efficiency across method. Bold values indicate the best performance in each column.
To further evaluate the cross-subject generalization ability of STCCA, we conduct a Leave-One-Subject-Out (LOSO) cross-validation, in which one subject is used as the test set while the remaining subjects are used for training. As shown in Table 8, the proposed STCCA consistently achieves the highest or near-highest accuracies across most subjects, with an overall improvement over the baselines. Specifically, STCCA attains the best performance for seven out of ten subjects (S01, S02, S04, S05, S06, S08, and S10), indicating its strong generalization capability when tested on unseen subjects. Compared with classical CNN-based architectures (ShallowConvNet, DeepConvNet, and EEGNet) and attention-based models (Deformer and D-FaST), STCCA exhibits more stable accuracy across subjects, demonstrating its ability to effectively capture invariant spatial–temporal representations from heterogeneous EEG patterns. These results validate that the proposed model maintains reliable performance under cross-subject conditions in EEG-based speech decoding.
Table 8. Accuracy (%) under the LOSO cross-subject evaluation on Dataset I. Bold values indicate the best performance in each subject.

3.4. Ablation Studies

We conduct ablation experiments to evaluate the contribution of each module in STCCA. As shown in Table 9, Table 10 and Table 11, removing either LFEM or GFEM results in a notable decline in performance, highlighting the critical contribution of both modules to the hierarchical architecture. To further validate the contribution of the CCA module, we replace it with three commonly used feature fusion strategies: summation (sum), concatenation (cat), and standard attention (atten). Standard attention computes a weighted combination of the temporal and spatial features by applying a linear projection, followed by softmax normalization to obtain attention weights that are then used to produce the fused feature. These commonly used feature fusion strategies treat temporal and spatial features independently, or fuse them statically. Each modified model is evaluated on the same datasets under identical experimental conditions, and the results are shown in Figure 6. The proposed CCA mechanism demonstrates superior performance, achieving the highest average accuracy (p < 0.05) across both datasets. In contrast, the sum, cat, and atten strategies yield lower accuracies, indicating their limited ability to capture the intricate relationships between temporal and spatial features. This highlights the distinct advantage of CCA in modeling bidirectional interactions between temporal and spatial features, thereby enhancing the spatial–temporal joint features. In addition to performance, we compare computational cost and inference efficiency, as shown in Table 12. All strategies have the same number of parameters, and while CCA has slightly higher FLOPs, its inference time remains comparable, indicating that the accuracy gain comes with minimal computational overhead. These results confirm the critical role of CCA in STCCA, emphasizing its role in driving performance improvements in EEG-based speech recognition.
Table 9. Ablation results of LFEM and GFEM on Dataset I. Bold values indicate the best performance in each subject.
Table 10. Ablation results of LFEM and GFEM on Dataset II. Bold values indicate the best performance in each subject.
Table 11. Ablation results of LFEM and GFEM on Dataset III. Bold values indicate the best performance in each subject.
Figure 6. Contrast results of ablation experiments on (a) Dataset I and (b) Dataset III.
Table 12. Summary of parameter counts, FLOPs, and inference efficiency across fusion strategies. Bold values indicate the best performance in each column.

4. Conclusions

In this study, we propose the STCCA network, a novel hierarchical framework designed to tackle the challenges of spatial–temporal feature fusion and local-global dependency modeling for EEG-based speech recognition. By integrating the CCA mechanism, the STCCA effectively captures the intricate bidirectional relationships between spatial and temporal features. Furthermore, the combination of convolutional layers and self-attention enables the network to extract both local and global features hierarchically, leading to a comprehensive feature representation. However, the current implementation is limited to offline analysis and has not yet been extended to real-time online systems, which restricts its applicability in practical scenarios. In addition, as the dataset scale increases, the model may face challenges in convergence due to increased data complexity, potentially requiring longer training and more careful parameter tuning to maintain stability and performance. Future research will aim to optimize the framework for real-time processing and deployment, while incorporating adaptive optimization strategies to improve convergence efficiency on larger EEG datasets, facilitating its integration into practical applications. This study highlights the potential of STCCA as a powerful tool for advancing EEG signal decoding in speech recognition, thus paving the way for more effective brain–computer interface applications.

Author Contributions

Conceptualization, L.L. and L.Z.; methodology, L.D.; software, L.D. and H.S.; validation, L.D.; formal analysis, L.D. and H.S.; investigation, H.S.; resources, L.L.; data curation, H.S. and L.D.; writing—original draft preparation, L.D.; writing—review and editing, H.S., L.L. and L.Z.; visualization, L.D.; supervision, L.L. and L.Z.; project administration, L.L.; funding acquisition, L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 62176024).

Institutional Review Board Statement

This study complied with the Declaration of Helsinki and was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of the Beijing University of Posts and Telecommunications (Ethic approval code: 202302003).

Data Availability Statement

Our code and model are available at https://github.com/buptantEEG/STCCA (accessed on 20 October 2025). The Dataset I presented in this study are available at https://openneuro.org/datasets/ds003626/versions/2.1.2 (accessed on 4 March 2025). The Dataset II presented in this study are available at http://www.cs.toronto.edu/~complingweb/data/karaOne/karaOne.html (accessed on 14 April 2025).

Acknowledgments

We sincerely thank the anonymous reviewers and associate editor for their constructive comments and feedback which significantly improved the technical quality of this manuscript.

Conflicts of Interest

Author Lin Zhang was employed by Beijing BigData Center. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
EEGelectroencephalogram
BCIbrain–computer interface
DLdeep learning
CNNconvolutional neural network
LFEMlocal feature extraction module
CCAcoupled cross-attention
GFEMglobal feature extraction module
Convconvolutional layers
AGaverage pooling layers
BNbatch normalization layers
LNlayer normalization
MHAmulti-head self-attention

Appendix A

Appendix A.1

Table A1. Number of trials per class per subject for Dataset I.
Table A1. Number of trials per class per subject for Dataset I.
 S01S02S03S04S05S06S07S08S09S10
Class0125150125150150135150125150150
Class1125150125150150135150125150150
Class2125150125150150135150125150150
Class3125150125150150135150125150150
Table A2. Number of trials per class per subject for Dataset II.
Table A2. Number of trials per class per subject for Dataset II.
 S01S02S03S04S05S06
Class0121212121512
Class1121212121512
Class2121212121512
Class3121212121512
Class4121212121512
Class5121212121512
Class6121212121512
Class7121212121512
Class8121212121512
Class9121212121512
Class10121212121512
Table A3. Number of trials per class per subject for Dataset III.
Table A3. Number of trials per class per subject for Dataset III.
 S01S02S03S04S05Sub6
Class0504050405040
Class1504050405040
Class2504050405040
Class3504050405040
Class4504050405040
Class5504050405040
Class6504050405040
Class7504050405040
Class8504050405040
Class9504050405040

Appendix A.2

Figure A1. Confusion matrices of S01–S06 (af) on Dataset I.
Figure A2. Confusion matrices of S07–S10 (ad) on Dataset I.

References

  1. Zhang, L.; Zhou, Y.; Gong, P.; Zhang, D. Speech imagery decoding using EEG signals and deep learning: A survey. IEEE Trans. Cogn. Dev. Syst. 2024, 17, 22–39. [Google Scholar] [CrossRef]
  2. Guetschel, P.; Ahmadi, S.; Tangermann, M. Review of deep representation learning techniques for brain–computer interfaces. J. Neural Eng. 2024, 21, 061002. [Google Scholar] [CrossRef]
  3. Musso, M.; Hübner, D.; Schwarzkopf, S.; Bernodusson, M.; LeVan, P.; Weiller, C.; Tangermann, M. Aphasia recovery by language training using a brain–computer interface: A proof-of-concept study. Brain Commun. 2022, 4, fcac008. [Google Scholar] [CrossRef]
  4. Lopez-Bernal, D.; Balderas, D.; Ponce, P.; Molina, A. Exploring inter-trial coherence for inner speech classification in EEG-based brain–computer interface. J. Neural Eng. 2024, 21, 026048. [Google Scholar] [CrossRef]
  5. Kamble, A.; Ghare, P.H.; Kumar, V. Optimized rational dilation wavelet transform for automatic imagined speech recognition. IEEE Trans. Instrum. Meas. 2023, 72, 4002210. [Google Scholar] [CrossRef]
  6. Rahman, N.; Khan, D.M.; Masroor, K.; Arshad, M.; Rafiq, A.; Fahim, S.M. Advances in brain-computer interface for decoding speech imagery from EEG signals: A systematic review. Cogn. Neurodyn. 2024, 18, 3565–3583. [Google Scholar] [CrossRef]
  7. Cai, Z.; Luo, T.j.; Cao, X. Multi-branch spatial-temporal-spectral convolutional neural networks for multi-task motor imagery EEG classification. Biomed. Signal Process. Control 2024, 93, 106156. [Google Scholar] [CrossRef]
  8. Li, X.; Tang, J.; Li, X.; Yang, Y. CWSTR-Net: A Channel-Weighted Spatial–Temporal Residual Network based on nonsmooth nonnegative matrix factorization for fatigue detection using EEG signals. Biomed. Signal Process. Control 2024, 97, 106685. [Google Scholar] [CrossRef]
  9. Nieto, N.; Peterson, V.; Rufiner, H.L.; Kamienkowski, J.E.; Spies, R. Thinking out loud, an open-access EEG-based BCI dataset for inner speech recognition. Sci. Data 2022, 9, 52. [Google Scholar] [CrossRef] [PubMed]
  10. Kamble, A.; Ghare, P.H.; Kumar, V. Deep-learning-based BCI for automatic imagined speech recognition using SPWVD. IEEE Trans. Instrum. Meas. 2022, 72, 4001110. [Google Scholar] [CrossRef]
  11. Li, C.; Wang, H.; Liu, Y.; Zhu, X.; Song, L. Silent EEG classification using cross-fusion adaptive graph convolution network for multilingual neurolinguistic signal decoding. Biomed. Signal Process. Control 2024, 87, 105524. [Google Scholar] [CrossRef]
  12. Li, C.; Liu, Y.; Li, J.; Miao, Y.; Liu, J.; Song, L. Decoding Bilingual EEG Signals With Complex Semantics Using Adaptive Graph Attention Convolutional Network. IEEE Trans. Neural Syst. Rehabil. Eng. 2024, 32, 249–258. [Google Scholar] [CrossRef]
  13. Liu, K.; Yang, M.; Yu, Z.; Wang, G.; Wu, W. FBMSNet: A filter-bank multi-scale convolutional neural network for EEG-based motor imagery decoding. IEEE Trans. Biomed. Eng. 2022, 70, 436–445. [Google Scholar] [CrossRef]
  14. Cao, L.; Yu, B.; Dong, Y.; Liu, T.; Li, J. Convolution spatial-temporal attention network for EEG emotion recognition. Physiol. Meas. 2024, 45, 125003. [Google Scholar] [CrossRef]
  15. Zhao, W.; Jiang, X.; Zhang, B.; Xiao, S.; Weng, S. CTNet: A convolutional transformer network for EEG-based motor imagery classification. Sci. Rep. 2024, 14, 20237. [Google Scholar] [CrossRef] [PubMed]
  16. Park, D.; Park, H.; Kim, S.; Choo, S.; Lee, S.; Nam, C.S.; Jung, J.Y. Spatio-temporal explanation of 3D-EEGNet for motor imagery EEG classification using permutation and saliency. IEEE Trans. Neural Syst. Rehabil. Eng. 2023. [Google Scholar] [CrossRef]
  17. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  18. Ding, Y.; Robinson, N.; Zhang, S.; Zeng, Q.; Guan, C. TSception: Capturing temporal dynamics and spatial asymmetry from EEG for emotion recognition. IEEE Trans. Affect. Comput. 2022, 14, 2238–2250. [Google Scholar] [CrossRef]
  19. Chang, Y.; Zheng, X.; Chen, Y.; Li, X.; Miao, Q. Spatiotemporal Gated Graph Transformer for EEG-based Emotion Recognition. IEEE Signal Process. Lett. 2024. [Google Scholar] [CrossRef]
  20. Chen, W.; Wang, C.; Xu, K.; Yuan, Y.; Bai, Y.; Zhang, D. D-FaST: Cognitive Signal Decoding with Disentangled Frequency-Spatial-Temporal Attention. IEEE Trans. Cogn. Dev. Syst. 2024. [Google Scholar] [CrossRef]
  21. Ding, Y.; Li, Y.; Sun, H.; Liu, R.; Tong, C.; Liu, C.; Zhou, X.; Guan, C. EEG-Deformer: A dense convolutional transformer for brain-computer interfaces. IEEE J. Biomed. Health Inform. 2024, 29, 1909–1918. [Google Scholar] [CrossRef]
  22. Song, Y.; Zheng, Q.; Liu, B.; Gao, X. EEG conformer: Convolutional transformer for EEG decoding and visualization. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 31, 710–719. [Google Scholar] [CrossRef]
  23. Zhang, X.; Cheng, X. A transformer convolutional network with the method of image segmentation for EEG-based emotion recognition. IEEE Signal Process. Lett. 2024, 31, 401–405. [Google Scholar] [CrossRef]
  24. Vaswani, A. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  25. Zhao, S.; Rudzicz, F. Classifying phonological categories in imagined and articulated speech. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, Australia, 19–24 April 2015; pp. 992–996. [Google Scholar]
  26. Wong, T.T.; Yeh, P.Y. Reliable Accuracy Estimates from k-Fold Cross Validation. IEEE Trans. Knowl. Data Eng. 2020, 32, 1586–1594. [Google Scholar] [CrossRef]
  27. Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Mapp. 2017, 38, 5391–5420. [Google Scholar] [CrossRef]
  28. Yang, Y.; Zhang, X.; Zhang, X.; Yu, C. MCMTNet: Advanced network architectures for EEG-based motor imagery classification. Neurocomputing 2024, 620, 129255. [Google Scholar] [CrossRef]
  29. Moctezuma, L.A.; Suzuki, Y.; Furuki, J.; Molinas, M.; Abe, T. GRU-powered sleep stage classification with permutation-based EEG channel selection. Sci. Rep. 2024, 14, 17952. [Google Scholar] [CrossRef]
  30. Fang, Z.; Wang, W.; Ren, S.; Wang, J.; Shi, W.; Liang, X.; Fan, C.C.; Hou, Z.G. Learning Regional Attention Convolutional Neural Network for Motion Intention Recognition Based on EEG Data. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, Yokohama, Japan, 11–17 July 2020; pp. 1570–1576. [Google Scholar]
  31. Palazzo, S.; Spampinato, C.; Kavasidis, I.; Giordano, D.; Schmidt, J.; Shah, M. Decoding brain representations by multimodal learning of neural activity and visual features. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3833–3849. [Google Scholar] [CrossRef] [PubMed]
  32. Miao, Z.; Zhao, M.; Zhang, X.; Ming, D. LMDA-Net: A lightweight multi-dimensional attention network for general EEG-based brain-computer interfaces and interpretability. NeuroImage 2023, 276, 120209. [Google Scholar] [CrossRef]
  33. Kamble, A.; Ghare, P.H.; Kumar, V.; Kothari, A.; Keskar, A.G. Spectral analysis of EEG signals for automatic imagined speech recognition. IEEE Trans. Instrum. Meas. 2023, 72, 4009409. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.