Next Article in Journal
Electroencephalogram Features of Perioperative Neurocognitive Disorders in Elderly Patients: A Narrative Review of the Clinical Literature
Previous Article in Journal
The Impact of Practicing Open- vs. Closed-Skill Sports on Executive Functions—A Meta-Analytic and Systematic Review with a Focus on Characteristics of Sports
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Applying Multiple Functional Connectivity Features in GCN for EEG-Based Human Identification

College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
These authors contributed to the work equally and should be regarded as co-first authors.
Brain Sci. 2022, 12(8), 1072; https://doi.org/10.3390/brainsci12081072
Submission received: 15 July 2022 / Revised: 10 August 2022 / Accepted: 11 August 2022 / Published: 12 August 2022
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
EEG-based human identification has gained a wide range of attention due to the further increase in demand for security. How to improve the accuracy of the human identification system is an issue worthy of attention. Using more features in the human identification system is a potential solution. However, too many features may cause overfitting, resulting in the decline of system accuracy. In this work, the graph convolutional neural network (GCN) was adopted for classification. Multiple features were combined and utilized as the structure matrix of the GCN. Because of the constant signal matrix, the training parameters would not increase as the structure matrix grows. We evaluated the classification accuracy on a classic public dataset. The results showed that utilizing multiple features of functional connectivity (FC) can improve the accuracy of the identity authentication system, the best results of which are at 98.56%. In addition, our methods showed less sensitivity to channel reduction. The method proposed in this paper combines different FCs and reaches high classification accuracy for unpreprocessed data, which inspires reducing the system cost in the actual human identification system.

1. Introduction

Due to the increasing security requirements for identification systems, biometric technologies have been extensively researched. Fingerprint, iris, face, and other biometric authentication methods have been extensively used in real life. However, these biometrics each possess their weaknesses [1]. For example, fingerprints can be easily stolen by touching the surface of an object. More importantly, these biometrics are non-renewable. Once the biometrics are stolen, the user cannot cancel them, as the user cannot simply change their face, iris, and fingerprints.
EEG is a recording of the electrical activity of the brain, and studies indicate that the EEG could be unique for each individual and thus has the potential to be a biometric [2]. EEG biometrics have unique advantages compared to traditional biometrics. Firstly, the acquisition of EEG cannot be performed without the users’ knowledge, therefore EEG biometrics are difficult to be stolen. Secondly, EEG biometric are reproducible. New EEG biometrics can be generated by simply changing the pattern of brain activity. Finally, EEG is sensitive to emotions such as stress, which makes it impossible to perform EEG identification in situations such as when the user receives coercion. For all these reasons, EEG-based identification holds the potential to be used in scenarios demanding a high security level [3].
Human identification is a prediction of a user’s identity among a group of people [4]. EEG-based human identification uses machine learning methods to classify EEG data from different people to identify different individuals. The most important thing in EEG-based identification is to select the EEG features that maximize the differences between different individuals [5]. EEG functional connectivity (FC) is a kind of index that describes the dependence and coupling relationship between different brain regions [6]. In the field of EEG analysis such as disease diagnosis, FC tools have been used to better observe the integration of distributed information of regular and synchronized multi-scale communication within and across inter-regional brain areas [7]. In previous studies, EEG-FC has been shown to have characterizing properties for different individuals [8]. Commonly used FC indicators in identity recognition are Pearson’s correlation coefficient (COR), coherence (COH) [9,10,11], phase lock value (PLV) [4,12,13,14,15,16], mutual information index (MI) [9], and Granger causality index (GC) [8], etc. These features are often used in isolation, which limits the accuracy improvement. When using delicate pre-processed data, the accuracy can reach 90.30% [8]–99.85% [4]. These FC indicators carry different information, which is complementary for the identification. Employing more FC features in identification may help to improve the accuracy. However, due to the redundancy and the unclear physiological mechanism of different FC features, efficiently integrating and utilizing these multiple features is challenging.
For traditional machine learning methods, simply incorporating more features will inevitably increase the dimension of the machine learning model. Especially when there is a lot of redundancy among different features, the model order is much higher than the real dimension of the feature, and this may result in overfitting.
Graph convolutional neural network (GCN) is a method for processing graph signals. Different from general machine learning algorithms, GCN not only needs to specify the features to be processed (node signals) but also needs to specify the connection relation-ship between features (graph structure) according to graph structure to aggregate neighbor node signals. In GCN, the dimension of the training parameters only depends on the column of the node signals rather than the graph structure.
This inspires solving the problem of multi-feature utilization: using concatenated multiple FCs as the structure matrix of GCN and using a homogenized vector as the node signal of GCN to aggregate information of the FC matrix. Because a larger structure matrix does not lead to more training parameters, the above GCN network can incorporate multiple FC features without increasing the model order.
In this work, a multiple-features integrating method based on GCN is proposed for identity recognition using multiple FC indicators. The new method was evaluated on a classic public dataset using different quantity of FCs and their different combination. The results showed that employing multiple FC features in GCN can improve the accuracy of the identification system compared to using a single FC feature. The best two-FC combination is MI + PLV with 98.51% accuracy, which is more than 3% better than the accuracy with a single FC. In addition, we evaluated the sensitivity of our method to channel reduction. The results showed that our method retained an accuracy of about 94.26% even when the number of channels was reduced to half of the original.
The rest of this paper is organized as follows: Section 2 presents the materials and a detailed approach to how multiple FC features can be used as graph structures. In Section 3, the evaluation results are reported. Section 4 explains the evaluation results and discusses the innovation and applicability of this work. Section 5 summarizes the work of this manuscript.

2. Materials and Methods

In this section, the dataset, FC computation, graph representation, and how our model is designed are introduced separately for this work.

2.1. Datasets and Preprocessing

We chose the widely used public dataset EEG Motor Movement/Imagery Dataset [17,18] to evaluate our method. This dataset comes from PhysioNet and contains more than 1500 one-minute and two-minute EEG recordings collected from more than 1000 subjects. Each subject performed 14 experiments: two one-minute baseline experiments with eyes open (EO) and eyes closed (EC), and three two-minute experiments with four motor or imagery tasks, four tasks, which are left or right fist open and closed (MOV-lr), imagine left or right fist open and closed (IMG-lr), double fist or both feet open and closed (MOV-b), imagine open and closed fists or feet (IMG-b). In this work, we used data from the baseline EO task for testing. The EEG was recorded from 64 electrodes as per the international 10–10 system (excluding electrodes A1 and A2). The sampling rate was 160 Hz. The description of Motor Movement/Imagery Dataset is illustrated in Table 1.
The main purpose of this work is to explore the classification capabilities of GCN networks for identity recognition when using multiple FCs as graph structures. To reduce the influence of irrelevant factors, only the sliding window method was adopted to segment the original EEG data, without any preprocessing methods such as removing bad segments that require manual participation. The sliding window used has a window length of 1 s and an overlap of 50%. Another reason to use sliding windows is for data augmentation.

2.2. Functional Connectivity Matrix

In this manuscript, four commonly used FCs are employed, namely Pearson’s correlation coefficient (COR), coherence (COH), phase locking value (PLV), and mutual information (MI). Detailed description and calculation process are listed below:
  • Pearson’s correlation coefficient (COR)
COR is the simplest FC metric. It measures the time–domain linear correlation between two signals x t and y t at zero lag. For zero mean, the unit variance signal is defined as:
R x y = 1 N k = 1 N x k y k
The value R x y range is [−1, 1], where a value of −1 represents a complete linear anti-correlation between the two signals, a value of 0 represents no linear dependence between the two signals, and a value of 1 represents a complete linear direct correlation between the two signals.
2.
Coherence (COH)
COH measures the linear dependence of two variables x t and y t in the frequency domain. COH can be calculated as follows:
C O H x y f = K x y f 2 = S x y f 2 S x x f S y y f
where S x y f is the cross power spectral density between x t and y t , S x x f and S y y f are the individual power spectral densities of x t and y t . The value C O H x y range is [0, 1]. A value of 0 means that the two signals have no linear correlation at the frequency f . A larger value means a stronger correlation between the two signals at the frequency f . A value of 1 means the correspondence between x t and y t at frequency f . Similar to COR, COH can only characterize the linear correlation between two signals.
3.
Phase Locking Value (PLV)
PLV estimates how the relative phase is distributed over the unit circle. It is defined as:
P L V = n 1 t = 1 n e i ϕ x t ϕ y t
where n represents the time point, ϕ x t and ϕ y t respectively represent the phase angles of the signals x and y at the time point t . The value range of PLV is [0, 1], and the larger the value, the stronger the phase synchronization between the two signals. The center frequency and bandwidth to be analyzed are usually specified in the PLV calculation and are usually divided according to the four bands of the EEG, that is, according to δ (1–3 Hz), θ (4–7 Hz), α (8–13 Hz), β (14–30 Hz) to calculate the PLV index.
4.
Mutual Information (MI)
MI is an important indicator of information theory. EEG is essentially a multi-dimensional time series signal; thus, it is meaningful to measure EEG signals with the complexity indexes of information theory. Complexity measures of time series, such as entropy and Lempel–Ziv complexity, have been employed to quantify the dynamical changes of EEG [19,20,21,22,23,24]. Based on concepts from information theory, MI measures the interdependence between two variables, and it quantifies the amount of information that can be obtained about a random variable by observing another. In the EEG functional connection, the mutual information index can represent the amount of information contained in each other between the time-series signals of the two channels. MI can be calculated as:
M I x y = i p x , y log p x , y p x p y
where p x , y is the joint probability distribution function of x t and y t . p x and p y are the marginal probability distribution functions of x t and y t . This equation represents the cross mutual information between x t and y t . The value range of MI is [0, 1]. When the value of MI is close to 1, it indicates that the two signals and the mutual information are completely correlated. When the value of MI is close to 0, it indicates that the two signals are completely independent.

2.3. Graph Representation

At present, the methods of analyzing EEG signals with GCN have been widely studied, and one of the basic tasks is to obtain the graph representation of multi-channel EEG signals. In this section, the general concepts of the graph and the graph representation of multiple FCs will be introduced.

2.3.1. General Concepts of Graph

The graph consists of vertices and edges connecting the vertices, denoted as G = V ,   E , where V is the vertex set and E is the edge set. A graph with n nodes whose nodes v i V , edges e i j E , where i , j 1 , 2 , , n .
A graph signal is a signal defined on a node of the graph, denoted as X , X = x 1 , x 2 , , x n T , where x i represents the signal strength on the node v i . When analyzing the properties of the graph, in addition to the strength of the graph signal, it is also necessary to consider the topological structure of the graph. Based on the transformation of adjacency matrix A, the renormalized Laplacian matrix L ˜ s y m is usually used to study the structural properties of the graph. The adjacency matrix A = a i j m × n represents the connection relationship between nodes. L ˜ s y m is defined as:
L ˜ s y m = D ˜ 1 / 2 A ˜ D ˜ 1 / 2  
where A ˜ = A + I , D ˜ i i = j A ˜ i j .

2.3.2. Graph Representation of Multiple FCs

In this work, the EEG electrodes are regarded as nodes. The graphs defined by a single FC are called subgraphs. Multiple subgraphs are fused into graphs by different fusion strategies, add and concat. For add-fusion, the adjacency matrices of subgraphs are averaged. For concat-fusion, the adjacency matrices of subgraphs are concatenated.
  • Add-fusion
The add-fusion is used to average the signal matrices of subgraphs containing different types of FC matrices, and the adjacency matrices are also averaged:
X a d d = m e a n X , X , A a d d = m e a n A , A ,
where X a d d and A a d d represent the signal strength matrix and adjacency matrix of the new graph obtained by add-fusion. X and A are the signal strength matrix and adjacency matrix of the first subgraph. X and A are the signal strength matrix and adjacency matrix of the second subgraph, and so on.
After the add-fusion, the signal strength of the graph and the connection strength of the graph structure are averaged, while the size of the adjacency matrix does not change; that is, no new connections are added.
2.
Concat-fusion
The concat-fusion preserves the signal values of all original nodes, concatenates the adjacency matrices of multiple subgraphs along the diagonal, and then fills the corresponding positions with the relationship matrices of multiple subgraphs:
X c o n c a t = X , X , T A c o n c a t = A R   R A      
where X c o n c a t and A c o n c a t represent the signal strength matrix and adjacency matrix of the new graph obtained by concat-fusion. R i j R represents the relationship matrix between the subgraphs.
Taking the fusion of two subgraphs as an example, as shown in Figure 1, two subgraphs and their adjacency matrix of different FC are represented as blue (subgraph 1 and A) and yellow (subgraph 2 and A′). The nodes are V = v a , v b , v c and V = v a , v b , v c , and the edges are E = e a b , e a c , e b c and E = e a b , e a c , e b c . A and A are the adjacency matrix within subgragh 1 and subgraph 2, respectively. The edges between the nodes from different subgraphs are colored in red. R is the relational matrix between the two subgraphs. In this work, we consider a simple case where the relational graph of two subgraphs is undirected. A c o n c a t is the adjacency matrix of the new graph obtained by concat-fusion. It can be seen from A c o n c a t that when the two subgraphs are combined by the concat-fusion, the number of nodes is doubled. The rows and columns of the adjacency matrix are doubled (the number of rows and columns of A c o n c a t in the graph changes to 6), and new connections appear between the nodes of the original two subgraphs (Matrix of the red part of the top right and bottom left corners of A c o n c a t ).
This work designs three types of relational matrices R for the concat method, which are zero matrix, identity matrix, and relational matrix based on Pearson correlation coefficient (concat_zero, concat_eye, and concat_corr).
The calculation method of the relationship matrix based on the Pearson correlation coefficient is:
r i j = ρ A i , : A j , :                         i f   i = j 0                                               o t h e r w i s e
where ρ x y represents the Pearson correlation coefficient of the two variables x and y , A i , : and A j , : represent the connection strength signals of nodes v i and v j to other nodes in the two subgraphs.

2.4. The Proposed Multi-FC Fusion GCN Model

The proposed multi-FC fusion GCN model is shown in Figure 2. Firstly, the original data are preprocessed and segmented. Then, different types of FC matrices are calculated as adjacency matrices. After that, we use the add or concat-fusion to combine multiple FCs and set the padding matrix to the three relationship matrices R described in Section 2.3.2. Through the above steps, we obtain a new adjacency matrix. Finally, a two-layer GCN network is employed to realize the classification of identities. The two layers of GCN are respectively defined as:
X 1 = σ l e a k y R e L U L ˜ s y m , X 0 | W 1 , b 1               W 1 n × 32 , b 1 32 X 2 = σ t a n h L ˜ s y m , X 1 | W 2 , b 2                                 W 2 32 × 8 , b 2 8
where σ is the activation function. W i   and b i are weights and bias matrices of the i th layer. L ˜ s y m is the renormalized laplacian matrix defined in Section 2.3.1. X is the signal strength matrix. In this work, the leakyReLU is applied to the first GCN layer with 32 convolutional neurons, and the tanh is applied to the second GCN layer with 8 convolutional neurons. For no other features except FC are employed, X 0 is set to one matrix 1 N × 1 .
After the two GCN layers, a dense layer is connected to map the feature space calculated by the previous layer to the sample label space, and finally, the softmax layer is used to realize the classification.
The model is written and implemented based on the python 3.7 and tensorflow 2.3.1 framework. NVIDIA GeForce RTX 2080Ti is used for model training. During training, a five-fold cross-validation method is used.

3. Results

3.1. The Evaluation Results by Using Different Combinations of Multiple FCs

In this work, we first evaluated the classification effect of four kinds of FC (COR, COH, PLV, and MI) used as adjacency matrices alone. The results are listed in Table 2. Then, we evaluated the classification accuracy of different FC combinations under different fusion methods (add, concat_zero, concat_eye, and concar_corr). There are a total of 11 combinations, and the detailed data are given in Table 3.
It can be seen from Table 2 that when the four FCs are used alone as the adjacency matrix of the graph, both COR and PLV achieve an accuracy of more than 95.26%, and COH alone achieves an accuracy of more than 91.21%. The worst performing MI can also achieve an accuracy of over 73.81%, which shows that the selected four FCs have a certain recognition ability for the identification task.
For convenience to compare the classification effect of single FC and multiple FCs, Figure 3 is reported, in which the left side shows the accuracy by using a single FC, and the right side shows the accuracy by using 11 different FC combinations with four different fusion methods.
As shown in Figure 3, in the case of two-FC combination, all four fusion methods are more effective than a single FC. The best two-FC combination is MI+PLV with the add-fusion, which achieves an accuracy of 98.51%. However, in the case of three-FC combination and four-FC combination, accuracy is not always higher than that in the two-FC combination. The accuracy of the combination COR+COH+MI with the add-fusion is 90.16%, which is lower than the combination COH+COR (96.36%), MI+COR (96.66%), and MI+COH (92.43%).
For further investigation of the phenomenon that the number of FCs increases while the accuracy decreases, Figure 4 is reported, where (a), (b), (c), and (d) show how the accuracy changes with the number of FCs in the four methods.
It can be seen in Figure 4a that under the add-fusion, when the number of FCs increases, the accuracy does not improve; the highest accuracy was achieved by two-FC combination. As shown in Figure 4b–d, in all three concat-fusions, the accuracy increases with the number of FCs. This may be due to the difference in the calculation method of add and concat, which is discussed in detail in the discussion section.

3.2. The Sensitivities to EEG Channel Numbers

In a human identification system, the number of channels is an important consideration. If the number of channels is too large, the deployment cost and complexity of the system will increase, which will limit the practical application [25]. Therefore, in the further study, we evaluate the performance of the proposed model in the case of channel reduction.
To evaluate the sensitivity of our method to the number of channels, we designed experiments to compare the classification accuracy by using single FC and multiple FCs at the different numbers of channels.
Firstly, we uniformly picked 32 channels from the original 64 channels by spatial distribution. Then, we evaluated the classification effect of single FC and multiple FCs of 32-channel-EEG. The results are shown in Figure 5, and detailed data are given in Table 4.
The left figure in Figure 5 shows the accuracy of four single FCs under 64 channels and 32 channels. It can be seen that when the number of channels is reduced to half of the original, the accuracy of the four single FCs decreases. PLV obtains the highest accuracy under 32 channels, reaching 82.58%.
The right figure in Figure 5 shows the accuracy of six two-FC combination under four different fusion methods under 32 channels. It can be seen that when the number of channels is reduced to half of the original, the accuracy of the FC combination using the concat-fusion is improved. Especially in the combination including PLV (PLV+COR, PLV+COH, and MI+PLV), the result is up to 94.26%, which is almost the same as the optimal single-channel under 64 channels. However, the result of using the add-fusion is not as good as a single FC, except for the combination PLV+COR.
In further exploration, to study the trend of performance when the number of channels is reduced, we give the accuracy of our method (PLV+COR) at a different level of channel reduction (64, 56, 48, 40, 32, 24, 20 and 16 channels remain); see Figure 6 and Table 5.
It can be seen from Figure 6 that the accuracy of both single FC and multiple FCs decreases as the number of channels decreases. The accuracy of multiple FCs with the concat-fusion declines slowly. When the number of channels is reduced to 40, the concat-fusion still maintains more than 97.13% accuracy, which is competent for the practical application of human identification.
Compared with the three methods of concat, when the number of channels is greater than or equal to 32, the accuracy of concat_eye is slightly higher than the other two methods. When the number of channels is less than 32, all three methods show a decline, but the accuracy of concat_corr decreases more slowly, which remains at 63.19% under 16 channels. This indicates that concat_corr is better than concat_zero and concat_eye in the ability to resist channel reduction.

4. Discussion

Many works are adopting GCN with FC in graph structure to deal with the individual differences of brain networks. Behrouzi et al. [4] set the signal feature as vector 1 and used single PLV or COR as the graph structure of GCN. This proves the feasibility of using FC as a graph structure to achieve stable identity recognition. In this paper, we use fused multiple FCs as the structure matrix of graphs. Because a larger structure matrix does not lead to more training parameters, our method can incorporate multiple FC features without increasing the model order.
The accuracy of using two-FC combination under 64 channels is up to 98.51% (MI+PLV). We used data without preprocessing. Most of the existing studies rely on delicate or even offline manual pre-processed data. With preprocessed data, the accuracy of the existing methods reached 90.30% (MI, SVM [8]), 98.75% (COR CNN [13]), and 99.85% (PLV, GCN [4]). Our method achieves an accuracy rate comparable to or exceeding existing methods even when using unpreprocessed data. This may be because our method reduces the effect of EEG artifacts by combining multiple FCs. The results show that our method has a higher classification capability for unpreprocessed data, which contributes to the cost reduction of the actual system deployment.
Based on the results under 64 channels reported in this paper, it can be seen that when a single FC could identify different people, the concat-fusion has a positive effect on improving the classification accuracy. With the increase in the number of multi-FC, the classification accuracy increases. This suggests that the concat-fusion can comprehensively utilize information of various types of FCs. Different from the concat-fusion, the evaluation of the add-fusion under 64 channels shows that in the case of two-FC combination, the add-fusion can bring some improvement to the accuracy. However, when the number of multiple FCs increases, the accuracy sometimes decreases, which is because the process of adding and averaging different types of FCs may cause a loss of features of conflicting changes between different FCs.
Based on the accuracy results of channel reduction reported in this paper, it can be seen that in the case of channel compression, as the number of channels decreases, the accuracy of the single FC, add and concat-fusions all decreases. Compared with a single FC and add, the three concat-fusions all decrease more slowly. This is because the graph structure obtained by the concat-fusion has more connections. When the number of channels is reduced (the number of nodes is reduced), more edge connections are beneficial to keep the aggregation path of information in the GCN network strong. In the comparison between the three concat-fusions, the concat_corr method shows higher accuracy when the number of channels is smaller (<32), which is because concat_corr artificially adds the relationship between the subgraphs of different FCs to express the relationship between each other. Compared to using zero matrix or eye matrix as the padding matrix, the subgraph correlation matrix itself also carries individual-specific information, which is conducive to further improve the classification ability.
Based on the above results, some application suggestions can be put forward for the actual deployment of the identification system. That is, using the concat-fusion to combine different types of FCs in the actual system can reduce the requirement for the minimum number of channels while ensuring a high accuracy rate. For example, when the dataset employed in this manuscript is used for identification, the accuracy of PLV+COR with the concat-fusion under 40 channels exceeds 97.13%, which is better than using any single FC alone under 64 channels This is of extraordinary significance for reducing the actual system cost.

5. Conclusions

At present, EEG identification using FC has become an extensively studied technique. However, how to apply multiple FCs in recognition is a challenging problem. This paper utilizes multiple FC features as the structure of the graph and explores the identification performance on a classical public dataset. The results show that utilizing multiple FCs can improve the accuracy of the identity authentication system, the best results of which are at 98.56%. In addition, this manuscript evaluates the sensitivity of our method to the number of channels. When the number of channels is reduced from 64 to 40, the combination PLV+COR with the concat-fusion maintains an accuracy of over 97.13%. This result suggests that multiple FCs can be adopted with the concat-fusion on fewer channels to reduce the system cost while maintaining high recognition performance, which inspires the actual deployment of the human identification system.

Author Contributions

Conceptualization, M.L.; methodology, W.T.; software, W.T. and X.J.; validation, Y.L. and M.L.; formal analysis, W.T. and X.J.; investigation, W.T. and M.L.; resources, M.L.; data curation, W.T. and M.L.; writing—original draft preparation, W.T.; writing—review and editing, W.T., Y.L., and M.L.; visualization, W.T.; supervision, M.L.; project administration, M.L.; funding acquisition, Y.L. and M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Natural Science Foundation of China, grant numbers u19a2083, 62076248 and by the Defense Industrial Technology Development Program, grant number JCKY2020550B003.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Jain, A.K.; Ross, A.; Prabhakar, S. An introduction to biometric recognition. IEEE Trans. Circuits Syst. Video Technol. 2004, 14, 4–20. [Google Scholar] [CrossRef]
  2. Campisi, P.; La Rocca, D. Brain waves for automatic biometric-based user recognition. IEEE Trans. Inf. Forensics Secur. 2014, 9, 782–800. [Google Scholar] [CrossRef]
  3. Gui, Q.; Ruiz-Blondet, M.V.; Laszlo, S.; Jin, Z. A Survey on Brain Biometrics. ACM Comput. Surv. 2019, 51, 1–38. [Google Scholar] [CrossRef]
  4. Behrouzi, T. Graph Learning Perspective on EEG-Based Human Identification. Master’s Thesis, University of Toronto, Toronto, ON, Canada, 2021. [Google Scholar]
  5. Del Pozo-Banos, M.; Alonso, J.B.; Ticay-Rivas, J.R.; Travieso, C.M. Electroencephalogram subject identification: A review. Expert Syst. Appl. 2014, 41, 6537–6554. [Google Scholar] [CrossRef]
  6. Bastos, A.M.; Schoffelen, J.-M. A tutorial review of functional connectivity analysis methods and their interpretational pitfalls. Front. Syst. Neurosci. 2016, 9, 175. [Google Scholar] [CrossRef]
  7. Varone, G.; Boulila, W.; Lo Giudice, M.; Benjdira, B.; Mammone, N.; Ieracitano, C.; Dashtipour, K.; Neri, S.; Gasparini, S.; Morabito, F.C.; et al. A Machine Learning Approach Involving Functional Connectivity Features to Classify Rest-EEG Psychogenic Non-Epileptic Seizures from Healthy Controls. Sensors 2021, 22, 129. [Google Scholar] [CrossRef] [PubMed]
  8. Zhang, Y.; Li, M.; Shen, H.; Hu, D. On the Specificity and Permanence of Electroencephalography Functional Connectivity. Brain Sci. 2021, 11, 1266. [Google Scholar] [CrossRef] [PubMed]
  9. Riera, A.; Soria-Frisch, A.; Caparrini, M.; Grau, C.; Ruffini, G. Unobtrusive biometric system based on electroencephalogram analysis. EURASIP J. Adv. Signal Process. 2007, 2008, 143728. [Google Scholar] [CrossRef]
  10. Safont, G.; Salazar, A.; Soriano, A.; Vergara, L. Combination of multiple detectors for EEG based biometric identification/authentication. In Proceedings of the IEEE International Carnahan Conference on Security Technology, Newton, MA, USA, 15–18 October 2012. [Google Scholar]
  11. Liew, S.H.; Choo, Y.H.; Yin, F.L. Fuzzy-Rough Nearest Neighbour classifier for person authentication using EEG signals. In Proceedings of the 2013 International Conference on Fuzzy Theory and Its Applications (iFUZZY), Taipei, Taiwan, 6–8 December 2013. [Google Scholar]
  12. Behrouzi, T.; Hatzinakos, D. Graph variational auto-encoder for deriving EEG-based graph embedding. Pattern Recognit. 2022, 121, 108202. [Google Scholar] [CrossRef]
  13. Wang, M.; El-Fiqi, H.; Hu, J.; Abbass, H.A. Convolutional Neural Networks Using Dynamic Functional Connectivity for EEG-Based Person Identification in Diverse Human States. IEEE Trans. Inf. Forensics Secur. 2019, 14, 3259–3272. [Google Scholar] [CrossRef]
  14. Liu, S.; Bai, Y.; Liu, J.; Qi, H.; Li, P.; Zhao, X.; Zhou, P.; Zhang, L.; Wan, B.; Wang, C.; et al. Individual feature extraction and identification on EEG signals in relax and visual evoked tasks. In Proceedings of the International Conference on Biomedical Informatics and Technology, Aizu-Wakamatsu, Japan, 16–17 September 2013; pp. 305–318. [Google Scholar]
  15. Kong, W.; Fan, Q.; Wang, L.; Jiang, B.; Peng, Y.; Zhang, Y. Task-free brainprint recognition based on degree of brain networks. In Proceedings of the International Conference on Neural Information Processing, Long Beach, CA, USA, 4–9 December 2017; pp. 709–717. [Google Scholar]
  16. Bai, Y.; Zhang, Z.; Ming, D. Feature selection and channel optimization for biometric identification based on visual evoked potentials. In Proceedings of the 2014 19th International Conference on Digital Signal Processing, Hong Kong, China, 20–23 August 2014; pp. 772–776. [Google Scholar]
  17. Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [PubMed]
  18. Schalk, G.; McFarland, D.J.; Hinterberger, T.; Birbaumer, N.; Wolpaw, J.R. BCI2000: A General-Purpose Brain-Computer Interface (BCI) System. IEEE Trans. Biomed. Eng. 2004, 51, 1034–1043. [Google Scholar] [CrossRef] [PubMed]
  19. Hornero, R.; Abásolo, D.; Escudero, J.; Gómez, C. Nonlinear analysis of electroencephalogram and magnetoencephalogram recordings in patients with Alzheimer’s disease. Philos. Trans. R. Soc. Math. Phys. Eng. Sci. 2009, 367, 317–336. [Google Scholar] [CrossRef] [PubMed]
  20. Adeli, H.; Ghosh-Dastidar, S. Automated EEG-Based Diagnosis of Neurological Disorders; CRC Press: Boca Raton, FL, USA, 2010; p. 75. [Google Scholar]
  21. Bai, Y.; Liang, Z.; Li, X. A permutation Lempel-Ziv complexity measure for EEG analysis. Biomed. Signal Process. Control 2015, 19, 102–114. [Google Scholar] [CrossRef]
  22. Bachmann, M.; Päeske, L.; Kalev, K.; Aarma, K.; Lehtmets, A.; Ööpik, P.; Lass, J.; Hinrikus, H. Methods for classifying depression in single channel EEG using linear and nonlinear signal analysis. Comput. Methods Programs Biomed. 2018, 155, 11–17. [Google Scholar] [CrossRef] [PubMed]
  23. Pregowska, A.; Proniewska, K.; van Dam, P.; Szczepanski, J. Using Lempel-Ziv complexity as effective classification tool of the sleep-related breathing disorders. Comput. Methods Programs Biomed. 2019, 182, 105052. [Google Scholar] [CrossRef] [PubMed]
  24. Bajić, D.; Đajić, V.; Milovanović, B. Entropy Analysis of COVID-19 Cardiovascular Signals. Entropy 2021, 23, 87. [Google Scholar] [CrossRef] [PubMed]
  25. Demir, A.; Koike-Akino, T.; Wang, Y.; Haruna, M.; Erdogmus, D. EEG-GNN: Graph Neural Networks for Classification of Electroencephalogram (EEG) Signals. arXiv 2021, arXiv:2106.09135. [Google Scholar]
Figure 1. The schematic of the fusion of two subgraphs. Two subgraphs and their adjacency matrix of different FC are represented as blue (subgraph 1 and A ) and yellow (subgraph 2 and A ). A and A are the adjacency matrices within subgragh 1 and subgraph 2, respectively. The edges between the nodes from different subgraphs are colored in red. R is the relational matrix between the two subgraphs. A c o n c a t is the adjacency matrix of the new graph obtained by concat-fusion.
Figure 1. The schematic of the fusion of two subgraphs. Two subgraphs and their adjacency matrix of different FC are represented as blue (subgraph 1 and A ) and yellow (subgraph 2 and A ). A and A are the adjacency matrices within subgragh 1 and subgraph 2, respectively. The edges between the nodes from different subgraphs are colored in red. R is the relational matrix between the two subgraphs. A c o n c a t is the adjacency matrix of the new graph obtained by concat-fusion.
Brainsci 12 01072 g001
Figure 2. The flow chart of the proposed multiple features integrating method based on GCN.
Figure 2. The flow chart of the proposed multiple features integrating method based on GCN.
Brainsci 12 01072 g002
Figure 3. Comparison between using single FC and multiple FCs on classification accuracy. Upper left: accuracy by using single FC. Upper right: accuracy by using 11 different FC combinations with four different fusion methods. Below: the enlarged view of the rectangle in the upper right figure. It can be seen that all four fusion methods are more effective than a single FC in the case of two-FC combination. The best two-FC combination is MI+PLV, which achieves an accuracy of 98.51% with the add-fusion. However, in the case of three-FC combination and four-FC combination, accuracy is not always higher than that in two-FC combination. The accuracy of the combination COR+COH+MI with the add-fusion is 90.16%, which is lower than the combination COH+COR (96.36%), MI+COR (96.66%), and MI+COH (92.43%).
Figure 3. Comparison between using single FC and multiple FCs on classification accuracy. Upper left: accuracy by using single FC. Upper right: accuracy by using 11 different FC combinations with four different fusion methods. Below: the enlarged view of the rectangle in the upper right figure. It can be seen that all four fusion methods are more effective than a single FC in the case of two-FC combination. The best two-FC combination is MI+PLV, which achieves an accuracy of 98.51% with the add-fusion. However, in the case of three-FC combination and four-FC combination, accuracy is not always higher than that in two-FC combination. The accuracy of the combination COR+COH+MI with the add-fusion is 90.16%, which is lower than the combination COH+COR (96.36%), MI+COR (96.66%), and MI+COH (92.43%).
Brainsci 12 01072 g003
Figure 4. Comparison between using different quantities of FCs with the four fusion methods ((a): add, (b): concat_zero, (c): concat_eye, and (d): concat_corr). In the add-fusion, when the number of FCs increases, the accuracy does not improve; the highest accuracy was achieved by two-FC combination (subfigure a). In all three concat-fusions, the accuracy increases with the number of FCs (subfigures bd).
Figure 4. Comparison between using different quantities of FCs with the four fusion methods ((a): add, (b): concat_zero, (c): concat_eye, and (d): concat_corr). In the add-fusion, when the number of FCs increases, the accuracy does not improve; the highest accuracy was achieved by two-FC combination (subfigure a). In all three concat-fusions, the accuracy increases with the number of FCs (subfigures bd).
Brainsci 12 01072 g004
Figure 5. Classification accuracy comparison by using single FC and multiple FCs in the case of channel reducing. Left: accuracy by using four single FCs under 64 channels (grey bars) and 32 channels (purple bars). Right: accuracy by using six two-FC combination under 32 channels. When the number of channels is reduced to half of the original, the accuracy of the FC combination using the concat-fusion is higher than by using single FC. Especially in the combination containing PLV (PLV+COR, PLV+COH, and MI+PLV), the best result is up to 94.26%, which is almost the same as the optimal single channel under 64 channels.
Figure 5. Classification accuracy comparison by using single FC and multiple FCs in the case of channel reducing. Left: accuracy by using four single FCs under 64 channels (grey bars) and 32 channels (purple bars). Right: accuracy by using six two-FC combination under 32 channels. When the number of channels is reduced to half of the original, the accuracy of the FC combination using the concat-fusion is higher than by using single FC. Especially in the combination containing PLV (PLV+COR, PLV+COH, and MI+PLV), the best result is up to 94.26%, which is almost the same as the optimal single channel under 64 channels.
Brainsci 12 01072 g005
Figure 6. Performance comparison of different methods at the different levels of channel reduction (64, 56, 48, 40, 32, 24, 20, and 16 channels remain). It can be seen that the accuracy of multiple FCs with the concat-fusion declines slowly. When the number of channels is reduced to 40, the concat-fusion still maintains more than 97.13% accuracy. When the number of channels is greater than 32, the concat method shows a significant advantage.
Figure 6. Performance comparison of different methods at the different levels of channel reduction (64, 56, 48, 40, 32, 24, 20, and 16 channels remain). It can be seen that the accuracy of multiple FCs with the concat-fusion declines slowly. When the number of channels is reduced to 40, the concat-fusion still maintains more than 97.13% accuracy. When the number of channels is greater than 32, the concat method shows a significant advantage.
Brainsci 12 01072 g006
Table 1. Description of the motor movement/imagery dataset.
Table 1. Description of the motor movement/imagery dataset.
Motor Movement/Imagery Dataset
#Channel *64Electrodes Brainsci 12 01072 i001
#Subject109
Fs (Hz)160
Tasks (duration × run)EO (60 s)
EC (60 s)
MOV-lr (120 s × 3)
IMG-lr (120 s × 3)
MOV-d (120 s × 3)
IMG-d (120 s × 3)
* #Channel means the number of channels. In the subsequent manuscript, the meaning of # is the same as here.
Table 2. Classification accuracy by using single FC (COR, COH, PLV, and MI).
Table 2. Classification accuracy by using single FC (COR, COH, PLV, and MI).
FCCORCOHPLVMI
ACC/%95.2791.2195.2673.81
Table 3. Classification accuracy by combining multiple FCs with different fusion methods.
Table 3. Classification accuracy by combining multiple FCs with different fusion methods.
#FCCombinationFusion Method (%)
AddConcat_ZeroConcat_EyeConcat_Corr
2COH+COR96.3696.5396.9996.42
PLV+COR98.2898.2998.4498.25
PLV+COH98.3198.3498.3198.28
MI+COR96.6696.7596.5996.62
MI+COH92.4392.5692.0092.52
MI+PLV98.5198.4598.4098.43
3COH+PLV+MI96.4098.4798.3798.45
COR+PLV+MI97.1098.5198.4698.56
COR+COH+MI90.1697.0297.3896.89
COR+COH+PLV97.4698.2298.4098.22
4COR+CPH+PLV+MI96.8298.4898.5698.48
Table 4. Accuracy of two-FC combination in different fusion methods under 32 channels.
Table 4. Accuracy of two-FC combination in different fusion methods under 32 channels.
#FCCombinationFusion Method (%)
AddConcat_ZeroConcat_EyeConcat_Corr
2COH+COR70.1483.4184.0482.90
PLV+COR84.7192.8094.1793.39
PLV+COH80.7990.4490.0889.96
MI+COR52.8089.0888.6689.19
MI+COH41.4477.3176.7777.51
MI+PLV76.0394.0693.1394.26
Table 5. Accuracy of single FC and multiple FCs under different number of channels.
Table 5. Accuracy of single FC and multiple FCs under different number of channels.
#ChannelFusion Method (%)
PLV_SingleCOR_SingleAddConcat_ZeroConcat_EyeConcat_Corr
6495.2695.2798.2898.3298.4498.29
5694.1693.2797.0698.0798.3098.05
4890.9590.1696.1797.6997.7797.53
4088.0284.9393.6996.5697.1396.63
3282.5872.7084.7192.8094.1793.39
2471.6855.5968.6384.7684.8386.25
2056.8837.5750.7473.4572.7277.27
1633.7223.3329.7254.8553.4163.19
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tian, W.; Li, M.; Ju, X.; Liu, Y. Applying Multiple Functional Connectivity Features in GCN for EEG-Based Human Identification. Brain Sci. 2022, 12, 1072. https://doi.org/10.3390/brainsci12081072

AMA Style

Tian W, Li M, Ju X, Liu Y. Applying Multiple Functional Connectivity Features in GCN for EEG-Based Human Identification. Brain Sciences. 2022; 12(8):1072. https://doi.org/10.3390/brainsci12081072

Chicago/Turabian Style

Tian, Wenli, Ming Li, Xiangyu Ju, and Yadong Liu. 2022. "Applying Multiple Functional Connectivity Features in GCN for EEG-Based Human Identification" Brain Sciences 12, no. 8: 1072. https://doi.org/10.3390/brainsci12081072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop