Next Article in Journal
Functional Connectivity Increases in Response to High-Definition Transcranial Direct Current Stimulation in Patients with Chronic Disorder of Consciousness
Previous Article in Journal
The Affect Misattribution in the Interpretation of Ambiguous Stimuli in Terms of Warmth vs. Competence: Behavioral Phenomenon and Its Neural Correlates
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Decoding Task-Based fMRI Data with Graph Neural Networks, Considering Individual Differences

1
Department of Industrial Engineering and Management Systems, University of Central Florida, Orlando, FL 32816, USA
2
Department of Biostatistics, Johns Hopkins University, Baltimore, MD 21218, USA
3
Department of Psychology, University of Central Florida, Orlando, FL 32816, USA
4
Institute of Advanced Consciousness Studies, Santa Monica, CA 90403, USA
5
School of Modeling, Simulation, and Training Computer Science, University of Central Florida, Orlando, FL 32816, USA
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(8), 1094; https://doi.org/10.3390/brainsci12081094
Submission received: 15 June 2022 / Revised: 1 August 2022 / Accepted: 6 August 2022 / Published: 17 August 2022
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
Task fMRI provides an opportunity to analyze the working mechanisms of the human brain during specific experimental paradigms. Deep learning models have increasingly been applied for decoding and encoding purposes study to representations in task fMRI data. More recently, graph neural networks, or neural networks models designed to leverage the properties of graph representations, have recently shown promise in task fMRI decoding studies. Here, we propose an end-to-end graph convolutional network (GCN) framework with three convolutional layers to classify task fMRI data from the Human Connectome Project dataset. We compared the predictive performance of our GCN model across four of the most widely used node embedding algorithms—NetMF, RandNE, Node2Vec, and Walklets—to automatically extract the structural properties of the nodes in the functional graph. The empirical results indicated that our GCN framework accurately predicted individual differences (0.978 and 0.976) with the NetMF and RandNE embedding methods, respectively. Furthermore, to assess the effects of individual differences, we tested the classification performance of the model on sub-datasets divided according to gender and fluid intelligence. Experimental results indicated significant differences in the classification predictions of gender, but not high/low fluid intelligence fMRI data. Our experiments yielded promising results and demonstrated the superior ability of our GCN in modeling task fMRI data.

1. Introduction

Functional magnetic resonance imaging (fMRI) is a non-invasive technology that provides high spatial resolution in determining the human brain’s responses [1]. Technically, fMRI estimates the regional brain activity by measuring metabolic changes in blood oxygen consumption associated with neural activity [2]. Modeling task fMRI data provides an opportunity to analyze the working mechanism of the human brain during performance of specific tasks. In task fMRI scanning, time series of the three-dimensional volume of the brain are acquired within a task block while the participant’s brain actively performs an explicit task. Pattern classification techniques are frequently applied to task fMRI data to determine if information is present in a particular brain region in a format the model can exploit, as in decoding studies. Encoding models can be used to make predictions about brain representations based on stimuli used in a particular task [3]. These models can also be applied to predict behavioral responses based on task fMRI [4], or for making group membership predictions based on graph theoretic features [5].
Graph neural networks (GNNs) have recently gained interest in the deep learning community, and have recently been applied to analyze task fMRI [6,7,8,9,10,11]. GNNs are neural network-based models that operate on graph-structured data, and have demonstrated state-of-the-art results in learning graph representations by embedding nodes in a low dimensional space [12]. Depending on the particular data structure of a graph, the performance of embedding methods in extracting the structural properties of graph nodes varies; thus, classification algorithms may perform differently according to the node embedding method used [13]. Although previous studies have attempted to increase the classification performance of task fMRI data by using various node embedding methods [8], less effort has been directed toward inferring how various node embeddings perform differently on classification task fMRI data. In this study, we used task fMRI data from the Human Connectome Project (HCP) dataset [14] to compare and adjudicate amongst embedding methods. Herein, we propose an end-to-end graph convolutional network (GCN) framework to classify task-evoked fMRI data. We conducted a series of experiments to evaluate the model’s classification performance by using four well-known node embedding algorithms: NetMF, RandNE, Node2Vec, and Walklets (Section 3.4).
Brain connectivity varies considerably amongst individuals [15,16,17,18,19], and even within individuals across time [20]. A number of studies have noted structural and functional connectivity differences across gender [21], developmental populations [5], intelligence [22], across the lifespan [23], and even level of empathic concern [24]. On the basis of such research indicating a close link between individual variability and brain connectivity derived measures, we aimed to examine the effects of individual differences (i.e., gender and fluid intelligence (gF)) on the classification performance of the proposed GCN model. For this purpose, we performed extensive experiments on four sub-datasets: gender-associated sub-datasets (female and male) and gF score-associated sub-datasets (sub-datasets associated with individuals with gF scores lower than the median value, denoted LM-gF, or higher than the median value, denoted HM-gF). We assessed significant changes in classification performance across each dataset. In general, our contributions through this work are summarized below:
  • We propose an end-to-end GCN framework to classify task-evoked fMRI data. The objective is to examine the performance of various node embeddings to generate topological embeddings of the graph’s nodes. To our knowledge, this is the first investigation of different node embeddings on task fMRI classification performance. The code is available at https://github.com/krzysztoffiok/gnn-classification-pipeline, accessed on 20 February 2022.
  • We demonstrate the performance of the proposed GCN framework according to individual differences (i.e., gender and fluid intelligence). To this end, we constructed four small sub-datasets of gender and gF score (LM-gF/HM-gF) with replacement.
This paper is organized as follows. In Section 2, we describe the background of the work. Section 3 describes our task fMRI data, the GCN architecture, and performance evaluation. In Section 4, we present the results from our experiments, followed by a discussion of our findings, the limitations of the work, and future directions. We end the current work with concluding remarks in Section 6.

2. Background

Over the past several decades, a variety of computational methods have been proposed to analyze fMRI time series data, such as the generalized linear model (GLM) [25,26], sparse dictionary learning [27,28,29,30], and blind source separation techniques including independent component analysis [31,32,33,34], non-negative matrix factorization [35], and tensor decomposition [36,37]. While useful, these techniques are either model free, or impose a particular inductive bias in the model. As such, their architectures do not resemble the structural or functional information processing in the human brain, limiting their capability of being used as brain computational models [38].
In the past several years, a growing body of literature has applied deep learning (DL) algorithms to fMRI data for decoding and encoding purposes. DL models leverage only a small subset of the dynamic capabilities of biological neurons, yet are functionally inspired by neurobiology. In DL methods, rather than using manual features, which are usually based on expert domain knowledge and heuristics [39], high level complex features can be automatically extracted from the original fMRI data, thus providing meaningful information to improve the performance of classification models. For example, Huang et al. [40] proposed a deep neural network model, consisting of both convolutional and recurrent layers, that automatically extracts spatial and temporal features of fMRI data. Their convolutional recurrent neural network model was used for the seven-class classification task, and the experimental results on the HCP dataset achieved an average accuracy performance of 94.3%. Wang et al. [41] applied a DL classifier with five convolutional layers and two fully connected layers on a large subset of task fMRI data from the HCP dataset and obtained an average accuracy of 93.7%.
Among DL models, convolutional autoencoders [42,43,44], recurrent autoencoders [45,46], and deep belief networks [47,48,49] have shown a superior ability to decode fMRI data. Huang et al. have developed a deep convolutional autoencoder to model fMRI data [42,43]; Zhao and colleagues used a spatio-temporal convolutional neural network to obtain both spatial and temporal features of functional networks [44]; Wang et al. have applied a deep sparse recurrent neural network on task fMRI data that has shown promising performance in extracting the temporal dependencies of input fMRI volumes [45]; and a deep belief network with a restricted Boltzmann machine [47] has been used to identify networks in fMRI data. Similarly, Jang et al. applied the deep belief network from [47] to initialize the weights of a fully connected deep learning architecture [48]. Despite the advances made by these methods, the DL models are yet to reach their full potential in the functional neuroimaging community due to the high dimensionality of the data and limited training data [50].
Graph-based network analyses capture information about the topological architecture of human brain networks [20]. Therefore, GNNs represent an attractive new tool for modeling brain information processing given that they are biologically inspired and leverage the hierarchical computing power from deep learning neural network models [51,52,53]. These models have been applied for fMRI decoding purposes using spectral-based GCN [6]. Li et al. [8] extended this work by proposing the BrainGNN framework with ROI-aware graph convolutional layers and ROI-selection pooling layers. These two types of layers were used to extract topological features of fMRI data and highlight the important nodes of the brain’s graph for prediction, respectively. The framework has been used to map regional and cross-regional functional activation patterns for decoding cognitive states in the HCP S1200 dataset. Furthermore, Kim et al. [10] considered the dynamic characteristics of the functional connectivity network and proposed the Spatio-Temporal Attention Graph Isomorphism Network for learning dynamic graph representation of the brain connectome with spatio-temporal attention.

3. Materials and Methods

According to the model used in [8], we applied a GCN framework for learning hierarchical representations of brain graphs to perform the node classification task. The topological and spatial feature vectors of brain functional graph nodes can also be automatically extracted by using node embedding methods. In this section, we first introduce the concept of convolutional operation on graph spectral domains, on the basis of Fourier transform and graph Laplacian. We then describe our proposed GCN model and the loss function that we intend to minimize. Finally, we present the dataset used as well as the brain network construction and feature extraction for graph nodes.

3.1. fMRI Dataset and Preprocessing

We obtained task fMRI data for 302 participants, consisting of 164 women and 138 men (22–35 years, mean = 28.7 ± 3.6 ) from the HCP 1200 Subject Release (S1200) [14]. HCP participants were randomly drawn from a population of healthy individuals, and fMRI data was collected while subjects performed seven different tasks—emotion, gambling, working memory, language, relational, social, and motor [26]. We used a subset of the HCP data collected at a single site, Washington University, to obviate the need for data harmonization [54]. Data were collected at 3 Tesla with TR = 0.72 s, TE = 33.1 msec, flip angle = 52 degree, FOV = 208 × 180 mm, and voxel size = 2.0 mm isotropic with opposite phase encoding directions (left-to-right and right-to-left). For further details, see [14].
To perform our experiments aimed at evaluation of the influence of individual differences, we considered two categories of task fMRI data: gender and fluid intelligence (gF). The first category consisted of two datasets in which task fMRI data for 164 and 138 participants were assigned to each sub-dataset according to gender. In the second category, we sorted gF scores of 302 participants in descending order and divided the dataset of 302 participants into two sub-datasets, LM-gF and HM-gF, of participants with gF scores lower than the median value (gF score < 18 ) and with gF score higher than the median value (gF score 18 ), respectively. Consequently, a total of 144 and 158 participants’ task fMRI data were assigned to the LM-gF and HM-gF sub-datasets, respectively. Table 1 presents the demographics and participant distribution of the four defined sub-datasets.
The preprocessing of the task fMRI volume time-series was performed by the HCP consortium, as previously described [55]. The preprocessing pipeline included artifact removal and gradient distortion correction, motion correction, and registration to the standard Montreal Neurological Institute space with a DARTEL and voxel size of 2 × 2 × 2   mm 3 . Spatial smoothing and activation map generation were performed with a GLM implemented in FSL’s FILM (FMRIB’s Improved Linear Model with autocorrelation) [56]. More details regarding the HCP preprocessing pipeline can be found in Barch et al. [26].

3.2. Graph Convolutional Network: Spectral

3.2.1. Notation

We used the basic notions described in [57]. A graph is defined as G = V , E that consists of the set of nodes v 1 , v 2 , , v n and set of edges such that e i j = v i   ,   v j E and E V × V . An edge e has two endpoints, v i and v j , that are said to be joined by e . In this case, these two nodes are adjacent. A graph can be either directed or undirected. With an undirected graph, edges have no orientation. In contrast to undirected graphs, directed graphs are the set of nodes connected by edges that have a direction associated with them. Furthermore, a graph is a weighted graph if a weight is assigned to each edge. These weights quantify the degree of interaction between the nodes or the volume of exchange.
Definition 1 (adjacency matrix).
The adjacency matrix A for a graph G with n -nodes is an n × n matrix representation with A i j = 1 if direct connections exist between v i and v j , and A i j = 0 if no direct connections exist. If the graph is weighted, the entry of the adjacency matrix is A i j > 0 if v i ,   v j E and A i j = 0 , otherwise.
Definition 2 (feature matrix).
The node feature matrix X R V × d , where V is the number of nodes in the graph, and d is the number of node features, is a matrix with x i R d representing the d -dimensional feature vector of the node v . Similarly, the edge feature matrix X e R M × p is a matrix with X v i , v j e R p representing the p -dimensional feature vector of the edge e i j .
Definition 3 (Laplacian matrix).
The Laplacian matrix (or graph Laplacian) L R N × N is defined as L = D A , where D is the degree matrix, D i j = j = 1 n A i j , and A is the adjacency matrix of the unweighted graph. Similarly, for a weighted graph, L = D W , where W is a weighted adjacent matrix. The symmetric normalized Laplacian matrix can be defined as L sym = I D 1 2 AD 1 2 , where I is the identity matrix.

3.2.2. Spectral-Based GCN

Spectral GCNs use the Laplacian matrix to compute the eigen-decomposition of the graph Laplacian in the Fourier domain. Let L s y m be the symmetric normalized Laplacian matrix of graph G .   L s y m can be decomposed into L s y m = U Λ U T , where U = U 0 , U 1 , , U n 1 R n × n is the eigenvector matrix, and Λ is the diagonal matrix of eigenvalues, Λ = d i a g ( λ 1 , λ 2 , , λ n ) . In graph signal processing, node features are mapped to feature vectors (i.e., x 0 , , x n 1 ), which may be formed as a feature vector of all nodes of a graph, X R n . The graph Fourier transform to a signal X is defined as X ^ = U T X , and the inverse graph Fourier transform is defined as X = U X ^ . The graph convolution operation of X in the Fourier domain is defined as follows:
X × g = U U T g U T X
where × represents convolution operation, represents the pointwise product, and g R N represents the learnable parameters of the graph convolutional kernel. By defining g θ = d i a g U T g as a spectral filter in the spectral domain, the graph convolution operation can be simply defined as follows:
X × g θ = U g θ Λ U T X
Equation (2) was used for the first spectral network proposed [58]. However, this operation was computationally expensive because of multiplication eigenvector matrix U , which is a full matrix with n Fourier functions. To avoid the quadratic complexity, Defferrard et al. [59] have proposed ChebNet model, which avoids the eigen-decomposition by using a learning function of the Laplacian. The ChebNet model uses Chebyshev polynomials of the diagonal matrix of eigenvalues to estimate the filter g θ as shown below:
g θ = i = 0 K θ i T k ( Λ ~ )
where Λ ~ = 2 Λ / λ m a x I N , and Λ 1 , 1 . The model uses a Chebyshev polynomial for recursive calculation as T k x = 2 x T k 1 x T k 2 x with T 0 x = 1 and T 1 x = x . Therefore, the definition of the convolution of the graph signal x with a filter g θ is as shown below:
X × g θ = U ( i = 0 K θ i T i ( L ˜ ) ) U T X
where L ~ = 2 L s y m / λ m a x I N , and maps the eigenvalues from 0 , λ m a x   to 1 , 1 [60].
The filters defined by ChebNet are unstable for localizing frequency bands of interest, which are essentially the graph communities. To improve the above-mentioned ChebNet model and reduce the overfitting problem [61], Kipf and Welling [62] have proposed the CayleyNet model to capture narrow frequency bands by using Cayley polynomials. ChebNet assumes a linear function with respect to K = 1 and λ m a x = 2 , which results in a simplification of Equation (4) as shown below:
X × g θ = f ( D ˜ 1 2 A ˜ D ˜ 1 2 X Θ )
where A ˜ = I + A , is an adjusted adjacency matrix A , D ˜ i j = j A ˜ i j , f is the activation function, and Θ is a matrix of filter parameters.

3.3. Functional Graph

The raw task fMRI data were preprocessed through the HCP minimal preprocessing pipeline [55] and denoised by using ICA-FIX [63] to remove spatial artifacts and to perform motion correction. Furthermore, we used a large-scale multimodal brain atlas to parcellate the brain regions into 360 anatomical areas by using HCP Multi-Modal Parcellation, which is based on a combination of cortical architecture, function, connectivity, and topography [64]. By parcellation, we define regions of interest that represent graph nodes for brain network construction. Theoretically, the construction of a functional graph involves two steps. Herein, we first averaged the time series of all voxels in the region. Then we computed the functional connectivity between each pair of averaged time series of brain region through Pearson’s correlation coefficient. We used Fisher’s z transformation to normalized r values to improve the normality, and obtained a 360 × 360 symmetric matrix A (adjacency matrix) for each participant.

3.4. Feature Engineering and Node Embedding Algorithms

Features from averaged time series of brain regions were extracted by using Time Series Feature Extraction on basis of Scalable Hypothesis tests (tsfresh), an efficient and scalable feature extraction algorithm for time series based on a Python package [65]. The tsfresh algorithm integrates the components from the hypothesis tests with the feature significance testing on the basis of the FRESH algorithm [66]. Each generated feature vector is independently assessed to identify its significance for the given target by quantifying p-values and is further evaluated through the Benjamini–Yekutieli procedure [67] to decide which features to keep. The features extracted by tsfresh consist of both basic and advanced characteristics of the time series, and a complete list of features along with their mathematical descriptions can be found in reference [66]. We selected a minimum set of relevant statistical features to prepare feature representations for each node as follows: “absolute_sum_of_changes”, “benford_correlation”, “c3” (i.e., a measure of non-linearity in the time series), “cid_ce” (i.e., a measure of complexity in the time series), “longest_strike_above_mean”, “variance”, “standard deviation”, “skewness”, and “quantile” (i.e., 0.25 quantile).
In addition to the statistical features obtained through the tsfresh algorithm, node embeddings were applied to automatically extract node attributes in graphs. Node embedding algorithms project nodes into low-dimensional vectors, such that nodes with similar topological structures are in proximity in the embedding space [68]. We used the Python framework Karate Club [69], which consists of at least 30 graph mining algorithms, for node and graph embedding. We compared the performance of four state-of-the-art node embedding algorithms: Walklets [70] and Node2Vec [71], which use sampled random walks to make the node embeddings; NetMF [72], a factorization-based model; and the recently proposed RandNE [73], which is based on a Gaussian random projection approach with the default dimension ordering.
Walklets. In this method, instead of the random walk process used in DeepWalk [74], sample node neighborhoods are approximated by skipping over nodes in each short random walk. Then the set of results of multiple skip lengths is used to train the model [13].
Node2Vec. This method is a modification of DeepWalk introducing parameters p and q to smoothly interpolate between breadth-first sampling and depth-first sampling. Parameter p controls the likelihood of immediately revisiting a node in the walk, whereas parameter q allows the search to differentiate between “inward” and “outward” nodes. In Node2Vec, a vector representation of a node is computed on the basis of the second order random walks in the graph, and the core assumption is that Node2Vec’s sampling strategy is based on a mixture of breadth-first sampling and depth-first sampling suited for structural equivalence (i.e., similar roles of nodes) and homophily (i.e., network community), respectively [75].
NetMF. This method is a matrix factorization-based algorithm based on the connection between DeepWalk’s implicit matrix and graph Laplacians [73]. NetMF uses a small subset of nodes and extracts embedding vectors by approximating the proximity between nodes and the subset with the help of graph Laplacians [76].
RandNE. This method of iterative random projection network embedding preserves high order proximity between nodes by using a Gaussian random projection method while reducing the time complexity [73].

3.5. Proposed Model

3.5.1. Modular Architecture

Our proposed model was developed by using PyTorch [77] and PyTorch Geometric [78]. The model takes a time series of fMRI volumes as input, in which each time series is a 2D matrix X of size T × N , where T is the number of time steps, and N is the number of brain regions. The tsfresh algorithm was used for statistical feature extraction for each node, and then high-level node features associated with each node were extracted with node embedding methods. The overall GCN model architecture for task fMRI classification is summarized in Figure 1.
The GCN model consists of three Conv layers with 92 neurons per layer. The Rectified Linear Unit (ReLU) and batch normalization layers are applied between each Conv layer to accelerate the convergence and enhance stability, and dropout layers are added after each Conv layer to reduce the inherent unnecessary complexity and redundant computation of our multilayer GCN model. Then a global mean pooling layer is applied to calculate the final graph representation vector. We performed experiments on the same computing machine equipped with a single NVIDIA Tesla T4 24 GB RAM GPU.

3.5.2. Training and Testing

This study used five-fold stratified cross-validation within a training/validation/test setup. Four-fifths of the available data were allocated to a training set within each fold. The remaining one-fifth of the data were partitioned with a 60:40 ratio into a validation set and a final test set. The hyperparameter search consisted of a grid of learning rate, dropout, and weight decay values. The model with the lowest loss in the validation set was considered the best model for the final test. The following ideal parameters were used: learning rate: 0.001, dropout: 0.65, and weight decay: 0.0. Furthermore, because batch size is among the most important hyperparameters to tune, we considered a set of values of batch sizes. The batch sizes used in all experiments were B = 16 ,   32 ,   48 ,   64 over 100 epochs, all using the Adam optimizer and reducing the learning rate on a plateau with a patience of 10. Furthermore, cross-entropy loss was used for the optimization function.

3.5.3. Evaluation Metrics

The metrics used for comparison embedding methods and evaluation of classification performance included accuracy, balanced accuracy, F1 scores (macro, micro, and weighted), Matthews correlation coefficient (MCC), precision, and recall. F1 macro and MCC have been widely considered as metrics to evaluate imbalanced datasets in which all classes are weighted equally [79,80]. Therefore, we applied accuracy, F1 macro, and MCC for further node embedding method comparisons and evaluation of GCN model performance. For statistical analysis, we used a significance threshold of 0.05. We also used the Shapiro–Wilk normality test [81] followed by the t-test to evaluate the statistical significance of the model’s classification performance in different scenarios.

4. Results

In this section, the experimental results are presented for the GCN model implementation and classification performance in different scenarios. Furthermore, the detailed information regarding the evaluation of node embeddings in the context of task fMRI decoding concerning gender and gF score differences is provided. Finally, we implemented classic univariate statistics to determine whether the difference in classification performance was statistically significant.

4.1. Classification of Task fMRI Data

The first set of results included the evaluation of our proposed GCN framework to classify which task the subject was performing during fMRI (7 classes) across node embedding techniques. The experiment was performed by using task fMRI data from the 302 participants, and the framework was set up by application of the four defined node embeddings regarding different batch sizes during training. The results are shown in Table 2. Table 2 illustrated that the RandNE and NetMF embedding methods outperformed the DeepWalk methods (Node2Vec and Walklets). This result might have been because DeepWalk-based methods require many sampled node neighborhoods to create node embedding vectors [82]. The F1 macro scores for RandNE and NetMF revealed similar performance across the GCN framework, and application of different batch sizes had a minor effect on the classification performance.
Figure 2 illustrates the effect of batch size on classification performance. As the number of batch sizes increased from 16 to 64, the F1 macro score and MCC increased. We also observed that using a batch size of 64 achieved superior results if any node embeddings were selected. However, our GCN model showed the best classification performance with NetMF when a batch size of 64 was chosen. We set up our GCN model and obtained the confusion matrix for task fMRI classification after the training step, as shown in Figure 3. The normalized confusion matrix indicated that the top confusions were between (1) the social and motor tasks and (2) the gambling and social tasks.

Performance Comparison

We compared the proposed GCN model with Logistic Regression (LR) that used L2 regularization, as our baseline model, to prove if the classification performance represented a noticeable improvement over the traditional machine learning model. LR works well as a baseline model since it is relatively easy to implement. The use of regularization prevents overfitting of the task fMRI data, so that the model features are shrunk towards zero and perform feature selection automatically. We evaluated the same brain decoding tasks and ran LR on our task fMRI dataset, splitting it into the train, validation, and test sets. To tune the regularization parameter, we used a range of values and perform a 5-fold cross-validation to achieve the optimal regularization parameter of 0.1. The result of L2-regularized LR showed a lower prediction accuracy in the seven-class classification task (97.7% vs. 86.4%, respectively, for GCN and LR with L2 regularization).

4.2. Effects of Group Membership on Classification

We performed experiments to evaluate the effects of gender and gF score on classification performance by using task fMRI data. The experiments were performed separately on the datasets described in Section 3.1. We applied the proposed GCN framework with the same hyperparameters above for all classification experiments.

4.2.1. Gender Predictions

Classification. We first assessed the predictive performance of our model on predicting gender. The classification performance of the GCN model was evaluated across the four node embedding methods, and batch sizes were varied during training (Table 3). Several observations were made. First, we observed that the average F1 macro of the classifier on both sub-datasets ranged from 79.5% to 97.9%. Second, the GCN model achieved the best classification performance with NetMF for both sub-datasets. Third, the GCN model was sensitive to the choice of batch size, such that the best performance was obtained with a batch size of 64 for male and female sub-datasets. Similar trends were observed in the performance of the GCN model for MCC in Figure 4. For MCC, the model performance across both sub-datasets ranged from 82% to 97% with batch sizes of 48 and 64, respectively.
Statistical analysis concerning MCC. According to the results of our model applied independently to female and male sub-datasets, the proposed GCN model had the best classification performance when NetMF was the node embedding method, and the batch size of 64 was selected during training. We set up the GCN model with NetMF and trained it with a batch size of 64, by using a learning rate of 0.001 for 100 epochs to classify task fMRI data for each sub-dataset separately. This process was performed iteratively a total of 35 times, and related MCC values were used to assess the statistical significance of the differences in classification performance. Figure 5A represents the results of this process. The GCN model performed relatively similarly, whereas each run showed varying performance between two sub-datasets.
To perform statistical significance testing, we used the Shapiro–Wilk normality test to assess normality. After assessing the statistical significance of the difference between classification performance of two sub-datasets (i.e., female and male), we performed a t-test, which indicated a significant difference ( p < 0.00001 ) . The null hypothesis for this test was that the mean of classification performance for two sub-datasets was identical. Together, these results revealed that differences between the male and female task fMRI data were significant, such that classification was more accurate for of female than male task fMRI data.

4.2.2. Fluid Intelligence Level Discrepancy

Classification. We evaluated the gF-score through the same procedures used for assessment of the influence of gender differences on classification task fMRI data. We set up the model and independently performed classification experiments on two sub-datasets: LM-gF and HM-gF. Table 4 shows the model’s performance regarding defined node embedding methods after training with a batch size range from 16 to 64 for LM-gF and HM-gF. Correspondingly, Figure 6 represents the visualization of the model’s performance for various node embeddings for MCC. The x-axis in the figures shows the batch sizes. The GCN classification showed high performance on LM-gF and HM-gF sub-datasets with RandNE and NetMF node embedding methods, which exhibited similar trends. In addition, the results indicated a change in performance of the model when the size of the batch increased from 16 to 64. The most striking observation was that for both sub-datasets, classification performance with RandNE achieved the best MCC.
Statistical analysis concerning MCC. To assess the influence of individuals’ gF-scores on the classification performance, we conducted the same procedure as the previous scenario for gender differences. However, we found that the GCN model had the best classification performance when RandNE was used as the node embedding method. Similarly, the model was set up and trained with a batch size of 48 and a learning rate of 0.001 for 100 epochs to classify task fMRI data by using LM-gF and HM-gF. We obtained two groups of values indicating the accuracy performance (i.e., MCC) of the classification model in different sub-datasets (Figure 5B). The Shapiro–Wilk normality test was performed to assess normality, and a t-test was used to assess the statistical significance of the differences in classification performance. The difference was found to be non-significant ( p = 0.604 ), at p < 0.05 . Therefore, the accuracy performance of classification task fMRI data for participants with lower and higher fluid intelligence was comparable. Thus, individuals’ gF scores do not affect task fMRI classification performance.

5. Discussion

5.1. Overview

In this study, we developed a GCN based model for classifying task fMRI data, or graph-structured data with associated nodal attributes. GCNs can aggregate higher-order information in “neighborhoods” from graph nodes representing regions of interest in the brain, and edges representing the functional connectivity [83,84,85,86]. Our study achieved an accuracy of 97.7% in a seven-class classification task, thus demonstrating a competitive classification performance for brain state decoding, with respect to those recently reported across task fMRI data by using the HCP dataset [8,40,41]. The comparisons of our multilayer GCN model with deep neural networks illustrate that node embedded features achieved better results than automatic feature extraction in DL. Our proposed method is clearly better than [40,41] that applied deep neural network models which obtained classification performances of 94.3% and 93.7%, respectively. Inspired by reference [8], our proposed model included three Conv layers, in which we first implemented several node embedding methods to extract the topological features of nodes and defined node weight. Then the first Conv layer was fed by using different node embedding weights instead of using the same weights for all nodes. To this end, we tested four node embeddings (i.e., NetMF, RandNE, Node2Vec, and Walklets) and observed that our GCN model using NetMF and RandNE tended to yield the best results for group membership classifying based on task fMRI data.
Furthermore, our findings confirmed the importance of selecting a proper node embedding method to extract topological features of graph nodes before feeding the GCN model, in agreement with previous research detecting influenza-like symptoms with a GNN model [82].

5.2. Effects of Individual Differences

We examined the effects of individual differences on task fMRI classification in terms of the gender and gF score discrepancy. With respect to gender differences, the performance of the proposed GCN model was tested on two sub-datasets (female/male) by considering four node embedding methods. The same procedure was applied on gF-associated sub-datasets (LM-gF/HM-gF). We observed a significant difference in task fMRI classification performance between gender sub-datasets. However, no significant difference was observed in classification performance at a 95% confidence interval, because the p-value was greater than 0.05.

5.3. Effects of Batch Size

Training a DL model involves selecting a large set of hyperparameters, among which batch size is important [87]. Batch size defines the number of training samples used in one iteration to update the internal network parameters. To achieve the best accuracy performance of the GCN model associated with the batch size values, we chose a sequence of batch sizes of 16, 32, 48, and 64, and applied GCN architectures to each dataset. This approach allowed us to obtain the best classification performance for each experiment. The trend of the batch size change influenced the classification performance for all considered datasets. The worst classification performance values were obtained with a batch size of 16, and the best results were achieved with batch sizes of 48 and 64.

5.4. Limitations and Future Work

Our current study has several limitations that should be considered in future research. Although this study examined several node embedding methods to represent the graph nodes as low dimensional vectors, we disregarded the influence of the dimensionality of the node embeddings. Although finding the optimal dimension for embedding methods is challenging, some studies have applied several embedding dimensions on various datasets and achieved varying performance [13,88]. Therefore, the node embedding method must be customized to our dataset in future work. Furthermore, although batch size is an important hyperparameter to be considered in training a DL model [87], and we refitted our GCN model with different batch sizes and analyzed the effects of the change in batch size on classification performance, more hyperparameters should be studied, such as the number of convolutional layers, pooling ratio, and different readout operations. Finally, we analyzed only the task fMRI dataset for 302 participants and concluded that gender differences can affect classification performance. However, the ability to generalize our findings should be studied over a large number of participants and evaluated using our decoding model for experimental conditions under each task fMRI.

6. Conclusions

We proposed a GCN model to decode task fMRI data from the HCP dataset. Four node embedding methods—NetMF, RandNE, Node2Vec, and Walklets—were used to extract the topological features of graph nodes. We compared the performance of the model with different node embeddings through experiments and assessed classification accuracy. Our GCN model not only performs better on classification than alternative methods but also offered a relatively simple GCN architecture in which dropout layers reduced the redundant computation of the model. We further examined whether individual differences affect task fMRI data classification performance. Several conclusions were drawn. First, the overall task fMRI classification of the GCN model resulted in an accuracy, F1 macro and MCC of 0.977, 0.978 and 0.974, respectively. Second, the most robust node embedding methods for task fMRI data were NetMF and RandNE, whereas the least robust node embedding method was Node2Vec. Third, the influence of gender differences on task fMRI classification performance was significant, whereas no significant difference was observed between gF score categories.
In general, the method of this study provides a robust graph neural network-based data analysis method and examined various node embedding methods to provide a more effective solution for analyzing task fMRI data. Developing methods to test and validate saliency methods used for explainable artificial intelligence is still an active area of research [89]. However, as these techniques mature, GCNs may represent an important new tool for modeling brain information processing, using architectures inspired by the structural and functional graph properties of the brain.

Author Contributions

M.S., W.K., F.V.F. and K.F. conducted data exploration and modeling. M.S. prepared the initial draft of the manuscript. W.K., F.V.F., K.F., P.A.H., B.D.S., L.C.-M. and P.K.D. supervised all aspects of manuscript preparation, revision, editing, and final content. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

P.K.D would like to thank the AFOSR for a CCN award, the Brain and Behavior Foundation for a Young Investigator Grant, and the NIH for K01 award.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goense, J.; Bohraus, Y.; Logothetis, N.K. fMRI at high spatial resolution implications for BOLD-models. Front. Comput. Neurosci. 2016, 10, 66. [Google Scholar] [CrossRef] [PubMed]
  2. Logothetis, N.K.; Pauls, J.; Augath, M.; Trinath, T.; Oeltermann, A. Neurophysiological investigation of the basis of the fMRI signal. Nature 2001, 412, 150–157. [Google Scholar] [CrossRef] [PubMed]
  3. Kriegeskorte, N.; Douglas, P.K. Interpreting encoding and decoding models. Curr. Opin. Neurobiol. 2019, 55, 167–179. [Google Scholar] [CrossRef] [PubMed]
  4. Douglas, P.K.; Lau, E.; Anderson, A.; Head, A.; Kerr, W.; Wollner, M.A.; Moyer, D.; Li, W.; Durnhofer, M.; Bramen, J.; et al. Single trial decoding of belief decision making from EEG and fMRI data using independent components features. Front. Hum. Neurosci. 2013, 7, 392. [Google Scholar] [CrossRef]
  5. Colby, J.B.; Rudie, J.D.; Brown, J.A.; Douglas, P.K.; Cohen, M.S.; Shehzad, Z. Insights into multimodal imaging classification of ADHD. Front. Syst. Neurosci. 2012, 6, 59. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Tetrel, L.; Thirion, B.; Bellec, P. Functional annotation of human cognitive states using deep graph convolution. NeuroImage 2021, 231, 117847. [Google Scholar] [CrossRef]
  7. Li, X.; Dvornek, N.C.; Zhou, Y.; Zhuang, J.; Ventola, P.; Duncan, J.S. Graph Neural Network for Interpreting Task-fMRI Biomarkers. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2019, Shenzhen, China, 13–17 October 2019; Volume 11768, pp. 485–493. [Google Scholar] [CrossRef]
  8. Li, X.; Zhou, Y.; Dvornek, N.; Zhang, M.; Gao, S.; Zhuang, J.; Scheinost, D.; Staib, L.H.; Ventola, P.; Duncan, J.S. BrainGNN: Interpretable Brain Graph Neural Network for fMRI Analysis. Med. Image Anal. 2021, 74, 102233. [Google Scholar] [CrossRef]
  9. Li, X.; Zhou, Y.; Dvornek, N.C.; Zhang, M.; Zhuang, J.; Ventola, P.; Duncan, J.S. Pooling Regularized Graph Neural Network for fMRI Biomarker Analysis. In Alzheimer’s Disease and Frontotemporal Dementia; Humana Press: Totowa, NJ, USA, 2020; Volume 12267, pp. 625–635. [Google Scholar] [CrossRef]
  10. Kim, B.-H.; Ye, J.C.; Kim, J.-J. Learning Dynamic Graph Representation of Brain Connectome with Spatio-Temporal Attention. In Proceedings of the Advances in Neural Information Processing Systems 34 (NeurIPS 2021), Virtual Event, 6–14 December 2021. [Google Scholar]
  11. Li, X.; Dvornek, N.C.; Zhuang, J.; Ventola, P.; Duncan, J. Graph embedding using Infomax for ASD classification and brain functional difference detection. In Proceedings of the Medical Imaging 2020: Biomedical Applications in Molecular, Structural, and Functional Imaging, Houston, TX, USA, 15–20 February 2020; Volume 11317, p. 1131702. [Google Scholar] [CrossRef]
  12. Xu, K.; Jegelka, S.; Hu, W.; Leskovec, J. How Powerful are Graph Neural Networks? arXiv 2018, arXiv:1810.00826. [Google Scholar] [CrossRef]
  13. Goyal, P.; Ferrara, E. Graph embedding techniques, applications, and performance: A survey. Knowl.-Based Syst. 2018, 151, 78–94. [Google Scholar] [CrossRef]
  14. Van Essen, D.C.; Smith, S.M.; Barch, D.M.; Behrens, T.E.J.; Yacoub, E.; Ugurbil, K. The WU-Minn Human Connectome Project: An overview. NeuroImage 2013, 80, 62–79. [Google Scholar] [CrossRef]
  15. Tomasi, D.; Volkow, N.D. Gender differences in brain functional connectivity density. Hum. Brain Mapp. 2011, 33, 849–860. [Google Scholar] [CrossRef]
  16. Bruzzone, S.E.P.; Lumaca, M.; Brattico, E.; Vuust, P.; Kringelbach, M.L.; Bonetti, L. Dissociated brain functional connectivity of fast versus slow frequencies underlying individual differences in fluid intelligence: A DTI and MEG study. Sci. Rep. 2022, 12, 4746. [Google Scholar] [CrossRef]
  17. Zhang, X.; Liang, M.; Qin, W.; Wan, B.; Yu, C.; Ming, D. Gender Differences Are Encoded Differently in the Structure and Function of the Human Brain Revealed by Multimodal MRI. Front. Hum. Neurosci. 2020, 14, 244. [Google Scholar] [CrossRef]
  18. Santarnecchi, E.; Emmendorfer, A.; Tadayon, S.; Rossi, S.; Rossi, A.; Pascual-Leone, A. Network connectivity correlates of variability in fluid intelligence performance. Intelligence 2017, 65, 35–47. [Google Scholar] [CrossRef]
  19. Jiang, R.; Calhoun, V.D.; Fan, L.; Zuo, N.; Jung, R.; Qi, S.; Lin, D.; Li, J.; Zhuo, C.; Song, M.; et al. Gender Differences in Connectome-based Predictions of Individualized Intelligence Quotient and Sub-domain Scores. Cereb. Cortex 2019, 30, 888–900. [Google Scholar] [CrossRef]
  20. Farahani, F.V.; Fafrowicz, M.; Karwowski, W.; Douglas, P.K.; Domagalik, A.; Beldzik, E.; Oginska, H.; Marek, T. Effects of chronic sleep restriction on the brain functional network, as revealed by graph theory. Front. Neurosci. 2019, 13, 1087. [Google Scholar] [CrossRef]
  21. Sen, B.; Parhi, K.K. Predicting Male vs. Female from Task-fMRI Brain Connectivity. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 4089–4092. [Google Scholar] [CrossRef]
  22. Anderson, A.E.; Diaz-Santos, M.; Frei, S.; Dang, B.H.; Kaur, P.; Lyden, P.; Buxton, R.; Douglas, P.K.; Bilder, R.M.; Esfandiari, M.; et al. Hemodynamic latency is associated with reduced intelligence across the lifespan: An fMRI DCM study of aging, cerebrovascular integrity, and cognitive ability. Brain Struct. Funct. 2020, 225, 1705–1717. [Google Scholar] [CrossRef]
  23. Betzel, R.F.; Byrge, L.; He, Y.; Goñi, J.; Zuo, X.-N.; Sporns, O. Changes in structural and functional connectivity among resting-state networks across the human lifespan. NeuroImage 2014, 102, 345–357. [Google Scholar] [CrossRef]
  24. Christov-Moore, L.; Reggente, N.; Douglas, P.K.; Feusner, J.D.; Iacoboni, M. Predicting Empathy from Resting State Brain Connectivity: A Multivariate Approach. Front. Integr. Neurosci. 2020, 14, 3. [Google Scholar] [CrossRef]
  25. Friston, K.J.; Holmes, A.P.; Worsley, K.J.; Poline, J.-P.; Frith, C.D.; Frackowiak, R.S.J. Statistical parametric maps in functional imaging: A general linear approach. Hum. Brain Mapp. 1995, 2, 189–210. [Google Scholar] [CrossRef]
  26. Barch, D.M.; Burgess, G.C.; Harms, M.; Petersen, S.E.; Schlaggar, B.L.; Corbetta, M.; Glasser, M.F.; Curtiss, S.; Dixit, S.; Feldt, C.; et al. Function in the human connectome: Task-fMRI and individual differences in behavior. NeuroImage 2013, 80, 169–189. [Google Scholar] [CrossRef]
  27. Hu, X.; Lv, C.; Cheng, G.; Lv, J.; Guo, L.; Han, J.; Liu, T. Sparsity-Constrained fMRI Decoding of Visual Saliency in Naturalistic Video Streams. IEEE Trans. Auton. Ment. Dev. 2015, 7, 65–75. [Google Scholar] [CrossRef]
  28. Lv, J.; Lin, B.; Li, Q.; Zhang, W.; Zhao, Y.; Jiang, X.; Guo, L.; Han, J.; Hu, X.; Guo, C.; et al. Task fMRI data analysis based on supervised stochastic coordinate coding. Med. Image Anal. 2017, 38, 1–16. [Google Scholar] [CrossRef]
  29. Zhang, W.; Lv, J.; Li, X.; Zhu, D.; Jiang, X.; Zhang, S.; Zhao, Y.; Guo, L.; Ye, J.; Hu, D.; et al. Experimental Comparisons of Sparse Dictionary Learning and Independent Component Analysis for Brain Network Inference From fMRI Data. IEEE Trans. Biomed. Eng. 2018, 66, 289–299. [Google Scholar] [CrossRef]
  30. Anderson, A.; Han, D.; Douglas, P.K.; Bramen, J.; Cohen, M.S. Real-time functional MRI classification of brain states using Markov-SVM hybrid models: Peering inside the rt-fMRI black box. In Machine Learning and Interpretation in Neuroimaging; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7263, pp. 242–255. [Google Scholar] [CrossRef]
  31. Beckmann, C.F.; DeLuca, M.; Devlin, J.T.; Smith, S.M. Investigations into resting-state connectivity using independent component analysis. Philos. Trans. R. Soc. B Biol. Sci. 2005, 360, 1001–1013. [Google Scholar] [CrossRef]
  32. Calhoun, V.D.; Liu, J.; Adali, T. A review of group ICA for fMRI data and ICA for joint inference of imaging, genetic, and ERP data. Neuroimage 2009, 45, S163–S172. [Google Scholar] [CrossRef]
  33. Calhoun, V.D.; Adali, T. Multisubject Independent Component Analysis of fMRI: A Decade of Intrinsic Networks, Default Mode, and Neurodiagnostic Discovery. IEEE Rev. Biomed. Eng. 2012, 5, 60–73. [Google Scholar] [CrossRef]
  34. Douglas, P.; Harris, S.; Yuille, A.; Cohen, M.S. Performance comparison of machine learning algorithms and number of independent components used in fMRI decoding of belief vs. disbelief. NeuroImage 2011, 56, 544–553. [Google Scholar] [CrossRef]
  35. Anderson, A.; Douglas, P.; Kerr, W.; Haynes, V.S.; Yuille, A.L.; Xie, J.; Wu, Y.N.; Brown, J.A.; Cohen, M.S. Non-negative matrix factorization of multimodal MRI, fMRI and phenotypic data reveals differential changes in default mode subnetworks in ADHD. NeuroImage 2013, 102, 207–219. [Google Scholar] [CrossRef]
  36. Sen, B.; Parhi, K.K. Extraction of common task signals and spatial maps from group fMRI using a PARAFAC-based tensor decomposition technique. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 1113–1117. [Google Scholar] [CrossRef]
  37. Sen, B.; Parhi, K.K. Predicting Tasks from Task-fMRI Using Blind Source Separation. In Proceedings of the 2019 53rd Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 3–6 November 2019; pp. 2201–2205. [Google Scholar] [CrossRef]
  38. Kriegeskorte, N.; Douglas, P. Cognitive computational neuroscience. Nat. Neurosci. 2018, 21, 1148–1160. [Google Scholar] [CrossRef]
  39. Xu, M. Understanding Graph Embedding Methods and Their Applications. SIAM Rev. 2021, 63, 825–853. [Google Scholar] [CrossRef]
  40. Huang, X.; Xiao, J.; Wu, C. Design of Deep Learning Model for Task-Evoked fMRI Data Classification. Comput. Intell. Neurosci. 2021, 2021, 6660866. [Google Scholar] [CrossRef] [PubMed]
  41. Wang, X.; Liang, X.; Jiang, Z.; Nguchu, B.A.; Zhou, Y.; Wang, Y.; Wang, H.; Li, Y.; Zhu, Y.; Wu, F.; et al. Decoding and mapping task states of the human brain via deep learning. Hum. Brain Mapp. 2019, 41, 1505–1519. [Google Scholar] [CrossRef] [PubMed]
  42. Huang, H.; Hu, X.; Zhao, Y.; Makkie, M.; Dong, Q.; Zhao, S.; Guo, L.; Liu, T. Modeling Task fMRI Data Via Deep Convolutional Autoencoder. IEEE Trans. Med. Imaging 2017, 37, 1551–1561. [Google Scholar] [CrossRef] [PubMed]
  43. Huang, H.; Hu, X.; Dong, Q.; Zhao, S.; Zhang, S.; Zhao, Y.; Quo, L.; Liu, T. Modeling task fMRI data via mixture of deep expert networks. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 82–86. [Google Scholar] [CrossRef]
  44. Zhao, Y.; Li, X.; Huang, H.; Zhang, W.; Zhao, S.; Makkie, M.; Zhang, M.; Li, Q.; Liu, T. Four-Dimensional Modeling of fMRI Data via Spatio–Temporal Convolutional Neural Networks (ST-CNNs). IEEE Trans. Cogn. Dev. Syst. 2019, 12, 451–460. [Google Scholar] [CrossRef] [PubMed]
  45. Wang, H.; Zhao, S.; Dong, Q.; Cui, Y.; Chen, Y.; Han, J.; Xie, L.; Liu, T. Recognizing Brain States Using Deep Sparse Recurrent Neural Network. IEEE Trans. Med. Imaging 2018, 38, 1058–1068. [Google Scholar] [CrossRef] [PubMed]
  46. Li, Q.; Dong, Q.; Ge, F.; Qiang, N.; Wu, X.; Liu, T. Simultaneous spatial-temporal decomposition for connectome-scale brain networks by deep sparse recurrent auto-encoder. Brain Imaging Behav. 2021, 15, 2646–2660. [Google Scholar] [CrossRef]
  47. Hjelm, R.D.; Calhoun, V.D.; Salakhutdinov, R.; Allen, E.A.; Adali, T.; Plis, S.M. Restricted Boltzmann machines for neuroimaging: An application in identifying intrinsic networks. NeuroImage 2014, 96, 245–260. [Google Scholar] [CrossRef]
  48. Jang, H.; Plis, S.M.; Calhoun, V.D.; Lee, J.-H. Task-specific feature extraction and classification of fMRI volumes using a deep neural network initialized with a deep belief network: Evaluation using sensorimotor tasks. NeuroImage 2016, 145, 314–328. [Google Scholar] [CrossRef]
  49. Dong, Q.; Ge, F.; Ning, Q.; Zhao, Y.; Lv, J.; Huang, H.; Yuan, J.; Jiang, X.; Shen, D.; Liu, T.; et al. Modeling Hierarchical Brain Networks via Volumetric Sparse Deep Belief Network. IEEE Trans. Biomed. Eng. 2019, 67, 1739–1748. [Google Scholar] [CrossRef]
  50. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef]
  51. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The Graph Neural Network Model. IEEE Trans. Neural Netw. 2009, 20, 61–80. [Google Scholar] [CrossRef]
  52. Bi, X.; Liu, Z.; He, Y.; Zhao, X.; Sun, Y.; Liu, H. GNEA: A Graph Neural Network with ELM Aggregator for Brain Network Classification. Complexity 2020, 2020, 8813738. [Google Scholar] [CrossRef]
  53. Zhu, S.; Pan, S.; Zhou, C.; Wu, J.; Cao, Y.; Wang, B. Graph geometry interaction learning. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Virtual Event, 6–12 December 2020. [Google Scholar]
  54. Ning, L.; Bonet-Carne, E.; Grussu, F.; Sepehrband, F.; Kaden, E.; Veraart, J.; Blumberg, S.B.; Khoo, C.S.; Palombo, M.; Kokkinos, I.; et al. Cross-scanner and cross-protocol multi-shell diffusion MRI data harmonization: Algorithms and results. NeuroImage 2020, 221, 117128. [Google Scholar] [CrossRef]
  55. Glasser, M.F.; Sotiropoulos, S.N.; Wilson, J.A.; Coalson, T.S.; Fischl, B.; Andersson, J.L.; Xu, J.; Jbabdi, S.; Webster, M.; Polimeni, J.R.; et al. The minimal preprocessing pipelines for the Human Connectome Project. NeuroImage 2013, 80, 105–124. [Google Scholar] [CrossRef]
  56. Beckmann, C.F.; Smith, S.M. Probabilistic Independent Component Analysis for Functional Magnetic Resonance Imaging. IEEE Trans. Med. Imaging 2004, 23, 137–152. [Google Scholar] [CrossRef]
  57. Gross, J.L.; Yellen, J.; Anderson, M. Graph Theory and Its Applications; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018. [Google Scholar] [CrossRef]
  58. Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral networks and deep locally connected networks on graphs. arXiv 2014, arXiv:1312.6203. [Google Scholar]
  59. Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  60. Ma, Y.; Hao, J.; Yang, Y.; Alibaba, H.L.; Alibaba, J.J.; Tencent, G.C. Spectral-based Graph Convolutional Network for Directed Graphs. arXiv 2019, arXiv:1907.08990. [Google Scholar]
  61. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  62. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017—Conference Track Proceedings, Toulon, France, 24–26 April 2017. [Google Scholar]
  63. Salimi-Khorshidi, G.; Douaud, G.; Beckmann, C.F.; Glasser, M.F.; Griffanti, L.; Smith, S.M. Automatic denoising of functional MRI data: Combining independent component analysis and hierarchical fusion of classifiers. NeuroImage 2014, 90, 449–468. [Google Scholar] [CrossRef]
  64. Glasser, M.F.; Coalson, T.S.; Robinson, E.C.; Hacker, C.D.; Harwell, J.; Yacoub, E.; Ugurbil, K.; Andersson, J.; Beckmann, C.F.; Jenkinson, M.; et al. A multi-modal parcellation of human cerebral cortex. Nature 2016, 536, 171–178. [Google Scholar] [CrossRef]
  65. Christ, M.; Braun, N.; Neuffer, J.; Kempa-Liehr, A.W. Time Series FeatuRe Extraction on basis of Scalable Hypothesis tests (tsfresh—A Python package). Neurocomputing 2018, 307, 72–77. [Google Scholar] [CrossRef]
  66. Christ, M.; Kempa-Liehr, A.W.; Feindt, M. Distributed and parallel time series feature extraction for industrial big data applications. arXiv 2016, arXiv:1610.07717. [Google Scholar] [CrossRef]
  67. Benjamini, Y.; Yekutieli, D. The control of the false discovery rate in multiple testing under dependency. Ann. Stat. 2001, 29, 1165–1188. [Google Scholar] [CrossRef]
  68. Cai, H.; Zheng, V.W.; Chang, K.C.-C. A Comprehensive Survey of Graph Embedding: Problems, Techniques, and Applications. IEEE Trans. Knowl. Data Eng. 2018, 30, 1616–1637. [Google Scholar] [CrossRef]
  69. Rozemberczki, B.; Kiss, O.; Sarkar, R. Karate Club: An API Oriented Open-Source Python Framework for Unsupervised Learning on Graphs. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2020; pp. 3125–3132. [Google Scholar] [CrossRef]
  70. Perozzi, B.; Kulkarni, V.; Chen, H.; Skiena, S. Don’t walk, skip! online learning of multi-scale network embeddings. In Proceedings of the 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2017, Sydney, Australia, 31 July–3 August 2017; pp. 258–265. [Google Scholar] [CrossRef]
  71. Grover, A.; Leskovec, J. Node2vec: Scalable Feature Learning for Networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2022. [Google Scholar] [CrossRef]
  72. Qiu, J.; Dong, Y.; Ma, H.; Li, J.; Wang, K.; Tang, J. Network embedding as matrix factorization: Unifying DeepWalk, LINE, PTE, and node2vec. In Proceedings of the WSDM 2018: The Eleventh ACM International Conference on Web Search and Data Mining, Marina Del Rey, CA, USA, 5–9 February 2018; pp. 459–467. [Google Scholar] [CrossRef]
  73. Zhang, Z.; Cui, P.; Li, H.; Wang, X.; Zhu, W. Billion-Scale Network Embedding with Iterative Random Projection. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; pp. 787–796. [Google Scholar] [CrossRef]
  74. Perozzi, B.; Al-Rfou, R.; Skiena, S. DeepWalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar] [CrossRef]
  75. Dalmia, A.; Ganesh, J.; Gupta, M. Towards Interpretation of Node Embeddings. In Proceedings of the WWW’18: Companion Proceedings of the Web Conference 2018, Lyon, France, 23–27 April 2013; pp. 945–952. [Google Scholar] [CrossRef]
  76. Yu, B.; Lu, B.; Zhang, C.; Li, C.; Pan, K. Node proximity preserved dynamic network embedding via matrix perturbation. Knowl.-Based Syst. 2020, 196, 105822. [Google Scholar] [CrossRef]
  77. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar] [CrossRef]
  78. Fey, M.; Lenssen, J.E. Fast Graph Representation Learning with PyTorch Geometric. arXiv 2019, arXiv:1903.02428. [Google Scholar] [CrossRef]
  79. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 1–13. [Google Scholar] [CrossRef]
  80. Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems; O’Reilly: Sebastopol, CA, USA, 2017. [Google Scholar]
  81. Shapiro, S.S.; Wilk, M.B. An Analysis of Variance Test for Normality (Complete Samples). Biometrika 1965, 52, 591. [Google Scholar] [CrossRef]
  82. Dong, G.; Cai, L.; Datta, D.; Kumar, S.; Barnes, L.E.; Boukhechba, M. Influenza-like symptom recognition using mobile sensing and graph neural networks. In Proceedings of the Conference on Health, Inference, and Learning, Virtual Event, 8–10 April 2021; Volume 21, pp. 291–300. [Google Scholar] [CrossRef]
  83. Zhang, Y.; Huang, H. New Graph-Blind Convolutional Network for Brain Connectome Data Analysis. In Proceedings of the 26th International Conference, IPMI 2019, Hong Kong, China, 2–7 June 2019; Volume 11492, pp. 669–681. [Google Scholar] [CrossRef]
  84. Yu, S.; Yue, G.; Elazab, A.; Song, X.; Wang, T.; Lei, B. Multi-scale Graph Convolutional Network for Mild Cognitive Impairment Detection. In Proceedings of the Graph Learning in Medical Imaging, GLMI 2019, Shenzhen, China, 13–17 October 2019; Volume 11849, pp. 79–87. [Google Scholar] [CrossRef]
  85. Yu, S.; Wang, S.; Xiao, X.; Cao, J.; Yue, G.; Liu, D.; Wang, T.; Xu, Y.; Lei, B. Multi-scale Enhanced Graph Convolutional Network for Early Mild Cognitive Impairment Detection. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2020, Lima, Peru, 4–8 October 2020; Volume 12267, pp. 228–237. [Google Scholar] [CrossRef]
  86. Xing, X.; Li, Q.; Wei, H.; Zhang, M.; Zhan, Y.; Zhou, X.S.; Xue, Z.; Shi, F. Dynamic spectral graph convolution networks with assistant task training for early MCI diagnosis. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2019, Shenzhen, China, 13–17 October 2019; Volume 11767, pp. 639–646. [Google Scholar] [CrossRef]
  87. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  88. Martin, C.; Riebeling, M. A Process for the Evaluation of Node Embedding Methods in the Context of Node Classification. arXiv 2020, arXiv:2005.14683. [Google Scholar] [CrossRef]
  89. Douglas, P.K.; Farahani, F.V. On the Similarity of Deep Learning Representations Across Didactic and Adversarial Examples. arXiv 2020, arXiv:2002.06816. [Google Scholar] [CrossRef]
Figure 1. Overview of our Graph Convolutional Network (GCN) model for task fMRI classification. After acquisition of the raw task fMRI data and identification of the brain’s divisions into various parcels, several time courses of each parcel were extracted (A) to create the functional connectivity matrix. To reduce the complexity of the graph, a threshold was applied to the connectivity matrix (B) and transferred to a graph. The initial representation of each node was extracted by using the FRESH algorithm and node embedding methods (C). Finally, the feature vectors were used to perform the classification task with the proposed GCN framework including three Conv layers followed by a dropout layer after each Conv layer (D).
Figure 1. Overview of our Graph Convolutional Network (GCN) model for task fMRI classification. After acquisition of the raw task fMRI data and identification of the brain’s divisions into various parcels, several time courses of each parcel were extracted (A) to create the functional connectivity matrix. To reduce the complexity of the graph, a threshold was applied to the connectivity matrix (B) and transferred to a graph. The initial representation of each node was extracted by using the FRESH algorithm and node embedding methods (C). Finally, the feature vectors were used to perform the classification task with the proposed GCN framework including three Conv layers followed by a dropout layer after each Conv layer (D).
Brainsci 12 01094 g001
Figure 2. F1 macro (A) and MCC (B) comparison of the GCN model, taking into account the influence of node embedding methods and batch sizes for the task fMRI classification task, by using the 302 participants’ fMRI data.
Figure 2. F1 macro (A) and MCC (B) comparison of the GCN model, taking into account the influence of node embedding methods and batch sizes for the task fMRI classification task, by using the 302 participants’ fMRI data.
Brainsci 12 01094 g002
Figure 3. Confusion matrix of the GCN classification results on the 302 participants’ task fMRI data, normalized to the seven tasks in the five-fold cross-validation. The top two confusions were caused by the social task versus motor task and the gambling task versus social task. The F1 macro and MCC of classification were 0.977 and 0.974, respectively.
Figure 3. Confusion matrix of the GCN classification results on the 302 participants’ task fMRI data, normalized to the seven tasks in the five-fold cross-validation. The top two confusions were caused by the social task versus motor task and the gambling task versus social task. The F1 macro and MCC of classification were 0.977 and 0.974, respectively.
Brainsci 12 01094 g003
Figure 4. MCC comparison of the GCN model, taking into account the effects of node embedding methods and batch sizes, by using female task fMRI data (A) and male task fMRI data (B).
Figure 4. MCC comparison of the GCN model, taking into account the effects of node embedding methods and batch sizes, by using female task fMRI data (A) and male task fMRI data (B).
Brainsci 12 01094 g004
Figure 5. Box plots of the classification performance of the GCN model in 35 independent runs, by using gender sub-datasets (A) and fluid intelligence sub-datasets (B). Significant differences in classification performance of task fMRI data were observed between female and male data, but not between high and low fluid intelligence data.
Figure 5. Box plots of the classification performance of the GCN model in 35 independent runs, by using gender sub-datasets (A) and fluid intelligence sub-datasets (B). Significant differences in classification performance of task fMRI data were observed between female and male data, but not between high and low fluid intelligence data.
Brainsci 12 01094 g005
Figure 6. MCC comparison of the GCN model, taking into account the effects of node embedding methods and batch sizes, by using LM-gF task fMRI data (A) and HM-gF task fMRI data (B).
Figure 6. MCC comparison of the GCN model, taking into account the effects of node embedding methods and batch sizes, by using LM-gF task fMRI data (A) and HM-gF task fMRI data (B).
Brainsci 12 01094 g006
Table 1. Demographics and participant distribution results.
Table 1. Demographics and participant distribution results.
GroupsNumber of Participants Age   ( Mean   ±   S D ) gF - Score   ( Mean   ±   S D )
Female 164 29.2   ±   3.6 -
Male 138 28.1   ±   3.6 -
LM-gF F = 91 ,   M = 53 29.1   ±   3.6 12.5   ±   3.4
HM-gF F = 83 ,   M = 75 28.4   ±   3.6 20.6   ±   1.7
Abbreviations: LM-gF, low median-gF score; HM-gF, high median-gF score; F, female; M, male.
Table 2. Two-factor performance comparison on predicting experimental task, taking into account the influence of node embedding methods and batch sizes for the task fMRI classification. The training processes were set with 100 epochs, 10 step patience for early stopping, and learning rate = 0.001 for Adam. The proposed GCN model showed impressive results with both RandNE and NetMF node embedding methods. Classification performance values for 302 participants’ task fMRI data were in the range of 94% to 98%. Bold values represent the best classification performance obtained for each batch size.
Table 2. Two-factor performance comparison on predicting experimental task, taking into account the influence of node embedding methods and batch sizes for the task fMRI classification. The training processes were set with 100 epochs, 10 step patience for early stopping, and learning rate = 0.001 for Adam. The proposed GCN model showed impressive results with both RandNE and NetMF node embedding methods. Classification performance values for 302 participants’ task fMRI data were in the range of 94% to 98%. Bold values represent the best classification performance obtained for each batch size.
Batch SizeNode EmbeddingsMetrics
AccuracyF1 MacroMCC
16Walklets0.8860.890.867
Node2Vec0.8540.8630.831
RandNE0.9390.9410.928
NetMF0.9330.9360.921
32Walklets0.9110.9150.895
Node2Vec0.8730.8860.854
RandNE0.9690.970.954
NetMF0.9740.9760.97
48Walklets0.9150.920.901
Node2Vec0.8980.90.882
RandNE0.9710.9730.966
NetMF0.9760.9770.971
64Walklets0.9320.9360.922
Node2Vec0.9020.9080.886
RandNE0.9750.9760.971
NetMF0.9770.9780.974
Table 3. Two-factor performance comparison, taking into account the influence of node embedding methods and batch sizes in the GCN model, by using both the female and the male fMRI data. Bold values represent the best classification performance obtained for each batch size.
Table 3. Two-factor performance comparison, taking into account the influence of node embedding methods and batch sizes in the GCN model, by using both the female and the male fMRI data. Bold values represent the best classification performance obtained for each batch size.
Batch SizeNode EmbeddingsFemale DatasetMale Dataset
MetricsMetrics
AccuracyF1 MacroMCCAccuracyF1 MacroMCC
16Walklets0.8810.8820.8620.8490.850.823
Node2Vec0.7920.7950.7640.8410.8450.817
RandNE0.9160.9160.9020.9070.9090.891
NetMF0.9270.9280.9150.9080.9110.892
32Walklets0.9190.9190.9050.8790.8820.859
Node2Vec0.8350.8370.8070.8780.8790.856
RandNE0.9380.9390.9270.9490.9510.94
NetMF0.9590.9590.9520.939 0.941 0.928
48Walklets0.9310.9320.920.8870.8890.862
Node2Vec0.8710.8690.8510.8520.8570.827
RandNE0.9520.9520.9440.9550.9570.947
NetMF0.9670.9670.9610.9620.9640.953
64Walklets0.9280.9280.9160.8710.8740.844
Node2Vec0.8590.8610.8370.8450.8490.817
RandNE0.9710.9720.9660.9580.9610.951
NetMF0.9790.9790.9740.9620.9650.953
Table 4. Two-factor performance comparisons, taking into account the influence of node embedding methods and batch sizes in the GCN model, by using both LM-gF and HM-gF task fMRI data. Bold values represent the best classification performance obtained for each batch size.
Table 4. Two-factor performance comparisons, taking into account the influence of node embedding methods and batch sizes in the GCN model, by using both LM-gF and HM-gF task fMRI data. Bold values represent the best classification performance obtained for each batch size.
Batch SizeNode EmbeddingsLM-gF DatasetHM-gF Dataset
MetricsMetrics
AccuracyF1 MacroMCCAccuracyF1 MacroMCC
16Walklets0.8690.8730.8470.8760.8760.855
Node2Vec0.8950.8960.8770.8760.8780.855
RandNE0.9360.9370.9250.9450.9440.936
NetMF0.9060.9080.890.9210.9210.909
32Walklets0.8990.9020.8820.9010.9010.883
Node2Vec0.8910.8930.8690.920.920.907
RandNE0.9770.9770.9730.980.9790.975
NetMF0.930.930.9180.9420.9420.932
48Walklets0.9080.910.8910.920.9190.906
Node2Vec0.90.9020.8830.9150.9150.901
RandNE0.9910.9910.9890.9880.9880.983
NetMF0.9340.9340.9220.9480.9470.939
64Walklets0.8990.9010.8810.920.9180.907
Node2Vec0.9010.9030.8850.9060.9050.891
RandNE0.9910.9910.990.9880.9880.986
NetMF0.9360.9380.9250.9440.9430.935
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Saeidi, M.; Karwowski, W.; Farahani, F.V.; Fiok, K.; Hancock, P.A.; Sawyer, B.D.; Christov-Moore, L.; Douglas, P.K. Decoding Task-Based fMRI Data with Graph Neural Networks, Considering Individual Differences. Brain Sci. 2022, 12, 1094. https://doi.org/10.3390/brainsci12081094

AMA Style

Saeidi M, Karwowski W, Farahani FV, Fiok K, Hancock PA, Sawyer BD, Christov-Moore L, Douglas PK. Decoding Task-Based fMRI Data with Graph Neural Networks, Considering Individual Differences. Brain Sciences. 2022; 12(8):1094. https://doi.org/10.3390/brainsci12081094

Chicago/Turabian Style

Saeidi, Maham, Waldemar Karwowski, Farzad V. Farahani, Krzysztof Fiok, P. A. Hancock, Ben D. Sawyer, Leonardo Christov-Moore, and Pamela K. Douglas. 2022. "Decoding Task-Based fMRI Data with Graph Neural Networks, Considering Individual Differences" Brain Sciences 12, no. 8: 1094. https://doi.org/10.3390/brainsci12081094

APA Style

Saeidi, M., Karwowski, W., Farahani, F. V., Fiok, K., Hancock, P. A., Sawyer, B. D., Christov-Moore, L., & Douglas, P. K. (2022). Decoding Task-Based fMRI Data with Graph Neural Networks, Considering Individual Differences. Brain Sciences, 12(8), 1094. https://doi.org/10.3390/brainsci12081094

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop