Next Article in Journal
Categorical Smoothness of 4-Manifolds from Quantum Symmetries and the Information Loss Paradox
Previous Article in Journal
T-cell Receptor Is a Threshold Detector: Sub- and Supra-Threshold Stochastic Resonance in TCR-MHC Clusters on the Cell Surface
Previous Article in Special Issue
Nonparametric Causal Structure Learning in High Dimensions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Functional Connectivity Methods and Their Applications in fMRI Data

by
Yasaman Shahhosseini
and
Michelle F. Miranda
*
Department of Mathematics and Statistics, University of Victoria, Victoria, BC V8W 2Y2, Canada
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(3), 390; https://doi.org/10.3390/e24030390
Submission received: 15 January 2022 / Revised: 23 February 2022 / Accepted: 8 March 2022 / Published: 11 March 2022

Abstract

:
The availability of powerful non-invasive neuroimaging techniques has given rise to various studies that aim to map the human brain. These studies focus on not only finding brain activation signatures but also on understanding the overall organization of functional communication in the brain network. Based on the principle that distinct brain regions are functionally connected and continuously share information with each other, various approaches to finding these functional networks have been proposed in the literature. In this paper, we present an overview of the most common methods to estimate and characterize functional connectivity in fMRI data. We illustrate these methodologies with resting-state functional MRI data from the Human Connectome Project, providing details of their implementation and insights on the interpretations of the results. We aim to guide researchers that are new to the field of neuroimaging by providing the necessary tools to estimate and characterize brain circuitry.

1. Introduction

Functional magnetic resonance imaging (fMRI) techniques have emerged as a powerful tool for the characterization of human brain connectivity and its relationship to health, behavior, and lifestyle [1]. The fMRI measurements comprise of an indirect and non-invasive measurement of brain activity based on the blood oxygen level dependent (BOLD) contrast [2]. Compared to alternative brain imaging modalities such as positron emission tomography (PET) and eletroencephalography (EEG), fMRIs are non-invasive and have a high spatial resolution, which makes them a popular choice in large brain imaging studies. An example of such studies is the Human Connectome Project that aims at understanding the underlying function of the brain by describing the patterns of connectivity in the healthy adult human brain [3].
There are mainly two goals in such studies: first, to identify location signatures in the brain that respond to external stimuli, and second, to identify brain space–time association patterns that emerge when the brain is either at rest or performing a task. These association patterns are measures of co-activation in functionally connected time series of anatomically different brain regions, known as functional connectivity [4,5]. There is evidence that individual differences in these connectivity patterns are responsible for important differences in cognitive and behavioral functions. Therefore, understanding these patterns can play an important role in predicting the early onset of neurodegenerative diseases and in monitoring disease care and treatment [6,7].
Functional MRI data is often high-dimensional and consists of images of 3D brain volumes collected over a period of time. In a typical study, the number of voxels N v is in the hundred of thousands, and the number of time points T is in the hundreds. Therefore, estimating the N v × N v correlation matrix of voxelwise spatial connectivities is challenging and requires a few strategies and assumptions. A simple technique is to first pre-specify regions called seeds and then compute the cross-correlation of seeds and the functional time series of every other voxel in the brain. This seed-based approach became popular due to its straightforward calculation and interpretation. Seeds can vary in size and be as small as a single voxel. If the seed is a region, it is customary to average the time courses of the region and use that as the reference time course to be correlated with all the other voxels. In order to improve scalability, it is also common to first parcellate the brain into small regions and use the average time series of these regions to estimate the networks. The seed-based method can be a helpful resource when comparing patterns of neuropathologies and the normal brain. For example, ref. [8] uses this method to show that connectivity between the hippocampus and other brain regions change with the early signs of Alzheimer’s disease when compared to control subjects. Despite the popularity of these approaches, there are various criticisms to the method. First, by focusing on pre-determined seeds, potential patterns that emerge in different brain regions are ignored [9]. Second, the method neglects the variability across voxels by averaging the time series in the ROIs. Third, the approach computes correlations between pairs of nodes and ignores the potential influence of other nodes in the entire network.
In contrast to region pre-specification, dimension reduction approaches characterize the spatial and/or temporal connectivity patterns by representing the data through a small number of latent components [10]. Principal component analysis (PCA) and independent component analysis (ICA) are the two most common of these methods. Both methods project the high-dimensional imaging data into a low-dimensional subspace. In PCA, this projection consists of orthogonal components that maximize the variance of the data projected into the low-dimensional subspace [11]. In ICA, the projection consists of components that are as spatially independent as possible [12]. Each of these components is then assembled into brain maps with the value in each voxel representing the relative amount of that particular voxel, which is modulated by the activation of that component. Compared to the seed-based approach, both PCA and ICA have the advantage of providing automated components with no need for the pre-specification of a seed region, i.e., these methods are data-driven. The authors in reference [13] used ICA to decompose brain networks into spatial sub-networks with similar functions in both the resting state and task fMRI data.
Other methods combine the brain parcellation strategy used in seed-based methods with dimension reduction approaches to characterize brain circuitry. Reference [10] uses an anatomical atlas to pre-determine clusters (ROIs) and then extract features from each cluster via principal components. Multiple extracted components were then used to estimate connectivity between these ROIs using the RV coefficient, a measure that summarizes the correlation among sets of features.
In addition to the methods utilized to estimate connectivity, it is common to characterize the functional networks by using the tools of graph theory. In a graph, brain networks are treated as a collection of nodes connected by edges. Commonly, the edges are defined by an estimated connectivity. Following the specification of the nodes, a binary matrix is obtained by thresholding the connectivity matrix. The binary graph is then used to compute various graph parameters that describe the nature of the brain network. These parameters express key characteristics of the network and usually include quantities that help determine if the graph nodes are connected in a random or small-world order. Random networks have a more globally connected pattern while small-world networks show a high level of local ordering [14]. Statistical network models take these graph measures as inputs for the prediction of global networks that characterize multiple individuals.
The goal of this paper is to provide an overview of the most commonly used methods to estimate and characterize functional connectivity in resting-state fMRI data. We illustrate these methods by analyzing data from a single-subject in the Human Connectome Project. Although we do not attempt to offer an exhaustive presentation of the rapidly evolving methods in the field, we expect that the information provided here will guide researchers that are new to the field of neuroimaging in exploring these data.
The remainder of the paper is organized as follows: In Section 2, we describe the different methods of estimating functional connectivity, focusing on data from a single subject. In Section 3, we estimate functional connectivity for a single-subject resting-state data from the Human Connectome Project, using the methods described in Section 2. In Section 4, we present a few multiple-subject estimation methods. In Section 5, we describe statistical network models. Finally, in Section 6, we present some final remarks on the topic.

2. Methods for Functional Connectivity

In this section, we review the different methods to estimate functional connectivity for single-subject data. We focus on data for a single subject and discuss group connectivity in Section 4. For all calculations, let the matrix Y be a matrix of size T × N v , consisting of N v time courses representing the BOLD signal at each voxel v = 1 , , N v [2] for a single subject. Here, for simplicity, we centralized the matrix Y by subtracting each voxel data (column) by the average of its corresponding time course. The goal of a connectivity-based analysis is to describe how various brain regions interact, either when the brain is resting or performing a task [15].

2.1. Seed-Based Analysis

It is computationally expensive to compute all pairwise correlations among a large number of voxels as it would require N v 2 operations. Seed-based analysis (SBA) relies on estimating pairwise correlations between regions of interest (ROIs) or between a seed region and all the other voxels across the brain.
To estimate correlations between ROIs, a common approach is to divide the brain volume according to anatomical templates, usually called brain atlases [16]. There are several human brain atlases available, including the popular Automated Anatomical Labelling (AAL), Tailarach Atlas, and the MN1-152 atlas [16,17]. To estimate correlations between a seed region and the other voxels, a seed is usually selected either by expert opinion or by choosing the voxel that shows greater activation during the fMRI experiment as the seed. The latter is more common during experiments involving tasks. After selection, the connectivity is estimated through a measure that quantifies correlation. Various measures were proposed in the literature. We provide more details about these measures in the Appendix A. We can summarize the seed-based approach the following way.
(i)
Choose a seed region or voxel;
(ii)
Correlate the time series of the region or voxel with all other voxels in the brain. If the seed is a region, average the time series of the region prior to correlating that with all other voxels in the brain. Use one of the measures described in Appendix A;
(iii)
Display the 3D volumes of the correlation measure or display the thresholded correlations (just the ones that are significant). Note: To determine significance, we need to account for multiple comparisons. Bonferroni and FDR are widely used procedures.
Alternatively, after dividing the brain into various ROIs using an atlas, we can summarize the time series of that region, either by averaging across voxels or by calculating the first principal component [15]. Next, we use those summary time series to be correlated between all regions. We illustrate both options in Section 3.

2.2. Decomposition Methods

Although seed-based methods have a straightforward interpretation, they are biased to the seed selection technique [18] and, therefore, should be used with caution. Principal component analysis and independent component analysis aim at solving the issue by providing a data-driven approach to functional connectivity. These decomposition methods play many roles in functional neuroimaging applications. They are used in the pre-processing steps to remove data artifacts and to reduce data dimensionality, and they will likely appear in at least one step of estimating functional connectivity in various populations. In this section, we will focus on their role as a method to describe functional connectivity in single-subject fMRI data, while in Section 4, we explore their contribution in multi-subject analysis.
As an alternative to seed-based analysis, the goal of the decomposition methods is to represent the voxel domain as a smaller subset of spatial components. Each spatial component has a separate time course and represents simultaneous changes in the fMRI signals of many voxels [12]. In this section, we assume that for each column of Y the average was subtracted from the data.

2.2.1. Principal Component Analysis (PCA)

PCA is a common method to reduce data dimensionality while minimizing the loss of data information and maximizing data variability [11]. The principal components are obtained either by the eigendecomposition of the sample covariance matrix Y T Y or by finding the eigenvectors of the data matrix Y using the theory of singular value decomposition (SVD). The rank of the data matrix is r = m i n { T , N v } (usually T < N v and r = T ) and therefore we can find r principal components through the decomposition
Y = U Σ W T = k = 1 r σ k u k w k T ,
where the T × r matrix U contains an orthonormal left singular vector u k T , the r × N v matrix W contains orthonormal right singular vectors w k N v , and the r × r diagonal matrix Σ contains the ordered singular values [11,15,19]. Notice that the eigendecomposition of Y T Y is defined as W T Σ 2 W . The orthonormal rows of the r × N v matrix W are referred to as eigenimages and can be assembled into brain maps, each representing the relative amount from a given voxel that is modulated by the activation of that component.
A different approach is to project the original fMRI data into the space spanned by the first p principal components, where the choice of p is based on the amount of data variability explained by the component. The projected data matrix, Y W , consists of the time series of regions in this new subspace. The authors in reference [20] used this idea to reduce the dimensionality of the fMRI data in certain ROIs and then applied a Granger causality analysis on the block time series of two brain regions to infer directional connections. Although it is possible to compute correlations using the time series of these projected data points, the results have no clear interpretation since each of these spatial regions in the new subspace represent a linear combination of different voxels in the original data space.

2.2.2. Independent Component Analysis (ICA)

ICA aims at representing the brain data using a latent representation of independent factors. Differently from PCA, the goal is to decompose Y as a product of a mixing matrix and a combination of spatially independent components (ICs).
Y = M C + E = k = 1 K m k c k + E ,
where M is the T × K mixing matrix with columns m k , and the K × N v matrix C is the matrix of independent components with rows c k , where each c k contains brain networks corresponding to component k for a total of K independent components. These components represent the networks of various functions, such as motor, vision, auditory, etc. The elements of the matrix E are independent Gaussian noises.
It is assumed that the component maps, c k , k = 1 , , K represent possible overlapping and statistically dependent signals, but the individual component map distributions are independent, i.e., if P ( c k ) represents the probability distribution of the voxels values in the kth component map, we have
P ( c 1 , c 2 , , c K ) = k = 1 K P ( c k ) .
Each independent component c k is a vector of size N v and can be assembled into brain maps. As in PCA, these maps represent the relative amount of a given voxel that is modulated by the activation of that component.

2.3. Computational Aspects

In imaging applications, estimating the principal components requires the decomposition of the N v × N v matrix Y T Y , which is usually unfeasible. Many algorithms were proposed in the literature to efficiently estimate the components in such high-dimensional settings. Ref. [21] develops the sparse PCA (SPCA), which is based on a regression optimization problem using a lasso penalty, [22] a multilevel functional principal component for high-dimensional settings, and [23] estimate a sparse set of principal components through an iterative thresholding algorithm. Some of these toolboxes are easy to access and available for downloading at the authors’ website.
Similarly, estimating the independent components is not straightforward, and ranking the components is challenging because the ICs are usually not orthogonal, and the sum of the variances explained by each component will not sum to the variance of the original data. One of the first algorithms was the Infomax, which aims at maximizing the joint entropy of suitably transformed component maps [12,24]. Recently, more modern algorithms focus on extracting a sparse set of features from data matrices containing a very large number of features. Examples are the ICA with a reconstruction cost (RICA) proposed by [25], which is available as a Matlab toolbox.

2.4. A Hybrid Method

A different approach to estimate functional connectivity is given by reference [10]. The authors propose a multi-scale model based on networks at multiple topological scales, from voxel level to regions consisting of clusters of voxels, and larger networks consisting of collections of those regions. In practice, these collections of voxels are pre-specified and usually taken as anatomical ROIs. These anatomical ROIs can be then combined to form clusters of ROIs. Their approach consists of a dimension reduction step through to a factor model within each ROI. Let r represent the r-th ROI and Y r be a T × p r data matrix consisting of the time series of voxels belonging to the r-th ROI (containing a total of p r voxels, where r = 1 R p r = N v and R is the total number of ROIs). Then, we write
Y r ( t ) = Q r f r ( t ) + E r ( t ) ,
where Y r ( t ) is a column vector of size p r , f r ( t ) is a m r × 1 vector of latent common factors with a number of factors m r p r , Q r is a p r × m r factor-loading matrix that defines the dependence between the p r voxels through the mixing of f r , and E r ( t ) = [ e r 1 ( t ) , , e r p r ( t ) ] is a p r × 1 vector of white noise with E ( E r ( t ) ) = 0 and Σ E r , E r = C o v ( E r ( t ) ) = diag ( σ e r 1 2 , , σ e r p r 2 ) .
These factor models are then concatenated to define
Y ( t ) = Q f ( t ) + E ( t ) ,
where Y ( t ) is a column vector of size r = 1 R p r = N v , Q = diag ( Q 1 , , Q R ) is a r = 1 R p r × r = 1 R m r block-diagonal mixing matrix and f ( t ) = [ f r ( t ) , , f R ( t ) ] is a r = 1 R m r × 1 vector of aggregated latent factors.
Network covariance matrices in these different topological scales are estimated using the low-rank matrix in the following way. Let Σ Y r Y r be the covariance matrix within ROI r. Model (4) implies the following decomposition
Σ Y r Y r = Q r Σ f r f r Q r + Σ E r E r .
Similarly, from Model (5) we have
Σ Y Y = Q Σ f f Q + Σ E E .
The low-dimensional factor covariance matrix Σ f f is a block matrix used to estimate the lag-zero dependency between ROIs as follows.
Σ f f = Σ f 1 f 1 Σ f 1 f R Σ f R f 1 Σ f R f R
The diagonal blocks Σ f r f r , r = 1 , , R are diagonal covariance matrices, capturing the total variance of factors within each ROI. The off-diagonal blocks Σ f k f j , j j are cross-covariance matrices between factors and summarize the dependence between clusters j and k.
The authors summarize the dependence between ROIs using the RV coefficient, a multivariate generalization of the squared correlation coefficient. The RV coefficient between factors in clusters j and k is defined by
R V = tr ( C f k f j , C f j f k ) tr ( C j k f j , C f j f j ) tr ( C f k f k , C f k f k ) ,
where C f j f k = ( Σ f j f j ) 1 2 Σ f j f k ( Σ f k f k ) 1 2 .
In practice, the authors apply this model to estimate resting-state networks. They estimate the factors f r and matrices Q r using PCA and pre-specify the ROIs based on an anatomical atlas. The authors in reference [26] use this approach to estimate background functional connectivity between ROIs using data from the Working Memory Task in the Human Connectome Project.

2.5. Brain Networks

It is common to represent the brain using tools from graph theory. In this framework, we can think of functional connectivity as a network represented by a graph, where the spatial units are the nodes and the connection between them are the edges. Networks are treated as a collection of nodes (vertices) connected by links (edges). The graph (network) is represented as the pair G = ( V , E ) , where V and E are the sets of vertices and edges, respectively. In addition, graphs may be weighted and, in such cases, will be denoted by the triple G = ( V , E , W ) , with W ( E ) indicating the weight for each edge.
The first decision to make is the selection of the nodes of the network. Similar to the seed-based connectivity, these nodes are defined by either voxels or the ROI parcellations given by anatomical atlases. Following the specification of the nodes, their edges (links) must be determined. These edges quantify the strength of association between these different nodes, i.e., they are the functional connectivity. The same measures discussed previously for functional connectivity and described in Appendix A are used to quantify the strength of the edges.
Most of the standard tools of graph theory have been developed for binary networks, where each edge is either present or not. A binary matrix, usually called an adjacency matrix, is obtained by thresholding the connectivity matrix. Although it is convenient to threshold the weighted graphs to apply the standard graph theoretical machinery, information about the original signal is discarded in the process. Moreover, in most situations, the choice of a threshold is not unique, and such a decision may be difficult to justify. One strategy is the use of a mass-univariate approach, in which a statistical test is performed for every possible edge in the network and then corrected for multiple comparisons using standard techniques, such as the Bonferroni correction or the false discovery rate (FDR) [27,28].
After the network is estimated, some descriptive measures are calculated as means to describe the topological graph properties. In brain networks, the popular metrics are the characteristic path length, the clustering coefficient, and the degree distribution. For a list of the comprehensive topological measures used in neuroimaging, see reference [29].
  • Characteristic path length. Paths are the sequences of distinct nodes that represent the potential flow of information between pairs of brain regions with shorter paths, implying stronger potential for integration. The length of a path estimates the potential for functional integration between brain regions. One of the most commonly used measures of functional integration is the average shortest path length between all pairs of nodes in the network, which is defined as the characteristic path length [15]. Paths between disconnected nodes are defined to have infinite length, which is a problem when calculating this measure, especially in sparse networks such as in functional connectivity. In practice, we take the average only between the existing paths, which can be a problem. For a discussion on this issue please refer to reference [29].
  • Degree distribution. A measure of centrality, the degree of an individual node is equal to the number of links connected to that node, i.e., the number of neighbors of the node. The degree distribution is, therefore, the distribution of the degrees of all the nodes in the network. In functional connectivity, nodes with a high degree are interacting functionally with many other nodes in the network [29] and are referred to as hubs.
  • Clustering coefficient. A measure of segregation, the clustering coefficient is the fraction of the node’s neighbors that are also neighbors of each other, which in graph theory is the fraction of triangles around an individual node. The presence of clusters in functional networks suggests an organization of statistical dependencies indicative of segregated functional neural processing, which is the ability for specialized processing to occur within densely interconnected groups of brain regions. The mean clustering coefficient for the network reflects, on average, the prevalence of clustered connectivity around individual nodes. The mean clustering coefficient is normalized individually for each node and can disproportionately be influenced by nodes with a low degree.
Many other network measures are greatly influenced by basic network characteristics, such as the number of nodes and links and the degree of distribution presented in this section.

3. Real Data Example

We analyzed the resting-state data from the Human Connectome Project (HCP). We chose to work with the data that had been previously denoised using the FIX pipeline [30]. This pipeline uses a gentle high-pass temporal filter, performs motion regression (i.e., the regression of 24 movement parameters: six rigid-body motion parameters, their backward temporal derivatives, and squares of those 12 time series), and applies a regression based on ICA to remove the variance in noise components that was orthogonal to signal components [31]. For the single-subject analysis, we considered the volumes collected from the right–left phase of the example, Subject 100307. Volumes of fMRI were obtained every 720 ms. Each volume consisted of images of size 91 × 109 × 91 for a total of 1200 time frames.

3.1. Single-Subject Examples

3.1.1. ROI-Based Connectivity

We partitioned the brain into ROIs using the AAL Atlas version that was registered into the MNI152 space. We considered a total of 166 ROIs and estimated the connectivity using the following methods:
(a)
Cross correlation of the average time series in each ROI;
(b)
Partial correlation of the average time series in each ROI;
(c)
Cross correlation of the time series of the ROI data projected into the space of its first principal component;
(d)
Partial correlation of the time series of the ROI data projected into the space of its first principal component;
(e)
For each ROI, we consider the principal components that account for 20% of the ROI variability and calculate the RV coefficient as described in Equation (8).
The results for the estimated connectivity values are shown in Figure 1. Inspecting Figure 1 reveals that cross-correlation measures in panels (a) and (c) capture larger correlations than their corresponding partial cross-correlation measures (panels (b) and (d)). The RV coefficient from the method described in (e) seems to be able to capture a large number of small correlations among ROIs. Before drawing any conclusions from the figure, we should first test whether these values are significant. For the first four matrices, the test is done by first transforming these values to z-scores and then thresholding them to identify important correlations. For the RV coefficient in panel (e), significance is based on the asymptotic distribution of the coefficient as detailed in reference [10].
Next, we used these connectivity matrices to obtain a binary graph with the edges determined based on the p-values obtained from the z-scores of the correlation coefficient, as described in Appendix A, Equation (A2). The p-values were thresholded based on the Bonferroni correction and a significance of 5%. For the RV coefficient in panel (e), we use the asymptotic distribution of the coefficients to convert the values to z-scores and thresholded based on the Bonferroni correction to find the quantile of the standard normal distribution with a significance of 5%. Considering this criteria, we compute the adjacency matrices shown in Figure 2.
Inspecting Figure 2 reveals the presence of a large number of edges for both (a) and (c) graphs. This indicates a high level of interaction between the different anatomical regions. This high-interaction level was not found in graphs (b) and (d). In panel (e), we observe a moderate level of interaction with a few ROIs connecting with many others, while some regions are quiet during the resting-state experiment.

3.1.2. Network Summary Measures

We used the binary graphs obtained above to estimate a few summary measures, using graph theory as described in Section 2.5. The formulas used in each calculation are detailed in Appendix B. Table 1 shows the results. CPL is the characteristic path length excluding all infinity paths from the network, DG is the average degree of the network, where the degree indicates the number of links in each node, CC is the average clustering coefficient of the network, and Inf is the number of infinity paths in the network. The quantities in Table 1 reflect what we observe in Figure 2. The degree indicates the number of connections between regions. As noticed before, the graphs in panels (a) and (c) indicate a high degree, with many interactions between ROIs. The characteristic path length (CPL) of the RV coefficient indicates that on average the network has a short path length, with a value that is comparable to the networks in panels (a) and (c) of Figure 2. This indicates that despite few regions being connected, the ones that are connected are near each other.

3.1.3. Volume-Based Connectivity

Seed-Based Analysis. For seed-based analysis, we chose the left pars opercularis (left interior frontal gyrus) as the seed [32]. We take the average time series for this region and compute the cross correlation with the remaining voxels. We perform a Bonferroni correction considering α = 0.05 to threshold the correlation values. Figure 3 shows the resulting brain map. We display clusters bigger than 125 as significant voxels, and their mask is overlaid on a template brain consisting of the average time points of the example subject data used here.
Decomposition Methods. We first obtain the principal components of the data matrix Y. It is important to notice that a large number of principal components is needed to represent data variability and that traditional principal components have the issues discussed in Section 2.3. For this particular data, 150 components are needed to represent 20% of the data variability and 463 are needed to represent 50%. We illustrate the first five components scaled by their eigenvalues (i.e., the loadings) in Figure 4.
Next, to estimate the independent components, we use the probabilistic independent components analysis proposed in reference [33] and implemented in the MELODIC (multivariate exploratory linear optimized decomposition into independent components) function in FSL. Figure 5 depicts the results.
For illustration purposes, we present the components without thresholding their values. It is more common to use the individual components’ maps as inputs in a multi-subject approach and then perform thresholding in the group components to identify a group network. We comment more on the topic in the next section.

4. Multiple-Subject Functional Connectivity

When modeling functional MRI, an important goal is to identify the functional connectivity structure in multi-subject data by leveraging a shared structure across subjects. Multi-subject functional connectivity models can range from constrained tensor decomposition models, e.g., PARAFAC, to more flexible approaches where subject-specific connectivity matrices or PCA and ICA models are estimated first, and their concatenated results are used as inputs on a group-based estimation. The optimal model will depend on which level of flexibility best captures the functional connectivity features within the group [34].
In multi-subject ICA models, a simple procedure is to estimate the single-subject connectivity matrix using pre-specified ROIs, as in the seed-based approach described in Section 2, and then aggregate those results into a single matrix, subsequently further decomposing this matrix using principal components. The principal components can then be mapped to estimate a group-based functional connectivity. Ref. [35] used this idea to estimate a dynamical group-based resting-state connectivity of minimally disabled relapsing–remitting patients.
A multi-stage approach is implemented in reference [36] to compare functional connectivity between subjects at a high familial risk for Alzheimer’s disease that are clinically asymptomatic versus matched controls. The method follows four steps, including subject-specific SVD, a population-level decomposition of aggregated subject-specific eigenvectors, a projection of the subject-level data onto the population eigenvectors to obtain subject-specific loadings, and the use of the subject-specific loadings in a functional logistic regression model.
A group of methods propose a group ICA approach, where fMRI data is either temporally concatenated across subjects or taken as a multi-dimensional array. The FMRIB Software Library (FSL), a software library containing image analysis and statistical tools for various imaging data, provides group ICA and tensorial ICA in its MELODIC function.This section will focus on these two approaches.

4.1. Group ICA

Ref. [37] proposed for the first time an approach to perform ICA on fMRI data from a group of subjects. Suppose we observe fMRI data from n subjects. Let Y i be a matrix of size T × N v consisting of N v time courses representing the BOLD signal at each voxel v = 1 , , N v for subject i = 1 , , n . Their model involves a multi-stage approach as follows.
  • Subject-level data reduction. In this step, reduction is applied in the temporal domain. For each subject i = 1 , . n , the reduced data is given by
    X i = F i 1 Y i ,
    where F i 1 is a L × T reducing matrix and X i is a L × N v matrix representing the reduced data. In practice, F 1 is obtained by PCA decomposition;
  • Data reduction of the aggregated subject-level data. Data reduction is applied to the N L × N v matrix [ X 1 T , , X N T ] T to obtain a K × N v matrix X = G 1 [ X 1 T , , X N T ] T , where K is the number of components to be obtained and G 1 is a K × N L -reducing matrix that is in practice obtained by principal components;
  • Estimation of independent sources. An ICA decomposition is applied to the matrix X, as described in Section 2.2.2.
    X = M C ,
    where M is a K × K -mixing matrix and C is a K × N v component map matrix. The resulting group ICA components can be thresholded by first converting them into Z-scores.
Individual maps can be obtained by partitioning G M (where G = ( G 1 ) T ) by subject and going back along the previous steps as follows.
G X = G M C = F 1 1 Y 1 F N 1 Y N .
Based on these steps, the matrix G M C is a matrix of size N L × N v of individual maps and can be partitioned such that G i M i C i = F i 1 Y i , and C i contains the individual maps.

4.2. Tensorial ICA

The tensor ICA is based on tensor decomposition, which obtains a low-rank representation of a multi-dimensional array. PARAFAC is a common decomposition method [38]. Let X R T × N v × N be an array with fMRI data and dimension times, voxels, and subjects, respectively. The three-dimensional array X can be decomposed as a sum of R outer products in the following way
X = r = 1 R a r b r c r ,
where a r R T , B r R N V , and c r R N . This decomposition implies that each element in the array X can be written as
{ x i j k } = r = 1 R a i r b j r c k r .
The vectors in the decomposition can be represented in matrices, e.g., A = [ a 1 a 2 a R ] , and likewise to obtain matrices B and C. The three-dimensional array can be unfolded into matrices in a process called matricization. The unfolding can happen in any of the three dimensions. On the second dimension, X ( 2 ) R N v × N T is the mode-two matricization of X. Similarly, we can use the unfolding to generate mode-two and mode-three matrices. For details on the PARAFAC decomposition and matricization, please refer to reference [38]. Using these definitions, we can write:
X ( 2 ) = B ( C A ) T ,
where ⊙ denotes the Katri–Rao product. In reference [39], the authors propose an ICA decomposition of the form
X = ( C A ) B T + E ,
where X = X ( 2 ) T and the mixing matrix M = ( C A ) and component matrix B T are estimated as in reference [33].

5. Statistical Network Models

In this section, we follow the notation in reference [40] to describe statistical network models with the purpose of characterizing brain circuitry. In these models, individual functional connectivity is estimated first, using the techniques described in Section 2. After individual estimation, the effects of multiple variables of interest and topological network features are taken into account on the overall network structure.
Let ( Y i , X i ) represent the network and covariates for subject i, respectively. The probability density function of the network given the covariates is denoted by P ( Y i | X i , θ i ) , where θ i describes the relationship of Y i and X i . These covariates can be node-specific covariates, such as brain location and also functions of the network Y i , such as the path length or other metrics described in Section 2.5. Popular ways of modeling the density function include exponential random graph models (ERGMs) and mixed models [40].
In ERGMs, we consider binary graphs and the models are fitted for each subject individually as follows. Let Y i be a network consisting of R × R nodes. Then, Y i j k = 1 if a link exists between nodes j and k, and Y i j k = 0 otherwise. The probability mass function has the form of a regular exponential family:
P ( Y i = y i | X i ) = κ ( θ ) 1 e x p θ T g ( y i , X i ) .
The estimation of the parameters θ is done by MCMC MLE. In reference [41], they identify the most important explanatory metrics g ( y i ) for each subject’s network. Next, the authors create a group-based summary measure of the fitted parameter values θ for all subjects. They use these group-based explanatory metrics and parameters to fit a group-based representative network via ERGMs.
One limitation of the current estimation methods for ERGMs is scalability. The major issue is not the number of ROIs per se but the edge structure of the network, which can cause convergence problems. Moreover, most models were developed for binary graphs and are not well-suited for link-level examination [40].
As an alternative to ERGMs, mixed models allow for both link-level examination and multiple-subject comparisons. The framework defines a two-part mixed effect that models both the probability of a connection being present or absent and the strength of a connection if it exists. Let Y i be a representation of the functional connectivity strength given by one of the correlation measures listed in Appendix A, and let R i j k be an indicator of whether a connection between j and k is present. Then the conditional probabilities are
P ( R i j k = r i j k | β r ; b r i ) = 1 p i j k ( β r ; b r i ) , if r i j k = 0 p i j k ( β r ; b r i ) , if r i j k = 1 ,
where β r are the vector of fixed effects that relate the covariates X i j k for each participant and pair of nodes, and b r i are random effects representing subject-specific and node-specific parameters.
Let Z i j k be the design matrix associated with the random effects b r i ; the models are divided into two parts. The first part of the model uses a logit link function to relate the probability of connection between nodes j and k to the covariates as follows.
logit ( p i j k ) = X i j k β r + Z i j k b r i .
The second part models the strength of the connection between nodes j and k given that the connection is present, by linearly linking the Fisher’s Z-transform of the correlation coefficient between nodes i and j and the covariates. Let S i j k = Y i j k | R i j k = 1 , then
Fisher s   Z transform ( S i j k ) = X i j k β s + Z i j k b s i + e i j k ,
where β r is a vector of population parameters that related the strength of connection to the same set of covariates X i j k for each participant and pair of nodes, b s i is a vector of subject and node-specific parameters that capture how this relationship varies about the population average β s , and e i j k is the random noise for subject i and nodes j and k. Details of the two-parts modeling approach is presented in reference [42].
One issue that arises from these models is that thresholding choices based on the connectivity weights will impact the network topology, even if multiple comparisons are taken into account. The authors in reference [40] argue that persistent homology provides a multi-scale hierarchical framework that addresses the threshold issue. The method is a technique of computational topology that provides a coherent mathematical framework for comparing networks. Instead of looking at the networks at a fixed threshold, persistent homology records the changes in topological network features over multiple resolutions and scales. By doing so, it reveals the features that are robust to noise, i.e., the most ‘persistent’ topological features.

6. Summary

In this paper, we have reviewed the most common methods to estimate functional connectivity in fMRI data. For single-subject data, estimation can be done by directly quantifying correlations across regions of interest and/or seed regions, or by finding a set of latent components that represent simultaneous activity, and while interpretation is straightforward for the former approach, it is not as clear for the later. In the example provided, the number of component maps needed to represent the data variability is very high and, therefore, the investigation of only a few components might not reflect the whole picture of the brain network.
The results obtained in Section 2 indicate that even if the regions are defined in an equivalent way, different estimation procedures of connectivity will lead to different interpretations of the networks. Therefore, it is of great importance to be aware of the limitations of each approach, especially when interpreting results from individual datum.
Despite the challenges with the single-subject analysis, a consistent procedure, applied to various subjects, might translate into a successful representation of multiple-subject networks. This is specially true if the method does not require a multi-stage approach and performs, instead, a joint estimation as in the tensorial ICA framework. Other emerging multi-subject network methods, such as persistent homology, are a promising way to estimate brain circuitry, especially if scalability can be achieved.

Funding

This research was funded by Natural Sciences and Engineering Research Council grant number RGPIN-2020-06941.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Methods to Quantify Correlation

Cross Correlation. Cross correlation measures the (lagged) temporal dependencies between two signals, and it was first proposed by reference [43] as an effective way to describe functional connectivity. Suppose we want to calculate the correlation between the BOLD time series for a given voxel v, i.e., y v ( t ) , t = 1 , , T and a reference time series r v ( t ) , t = 1 , , T for v v . Let μ y and μ r be the average value of the vectors y v and r v , respectively. Then, the cross correlation between the vectors y v and r v is defined as
c c y , r = t = 1 T ( y v ( t ) μ y ) ( r v ( t ) μ r ) t = 1 T ( y v ( t ) μ y ) 2 t = 1 T ( r v ( t ) μ r ) 2
The reference vector can be a pre-selected voxel, the seed, or it can be an average of time series in a certain region. For cross correlation between ROIs, both y ( t ) , t = 1 , , T and r ( t ) , t = 1 , , T can be the average time series in the pre-determined regions y and r, respectively.
It is common to transform the correlation coefficient obtained in (A1) using a Fisher’s Z-transformation for each correlation coefficient as follows
z score = l n ( 1 + c c y , r ) l n ( 1 c c y , r ) 2 .
These coefficients are approximately normally distributed, and cutoff values are obtained from the standard normal distribution.
Partial Cross Correlation. Cross correlation quantifies only the marginal linear dependence between two signals and does not consider the effect of a third signal [15,44]. To remove the linear influence of a third signal k ( t ) we define the partial correlation as follows.
P C C y , r | k = c c y , r c c y , k c c r , k 1 c c y , k 2 1 c c r , k 2 .
Partial cross correlation is a valuable metric for estimating brain networks because it can estimate the direct relationship between two signals [15].
The calculation of cross correlation and partial cross-correlation measures assumes the signals to be stationary. When this assumption is not satisfied, detrended cross correlation and detrended partial cross correlation should be used instead [45].
Time-varying connectivity. It is possible to obtain a dynamical functional connectivity to understand its pattern over time. Both static measures mentioned in this section have a natural time-varying analogue in conjunction with a sliding window [15].
The concept of the sliding window is simple. Starting from the first time point, a window (a fixed number of time points) is selected, and all data points within the window are used to estimate the FC. This window is then shifted a certain number of time points, and the FC is estimated on the new set of data points. The process is repeated until the end of the time course. The series of estimated FC over time is the time-varying FC.

Appendix B. Calculation of Network Measures

For completeness, we present the mathematical definitions of the network measures presented in Section 2.5. For a complete list of network measures, please refer to reference [29].
We use the graph notation as defined in Section 2.5. Let n be the number of nodes in the network and N be the set of all nodes. Let l be the number of links in the network and L be the set of all links. Then, ( i , j ) is a link between nodes i and j and a i j = 1 when there is a link ( i , j ) . We define l = i , j a i j (counting each indirect link twice).
  • Degree of a node. The degree of a node i is the sum of all the links connected to the node and is defined as
    k i = j N a i j .
  • Shortest path length. The shortest path length measures the shortest distance between nodes i and j and is defined as:
    d i j = a u v g i j a u v ,
    where g i j is the shortest distance between i and j. For all disconnected pairs ( i , j ) , d i j = .
  • Characteristic path length. Let L i be the average distance between node i and all other nodes. The characteristics path length is defined as
    L = 1 n i N L i = 1 n i N j N , j i d i j n 1 .
  • Number of triangles. The number of triangles of a node i is defined as
    t i = 1 2 j , h N a i j a i h a j h .
  • Clustering coefficient. The clustering coefficient of the network is defined as
    C = 1 n i N C i = 1 n i N 2 t i k i ( k i 1 ) .

References

  1. Elam, J.S.; Glasser, M.F.; Harms, M.P.; Sotiropoulos, S.N.; Andersson, J.L.; Burgess, G.C.; Curtiss, S.W.; Oostenveld, R.; Larson-Prior, L.J.; Schoffelen, J.M.; et al. The Human Connectome Project: A retrospective. NeuroImage 2021, 244, 118543. [Google Scholar] [CrossRef] [PubMed]
  2. Ogawa, S.; Lee, T.M.; Kay, A.R.; Tank, D.W. Brain magnetic resonance imaging with contrast dependent on blood oxygenation. Proc. Natl. Acad. Sci. USA 1990, 87, 9868–9872. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Barch, D.M.; Burgess, G.C.; Harms, M.P.; Petersen, S.E.; Schlaggar, B.L.; Corbetta, M.; Glasser, M.F.; Curtiss, S.; Dixit, S.; Feldt, C.; et al. Function in the human connectome: Task-fMRI and individual differences in behavior. NeuroImage 2013, 80, 169–189. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Van den Heuvel, M.P.; Hulshoff Pol, H.E. Exploring the brain network: A review on resting-state fMRI functional connectivity. Eur. Neuropsychopharmacol. 2010, 20, 519–534. [Google Scholar] [CrossRef] [PubMed]
  5. Biswal, B.B.; Kylen, J.V.; Hyde, J.S. Simultaneous assessment of flow and BOLD signals in resting-state functional connectivity maps. NMR Biomed. 1997, 10, 165–170. [Google Scholar] [CrossRef]
  6. Belliveau, J.W.; Cohen, M.S.; Weisskoff, R.M.; Buchbinder, B.R.; Rosen, B.R. Functional studies of the human brain using high-speed magnetic resonance imaging. J. Neuroimaging 1991, 1, 36–41. [Google Scholar] [CrossRef] [PubMed]
  7. Glover, G.H. Overview of functional magnetic resonance imaging. Neurosurg. Clin. 2011, 22, 133–139. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, L.; Zang, Y.; He, Y.; Liang, M.; Zhang, X.; Tian, L.; Wu, T.; Jiang, T.; Li, K. Changes in hippocampal connectivity in the early stages of Alzheimer’s disease: Evidence from resting state fMRI. NeuroImage 2006, 31, 496–504. [Google Scholar] [CrossRef]
  9. Rajamanickam, K. A Mini Review on Different Methods of Functional-MRI Data Analysis. Arch. Intern. Med. Res. 2020, 3, 44–60. [Google Scholar] [CrossRef]
  10. Ting, C.; Ombao, H.; Salleh, S.; Latif, A.Z.A. Multi-Scale Factor Analysis of High-Dimensional Functional Connectivity in Brain Networks. IEEE Trans. Netw. Sci. Eng. 2020, 7, 449–465. [Google Scholar] [CrossRef] [Green Version]
  11. Jolliffe, I.T.; Cadima, J. Principal component analysis: A review and recent developments. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2016, 374, 20150202. [Google Scholar] [CrossRef] [PubMed]
  12. Mckeown, M.J.; Makeig, S.; Brown, G.G.; Jung, T.P.; Kindermann, S.S.; Bell, A.J.; Sejnowski, T.J. Analysis of fMRI data by blind separation into independent spatial components. Hum. Brain Mapp. 1998, 6, 160–188. [Google Scholar] [CrossRef]
  13. Smith, S.M.; Fox, P.T.; Miller, K.L.; Glahn, D.C.; Fox, P.M.; Mackay, C.E.; Filippini, N.; Watkins, K.E.; Toro, R.; Laird, A.R.; et al. Correspondence of the brain’s functional architecture during activation and rest. Proc. Natl. Acad. Sci. USA 2009, 106, 13040–13045. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Van den Heuvel, M.; Stam, C.; Boersma, M.; Hulshoff Pol, H. Small-world and scale-free organization of voxel-based resting-state functional connectivity in the human brain. NeuroImage 2008, 43, 528–539. [Google Scholar] [CrossRef] [PubMed]
  15. Ombao, H.; Lindquist, M.; Thompson, W.; Aston, J. Handbook of Neuroimaging Data Analysis; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  16. O’Reilly, J.X.; Woolrich, M.W.; Behrens, T.E.; Smith, S.M.; Johansen-Berg, H. Tools of the trade: Psychophysiological interactions and functional connectivity. Soc. Cogn. Affect. Neurosci. 2012, 7, 604–609. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Tzourio-Mazoyer, N.; Landeau, B.; Papathanassiou, D.; Crivello, F.; Etard, O.; Delcroix, N.; Mazoyer, B.; Joliot, M. Automated Anatomical Labeling of Activations in SPM Using a Macroscopic Anatomical Parcellation of the MNI MRI Single-Subject Brain. NeuroImage 2002, 15, 273–289. [Google Scholar] [CrossRef] [PubMed]
  18. Wu, L.; Caprihan, A.; Bustillo, J.; Mayer, A.; Calhoun, V. An approach to directly link ICA and seed-based functional connectivity: Application to schizophrenia. NeuroImage 2018, 179, 448–470. [Google Scholar] [CrossRef] [PubMed]
  19. Andersen, A.H.; Gash, D.M.; Avison, M.J. Principal component analysis of the dynamic response measured by fMRI: A generalized linear systems framework. Magn. Reson. Imaging 1999, 17, 795–815. [Google Scholar] [CrossRef]
  20. Zhou, Z.; Ding, M.; Chen, Y.; Wright, P.; Lu, Z.; Liu, Y. Detecting directional influence in fMRI connectivity analysis using PCA based Granger causality. Brain Res. 2009, 1289, 22–29. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Zou, H.; Hastie, T.; Tibshirani, R. Sparse Principal Component Analysis. J. Comput. Graph. Stat. 2006, 15, 265–286. [Google Scholar] [CrossRef] [Green Version]
  22. Zipunnikov, V.; Caffo, B.; Yousem, D.M.; Davatzikos, C.; Schwartz, B.S.; Crainiceanu, C. Multilevel Functional Principal Component Analysis for High-Dimensional Data. J. Comput. Graph. Stat. 2011, 20, 852–873. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Ma, Z. Sparse principal component analysis and iterative thresholding. Ann. Stat. 2013, 41, 772–801. [Google Scholar] [CrossRef]
  24. Bell, A.J.; Sejnowski, T.J. An information-maximization approach to blind separation and blind deconvolution. Neural Comput. 1995, 7, 1129–1159. [Google Scholar] [CrossRef] [PubMed]
  25. Le, Q.; Karpenko, A.; Ngiam, J.; Ng, A. ICA with Reconstruction Cost for Efficient Overcomplete Feature Learning. In Advances in Neural Information Processing Systems; Shawe-Taylor, J., Zemel, R., Bartlett, P., Pereira, F., Weinberger, K.Q., Eds.; Curran Associates, Inc.: New York, NY, USA, 2011; Volume 24. [Google Scholar]
  26. Miranda, M.F.; Morris, J.S. Novel Bayesian method for simultaneous detection of activation signatures and background connectivity for task fMRI data. arXiv 2021, arXiv:2109.00160. [Google Scholar]
  27. He, Y.; Chen, Z.J.; Evans, A.C. Small-World Anatomical Networks in the Human Brain Revealed by Cortical Thickness from MRI. Cereb. Cortex 2007, 17, 2407–2419. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Achard, S.; Salvador, R.; Whitcher, B.; Suckling, J.; Bullmore, E. A Resilient, Low-Frequency, Small-World Human Brain Functional Network with Highly Connected Association Cortical Hubs. J. Neurosci. 2006, 26, 63–72. [Google Scholar] [CrossRef] [PubMed]
  29. Rubinov, M.; Sporns, O. Complex network measures of brain connectivity: Uses and interpretations. NeuroImage 2010, 52, 1059–1069. [Google Scholar] [CrossRef] [PubMed]
  30. Salimi-Khorshidi, G.; Douaud, G.; Beckmann, C.F.; Glasser, M.F.; Griffanti, L.; Smith, S.M. Automatic denoising of functional MRI data: Combining independent component analysis and hierarchical fusion of classifiers. NeuroImage 2014, 90, 449–468. [Google Scholar] [CrossRef] [Green Version]
  31. Burgess, G.C.; Kandala, S.; Nolan, D.; Laumann, T.O.; Power, J.D.; Adeyemo, B.; Harms, M.P.; Petersen, S.E.; Barch, D.M. Evaluation of Denoising Strategies to Address Motion-Correlated Artifacts in Resting-State Functional Magnetic Resonance Imaging Data from the Human Connectome Project. Brain Connect. 2016, 6, 669–680. [Google Scholar] [CrossRef]
  32. Smitha, K.; Raja, K.A.; Arun, K.; Rajesh, P.; Thomas, B.; Kapilamoorthy, T.; Kesavadas, C. Resting state fMRI: A review on methods in resting state connectivity analysis and resting state networks. Neuroradiol. J. 2017, 30, 305–317. [Google Scholar] [CrossRef]
  33. Beckmann, C.; Smith, S. Probabilistic independent component analysis for functional magnetic resonance imaging. IEEE Trans. Med. Imaging 2004, 23, 137–152. [Google Scholar] [CrossRef] [PubMed]
  34. Madsen, K.H.; Churchill, N.W.; Mørup, M. Quantifying functional connectivity in multi-subject fMRI data using component models. Hum. Brain Mapp. 2017, 38, 882–899. [Google Scholar] [CrossRef] [PubMed]
  35. Leonardi, N.; Richiardi, J.; Gschwind, M.; Simioni, S.; Annoni, J.M.; Schluep, M.; Vuilleumier, P.; Van De Ville, D. Principal components of functional connectivity: A new approach to study dynamic brain connectivity during rest. NeuroImage 2013, 83, 937–950. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Caffo, B.S.; Crainiceanu, C.M.; Verduzco, G.; Joel, S.; Mostofsky, S.H.; Bassett, S.S.; Pekar, J.J. Two-stage decompositions for the analysis of functional connectivity for fMRI with application to Alzheimer’s disease risk. NeuroImage 2010, 51, 1140–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Calhoun, V.; Adali, T.; Pearlson, G.; Pekar, J. A method for making group inferences from functional MRI data using independent component analysis. Hum. Brain Mapp. 2001, 14, 140–151. [Google Scholar] [CrossRef] [PubMed]
  38. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  39. Beckmann, C.; Smith, S. Tensorial extensions of independent component analysis for multisubject FMRI analysis. NeuroImage 2005, 25, 294–311. [Google Scholar] [CrossRef]
  40. Solo, V.; Poline, J.B.; Lindquist, M.A.; Simpson, S.L.; Bowman, F.D.; Chung, M.K.; Cassidy, B. Connectivity in fMRI: Blind Spots and Breakthroughs. IEEE Trans. Med. Imaging 2018, 37, 1537–1550. [Google Scholar] [CrossRef]
  41. Simpson, S.L.; Moussa, M.N.; Laurienti, P.J. An exponential random graph modeling approach to creating group-based representative whole-brain connectivity networks. NeuroImage 2012, 60, 1117–1126. [Google Scholar] [CrossRef] [Green Version]
  42. Simpson, S.L.; Laurienti, P.J. A two-part mixed-effects modeling framework for analyzing whole-brain network data. NeuroImage 2015, 113, 310–319. [Google Scholar] [CrossRef] [Green Version]
  43. Bandettini, P.A.; Jesmanowicz, A.; Wong, E.C.; Hyde, J.S. Processing strategies for time-course data sets in functional MRI of the human brain. Magn. Reson. Med. 1993, 30, 161–173. [Google Scholar] [CrossRef] [PubMed]
  44. Marrelec, G.; Krainik, A.; Duffau, H.; Pélégrini-Issac, M.; Lehéricy, S.; Doyon, J.; Benali, H. Partial correlation for functional brain interactivity investigation in functional MRI. NeuroImage 2006, 32, 228–237. [Google Scholar] [CrossRef] [PubMed]
  45. Podobnik, B.; Stanley, H.E. Detrended cross-correlation analysis: A new method for analyzing two nonstationary time series. Phys. Rev. Lett. 2008, 100, 084102. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Estimated connectivity for the ROIs based on the AAL parcellation. Panel (a) depicts the cross-correlation for the average time series of the ROIs, panel (b) depicts the partial cross correlation for the average time series of the ROIs, panel (c) depicts the cross correlation for the time series of the ROI data projected with the first PC, panel (d) depicts the partial cross correlation for the time series of the ROI data projected with the first PC, and panel (e) represents the RV coefficient with each ROI retaining the principal components that explain 20% of its variability.
Figure 1. Estimated connectivity for the ROIs based on the AAL parcellation. Panel (a) depicts the cross-correlation for the average time series of the ROIs, panel (b) depicts the partial cross correlation for the average time series of the ROIs, panel (c) depicts the cross correlation for the time series of the ROI data projected with the first PC, panel (d) depicts the partial cross correlation for the time series of the ROI data projected with the first PC, and panel (e) represents the RV coefficient with each ROI retaining the principal components that explain 20% of its variability.
Entropy 24 00390 g001
Figure 2. Binary Graphs obtained from the thresholded connectivities matrices of Figure 1. For all panels, the white color indicates an edge between the ROIs. Panel (a) is the graph obtained by thresholding the cross correlation of the average time series of the ROIs, panel (b) depicts the graph from the thresholded partial cross correlation for the average time series of the ROIs, panel (c) depicts the graph obtained by thresholding the cross correlation for the time series of the ROI data projected with the first PC, panel (d) depicts the graph obtained by thresholding the partial cross correlation for the time series of the ROI data projected with the first PC, and panel (e) represents the graph obtained by thresholding the RV coefficient.
Figure 2. Binary Graphs obtained from the thresholded connectivities matrices of Figure 1. For all panels, the white color indicates an edge between the ROIs. Panel (a) is the graph obtained by thresholding the cross correlation of the average time series of the ROIs, panel (b) depicts the graph from the thresholded partial cross correlation for the average time series of the ROIs, panel (c) depicts the graph obtained by thresholding the cross correlation for the time series of the ROI data projected with the first PC, panel (d) depicts the graph obtained by thresholding the partial cross correlation for the time series of the ROI data projected with the first PC, and panel (e) represents the graph obtained by thresholding the RV coefficient.
Entropy 24 00390 g002
Figure 3. Seed- based connectivity of the left pars opercularis. Figure shows sagittal slices with voxels that have a significant connection with the seed ROI depicted in red.
Figure 3. Seed- based connectivity of the left pars opercularis. Figure shows sagittal slices with voxels that have a significant connection with the seed ROI depicted in red.
Entropy 24 00390 g003
Figure 4. Sagittal view of the ordered principal components’ maps from first (top) to fifth (bottom).
Figure 4. Sagittal view of the ordered principal components’ maps from first (top) to fifth (bottom).
Entropy 24 00390 g004
Figure 5. Sagittal view of the independent components’ maps ordered based on increasing amounts of uniquely explained variance from first (top) to fifth (bottom).
Figure 5. Sagittal view of the independent components’ maps ordered based on increasing amounts of uniquely explained variance from first (top) to fifth (bottom).
Entropy 24 00390 g005
Table 1. Network summary measures.
Table 1. Network summary measures.
(a) Av. CCorr(b) Av. Pcorr(c) PC1 Ccorr(d) PC1 Pcorr(e) RV
CPL2.1375.0291.5083.96851.8789
DG34.1931.31369.5181.3864.217
CC0.6400.0720.8250.1790.820
Inf296311,239296411,78912,366
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shahhosseini, Y.; Miranda, M.F. Functional Connectivity Methods and Their Applications in fMRI Data. Entropy 2022, 24, 390. https://doi.org/10.3390/e24030390

AMA Style

Shahhosseini Y, Miranda MF. Functional Connectivity Methods and Their Applications in fMRI Data. Entropy. 2022; 24(3):390. https://doi.org/10.3390/e24030390

Chicago/Turabian Style

Shahhosseini, Yasaman, and Michelle F. Miranda. 2022. "Functional Connectivity Methods and Their Applications in fMRI Data" Entropy 24, no. 3: 390. https://doi.org/10.3390/e24030390

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop