1. Introduction
Knowledge of the brain structure, function, and mechanisms of underlying neural processes have advanced significantly in recent decades [
1,
2]. These breakthroughs have been driven by the rapid development of brain imaging techniques, including functional magnetic resonance imaging (fMRI), electroencephalography (EEG), electrocorticography (ECoG), local field potentials (LFPs), and calcium imaging, among others [
3]. Equally important are advances in statistical and computational methodologies, enabling the efficient estimation and robust analysis of the complex datasets generated by these imaging techniques [
4,
5].
Neuroscientists have dedicated immense efforts to understanding both localized brain region function and integration across brain regions during resting state and while responding to external stimuli. It is indeed paramount to estimate and analyze the connectivity patterns between signals, recorded at different brain regions, to uncover the neural mechanisms that underlie perception, action, and cognition [
6]. These connectivity patterns not only reveal the strength of connectivity between different brain regions but can also provide information on how disruptions can lead to neurological disorders [
7,
8,
9].
The hippocampus plays a pivotal role in memory formation, spatial navigation, and information processing [
10,
11,
12]. Its importance as a hub for neural connectivity makes it an essential target for studying brain networks. To investigate hippocampal function, researchers often rely on animal models (e.g., macaques and rats) whose brain structures bear a strong resemblance to those of humans [
13]. These models allow for precise and invasive experimental designs that are considered unethical in human studies.
This paper explores some advanced methods commonly used to analyze brain networks. We apply them to hippocampal LFP data from the CA1 region to investigate the encoding of nonspatial olfactory information in rats. Our aim is to introduce these methods to readers while highlighting the insights they can offer into neural connectivity. We will present and discuss the differences between classical methods and state-of-the-art approaches in modeling brain connectivity. Beyond showcasing the strengths, we also discuss their limitations and identify areas for potential future improvement. By presenting a wide set of techniques, we provide readers with tools to analyze different aspects of brain connectivity, offering diverse perspectives and insights into complex neural systems.
The structure of this paper is organized as follows:
Section 2 introduces the dataset and provides formal definitions of foundational concepts such as correlation, partial correlation, and coherence, which are essential for understanding basic connectivity patterns.
Section 3 details robust canonical coherence, a method for assessing more complex interdependencies.
Section 4 presents a hybrid approach combining Spectral Dynamic Principal Component Analysis (sDPCA) with Granger causality (GC) to analyze directional influences among specific channels and mitigate confounding effects from the broader network.
Section 5 explores spectral transfer entropy (STE), an information-theoretic method that examines frequency-specific influence and information flow in the brain.
Section 6 discusses wavelet coherence, which captures dynamic and nonlinear interactions between brain regions.
Section 7 introduces Persistence Homology (PH), a topological method that avoids the need for thresholding in weighted networks and extracts multi-scale connectivity patterns, thus detecting higher-order interactions.
Section 8 summarizes and discusses the strengths and limitations of the methods presented, emphasizing their potential for enhancing brain connectivity analysis and identifying promising directions for future research. Finally,
Section 9 offers concluding remarks.
2. Exploratory Data Analysis
In this section, we turn our attention to the experimental data that underpin our analyses. The dataset consists of local field potential (LFP) recordings from the CA1 region of the hippocampus in rats performing a nonspatial sequence memory task, a paradigm chosen for its strong behavioral parallels between rats and humans. As detailed in Allen et al. [
14], LFP signals were recorded from five male Long–Evans rats. The animals were individually housed, with water access controlled during weekdays (serving as a reward in odor memory tasks).
The experimental protocol involved a nonspatial sequence memory task, in which rats were required to memorize and recognize a fixed sequence of five odors: lemon (A), anise (B), rum (C), vanilla (D), and banana (E). Rats underwent an incremental training protocol over 6–8 weeks. Initially, naive rats were trained to nosepoke and maintain their nose in the odor port for a water reward. The required nosepoke duration was gradually increased from 50 ms in 15 ms steps until reaching 1.2 s, with a criterion of 80% correct responses over three consecutive sessions (100–200 nosepokes per session). Subsequently, the animals were habituated to odor presentations, first with a single odor (Odor A) and then with a two-odor sequence (Odors A and B), both requiring a 1.2 s nosepoke for a reward. Once performance was stabilized, the rats were trained to discriminate between in-sequence and out-of-sequence presentations, starting with a two-item sequence (e.g., “AB” for in-sequence versus “AA” for out-of-sequence) and progressing to sequences of three, four, and finally five odors. After achieving criterion performance on the five-item sequence, the rats underwent microdrive implantation surgery for subsequent electrophysiological recordings. The odors were delivered through a single odor port as described in
Figure 1, with each session featuring odors presented either in the correct sequential order or with at least one item out-of-sequence.
In this study, the dataset is organized into four-second trials, with odor presentation occurring at the midpoint (two seconds) and initiated by a nosepoke. LFP signals were recorded at a sampling rate of 1000 Hz from five rats using a microdrive equipped with approximately 20–22 tetrodes per rat, each positioned in either the proximal or distal region of the CA1 layer. On average, each rat completed between 170 and 300 trials (approximately 170–260 in-sequence and 20–45 out-of-sequence trials). Notably, the ‘Barat’ rat demonstrated the highest accuracy in recognizing in-sequence odors, while ‘Superchris’ excelled in identifying out-of-sequence presentations. Detailed variations in the number of tetrodes and trials across subjects are provided in
Table 1.
2.1. Classical Dependence Measures
For the remainder of this manuscript, we adopt the following unified notation to facilitate the presentation of analytical methods. Consider LFP signals measured over time from P tetrodes. Let represent the LFP signal recorded from the p-th tetrode at time t, where . For multiple trials of the same experiment, we use the superscript notation to denote the r-th trial. For clarity, we refer to tetrodes as , respectively.
Functional connectivity (FC) refers to the statistical association between neurophysiological events measured across various scales, microscale (individual neurons), mesoscale (neuronal populations), and macroscale (brain regions) [
15]. In the context of brain connectivity analysis, FC is almost always measured using Pearson correlations [
16]. In this study, we focus on LFP signals, which serve as mesoscale measurements that capture the collective activity of multiple neurons. Specifically, we investigate FC between tetrodes using correlation and partial correlation as defined below.
Consider the LFP signals recorded during the
r-th trial from two tetrodes
p and
q, denoted by
and
. For this section, we assume the LFP during trial
r and
to be the zero-mean second-order stationary time series (see [
17] for definition). Then, the
cross-covariance between
and
at time delay
k is written as
Given
P tetrodes in the system, all pairwise covariances can be compactly written as a
cross-covariance matrix at lag
k, denoted as
, i.e.,
Although cross-covariance quantifies the dependence between two time series, it is often difficult to interpret because its magnitude depends on the scale (level of variability) of the data. Thus, it is more common to use its scaled version, called
cross-correlation or simply correlation, which takes values in the interval
. This is more useful especially when comparing strengths of connectivity across different tetrode pairs. More precisely, the correlation between
and
is defined as
The correlation index measures the linear association between the signals and . Moreover, when the LFPs have a normal distribution, implies unconditional independence between them. However, one limitation is that it may include confounding variables that influence the interaction between and , e.g., another signal from the same system, say . Thus, an alternative approach is to quantify the direct dependence between a pair of signals after taking into account the contributions of other components in the brain network. This is offered by the partial correlation measure, which we define below.
Define
to be the set of
tetrodes excluding the
p-th and
q-th tetrodes, and
to be the multivariate time series recorded during the
r-th trial from all tetrodes in
. Note that
and
are excluded in
. Consider the variance–covariance matrix at lag 0, which can be derived from Equation (
2), and denote it by
. The precision matrix, denoted by
, is the inverse of
, i.e.,
Then, the partial correlation between two tetrodes
and
, after removing the linear contributions of the remaining tetrodes in the system
, is defined to be
The quantity represents the element in the p-th row, q-th column of the precision matrix, which in practice, may be obtained as the inverse of an estimated variance–covariance matrix. A caveat, however, is that the covariance matrix should be positive definite (and hence non-singular) for the precision matrix to exist. In cases of perfect collinearity between at least one pair of signals, the covariance matrix is singular, preventing the signals from being de-confounded.
2.1.1. Permutation Test
For a given odor, consider the correlation between the LFP signals, recorded during the
r-th trial, from tetrodes
p and
q and denote it by
. Here, we compare the two groups of trials (in-sequence vs. out-of-sequence). Denote by
and
the respective true means of the correlations and variances across the entire distribution all possible realizations of in-sequence and out-of-sequence trials. Our goal is to determine whether there are differences in the mean correlations between in vs. out-of sequence states, e.g.,
in-sequence correct trials and
out-of-sequence correct trials with zero time delay (
). Hence, for a given
-tetrode pair, we wish to test the following hypothesis:
We consider the test statistic
where
and
are the group
sample averages and group
sample variances, respectively, of the correlations. Given the observed data, let
denote the calculated value of the test statistics
. As a decision rule, we reject the null hypothesis
if the
p-value, i.e.,
, is less than the significance level
.
One approach is to empirically derive the unknown distribution of under the null and thus estimate the p-value through a permutation testing scheme. Under the null hypothesis, the correlations from all correct trials, whether in-sequence or out-of-sequence, come from the same distribution. Such an assumption allows the reassignment or relabeling of the correlations as in-sequence or out-of-sequence, which corresponds to one permutation. In an iterative manner, several permutations of labels for the observed correlations are obtained, and for each permutation, a is calculated. The collection of values comprises the empirical null distribution of , from which we obtain the p-value for the two-sample t-test.
2.1.2. Correlation Analysis of the LFP Dataset
We now implement the permutation test (discussed above) on the correlations of LFP signals from the rat named “Superchris” for in-sequence and out-of-sequence trials of all combinations of odors and pairs of tetrodes (
Figure 2). In particular, we examine trials where the odor presented is
vanilla. For recordings from T11 and T21, there is a significant difference between the mean correlations of the in-sequence trials and the out-of-sequence trials at
, with
and
.
The same analysis can be performed for partial correlation (
Figure 3), accounting for the confounding variable results in sparse correlation matrices. For rum and tetrode-pair T5–T20, mean partial correlations between in-sequence and out-of-sequence trials have a significant difference (
), revealing a change in interaction depending on the accuracy of the odor sequence.
2.2. Spectral Dependence Measures
Correlation and partial correlation are simple yet effective measures for capturing the linear dependence between signals in the time domain. In contrast, assessing synchronization in the frequency domain provides a more detailed understanding of the oscillatory dynamics that drive neural interactions. Coherence analysis, the frequency-domain counterpart of correlation, has proven highly effective in evaluating brain connectivity by yielding results that are directly interpretable in terms of frequency components [
18].
Let
be a
P-dimensional second-order stationary time series, meaning that its mean vector,
remains constant over time, and its covariance matrix,
depends only on the lag
k rather than the specific time index
t. In addition, we assume that the elements of
are absolutely summable:
These conditions ensure the existence of a well-defined spectral matrix for .
The essence of the spectral analysis is to represent brain signals as a superposition of oscillatory components across various frequency bands. This is achieved by decomposing the signal into complex exponentials, where
serves as the fundamental building block. This idea is formalized in the
Cramér representation:
with
denoting a zero-mean, orthogonal increment process. This representation holds under the aforementioned stationarity conditions.
The spectrum of an individual component
is defined as the Fourier transform of its autocovariance function:
and the cross-spectrum between components
and
is similarly given by the Fourier transform of their cross-covariance function:
Collecting all auto- and cross-spectral quantities from the spectral matrix
The spectral matrix can provide insight into the spectral power distribution of the signals. However, by examining dominant specific frequency bands that are present in the signals, one can develop a better understanding of the brain connectivity and mental state of a subject. Studies have demonstrated that the five traditional frequency bands are associated with cognitive states [
19] and can also be used as potential biomarkers for neurological diseases (e.g., autism, and attention deficit-hyperactivity disorder (ADHD)) [
20]. These frequency bands, which are defined below, can be adapted in the analysis of LFP signals.
The most common frequency bands of interest in EEG and LFP analysis are
Hertz,
Hertz,
Hertz,
Hertz, and
Hertz. There is a 1-1 mapping between the frequency band in the generic interval
and the bands in practical EEG/LFP, which is defined as follows. Let
s be the sampling rate, and let
be the generic band of interest. This corresponds to
. Using these bands of interest, we can decompose the observed LFP to be
where
are weights associated to the contribution of each frequency band in the signal. Additionally,
can be derived from the observed signal via linear filtering (e.g., Butterworth filter)
where the filter
is selected so that the power of
is concentrated at the frequency band
. The filtered decomposition for a sample LFP signal from a trial is shown in
Figure 4, at which the left-hand side showcases the spectral power and the right is the decomposed signals for each of the signals.
Dependence between tetrodes can be characterized via
coherence which is a frequency domain measure of linear correlation between two signals of the same frequency band
. For tetrodes
p and
q, the coherence at frequency
is defined as
where
and
are the auto-spectrum and cross-spectrum at band
. The values of coherence lie between 0 and 1, with 0 indicating that there is no linear correlation at that frequency and 1 indicating a perfect linear relationship at that frequency.
It is of interest to see that the resulting coherence is also clustered within the same tetrodes as in
Figure 2. Nonetheless,
Figure 5 shows that there is a difference in the intensity of synchronization between tetrode signals at the alpha and gamma frequency bands. It appears that even though the alpha band has lower within-cluster coherence, it shows overall higher coherence in many tetrodes further from the diagonal. On the other hand, the gamma frequency band shows clearer coherence intensities; the clusters near the diagonal have high coherence, and clusters far from the diagonal have low coherence.
Coherence analysis can be a useful tool for looking at brain connectivity. However, it comes with the assumption of the stationarity of the time series signals, which can be the case for a short time frame but cannot hold for a longer time, as many real-world signals are non-stationary. Additionally, the coherence of the frequency domain does not indicate the temporal information when the signal pairs are coherent, and coherence analysis can require averaging over time windows (trials) in order to estimate the spectral and cross-spectral power. Consequently, coherence gives a global frequency relationship and is sensitive to noise, as it can inflate or deflate the coherence between signals.
Nonetheless, in
Section 6, wavelet analysis, which does not require stationarity for the time series, is discussed as another spectral domain analysis method. The wavelets are less sensitive to noise and can capture multi-scale relationships. Additionally, looking at pairwise analysis can result in redundancies in real-life applications. Therefore, in
Section 3, coherence is used to obtain connectivity measures between a cluster of tetrodes from major brain regions.
4. Granger Causality Across Node/Region Subsets
Recent brain connectivity analyses frequently involve high-dimensional signals, such as large sets of LFPs or EEG recordings. Dissecting the directional influence between specific nodes (i.e., channels, electrodes, tetrodes) or sub-regions (e.g., pre-defined cluster of nodes) within such high-dimensional and possibly complex networks can serve as a methodological basis for explaining the neural mechanisms at a detailed level. However, analyzing a pair of nodes or sub-regions under the possible confounding effects of other channels in the network leads to inferential complications. Various methods have addressed similar problems in high-dimensional networks [
33,
34,
35,
36,
37,
38], yet the difficulty of isolating a subset of nodes from the rest of the network remains a critical problem [
39,
40].
This section introduces an approach designed to overcome these challenges using spectral domain dynamic principal component analysis (sDPCA) [
21,
41,
42]. The methodology aims to isolate two nodes (or sub-regions, depending on the context) of interest within a high-dimensional network by removing the aggregate influence of all other nodes, facilitating the subsequent application of conventional examinations for Granger causality (GC). By collapsing the complexity of the entire network into a low-dimensional representation and partialling out the effects of other nodes/subregions, the resulting node-specific signals are isolated from the confounding effects. This provides a practical medium for inferring directional interactions.
4.1. Inference in High-Dimensional Setting
Consider a high-dimensional network, , of P nodes (i.e., channels, electrodes, and tetrodes), each corresponding to a signal. Let the network, , be , where and are two nodes of interest (NOIs) and represent a large set of other nodes whose influence we wish to control.
We propose combining GC with sDPCA specifically to focus on pairwise interactions among NOI in LFP data. In general, LFP or brain imaging data often involve many channels/nodes measuring neural activity across multiple frequency bands, making direct pairwise GC analysis difficult because of the curse of dimensionality and the risk of overfitting or spurious connections. Applying sDPCA provides an advertent dimensionality reduction step with respect to the spectral structure of the data: sDPCA operates in the frequency domain and accounts for the dominant oscillatory and frequency-specific patterns in the signals. This means that important neural dynamics (such as rhythmic oscillations or cross-frequency interactions) are preserved in a few components rather than averaged out. Following the implementation of sDPCA to extract and regress the network’s interfering background from each node of interest, the exploration of GC between the residual signals becomes feasible. While alternative approaches exist (for example, one could apply GC with sparse regularization to the complete set of channels), the sDPCA+GC methodology is proposed for its ability to maintain frequency-specific information and improve reliability in detecting neural interactions. This balanced and easy-to-implement strategy addresses the complexity of LFP data by focusing on physiologically meaningful components, thus providing a clearer and more interpretable pairwise causal connectivity analysis in a high-dimensional neural recording context.
To isolate the NOI and uncover their causal relationships, one can follow a strategy that involves transforming the signals as follows:
where
is a function summarizing the collective influence of the remaining nodes excluded. After this step,
and
become isolated versions of the original signals, where the network’s background variability has been partially removed. The key question is how to construct this function
to capture the large network’s dynamics without overfitting or losing crucial frequency-dependent structures.
4.2. Using Spectral Dynamic PCA to Represent the Background Network
Conventional principal component analysis (PCA) focuses on reducing dimensionality by finding linear combinations of variables that explain the most significant variance. However, PCA operates on covariance structures that do not directly incorporate temporal dependency or frequency-specific patterns. Neural signals often have rich spectral content—specific frequencies may carry more meaningful interactions than others. Frequency-domain dynamic PCA (sDPCA) [
21,
42,
43] addresses this need by operating in the frequency domain and extracting components that are informative about temporal associations.
To apply sDPCA, we first estimate the cross-spectral density matrix of the background signals
. Let
be the cross-spectral density matrix at frequency
. This matrix encodes frequency-specific variances and covariances. It can be estimated by
where
is a window function,
M is the window size, and
is the empirical lag-
h covariance matrix given by
for
, and
for
.
At each frequency
, we solve the eigenvalue problem:
The eigenvectors represent frequency-domain principal directions, and are the corresponding eigenvalues. These frequency-specific eigenvectors reflect how variability is arranged throughout the spectral domain.
To return to the time domain, we compute filters from the eigenvectors via inverse Fourier transform:
for integer shifts
m. Each set of filters,
defines a dynamic principal component in the time domain.
Applying these filters to
yields a reduced set of dynamic principal component scores:
Only a few dynamic components are usually necessary to capture a significant fraction of the total variance. These scores represent a low-dimensional snapshot of the entire background network’s activity, integrated over time and frequency.
4.3. Partialling Out Background Influence and Applying Granger Causality
Once we have obtained principal scores
, where
, treated as covariates summarizing the background nodes, we model
Subtracting these fitted values from and yields and , which are now approximately isolated from the rest of the network’s influence. In practice, these conditional expectations are approximated by regressing each node of interest (e.g., ) on the set of dynamic principal component scores and their interactions. Using a linear model (or a nonlinear model by preference), the fitted values and are used to approximate and , respectively.
With
and
in hand, we return to a more conventional Granger causality framework [
44]. Testing for GC typically consists of comparing a restricted model that predicts
using only its own past against an unrestricted model that also includes
’s past values:
If adding past values of significantly reduces variability in the predictive errors, we conclude that Granger causes . Similarly, we can test whether Granger causes .
Working with the isolated channels,
and
, makes the GC results less likely to be distorted by unmodeled interactions from other network nodes. Thus, this approach transforms a high-dimensional problem into a more tractable one, employing the sDPCA spectrum-aware dimension reduction.
Figure 9 exemplifies the proposed approach for GC in high-dimensional networks.
4.4. Practical Advantages and Limitations
The proposed methodology avoids the complexity of simultaneously fitting a high-dimensional vector autoregressive (VAR) model to the entire network. Instead, it establishes a controlled setting where the causal interactions between selected nodes can be tested more directly rather than being potentially affected by the influence of the entire network. It does so via bridging the original high-dimensional network and a low-dimensional summary captured by dynamic principal components. By partialling out these low-dimensional summaries, the nodes of interest recover with reduced confounding.
Thus, the sDPCA-based procedure is well suited for scenarios where interest focuses on the directional interaction between a small subset of channels embedded in an extensive network. It can also be adapted when the nodes of interest represent not just single channels but clusters of channels (i.e., regions), each represented by their sDPCA-derived summary scores.
However, several limitations could be noted. First, selecting an appropriate number of principal components (here denoted ) is critical; poor choices can lead to the omission of essential dynamics or the exclusion/inclusion of informative/uninformative components. Second, spectral estimation relies on smoothing parameters and windowing techniques, which may discard sharp spectral features or hidden localized frequency-specific phenomena. Third, conventional tests for GC assume linearity and stationarity, yet neural signals often violate these assumptions, thereby potentially invalidating the inferred causal connections. Finally, while sDPCA reduces the dimensionality in a frequency-aware mode, interpreting the resulting dynamic principal components and linking them to specific bio-physiological processes can remain a significant challenge.
4.5. Granular Level GC in Olfactory LFP Network
We considered trials where the rats correctly identified a given odor’s sequence status (In-sequence—Correct,
Table 1). For each trial of interest, segments of LFP signals are extracted starting at the onset of the odor. These segments are then differenced to ensure stationarity and combined across odor types and trials, ultimately yielding a separate collection of LFP samples for each odor. Because each subject’s recording involved multiple tetrodes placed along the proximal to distal axis of CA1, the goal is set to examine how specific tetrode pairs might exhibit directional influences/connectivities (i.e., Granger causality) unique to a particular odor or common across all odors. In practice, we conduct GC analysis for every distal–proximal (i.e., proximal–distal) pair and count how frequently one tetrode “drove” another across the trials for that odor. If a directed association is significant on at least 99% of trials, we label that connection as consistently present. This procedure “votes” on each possible directed relation in the network and flags only those consistently emerging in every trial. Finally, we compile these odor-specific GC connections into plots. The results are visualized as in
Figure 10: each row corresponds to one subject, each column to odor, plus the first column shows a connectivity pattern recurring under all odor conditions.
Figure 10 reveals that each subject has certain distal (i.e., blue-colored region) to proximal (i.e., orange-colored region) channels linking up under every odor. Mitt’s distal tetrodes
consistently influence proximal channels
. These connections emerge throughout odors A–E, whereas
→
or
→
occur exclusively in certain odors (e.g., odor A, C, E), implying that these unique routes are distinct circuit elements engaged by particular olfaction. Stella features
→
,
→
, and
→
across all odors, with distinct connectivities such as
→
or
→
emerging in certain odor, suggesting that while Stella’s distal to proximal drive persists, some odor-specific flows may support specific olfaction demands. Buchanan displays a smaller cluster of common edges than Mitt or Stella, but we still see consistent patterns like
→
or
→
or
→
. Odor-specific differences appear in connections such as
for odor A/C/D,
for odor B/E, and
for odor D/E. While these odor-level connections vary, the overall structure still centers on distal nodes (
,
,
), influencing its several proximal tetrodes
,
,
, and
. Barat has relatively a large cluster of distal-to-proximal associations, with
consistently targeting
, and
targeting
. These connections remain steady across odors. A few bidirectional connections (e.g.,
↔
and
↔
) also appear. Some extra pathways show up in certain odors (e.g.,
→
in odor C,
→
in odor C, and
→
in odor A/E), but the broader pattern remains
to
. The subject Superchris stands apart by exhibiting a large cluster of bidirectional connectivity between its distal tetrodes (i.e.,
) and proximal tetrodes (i.e.,
,
, …,
). Unlike Mitt or Stella, who rely more on unidirectional distal → proximal pathways (for example,
→
in Mitt or
→
in Stella), Superchris shows multiple two-way interactions, such as
↔
and
↔
, that persist under all odors. Superchris also exhibits several odor-specific connections involving
,
,
, or
, which emerge only in certain odors, whereas other subjects generally employ fewer links when shifting from one odor to another. Therefore, while the common pattern of distal → proximal influences remains visible, Superchris integrates richer reciprocal activity, suggesting a denser or interconnected CA1 network compared to the predominantly one-directional flows seen in Mitt, Stella, Buchanan, or Barat.
A crucial question here is whether odor-specific connections show a clear shift in CA1 circuitry. Considering
Figure 10, most subjects display considerable similarities across “common to all odors” patterns (i.e., distal to proximal). This overlap suggests that hippocampal olfaction processing relies on shared associations of distal to proximal transmissions, with modest variations in direction or presence depending on odor identity. The new connections that do appear exclusively in a single odor could serve a more specialized role.
Consequently, a basic understanding is that each subject’s dorsal CA1 circuit favors a core route of information flow, with distinctive odor-related changes superimposed. In all five subjects, sub-regions along distal CA1 often serve as a “source” area projecting into multiple proximal tetrodes, while back-influences from proximal to distal also emerge in the data. This observation aligns with existing knowledge that CA1 exhibits prominent directionality along its anatomical axis [
45,
46].
5. Spectral Transfer Entropy
Unlike GC, which relies on model-based assumptions, transfer entropy (TE) is an information-theoretic measure that captures directional and potentially nonlinear dependencies between signals. This makes it particularly valuable for analyzing complex interactions in neural systems.
Consider two signals, denoted by
and
, observed from distinct nodes, voxels, channels, or tetrodes in a brain network. Let
for some time lags
k and
ℓ. Developed by Schreiber [
47], TE quantifies the information transfer from
to
by measuring the conditional mutual information (CMI) between
and
given
. This is expressed as
where
represents the conditional mutual information. This metric reflects the directed influence of
on
while accounting for the past states of
. More precisely,
with
,
and
. For a comprehensive discussion on other information-theoretic measures, including entropy and mutual information, refer to Cover and Thomas [
48]. In contrast with GC, which looks at the improvement in the prediction variance due to the additional information provided by another series’ history, TE measures the causal impact of a series to another series directly from their joint and conditional distributions. Explicitly, TE quantifies the statistical conditional dependence of
on the past
given its own history
. This formulation does not require any assumption on the distribution (e.g., Gaussianity) or the type of relationship (e.g., linear) between the two series, hence making the TE framework more general and applicable to analyzing complex data like LFP signals.
When the interest lies in relating effective connectivity to various frequency bands with well-explored cognitive interpretations, one strategy is to apply a bandpass filter on the observed signals to extract the band-specific oscillations of interest and conduct investigations via the GC or TE framework. For example, one may consider the filtered series
and
, where their respective spectral densities concentrate only on frequency band
, and calculate TE from
to
. However, the smooth oscillatory behavior of these band-specific oscillations often leads to erroneous results for approaches like GC and TE as linear filtering induces potential temporal dependence distortion and the false extraction of spectral influence [
49,
50,
51]. The problem does not stem from these causal frameworks (i.e., GC and TE) but rather from the direct use of smoothly oscillating filtered signals.
To address this issue, Redondo et al. [
52] formulated the spectral transfer entropy (STE) measure. Instead of capturing the direction and magnitude of information flow directly between two band-specific series, STE defines the information transfer between two nodes of a brain network based on a series of maximum amplitudes over non-overlapping time blocks. Let
and
where
is the time point preceding the
b-th time block of length
m. Concisely,
and
represent the block maxima series of amplitudes of the oscillations
and
, respectively. Specifically, STE from
to
, denoted by
, is defined as
and is shown to be robust (empirically) to the inherent issues associated with linear filtering, i.e., it adequately captures spectral causal influences with controlled false positive rates, which provides evidence of the practical advantages of such a formulation.
Aggregating band-specific signals into a series of maximum amplitudes over time blocks takes inspiration from communication theory, where the information transfer between two devices occurs through signal modulation. That is, an
-band oscillation
can be expressed as a product of a carrier signal
, whose spectral density is concentrated in the frequency band
and serves as the information pathway for the flow of information, and a modulating signal
that carry the information being transferred from one node to another (see
Figure 11 for illustration). However, there is a shift in the temporal resolution of causality defined by the STE measure. For instance, if the signals are observed at a sampling rate of 1000 Hz (i.e., at 1000 time points per second) and the specified block size is
, STE quantifies the causal interactions that occur in about every one-tenth of a second (which is slower than the original temporal scale of the observed data). For more details on the interpretation, choice of tuning parameters, and vine copula-based inference for STE, which we employ in the subsequent analysis, refer to [
52].
Let
and
be the LFP signals from tetrode
q in the distal region and tetrode
p in the proximal region, respectively. In this section, our primary objective is to identify differences in effective connectivity between tetrodes placed in the distal and proximal regions during correct and incorrect trials of mice given the olfactory-based task. Correct trials refer to trials where a mouse received a reward for recognizing an in-sequence odor, while incorrect trials result in having no reward. Here, we consider in-sequence trials from two subjects (namely
Superchris and
Mitt) given the odors
rum and
lemon, respectively. Since the number of incorrect trials is much smaller than that of the correct trials, we randomly sample an equal number of correct trials to make balanced cases. To be exact, we include 9 trials each from Superchris and 13 trials each from Mitt in the analysis. However, each trial contains roughly 1.2 s of LFP recordings after the odor presentation. Since the STE framework requires aggregating data over non-overlapping time blocks of length
(which we specify to define a practical temporal resolution of causality), we use all block maxima series obtained from all correct trials and from all incorrect trials to calculate the STE measure. Finally, we focus our attention on two frequency bands, namely the alpha and beta bands, as several spectral analyses on individual LFP signals reveal changes in the latter related to olfactory functions [
53,
54,
55], while differences in the former are yet to be discovered. Our goal is to provide insights on the causal influence of these band-specific oscillations on one another via the distributions of
and
across relevant
-tetrode pairs, complimenting the existing results based on univariate spectral methods.
In
Figure 12, we observe a high magnitude of information transfer, as quantified by the STE measure, between the distal and proximal regions during correct in-sequence trials while lower STE values during incorrect in-sequence trials. In addition, the differences in the magnitude of captured causal influence between correct and incorrect trials are highly prominent for subject Superchris while being less yet arguably still prominent for subject Mitt. This suggests that the STE approach is able to reveal prominent differences in connectivity patterns among the multivariate LFP signals in the alpha band, even though there is limited work on univariate spectral density methods that detect any differences in the same frequency band. By contrast, the flow of information in the beta band from the distal to the proximal region of Superchris has higher magnitudes during incorrect trials than during correct trials, while the STE values from proximal to distal region have larger variability among correct trials than among incorrect trials (see
Figure 13). Further, there are very minimal differences in the beta band for subject Mitt. Such inconsistency in differences may be related to the odors presented to the respective subjects, as one odor may have a stronger or weaker impact on the subjects than the other. Nonetheless, the STE framework is a promising new tool for investigating effective brain connectivity, which we illustrate as successful in providing insights into how node interactions in the frequency domain may vary among different brain networks.
6. Wavelet Coherence Analysis
A key challenge in analyzing brain signals, such as LFPs, is their inherent non-stationarity; that is, statistical properties like the spectrum (or covariance) evolve over time. Wavelet analysis has proven exceptionally useful in capturing the transient features of non-stationary signals due to the compact support and flexibility of wavelet functions [
56,
57]. The compact support of wavelets allows for dynamic scaling—through compression or stretching as illustrated in
Figure 14—which enables them to adapt to changing signal characteristics. In contrast, traditional Fourier methods, which lack time localization and the ability to adapt to a signal’s dynamic behavior, often struggle to capture these transient properties.
To address these limitations, Nason et al. [
58] introduced a scale-specific stochastic representation of time series that leverages the multi-resolution property of wavelets to estimate evolving wavelet coherence. Building on this foundation, Park et al. [
59] extended the framework to multivariate locally stationary wavelet (LSW) processes, enabling precise characterization of single-scale coherence among different channels. More recently, Wu et al. [
60] proposed an innovative modeling framework that effectively captures the cross-scale dependence structure between channels in multivariate non-stationary time series. This advancement further enhances the capability of wavelet-based methods to uncover complex dependencies and evolving connectivity patterns in neural signals.
In this section, we introduce the framework of LSW and demonstrate its application to analyzing LFP data from different brain regions. We implement both single- and cross-scale coherence to capture the time-varying dependence structure across regions. This approach also allows us to examine how fluctuations in longer-term dynamics influence the amplitude of shorter-term dynamics, providing deeper insights into the multi-scale interactions within the brain.
6.1. LSW Model and Wavelet Coherence
Wavelets are powerful mathematical tools that enable the decomposition of signals into components containing both time and frequency (or scale) information. Unlike the Fourier transform, which represents signals as combinations of infinite sinusoids and provides only global frequency information, wavelets are uniquely suited for analyzing localized variations in signals. This makes wavelets particularly valuable for studying non-stationary data where signal characteristics may vary over time.
Wavelet analysis is built on two foundational functions: the
father wavelet and the
mother wavelet . The father wavelet
is designed to capture smooth, low-frequency components of a signal and integrate them into one, ensuring a focus on the overall trend. In contrast, the mother wavelet
integrates to zero and is responsible for extracting detailed, high-frequency components, thereby highlighting localized variations in the signal. To analyze signals at different resolutions, the mother wavelet is compressed and shifted to generate a family of
child wavelets. These wavelets, indexed by a scale parameter
j and a shift parameter
k, are defined as
where
J denotes the maximum number of scales, and the scale
j determines the resolution or level of detail captured by the wavelet, with smaller
j values corresponding to coarser scales and larger
j values corresponding to finer scales. The parameter
j determines the translation of the wavelet in time, allowing for localized analysis across the signal.
Nason et al. [
58] introduced the LSW framework, a novel representation for stochastic processes exhibiting complex, time-varying dynamics. Unlike traditional wavelet decomposition, which typically relies on decimated and orthogonal wavelet bases, the LSW framework employs non-decimated wavelet bases. This means that the wavelet bases in LSW are non-orthogonal across different scales and shifts, allowing for a more flexible representation of signals with intricate temporal structures. This unique feature makes LSW particularly well suited for analyzing non-stationary processes where traditional methods may fall short. Park et al. [
59] extended the LSW to a multivariate setting to capture the time-varying scale-specific cross-dependence between the components from the signals among different channels. Here, we directly start by introducing this multivariate LSW (MvLSW).
The P-variate locally stationary wavelet process
, defined by [
58], where
can be represented by,
where
is a set of discrete non-decimated wavelets;
is the time-dependent transfer function matrix.
are uncorrelated random vectors with mean vector
and variance–covariance matrix equal to the
identity matrix. Furthermore, the scale-
j subprocess of
is defined as
The evolutionary spectrum matrix is defined based on the transfer function, which is used for quantifying the time-scale power of
. This local wavelet spectral (LWS) matrix is given by
where
denotes the transpose of
. The random innovation term
is assumed to be uncorrelated across different scales
j and shifts
k:
where
denotes the Kronecker delta function; however, one of the main limitations of this framework is its inability to capture cross-dependence between subprocesses at different scales, which can be an important measure of dependence.
To address this limitation, ref. [
60] relaxed this assumption by defining the covariance matrix of
using a general matrix
. The dual-scale LWS matrix is then formulated as
Based on the single- and cross-scale LWS matrix, the time-varying wavelet coherence is defined as
where the matrices
and
are diagonal with elements
and
, respectively (see details in [
60]). The wavelet coherence values range from
to 1, measuring the local single- or cross-scale dependence structure between channel
p and channel
q in multivariate time series. This framework offers a powerful way to study the time-varying connectivity between different brain regions, enabling insights into dynamic neural interactions. An empirical way to calculate localized measures of cross-scale dependence in the time domain is
It is easy to see that if , then , i.e., the cross-scale coherence becomes equal to the single-scale coherence.
6.2. Wavelet Coherence Analysis with LFP Data
In this part, we implement both single-scale and cross-scale wavelet coherence across different channels of the LFP data recorded from Superchris. The primary objective is to determine whether wavelet coherence can effectively reveal alterations in brain connectivity when the rat makes mistakes in odor sequence discrimination.
The initial step involves decomposing the LFP time series into subprocesses at each scale. Here, we set the decomposition level to
J, which can be adjusted based on the desired resolution of the analysis.
Figure 15 presents the wavelet coherence between subprocesses at the same scale across eight channels recorded from Superchris, averaged over 24 trials of correct and incorrect responses, respectively. The heatmaps for the two groups of trials largely exhibit similar values across most blocks. Additionally,
Figure 16 illustrates several pairs of wavelet coherence between subprocesses at different scales during trials where the rat responded correctly to the stimulus. The heatmap matrices are not symmetric as single-scale cases because of the changes in scales.
To investigate whether there are changes in brain connectivity when the rat makes mistakes compared to correct responses, we conducted a permutation test with 1000 replicates. This test aimed to identify significant differences in the average wavelet coherence between trials with correct and incorrect responses.
Figure 17 presents the
p-value results across channels and specific pairs of subprocesses, corresponding to the cross-scale coherence shown in
Figure 16.
Based on the results of the permutation test, we selected several pairs of scale-specific subprocesses from different channels that correspond to significant
p-values in
Figure 17. This selection aimed to verify whether the differences between correct and incorrect trials are clearly observable.
Figure 18 demonstrates that the cross-scale coherence between these selected subprocesses shows significant differences across the two types of trials. Our framework effectively captures the time-evolving coherence, revealing that, in most cases, the coherence during incorrect trials is substantially higher than that observed in correct trials.
The analysis reveals numerous alterations in brain connectivity during incorrect responses, even in cross-scale interactions between different regions. These findings highlight the effectiveness of wavelet coherence as a powerful tool for capturing critical dynamics in brain activity.
7. Topological Data Analysis
As described in the previous sections, numerous methods have been proposed to estimate brain connectivity, spanning from correlation- and coherence-based measures to Granger causality, transfer entropy, and wavelet-based approaches for non-stationary data. The subsequent step typically involves performing a brain network analysis under different scenarios.
Brain network analysis has emerged as a vital area of research for understanding neural connectivity and its role in cognitive and physiological processes [
5]. Over the past few decades, this field has been shaped by foundational studies in network science, such as the concepts of small-world networks [
61] and scale-free networks [
62], which highlighted the key organizational principles of brain networks. These studies have motivated the application of graph-theoretic approaches to analyze brain connectivity, providing valuable insights into the structural and functional organization of the brain [
63,
64].
Graph-theoretic measures, such as the clustering coefficient and modularity, have been extensively used in brain network analysis. The clustering coefficient quantifies the extent to which a node is connected to or influences its neighbors, providing insights into how different brain regions collaborate to process information. Modularity, in contrast, measures the extent to which a network can be divided into distinct communities or modules with dense intra-community connections and sparse inter-community links, offering a deeper understanding of functional segregation in the brain. Physiologically, these measures have been linked to cognitive processes such as information integration and function segregation [
65,
66].
Despite their utility, graph-theoretic approaches have notable limitations. One major challenge is the thresholding problem, where the process of binarizing or sparsifying (creating an adjacency matrix from edge weights) connectivity matrices can significantly affect results, introducing subjectivity and potential bias [
67,
68]. Additionally, measures like clustering coefficient and modularity are summaries of the graph and may overlook more intricate, multi-scale interactions beyond pairwise relationships within the brain network. This limitation has prompted the exploration of alternative methods that can capture the richer and more detailed features of connectivity.
Topological data analysis (TDA) methods have gained significant momentum in recent years, especially in the analysis of brain signals due to its ability to characterize the shape and structure of multivariate time series data across multiple scales [
69,
70]. One of the key tools in TDA, persistent homology (PH), has proven particularly powerful for understanding the topological structure of data. For example, Lee et al. [
71] was among the first to introduce PH to brain network analysis, comparing functional networks across groups of children with ADHD, ASD, and typical development. Similarly, Wang et al. [
72] applied PH to event-related potentials, successfully detecting differences between post-stroke aphasic individuals and healthy controls under conditions of altered auditory feedback. Additionally, Saggar et al. [
73] demonstrated the utility of the Mapper algorithm for reducing the dimensionality of connectivity graphs, thereby facilitating the analysis of dynamic brain networks and task-related effects.
PH analyzes the evolution of topological features, such as connected components (clusters), loops (cycles), and higher-dimensional voids, across a scale parameter. As illustrated in
Figure 19, PH constructs a filtration, that is, a nested sequence of simplicial complexes that extend the notion of networks beyond pairwise interactions. By tracking the “birth” and “death” of these features, PH reveals the scales at which significant topological structures emerge, offering a nuanced view of neural connectivity. The figure demonstrates this process with two examples: the top row represents a dataset with two distinct clusters, while the bottom row illustrates a dataset with a single prominent cycle. As the parameter
increases (illustrated by growing balls around the data points in the four rightmost columns), the filtration encodes the topology at different scales. For the clusters, the features persist until they merge, while for the cycle, the loop appears at a certain scale and disappears at another. PH typically summarizes this information using visual tools such as barcodes [
74], persistence landscapes [
75], persistence images [
76], or persistence diagrams as illustrated in the left column of
Figure 19, where for each dimension, the birth–death pairs
of topological features are plotted as points in the
-coordinate system, with different dimensions represented by distinct colors. For instance, connected components (
) are shown as points on the y-axis (
), while cycles (
) are points in the upper triangle (
) with points farther away from the diagonal, indicating a longer persistence.
By tracking these features and representing them in diagrams, PH offers a deeper understanding of the underlying data structure. In the context of brain networks, it reveals connectivity patterns that extend beyond traditional graph-theoretic measures, providing a robust framework for studying neural dynamics and organization. For time series data, the Vietoris–Rips filtration can be constructed in various ways [
70].
To analyze the connectivity patterns in the rat’s LFP data, we focus on in-sequence trials (A, B, C, D, E) where the rat made correct decisions recognizing the odor, resulting in a total of 190 trials. For each trial, we estimate coherence matrices across different frequency bands and construct the corresponding persistence diagrams (PDs).
Figure 20 illustrates the results for two selected trials (Trial 1 for odor A and Trial 100 for odor B) at two frequency bands (0–12 Hz and 12–30 Hz). The first row displays the coherence matrices for each trial and frequency band, offering insights into the pairwise interactions between regions. The second row showcases the associated persistence diagrams, which summarize the birth and death times of the 0D and 1D topological features, providing a compact representation of the topological structure inherent in the coherence matrices.
All four coherence matrices appear to display similar information, with three main clusters visible across the trials. Within the rat hippocampus, the distal (first half of the tetrodes) and proximal (second half of the tetrodes) regions exhibit distinct patterns: two clusters are evident in the distal region, while only a single cluster is apparent in the proximal region. However, the visually distinguishing differences between the coherence matrices across trials and frequency bands remains challenging. In contrast, the PDs provide more precise information on the birth and death times of topological features, clearly highlighting differences between the trials that are not easily observable in the coherence matrices.
A persistence diagram
for a dimension
k is a multiset of birth–death pairs:
where each
encodes the appearance (
) and disappearance (
) scales of a topological feature of dimension
k.
This approach enables a rigorous comparison of the topological structures across trials, offering insights that surpass traditional graph-theoretic measures. By estimating dependence for each trial and applying PH, we assess the topological features in connectivity graphs. Using the Wasserstein distance (see Equation (
17)), we quantify changes in connected components (dimension 0) and cycles (dimension 1) across trials, uncovering subtle differences in neural dynamics. This approach enables a rigorous comparison of the topological structures across trials, offering insights that surpass traditional graph-theoretic measures. By estimating dependence for each trial and applying PH, we assess the topological features in connectivity graphs. Using the Wasserstein distance (see Equation (
17)), we quantify changes in connected components (dimension 0) and cycles (dimension 1) across trials, uncovering subtle differences in the neural dynamics:
where
ranges over all bijective matchings between points in
and points in
(possibly adding diagonal points if needed).
To analyze the partial correlation dependence, we compute the Wasserstein distances between persistence diagrams across trials. The results are presented in
Figure 21, where the top row shows the full distance matrices for dimensions 0 (left) and 1 (right), and the bottom row summarizes averages grouped by odors (lemon, anise, rum, vanilla, and banana). For dimension 0, the variability in connected components is notably smaller for trials corresponding to the first odor (lemon) compared to the others. Conversely, for dimension 1, the second odor (anise) shows the lowest variability, suggesting distinct patterns in topological features depending on the odor and dimension.
The coherence-based analysis highlights frequency-specific topological patterns across trials and odors.
Figure 22 focuses on dimension 0, showing that the lemon odor exhibits lower variability in connected components within the delta and theta bands but not in the beta band. This indicates that distinct frequency bands capture different aspects of connectivity. Meanwhile,
Figure 23 explores cycles (dimension 1) across the same frequency bands, showing relatively consistent patterns with no strong odor-specific differences.
By combining these results, we highlight the value of PH in identifying nuanced patterns in brain connectivity networks. The coherence-based analysis underscores the ability to pinpoint frequency bands where significant topological changes occur, while the partial correlation analysis reveals odor-specific differences in neural connectivity. Together, these findings demonstrate the potential of persistent homology to provide a deeper understanding of the organization and dynamics of complex neural systems.
In summary, one of the key strengths of TDA lies in its ability to move beyond pairwise interactions and capture higher-order structures in complex networks. While many traditional graph-based tools focus on edges between pairs of nodes, TDA leverages simplicial complexes to incorporate multi-node (higher-order) dependencies. In parallel, new frameworks have emerged to integrate higher-order interactions into measures of dependence such as transfer entropy [
77], and recent investigations of brain connectivity underscore the importance of these high-order relationships [
78,
79]. Advanced TDA approaches like Hodge decomposition have extended standard methodologies (e.g., persistent homology) from handling only symmetric connectivity to accommodating non-symmetric dependency measures [
80]. This expansion enables TDA to capture global topological patterns (gradient, local, and global loops) when the underlying networks arise from effective connectivity. The ability to incorporate such higher-order interactions in non-symmetric, directed settings highlights a promising avenue for future research.
8. Discussion
As fundamental measures of linear association, correlation and coherence have been extensively utilized to assess functional connectivity in neuroscience research. Coherence, in particular, offers a more nuanced analysis when dependence is driven by specific oscillations, forming the foundational elements of most brain connectivity analyses.
Building upon these principles, KenCoh has been developed to address some of the limitations inherent in coherence. Specifically, it enhances the ability to discern more complex patterns of connectivity that are not readily apparent with traditional coherence measures. Moreover, it provides region-to-region analyses that align with the spatial orientation of most brain imaging data. In
Section 3, we apply KenCoh to LFP data to investigate Buchanan’s brain connectivity during in-sequence and out-of-sequence trials. The results indicate that the same key tetrodes contribute more to global coherence in the beta band during in-sequence trials, contrasting with the findings from out-of-sequence trials. This suggests a potential role in the pattern recognition skills of rats.
Granger causality analysis is designed to capture directional interactions between brain regions, offering deeper insights into effective connectivity. Unlike functional connectivity which reflects statistical associations without implying causation, GC indicates the direction of information flow between regions. This directional information is especially valuable for studying specific pathways, such as those involved in sensory integration or memory formation, without needing to map the entire connectivity structure. Moreover, when combined with sDPCA, GC preserves the dominant oscillatory activity in the broader network before focusing on pairwise interactions, thereby enhancing the reliability of inferred direct influences.
Pairwise GC analysis of the LFP recordings indicate a dominant flow from distal to proximal CA1 across odors, although certain subjects show additional reverse or two-way interactions. One subject exhibits extensive reciprocal connectivity, contrasting the more unilateral patterns observed in others. These findings suggest that hippocampal olfactory processing depends on a shared distal-to-proximal route, with odor-specific link changes superimposed, reflecting subtle variations in how each subject’s CA1 circuitry responds to different odor conditions.
One major advantage of the STE method in analyzing LFP data is that it enables the capturing of nonlinear (possibly cross-frequency) information transfer between nodes in a brain network, with minimal assumptions on the distribution or type of relationship between the signals. That is, it allows for quantifying effective brain connectivity that concentrates on specific frequency bands, which makes it straightforward to link results to well-established findings in cognitive neuroscience. Also, its application to understanding effective brain connectivity is not limited only to LFP data but to other brain imaging modalities such as EEG and functional near infrared spectroscopy (fNIRS). Moreover, its estimation is simple and computationally efficient, as it employs a vine copula approach as illustrated in [
52]. Since STE is defined over maximum amplitudes of non-overlapping time blocks, it is fairly robust to spontaneous noise artifacts which may primarily take effect at high-frequency oscillations. These advantages enable us to identify major differences in the magnitude of information flow between the distal and proximal regions of the subjects during in-sequence and out-of-sequence trials in the alpha and beta frequency bands.
A caveat, however, is that the STE approach assumes the stationarity of signals because it requires the extraction of band-specific oscillations through bandpass filtering (e.g., Butterworth filter). The stationarity assumption ensures the extracted signals appropriately capture the oscillations of interest. In our LFP analysis, this is not an issue since the 1.2 s segments we analyze exhibit quasi-stationary behavior. In addition, the temporal resolution of causality captured by STE is relatively slower than the actual sampling rate of the signal due to the aggregation over time blocks. Depending on the choice of block size m, the causal interpretations for the connections measured by STE change. Thus, practical considerations, aligning with the goals of the study, should be made before implementation to achieve its best performance.
Often brain signals are assumed to be stationary. However, this is not the case in many practical scenarios. Wavelet coherence addresses these challenges effectively by analyzing non-stationary time series and capturing time-varying statistical properties within these signals. The application of wavelet coherence analysis to LFP data helps us identify intriguing interactions between components at different scales across various channels. Furthermore, we observe distinct differences among channels when the rat makes mistakes compared to correct responses, providing valuable insights into the neural dynamics.
Persistent homology provides a robust framework for analyzing brain connectivity in LFP recordings by quantitatively assessing the shape and structure of high-dimensional brain networks. This capability is particularly valuable for investigating how different tasks or conditions affect brain organization and for comparing individuals with varying neurological disorders. Our analysis demonstrates that TDA can reveal subtle, yet meaningful, variations in neural connectivity that conventional methods often overlook. Specifically, persistence diagrams derived from both coherence and partial correlation matrices highlight the variations that are both odor specific and frequency specific. For instance, in the partial correlation analysis, the persistence diagrams show that trials associated with the lemon odor exhibit reduced variability in connected components (dimension 0), suggesting a more stable clustering of neural activity. In contrast, trials linked to the anise odor display lower variability in cyclic features (dimension 1), indicating more consistent loop structures. Similarly, the coherence-based analysis reveal that low-frequency bands (delta and theta) capture more stable connectivity patterns for the lemon odor compared to the beta band.
8.1. Advantages and Limitations
The presented methods present distinct strengths and limitations in capturing the various aspects of neural connectivity. To summarize these findings and aid in method selection,
Table 2 presents a comprehensive comparison of the advantages and limitations of each approach, addressing the specific challenges inherent in neural data analysis.
8.2. Future Directions
The theoretical properties of KenCoh remain an open question and warrant further investigation. In particular, it would be interesting to examine the performance of KenCoh when the group sizes of variables become large, i.e., as
P and
Q tend to infinity. A natural approach to addressing the high dimensionality of the problem is to regularize the canonical directions
and
. For instance, one could impose additional constraints, such as
, to obtain a sparse solution while solving the maximization problem in (
10).
In the relatively novel field of neural network-based Granger causality (NN-GC), we see an extension of traditional causality concepts. Conventional approaches assume a linear dependence structure in data or provide hand-selected basis functions or kernel transformations which requires domain knowledge and expertise. NN-GC approaches leverage the function approximation power of neural networks to model complex, nonlinear interactions that are not easily captured by standard statistical methods. These methods relax the linearity assumption of standard methods and learns data-driven features in an end-to-end manner through error backpropagation. Several methods based on NN have been recently proposed based on sparse regression. We refer the interested reader to [
81,
82,
83,
84] for more details.
Deep learning-based approaches to GC discovery from observational time series data have considerable potential due to the neural network models’ ability to learn task-specific, data-driven representations. Although sparse regression-based techniques have been proposed in the literature [
81,
82,
83,
84], these methods do not provide uncertainty quantification in their estimates. In addition, developing efficient Auto-ML and sensitivity analysis techniques to optimize hyperparameters (e.g., regularization, sparsity, and optimizer settings) and to reduce computational cost, as well as exploring time-varying and multi-scale GC analyses (where neuronal states can switch between connectivity patterns over different frequency bands or behavioral conditions), represent further directions to enhance these methods.
In its current formulation, STE addresses effective connectivity between nodes in a brain network in a pairwise manner. That is, it does not account for how other parts of the network, say signals from a third node, affect the strength of information transfer between the pair of nodes being investigated. Thus, one interest extension is to develop a new metric based on causation entropy, which is another information-theoretic measure, that captures the magnitude and direction of information flow between two variables after taking into account the contributions of other variables in the system.
The definition of wavelet coherence can be extended to address more complex scenarios, such as locally stationary partial coherence in the presence of high-dimensional confounders. This extension would further capitalize on the time–frequency localization capabilities of wavelets, offering improved sensitivity in detecting dynamic connectivity patterns.
While most TDA techniques focus on functional connectivity, Hodge decomposition, a method rooted in algebraic topology offers a complementary perspective on effective connectivity. By decomposing brain connectivity into gradient, curl, and harmonic components, this approach can reveal subtle dynamics in the flow of information that are often disrupted in neurological disorders [
80]. Furthermore, integrating this decomposition with machine learning techniques holds promise for detecting abnormal connectivity patterns associated with specific conditions, potentially paving the way for improved diagnostics and targeted therapies (e.g., in epilepsy).
9. Conclusions
In this paper, we analyzed brain connectivity data from the hippocampal region of rats using a diverse set of methods. Our approach spanned traditional techniques such as correlation, partial correlation, and coherence—and advanced methods including Granger causality, robust canonical coherence, spectral transfer entropy, wavelet coherence, and persistent homology. By comparing these techniques, we provided a detailed examination of their strengths, limitations, and their applicability to uncovering the complex interactions within neural systems.
Our findings demonstrate that classical methods serve as a reliable foundation for capturing linear and stationary relationships, while advanced techniques are better suited to capture nonlinear, dynamic, multi-scale and higher-order interactions.
The application of these methods to hippocampal LFP data revealed nuanced, odor-specific, and frequency-specific patterns in connectivity, which underscore the complex organization of neural circuits underlying nonspatial olfactory processing. Despite these promising results, several challenges remain, including the need for careful parameter tuning, computational efficiency, and improved interpretability of some of these advanced techniques.
Integrating these diverse methods into unified frameworks that leverage their complementary strengths could offer even deeper insights into brain connectivity. Moreover, the development of scalable algorithms and user-friendly software tools is essential for translating these advanced techniques into practical applications for neuroscience research.
By presenting a comprehensive suite of methods and applying them to hippocampal LFP data, this study aims to pave the way for further exploration and innovation in brain connectivity analysis.