Next Article in Journal
Quantum Attacks on Sum of Even–Mansour Construction with Linear Key Schedules
Previous Article in Journal
Closed-System Solution of the 1D Atom from Collision Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Representation Learning for Dynamic Functional Connectivities via Variational Dynamic Graph Latent Variable Models

College of Automation Science and Technology, South China University of Technology, Guangzhou 510641, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(2), 152; https://doi.org/10.3390/e24020152
Submission received: 6 December 2021 / Revised: 14 January 2022 / Accepted: 14 January 2022 / Published: 19 January 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Latent variable models (LVMs) for neural population spikes have revealed informative low-dimensional dynamics about the neural data and have become powerful tools for analyzing and interpreting neural activity. However, these approaches are unable to determine the neurophysiological meaning of the inferred latent dynamics. On the other hand, emerging evidence suggests that dynamic functional connectivities (DFC) may be responsible for neural activity patterns underlying cognition or behavior. We are interested in studying how DFC are associated with the low-dimensional structure of neural activities. Most existing LVMs are based on a point process and fail to model evolving relationships. In this work, we introduce a dynamic graph as the latent variable and develop a Variational Dynamic Graph Latent Variable Model (VDGLVM), a representation learning model based on the variational information bottleneck framework. VDGLVM utilizes a graph generative model and a graph neural network to capture dynamic communication between nodes that one has no access to from the observed data. The proposed computational model provides guaranteed behavior-decoding performance and improves LVMs by associating the inferred latent dynamics with probable DFC.

1. Introduction

Recent progress in invasive recording technologies, such as high-density microelectrode arrays [1,2,3] and optical fibers [4], allows access to large-scale neural population activities with the precision of single neurons. While these data are usually high-dimensional [5,6], the literature has shown that dimensionality-reduction approaches and generative models can faithfully explain population spike activities [7] and infer single-trial neural firing rates [8] with stable low-dimensional latent dynamics. At present, with rapid developments in machine learning and deep learning, the community has proposed several latent variable models (LVMs) with better efficiency and performance to extract the low-dimensional structure [9,10,11]. They bring novel insights into neuroscience [12,13] and facilitate the development of brain–computer interfaces [14].
However, existing LVMs consider the latent dynamics as abstract trajectories in the vector space without neurophysiological meaning. We are interested in how the coordination of different brain functions produces such low-dimensional trajectories. In other words, can we specify the exact meaning for each dimension of the trajectories based on neurophysiology? This question may help with understanding how cognitive or behavioral processes emerge. Recently, advances in network neuroscience [15,16] suggest that dynamic functional connectivities (DFC) may be a more accurate representation of functional brain networks [17,18,19,20]. This refers to dynamic relationships between distributed signals. DFC may be the mechanism for the coordination of activity between different neural networks to accomplish a complex task. Hence, it may also be a proper explanation for the latent dynamics discovered by LVMs. Nevertheless, while LVMs assume that latent dynamics follow a point process, DFC involve modeling dynamic relationships between variables. The community needs a more general framework to associate LVMs with DFC.
In this work, we improve LVMs and bridge LVMs and DFC by drawing on recent progress in graph representation learning and graph generative models [21], proposing a Variational Dynamic Graph Latent Variable Model (VDGLVM). We first generalize LVMs with the variational information bottleneck [22] framework to enable the design of non-Euclidean latent variables. We then leverage a dynamic graph as the latent variable to model population spiking and behavior simultaneously. The inferred evolution of relationships between nodes in the inferred dynamic graph can be associated with DFC implicitly. Thus, our model improves LVMs via specifying a probable explanation for the latent dynamics based on DFC. We evaluate our framework with real-world neural data. The results show that our method can not only decode behaviors with high performance but also identify simple yet interpretable representations that are informative of the task.

2. Background and Related Work

In this section, we describe the notations used in the present work, and we provide some background and related work on latent variable models, dynamic functional connectivity, and graph neural networks.

2.1. Latent Variable Models (LVMs)

We consider recorded spike count data from N neurons at T time steps, which results in an observation matrix X N N × T , with elements x i , t denoting the spike count of neuron i { 1 , , N } at time t { 1 , , T } , and the corresponding behavior Y R B × T , with B as the dimension of the behavior. The literature assumes that the spikes are generated from an inhomogeneous Poisson process based on an underlying non-negative value firing rate. Most LVMs try to find a lower-dimensional trajectory Z R L × T with L N , which explains the observed neural data. In this paper, we focus on dynamical models for behavior decoding. Usually, this is accomplished within the auto-encoder framework. Different choices of the encoder and the decoder may result in different LVMs, including the Gaussian Process (GP) [23,24,25,26,27], linear dynamical system (LDS) [28,29], multi-layer perceptrons (MLP) [30,31], recurrent neural networks (RNN) [8,26], etc. Moreover, since most LVMs are generative models, some approaches extract complex latent structures by carefully designing constraints on the latent space. For example, pi-VAE explicitly uses brain states as labels and models them with the latent variable in a supervised manner [31] and Swap-VAE leverages self-supervised learning techniques to decompose the latent space [30].
In this work, we consider the dynamic communications of nodes as the underlying dynamics. While existing LVMs only use a point process without exact neurophysiological meanings, our model introduces DFC to explain the latent dynamics. A slight yet significant difference between the proposed model and LVMs mentioned above is that we train our model to decode behavior in a supervised manner. This fits more closely with the information-based framework described below.

2.2. Dynamic Functional Connectivities (DFC)

Recently, the focus of the systems neuroscience community tends to shift from the analysis of the static correlation between signals to the study of dynamic relationships between different brain functions evolving with time [32]. The communications are often referred to as dynamic functional connectivities (DFC). To capture undirectional statistical dependencies between different neural signals, a common approach is investigating the correlations based on sliding windows [33,34,35,36]. To model directional dependencies, the most frequently used method is to estimate Granger causality [37,38,39]. Since DFC can be conveniently described using a dynamic graph, some works on optimizing the dynamic graph structure based on graph measures or designed constraints have been developed [40,41,42]. These approaches have been applied on modeling DFC based on functional magnetic resonance imaging (fMRI) [36], electroencephalograms [43,44], and neurons [42].
For these methods of studying DFC, the nodes are directly defined in the data, and the challenge accurately estimates the correlations or edges based on the data. However, to associate LVMs with DFC, the nodes and node information for determining the edges are not available. To bridge this gap and reveal a probable neurophysiological meaning to improve LVMs, the present work extends LVMs and explores a computational model to infer a dynamic graph that can be associated with DFC.

2.3. Graph Neural Networks (GNNs)

A graph G = ( V , E ) is defined by its node set V = { v 1 , v 2 , , v | V | } and edge set E V × V . The numbers of nodes and edges are denoted by | V | and | E | , respectively. The structure of the graph can be described by an adjacency matrix, A { 0 , 1 } | V | × | V | . We also have d-dimensional features for each node in the graph, denoted as X g R | V | × d .
GNNs are representation learning neural networks for graph-structured data [21,45,46]. They have achieved remarkable success in different areas, such as molecular graph generation [47], brain connectome analysis [48], etc. GNNs leverage a message-passing [49] procedure to obtain a representation for each node v i . It first aggregates features of its neighbor nodes N i to obtain a message m i for v i , with a multi-set function A G G R E G A T E  as:
m i = A G G R E G A T E h j , v j N i ,
where h j is the intermediate representation of node v j . Second, GNNs update the representation for each node v i with the message via a function C O M B I N E as:
h i = C O M B I N E ( h i , m i ) .
GNNs repeat these two steps to iteratively aggregate neighbor information to obtain better node representations. For graph-level representation learning, GNNs use an additional multi-set function R E A D O U T on the node embeddings to have a graph representation vector h g :
h g = R E A D O U T ( h k , v k V ) .
In this work, we define DFC as the latent variable instead of constructing edges on nodes predefined on the data. Thus, we leverage graph generative models to infer a dynamic graph as the latent variable. We also use GNNs to map the dynamic relationships between nodes and behavior.

3. Methodology

In this section, we introduce VDGLVM. We start by introducing the learning objectives of LVMs based on the variational information bottleneck framework to understand LVMs at a higher level. Within the framework, we build a generative model to utilize a dynamic graph as the latent variable. VDGLVM inherits the expressiveness of deep neural networks and attempts to interpret neural latent dynamics by LVMs based on DFC.

3.1. Variational Information Bottleneck (VIB)

The information-theoretical framework explains deep neural networks as an information bottleneck [50]. A network is optimized to find the representation mapping that maintains the maximum mutual information between the input and output [51]. The literature has shown that deep neural networks that constrain information from the input to an intermediate representation tend to be more robust [22].
For LVMs, we find the latent point process Z by maximizing the mutual information between the latent variable Z and the behavior variable Y , restricting the mutual information between the spikes variable X and the latent variable Z with a constant I c , resulting in a constrained optimization problem:
max M I ( Z ; Y ) s . t . M I ( X ; Z ) I c ,
where the exact form of M I ( Z ; Y ) is:
M I ( Z ; Y ) = E p ( Z , Y ) log p ( Z | Y ) p ( Z ) = E p ( Z , Y ) log p ( Y | Z ) p ( Y ) .
We can be further formulate the objective by introducing a Lagrange multiplier β > 0 as:
min M I ( Z , Y ) + β M I ( X , Z ) .
This suggests that we can tune β to find a suitable bottleneck for deep neural networks to restrict the mutual information between X and Z . However, the direct optimization of the above objective function is intractable since mutual information is notoriously difficult to compute when we only have access to samples but not the exact distributions.
VIB [22] solves this by maximizing a lower bound of M I ( Z , Y ) and minimizing an upper bound of M I ( X , Z ) simultaneously. The key is to use a variational distribution q ( Y | Z ) to approximate the intractable conditional distribution p ( Y | Z ) . This approximation indicates that K L ( p ( Y | Z ) | | q ( Y | Z ) ) 0 , where K L ( · | | · ) is the Kullback–Leibler divergence. Therefore, we have:
M I ( Z ; Y ) = E p ( Z , Y ) log p ( Y | Z ) p ( Y ) E p ( Z , Y ) log q ( Y | Z ) p ( Y ) = E p ( Z , Y ) log q ( Y | Z ) + H ( Y ) ,
where H ( Y ) is the differential entropy of the output variable Y . We neglect this term in optimization since it is irrelevant to the model once given the data. This bound is tight when q ( Y | Z ) = p ( Y | Z ) . Based on the graphical model of the information bottleneck principle, we have a lower bound of M I ( Z ; Y ) as:
M I ( Z ; Y ) X Y Z p ( X , Y ) p ( Z | X ) log q ( Y | Z ) d X d Y d Z .
Similarly, an upper bound of M I ( X ; Z ) is:
M I ( X ; Z ) E p ( X ) K L ( p ( Z | X ) | | q ( Z ) ) ,
where q ( Z ) is a variational approximation of the latent distribution p ( Z ) . The bound is tight when q ( Z ) = p ( Z ) . Using the following empirical data distribution,
p ( X , Y ) = 1 N s n = 1 N s δ X ( i ) ( X ( i ) ) δ Y ( i ) ( Y ( i ) ) ,
where N s is the number of samples and δ ( · ) is the Dirac delta function, we have the objective for LVMs based on VIB when the latent variable is a point process:
min 1 N s n = 1 N s Z p ( Z | X ( i ) ) log q ( Y ( i ) | Z ) d Z + β K L ( p ( Z | X ( i ) ) | | q ( Z ) ) ,
where p ( Z | X ) is the encoder and q ( Y | Z ) is the decoder. The first term is the likelihood of output Y , given latent variable Z , inferred from input X . In this work, Y is the behaviors of the subject, and thus, we formulate this term based on a Gaussian distribution. The second term is regarded as a regularization term for the latent variable Z , given the prior distribution q ( Z ) .

3.2. Dynamic Graphs as Dynamic Functional Connectivities

Most LVMs study neural activities in a particular region based on the behavior, while the behavior may be completed as a complex cognitive process involving communications between different brain regions. However, for LVMs, nodes and edges for defining a graph are not available on the data, and thus, one needs to infer them, given the data. This can be formulated as inferring the dynamic relationships between variables that are not defined in the observed data. Meanwhile, graphs are suitable to model dependencies. Recent progress on graph-based deep learning significantly promotes graph modeling both in performance and efficiency. We then wonder whether a graph is feasible as the latent variable of LVMs, such that it corresponds to DFC and provides a probable explanation for the latent dynamics based on graph generative models.
To simplify, we assume that the dynamic relationships faithfully capture the evolving nature of the observations. The present work aims to capture pair-wise dependencies between variables and emphasize the dynamic relationships in generating neural activities and behaviors. Thus, we do not consider node dynamics, but assume that the dynamic relationships already summarize all necessary information about the data. The nodes only serve as labels for the variables we are interested in.
Let G t = ( V , E t ) be a dynamic graph, described by its dynamic adjacency matrix A t and static node features X g , and let G = { G 1 , , G T } be a graph process. Our objective is to replace the point process Z with a graph process G within the VIB framework. However, it is intractable to directly solve the problem due to the discrete nature of graphs. We start by assuming that the graph is a Gilbert random graph [52], such that every possible edge of a graph is conditionally independent of each other. Let e i j B e r n o u l l i ( θ i j ) be the binary variable indicating whether the edge ( i , j ) exists, where θ i j is the probability that edge ( i , j ) occurs in a graph G. Based on the theory of Gilbert random graphs, the distribution of random graph variable G can be factorized as:
p ( G ) = ( i , j ) E p ( e i j ) .
We relax the edge weights from binary variables to continuous variables in the range ( 0 , 1 ) , such that we can optimize the objective based on gradient descent efficiently. Specifically, we give a weight e ^ i j to every edge ( i , j ) to reformulate G as a smoothed-graph variable G ^ . The edge weights can be summarized in a modified adjacency matrix A ^ [ 0 , 1 ] n × n . Therefore, with a smoothed-graph process G ^ = { G ^ 1 , , G ^ T } , the objective function becomes:
min 1 N s n = 1 N s G ^ p ( G ^ | X ( i ) ) log q ( Y ( i ) | G ^ ) d G ^ + β K L ( p ( G ^ | X ( i ) ) | | q ( G ^ ) ) ,
where p ( G ^ | X ( i ) ) is a sequential graph generative model serving as the encoder, and  q ( Y ( i ) | G ^ ) is a sequential graph neural network serving as the decoder.

3.3. VDGLVM

In the VIB framework, we propose the use of dynamic graphs as the latent variable to study the relationship dynamics between static variables with VDGLVM. The architecture of VDGLVM is shown in Figure 1. The network is comprised of 4 layers. The encoder p ( G ^ | X ( i ) ) includes the encoding block, extracting features from neural population spikes, and the graph generative model, generating graphs. The decoder q ( Y ( i ) | G ^ ) includes the graph neural network, which performs graph-level representation learning, and the output layer, which maps graph embeddings to output behavior.

3.3.1. Encoder

We instantiate p ( G ^ | X ( i ) ) based on dynamical models. To capture nonlinear, non-Markovian, long short-term time-dependent dynamics, we utilize the RNN to extract features. At each time step t, RNN reads the spikes observation x t and updates its hidden state h t R H by:
h t + 1 = R N N θ ( x t , h t ) ,
where R N N θ is a deterministic non-linear transition function parameterized by θ . In practice, the vanilla RNN may suffer from gradient issues; it becomes hard to train and leads to performance degeneration. Thus, normally, R N N θ is implemented via long short-term memory (LSTM) or a gated recurrent unit (GRU). It models the dynamic graph G ^ t by parameterizing a factorization of the joint sequence probability distribution as a product of conditional probabilities, such that:
p ( G ^ 1 , , G ^ T ) = t = 1 T p ( G ^ t | G ^ < t ) p ( G ^ t | G ^ < t ) = p ( G ^ t | h t ) ,
where we instantiate p ϕ ( G ^ t | h t ) as a graph generative model parameterized by ϕ , which maps the RNN hidden state h t to a probability distribution of the graph variable G ^ t , which can be factorized as:
p ( G ^ t ) = i = 1 n j = 1 n p ( e ^ i j t ) .
To compute egde weights, we first use n MLPs to generate n Gaussian latent variables v i t R d , summarized in V t R n × d , from the RNN hidden state, as:
p v i t | h t = N μ v i t , σ v i t 2 μ v i t , σ v i t 2 = M L P i ( h t ) ,
where, in practice, we share the parameters of M L P i , except for the last layer. We apply the reparameterization trick to optimize the objective function with gradient-based methods. Specifically, we sample ϵ from a standard normal distribution, and then we generate v i t via:
ϵ N ( 0 , I ) v i t = μ v i t + ϵ σ v i t .
To perform link prediction, we compute the edge weights e ^ i j t based on pair-wise distance defined by cosine similarity:
p ( G ^ t | V t ) = i = 1 n j = 1 n p ( e ^ i j t | v i t , v j t ) e ^ i j t = σ log w i j t log ( 1 w i j t ) τ w i j t = 1 2 v i t T v j t v i t × v j t + 1 ,
where σ ( x ) = 1 / 1 + e x is the sigmoid function and τ ( 0 , 1 ) is temperature. When τ 0 , we have:
lim τ 0 p ( e ^ i j t | v i t , v j t ) = 0 , i f w i j t < 0.5 , lim τ 0 p ( e ^ i j t | v i t , v j t ) = 1 , i f w i j t > 0.5 ,
and p ( e ^ i j t | v i t , v j t ) becomes more sensitive when w i j t is around 0.5 . When τ = 1 , p ( e ^ i j t | v i t , v j t ) = w i j t . Our solution differs from previous works, generating edges such as: p ( e ^ i j t | v i t , v j t ) = σ v i t T v j t . Although both procedures use a continuous distribution to approximate the Bernoulli distribution, the logarithm and the normalization provide more stable gradients. With a proper temperature τ , the objective function is smoothed with a well-defined gradient e ^ i j t w i j t . This is similar to the binary concrete distribution, but we do not involve the reparameterization trick and sample from the distribution, and the formula is slightly different. We provide this scheme as a better alternative to generate connectivities from a set of vectors.
Moreover, when τ , the weight e ^ i j t tends to be irrelevant to w i j t as:
lim τ p ( e ^ i j t | v i t , v j t ) = 0.5 .
This means that, when τ > 1 , this computation is similar to the VIB, which controls information from w i j t to e ^ i j t given a prior q ( e ^ i j t ) = 0.5 . Nevertheless, a focal solution for G ^ t , such that e ^ i j t stays around 0 or 1, may be more informative about the data, since each functional connectivities may distribute more separately. Hence, here, we study τ ( 0 , 1 ) .

3.3.2. Decoder

We design q ( Y ( i ) | G ^ ) as a combination of a sequential graph-level GNN p ω 0 ( r t | G ^ t ) parameterized by ω 0 , and a multi-layer neural network M L P ω 1 parameterized by ω 1 , where r t R d is the graph-embedding vector for G ^ t . We implement p ω 0 ( r t | G ^ t ) based on a 1-layer Graph Isomorphism Network (GIN) [53]:
X g = M L P η A ^ t + ( 1 + γ ) I X g r t = s u m ( X g ) ,
where M L P η is a multi-layer neural network parameterized by η , γ is a hyperparameter, X g R n × d is the matrix for n d-dimensional static variables defined by the delta function, A ^ t is the dynamic adjacency matrix, and  s u m is a summation across nodes serving as the R E A D O U T function. Without a loss of generality, we treat static variables X g as learnable parameters, optimized jointly with the model.
Finally, we map the graph-level representation to the output behaviors via a multi-layer neural network M L P ω :
y t = M L P ω ( r t ) .
We do not utilize time information in the decoder explicitly. Instead, we decode output behaviors y t from graphs G ^ t for every time step t separately, since the time-dependent dynamics are already extracted by R N N θ .

3.4. Differences from Existing Work

In this subsection, we briefly discuss the differences between our work and existing dynamic graph-learning methods based on optimization.
It is of great interest to learn the underlying graph structure of observed multivariate time series data. Chepuri et al. [54] propose a general learning scheme for noisy signals with a prior smoothness. Specifically, this learning approach finds denoised signals and a K-sparse graph structure from the data via optimizing:
arg min X ^ , A ^ X ^ X 2 + γ t r ( X ^ T L ( A ^ ) X ^ ) s . t . i , j e ^ i j = K ,
where X R N × T is the noisy data, X ^ is the denoised signals, L ( A ^ ) is the graph Laplacian matrix given the structure A ^ , and  γ and K are hyperparameters. To generalize this learning procedure to dynamic graphs learning to study DFC, Jiang et al. [42] find the solution based on:
arg min { x ^ t } , A ^ t t = 1 T x ^ t x t 2 + γ 1 t r ( x ^ t T L ( A ^ t ) x ^ t ) + γ 2 t = 1 T 1 A ^ t A ^ t + 1 1 s . t . i , j e ^ i j t = K ,
where x t R N is the noisy signals on time step t, and the additional constraint serves as temporal smoothness. One should notice two important differences between the optimization approaches and VDGLVM, namely, the modeling goal and the learning framework.
First, we define the graph structure on Gaussian latent variables V t R n × d inferred from the observed data x t , rather than on the data directly. The interest of the present work is not learning correlations of the observed signals, but finding probable dynamic communications between unseen latent variables underlying the observed data to interpret latent dynamics by LVMs. Since variables for DFC are less than the dimension of the data ( n < N ), we find the variables via a sequential VAE and perform link prediction on the variables to infer a graph from the data. This is why we introduce a dynamic graph as the latent variable and approximate the intractable posterior distribution with graph generative models. This is the most predominant difference.
Second, the learning framework is different. The optimization approach attempts to find the graph structure and eliminate additive Gaussian noise directly. To find the solution, the optimization framework requires varying forms of prior assumptions to constrain the solution space of the optimization problem. Consequently, it relies heavily on the formulation of the regularization terms, reflecting simplified prior knowledge of graph signals. Moreover, as indicated in Jiang et al. [42], computational complexity may be increased considerably with a more accurate modeling analysis. On the other hand, deep generative models can learn sophisticated mappings to find a more informative graph structure for its expressiveness and efficiency in a data-driven manner.
Nevertheless, the graph construction procedures are both based on the smoothness prior to the variables. t r ( X ^ T L ( A ^ ) X ^ ) is the Dirichlet energy of a graph, given the structure A ^ , which encourages similar graph signals to be connected. It describes the distance between two signals as e i j x i x j 2 , which is essentially consistent with our link-prediction scheme based on cosine similarity in Equation (19). Furthermore, the temporal smoothness priors are fundamentally consistent as well. While the optimization approach applies the assumption on the temporal graph property directly, VDGLVM merges the assumption in the RNN with an appropriate regularization on the parameters.

4. Experimental Setup

The validity and usefulness of our model depend on its performance when decoding behavior from neural spikes, and on whether the inferred representations preserve task-relevant structures. In this section, we first describe the datasets we use in the experiments. Then, we compare VDGLVM’s performance in decoding behaviors against other typical methods to demonstrate its feasibility. We further analyze qualitatively the representations that VDGLVM learned. Finally, we show the effects of two important hyperparameters on VDGLVM.

4.1. Dataset Description

We used two real-world neural datasets from the Neural Latents Benchmark [55] to demonstrate the interpretability of VDGLVM. All datasets are publicly available.
The primary dataset we conducted experiments on is Area2_Bump [56]. The data were recorded as a macaque was performing a center–out reaching task. Area2_Bump includes hand kinematics as behaviors and corresponding neural population spikes of neurons from area 2 of the somatosensory cortex.
We also used MC_Maze [57], and its scaled versions, MC_Maze_L, MC_Maze_M, and MC_Maze_S, as the secondary datasets to compare model performance on different datasets. The datasets consist of neural population spikes and the associated hand kinematics when a monkey made reaches with an instructed delay to visually presented targets while avoiding the boundaries of a virtual maze. The neurons were from the primary motor and dorsal premotor cortices, which directly involve motor control.
We used the recorded hand position as target behaviors and built models to map the neural spikes to the behaviors. We binned the ensemble spike activities into 20-ms bins.

4.2. Model Configurations

We used two unidirectional GRU layers as the feature extractor. The space of the features was 128-dimensional. We applied dropout with p = 0.5 and layer normalization to avoid over-fitting. MLPs were implemented with two hidden layers. To further accelerate model training, we used ELU as the nonlinear activation function. In our model, we used three nodes for the graph, resulting in three dynamic edges. The static node features were in 128-dimensional space. We set the temperature τ = 0.03 and VIB regularization multiplier β = 1 × 10 4 as defaults. The model was trained in a single Nvidia Titan RTX GPU for 256 epochs. We used the Adam optimizer with a learning rate of 0.001 to optimize the parameters. In our experiments, we split the dataset into 75 % as the training set and 25 % as the test set.

5. Results

5.1. Performance of Decoding Kinematics

We first wondered whether VDGLVM decodes behavior from neural spikes with a promised performance. We measured the performance by computing R 2 between the decoded behavior and normalized ground-truth behavior. To validate the expressiveness of VDGLVM, we assessed our model on different datasets and compared it against classic LVMs, including spike smoothing, GPFA, and SLDS. Since our model is based on nonlinear dynamical models and deep neural networks, it should significantly outperform these baselines. We also included a GRU-based RNN to compare, which is the backbone of VDGLVM without graphs and VIB. We trained these models in a supervised manner. The results are summarized in Table 1 (best results presented in bold font).
Spike smoothing, GPFA, and SLDS are essentially linear dynamical models. GPFA gives better results than spike smoothing because it accounts for the spikes across neurons and time in a probabilistic framework simultaneously. While both spike smoothing and GPFA benefit from the smoothness, SLDS uses a switching mechanism to approximate nonlinear dynamics with disparate linear dynamics and, thus, better models complex neural dynamics. RNN can capture more complex dynamics of nonlinear systems. Thus, we have the following observation from the table: in general, both RNN and VDGLVM outperform the methods based on linear models significantly. Another implicit advantage of RNN and VDGLVM is that they are end-to-end data-driven models, which are convenient to learn.
Furthermore, VDGLVM consistently achieves better performance than RNN. The predominant difference between VDGLVM and RNN is the design of the latent variable. This result proves the feasibility of utilizing graphs as the latent variable. It also implies that the introduced graph generative model and graph neural network may induce additional inductive bias. These modules ultimately improve expressiveness.
Deep neural networks tend to overfit. The results on MC_Maze and its scaled versions illustrate that both RNN and VDGLVM suffer from performance degeneration when samples are insufficient. However, VDGLVM is relatively alleviated, even though the model has extra trainable parameters for tackling graphs. Owing to the VIB framework, VDGLVM is less affected by noise and, thus, provides better generalization capability.

5.2. Latent Structure for Behavior Decoding

We are interested in the latent structure learned by VDGLVM. We include a three-dimensional latent space learned by RNN to compare, which represents existing LVMs based on a point process. Since we do not apply additional assumptions about the trajectories, the latent dynamics learned by VDGLVM should be similar to that by RNN. We visualize the dynamics after training VDGLVM and RNN on the A r e a 2 _ B u m p dataset with active trials in Figure 2.
Deep neural networks learn representations favorable to prediction. Since our main objective is decoding behavior from neural spikes, the actual dynamic functional connectivities may be further mapped to the representations favor for predictions. Therefore, as can be seen in Figure 2b,f, the latent trajectories learned by both VDGLVM and RNN preserve the task structure (reaching trajectories) in three-dimensional space.
However, the representations by RNN are only a low-dimensional embedding of the prediction, while the representations by VDGLVM contain a more informative structure. Looking at Figure 2a, we found that at least one dimension of the latent trajectories stays around 0 or 1 and provides no information about the dynamics. The results indicate that the dynamic relationships are local, governed by a subset of relationships. This sparsity agrees with the results of the constructed graph in previous work studying DFC [20]. The literature has shown that communication efficiency is one of the fundamental attributes that support neural function. In the language of complex networks, sparser connections between modules enable efficient inter-module integrations of information. We capture this by properly parameterizing the edge weights. The property is missed in most LVMs based on a point process, as shown in Figure 2e. Therefore, the dynamics graph inferred by our model provides a probable explanation, close to DFC, for latent dynamics by LVMs.
This illustrates that using a dynamic graph as the latent variable does associate latent dynamics with probable DFC. Meanwhile, as shown in Figure 2h, one may interpret the point process as a complete graph with node dynamics, where each node corresponds to one dimension of the trajectories. However, this trivial consideration does not have useful and practical insight into neural dynamics. On the contrary, as shown in Figure 2d, the graph by VDGLVM specifies an exact meaning for the dynamics as communications between nodes. One can conveniently associate the nodes with variables defined in the brain. Thus, we conclude that VDGLVM provides feasible neurophysiological interpretability to the inferred latent dynamics by existing LVMs.

5.3. Hyperparameter Tuning

We empirically analyzed two key hyperparameters by adjusting them and evaluating the resulting decoding performance.

5.3.1. β in VIB Objective Function

Our model is based on the VIB framework. It restricts the information flow from the observations X to the latent variable G ^ to minimize the VIB objective function. Given a proper β , this encourages the model to focus on information relevant to output Y in X . Meanwhile, the model learns to drop useless information, such as noise. Therefore, compared with the vanilla auto-encoder, the VIB-based model improves the generalization performance. However, an inappropriately large β will be harmful to the model. Since VIB forces the latent coding towards a given prior, the bottleneck becomes so small that the information flow is blocked. With extremely large β values, the model learns to transform the distribution of input p ( X ) to the prior q ( G ^ ) , which is completely uninformative for decoding.
The property of β has been extensively studied in the literature. We verified this by studying the relationship between β and model performance. We present the result in Figure 3a. We manually adjusted β from typical settings to illustrate its effects on the model performance revealed by R 2 . As shown in the figure, when we use β 0 , increasing β improves the decoding performance against the baseline. The network attains the best when 10 5 β 10 4 . Nevertheless, with a larger β , more and more discriminative information about input and output are blocked from input to latent coding. Thus, the model becomes seriously degenerated.

5.3.2. τ in Graph Generation

Another hyperparameter significant for the model is the temperature τ . As discussed above, we only study τ ( 0 , 1 ) . A small τ makes the sigmoid function steeper; it tends to be a step function and, thus, is more sensitive for the input around 0.5 . More mathematically, with a smaller τ , the function has a higher Lipschitz constant. Since we directly involve this term in the propagation process, the Lipschitz constant of the model increases accordingly. While a smaller τ extracts more stable graph structures, an improperly small τ may make the model over-confident and, thus, less robust. We studied the association between τ and model performance. We present the result in Figure 3b. We found that the model may become unstable when τ < 0.01 , as the objective function becomes hard to optimize. Moreover, necessary dynamics are severely suppressed and thus insufficient to decode the target behavior. Therefore, the model tends to learn a trivial solution for the objective function.

6. Conclusions

Latent variable models (LVMs) for modeling neural data fail to determine the exact meaning of the inferred latent dynamics, while the literature has shown that dynamic functional connectivities (DFC) may be a possible mechanism of governing cognitive or perceptual tasks. In the present work, we propose Variational Dynamic Graph Latent Variable Model (VDGLVM), a representation learning model improving existing LVMs by interpreting the latent dynamics as DFC. To accomplish this, we design our model based on the variational information bottleneck and propose the use of a dynamic graph as the latent variable. Our model provides a probable explanation for latent dynamics captured by LVMs based on neurophysiology. In addition to the behavior dataset tested in the paper, we hope that the proposed model has the potential to apply to the analysis of neural activities in other perceptive and cognitive processes.

Author Contributions

Conceptualization, Y.H.; methodology, Y.H.; software, Y.H.; validation, Y.H.; formal analysis, Y.H.; investigation, Y.H.; resources, Z.Y.; data curation, Y.H.; writing—original draft preparation, Y.H.; writing—review and editing, Y.H. and Z.Y.; visualization, Y.H.; supervision, Z.Y.; project administration, Y.H.; funding acquisition, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grants 61836003, 61573150, and 61573152, as well as the International Cooperation open Project of the State Key Laboratory of Subtropical Building Science, South China University of Technology (2019ZA01).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The datasets can be found in https://gui.dandiarchive.org/#/dandiset (accessed on 5 December 2021).

Acknowledgments

The authors acknowledge all of the anonymous reviewers for their constructive comments that helped to improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jun, J.J.; Steinmetz, N.A.; Siegle, J.H.; Denman, D.J.; Bauza, M.; Barbarits, B.; Lee, A.K.; Anastassiou, C.A.; Andrei, A.; Aydın, Ç.; et al. Fully integrated silicon probes for high-density recording of neural activity. Nature 2017, 551, 232–236. [Google Scholar] [CrossRef] [Green Version]
  2. Hong, G.; Lieber, C.M. Novel electrode technologies for neural recordings. Nat. Rev. Neurosci. 2019, 20, 330–345. [Google Scholar] [CrossRef]
  3. Steinmetz, N.A.; Aydin, C.; Lebedeva, A.; Okun, M.; Pachitariu, M.; Bauza, M.; Beau, M.; Bhagat, J.; Böhm, C.; Broux, M.; et al. Neuropixels 2.0: A miniaturized high-density probe for stable, long-term brain recordings. Science 2021, 372, 6539. [Google Scholar] [CrossRef] [PubMed]
  4. Sych, Y.; Chernysheva, M.; Sumanovski, L.T.; Helmchen, F. High-density multi-fiber photometry for studying large-scale brain circuit dynamics. Nat. Methods 2019, 16, 553–560. [Google Scholar] [CrossRef] [Green Version]
  5. DiCarlo, J.J.; Zoccolan, D.; Rust, N.C. How does the brain solve visual object recognition? Neuron 2012, 73, 415–434. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Stringer, C.; Pachitariu, M.; Steinmetz, N.; Carandini, M.; Harris, K.D. High-dimensional geometry of population responses in visual cortex. Nature 2019, 571, 361–365. [Google Scholar] [CrossRef]
  7. Cunningham, J.P.; Byron, M.Y. Dimensionality reduction for large-scale neural recordings. Nat. Neurosci. 2014, 17, 1500–1509. [Google Scholar] [CrossRef]
  8. Pandarinath, C.; O’Shea, D.J.; Collins, J.; Jozefowicz, R.; Stavisky, S.D.; Kao, J.C.; Trautmann, E.M.; Kaufman, M.T.; Ryu, S.I.; Hochberg, L.R.; et al. Inferring single-trial neural population dynamics using sequential auto-encoders. Nat. Methods 2018, 15, 805–815. [Google Scholar] [CrossRef] [Green Version]
  9. Keshtkaran, M.R.; Pandarinath, C. Enabling hyperparameter optimization in sequential autoencoders for spiking neural data. Adv. Neural Inf. Process. Syst. 2019, 32, 15937–15947. [Google Scholar]
  10. Ye, J.; Pandarinath, C. Representation learning for neural population activity with Neural Data Transformers. Neurons Behav. Data Anal. Theory 2021. [Google Scholar] [CrossRef]
  11. Hurwitz, C.; Kudryashova, N.; Onken, A.; Hennig, M.H. Building population models for large-scale neural recordings: Opportunities and pitfalls. arXiv 2021, arXiv:2102.01807. [Google Scholar] [CrossRef]
  12. Gallego, J.A.; Perich, M.G.; Naufel, S.N.; Ethier, C.; Solla, S.A.; Miller, L.E. Cortical population activity within a preserved neural manifold underlies multiple motor behaviors. Nat. Commun. 2018, 9, 1–13. [Google Scholar] [CrossRef] [Green Version]
  13. Perich, M.G.; Gallego, J.A.; Miller, L.E. A neural population mechanism for rapid learning. Neuron 2018, 100, 964–976. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Degenhart, A.D.; Bishop, W.E.; Oby, E.R.; Tyler-Kabara, E.C.; Chase, S.M.; Batista, A.P.; Byron, M.Y. Stabilization of a brain–computer interface via the alignment of low-dimensional spaces of neural activity. Nat. Biomed. Eng. 2020, 4, 672–685. [Google Scholar] [CrossRef] [PubMed]
  15. Bassett, D.S.; Sporns, O. Network neuroscience. Nat. Neurosci. 2017, 20, 353–364. [Google Scholar] [CrossRef] [Green Version]
  16. Bassett, D.S.; Zurn, P.; Gold, J.I. On the nature and use of models in network neuroscience. Nat. Rev. Neurosci. 2018, 19, 566–578. [Google Scholar] [CrossRef] [PubMed]
  17. Breakspear, M. “Dynamic” connectivity in neural systems. Neuroinformatics 2004, 2, 205–224. [Google Scholar] [CrossRef]
  18. Hutchison, R.M.; Womelsdorf, T.; Allen, E.A.; Bandettini, P.A.; Calhoun, V.D.; Corbetta, M.; Della Penna, S.; Duyn, J.H.; Glover, G.H.; Gonzalez-Castillo, J.; et al. Dynamic functional connectivity: Promise, issues, and interpretations. Neuroimage 2013, 80, 360–378. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Gonzalez-Castillo, J.; Bandettini, P.A. Task-based dynamic functional connectivity: Recent findings and open questions. Neuroimage 2018, 180, 526–533. [Google Scholar] [CrossRef]
  20. Avena-Koenigsberger, A.; Misic, B.; Sporns, O. Communication dynamics in complex brain networks. Nat. Rev. Neurosci. 2018, 19, 17–33. [Google Scholar] [CrossRef]
  21. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [Green Version]
  22. Alemi, A.A.; Fischer, I.; Dillon, V.J.; Murphy, K. Deep Variational Information Bottleneck. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  23. Byron, M.Y.; Cunningham, J.P.; Santhanam, G.; Ryu, S.I.; Shenoy, K.V.; Sahani, M. Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity. In Proceedings of the Advances in Neural Information Processing Systems 22 (NIPS 2009), Vancouver, BC, Canada, 7–10 December 2009; pp. 1881–1888. [Google Scholar]
  24. Zhao, Y.; Park, I.M. Variational latent gaussian process for recovering single-trial dynamics from population spike trains. Neural Comput. 2017, 29, 1293–1316. [Google Scholar] [CrossRef] [PubMed]
  25. Wu, A.; Roy, N.A.; Keeley, S.; Pillow, J.W. Gaussian process based nonlinear latent structure discovery in multivariate spike train data. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 3499–3508. [Google Scholar]
  26. She, Q.; Wu, A. Neural dynamics discovery via gaussian process recurrent neural networks. In Proceedings of the 35th Uncertainty in Artificial Intelligence Conference, Virtual online, 3–6 August 2020; pp. 454–464. [Google Scholar]
  27. Liu, D.; Lengyel, M. A universal probabilistic spike count model reveals ongoing modulation of neural variability. In Proceedings of the Thirty-Fifth Conference on Neural Information Processing Systems, Online, 6–14 December 2021. [Google Scholar]
  28. Macke, J.H.; Buesing, L.; Cunningham, J.P.; Yu, B.M.; Shenoy, K.V.; Sahani, M. Empirical models of spiking in neural populations. In Proceedings of the Advances in Neural Information Processing Systems 24: 25th Conference on Neural Information Processing Systems (NIPS 2011), Granada, Spain, 12–15 December 2011; pp. 1350–1358. [Google Scholar]
  29. Gao, Y.; Archer, E.W.; Paninski, L.; Cunningham, J.P. Linear dynamical neural population models through nonlinear embeddings. Adv. Neural Inf. Process. Syst. 2016, 29, 163–171. [Google Scholar]
  30. Liu, R.; Azabou, M.; Dabagia, M.; Lin, C.H.; Gheshlaghi Azar, M.; Hengen, K.; Valko, M.; Dyer, E. Drop, Swap, and Generate: A Self-Supervised Approach for Generating Neural Activity. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–14 December 2021; Volume 34. [Google Scholar]
  31. Zhou, D.; Wei, X.X. Learning identifiable and interpretable latent models of high-dimensional neural activity using pi-VAE. Adv. Neural Inf. Process. Syst. 2020, 33, 7234–7247. [Google Scholar]
  32. Bastos, A.M.; Schoffelen, J.M. A tutorial review of functional connectivity analysis methods and their interpretational pitfalls. Front. Syst. Neurosci. 2016, 9, 175. [Google Scholar] [CrossRef] [Green Version]
  33. Handwerker, D.A.; Roopchansingh, V.; Gonzalez-Castillo, J.; Bandettini, P.A. Periodic changes in fMRI connectivity. Neuroimage 2012, 63, 1712–1719. [Google Scholar] [CrossRef] [Green Version]
  34. Thompson, G.J.; Magnuson, M.E.; Merritt, M.D.; Schwarb, H.; Pan, W.J.; McKinley, A.; Tripp, L.D.; Schumacher, E.H.; Keilholz, S.D. Short-time windows of correlation between large-scale functional brain networks predict vigilance intraindividually and interindividually. Hum. Brain Mapp. 2013, 34, 3280–3298. [Google Scholar] [CrossRef]
  35. Zalesky, A.; Fornito, A.; Cocchi, L.; Gollo, L.L.; Breakspear, M. Time-resolved resting-state brain networks. Proc. Natl. Acad. Sci. USA 2014, 111, 10341–10346. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Hindriks, R.; Adhikari, M.H.; Murayama, Y.; Ganzetti, M.; Mantini, D.; Logothetis, N.K.; Deco, G. Can sliding-window correlations reveal dynamic functional connectivity in resting-state fMRI? Neuroimage 2016, 127, 242–256. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Granger, C.W. Investigating causal relations by econometric models and cross-spectral methods. Econometrica 1969, 37, 424–438. [Google Scholar] [CrossRef]
  38. Dhamala, M.; Rangarajan, G.; Ding, M. Analyzing information flow in brain networks with nonparametric Granger causality. Neuroimage 2008, 41, 354–362. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. West, T.O.; Halliday, D.M.; Bressler, S.L.; Farmer, S.F.; Litvak, V. Measuring directed functional connectivity using non-parametric directionality analysis: Validation and comparison with non-parametric Granger Causality. NeuroImage 2020, 218, 116796. [Google Scholar] [CrossRef] [PubMed]
  40. Fallahi, A.; Pooyan, M.; Lotfi, N.; Baniasad, F.; Tapak, L.; Mohammadi-Mobarakeh, N.; Hashemi-Fesharaki, S.S.; Mehvari-Habibabadi, J.; Ay, M.R.; Nazem-Zadeh, M.R. Dynamic functional connectivity in temporal lobe epilepsy: A graph theoretical and machine learning approach. Neurol. Sci. 2021, 42, 2379–2390. [Google Scholar] [CrossRef]
  41. Qiao, C.; Hu, X.Y.; Xiao, L.; Calhoun, V.D.; Wang, Y.P. A deep autoencoder with sparse and graph Laplacian regularization for characterizing dynamic functional connectivity during brain development. Neurocomputing 2021, 456, 97–108. [Google Scholar] [CrossRef]
  42. Jiang, B.; Huang, Y.; Panahi, A.; Yu, Y.; Krim, H.; Smith, S.L. Dynamic Graph Learning: A Structure-Driven Approach. Mathematics 2021, 9, 168. [Google Scholar] [CrossRef]
  43. Dimitriadis, S.I.; Laskaris, N.A.; Del Rio-Portilla, Y.; Koudounis, G.C. Characterizing dynamic functional connectivity across sleep stages from EEG. Brain Topogr. 2009, 22, 119–133. [Google Scholar] [CrossRef]
  44. Allen, E.; Damaraju, E.; Eichele, T.; Wu, L.; Calhoun, V.D. EEG signatures of dynamic functional network connectivity states. Brain Topogr. 2018, 31, 101–116. [Google Scholar] [CrossRef]
  45. Gori, M.; Monfardini, G.; Scarselli, F. A new model for learning in graph domains. In Proceedings of the 2005 IEEE International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005; Volume 2, pp. 729–734. [Google Scholar]
  46. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef] [Green Version]
  47. Jin, W.; Barzilay, R.; Jaakkola, T. Junction tree variational autoencoder for molecular graph generation. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 2323–2332. [Google Scholar]
  48. Kawahara, J.; Brown, C.J.; Miller, S.P.; Booth, B.G.; Chau, V.; Grunau, R.E.; Zwicker, J.G.; Hamarneh, G. BrainNetCNN: Convolutional neural networks for brain networks; towards predicting neurodevelopment. NeuroImage 2017, 146, 1038–1049. [Google Scholar] [CrossRef] [PubMed]
  49. Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural message passing for quantum chemistry. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1263–1272. [Google Scholar]
  50. Tishby, N.; Zaslavsky, N. Deep learning and the information bottleneck principle. In Proceedings of the 2015 IEEE Information Theory Workshop (ITW), Jerusalem, Israel, 26 April–1 May 2015; pp. 1–5. [Google Scholar]
  51. Shwartz-Ziv, R.; Tishby, N. Opening the black box of deep neural networks via information. arXiv 2017, arXiv:1703.00810. [Google Scholar]
  52. Gilbert, E.N. Random graphs. Ann. Math. Stat. 1959, 30, 1141–1144. [Google Scholar] [CrossRef]
  53. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful are Graph Neural Networks? In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  54. Chepuri, S.P.; Liu, S.; Leus, G.; Hero, A.O. Learning sparse graphs under smoothness prior. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 6508–6512. [Google Scholar]
  55. Pei, F.C.; Ye, J.; Zoltowski, D.M.; Wu, A.; Chowdhury, R.H.; Sohn, H.; O’Doherty, J.E.; Shenoy, K.V.; Kaufman, M.; Churchland, M.M.; et al. Neural Latents Benchmark ‘21: Evaluating latent variable models of neural population activity. In Proceedings of the Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2), Virtual, 6–14 December 2021. [Google Scholar]
  56. Chowdhury, R.H.; Glaser, J.I.; Miller, L.E. Area 2 of primary somatosensory cortex encodes kinematics of the whole arm. eLife 2020, 9, e48198. [Google Scholar] [CrossRef] [PubMed]
  57. Churchland, M.M.; Cunningham, J.P.; Kaufman, M.T.; Ryu, S.I.; Shenoy, K.V. Cortical preparatory activity: Representation of movement or first cog in a dynamical machine? Neuron 2010, 68, 387–400. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Illustration of VDGLVM using graphs to model the dynamic relationships between static variables. The left side is the encoder, consisting of a dynamical model extracting features from spikes data and a graph generative network building graphs for every time step. The right side is the decoder, including a GNN computing embeddings for graphs and a MLP projector mapping embeddings to behaviors.
Figure 1. Illustration of VDGLVM using graphs to model the dynamic relationships between static variables. The left side is the encoder, consisting of a dynamical model extracting features from spikes data and a graph generative network building graphs for every time step. The right side is the decoder, including a GNN computing embeddings for graphs and a MLP projector mapping embeddings to behaviors.
Entropy 24 00152 g001
Figure 2. Inferred edge weights dynamics and behavior reconstructions by VDGLVM and inferred dynamics by RNN after training, where we use RNN to represent LVMs based on a point process. We visualize the edges evolving with time in (a). We combine them and show three-dimensional trajectories in (b), where each axis corresponds to an edge. We show behavior reconstructions for VDGLVM in (c), where the ground-truth trajectories are dash lines and decoding ones are solid lines. The graph with three dynamic edges and three nodes is shown in (d), where blue circles are static node variables and orange squares are dynamic edges. We show the three-dimensional trajectories inferred by RNN in (e,f). We show behavior reconstructions for RNN in (g). The equivalent graph structure of RNN is presented in (h), where orange circles are dynamic nodes.
Figure 2. Inferred edge weights dynamics and behavior reconstructions by VDGLVM and inferred dynamics by RNN after training, where we use RNN to represent LVMs based on a point process. We visualize the edges evolving with time in (a). We combine them and show three-dimensional trajectories in (b), where each axis corresponds to an edge. We show behavior reconstructions for VDGLVM in (c), where the ground-truth trajectories are dash lines and decoding ones are solid lines. The graph with three dynamic edges and three nodes is shown in (d), where blue circles are static node variables and orange squares are dynamic edges. We show the three-dimensional trajectories inferred by RNN in (e,f). We show behavior reconstructions for RNN in (g). The equivalent graph structure of RNN is presented in (h), where orange circles are dynamic nodes.
Entropy 24 00152 g002
Figure 3. Decoding performance with the changing (a) β and (b) τ . Results are assessed with typical settings of the hyperparameters across 32 Monte Carlo simulations. The results are shown on the left side as two box plots, and the means are shown as lines. To present a more clear illustration, the results are zoomed in and shown on the right side.
Figure 3. Decoding performance with the changing (a) β and (b) τ . Results are assessed with typical settings of the hyperparameters across 32 Monte Carlo simulations. The results are shown on the left side as two box plots, and the means are shown as lines. To present a more clear illustration, the results are zoomed in and shown on the right side.
Entropy 24 00152 g003
Table 1. Comparison of different model performances with R 2 .
Table 1. Comparison of different model performances with R 2 .
ModelArea2_BumpMC_MazeMC_Maze_LMC_Maze_MMC_Maze_S
Smoothing0.5950.6420.5970.5460.481
GPFA0.6130.6690.5980.5710.533
SLDS0.7620.8120.7920.7720.667
RNN0.9010.8960.8620.7940.710
VDGLVM0.9270.9120.8980.8180.794
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, Y.; Yu, Z. Representation Learning for Dynamic Functional Connectivities via Variational Dynamic Graph Latent Variable Models. Entropy 2022, 24, 152. https://doi.org/10.3390/e24020152

AMA Style

Huang Y, Yu Z. Representation Learning for Dynamic Functional Connectivities via Variational Dynamic Graph Latent Variable Models. Entropy. 2022; 24(2):152. https://doi.org/10.3390/e24020152

Chicago/Turabian Style

Huang, Yicong, and Zhuliang Yu. 2022. "Representation Learning for Dynamic Functional Connectivities via Variational Dynamic Graph Latent Variable Models" Entropy 24, no. 2: 152. https://doi.org/10.3390/e24020152

APA Style

Huang, Y., & Yu, Z. (2022). Representation Learning for Dynamic Functional Connectivities via Variational Dynamic Graph Latent Variable Models. Entropy, 24(2), 152. https://doi.org/10.3390/e24020152

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop