Next Article in Journal
Analysing All-Optical Random Bit Sequences Using Gap-Based Approaches
Previous Article in Journal
Estimating the Workability of Concrete with a Stereovision Camera during Mixing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiview Spatial-Temporal Meta-Learning for Multivariate Time Series Forecasting

1
School of Computer Science and Technology, Dalian University of Technology, Dalian 116024, China
2
School of Innovation and Entrepreneurship, Dalian University of Technology, Dalian 116024, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(14), 4473; https://doi.org/10.3390/s24144473
Submission received: 26 May 2024 / Revised: 5 July 2024 / Accepted: 8 July 2024 / Published: 10 July 2024
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Multivariate time series modeling has been essential in sensor-based data mining tasks. However, capturing complex dynamics caused by intra-variable (temporal) and inter-variable (spatial) relationships while simultaneously taking into account evolving data distributions is a non-trivial task, which faces accumulated computational overhead and multiple temporal patterns or distribution modes. Most existing methods focus on the former direction without adaptive task-specific learning ability. To this end, we developed a holistic spatial-temporal meta-learning probabilistic inference framework, entitled ST-MeLaPI, for the efficient and versatile learning of complex dynamics. Specifically, first, a multivariate relationship recognition module is utilized to learn task-specific inter-variable dependencies. Then, a multiview meta-learning and probabilistic inference strategy was designed to learn shared parameters while enabling the fast and flexible learning of task-specific parameters for different batches. At the core are spatial dependency-oriented and temporal pattern-oriented meta-learning approximate probabilistic inference modules, which can quickly adapt to changing environments via stochastic neurons at each timestamp. Finally, a gated aggregation scheme is leveraged to realize appropriate information selection for the generative style prediction. We benchmarked our approach against state-of-the-art methods with real-world data. The experimental results demonstrate the superiority of our approach over the baselines.

1. Introduction

Advanced data acquisition technologies lead to the real-time collection of sequential signals, thus inducing widespread data analysis and downstream applications, such as electricity forecasting [1,2], traffic flow prediction [3,4,5], inventory control [6], healthcare management [7], etc. In essence, due to inherent connections among multiple univariate time series, most of the aforementioned scenarios can be regarded as multivariate time series modeling tasks, which play a key role in the decision-making process of digital management. For instance, the sales data-driven prediction of product supply for chain stores can optimize inventory management, helping save millions of dollars in each year. However, capturing the complex temporal dynamics of the multivariate time series data is a very challenging task.
First, there exist complex intra- and inter-variable relations. For each variable, temporal dependencies might be entangled with different evolving patterns, e.g., various trends and seasonality [2]. Existing autoregressive methods, filtering methods, and deep forecasting models via recurrent connections, temporal attention, and causal convolution can only capture limited modalities. In addition, there exist interdependent and even causal relations [8,9], i.e., each variable depends not only on its past value but also on other variables. Nonetheless, these relations are not always available as prior knowledge. For example, pairwise relations of different stations might not be given in the electricity forecasting task. Hence, it is mission-critical to automatically extract underlying relation structures and then develop structure-enhanced temporal modeling techniques, especially about how to fuse temporal and spatial information. Recent advances [8,9,10] seek to couple graph convolution with temporal recurrent processing at each timestamp. Instead, inspired by the effectiveness of separate autoregressive models [11] and graph structures [12,13] and by the disentangled [14] and fine-grained multiview learning [15] in time series modeling, we argue that the multiview fine-grained fusion of temporal view- and spatial view-specific information are essential to make up for the information loss brought by tightly coupled schemes.
Second, mainstream learning paradigms assume the distribution consistency of the training and inference stages. Nonetheless, in most multivariate time series forecasting tasks, temporal patterns keep evolving [16] and can be significantly different (e.g., temporal burst, multivariate relationship changes), such that it can be regarded as time series modeling in diverse learning environments. One promising direction is the application of meta-learning strategies, e.g., MAML [17], in time series prediction tasks [18,19], which captures meta knowledge shared by different tasks for fast adaption in new few-shot learning tasks. For example, diverse spatio-temporal purchasing patterns brought by heterogeneous external factors require high-capacity multiple-task-aware meta-learning models [6] that can adapt to different learning environments via knowledge transfer. Despite some advances in spatio-temporal meta-learning [3,6,19], it is still unclear how to meta-fuse temporal and spatial information for better task-transfer performances. Furthermore, the evolving complex inter-variable relationships and intra-variable temporal information add further challenges to the fast-changing learning tasks.
Third, a fast adaptation capability to new or recurring patterns while being aware of the diversity of distributions is crucial in most multivariate time series forecasting tasks. For instance, the sudden change in electricity production might be caused by a change in the natural environment or mechanical problems, and a model should be able to capture such diversity while quickly adapting to current data distributions. While some recent works have successfully improved learning efficiency via either time series temporal modeling itself [1,2] or meta-learning schemes [6] and capture diversity via stochastic neurons in conditional generative models [20], the inherent multi-step nature of time series present challenges about how to realize position-specific diversities.
To tackle all the above challenges, we propose a holistic spatial-temporal meta-learning probabilistic inference framework, entitled ST-MeLaPI, for the efficient and versatile learning of complex dynamics of multivariate time series data. Unlike previous meta-learning scenarios where different time series learning tasks have clear boundaries, we focus on a general setting in which only time series data are provided without task specification. Recent progress in online time series forecasting [16] and continuous spatial-temporal meta-learning [21] inspired us to consider that samples in local consecutive time slots can be regarded as a specific learning task given the modality-rich and time-evolving nature of multivariate time series data. We chose the meta-learning framework VERSA [22] because it realizes multi-task probabilistic learning across all tasks through an amortization network that can directly take support datasets (with any number of shots, i.e., flexibility) as inputs, and outputs a distribution (rather than fixed value, i.e., versatility) over task-specific parameters in a single forward pass (i.e., rapid inference). Hence, it is more appropriate for capturing diverse temporal patterns spanning different periods of time series data. In essence, ST-MeLaPI can be regarded as a sequential structure derivation of VERSA.
More concretely, we first formalize diverse learning tasks by dividing the original data into different batches. For each task, a multivariate relationship identification module (MRI) [9] is leveraged to learn the task-specific inter-variable dependencies. Then, a multiview meta-learning and probabilistic inference (MV-MeLaPI) strategy is developed. Specifically, first, given the support set and query set in a task, a multiview feature encoder (MFE) captures spatial dependency- and temporal pattern-induced view-specific information, respectively. Second, a multiview position-aware network (MV-PAN) was designed to map the spatial dependency- and temporal pattern-induced information (after concatenation with prediction window data) of the support set information into stochastic input at each timestamp, respectively. Third, a cross-view gated generator (CV-GAG) is leveraged to realize cross-view position-aware fusion and generative style predictions at one forward operation.
To sum up, our main contributions are the following:
  • We propose a novel framework based on meta-learning approximate probabilistic inference, entitled ST-MeLaPI, for the efficient and versatile learning of multivariate time series data. ST-MeLaPI regards that each short time span serves as the dataset for a specific task and then leverages a pairwise variable relationship learner (MRI) to capture task-specific inter-variable dependencies, hence enabling a spatial dependency- and temporal pattern-induced approximate posterior for one forward all-timestamp prediction (MV-MeLaPI).
  • We developed a multiview position-aware meta-learning and inference strategy: MV-MeLaPI. It enables fast and versatile learning via position-aware stochastic neurons. Spatial dependency- and temporal pattern-induced view-specific stochastic inputs from the support set are generated. Then, a cross-view information concatenation of stochastic inputs and encoded (view-specific) query set was designed to the approximate posterior predictive distribution, which realizes multiview meta-probabilistic inference learning at each timestamp.
  • ST-MeLaPI can be used for both long-sequence and short-sequence time-series forecasting tasks. In this study, we conducted several experiments with real-world data. The extensive experimental results show improved learning performances.

2. Related Work

Multivariate time series forecasting techniques have attracted much attention in various application areas. Mainstream techniques can be divided into three categories:
(1)
Classic statistical tools. Typical approaches, such as ARIMA [23], Kalman Filters [24], and linear state-space models [25], are essentially linear, thus only capturing low-order temporal dynamics. Moreover, they cannot benefit from shared temporal patterns in the whole dataset.
(2)
Deep sequential neural networks. Nonlinear recurrent neural network (RNN) models [26] have shown advantages in capturing high-order complex dynamics. For instance, LSTNet [27] combines the convolutional neural network with recurrent skip connections to extract short-term and long-term temporal patterns. DeepAR [11] performs probabilistic prediction through the combination of an autoregressive method and RNN. Transformer-based models [1,2,28,29,30] utilize self-attention mechanisms [31] to capture the long-range dependencies for prediction. In addition, another popular and competitive architecture is the causal convolution or temporal convolutional network (TCN) [32]. For example, DeepGLO [33] combines a global matrix factorization model with temporal convolutional networks. Even though recent state-of-the-art decomposition-based models [2,34,35,36] achieve superior performances, they are mainly designed for capturing the seasonal or trend-cyclical patterns in time series, which is far from satisfactory due to the versatile patterns of time series, especially considering the evolving nature of the data. In light of this, some works adopt meta-learning strategies [18] to realize quick task adaption, or choose multitask learning [37] for adaptive learning and forecasting in non-stationary data contexts.
(3)
Multivariate relationship-aware techniques. The underlying inter-variable dependencies, e.g., correlation and causation, have shown promise in enhancing the prediction performance [9]. Therefore, recent advances focus on spatio-temporal learning; in particular, traffic forecasting models [3,4,10,38,39,40,41,42] consider spatial connections through the combination of graph neural networks (GNNs) [43] and recurrent neural networks, thus forecasting all time series simultaneously. However, the aforementioned methods assume the spatial connections of roads as prior knowledge, which is often not available in most multivariate time series forecasting tasks. Therefore, automatic inter-variable dependency structure learning has been integrated into temporal dynamics modeling [8,9,12,44]. For instance, NRI [8] uses an autoencoder approach that learns different structures given different multivariate time series inputs. GTS [9] first learns a probabilistic graph for the whole time series and then combines the graph neural network and temporal recurrent processing to model inter-variable structure-aware temporal dynamics. Nevertheless, these models either only learn a single structure given one set of K series spanning a whole time window [9] or cannot capture the diverse temporal patterns with fixed model parameters despite dynamic structures [8]. Recently, ML-PIP [6,22] was proposed, adopting a multitask meta-learning framework to perform approximate probabilistic inference under low-resource learning environments. However, on the one hand, ML-PIP assumes that spatial information is given and divides all data into different learning tasks according to spatial and temporal information simultaneously; on the other hand, it does not consider the multivariate time series prediction problem, much less a position-aware multiview meta-learning strategy. Hence, it is very different from our learning tasks and proposed model. But it provides us the insight that each short time period can be naturally regarded as a specific learning task given the modality-rich and time-evolving nature of multivariate time series, thus realizing data-efficient and versatile learning. However, the inherent multi-step nature of time series present challenges about how to realize the time-specific diversity. In particular, the effectiveness of sequential dynamics-only [1,11,27,33], graph structure-only [12,13] and entangled spatio-temporal modeling [9,44] motivated us to design multivariate relationships and temporal dynamics in a more sophisticated way. For example, Qu et al. [13] proposed a sparse connectivity network to identify temporal dynamics using an adaptive regularization approach; Informer [1] shows the power of self-attention in sequential modeling. In particular, both temporal and contextual information [15] enhance the representation learning of multivariate time series data, strengthening the necessities of the fine-grained multiview fusion of temporal and spatial information.

3. Methodology

In this section, we provide an explicit problem definition and then describe the architecture of our proposed framework, ST-MeLaPI (Figure 1), which consists of Multivariate Relationship Identification (MRI) and Multiview Meta-Learning and Probabilistic Inference (MV-MeLaPI).

3.1. Problem Definition

A multivariate time series forecasting task aims at modeling the conditional distribution P ( Y ( i ) | X ( i ) ) of the ith dynamic system with a set of K series, where X ( i ) R T x × K represents the conditioning range (past T x timestamps) data, and Y ( i ) R T y × K represents the prediction range (future T y timestamps) data. Note that we assume that the K series have already been aligned. In particular, our key motivation is that samples in a short time span serve as a specific learning task. Hence, we split the whole time series data into different batches. Each batch consisting of X R B × T x × K and Y R B × T y × K contains N support sets and M query sets with B = N + M . In our work, we let X b = { x b , 1 , , x b , T x | x b , t R K } and Y b = { y b , 1 , , y b , T y | y b , t R K } represent the input and output of the bth sample in the batch, respectively. By learning from different short time periods in the past, our model aims at capturing shared patterns (or parameters) via knowledge transfer while realizing fast and versatile learning and inference for future predictions via stochastic neurons at each timestamp.

3.2. Multivariate Relationship Identification

The Multivariate Relationship Identification module (MRI) aims at improving the forecasting performance by exploiting inter-variable relationships. In our work, we implemented MRI by following GTS [9]. It consists of an univariate time series feature extractor, an inter-variable linkage predictor, and a differentiable structure sampler based on the Gumbel-Softmax reparameterization trick.
To be specific, first, a univariate feature extractor maps each variable X i to a feature vector c i :
c i = W c ( c o n v 1 d ( X i ) ) + b c ,
where W c and b c are trainable parameters, and c o n v 1 d is a one-dimensional convolutional layer that captures a global intra-variable temporal pattern. Then, the link predictor yields a link probability δ i j [ 0 , 1 ] with Bernoulli distribution for each pair of feature vectors c i and c j :
δ i j = σ ( M L P ( c i | | c j ) ) ,
where | | denotes the concatenation operation, M L P represents a multi-layer perceptron, and σ represents a sigmoid activation function. Finally, the Gumbel reparameterization technique [45] is leveraged to solve the differentiability problem brought by the sampling operation, while generating global structure diversity. The final multivariate relationship structure A is learned as follows:
A i j = σ ( ( log ( δ i j / ( 1 δ i j ) ) + ( g i j 1 g i j 2 ) ) / τ ) ,
where g i j 1 , g i j 2 G u m b e l ( 0 , 1 ) for all i, j. When the temperature τ tends to 0, δ i j is used as the probability of A i j = 1 .
Since each task spans short time periods, the inter-variable relationship can be regarded as invariant in each task. Therefore, each sample shares the same structure, which is generated by inputting the whole conditioning range data in the support set into the MRI module. Moreover, other pairwise relationship structure learning approaches can be used as well, e.g., NRI [8] and MTPool [44]. We followed GTS since the end-to-end structure learning and diverse structure sampling strategy are more intuitive for flexible and efficient learning in multivariate time series prediction tasks.

3.3. Multiview Meta-Learning and Probabilistic Inference

Based on the task-specific multivariate relationship structure A, we can conduct inter-variable relationship-aware temporal recurrent processing. Different from coupled combination [9,10,39,44], we seek a fine-grained fusion of view-specific information. At the core is a multiview position-aware spatial-temporal meta-learning strategy MV-MeLaPI by following VERSA [22]. It uses meta-learning to capture shared information between tasks, and enables fast and versatile learning through an amortization network. As a temporal variant of VERSA, MV-MeLaPI realizes efficient and versatile model learning and approximate prediction via stochastic neurons at each timestamp. Fundamentally, it is composed of a Multiview Feature Encoder (MFE), a Multiview Position-aware Amortization Network (MV-PAN), and a Cross-view Gated Generator (CV-GAG). Both MFE and CV-GAG are shared parameters (i.e., θ in VERSA) that are common to all tasks, while MV-PAN (i.e., ϕ in VERSA) helps generate task-specific stochastic inputs ψ (parameters) for a regression (classification) task.
Note that global auxiliary information, such as hierarchical time stamps (week, month, and year) and agnostic time stamps (holidays, events), offers guidance in capturing temporal dynamics [1]. Therefore, we incorporate global auxiliary information through a linear mapping operation and concatenation with the original multivariate time series data. Finally, we obtain the input Z R B × T × K × F e of the MFE through a linear mapping:
Z = W H ( X e n | | T i m e ) + b H ,
where T i m e R B × T × K × 1 ; W H and b H are trainable parameters; and F e represents the dimension of the initial embedding for each variable. We take X e n = C o n c a t ( X , X 0 ) R B × T × K × 1 as input data, where X 0 R B × T y × K is a placeholder (set as 0) for the target sequence, and T = T x + T y .

3.3.1. Multiview Feature Encoder

The Multiview Feature Encoder (MFE) was designed to extract spatial dependency- and temporal pattern-induced view-specific features, leading to comprehensive information utilization. Specifically, it consists of a coarse-grained graph convolutional network (C-GCN) and a fine-grained Gated Recurrent Unit (F-GRU).
Coarse-grained Graph Convolutional Network. The C-GCN was designed to capture the spatial dependency-induced temporal dynamics. Note that we call it “coarse-grained” since it only captures inter-variable relationship-enhanced feature learning at each timestamp without modeling sequential temporal transitions. In effect, it learns a structure-enhanced feature embedding for each variable through a graph convolution. Specifically, based on the learned batch-specific (task-specific) adjacency matrix A, we used the GCN to obtain an updated representation for each variable, formalizing Z c :
Z c = σ ( D ˜ 1 2 A ˜ D ˜ 1 2 Z W C ) ,
where W C is the trainable parameter, A ˜ = A + I is the adjacency matrix joining the self-recursive edges, I is the unit matrix, and D ˜ is the degree matrix of A ˜ .
Fine-grained Gated Recurrent Unit. Most of the existing temporal prediction models rely on discrete snapshots, i.e., the model input is only the recorded data at each time stamp, which fails to explicitly consider the fluctuations of the data within two consecutive time points. To fill this gap, as shown in Figure 2, we propose a fine-grained Gated Recurrent Unit Network (F-GRU), which adds a fine-grained feature-forgetting gate to the traditional GRU model. Each F-GRU cell takes as input the feature vector z t 1 with z t of continuous timestamps in Z = { z 1 , , z T } :
z t 1 , t = W h f ( z t 1 | | z t ) + b h f , v t f = σ ( W v f · [ z t 2 , t 1 , z t 1 , t ] + b v f ) , r t f = σ ( W r f · [ z t 1 f , z t ] + b r f ) , u t f = σ ( W u f · [ z t 1 f , z t ] + b u f ) , z ^ t f = t a n h ( W f · [ r t f z t 1 f , v t f z t 2 , t 1 , z t 1 , t ] + b f ) , z t f = ( 1 u t f ) z t 1 f + u t f z ^ t f ,
where W h f , W v f , W r f , W u f , W f , b h f , b v f , b r f , b u f , and b f are all trainable parameters. We first concatenate z t 1 with z t as the input of a linear layer to obtain z t 1 , t , which represents continuous local context information at the tth step. The fine-grained forgetting gate v t f controls the degree of forgetting from previous local context information. When updating the memory content z ^ t f at the tth step, our model incorporates the fine-grained features z t 1 , t , in addition to the reset gate r t f and the fine-grained forgetting gate v t f . Finally, at each timestamp, the fine-grained temporal feature vector z t f is generated by the update gate u t f .

3.3.2. Multiview Position-Aware Amortization Network

Based on the aforementioned view-specific information, i.e., two view-specific feature embedding for each variable at each timestamp, an MV-PAN was developed to fully exploit shared information among tasks and generate task-specific parameters. Essentially, this amortization network realizes fast and versatile learning through outputting a distribution over task-specific parameters or stochastic inputs in a single forward pass. Specifically, for each task, we have support set D b c / f = { ( Z b , n c / f , y b , n ) } n = 1 N , and query set { ( Z ˜ b , m c / f , y ˜ b , m ) } m = 1 M . As mentioned in VERSA [22], forming a posterior distribution over the task-specific parameters p ( ψ b s / d | D b c / f , θ ) or stochastic input is usually intractable; hence, approximate posterior distribution q ϕ s ( ψ s | D c ) and q ϕ d ( ψ d | D f ) are learned via amortization networks ϕ s (spatial) and ϕ d (temporal), respectively. In practice, we employ feed forward neural networks for the amortization networks. First, the view-specific spatial-temporal representations of the support set are concatenated with the corresponding prediction window data. Then, an average instance-pooling operation ensures that the network can process any number of training observations while being permutation-invariant. After that, a linear operation helps generate the means and variances of a factorized Gaussian distribution, which approximates the real spatial-temporal posterior distributions over the task-specific stochastic inputs (or task-specific parameter information in a classification task):
μ s / d = W μ s / d ( 1 N i = 1 N ( Z i c / f | | y i ) ) + b μ s / d , σ s / d = W σ s / d ( 1 N i = 1 N ( Z i c / f | | y i ) ) + b σ s / d ,
where W μ s / d , W σ s / d , b μ s / d , and b σ s / d are trainable parameters. Finally, feature vectors Z ˜ c / f (generated through the MFE) of the conditioning window from the query set will be concatenated with the above stochastic inputs to form an approximate posterior predictive distribution:
q ϕ ( y ˜ | D ) = p ( y ˜ | ψ s , ψ d ) q ϕ s ( ψ s | D c ) q ϕ d ( ψ d | D f ) d ψ s d ψ d ,
where the neural network ϕ contains ϕ s and ϕ d . The use of amortized variational inference and neural networks enable fast predictions at test time without traditional second-order derivatives during training. Further details can be seen in VERSA [22]. We will introduce, in detail, how to model p ( y ˜ | ψ s , ψ d ) and q ϕ ( y ˜ | D ) in the following subsection.

3.3.3. Cross-View Gated Generator

After obtaining posterior distributions q ϕ s ( ψ s | D c ) and q ϕ d ( ψ d | D f ) of the spatial view- and temporal view-specific stochastic inputs through the MV-PAN, the Cross-view Gated Generator is used to realize posterior predictions. Here, the cross-view information fusion between view-specific stochastic inputs (from the support set) and encoded conditioning window data (from the query set) enables the query set to capture multiple views of temporal dynamics from the training dataset, while making up for the information utilization loss in each view.
Figure 3 demonstrates how to leverage the shared parameters θ of all tasks and support set to learn task-specific parameter { ψ b s / d } b = 1 B and to perform inference for the query set. To be specific, the spatial relationship-induced parameter ψ s is concatenated to the temporal pattern-induced query set information Z ˜ f and then sent to a fine-grained graph convolutional network (F-GCN) for predictive predictions. In addition, the temporal pattern-induced parameter ψ d is concatenated to the spatial relationship-induced query set information Z ˜ c and then sent to a coarse-grained gated recurrent unit network (C-GRU) for further learning.
We use a fusion function Ψ ( ) to complete the initialization of the information:
Z ˜ c , ψ = Ψ ( Z ˜ c , ψ d ) , Z ˜ f , ψ = Ψ ( Z ˜ f , ψ s ) .
Coarse-grained Gated Recurrent Unit. To broaden the receptive field when learning sequential dynamics, we use an attention mechanism to directly aggregate temporal information from the past w 1 time steps. Then, the reset gate in the Gated Recurrent Unit (GRU) can control the information flow from more steps away. Specifically, as shown in Figure 4,
α i = exp ( z ˜ i c , ψ · z ˜ t c , ψ ) j = t w t 1 exp ( z ˜ j c , ψ · z ˜ t c , ψ ) , f t c = i = t w t 1 α i z ˜ i c , ψ , r t c = σ ( W r c · [ f t c , z ˜ t c , ψ ] + b r c ) , u t c = σ ( W u c · [ h t 1 c , z ˜ t c , ψ ] + b u c ) , h ^ t c = t a n h ( W c · [ r t c f t c , z ˜ t c , ψ ] + b c ) , h t c = ( 1 u t c ) h t 1 c + u t c h ^ t c ,
where W r c , W u c , W c , b r c , b u c , and b c are trainable parameters. f t c is a coarse-grained vector generated by the aforementioned attention mechanism, and r t c is a coarse-grained reset gate. The coarse-grained feature vector h t c is obtained by updating u t c .
Fine-grained Graph Convolutional Network. In Section 3.2, we present a batch-specific (task-specific) inter-variable dependency structure A. However, the fluctuation nature of the time-series data makes the inter-variable relationship change slightly at each timestamp, leading to diversity in temporal dynamics. In order to capture the slight changes, we calculate the cosine similarity between different variables at each timestamp. And the adjacency matrix is controlled within [ 0 , 1 ] through the ReLU activation function:
A i j t = 1 R e L U ( 1 R e L U ( A i j + cos ( z ˜ i f , ψ , z ˜ j f , ψ ) / β ) ) ,
where c o s ( ) is the cosine similarity function, and β is the perturbation parameter, which controls the perturbation strength. The slightly adjusted A t can be regarded as the fine-grained adjacency matrix at the tth timestamp, which is fed into the F-GCN to learn the fine-grained variable relationships at each timestamp and finally obtain the updated fine-grained feature vectors h t f :
h t f = σ ( D ˜ t 1 2 A ˜ t D ˜ t 1 2 z ˜ t f , ψ W F ) ,
where W F is the trainable parameter, A ˜ t = A t + I is the adjacency matrix of the self-recursive edges added at each timestamp, and D ˜ t is the degree matrix of A ˜ t .
Gated Aggregation Scheme. Based on cross-view position-aware concatenation (i.e., the concatenation of the query set’s conditioning range data and task-specific contextual information extracted from the support set by the MV-PAN), the F-GCN and C-GRU capture temporal dynamics from two views, respectively. Then, we use a gated aggregation scheme to achieve appropriate information selection for generative predictions, thus generating the final prediction h t :
g f = σ ( W c 1 h t c + W f 1 h t f + b 1 ) , g c = σ ( W c 2 h t c + W f 2 h t f + b 2 ) , h t = g f h t f + g c h t c ,
where W c 1 , W 2 c , W f 1 , W 2 f , b 1 , and b 2 are trainable parameters, σ is the sigmoid activation function, and g f and g c are gating units. Finally, the vector h t R F e realizes the prediction of y ^ through a trainable parameter W y R 1 × F e , i.e., y ^ t = W y h t .

3.3.4. End-to-End Stochastic Training

The training process learns shared parameters θ (i.e., the parameters of the MRI, MFE, and CV-CAG) across all tasks and parameters ϕ of the amortization networks (i.e., the parameters of the MV-PAN) through the loss optimization of the query set in different tasks. Given a new task, our model takes few-shot training datasets as inputs and then outputs a distribution over task-specific stochastic inputs ψ in a single forward pass, realizing fast and versatile learning and inference.
We layout the specific loss function for end-to-end stochastic training. First, ST-MeLaPI realizes learning via minimizing the expected value of KL divergence (between the approximate posterior predictive distribution and true distribution of the query set) averaged over the tasks:
ϕ = a r g m i n ϕ E p ( D ) [ K L [ p ( y ˜ | D ) | | q ϕ ( y ˜ | D ) ] ] = a r g m a x ϕ E p ( y ˜ | D ) [ log p ( y ˜ | D , ψ s , ψ d ) q ϕ s ( ψ s | D c ) q ϕ d ( ψ d | D f ) d ψ s d ψ d ] .
After introducing inputs Z ˜ and shared parameters θ , we have
L ( ϕ ) = E p ( D , Z ˜ , y ˜ ) [ log q ϕ ( y ˜ | Z ˜ , θ ) ] = E p ( D , Z ˜ , y ˜ ) [ log p ( y ˜ | Z ˜ , ψ s , ψ d , θ ) q ϕ s ( ψ s | D c , θ ) q ϕ d ( ψ d | D f , θ ) d ψ s d ψ d ] .
Finally, we leverage L s / d Monte Carlo samples to approximate the expectations of ψ s and ψ d , formalizing the final end-to-end stochastic training loss function:
L ( θ , ϕ ) = 1 M m = 1 M log 1 L s L d l s = 1 L s l d = 1 L d p ( y ˜ m | Z ˜ m , ψ s , ψ d , θ ) .

4. Experimental Settings

4.1. Datasets

We conducted extensive experiments on three traffic benchmark datasets.
METR-LA: This is a traffic dataset collected by 207 loop detectors on highways in Los Angeles [46]. It contains data recorded every five minutes for 4 months between March 2012 and June 2012. The training/validation/testing data ratio was 7/1/2.
PEMS-BAY: This is a traffic dataset collected by the California Transportation Agency Performance Measurement System. In total, 325 sensors recorded data in aggregated 5 min intervals for 6 months. The training/validation/testing data ratio was 7/1/2.
EXPY-TKY: The EXPY-TKY dataset includes traffic speed information and corresponding accident reports for 1843 highways in Tokyo over three months (October to December 2021), with data collected every ten minutes. Consistent with the MegaCRN, we utilized the first two months (October and November 2021) as the training and validation datasets and the last month (December 2021) as the test dataset. Table 1 summarizes the specific spatio-temporal details of these datasets.

4.2. Experimental Details

4.2.1. Baselines

Given the intrinsic dependence of traffic forecasting tasks on graph-structured relationships, our selection of baselines extends beyond the basic time series forecasting algorithm, Historical Average (HA), to include a focused comparison with recent graph-based time series forecasting methods. This includes representative spatio-temporal fusion algorithms such as the STGCN [38], DCRNN [39], and GW-Net [47]; Transformer-based forecasting algorithms like the STTN [48] and GMAN [49]; and models built upon the GCRN [50], including the CCRNN [51], GTS [9], and PM-MemNet [52]. Furthermore, we examined approaches based on meta-learning, specifically ST-GFSL [19] and the meta-graph-based MegaCRN [53].

4.2.2. Hyperparameter Tuning

We employed the ADAM optimizer with a learning rate set to 0.0025 and a batch size of 16. We set the dimensionality of all feature vectors to 64. To ensure a certain degree of relevance within each batch, while simultaneously avoiding the premature exposure of test set labels through the training set in meta-learning, we randomized all samples in a specific pattern. For instance, when predicting future consecutive time points, samples were shuffled every 12 points, and after shuffling, every 16 samples were grouped into a batch, with the first four samples serving as the support set. On the METR-LA and PEMS-BAY traffic datasets, the model input length was fixed at 12. In the C-GRU, we set the coarse-grained receptive field to 10 and filled missing data with zeros. The fusion function, Ψ ( ) , in this study was implemented using a concatenation operation. A generative forecasting approach was utilized; for example, in the METR-LA dataset, based on 12 consecutive time points, we predicted the next 6 consecutive time points by appending six zeros after the 12 time points. The positions of the appended zeros in the model’s final output represent our forecast results. We employed the MAE, RMSE, and MAPE as evaluation metrics. All models were trained/tested on an NVIDIA Tesla V100 32 G GPU.

4.2.3. Evaluation Metrics

For the three traffic forecasting datasets, METR-LA, PEMS-BAY, and EXPY-TKY, we used three evaluation metrics, including the MAE, RMSE, and MAPE. The MAE, RMSE, and MAPE are calculated as follows:
M A E = 1 T y i = 1 T y ( y i y ^ i ) , R M S E = 1 T y i = 1 T y ( y i y ^ i ) 2 , M A P E = 100 % T y i = 1 T y y i y ^ i y i .

4.3. Meta-Task Construction

Sequential Construction. An intuitive strategy is to sequentially split the data into different batches (tasks). In each task, the support set is formed by historical data, and the query set is formed by newly accumulated data, which is consistent with the training and inference requirement in time series prediction tasks. Note that there is no time-window overlap between the support set and query set in case of the information leakage problem. Specifically, based on the grid search method in hyperparameter learning, we set the batch size to 16, including 4 shots for the support set and 12 shots for query set. This was the default setting in our learning process.
Generative Construction. One drawback of the sequential construction strategy is the large time interval between the support set and the query set in the long-sequence prediction task. As a result, it easily generates inconsistent (even non-stationary) distributions, leading to a negative knowledge transfer issue. For example, in the task of predicting 12 future time points on the METR-LA and PEMS-BAY datasets, the interval between the support set and the query set will reach at least 60 min under our split method. In addition, the large time interval cannot be utilized as the conditioning window, thus failing to fully leverage all the historical data. To this end, we propose a generative strategy. An off-the-shelf time series prediction model can be pretrained to generate prediction window data to replace the original data. Hence, each conditioning window data can generate its own prediction window data without access to real data, thus bringing consistency between the support set and query set.

5. Results and Analysis

5.1. Traffic Flow Forecast

In the datasets for traffic flow prediction tasks, a large number of sensor nodes constitute a highly complex spatial network. This necessitates that ST-MeLaPI possesses strong capabilities for learning and capturing spatial features. Our model constructs its spatial structure from both macroscopic and microscopic perspectives to capture a richer set of structural features. As observed in Table 2, graph-structured algorithms significantly outperform traditional time series forecasting algorithms like the HA, validating the role of graph structures in multivariate time series prediction. Compared to the current advanced algorithm MegaCRN, on the METR-LA dataset, ST-MeLaPI achieved an average reduction of 16.3 % in the MAE, 20.8 % in the MAPE, and 32.5 % in the RMSE. On the PEMS-BAY dataset, it showed an average reduction of 20.7 % in the MAE, 27.4 % in the MAPE, and 39.0 % in the RMSE. On the EXPY-TKY dataset, there was an average decrease of 7.6 % in the MAE, 5.9 % in the MAPE, and 6.3 % in the RMSE.
As a spatio-temporal meta-learning probabilistic inference method, ST-MeLaPI demonstrates rapid adaptability to complex and fluctuating multivariate time series forecasting scenarios. Compared with the latest meta-graph-based MegaCRN and the spatio-temporal meta-learning method ST-GFSL, ST-MeLaPI exhibits a superior performance. Utilizing the VERSA framework, ST-MeLaPI generates task-specific meta-parameters, circumventing the time-consuming second-order gradient computations found in methods like MAML, thereby enhancing the efficiency of meta-learning.

5.2. Ablation Studies

To further assess the effectiveness of each component within the model, we conducted ablation experiments on the PEMS-BAY dataset. The final results of these experiments are presented in Table 3.
Specifically, we first removed the MRI component from ST-MeLaPI (denoted as ST-MeLaPIMRI) to validate whether the adjacency matrices captured via MRI effectively enhance the model performance in multivariate time series forecasting tasks. In ST-MeLaPIMRI, a fully connected adjacency matrix was utilized as a substitute for the adjacency matrix generated via MRI. The results indicate that the dynamically generated adjacency matrices by MRI indeed improve prediction performance, especially evident within the traffic dataset (PEMS-BAY). Thus, the dynamic learning of inter-variable relationships is crucial in multivariate spatio-temporal dynamic analysis.
Subsequently, we removed structures related to the GCN (i.e., C-GCN and F-GCN) and GRU (i.e., C-GRU and F-GRU) from ST-MeLaPI to verify the necessity of multi-perspective learning. From a temporal perspective (ST-MeLaPIGCN), the task-specific random input ψ d generated in the F-GRU from the support set was concatenated with the query set (control window data) and then inputted into the C-GRU for the final prediction. Similarly, from a spatial perspective (ST-MeLaPIGRU−), the task-specific random input ψ s generated by the C-GCN was added to the query set before being inputted into the F-GCN for the final prediction outcome. The significant performance decrease in single-view scenarios underscores the importance of considering both perspectives in the meta-learning framework for multivariate time series modeling.
Next, we removed MV-PAN from ST-MeLaPI (ST-MeLaPIMVPAN) to validate the effectiveness of the position-aware meta-learning strategy. As Table 3 shows, ST-MeLaPIMVPAN performs even better than ST-MeLaPI in short-term forecasting (15 min). However, as the forecasting horizon extends (30 min, 60 min), the performance of ST-MeLaPIMVPAN deteriorates, highlighting the advantages of ST-MeLaPI. We believe that longer sequence prediction tasks have more similar patterns that can be captured through the position-aware meta-learning strategy, demonstrating its necessity and effectiveness for long-sequence prediction tasks.
Finally, we removed the F-GCN and C-GRU from the CV-GAG in ST-MeLaPI to test the necessity of cross-view modeling. In the model withoutthe F-GCN (ST-MeLaPIFGCN), the final outcome was generated by the C-GRU. In the model without the C-GRU (ST-MeLaPICGRU), the final outcome was generated by the F-GCN. As shown in Table 3, the cross-view information fusion strategy indeed significantly improves the performance, indicating that independent views lead to a decrease in information utilization.
Through comprehensive ablation experiments, we found that removing the GRU-related components used for temporal encoding significantly degraded the model performance. This result demonstrates the crucial role of capturing temporal features in traffic flow prediction tasks. Additionally, in a lateral comparison analyzing the impact of adjacency relationship modeling and graph convolution on the model, we observed that removing the MRI structure and using a fully connected matrix introduced more noise, significantly reducing model performance. This indicates that structured modeling is vital in reducing noise and enhancing model performance.

5.3. Few-Data Training

We conducted extensive experiments on the METR-LA dataset to validate our model’s learning capabilities in a low-resource learning environment. Specifically, we trained the model using only 30 % , 20 % , 10 % , and 5 % of the original training data. The final results are illustrated in Figure 5. For instance, ST-MeLaPI-10% indicates that only 10 % of the original training data were utilized for multi-task meta-learning. We observed that ST-MeLaPI maintained a stable performance despite the reduction in the training data volume, and in many cases, models trained on a limited amount of data even outperformed those trained on the full dataset. Future research could focus on selecting the most informative time spans for training.

5.4. Dynamic Graph Generation

To explore the model’s capability in modeling multivariate relationships, this study employed a segment of multivariate time series data selected from the METR-LA dataset. Through the MRI module, a corresponding adjacency matrix was generated for in-depth visual analysis. Figure 6 displays the interconnectivity among 10 sensors captured during the learning process of the model. To further validate the practical utility of these adjacency relationships, we simulated a traffic accident scenario. Specifically, we mimicked the occurrence of a traffic accident by reducing the traffic flow at sensor 2 to observe its impact on the adjacency relationships. The analysis results clearly show that when traffic flow at an intersection decreases, the connection strength with other intersections weakens accordingly, thereby corroborating the effectiveness and accuracy of our model in understanding and depicting complex multivariate relationships.

5.5. Task Relevance Verification

The meta-learning paradigm assumes that tasks are both inherently related and stationary such that inductive transfer can improve sample efficiency and generalization. Otherwise, it easily leads to the negative knowledge transfer issue and catastrophic forgetting issue. Therefore, in this work, we split the sequential training data into different tasks, where each task spanned short time periods. As such, this generates a large number of meta-training tasks, bringing diversity as well as inherent relations. Since task-specific stochastic inputs (or parameters in a classification task) are generated based on support set for each task in ST-MeLaPI, we can compare the mean and variance between two tasks for a task relationship evaluation. In effect, we conducted the evaluation by analyzing the mean and variance of the spatial dependency-induced and temporal pattern-induced stochastic inputs for two consecutive learning tasks in the METR-LA dataset, respectively. As shown in Figure 7, the means and variances at each timestamp show strong similarities between the two consecutive batches and have some similarities. This provides insights for the effectiveness of the meta-learning framework.

6. Conclusions

In this study, we studied the multivariate time series prediction problem and proposed ST-MeLaPI to realize the efficient and versatile learning of complex temporal dynamics. Specifically, we designed a multiview position-aware spatial-temporal meta-learning probabilistic inference strategy, which essentially utilizes meta-learning to capture both shared and task-specific temporal dynamics. In particular, the multiview meta-learning of pairwise variable relationship-induced information and temporal pattern-induced view-specific information highlights the importance of sophisticated consideration for temporal dynamics-only and graph structure-only information. In addition, the proposed learning method realizes fast learning with forward passes trough inference networks without second-derivative computations during training, while harvesting diversity via stochastic neurons in each timestamp. The experiments on real-world data demonstrate the advantages of the model in multivariate time series prediction tasks.

Author Contributions

Conceptualization, L.Z.; methodology, L.Z. and J.Z.; software, J.Z.; validation, B.J.; formal analysis, L.Z. and X.W.; investigation, J.Z.; data curation, J.Z.; writing—original draft preparation, L.Z. and J.Z.; writing—review and editing, B.J. and X.W.; visualization, J.Z.; supervision, B.J. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (No. 62202085, No. 61772110), and the Fundamental Research Funds for the Central Universities (No. DUT23RC(3)063).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study come from public resources.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2021; Volume 35, pp. 11106–11115. [Google Scholar]
  2. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  3. Pan, Z.; Zhang, W.; Liang, Y.; Zhang, W.; Yu, Y.; Zhang, J.; Zheng, Y. Spatio-temporal meta learning for urban traffic prediction. IEEE Trans. Knowl. Data Eng. 2020, 34, 1462–1476. [Google Scholar] [CrossRef]
  4. Han, L.; Du, B.; Sun, L.; Fu, Y.; Lv, Y.; Xiong, H. Dynamic and multi-faceted spatio-temporal deep learning for traffic speed forecasting. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–19 August 2021; pp. 547–555. [Google Scholar]
  5. Cao, D.; Wang, Y.; Duan, J.; Zhang, C.; Zhu, X.; Huang, C.; Tong, Y.; Xu, B.; Bai, J.; Tong, J.; et al. Spectral temporal graph neural network for multivariate time-series forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17766–17778. [Google Scholar]
  6. Qin, H.; Ke, S.; Yang, X.; Xu, H.; Zhan, X.; Zheng, Y. Robust spatio-temporal purchase prediction via deep meta learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2021; Volume 35, pp. 4312–4319. [Google Scholar]
  7. An, Y.; Zhang, L.; Yang, H.; Sun, L.; Jin, B.; Liu, C.; Yu, R.; Wei, X. Prediction of treatment medicines with dual adaptive sequential networks. IEEE Trans. Knowl. Data Eng. 2021, 34, 5496–5509. [Google Scholar] [CrossRef]
  8. Kipf, T.; Fetaya, E.; Wang, K.C.; Welling, M.; Zemel, R. Neural relational inference for interacting systems. In Proceedings of the International Conference on Machine Learning—PMLR, Stockholm Sweden, 10–15 July 2018; pp. 2688–2697. [Google Scholar]
  9. Shang, C.; Chen, J.; Bi, J. Discrete graph structure learning for forecasting multiple time series. arXiv 2021, arXiv:2101.06861. [Google Scholar]
  10. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-gcn: A temporal graph convolutional network for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3848–3858. [Google Scholar] [CrossRef]
  11. Salinas, D.; Flunkert, V.; Gasthaus, J.; Januschowski, T. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
  12. Choi, E.; Xu, Z.; Li, Y.; Dusenberry, M.; Flores, G.; Xue, E.; Dai, A. Learning the graphical structure of electronic health records with graph convolutional transformer. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 606–613. [Google Scholar]
  13. Qu, Y.; Liu, C.; Zhang, K.; Xiao, K.; Jin, B.; Xiong, H. Diagnostic sparse connectivity networks with regularization template. IEEE Trans. Knowl. Data Eng. 2021, 35, 307–320. [Google Scholar] [CrossRef]
  14. Li, Y.; Chen, Z.; Zha, D.; Du, M.; Ni, J.; Zhang, D.; Chen, H.; Hu, X. Towards Learning Disentangled Representations for Time Series. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 3270–3278. [Google Scholar]
  15. Eldele, E.; Ragab, M.; Chen, Z.; Wu, M.; Kwoh, C.K.; Li, X.; Guan, C. Time-Series Representation Learning via Temporal and Contextual Contrasting. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, Montreal, QC, Canada, 19–27 August 2021; pp. 2352–2359. [Google Scholar]
  16. Pham, Q.; Liu, C.; Sahoo, D.; Hoi, S.C. Learning Fast and Slow for Online Time Series Forecasting. arXiv 2022, arXiv:2202.11672. [Google Scholar]
  17. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning—PMLR, Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar]
  18. Oreshkin, B.N.; Carpov, D.; Chapados, N.; Bengio, Y. Meta-learning framework with applications to zero-shot time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2021; Volume 35, pp. 9242–9250. [Google Scholar]
  19. Lu, B.; Gan, X.; Zhang, W.; Yao, H.; Fu, L.; Wang, X. Spatio-Temporal Graph Few-Shot Learning with Cross-City Knowledge Transfer. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 1162–1172. [Google Scholar]
  20. Sohn, K.; Lee, H.; Yan, X. Learning structured output representation using deep conditional generative models. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar]
  21. Zhang, Y.; Li, Y.; Zhou, X.; Luo, J.; Zhang, Z.L. Urban traffic dynamics prediction—A continuous spatial-temporal meta-learning approach. ACM Trans. Intell. Syst. Technol. 2022, 13, 1–19. [Google Scholar] [CrossRef]
  22. Gordon, J.; Bronskill, J.; Bauer, M.; Nowozin, S.; Turner, R. Meta-Learning Probabilistic Inference for Prediction. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  23. Makridakis, S.; Hibon, M. ARMA models and the Box–Jenkins methodology. J. Forecast. 1997, 16, 147–163. [Google Scholar] [CrossRef]
  24. de Bézenac, E.; Rangapuram, S.S.; Benidis, K.; Bohlke-Schneider, M.; Kurle, R.; Stella, L.; Hasson, H.; Gallinari, P.; Januschowski, T. Normalizing kalman filters for multivariate time series analysis. Adv. Neural Inf. Process. Syst. 2020, 33, 2995–3007. [Google Scholar]
  25. Hyndman, R.; Koehler, A.B.; Ord, J.K.; Snyder, R.D. Forecasting with Exponential Smoothing: The State Space Approach; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  26. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to Sequence Learning with Neural Networks. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  27. Lai, G.; Chang, W.C.; Yang, Y.; Liu, H. Modeling long-and short-term temporal patterns with deep neural networks. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 95–104. [Google Scholar]
  28. Li, S.; Jin, X.; Xuan, Y.; Zhou, X.; Chen, W.; Wang, Y.X.; Yan, X. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  29. Kitaev, N.; Kaiser, L.; Levskaya, A. Reformer: The Efficient Transformer. In Proceedings of the 8th International Conference on Learning Representations, ICLR, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  30. Madhusudhanan, K.; Burchert, J.; Duong-Trung, N.; Born, S.; Schmidt-Thieme, L. Yformer: U-Net Inspired Transformer Architecture for Far Horizon Time Series Forecasting. arXiv 2021, arXiv:2110.08255. [Google Scholar]
  31. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  32. Oord, A.v.d.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  33. Sen, R.; Yu, H.F.; Dhillon, I.S. Think globally, act locally: A deep neural network approach to high-dimensional time series forecasting. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar]
  34. Oreshkin, B.N.; Carpov, D.; Chapados, N.; Bengio, Y. N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. In Proceedings of the 8th International Conference on Learning Representations, ICLR, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  35. Woo, G.; Liu, C.; Sahoo, D.; Kumar, A.; Hoi, S. CoST: Contrastive Learning of Disentangled Seasonal-Trend Representations for Time Series Forecasting. arXiv 2022, arXiv:2202.01575. [Google Scholar]
  36. Challu, C.; Olivares, K.G.; Oreshkin, B.N.; Garza, F.; Mergenthaler, M.; Dubrawski, A. N-hits: Neural hierarchical interpolation for time series forecasting. arXiv 2022, arXiv:2201.12886. [Google Scholar]
  37. Du, Y.; Wang, J.; Feng, W.; Pan, S.; Qin, T.; Xu, R.; Wang, C. Adarnn: Adaptive learning and forecasting of time series. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Gold Coast, QLD, Australia, 1–5 November 2021; pp. 402–411. [Google Scholar]
  38. Yu, B.; Yin, H.; Zhu, Z. Spatio-Temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  39. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting. In Proceedings of the International Conference on Learning Representations (ICLR’18), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  40. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph WaveNet for Deep Spatial-Temporal Graph Modeling. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence IJCAI-19, Macao, China, 10–16 August 2019. [Google Scholar]
  41. Li, M.; Zhu, Z. Spatial-temporal fusion graph neural networks for traffic flow forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2021; Volume 35, pp. 4189–4196. [Google Scholar]
  42. Jin, Y.; Chen, K.; Yang, Q. Selective cross-city transfer learning for traffic prediction via source city region re-weighting. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 731–741. [Google Scholar]
  43. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful are Graph Neural Networks? In Proceedings of the 7th International Conference on Learning Representations, ICLR, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  44. Duan, Z.; Xu, H.; Wang, Y.; Huang, Y.; Ren, A.; Xu, Z.; Sun, Y.; Wang, W. Multivariate time-series classification with hierarchical variational graph pooling. Neural Netw. 2022, 154, 481–490. [Google Scholar] [CrossRef] [PubMed]
  45. Jang, E.; Gu, S.; Poole, B. Categorical Reparameterization with Gumbel-Softmax. In Proceedings of the 5th International Conference on Learning Representations, ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  46. Jagadish, H.V.; Gehrke, J.; Labrinidis, A.; Papakonstantinou, Y.; Patel, J.M.; Ramakrishnan, R.; Shahabi, C. Big Data and Its Technical Challenges. Commun. ACM 2014, 57, 86–94. [Google Scholar] [CrossRef]
  47. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph wavenet for deep spatial-temporal graph modeling. arXiv 2019, arXiv:1906.00121. [Google Scholar]
  48. Xu, M.; Dai, W.; Liu, C.; Gao, X.; Lin, W.; Qi, G.J.; Xiong, H. Spatial-temporal transformer networks for traffic flow forecasting. arXiv 2020, arXiv:2001.02908. [Google Scholar]
  49. Zheng, C.; Fan, X.; Wang, C.; Qi, J. Gman: A graph multi-attention network for traffic prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1234–1241. [Google Scholar]
  50. Seo, Y.; Defferrard, M.; Vandergheynst, P.; Bresson, X. Structured sequence modeling with graph convolutional recurrent networks. In Proceedings of the Neural Information Processing: 25th International Conference, ICONIP 2018, Proceedings, Part I 25, Siem Reap, Cambodia, 13–16 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 362–373. [Google Scholar]
  51. Ye, J.; Sun, L.; Du, B.; Fu, Y.; Xiong, H. Coupled layer-wise graph convolution for transportation demand prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2021; Volume 35, pp. 4617–4625. [Google Scholar]
  52. Lee, H.; Jin, S.; Chu, H.; Lim, H.; Ko, S. Learning to remember patterns: Pattern matching memory networks for traffic forecasting. arXiv 2021, arXiv:2110.10380. [Google Scholar]
  53. Jiang, R.; Wang, Z.; Yong, J.; Jeph, P.; Chen, Q.; Kobayashi, Y.; Song, X.; Fukushima, S.; Suzumura, T. Spatio-temporal meta-graph learning for traffic forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 8078–8086. [Google Scholar]
Figure 1. General overview of the ST-MeLaPI model framework. (1) Multivariate Relationship Identification (MRI) generates a probabilistic structure matrix through a convolution-based intra-variable temporal dynamic extractor and an inter-variable pairwise relationship predictor (MLP). The Gumbel-Softmax sampling technique is utilized to obtain an inter-variable pairwise relationship matrix A. (2) Multiview Meta-Learning and Probabilistic Inference (MV-MeLaPI). First, a multiview feature encoder (MFE) captures pairwise variable relationship-induced information and temporal pattern-induced view-specific information through a C-GCN and F-GRU, respectively. Note that only conditioning range data are fed into the encoder while prediction range data are replaced with zeros. Second, a multiview position-aware amortization network (MV-PAN) concatenates the view-specific output vectors with the corresponding prediction data of the support set, generating the task-specific stochastic input ψ b s and ψ b d , at each timestamp. Third, a cross-view gated generator (CV-GAG) utilizes a cross-view mechanism that concatenates the view-specific feature vectors (generated through the F-GRU and C-GCN, respectively) of a query set with ψ b s and ψ b d . Then, it further performs spatio-temporal representation learning using the F-GCN and C-GRU. Finally, a gated aggregation scheme (GAS) is used to achieve appropriate information selection for generative predictions.
Figure 1. General overview of the ST-MeLaPI model framework. (1) Multivariate Relationship Identification (MRI) generates a probabilistic structure matrix through a convolution-based intra-variable temporal dynamic extractor and an inter-variable pairwise relationship predictor (MLP). The Gumbel-Softmax sampling technique is utilized to obtain an inter-variable pairwise relationship matrix A. (2) Multiview Meta-Learning and Probabilistic Inference (MV-MeLaPI). First, a multiview feature encoder (MFE) captures pairwise variable relationship-induced information and temporal pattern-induced view-specific information through a C-GCN and F-GRU, respectively. Note that only conditioning range data are fed into the encoder while prediction range data are replaced with zeros. Second, a multiview position-aware amortization network (MV-PAN) concatenates the view-specific output vectors with the corresponding prediction data of the support set, generating the task-specific stochastic input ψ b s and ψ b d , at each timestamp. Third, a cross-view gated generator (CV-GAG) utilizes a cross-view mechanism that concatenates the view-specific feature vectors (generated through the F-GRU and C-GCN, respectively) of a query set with ψ b s and ψ b d . Then, it further performs spatio-temporal representation learning using the F-GCN and C-GRU. Finally, a gated aggregation scheme (GAS) is used to achieve appropriate information selection for generative predictions.
Sensors 24 04473 g001
Figure 2. The structure of F−GRU.
Figure 2. The structure of F−GRU.
Sensors 24 04473 g002
Figure 3. A directed graphical representation of the meta-learning and inference strategy.
Figure 3. A directed graphical representation of the meta-learning and inference strategy.
Sensors 24 04473 g003
Figure 4. The structure of C-GRU.
Figure 4. The structure of C-GRU.
Sensors 24 04473 g004
Figure 5. Model stability comparison using different amounts of training data on METR-LA dataset.
Figure 5. Model stability comparison using different amounts of training data on METR-LA dataset.
Sensors 24 04473 g005
Figure 6. Connectivity pattern changes between sensors when sensor 2 experiences a sudden condition.
Figure 6. Connectivity pattern changes between sensors when sensor 2 experiences a sudden condition.
Sensors 24 04473 g006
Figure 7. Distribution of the means and variances of the task-specific temporal pattern- and spatial dependency-induced stochastic inputs of the support set at each timestamp.
Figure 7. Distribution of the means and variances of the task-specific temporal pattern- and spatial dependency-induced stochastic inputs of the support set at each timestamp.
Sensors 24 04473 g007
Table 1. Statistics of datasets.
Table 1. Statistics of datasets.
DatasetMETR-LAPEMS-BAYEXPY-TKY
Start Time1 March 20121 January 20171 October 2021
End Time30 June 201231 May 201731 December 2021
Time Interval5 min5 min10 min
Timesteps34,27252,11613,248
Spatial Units207 sensors325 sensors1843 road links
Table 2. Performance comparison on three traffic forecasting datasets.
Table 2. Performance comparison on three traffic forecasting datasets.
DatasetsModels15 min/Horizon 330 min/Horizon 660 min/Horizon 12
MAE MAPE (%) RMSE MAE MAPE (%) RMSE MAE MAPE (%) RMSE
METR-LAHA [39]4.1613.07.804.1613.07.804.1613.07.80
STGCN [38]2.887.65.743.479.57.244.5912.79.40
DCRNN [39]2.777.35.383.158.86.453.6010.57.60
GW-Net [47]2.696.95.153.078.46.223.5310.07.37
STTN [48]2.797.25.483.168.56.503.6010.27.60
GMAN [49]2.807.45.553.128.76.493.4410.17.35
CCRNN [51]2.857.55.543.248.96.543.7310.67.65
GTS [9]2.656.85.203.058.36.223.479.87.29
PM-MemNet [52]2.657.05.293.038.46.293.4610.07.29
ST-GFSL [19]2.90N/A5.593.56N/A6.96N/AN/AN/A
MegaCRN [53]2.526.44.942.937.96.063.389.77.23
ST-MeLaPI2.125.33.722.606.74.132.646.94.25
Generate-ST-MeLaPI2.506.34.992.696.65.582.616.74.83
    15 min/horizon 330 min/horizon 660 min/horizon 12
MAEMAPE (%)RMSEMAEMAPE (%)RMSEMAEMAPE (%)RMSE
PEMS-BAY HA [39]2.886.85.592.886.85.592.886.85.59
STGCN [38]1.362.92.961.814.14.272.495.75.69
DCRNN [39]1.382.92.951.743.93.972.074.94.74
GW-Net [47]1.302.72.741.633.63.701.954.64.52
STTN [48]1.362.82.871.673.73.791.954.54.50
GMAN [49]1.352.82.901.653.73.821.924.54.49
CCRNN [51]1.382.92.901.743.93.872.074.84.65
GTS [9]1.342.82.841.673.73.831.984.54.56
PM-MemNet [52]1.342.82.821.653.73.761.954.54.49
ST-GFSL [19]1.56N/A3.182.07N/A4.58N/AN/AN/A
MegaCRN [53]1.282.62.721.603.53.681.884.44.42
ST-MeLaPI1.082.21.811.212.62.241.463.02.44
Generate-ST-MeLaPI0.931.91.631.132.42.091.413.12.83
    10 min/horizon 130 min/horizon 360 min/horizon 6
MAEMAPE (%)RMSEMAEMAPE (%)RMSEMAEMAPE (%)RMSE
EXPY-TKY HA [39]7.6331.211.967.6331.211.967.6331.211.96
STGCN [38]6.0924.89.606.9130.210.998.4132.912.70
DCRNN [39]6.0425.59.446.8531.010.877.4534.611.86
GW-Net [47]5.9125.29.306.5929.710.546.8931.711.07
STTN [48]5.9025.69.276.5329.810.406.9932.511.23
GMAN [49]6.0926.59.496.6430.110.557.0532.911.28
CCRNN [51]5.9024.59.296.6829.910.777.1132.511.56
GTS [9]N/AN/AN/AN/AN/AN/AN/AN/AN/A
PM-MemNet [52]5.9425.19.256.5228.910.426.8731.211.14
ST-GFSL [19]N/AN/AN/AN/AN/AN/AN/AN/AN/A
MegaCRN [53]5.8124.49.206.4428.910.336.8331.011.04
ST-MeLaPI5.0222.38.415.8926.19.456.7731.310.83
Generate-ST-MeLaPI5.3622.98.795.9226.09.666.8831.611.29
Table 3. Contribution of different components to ST-MeLaPI performance on PEMS-BAY dataset.
Table 3. Contribution of different components to ST-MeLaPI performance on PEMS-BAY dataset.
DatasetsModels15 min30 min60 min
MAE MAPE (%) RMSE MAE MAPE (%) RMSE MAE MAPE (%) RMSE
PEMS-BAYST-MeLaPI1.082.21.811.212.62.241.463.02.44
ST-MeLaPI   M R I 1.783.72.872.154.93.553.188.65.41
ST-MeLaPI   G C N 1.282.62.281.733.63.112.325.04.28
ST-MeLaPI   G R U 3.087.34.863.148.15.194.7013.47.26
ST-MeLaPI   M V P A N 0.881.71.731.232.62.511.804.13.52
ST-MeLaPI   C G R U 1.603.62.572.265.13.472.445.63.78
ST-MeLaPI   F G C N 1.422.92.451.523.12.531.613.32.60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Zhu, J.; Jin, B.; Wei, X. Multiview Spatial-Temporal Meta-Learning for Multivariate Time Series Forecasting. Sensors 2024, 24, 4473. https://doi.org/10.3390/s24144473

AMA Style

Zhang L, Zhu J, Jin B, Wei X. Multiview Spatial-Temporal Meta-Learning for Multivariate Time Series Forecasting. Sensors. 2024; 24(14):4473. https://doi.org/10.3390/s24144473

Chicago/Turabian Style

Zhang, Liang, Jianping Zhu, Bo Jin, and Xiaopeng Wei. 2024. "Multiview Spatial-Temporal Meta-Learning for Multivariate Time Series Forecasting" Sensors 24, no. 14: 4473. https://doi.org/10.3390/s24144473

APA Style

Zhang, L., Zhu, J., Jin, B., & Wei, X. (2024). Multiview Spatial-Temporal Meta-Learning for Multivariate Time Series Forecasting. Sensors, 24(14), 4473. https://doi.org/10.3390/s24144473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop