Next Article in Journal
Machine Learning-Based Satellite Routing for SAGIN IoT Networks
Next Article in Special Issue
Time Series Forecasting of Software Vulnerabilities Using Statistical and Deep Learning Models
Previous Article in Journal
Bridging the Gap between Physical and Circuit Analysis for Variability-Aware Microwave Design: Modeling Approaches
Previous Article in Special Issue
Time Series Network Data Enabling Distributed Intelligence—A Holistic IoT Security Platform Solution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Traffic Forecasting Based on Integration of Adaptive Subgraph Reformulation and Spatio-Temporal Deep Learning Model

Shandong Provincial Key Laboratory of Network Based Intelligent Computing, University of Jinan, Jinan 250022, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(6), 861; https://doi.org/10.3390/electronics11060861
Submission received: 1 February 2022 / Revised: 27 February 2022 / Accepted: 3 March 2022 / Published: 9 March 2022

Abstract

:
Traffic forecasting provides the foundational guidance for many typical applications in the smart city management, such as urban traffic control, congestion avoidance, and navigation guidance. Many researchers have focused on the spatio-temporal correlations under fixed topology structure in traffic network to improve the traffic forecasting accuracy. Despite their advantages, the existing approaches are not completely discussed that the association relationship among traffic network nodes are not invariable under different traffic conditions. In this paper, a novel traffic forecasting framework is proposed by integrating the dynamic association of traffic nodes with the spatio-temporal deep learning model. To be specific, an adaptive subgraph reformulation algorithm is designed first based on the specific forecasting interval to reduce the interference of irrelevant spatio-temporal information. After that, by enhancing the attention mechanism with the generative decoder, a spatio-temporal deep learning model with only one forward operation is proposed to avoid the degradation of accuracy in the long-term prediction, in which the spatio-temporal information and the external factors (such as weather and holiday) are fused together to be as an input vector. Based on the reformulated subgraph constructed of traffic nodes with closer spatio-temporal correlation, experiments show that the proposed framework consistently outperforms other GNN (Graph Neural Network)-based state-of-the-art baselines for various forecasting intervals on a real-world dataset.

1. Introduction

As the increasing complexity of urban transportation, many social problems are becoming more and more serious such as air pollution and traffic congestion. As one of the most important infrastructure facilities in modern cities, ITSs (Intelligent Transforation Systems) provide the great conveniences in daily travel [1]. Intervention in advance based on traffic forecasting is regarded as a key for improving the efficiency of transportation, where the traffic condition is mainly detected by sensors installed on roads, subways and buses. However, there are complex spatial dependencies among sensors distributed on a road network, even the nearest two sensors in Euclidean space. Meanwhile, mass traffic data are often with high dimension and affected by emergencies, such as traffic accident and holidays. Therefore, the major challenge for traffic forecasting is that the spatial–temporal correlation must be captured simultaneously to improve the traffic forecasting accuracy.
The early methods of traffic forecasting are usually based on historical traffic data. Traditional statistical analysis methods with regression functions are adopted to analyze the historical traffic data, such as linear regression prediction method and ARIMA (Auto-Regressive Integrated Moving Average) method [2]. The most statistical analysis methods are simple and easy to be implemented under the assumption that the traffic time-series data is linearly modeled. However, due to the high nonlinearity of traffic data, statistical analysis models are not completely capture the changing trends of traffic flow in the next period. Meanwhile, the traditional machine learning methods rely on manually extracted traffic features under the simple traffic conditions or small volume of traffic data. Recently, deep learning methods have been widely used in various complex prediction tasks to improve the forecasting accuracy, such as feedforward neural networks [3], deep belief networks [4] and RNN (Recurrent Neural Networks) [5]. Specially, ref. [6] uses LSTM (Long Short-Term Memory) network for traffic speed forecasting; [7] combines LSTM and Gaussian process regression to forecast the traffic load of network. However, the above mentioned methods are only adopted to study the temporal characteristics of mass traffic data.
Inspired by image feature extraction, CNNs (Convolutional Neural Networks) are used to extract spatial features of road network [8,9], in which the traffic network and the irregular road structure are converted into the image grid data with fixed adjacency relationship. However, the spatial correlation among traffic nodes will be destroyed. To overcome the shortcomings of CNNs, GNNs (Graph Neural Networks) are employed to predict the traffic flow on non-grid road networks [10,11,12,13], in which a non-Euclidean structure is used to capture the spatial correlation based on the adjacency matrix. Recently, various enhanced GNNs have been designed for traffic forecasting under the complex, nonlinear and spatio-temporal correlated traffic networks [14,15,16,17,18]. Specially, ref. [14] proposes an optimized graph convolutional recurrent neural network to learn an optimized graph in a data-driven manner and to reveal the potential relationships between the road segments with the traffic data. Ref. [15] introduces an attention mechanism based on DCRNN (Diffusion Convolutional Recurrent Neural Network) to obtain the adaptive weight matrix of nodes, in which the weight matrix assigned to the nodes by graph convolution is fixed. Those studies have fully demonstrated the effectiveness of GNN-based models to reveal the spatio-temporal correlation on complex traffic network. However, their common limitations have been recognized as follows:
(1)
No matter what the forecasting intervals and the traffic condition, most existing models often require the extraction of spatial features on the entire traffic network with fixed nodes and traffic network topology. However, the association relationships among traffic nodes are not invariable.
(2)
The traffic forecasting methods based on deep learning models are with the forms of step by step, such as LSTM, Seq2Seq (Sequence to Sequence) model, and Transformer. While dealing with long sequence of traffic data, the prediction accuracy will deteriorate caused by the aggravated propagation of accumulated errors.
Motivated by the above discussion, by integrating the dynamic association of traffic nodes with the enhanced attention mechanism, a traffic forecasting framework with adaptive subgraph reformulation is proposed to improve forecasting accuracy involving with different forecasting intervals. The main contributions of this research are as follows:
(1)
An algorithm of adaptive subgraph reformulation is proposed to reduce the impact of irrelevant spatiotemporal information on the forecast accuracy, in which the traffic nodes and network topology in traffic network are selected and reconstructed adaptively based on the reachability analysis and similarity quantization among traffic nodes within a specific forecasting interval.
(2)
A spatio-temporal deep learning model with self-attention mechanism is designed to avoid the cumulative error propagation caused by the long sequence of traffic data. The fusion of the time-series information, the spatial information and the time-stamp information sets as the input, and all forecasting results are output by performing one forward operation based on the generative decoder.
The framework of this paper is as follows: In Section 2, the proposed traffic forecasting framework is presented. The corresponding experimental results are given in Section 3. This paper will end with the conclusion and future work in Section 4.

2. Methodology

The traffic forecasting framework with adaptive subgraph reformulation is shown in Figure 1, which contains two main components: the adaptive subgraph reformulation algorithm and the spatio-temporal deep learning model. In adaptive subgraph reformulation module, the number of nodes and network topology in traffic network are constructed based on the the forecasting interval. Based on the reformulated network topology, the spatio-temporal deep learning model is intended to fuse the spatio-temporal data with the external factors and then use the encoder–decoder to predict the traffic flow.

2.1. Adaptive Subgraph Reformulation Algorithm

Different from the pre-defined static graph adapted in GCN model, an effective subgraph is reformulated to reduce the interference of irrelevant spatio-temporal information by using the designed adaptive subgraph reformulation algorithm. First, the adjacent matrix of the maximum traveling time from the original graph of traffic network is constructed. Then, a temporary connected subgraph is obtained by using the shortest-path algorithm, in which the nodes are selected based on the reachability analysis and similarity quantization in the specific forecasting interval. Finally, by employing the spectral clustering on the Laplacian matrix of the temporary connected subgraph, the target subgraph is obtained by remove nodes with low spatio-temporal similarity. The specific process of adaptive subgraph reformulation is shown in Figure 2.
Leong [19] observes that the speed distribution of vehicles on the road obeys a normal distribution. He has obtained the results after using radar speedometers to measure the traffic speed of smooth flow at 31 observation points in Australia during a three-year period. Katti and Raghavachari [20] have found that the speeds of cars, heavy commercial vehicles, light commercial vehicles and motorcycles, all follow the normal distribution. Kumar and Rao [21] have also demonstrated that the vehicle speed obeys a normal distribution.
First, under the assumption that the vehicle speed satisfies the normal distribution N ( μ , σ 2 ) , the minimum speeds among different traffic nodes could be computed. In details, the average vehicle speed during the latest t hours is brought into the normal distribution as μ . 120 % of the maximum speed limit is used as the upper bound of the 95 % of the confidence interval ( μ + 2 σ ) . The lower confidence bound is obtained as the minimum speed on the road by ( μ 2 σ ) . Thus, the distance matrix D and minimum speed matrix S are constructed from the traffic dataset:
D = 0 d 12 d 1 n d 21 0 d 2 n d n 1 d n 2 0 , S = 0 s 12 s 1 n s 21 0 s 2 n s n 1 s n 2 0 ,
where n is the number of nodes in the overall traffic network; d i j in D represents the actual distance from node i to node j; s i j in S represents the minimum speed from node i to node j. The maximum traveling time matrix T is obtained by dividing the corresponding elements of the D and S matrices, which is described as:
T = 0 t 12 t 1 n t 21 0 t 2 n t n 1 t n 2 0 = 0 d 12 / s 12 d 1 n / s 1 n d 21 / s 21 0 d 2 n / s 2 n d n 1 / s n 1 d n 2 / s n 2 0 ,
where t i j represents the maximum traveling time from node i to node j.
After that, by employing the shortest-path algorithm of Bellman–Ford method in [22], the traveling time T p a s s i from the ith node to the target forecasting node could be computed. The nodes with the traveling time shorter than the forecasting interval T f will be selected to construct a temporary connected subgraph. Obviously, those dropped nods will not affect the forecasting target node within the specific forecasting interval. The adjacent matrix G T of the designed temporary connected subgraph will be constructed based on the selected nodes and the corresponding distances.
By applying the spectral clustering on the Laplacian matrix of the adjacent matrix G T of the reformulated temporary connected subgraph, the similarity between the ith node to the target forecasting node could be quantized as H s i . Filtrating the nodes with less similarity quantitative value H s i than a reasonable threshold H, the reformulated subgraph could be obtained with the adjacent matrix G, which is described as
G = 0 g 12 g 1 m g 21 0 g 2 m g m 1 g m 2 0 ,
where m is the number of nodes in the reformulated subgraph; g i j represents the actual distance from node i to node j. Due to the difference in road connectivity, D is usually an asymmetric matrix.
Based on the above description, the designed adaptive subgraph reformulation Algorithm 1 is given as follows.
Algorithm 1 Adaptive subgraph reformulation algorithm
Require: The actual distances between each traffic nodes; the maximum speed limits in each road segment; the forecasting interval T f ; the threshold of similarity H.
  • Obtain the initial distance matrix D and minimum speed matrix S from the dataset;
  • Calculate the maximum travelling time matrix T;
  • Get the vector T p a s s = [ T p a s s 1 , T p a s s 2 , , T p a s s i , , T p a s s n ] based on Bellman–Ford method, where T p a s s i denotes the shortest traveling time from the ith node to the forecasting target node;
  • Construct the temporary connected subgraph with the adjacent matrix G T by dropping those nodes with T p a s s i > T f from original traffic network;
  • Calculate the similarity vector H s = [ H s 1 , H s 2 , , H s i , , H s k ] among the nodes in the temporary connected subgraph based on the spectral clustering;
  • Construct the reformulated subgraph with adjacent matrix G, where the nodes with H s i < H are dropped in the temporary connected subgraph;
  • Output: The adjacent matrix G of the reformulated subgraph.

2.2. Spatio-Temporal Deep Learning Model

There is an inherent limitation of the encoder–decoder architecture with Transformer in [23] that prevents it from being directly applicable to the time-series forecasting. To address this issue, we have changed the decoder’s structure. The generative style decoder, though conceptually simple, predicts the long time-series sequences in one forward operation rather than in a step-by-step way. The one forward operation will drastically improve the inference speed of the long-sequence predictions.
Referring to the structure of Transfomer model, the proposed spatio-temporal deep learning model consists of three components, including the embedding module used to preprocess the traffic data, the encoder module used to capture the dependencies between input sequences, and the decoder module used to forecast the traffic flow through one forward calculation. What’s following, the designed three components will be described in detailed.

2.2.1. Embedding Module of Traffic Data

The embedding module of traffic data consists of four parts:
  • Scalar: contains the traffic data of each nodes in the reformulated subgraph;
  • Network Topology: contains the spatial characteristics derived from the obtained reformulated subgraph;
  • Position coding: contains the traffic flow data with the position coding of each nodes in the reformulated subgraph;
  • Time Stamp: contains the external information expressed as n-dimensional variable, including the information about year, month, day, hour, minute, week, weather, and holiday.
As shown in Figure 3, the embedding of traffic data expresses as a 512-dimensional vector converted from the above four parts, which is defined as the input of the proposed spatio-temporal deep learning model.

2.2.2. Self-Attention Mechanism

The attention mechanism is first proposed in the field of natural language processing (NLP) [24,25], and gradually applied to the computer vision [26,27] and the time-series prediction [28,29]. As is well known, Recurrent Neural Networks (RNNs) for prediction with short sequences show good performance, in which all steps are encoded in the sequence and a final output of a sequence embedding is given out. In order to make up the long sequences, the proposed model in [25] uses all the hidden states of the encoder at the decoder end, and gives attention to certain hidden states when decoding each element. Specially, the self-attention mechanism is proposed and supposed to satisfy all requirements for the encode sequences, where each elements attend to every other element in the same sequence or other sequences.
In order to capture the spatio-temporal correlation among different traffic sequences in the reformulated subgraph, the self-attention mechanism and the multi-headed self-attention mechanism are employed and described as follows.
Self-attention mechanism: Based on the same input X, three matrices are designed to realize the self-attention mechanism, including the query of Q = X W Q , the key K = X W K and the value of V = X W V , where W Q , W K and W V are the trained weight matrices, and the superscripts of W Q , W K and W V are the outputs obtained after the operation. In order to stabilize the gradient, attention function is designed by using S o f t m a x function, which is described as
A t t e n t i o n ( Q , K , V ) = S o f t m a x Q K T d V ,
where d denotes the dimension of Q K T .
Multi-headed self-attention mechanism: In order to focus on different features of the traffic information, the deep learning model is divided into multiple heads to form the multiple subspaces. It means that the self-attention operation will be run N times simultaneously, where N is the number of heads. Meanwhile, the outputs of { h e a d 1 , h e a d 2 , , h e a d n } obtained from the multi-head are stacked together. Then, by using the splicing function C o n c a t ( ) , a linear transformation is designed to realize the multi-head attention, which is described as:
M u l t i H e a d ( Q , K , V ) = C o n c a t ( h e a d 1 , , h e a d n ) W O ,
where h e a d i = A t t e n t i o n ( Q W i Q , K W i V , V W i V ) with i = 1 , , N denotes the attention of the ith head, in which W i Q R d m o d e l × d k , W i K R d m o d e l × d k , W i V R d m o d e l × d v , and W O R d N m o d e l × d v are the trained weight matrices; d m o d e l is the dimension of input vector; d k = d v = d m o d e l / N . The structure of multi-headed self-attention mechanism is displayed in Figure 4.

2.2.3. Encoder–Decoder Component

In the proposed spatio-temporal deep learning model, a self-attention mechanism is adopted in the encoder module. A multi-head self-attention is employed in the decoder module to obtain all prediction results by one forward operation. The structure of the encoder–decoder component is shown in Figure 5.
The inputs of the encoder–decoder component is shown in Figure 6.
In order to extract the robust long-term dependency of the long sequential inputs, the input X e n of encoder module is designed as:
X e n = { X e n t n , X e n t n + 1 , , X e n t } ,
where X e n t i R d m o d e l × L x with i = 0 , 1 , , n denotes the ( t i )th sequence input, in which L x is the length of input. Meanwhile, the input of the decoder module is formulated as X i n p u t = { X i n p u t t n , X i n p u t t n + 1 , , X i n p u t t } , where the element X i n p u t t i with i = 0 , 1 , 2 , , n is described as:
X i n p u t t i = C o n c a t ( X d e t i , X 0 t i ) R ( L d e + L o ) × d m o d e l ,
in which X d e t i R L d e × d m o d e l is the start token; X 0 t i R L 0 × d m o d e l is a placeholder for the target sequence; L d e and L o represent the lengths of the start token and the placeholder, respectively.
Encoder module receives the embedding values of traffic data and outputs the Feature Map to the decoder module, which consists of a 3-layer encoder layer and a fully connected layer. More specially, the encoder layer mainly contains two sub-layers: a self-attention layer and a feedforward layer composed of two linear mappings. Both sub-layers are followed by a batch normalization layer. Between the sub-layers, the jump connections are employed to avoid the excessive fitting. Between every two attention layers, the distilling operation is performed to halve the length dimension, which involves with 1-dimensional convolution, E L U activation function, and the maximum pooling. The distilling procedure forwarded from the jth layer into the ( j + 1 )th layer is expressed as:
X j + 1 t = M a x P o o l ( E L U ( C o n v 1 d [ X j t ] ) ) ,
where X j t R d m o d e l × L x represents the tth sequence input in the jth layer; · A B represents the attention layer.
Decoder module employs a structure of generative style decoder [30] to perform one forward step on the time-series and output all the forecasting results, which includes two attention layers and a two-layer feedforward part of linear mapping. In the first attention layer, by multiplying the embedding input with the trained weight matrix of decoder module, the values of Q, K, and V are obtained. Meanwhile, the triangular causal mask mechanism is employed in this layer to ensure that the output decoded at time t depends only on the output before time t. In the second attention layer, the value of Q is obtained by multiplying the output of the first attention layer; the values of K and V are obtained by multiplying the output of encoder.
Based on the above description, the proposed spatio-temporal deep learning model is described in Algorithm 2 in details.
Algorithm 2 Spatio-temporal deep learning model
Require: The dataset of traffic flow data; the adjacent matrix G of the reformulated subgraph.
  • Read the traffic flow data T F , the position information P E , and the time stamps T S from the traffic dataset;
  • Construct the network topology N T from G based on the reformulated subgraph;
  • Obtain the input X e n of the encoder module and X i n p u t of the decoder module by embedding T F , P E , N T , and T S ;
  • Get the feature map F M based on the output of the encoder module;
  • Calculate the output of the decoder module based on X i n p u t and F M ;
  • Return the Output of the decoder module.

3. Experiments

3.1. Dataset

The effectiveness of the proposed traffic forecasting framework is validated on a dataset collected by the Didi Chuxing GAIA Initiative (https://GAIA.didichuxing.com (accessed on 13 March 2021)) in Jinan, China. The employed dataset contains the average speed of vehicles collected from 61 road monitors in Jinan ranging from 1 January 2018 to 31 December 2018 for the experiment. The sample period sets as 10 min. After that, the length and the connection relationship of each traffic node are calculated based on the longitude and latitude coordinates of each traffic node in the traffic dataset.
For data preprocessing of the traffic dataset, the normalization with Z scores is employed first to standardize the traffic data input. The time stamp information of the traffic data set is with the 4-dimensional information, including [month, date, hour, minute]. Meanwhile, the time-stamp information distribution is in the range of [ 0.5 , 0.5 ] . Meanwhile, the dataset is divided into 60 % for training, 20 % for validation, and 20 % for testing. The proposed traffic forecasting framework is trained on the training set and tested on the test set. It is worth noting that we filled in the missing values in the data set with the average speed of one hour before.

3.2. Experimental Settings and Results

Experiments are conducted under the environment with one Intel(R) Core(R) i9-9900K CPU @ 3.60 GHz (Intel, Santa Clara, CA, USA) and a NVIDIA GeForce 2080Ti GPU 11 GB card (NVIDIA, Santa Clara, CA, USA). The batch size is 32 and the training epoch is 100. The head size of the multi-head attention is 8. The Gaussian error linear units is defined as activation function. The learning rate sets as 0.001 . M S E (Mean Squared Error) is defined as the loss function.
Comparison results with seven deep-learning traffic forecasting methods are employed to verify the advancements and effectiveness of the proposed framework under the same traffic dataset, including the HA (Historical Average), ARIMA, LSTM in [31], GRU in [32], GCRN in [33], Gated-STGCN in [34], and GWNET in [35]. To estimate the impact of adaptive subgraph reformulation on the graph, we make a version of our model without structure, i.e., only using spatio-temporal deep learning model, denoted by Ours_WA. To further evaluate the efficiency of the proposed Spatio-Temporal Deep Learning Model, we replace the proposed deep learning model with Transformer, denoted by Ours_WS. Considering the different forecasting intervals with 30 min, 60 min and 120 min, three performance indicators are measured under the above methods and the proposed framework, including M A E (Mean Absolute Error), M A P E (Mean Absolute Percent Error), and R M S E (Root Mean Square Error). More specially, the comparison results of the above three performance indicators are displayed in Table 1, Table 2 and Table 3. Meanwhile, in order to show the effectiveness of the proposed traffic forecasting framework more intuitively, the comparison curves between the forecasting results and the actual values are shown in Figure 7, Figure 8 and Figure 9 under different forecasting intervals with 30 min, 60 min and 120 min, where the x-axis and y-axis represent the time step and the speed value, respectively.
From Table 1, Table 2 and Table 3, it could be noticed that the proposed traffic forecasting framework performs better than other prediction models in both the short-term and the long-term forecasting interval. In particular, as to the long-term traffic forecasting, the values of M A E , R M S E and M A P E are improved more markedly than other GNN-based state-of-the-art baselines in the short-term traffic forecasting. This demonstrates that the proposed traffic forecasting framework can capture the spatio-temporal correlations of urban road networks under different forecasting intervals effectively.
To estimate the effect of adaptive subgraph reformulation and spatio-temporal deep learning model, we design the validation experiments and show the forecasting results of Ours_WA and Ours_WS in Table 1, Table 2 and Table 3. From the results, we can conclude that both of adaptive subgraph reformulation and spatio-temporal deep learning model of the proposed traffic forecasting framework are effective as the results outperform both of Ours_WA and Ours_WS.
Particularly, it is necessary to point out that the original dataset includes 561 traffic nodes. This means that a graph consisting of 561 nodes needs to be constructed for training the traditional GNN-based prediction models. However, while applying the designed adaptive subgraph reformulation algorithm, the nodes in reformulated subgraph are 82, 164, and 220 under different forecasting intervals with 30 min, 60 min, and 120 min, respectively. The number of nodes will be reduced by 85.38%, 70.77%, 60.78%. From Table 1, Table 2 and Table 3 and Figure 7, Figure 8 and Figure 9, with relatively few traffic nodes, the forecasting accuracies still retain a satisfactory level. On the contrary, the proposed traffic forecasting framework with adaptive subgraph reformulation outperforms the employed state-of-the-art traffic forecasting models. This means that, on the premise of not reducing the forecasting accuracy, the adaptive subgraph reformulation algorithm can reduce the amount of input data under the dynamic urban traffic flow effectively.
From the above analysis, under different forecasting intervals, the proposed traffic forecasting framework performs better for improving the forecasting accuracy with less traffic information.

4. Conclusions

In this paper, we proposed a new traffic forecasting framework to improve the forecasting performance on traffic road network under different forecasting intervals, which consisted of the adaptive subgraph reformulation algorithm and the spatio-temporal deep learning model. Based on the accessibility evaluation and the similarity quantification method among nodes in traffic network, the designed adaptive subgraph reformulation algorithm reduced the interference of irrelevant spatio-temporal information, in which the traffic nodes and their links with the weight were adaptively selected to reformulate the traffic network topology. Meanwhile, different from the step-by-step prediction in traditional deep-learning models, the designed spatio-temporal deep learning model with only one forward operation could obtain all the prediction results, where the attention mechanism was enhanced by applying the generative decoder. However, considering the dynamic complexity of the road network and the interference of weather or other factors, more data sources should be introduced in the traffic forecasting in the future.
For future work, we will integrate different GCN (Graph Convolutional Network)-based deep-learning models for fusing the captured features either the short-term traffic forecasting or the long-term traffic forecasting, such as GCN with LSTM for short-term traffic forecasting, GCN with Seq2Seq for middle-term traffic forecasting, and GCN with Informer for long-term traffic forecasting. In addition, the occasion of subgraph reformulation could be discussed with traffic factors, such as traffic congestion and accident.

Author Contributions

Methodology, S.-Y.H.; Supervision, S.-Y.H. and Y.-H.C.; Validation, Q.Z.; Writing—original draft, Q.-W.S.; Writing—review and editing, R.-Z.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grants 61903156 and 61873324, the Natural Science Foundation of Shandong Province for Key Project under Grant ZR2020KF006, the Natural Science Foundation of Shandong Province under Grant ZR2019MF040, the University Innovation Team Project of Jinan under Grant 2019GXRC015, the Higher Educational Science and Technology Program of Jinan City under Grant 2020GXRC057, and the State Scholarship Fund of the China Scholarship Council.

Data Availability Statement

The data presented in this study are openly available at https://GAIA.didichuxing.com (accessed on 13 March 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mori, U.; Mendiburu, A. A review of travel time estimation and forecasting for advanced traveller information systems. Transp. A Transp. Sci. 2015, 11, 119–157. [Google Scholar] [CrossRef]
  2. Chen, L.-W.; Hu, T.-Y. Traffic flow prediction with big data: A deep learning approach. IEEE Trans. Intell. Transp. Syst. 2015, 16, 865–873. [Google Scholar]
  3. Sanger, T.D. Optimal unsupervised learning in a single-layer linear feedforward neural network. Neural Netw. 1989, 2, 459–473. [Google Scholar] [CrossRef]
  4. Huang, W.; Song, G. Deep architecture for traffic flow prediction: Deep belief networks with multitask learning. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2191–2201. [Google Scholar] [CrossRef]
  5. Li, Y.; Yu, R. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. arXiv 2017, arXiv:1707.01926. [Google Scholar]
  6. Ma, X.-L.; Tao, Z.-M. Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp. Res. Part C Emerg. Technol. 2015, 54, 187–197. [Google Scholar] [CrossRef]
  7. Wang, W.; Zhou, C. Cellular traffic load prediction with LSTM and Gaussian process regression. In Proceedings of the 2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020. [Google Scholar]
  8. Jiang, W.; Zhang, L. Geospatial data to images: A deep-learning framework for traffic forecasting. Tsinghua Sci. Technol. 2019, 24, 52–64. [Google Scholar] [CrossRef]
  9. Han, L.; Zheng, K.; Zhao, L. Short-term traffic prediction based on deep cluster in large-scale road networks. IEEE Trans. Veh. Technol. 2019, 68, 12301–12313. [Google Scholar] [CrossRef]
  10. Battaglia, P.W.; Hamrick, J.B.; Bapst, V.; Sanchez-Gonzalez, A.; Zambaldi, V.; Malinowski, M.; Tacchetti, A.; Raposo, D.; Santoro, A.; Faulkner, R.; et al. Relational inductive biases, deep learning, and graph networks. arXiv 2018, arXiv:1806.01261. [Google Scholar]
  11. Scarselli, F.; Gori, M. The graph neural network model. IEEE Trans. Neural Netw. 2009, 20, 61–80. [Google Scholar] [CrossRef] [Green Version]
  12. Wu, Z.-H.; Pan, S.-R. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Zhang, J.-W. Graph neural networks for small graph and Giant network representation learning: An overview. arXiv 2019, arXiv:1908.00187. [Google Scholar]
  14. Guo, K. Optimized graph convolution recurrent neural network for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2019, 22, 1138–1149. [Google Scholar] [CrossRef]
  15. Cirstea, R.G.; Guo, C.-J. Graph attention recurrent neural networks for correlated time series forecasting. arXiv 2021, arXiv:2103.10760. [Google Scholar]
  16. Cui, Z.; Henrickson, K.; Ke, R. Traffic graph convolutional recurrent neural network: A deep learning framework for network-scale traffic learning and forecasting. IEEE Trans. Intell. Transp. Syst. 2020, 21, 4883–4894. [Google Scholar] [CrossRef] [Green Version]
  17. Ke, J.-T.; Feng, S.-Y. Joint predictions of multi-modal ride-hailing demands: A deep multi-task multi-graph learning-based approach. Transp. Res. Part C Emerg. Technol. 2021, 127, 103063. [Google Scholar] [CrossRef]
  18. Mohanty, S.; Pozdnukhov, A. Region-wide congestion prediction and control using deep learning. Transp. Res. Part C Emerg. Technol. 2020, 116, 102624. [Google Scholar] [CrossRef]
  19. Leong, H.J.W. The distribution and trend of free speeds on two-lane rural highways in New South Wales. Aust. Road Res. Board Conf. 1968, 4, 791–814. [Google Scholar]
  20. Katti, B.K.; Raghavachari, S. Modelling of mixed traffic speed data as inputs for the traffic simulation modelsd. Highw. Res. Bull. 1986, 28, 35–48. [Google Scholar]
  21. Kumar, V.M.; Rao, S.K. Headway and speed studies on two-lane highways. Indian Highw. 1998, 26, 23–26. [Google Scholar]
  22. Dreyfus, S.E. Dynamic programming and the calculus of variations. J. Math. Anal. Appl. 1960, 1, 228–239. [Google Scholar] [CrossRef] [Green Version]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017.
  24. Luong, M.-T.; Pham, H. Effective approaches to attention-based neural machine translation. arXiv 2015, arXiv:1508.04025. [Google Scholar]
  25. Bahdanau, D.; Cho, K. Neural machine translation by jointly learning to align and translate. arXiv 2016, arXiv:1409.0473. [Google Scholar]
  26. Hu, J.; Shen, L. Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [Green Version]
  27. Sun, Y.; Fisher, R. Object-based visual attention for computer vision. Artif. Intell. 2003, 146, 77–123. [Google Scholar] [CrossRef] [Green Version]
  28. Mnih, V.; Heess, N. Recurrent models of visual attention. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 27, pp. 2204–2212. [Google Scholar]
  29. Lin, Z.; Li, M. Self-attention convlstm for spatiotemporal prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11531–11538. [Google Scholar]
  30. Zhou, H. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 11106–11115. [Google Scholar]
  31. Cui, Z.; Ke, R.; Pu, Z.; Wang, Y. Deep bidirectional and unidirectional lstm recurrent neural network for network-wide traffic speed prediction. arXiv 2018, arXiv:1801.02143. [Google Scholar]
  32. Agarap, A.F. A neural network architecture combining gated recurrent unit (GRU) and support vector machine (SVM) for intrusion detection in network traffic data. arXiv 2019, arXiv:1709.03082. [Google Scholar]
  33. Seo, Y.; Defferrard, M.; Vandergheynst, P.; Bresson, X. Structured sequence modeling with graph convolutional recurrent networks. In Proceedings of the International Conference on Neural Information Processing, Siem Reap, Cambodia, 13–16 December 2018; pp. 362–373. [Google Scholar]
  34. Yu, B.; Yin, H.-T. Spatio-temporal graph convolutional neural network: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  35. Wu, Z.-H.; Pan, S.-R. Graph waveNet for deep spatial-temporal graph modeling. arXiv 2019, arXiv:1906.00121. [Google Scholar]
Figure 1. The structure of traffic forecasting framework with adaptive subgraph reformulation.
Figure 1. The structure of traffic forecasting framework with adaptive subgraph reformulation.
Electronics 11 00861 g001
Figure 2. The process of the adaptive subgraph reformulation.
Figure 2. The process of the adaptive subgraph reformulation.
Electronics 11 00861 g002
Figure 3. The process of embedding of traffic data.
Figure 3. The process of embedding of traffic data.
Electronics 11 00861 g003
Figure 4. The structure of multi-head attention mechanism.
Figure 4. The structure of multi-head attention mechanism.
Electronics 11 00861 g004
Figure 5. The structure of encoder–decoder component.
Figure 5. The structure of encoder–decoder component.
Electronics 11 00861 g005
Figure 6. The inputs of the encoder–decoder component.
Figure 6. The inputs of the encoder–decoder component.
Electronics 11 00861 g006
Figure 7. The comparison cavers under 30 min forecasting interval.
Figure 7. The comparison cavers under 30 min forecasting interval.
Electronics 11 00861 g007
Figure 8. The comparison cavers under 60 min forecasting interval.
Figure 8. The comparison cavers under 60 min forecasting interval.
Electronics 11 00861 g008
Figure 9. The comparison cavers under 120 min forecasting interval.
Figure 9. The comparison cavers under 120 min forecasting interval.
Electronics 11 00861 g009
Table 1. The traffic forecasting results under different models with 30 min forecasting interval.
Table 1. The traffic forecasting results under different models with 30 min forecasting interval.
MethodHAARIMALSTMGRUGCRNGated-STGCNGWNETOurs_WAOurs_WSOurs
M A E 5.693.963.213.202.992.962.892.932.892.93
R M S E 7.606.144.854.854.544.484.374.484.374.13
M A P E 20.02%14.12%12.85%12.82%12.11%11.85%11.49%11.88%11.89%11.84%
Table 2. The traffic forecasting results under different models with 60 min forecasting interval.
Table 2. The traffic forecasting results under different models with 60 min forecasting interval.
MethodHAARIMALSTMGRUGCRNGated-STGCNGWNETOurs_WAOurs_WSOurs
M A E 5.694.483.673.673.303.293.163.153.032.98
R M S E 7.606.565.485.474.964.924.713.293.164.09
M A P E 20.02%16.10%14.92%14.86%13.53%13.33%12.52%11.58%11.47%11.38%
Table 3. The traffic forecasting results under different models with 120 min forecasting interval.
Table 3. The traffic forecasting results under different models with 120 min forecasting interval.
MethodHAARIMALSTMGRUGCRNGated-STGCNGWNETOurs_WAOurs_WSOurs
M A E 5.695.094.304.303.713.713.523.303.343.29
R M S E 7.607.016.266.265.455.445.165.054.994.85
M A P E 20.02%18.24%17.38%17.41%15.13%15.03%13.77%13.25%12.63%12.36%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Han, S.-Y.; Sun, Q.-W.; Zhao, Q.; Han, R.-Z.; Chen, Y.-H. Traffic Forecasting Based on Integration of Adaptive Subgraph Reformulation and Spatio-Temporal Deep Learning Model. Electronics 2022, 11, 861. https://doi.org/10.3390/electronics11060861

AMA Style

Han S-Y, Sun Q-W, Zhao Q, Han R-Z, Chen Y-H. Traffic Forecasting Based on Integration of Adaptive Subgraph Reformulation and Spatio-Temporal Deep Learning Model. Electronics. 2022; 11(6):861. https://doi.org/10.3390/electronics11060861

Chicago/Turabian Style

Han, Shi-Yuan, Qi-Wei Sun, Qiang Zhao, Rui-Zhi Han, and Yue-Hui Chen. 2022. "Traffic Forecasting Based on Integration of Adaptive Subgraph Reformulation and Spatio-Temporal Deep Learning Model" Electronics 11, no. 6: 861. https://doi.org/10.3390/electronics11060861

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop