Traffic Volume Estimation Based on Spatiotemporal Correlation Adaptive Graph Convolutional Network
Abstract
:1. Introduction
- We employ a dynamic adjacency matrix and DGCR module to adaptively model the spatial correlations of road traffic. The ASFC module adapts to historical speed data to extract road patterns and capture global spatial relationships, improving model performance.
- We design the MSTF module with a pyramid structure. This structure efficiently captures periodic and long-term dependencies in traffic flow by segmenting time-series data from the state matrix.
- Experiments on the PEMS-BAY real-world traffic dataset demonstrate that our model achieves improvements with sensor coverage rates ranging from 30% to 50%, outperforming all baseline models. Ablation and hyperparameter experiments further validate the rationality of the model architecture.
2. Literature Review
2.1. Method of Traffic Volume Estimation
- Model-driven approaches require large amounts of accurate and complete traffic data for parameter calibration and validation. This results in high data requirements and complex tasks.
- In real-world traffic networks, traffic flow often does not meet the model’s assumption of evenly distributed vehicles. This limits the accuracy of these methods and makes them less adaptable to dynamic changes in the road network.
- Machine Learning Methods: In recent years, methods such as deep learning, support vector machines (SVM), and random forests (RF) have been widely used for traffic volume estimation. By training on datasets, machine learning models can automatically identify spatiotemporal features in traffic volume and make accurate estimations. These methods offer greater adaptability. With advancements in sensor technology and the increase in the amount of data, machine learning-based methods have gradually become the mainstream in traffic volume estimation.
2.2. Application of Transformer and Its Variants in Traffic Time-Series Modeling
3. Problem Formulation
4. Methodology
- Spatial Correlation: The dynamic graph convolution recurrent module and the adaptive speed-flow correlation module work together to capture the spatial correlations of traffic volume. The DGCR models the traffic volume relationships between roads based on their relative static distances. The adjacency matrix is dynamically updated according to the changes in traffic volume. This enhances the model’s expressiveness. The ASFC dynamically captures different speed-flow patterns across various roads, significantly improving accuracy compared to static linear methods.
- Temporal Correlation: The Multi-Scale Transformer (MSTF) module is used to extract the temporal dependencies of traffic volume. The MSTF employs a multi-head attention pyramid structure that effectively captures periodic and long-term dependencies in traffic volume. It also retains the efficient training characteristics of the transformer structure, allowing for more effective handling of large datasets.
4.1. Spatial Correlations
4.1.1. Graph Convolution and Local Spatial Dependency
4.1.2. Adaptive Speed-Flow Correlation Module
4.1.3. Dynamic Graph Convolution Recurrent Module
4.2. Temporal Correlations
4.3. Model Implementations
4.3.1. Output Layer
4.3.2. Training Strategy and Loss Function
- Traffic volume reconstruction loss
- 2.
- Learnable graph structure regularization
4.4. Time Complexity Analysis
5. Experiments and Result Analysis
5.1. Data
5.2. Parameter Settings
5.3. Baselines
- KNN: The K-Nearest Neighbors method is a simple and intuitive algorithm that calculates the correlation between nodes based on their distance. This serves as the benchmark for all traffic volume estimation methods, reflecting the performance improvements of different models.
- ST-SSL [6]: This is a pioneering work in the field of traffic volume estimation. It first introduced a semi-supervised learning model and used speed information to supplement missing volume data. This model eliminates the reliance on complete labeled data for traffic volume estimation.
- TGMC-F [8]: This method integrates traffic and speed data into a geometric matrix completion model and designs a loss function that includes spatiotemporal regularization, further improving the accuracy of traffic volume estimation.
- GCBRNN [2]: This model designs a graph convolutional gated recurrent module to capture the spatiotemporal correlations in the data, and it performs both traffic volume estimation and traffic flow prediction tasks.
5.4. Volume Estimation Results at Different Coverage Rates
5.5. Model Analysis
5.5.1. Ablation Study
- DTSAGCN w/o DGCR: We replaced the dynamic adjacency matrix in DTSAGCN with a static adjacency matrix based on distance calculations between roads.
- DTSAGCN w/o ASFC: We directly computed the adjacency matrix using the Pearson correlation coefficient between road speeds, without considering the temporal delay of traffic volume.
- DTSAGCN w/o MSTF: We replaced the multi-scale transformer module in DTSAGCN with a standard transformer, without explicitly modeling the periodicity of traffic volume.
- After replacing the DGCR module, the model’s performance significantly declined, demonstrating the necessity of capturing the dynamic spatial correlations of traffic volume. Both static speed–volume adjacency matrices and distance-based correlation coefficient adjacency matrices fail to fully reflect the time-varying, unbalanced spatial correlations between roads.
- After replacing the ASFC module, the performance decreased, but overall, its impact on the entire model was minimal. This may be because the capturing of spatial correlations in the overall model is somewhat redundant.
- After replacing the MSTF module, the model’s performance showed a noticeable decline. This is because the standard transformer structure cannot retain the sequence positions and lacks the ability to capture long-term dependencies, which is crucial in traffic volume tasks, especially for capturing periodicity with longer time steps, such as daily and weekly cycles.
5.5.2. Hyperparameter Study
- Dimension of hidden node states, ranging from 32 to 256. The results are shown in Figure 6a. The dimension of the hidden node states determines the complexity of each node’s feature representation in the graph convolution network. The estimation accuracy of the DTSAGCN model increases significantly as the dimension of hidden node states increases, indicating that accommodating more features can enhance model performance. However, when the dimension exceeds 128, the model’s estimation accuracy decreases, suggesting that excessively large dimensions may lead to overfitting and reduced model training efficiency.
- Number of graph convolutional layers, ranging from one to four. The results are shown in Figure 6b. The number of graph convolutional layers determines the ability to aggregate information from neighboring nodes. The estimation accuracy of the DTSAGCN model improves initially as the number of layers increases, suggesting that deeper learning can focus on more correlations between nodes. However, when the number of layers exceeds three, the model’s estimation accuracy decreases, indicating that too many layers may introduce excessive information, causing the model to become overly smooth and risk underfitting.
- Number of multi-head attention heads in the transformer, ranging from one to eight. The results are shown in Figure 6c. The number of attention heads in the temporal extraction module determines the model’s flexibility in capturing long-term dependencies in time series. The estimation accuracy of the DTSAGCN model improves as the number of heads increases, showing that more attention heads can effectively capture multi-scale temporal patterns. However, when the number of heads exceeds four, the model’s estimation accuracy declines, suggesting that an excessively high model complexity may weaken the model’s performance.
5.5.3. Temporal Interpretability Study
6. Conclusions and Future Directions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
ITS | Intelligent Transportation Systems |
GCN | Graph Convolutional Network |
GNN | Graph Neural Network |
DTSAGCN | Dynamic Temporal-Spatial Attention Graph Convolutional Network |
DGCR | Dynamic Graph Convolution Recurrent |
ASFC | Adaptive Speed-Flow Correlation |
MSTF | Multi-Scale Transformer |
PIDL | Physical Information Deep Learning |
CNN | Convolutional Neural Network |
LSTM | Long Short-Term Memory |
RNN | Recurrent Neural Network |
MFD | Macroscopic Fundamental Diagram |
References
- Ren, Y.; Yin, H.; Wang, L.; Ji, H. Data-Driven RBFNN-Enhanced Model-Free Adaptive Traffic Symmetrical Signal Control for a Multi-Phase Intersection with Fast-Changing Traffic Flow. Symmetry 2023, 15, 1235. [Google Scholar] [CrossRef]
- Zhang, Z.; Lin, X.; Li, M.; Wang, Y. A Customized Deep Learning Approach to Integrate Network-Scale Online Traffic Data Imputation and Prediction. Transp. Res. Part C Emerg. Technol. 2021, 132, 103372. [Google Scholar] [CrossRef]
- Nigam, N.; Singh, D.P.; Choudhary, J. A Review of Different Components of the Intelligent Traffic Management System (ITMS). Symmetry 2023, 15, 583. [Google Scholar] [CrossRef]
- Aslam, J.; Lim, S.; Pan, X.; Rus, D. City-Scale Traffic Estimation from a Roving Sensor Network. In Proceedings of the 10th ACM Conference on Embedded Network Sensor Systems, Toronto, ON, Canada, 6–9 November 2012; Association for Computing Machinery: New York, NY, USA, 2012; pp. 141–154. [Google Scholar]
- Zhan, X.; Zheng, Y.; Yi, X.; Ukkusuri, S.V. Citywide Traffic Volume Estimation Using Trajectory Data. IEEE Trans. Knowl. Data Eng. 2017, 29, 272–285. [Google Scholar] [CrossRef]
- Meng, C.; Yi, X.; Su, L.; Gao, J.; Zheng, Y. City-Wide Traffic Volume Inference with Loop Detector Data and Taxi Trajectories. In Proceedings of the 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Redondo Beach, CA, USA, 7–10 November 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 1–10. [Google Scholar]
- Tang, X.; Gong, B.; Yu, Y.; Yao, H.; Li, Y.; Xie, H.; Wang, X. Joint Modeling of Dense and Incomplete Trajectories for Citywide Traffic Volume Inference. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 1806–1817. [Google Scholar]
- Zhang, Z.; Li, M.; Lin, X.; Wang, Y. Network-Wide Traffic Flow Estimation with Insufficient Volume Detection and Crowdsourcing Data. Transp. Res. Part C Emerg. Technol. 2020, 121, 102870. [Google Scholar] [CrossRef]
- Atwood, J.; Towsley, D. Diffusion-Convolutional Neural Networks. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Newry, UK, 2016; Volume 29. [Google Scholar]
- Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2017, arXiv:1609.02907. [Google Scholar] [CrossRef]
- Guo, S.; Lin, Y.; Feng, N.; Song, C.; Wan, H. Attention Based Spatial-Temporal Graph Convolutional Networks for Traffic Flow Forecasting. Proc. AAAI Conf. Artif. Intell. 2019, 33, 922–929. [Google Scholar] [CrossRef]
- Jiang, W.; Luo, J. Graph Neural Network for Traffic Forecasting: A Survey. Expert Syst. Appl. 2022, 207, 117921. [Google Scholar] [CrossRef]
- Qu, Z.; He, S. A Time-Space Network Model Based on a Train Diagram for Predicting and Controlling the Traffic Congestion in a Station Caused by an Emergency. Symmetry 2019, 11, 780. [Google Scholar] [CrossRef]
- Zhou, H.; Zhang, D.; Xie, K. Accurate Traffic Matrix Completion Based on Multi-Gaussian Models. Comput. Commun. 2017, 102, 165–176. [Google Scholar] [CrossRef]
- Kampitakis, E.P.; Fafoutellis, P.; Oprea, G.-M.; Vlahogianni, E.I. Shared Space Multi-Modal Traffic Modeling Using LSTM Networks with Repulsion Map and an Intention-Based Multi-Loss Function. Transp. Res. Part C Emerg. Technol. 2023, 150, 104104. [Google Scholar] [CrossRef]
- Yuan, Y.; Zhang, Y.; Wang, B.; Peng, Y.; Hu, Y.; Yin, B. STGAN: Spatio-Temporal Generative Adversarial Network for Traffic Data Imputation. IEEE Trans. Big Data 2023, 9, 200–211. [Google Scholar] [CrossRef]
- Richards, P.I. Shock Waves on the Highway. Oper. Res. 1956, 4, 42–51. [Google Scholar] [CrossRef]
- Aw, A.; Rascle, M. Resurrection of “Second Order” Models of Traffic Flow. SIAM J. Appl. Math. 2000, 60, 916–938. [Google Scholar] [CrossRef]
- Zhang, H.M. A Non-Equilibrium Traffic Model Devoid of Gas-like Behavior. Transp. Res. Part B Methodol. 2002, 36, 275–290. [Google Scholar] [CrossRef]
- Zhang, J.; Mao, S.; Yang, L.; Ma, W.; Li, S.; Gao, Z. Physics-Informed Deep Learning for Traffic State Estimation Based on the Traffic Flow Model and Computational Graph Method. Inf. Fusion 2024, 101, 101971. [Google Scholar] [CrossRef]
- Huang, A.J.; Agarwal, S. Physics Informed Deep Learning for Traffic State Estimation. In Proceedings of the 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), Rhodes, Greece, 20–23 September 2020; pp. 1–6. [Google Scholar]
- Shahriari, S.; Ghasri, M.; Sisson, S.A.; Rashidi, T. Ensemble of ARIMA: Combining Parametric and Bootstrapping Technique for Traffic Flow Prediction. Transp. A Transp. Sci. 2020, 16, 1552–1573. [Google Scholar] [CrossRef]
- Zhao, M.; Yu, H.; Wang, Y.; Song, B.; Xu, L.; Zhu, D. Real-Time Freeway Traffic State Estimation for Inhomogeneous Traffic Flow. Phys. A Stat. Mech. Its Appl. 2024, 639, 129633. [Google Scholar] [CrossRef]
- Nie, L.; Li, Y.; Kong, X. Spatio-Temporal Network Traffic Estimation and Anomaly Detection Based on Convolutional Neural Network in Vehicular Ad-Hoc Networks. IEEE Access 2018, 6, 40168–40176. [Google Scholar] [CrossRef]
- Yu, B.; Yin, H.; Zhu, Z. Spatio-Temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 3634–3640. [Google Scholar]
- Zhang, M.; Chen, Y. Link Prediction Based on Graph Neural Networks. arXiv 2018, arXiv:1802.09691. [Google Scholar] [CrossRef]
- Abdelraouf, A.; Abdel-Aty, M.; Mahmoud, N. Sequence-to-Sequence Recurrent Graph Convolutional Networks for Traffic Estimation and Prediction Using Connected Probe Vehicle Data. IEEE Trans. Intell. Transp. Syst. 2023, 24, 1395–1405. [Google Scholar] [CrossRef]
- Li, R.; Wang, S.; Zhu, F.; Huang, J. Adaptive Graph Convolutional Neural Networks. Proc. AAAI Conf. Artif. Intell. 2018, 32, 11691. [Google Scholar] [CrossRef]
- Ta, X.; Liu, Z.; Hu, X.; Yu, L.; Sun, L.; Du, B. Adaptive Spatio-Temporal Graph Neural Network for Traffic Forecasting. Knowl. -Based Syst. 2022, 242, 108199. [Google Scholar] [CrossRef]
- Lu, Z.; Lv, W.; Cao, Y.; Xie, Z.; Peng, H.; Du, B. LSTM Variants Meet Graph Neural Networks for Road Speed Prediction. Neurocomputing 2020, 400, 34–45. [Google Scholar] [CrossRef]
- BAI, L.; Yao, L.; Li, C.; Wang, X.; Wang, C. Adaptive Graph Convolutional Recurrent Network for Traffic Forecasting. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Newry, UK, 2020; Volume 33, pp. 17804–17815. [Google Scholar]
- Zhang, Z.; Li, Y.; Song, H.; Dong, H. Multiple Dynamic Graph Based Traffic Speed Prediction Method. Neurocomputing 2021, 461, 109–117. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2016, arXiv:1706.03762. [Google Scholar] [CrossRef]
- Yan, H.; Ma, X.; Pu, Z. Learning Dynamic and Hierarchical Traffic Spatiotemporal Features With Transformer. IEEE Trans. Intell. Transp. Syst. 2022, 23, 22386–22399. [Google Scholar] [CrossRef]
- Nie, Y.; Nguyen, N.H.; Sinthong, P.; Kalagnanam, J. A Time Series Is Worth 64 Words: Long-Term Forecasting with Transformers. arXiv 2022, arXiv:2211.14730. [Google Scholar] [CrossRef]
- Zhang, Y.; Yan, J. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
- Zhang, Y.; Wu, R.; Dascalu, S.M.; Harris, F.C. Multi-Scale Transformer Pyramid Networks for Multivariate Time Series Forecasting. IEEE Access 2024, 12, 14731–14741. [Google Scholar] [CrossRef]
- Liang, W.; Li, Y.; Xie, K.; Zhang, D.; Li, K.-C.; Souri, A.; Li, K. Spatial-Temporal Aware Inductive Graph Neural Network for C-ITS Data Recovery. IEEE Trans. Intell. Transp. Syst. 2023, 24, 8431–8442. [Google Scholar] [CrossRef]
- Bi, J.; Yuan, H.; Xu, K.; Ma, H.; Zhou, M. Large-Scale Network Traffic Prediction with LSTM and Temporal Convolutional Networks. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 3865–3870. [Google Scholar]
- Zhang, R.; Sun, F.; Song, Z.; Wang, X.; Du, Y.; Dong, S. Short-Term Traffic Flow Forecasting Model Based on GA-TCN. J. Adv. Transp. 2021, 2021, 1338607. [Google Scholar] [CrossRef]
- Gao, H.; Jia, H.; Yang, L. An Improved CEEMDAN-FE-TCN Model for Highway Traffic Flow Prediction. J. Adv. Transp. 2022, 2022, 2265000. [Google Scholar] [CrossRef]
- Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting. arXiv 2017, arXiv:1707.01926. [Google Scholar] [CrossRef]
- Hong, R.; Liu, H.; An, C.; Wang, B.; Lu, Z.; Xia, J. An MFD Construction Method Considering Multi-Source Data Reliability for Urban Road Networks. Sustainability 2022, 14, 6188. [Google Scholar] [CrossRef]
- Yildirimoglu, M.; Ramezani, M.; Geroliminis, N. Equilibrium Analysis and Route Guidance in Large-Scale Networks with MFD Dynamics. Transp. Res. Part C Emerg. Technol. 2015, 59, 404–420. [Google Scholar] [CrossRef]
- Jiang, J.; Han, C.; Zhao, W.X.; Wang, J. PDFormer: Propagation Delay-Aware Dynamic Long-Range Transformer for Traffic Flow Prediction. Proc. AAAI Conf. Artif. Intell. 2023, 37, 4365–4373. [Google Scholar] [CrossRef]
- Ma, D.; Song, X.; Li, P. Daily Traffic Flow Forecasting Through a Contextual Convolutional Recurrent Neural Network Modeling Inter- and Intra-Day Traffic Patterns. IEEE Trans. Intell. Transp. Syst. 2021, 22, 2627–2636. [Google Scholar] [CrossRef]
- Han, L.; Du, B.; Sun, L.; Fu, Y.; Lv, Y.; Xiong, H. Dynamic and Multi-Faceted Spatio-Temporal Deep Learning for Traffic Speed Forecasting. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 547–555. [Google Scholar]
- Wu, Y.; Tan, H. Short-Term Traffic Flow Forecasting with Spatial-Temporal Correlation in a Hybrid Deep Learning Framework. arXiv 2016, arXiv:1612.01022. [Google Scholar] [CrossRef]
- Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Newry, UK, 2021; Volume 34, pp. 22419–22430. [Google Scholar]
- Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. FEDformer: Frequency Enhanced Decomposed Transformer for Long-Term Series Forecasting. In Proceedings of the 39th International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 27268–27286. [Google Scholar]
Coverage | Evaluation | DTSAGCN | KNN | ST-SSL | TGMC-F | GCBRNN |
---|---|---|---|---|---|---|
30% | MAE | 44.13 | 54.9 | 52.45 | 46.62 | 47.66 |
RMSE | 61.93 | 83.47 | 78.57 | 70.91 | 68.99 | |
MAPE | 35.42% | 42.75% | 41.08% | 38.48% | 37.83% | |
40% | MAE | 40.75 | 52.39 | 50.28 | 47.81 | 45.92 |
RMSE | 60.73 | 79.01 | 76.51 | 70.89 | 67.18 | |
MAPE | 32.41% | 41.69% | 37.20% | 36.82% | 35.54% | |
50% | MAE | 38.24 | 50.33 | 47.37 | 47.06 | 43.76 |
RMSE | 58.02 | 75.26 | 71.05 | 69.43 | 65.89 | |
MAPE | 44.13% | 54.9% | 52.45% | 46.62% | 47.66% |
DTSAGCN | w/o DGCR | w/o ASFC | w/o MSTF | |
---|---|---|---|---|
MAE | 40.75 | 47.14 | 43.27 | 45.76 |
RMSE | 60.73 | 72.26 | 64.22 | 66.86 |
MAPE | 32.41% | 39.91% | 35.43% | 37.15% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ding, S.; Yan, F.; Yi, Y. Traffic Volume Estimation Based on Spatiotemporal Correlation Adaptive Graph Convolutional Network. Symmetry 2025, 17, 599. https://doi.org/10.3390/sym17040599
Ding S, Yan F, Yi Y. Traffic Volume Estimation Based on Spatiotemporal Correlation Adaptive Graph Convolutional Network. Symmetry. 2025; 17(4):599. https://doi.org/10.3390/sym17040599
Chicago/Turabian StyleDing, Sheng, Fei Yan, and Yingmin Yi. 2025. "Traffic Volume Estimation Based on Spatiotemporal Correlation Adaptive Graph Convolutional Network" Symmetry 17, no. 4: 599. https://doi.org/10.3390/sym17040599
APA StyleDing, S., Yan, F., & Yi, Y. (2025). Traffic Volume Estimation Based on Spatiotemporal Correlation Adaptive Graph Convolutional Network. Symmetry, 17(4), 599. https://doi.org/10.3390/sym17040599