Travel Time Prediction and Explanation with Spatio-Temporal Features: A Comparative Study
Abstract
:1. Introduction
- RQ1:
- To what extent can data-driven methods be applied for predicting travel time using spatiotemporal features?
- RQ2:
- To what extent can XAI methods be applied for explaining travel time predictions?
2. Travel Time Prediction Methods
2.1. An Overview
2.2. Related Work: Comparative Analysis of Travel Time Prediction Methods
3. Materials and Methods
3.1. Data Understanding
3.2. Data Preparation
3.3. Model Training and Tuning
- Ensemble Learning Models. Ensemble learning enhances the prediction performance of one model by training multiple models simultaneously and combining their predictive power to achieve the best performance possible [43]. Many ensemble learning methods are available, and this study considers two widely used gradient boosting methods [44,45]: XGBoost (eXtreme Gradient Boosting) and LightGBM (Light Gradient Boosting Machine). Boosting models consist of a sequence of regression trees, where every successive tree tries to correct the previous tree’s mistakes. Hence, increasing the prediction accuracy of the overall model [22]. XGBoost applies level-wise (horizontal) tree growth, whereas LightGBM applies leaf-wise (vertical) tree growth. Compared with XGBOOST, LightGBM is computationally less expensive and has better prediction accuracy [44]. We used the standard hyper-parameters of the two learning algorithms [40,41,45]: learning_rate, colsample_bytree, n_estimators, and max_depth for XGBoost, and learning_rate, bagging_frequency, n_estimators, and max_dept for LightGBM.
- Deep Neural Network Models. Neural networks are one of the most popular machine learning techniques [46]. They are represented as layered organizations of neurons with connections to other neurons, mimicking how biological neurons signal to one another. Neural networks can be used for travel time prediction as they can learn non-linear relations among variables [27,28,29,30,31,32]. This study uses long short-term memory (LSTM) and gated recurrent units (GRU) techniques of neural networks as they are more suitable for long sequence data [35]. We use the traditional LSTM and its extension, namely bidirectional LSTM, which combines a forward and a backward pass of operations, enabling considering past instances and future ones.
- Hybrid Models. Following the TT prediction literature [33,35], we selected multiple hybrid models by combining one deep learning model and one ensemble learning model in combination with a linear model for final prediction. Figure 2 shows the architecture of the hybrid model used. In this architecture, two different types of machine learning models are combined, and then the output of those two models is passed through a linear regression model to get the final result. In this study, four hybrid models are considered: a GRU model in combination with LightGBM, a GRU model in combination with XGBoost, an LSTM model in combination with LightGBM, and an LSTM model in combination with XGBoost.
- Linear SVMR. Support vector machine regression (SVMR) is based on statistical learning theory and can improve the ability of generalization by seeking the minimum structural risk [47]. We use the Linear SVMR model as the baseline since several TTP prediction studies use linear regression models as baselines [35,36].
3.4. Model Evaluation
- Score. It is a statistical measure that determines the proportion of variance in the dependent variable that can be explained by one or more independent variables in a regression model. score indicates how well the trained model fits the data. The score lies between 0 and 1, where a score of 0 means that the model does not capture any pattern in the data, and the predictions will be random. On the other hand, if the score is 1, the model perfectly fits the data and generalizes very well. formula is:
- RMSE. Root mean square error or deviation is a measurement of the difference between model prediction and actual value. The deviations in predicted values from actual values are known as residual. It is calculated over the test set and is also known as prediction error. RMSE is always positive, and 0 is considered a perfect fit on the data. The formula for RMSE is:
- MAE. Mean absolute error is the mean of the absolute errors, differences between predicted and actual values. It indicates how big of an error we can expect from the prediction on average. MAE formula is:
3.5. Model Explanation
4. Results
4.1. RQ1: Comparison of TTP Methods
4.2. RQ2: Comparison of TTP Explanation Methods
4.3. Global Explanations
4.4. Local Explanations
5. Discussion
5.1. Summary of Answers to Research Questions
5.2. Threats to Validity
6. Conclusions and Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Bai, M.; Lin, Y.; Ma, M.; Wang, P. Travel-Time Prediction Methods: A Review. In Smart Computing and Communication; Qiu, M., Ed.; Springer International Publishing: Cham, Switzerland, 2018; pp. 67–77. [Google Scholar]
- Oh, S.; Byon, Y.J.; Jang, K.; Yeo, H. Short-term Travel-time Prediction on Highway: A Review of the Data-driven Approach. Transp. Rev. 2015, 35, 4–32. [Google Scholar] [CrossRef]
- Qiu, B.; Fan, W.D. Machine Learning Based Short-Term Travel Time Prediction: Numerical Results and Comparative Analyses. Sustainability 2021, 13, 7454. [Google Scholar] [CrossRef]
- Teresa, G.; Evangelos, G. Importance of logistics services attributes influencing customer satisfaction. In Proceedings of the 2015 4th International Conference on Advanced Logistics and Transport (ICALT), Valenciennes, France, 20–22 May 2015; pp. 53–58. [Google Scholar] [CrossRef]
- Li, S.; Ragu-Nathan, B.; Ragu-Nathan, T.; Rao, S.S. The impact of supply chain management practices on competitive advantage and organizational performance. Omega 2006, 34, 107–124. [Google Scholar] [CrossRef]
- Tang, J.; Zheng, L.; Han, C.; Yin, W.; Zhang, Y.; Zou, Y.; Huang, H. Statistical and machine-learning methods for clearance time prediction of road incidents: A methodology review. Anal. Methods Accid. Res. 2020, 27, 100123. [Google Scholar] [CrossRef]
- Cheng, J.; Li, G.; Chen, X. Research on travel time prediction model of freeway based on gradient boosting decision tree. IEEE Access 2018, 7, 7466–7480. [Google Scholar] [CrossRef]
- Abdollahi, M.; Khaleghi, T.; Yang, K. An integrated feature learning approach using deep learning for travel time prediction. Expert Syst. Appl. 2020, 139, 112864. [Google Scholar] [CrossRef]
- Petersen, N.C.; Rodrigues, F.; Pereira, F.C. Multi-output bus travel time prediction with convolutional LSTM neural network. Expert Syst. Appl. 2019, 120, 426–435. [Google Scholar] [CrossRef] [Green Version]
- Zhao, J.; Qu, Q.; Zhang, F.; Xu, C.; Liu, S. Spatio-Temporal Analysis of Passenger Travel Patterns in Massive Smart Card Data. IEEE Trans. Intell. Transp. Syst. 2017, 18, 3135–3146. [Google Scholar] [CrossRef]
- Chen, C.H. Temporal-Spatial Feature Extraction Based on Convolutional Neural Networks for Travel Time Prediction. arXiv 2021, arXiv:2111.00149. [Google Scholar]
- Molnar, C. Interpretable Machine Learning; Leanpub: Victoria, BC, Canada, 2020. [Google Scholar]
- Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4765–4774. [Google Scholar]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
- Khetarpaul, S.; Gupta, S.; Malhotra, S.; Subramaniam, L.V. Bus arrival time prediction using a modified amalgamation of fuzzy clustering and neural network on spatio-temporal data. In Australasian Database Conference; Springer: Cham, Switzerland, 2015; pp. 142–154. [Google Scholar]
- Reshadat, V.; Hoorali, M.; Faili, H. A hybrid method for open information extraction based on shallow and deep linguistic analysis. Interdiscip. Inf. Sci. 2016, 22, 87–100. [Google Scholar] [CrossRef] [Green Version]
- Reshadat, V.; Faili, H. A new open information extraction system using sentence difficulty estimation. Comput. Inform. 2019, 38, 986–1008. [Google Scholar] [CrossRef]
- Reshadat, V.; Feizi-Derakhshi, M.R. Studying of semantic similarity methods in ontology. Res. J. Appl. Sci. Eng. Technol. 2012, 4, 1815–1821. [Google Scholar]
- Reshadat, V.; Hourali, M.; Faili, H. Confidence Measure Estimation for Open Information Extraction. Inf. Syst. Telecommun. 2018, 1, 1–10. [Google Scholar]
- Kakani, V.; Nguyen, V.H.; Kumar, B.P.; Kim, H.; Pasupuleti, V.R. A critical review on computer vision and artificial intelligence in food industry. J. Agric. Food Res. 2020, 2, 100033. [Google Scholar] [CrossRef]
- Vyborny, C.J.; Giger, M.L. Computer vision and artificial intelligence in mammography. AJR Am. J. Roentgenol. 1994, 162, 699–708. [Google Scholar] [CrossRef]
- Zhang, Y.; Haghani, A. A gradient boosting method to improve travel time prediction. Transp. Res. Part C Emerg. Technol. 2015, 58, 308–324. [Google Scholar] [CrossRef]
- Zahid, M.; Chen, Y.; Jamal, A.; Mamadou, C.Z. Freeway Short-Term Travel Speed Prediction Based on Data Collection Time-Horizons: A Fast Forest Quantile Regression Approach. Sustainability 2020, 12, 646. [Google Scholar] [CrossRef] [Green Version]
- Cristóbal, T.; Padrón, G.; Quesada-Arencibia, A.; Alayón, F.; de Blasio, G.; García, C.R. Bus Travel Time Prediction Model Based on Profile Similarity. Sensors 2019, 19, 2869. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Chen, Z.; Fan, W. A Freeway Travel Time Prediction Method Based on an XGBoost Model. Sustainability 2021, 13, 8577. [Google Scholar] [CrossRef]
- Zhang, F.; Zhu, X.; Hu, T.; Guo, W.; Chen, C.; Liu, L. Urban Link Travel Time Prediction Based on a Gradient Boosting Method Considering Spatiotemporal Correlations. ISPRS Int. J. Geo-Inf. 2016, 5, 201. [Google Scholar] [CrossRef] [Green Version]
- Wang, Z.; Fu, K.; Ye, J. Learning to estimate the travel time. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 858–866. [Google Scholar]
- Li, X.; Wang, H.; Sun, P.; Zu, H. Spatiotemporal Features—Extracted Travel Time Prediction Leveraging Deep-Learning-Enabled Graph Convolutional Neural Network Model. Sustainability 2021, 13, 1253. [Google Scholar] [CrossRef]
- Yuan, Y.; Shao, C.; Cao, Z.; He, Z.; Zhu, C.; Wang, Y.; Jang, V. Bus Dynamic Travel Time Prediction: Using a Deep Feature Extraction Framework Based on RNN and DNN. Electronics 2020, 9, 1876. [Google Scholar] [CrossRef]
- Wu, J.; Wu, Q.; Shen, J.; Cai, C. Towards Attention-Based Convolutional Long Short-Term Memory for Travel Time Prediction of Bus Journeys. Sensors 2020, 20, 3354. [Google Scholar] [CrossRef]
- Ran, X.; Shan, Z.; Fang, Y.; Lin, C. A Convolution Component-Based Method with Attention Mechanism for Travel-Time Prediction. Sensors 2019, 19, 2063. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ran, X.; Shan, Z.; Fang, Y.; Lin, C. An LSTM-Based Method with Attention Mechanism for Travel Time Prediction. Sensors 2019, 19, 861. [Google Scholar] [CrossRef] [Green Version]
- Ting, P.Y.; Wada, T.; Chiu, Y.L.; Sun, M.T.; Sakai, K.; Ku, W.S.; Jeng, A.A.K.; Hwu, J.S. Freeway Travel Time Prediction Using Deep Hybrid Model–Taking Sun Yat-Sen Freeway as an Example. IEEE Trans. Veh. Technol. 2020, 69, 8257–8266. [Google Scholar] [CrossRef]
- Yang, S.; Qian, S. Understanding and Predicting Travel Time with Spatio-Temporal Features of Network Traffic Flow, Weather and Incidents. IEEE Intell. Transp. Syst. Mag. 2019, 11, 12–28. [Google Scholar] [CrossRef] [Green Version]
- Liu, Y.; Wang, Y.; Yang, X.; Zhang, L. Short-term travel time prediction by deep learning: A comparison of different LSTM-DNN models. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–8. [Google Scholar]
- Goudarzi, F. Travel Time Prediction: Comparison of Machine Learning Algorithms in a Case Study. In Proceedings of the 2018 IEEE 20th International Conference on High Performance Computing and Communications; IEEE 16th International Conference on Smart City; IEEE 4th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Exeter, UK, 28–30 June 2018; pp. 1404–1407. [Google Scholar] [CrossRef]
- Adewale, A.E.; Hadachi, A. Neural Networks Model for Travel Time Prediction Based on ODTravel Time Matrix. arXiv 2020, arXiv:2004.04030. [Google Scholar]
- Fiosina, J. Explainable Federated Learning for Taxi Travel Time Prediction. In Proceedings of the 7th International Conference on Vehicle Technology and Intelligent Transport Systems (VEHITS 2021), Online Streaming, 28–30 April 2021; pp. 670–677. [Google Scholar]
- Fan, W.D.; Chen, Z. Predicting Travel Time on Freeway Corridors: Machine Learning Approach; Technical Report; Center for Advanced Multimodal Mobility Solutions and Education, University of North Carolina at Charlotte: Charlotte, NC, USA, 2020. [Google Scholar]
- Mohammadi, M.R.; Hadavimoghaddam, F.; Pourmahdi, M.; Atashrouz, S.; Munir, M.T.; Hemmati-Sarapardeh, A.; Mosavi, A.H.; Mohaddespour, A. Modeling hydrogen solubility in hydrocarbons using extreme gradient boosting and equations of state. Sci. Rep. 2021, 11, 17911. [Google Scholar] [CrossRef]
- Janizadeh, S.; Vafakhah, M.; Kapelan, Z.; Mobarghaee Dinan, N. Hybrid XGboost model with various Bayesian hyperparameter optimization algorithms for flood hazard susceptibility modeling. Geocarto Int. 2021, 1–20. [Google Scholar] [CrossRef]
- Martínez-Plumed, F.; Contreras-Ochando, L.; Ferri, C.; Orallo, J.H.; Kull, M.; Lachiche, N.; Quintana, M.J.R.; Flach, P.A. CRISP-DM twenty years later: From data mining processes to data science trajectories. IEEE Trans. Knowl. Data Eng. 2019. [Google Scholar] [CrossRef] [Green Version]
- Sagi, O.; Rokach, L. Ensemble learning: A survey. WIREs Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
- Liang, W.; Luo, S.; Zhao, G.; Wu, H. Predicting Hard Rock Pillar Stability Using GBDT, XGBoost, and LightGBM Algorithms. Mathematics 2020, 8, 765. [Google Scholar] [CrossRef]
- Mei, Z.; Xiang, F.; Zhen-hui, L. Short-Term Traffic Flow Prediction Based on Combination Model of Xgboost-Lightgbm. In Proceedings of the 2018 International Conference on Sensor Networks and Signal Processing (SNSP), Xi’an, China, 28–31 October 2018; pp. 322–327. [Google Scholar] [CrossRef]
- Pouyanfar, S.; Sadiq, S.; Yan, Y.; Tian, H.; Tao, Y.; Reyes, M.P.; Shyu, M.L.; Chen, S.C.; Iyengar, S.S. A Survey on Deep Learning: Algorithms, Techniques, and Applications. ACM Comput. Surv. 2018, 51, 1–36. [Google Scholar] [CrossRef]
- Zhang, J.; Liao, Y.; Wang, S.; Han, J. Study on driving decision-making mechanism of autonomous vehicle based on an optimized support vector machine regression. Appl. Sci. 2018, 8, 13. [Google Scholar] [CrossRef] [Green Version]
- James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning; Springer: New York, NY, USA, 2013; Volume 112. [Google Scholar]
- Lever, J.; Krzywinski, M.; Altman, N. Model selection and overfitting. Nat. Methods 2016, 13, 703–704. [Google Scholar] [CrossRef]
- Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
- Zimmerman, D.W.; Zumbo, B.D. Relative power of the Wilcoxon test, the Friedman test, and repeated-measures ANOVA on ranks. J. Exp. Educ. 1993, 62, 75–86. [Google Scholar] [CrossRef]
- Benavoli, A.; Corani, G.; Mangili, F. Should we really use post-hoc tests based on mean-ranks? J. Mach. Learn. Res. 2016, 17, 152–161. [Google Scholar]
- Fawaz, H.I.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Deep learning for time series classification: A review. Data Min. Knowl. Discov. 2019, 33, 917–963. [Google Scholar] [CrossRef] [Green Version]
- Xie, Y.; Pongsakornsathien, N.; Gardi, A.; Sabatini, R. Explanation of Machine-Learning Solutions in Air-Traffic Management. Aerospace 2021, 8, 224. [Google Scholar] [CrossRef]
- Velmurugan, M.; Ouyang, C.; Moreira, C.; Sindhgatta, R. Evaluating Fidelity of Explainable Methods for Predictive Process Analytics. In Intelligent Information Systems; Nurcan, S., Korthaus, A., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 64–72. [Google Scholar]
- Wohlin, C.; Runeson, P.; Höst, M.; Ohlsson, M.C.; Regnell, B.; Wesslén, A. Experimentation in Software Engineering; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
Total Number of Trips | 14,135 |
Average Number of Trips per Month | 643 |
Average Number of Trips per Day | 24 |
Average Number of Stops per Trip | 10 |
Total Number of Trips | 5272 |
Average Number of Trips per Month | 195 |
Average Number of Trips per Day | 7 |
Average Number of Stops per Trip | 7 |
Total Number of Travels | 211,392 |
Average Number of Travels per Route | 30,198 |
Average Number of Travels per Month | 35,232 |
Average Number of Travels per Day | 1155 |
Data Set/Model | NextUp-1 | NextUp-2 | PeMS | ||||||
---|---|---|---|---|---|---|---|---|---|
RMSE | MAE | RMSE | MAE | RMSE | MAE | ||||
XGBoost | 26.51 | 14.57 | 0.8083 | 18.31 | 10.64 | 0.7646 | 0.61 | 0.39 | 0.9993 |
LightGBM | 26.54 | 14.40 | 0.8079 | 18.30 | 10.65 | 0.7647 | 0.64 | 0.41 | 0.9992 |
LSTM | 29.97 | 16.59 | 0.7551 | 24.74 | 13.41 | 0.5704 | 0.87 | 0.51 | 0.9987 |
BiLSTM | 29.96 | 16.30 | 0.7553 | 23.16 | 12.72 | 0.6234 | 0.93 | 0.55 | 0.9985 |
GRU | 29.97 | 16.555 | 0.7550 | 25.04 | 13.51 | 0.5597 | 0.80 | 0.49 | 0.9989 |
LinearSVMR | 49.48 | 25.77 | 0.3323 | 26.20 | 14.01 | 0.5180 | 3.40 | 1.12 | 0.9797 |
Hybrid-1 | 26.53 | 14.49 | 0.8080 | 18.80 | 11.34 | 0.7519 | 0.65 | 0.42 | 0.9993 |
Hybrid-2 | 26.55 | 14.35 | 0.8078 | 18.45 | 10.87 | 0.7611 | 0.67 | 0.43 | 0.9992 |
Hybrid-3 | 26.52 | 14.50 | 0.8082 | 18.56 | 11.13 | 0.7580 | 0.65 | 0.42 | 0.9993 |
Hybrid-4 | 26.55 | 14.35 | 0.8078 | 18.45 | 10.87 | 0.7611 | 0.67 | 0.43 | 0.9992 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ahmed, I.; Kumara, I.; Reshadat, V.; Kayes, A.S.M.; van den Heuvel, W.-J.; Tamburri, D.A. Travel Time Prediction and Explanation with Spatio-Temporal Features: A Comparative Study. Electronics 2022, 11, 106. https://doi.org/10.3390/electronics11010106
Ahmed I, Kumara I, Reshadat V, Kayes ASM, van den Heuvel W-J, Tamburri DA. Travel Time Prediction and Explanation with Spatio-Temporal Features: A Comparative Study. Electronics. 2022; 11(1):106. https://doi.org/10.3390/electronics11010106
Chicago/Turabian StyleAhmed, Irfan, Indika Kumara, Vahideh Reshadat, A. S. M. Kayes, Willem-Jan van den Heuvel, and Damian A. Tamburri. 2022. "Travel Time Prediction and Explanation with Spatio-Temporal Features: A Comparative Study" Electronics 11, no. 1: 106. https://doi.org/10.3390/electronics11010106
APA StyleAhmed, I., Kumara, I., Reshadat, V., Kayes, A. S. M., van den Heuvel, W. -J., & Tamburri, D. A. (2022). Travel Time Prediction and Explanation with Spatio-Temporal Features: A Comparative Study. Electronics, 11(1), 106. https://doi.org/10.3390/electronics11010106