Next Article in Journal
Circular Economy for Sustainability: A Bibliometric Analysis for the Tourism/Hospitality Field
Previous Article in Journal
Exploring Apulia’s Regional Tourism Attractiveness through the Lens of Sustainability: A Machine Learning Approach and Counterfactual Explainability Process
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Research on Dynamic Subsidy Based on Deep Reinforcement Learning for Non-Stationary Stochastic Demand in Ride-Hailing

School of Business, East China University of Science and Technology, Shanghai 200237, China
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(15), 6289; https://doi.org/10.3390/su16156289
Submission received: 17 June 2024 / Revised: 13 July 2024 / Accepted: 22 July 2024 / Published: 23 July 2024

Abstract

The ride-hailing market often experiences significant fluctuations in traffic demand, resulting in supply-demand imbalances. In this regard, the dynamic subsidy strategy is frequently employed by ride-hailing platforms to incentivize drivers to relocate to zones with high demand. However, determining the appropriate amount of subsidy at the appropriate time remains challenging. First, traffic demand exhibits high non-stationarity, characterized by multi-context patterns with time-varying statistical features. Second, high-dimensional state/action spaces contain multiple spatiotemporal dimensions and context patterns. Third, decision-making should satisfy real-time requirements. To address the above challenges, we first construct a Non-Stationary Markov Decision Process (NSMDP) based on the assumption of ride-hailing service systems dynamics. Then, we develop a solution framework for the NSMDP. A change point detection method based on feature-enhanced LSTM within the framework can identify the changepoints and time-varying context patterns of stochastic demand. Moreover, the framework also includes a deterministic policy deep reinforcement learning algorithm to optimize. Finally, through simulated experiments with real-world historical data, we demonstrate the effectiveness of the proposed approach. It performs well in improving the platform’s profits and alleviating supply-demand imbalances under the dynamic subsidy strategy. The results also prove that a scientific dynamic subsidy strategy is particularly effective in the high-demand context pattern with more drastic fluctuations. Additionally, the profitability of dynamic subsidy strategy will increase with the increase of the non-stationary level.
Keywords: ride-hailing; nonstationary stochastic demand; change point detection; non-stationary Markov decision; deep reinforcement learning ride-hailing; nonstationary stochastic demand; change point detection; non-stationary Markov decision; deep reinforcement learning

Share and Cite

MDPI and ACS Style

Huang, X.; Cheng, Y.; Jin, J.; Kou, A. Research on Dynamic Subsidy Based on Deep Reinforcement Learning for Non-Stationary Stochastic Demand in Ride-Hailing. Sustainability 2024, 16, 6289. https://doi.org/10.3390/su16156289

AMA Style

Huang X, Cheng Y, Jin J, Kou A. Research on Dynamic Subsidy Based on Deep Reinforcement Learning for Non-Stationary Stochastic Demand in Ride-Hailing. Sustainability. 2024; 16(15):6289. https://doi.org/10.3390/su16156289

Chicago/Turabian Style

Huang, Xiangyu, Yan Cheng, Jing Jin, and Aiqing Kou. 2024. "Research on Dynamic Subsidy Based on Deep Reinforcement Learning for Non-Stationary Stochastic Demand in Ride-Hailing" Sustainability 16, no. 15: 6289. https://doi.org/10.3390/su16156289

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop