Green, Intelligent Traffic System

Repositioning Strategy for Ride-Hailing Vehicles Based on Geometric Road Network Structure and Reinforcement Learning

Expand
  • 1.School of Civil Engineering and Transportation,South China University of Technology,Guangzhou 510640,Guangdong,China
    2.School of Transportation Engineering,Chongqing Jiaotong University,Chongqing 400000,China
    3.School of Automobile and Transportation,Shenzhen Polytechnic,Shenzhen 518055,Guangdong,China
许伦辉(1965-),男,博士,教授,主要从事智能交通系统研究。E-mail:lhxu@scut.edu.cn

Received date: 2023-03-28

  Online published: 2023-06-21

Supported by

the Basic and Applied Research Foundation of Guangdong Province(2020A1515111024)

Abstract

The inefficient and inaccurate bidirectional search by both ride-hailing drivers and passengers leads to a mismatch between supply and demand. Ride-hailing vehicle repositioning strategy can pre-dispatch vehicles to areas with future demand, improving supply-demand matching. However, existing research mostly uses network grids to represent the urban road environment, lacking geometric topological information and reducing the dispatch accuracy. To address this issue, a ride-hailing vehicle relocation algorithm called GA2C was proposed based on Graph Neural Networks (GNN) and Actor-Critic reinforcement learning algorithm. This algorithm has a smoother learning process and can perform high-dimensional sampling, and it is suitable for learning the best relocation strategy for a large number of ride-hailing vehicles as multi-agent systems. Moreover, the geometric network structure was used to represent the urban road environment by using a GNN as a function approximator to learn the geometric information of the road network. Additionally, an action sampling strategy based on action value function was introduced to increase the randomness of action selection, effectively preventing competition. A ride-hailing vehicle relocation simulation experiment was conducted using Python, and the results are as follows: (i) the order response rate of the GA2C algorithm is 84.2%, significantly higher than all the comparative experimental results; (ii) in the order distribution comparative experiment, GA2C’s relative improvements in uniform distribution, central distribution layout, block distribution layout, and checkerboard distribution layout are 1.17%, 6.02%, 13.12%, and 14.55%, respectively. The above experimental results demonstrate that the GA2C algorithm can effectively relocate ride-hailing vehicles. When the order distribution presents significant differences, and the distance between different demand areas is relatively close, it can better learn dynamic demand changes, and achieve maximum order response rate by relocating ride-hailing vehicles.

Cite this article

XU Lunhui, YU Jiaxin, PEI Mingyang, et al . Repositioning Strategy for Ride-Hailing Vehicles Based on Geometric Road Network Structure and Reinforcement Learning[J]. Journal of South China University of Technology(Natural Science), 2023 , 51(10) : 99 -109 . DOI: 10.12141/j.issn.1000-565X.230148

References

1 XU Z T, YIN Y F, CHAO X L,et al .A generalized fluid model of ride-hailing systems[J].Transportation Research Part B:Methodological2021150:587-605.
2 CHENG S F, JHA S S, RAJENDRAM R .Taxis strike back:a field trial of the driver guidance system[C]∥Proceedings of the 17th International Conference on Autonomous Agents and Multi-Agent Systems.Richland:ACM,2018:577-584.
3 LEI D C, WU Y S .Order-dispatching strategy induced by optimal transport plan for an online ride-hailing system[J].Transportation Research Record20222676:156-169.
4 ODA T, CARLEE J W .MOVI:a model-free approach to dynamic fleet management[C]∥Proceedings of the IEEE INFOCOM 2018-IEEE Conference on Computer Communications.New York:IEEE,2018:2708-2716.
5 温惠英,林译峰,吴昊书,等 .基于城市道路交通环境演变的ECEA路径规划算法[J].华南理工大学学报(自然科学版)202149(10):1-10.
  WEN Huiying, LIN Yifeng, WU Haoshu,et al .Extended co-evolutionary algorithm for path planning based on the urban traffic environment evolution[J].Journal of South China University of Technology (Natural Science Edition)202149(10):1-10.
6 XU J, ROUHOLLAH R, LADISLAU B,et al .A taxi dispatch system based on prediction of demand and destination[J].Journal of Parallel and Distributed Computing2021157(11):269-279.
7 ZHANG L Y, TAO H, MIN Y,et al .A taxi order dispatch model based on combinatorial optimization[C]∥Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.New York:ACM,2017:2151-2159.
8 罗瑞发,郝慧君,徐桃让,等 .考虑时延的网联车混合交通流基本图模型[J].华南理工大学学报(自然科学版)202351(1):106-113.
  LUO Ruifa, HAO Huijun, XU Taorang,et al .Fundamental diagram model of mixed traffic flow of connected vehicles considering time delay[J].Journal of South China University of Technology(Natural Science Edition)202351(1):106-113.
9 FEI M, HAN S, LIN S,et al .Data-driven robust taxi dispatch under demand uncertainties[J].IEEE Transactions on Control Systems Technology201927(1):175-191.
10 SUTTON R, BARTO A .Reinforcement learning:an introduction[J].IEEE Transactions on Neural Networks201516:285-286.
11 WEN J, ZHAO J, JAILLET P .Rebalancing shared mobility-on-demand systems:a reinforcement learning approach[C]∥2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC).Yokohama:IEEE,2017:220-225.
12 LIU Z D, LI J Z, WU K S .Context-aware taxi dispatching at city-scale using deep reinforcement learning[J].IEEE Transactions on Intelligent Transportation Systems202223:1996-2009.
13 RONG H G, ZHOU X, YANG C,et al .The rich and the poor:a Markov decision process approach to optimizing taxi driver revenue efficiency[C]∥Proceedings of the 25th ACM International on Conference on Information and Knowledge Management.New York:ACM,2016:2329-2334.
14 FENG J K, GLUZMAN M O, DAI J G .Scalable deep reinforcement learning for ride-hailing[J].IEEE Control Systems Letters20215:2060-2065.
15 LIU Z D, LI J Z and WU K .Context-aware taxi dispatching at city-scale using deep reinforcement learning[J].IEEE Transactions on Intelligent Transportation Systems202023(3):1996-2009.
16 LIN K X, ZHAO R Y, XU Z,et al .Efficient large-scale fleet management via multi-agent deep reinforcement learning[C]∥Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.New York:ACM,2018:1774-1783.
17 SHOU Z Y, DI X, YE J P,et al .Optimal passenger-seeking policies on e-hailing platforms using aarkov decision process and imitation learning[J].Transportation Research Part C:Emerging Technologies2020111(2):91-113.
18 王福建,程慧玲,马东方,等 .基于深度逆向强化学习的城市车辆路径链重构[J].华南理工大学学报(自然科学版)202351(7):120-128.
  WANG Fujian, CHENG Huiling, MA Dongfang,et al .Urban vehicle path chain reconstruction based on deep inverse reinforcement learning[J].Journal of South China University of Technology (Natural Science Edition)202351(7):120-128.
19 ZHOU M, JIN J R, ZHANG W N,et al .Multi-agent reinforcement learning for order-dispatching via order-vehicle distribution matching[C]∥Proceedings of the 28th ACM International Conference on Information and Knowledge Management.New York:ACM,2019:2645-2653.
20 TAN M .Multi-agent reinforcement learning:independent vs.cooperative agents[C]∥Proceedings of the 10th International Conference on Machine Learning.Amherst:Morgan Kaufmann Publishers Inc,1993:330-337.
21 JIAO Y, TANG X C, QIN Z W,et al .Real-world ride-hailing vehicle repositioning using deep reinforcement learning[J].Transportation Research Part C:Emerging Technologies2021130:103289.
22 YANG Y D, LUO R, LI M,et al .Mean field multi-agent reinforcement learning[C]∥Proceedings of the 35th International Conference on Machine Learning.Stockholm:JMLR,201880:5571-5580.
23 LI M, QIN Z W, JIAO Y,et al .Efficient ridesharing order dispatching with mean field multi-agent reinforcement learning[C]∥Proceeding of the World Wide Web Conference.New York:ACM,2019:983-994.
24 ZHU Z, KE J, WANG H,et al .A mean-field markov decision process model for spatial-temporal subsidies in ride-sourcing markets[J].Transportation Research Part B:Methodological2021150:540-565.
25 SHOU Z Y, DI X .Reward design for driver repositioning using multi-agent reinforcement learning[J].Transportation Research Part C:Emerging Technologies2020119:102738.
26 LOWE R, WU Y, TAMAR A,et al .Multi-agent actor-critic for mixed cooperative competitive environments[C]∥Proceedings of the 31st International Conference on Neural Information Processing Systems.New York:Curran Associates Inc,2017:6382-6393.
27 JIN J, ZHOU M, ZHANG W N,et al .Coride:joint order dispatching and fleet management for multi-scale ride-hailing platforms[C]∥Proceedings of the 28th ACM International Conference on Information and Knowledge Management.New York:ACM,2019:1983-1992.
28 XU Z, LI Z X, GUAN Q W,et al .Large-scale order dispatch in on-demand ride-hailing platforms:a learning and planning approach[C]∥Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.New York:ACM,2018:905-913.
29 DAI H J, KHAKIK E, ZAHNG Y Y,et al .Learning combinatorial optimization algorithms over graphs[C]∥Proceedings of the 31st International Conference on Neural Information Processing Systems.New York:ACM,2017:6351-6361.
30 闫军威,黄琪,周璇 .基于Double-DQN的中央空调系统节能优化运行[J].华南理工大学学报(自然科学版)201947(1):135-144.
  YAN Junwei, HUANG Qi, ZHOU Xuan .Energy-saving optimization operation of central air-conditioning system based on double-DQN algorithm[J].Journal of South China University of Technology (Natural Science Edition)201947(1):135-144.
31 BELLO I, PHAM H, LE Q,et al .Neural combinatorial optimization with reinforcement learning[J].Learning201618:487-494.
32 YAN C W, ZHU H L, KOROLKO N,et al .Dynamic pricing and matching in ride-hailing platforms[J].Naval Research Logistics(NRL)202067(8):705-724.
33 ZHOU J, CUI G Q, HU S D,et al .Graph neural networks:a review of methods and applications[J].AI Open20201:57-81.
34 WANG Y, SUN Y B, LIU Z W,et al .Dynamic graph CNN for learning on point clouds[J].ACM Transactions on Graphics201938(5):1-12.
35 TOMAR M, EFRONI Y, GHAVAMZADEH M .Multi-step greedy reinforcement learning algorithms[C]∥Proceedings of the 37th International Conference on Machine Learning.Vienna:PMLR,2020:9504-9513.
36 YE D H, LIU Z, SUN M F,et al .Mastering complex control in MOBA games with deep reinforcement learning[C]∥Proceedings of the AAAI Conference on Artificial Intelligence.Menlo Park:AAAI,2020:6672-6667.
Outlines

/