Intelligent Cruise Guidance and Vehicle Resource Management With Deep Reinforcement Learning

被引:8
作者
Sun, Guolin [1 ]
Liu, Kai [1 ]
Boateng, Gordon Owusu [1 ]
Liu, Guisong [2 ]
Jiang, Wei [3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
[2] Southwestern Univ Finance & Econ, Sch Econ Informat Engn, Chengdu 611130, Peoples R China
[3] German Res Ctr Artificial Intelligence DFKI GmbH, Sch Comp Sci & Engn, D-67663 Kaiserslautern, Germany
关键词
Public transportation; Resource management; Adaptation models; Urban areas; Vehicles; Real-time systems; Dispatching; Cruise guidance; deep reinforcement learning (DRL); resource management; transportation network companies (TNCs);
D O I
10.1109/JIOT.2021.3098779
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The emergence of new business and technological models for urban-related transportation has revealed the need for transportation network companies (TNCs). Most research works on TNCs optimize the interests of drivers and passengers, and the operator assuming vehicle resources remain unchanged, but ignore the optimization of resource utilization and satisfaction from the perspective of flexible and controllable vehicle resources. In fact, the load of the scene is variable in time, which necessitates the flexible control of resources. Drivers wish to effectively utilize their vehicle resources to maximize profits. Passengers desire to spend minimum time waiting and the platform cares about the commission they can accrue from successful trips. In this article, we propose an adaptive intelligent cruise guidance and vehicle resource management model to balance vehicle resource utilization and request success rate, while improving platform revenue. We propose an advanced deep reinforcement learning (DRL) method to autonomously learn the statuses and guide the vehicles to hotspot areas where they can pick orders. We assume the number of online vehicles in the scene is flexible and the learning agent can autonomously change the number of online vehicles in the system according to the real-time load to improve effective vehicle resource utilization. An adaptive reward mechanism is enforced to control the importance of vehicle resource utilization and request success rate at decision steps. The simulation results and analysis reveal that our proposed DRL-based scheme balances vehicle resource utilization and request success rate at acceptable levels while improving the platform revenue, compared with other baseline algorithms.
引用
收藏
页码:3574 / 3585
页数:12
相关论文
共 26 条
[1]   DeepPool: Distributed Model-Free Algorithm for Ride-Sharing Using Deep Reinforcement Learning [J].
Al-Abbasi, Abubakr O. ;
Ghosh, Arnob ;
Aggarwal, Vaneet .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (12) :4714-4727
[2]  
Busoniu L, 2010, AUTOM CONTROL ENG SE, P1, DOI 10.1201/9781439821091-f
[3]  
Csiszár C, 2015, 2015 SMART CITIES SYMPOSIUM PRAGUE (SCSP)
[4]   Optimize taxi driving strategies based on reinforcement learning [J].
Gao, Yong ;
Jiang, Dan ;
Xu, Yan .
INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE, 2018, 32 (08) :1677-1696
[5]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P163
[6]   Spatio-Temporal Capsule-Based Reinforcement Learning for Mobility-on-Demand Coordination [J].
He, Suining ;
Shin, Kang G. .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (03) :1446-1461
[7]   An effective taxi recommender system based on a spatio-temporal factor analysis model [J].
Hwang, Ren-Hung ;
Hsueh, Yu-Ling ;
Chen, Yu-Ting .
INFORMATION SCIENCES, 2015, 314 :28-40
[8]   Satisfaction and uncertainty in car-sharing decisions: An integration of hybrid choice and random regret-based models [J].
Kim, Jinhee ;
Rasouli, Soora ;
Timmermans, Harry .
TRANSPORTATION RESEARCH PART A-POLICY AND PRACTICE, 2017, 95 :13-33
[9]   Urban Traffic Coulomb's Law: A New Approach for Taxi Route Recommendation [J].
Lai, Yongxuan ;
Lv, Zheng ;
Li, Kuan-Ching ;
Liao, Minghong .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (08) :3024-3037
[10]  
Li HJ, 2017, ICCAD-IEEE ACM INT, P847, DOI 10.1109/ICCAD.2017.8203866