Spatial-Aware Deep Reinforcement Learning for the Traveling Officer Problem

被引:0
作者
Strauss, Niklas [1 ]
Schubert, Matthias [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, Munich Ctr Machine Learning, Munich, Germany
来源
PROCEEDINGS OF THE 2024 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM | 2024年
关键词
Reinforcement Learning; Deep Learning; Spatial Optimization; Traveling Officer Problem;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The traveling officer problem (TOP) is a challenging stochastic optimization task. In this problem, a parking officer is guided through a city equipped with parking sensors to fine as many parking offenders as possible. A major challenge in TOP is the dynamic nature of parking offenses, which randomly appear and disappear after some time, regardless of whether they have been fined. Thus, solutions need to dynamically adjust to currently fineable parking offenses while also planning ahead to increase the likelihood that the officer arrives during the offense taking place. Though various solutions exist, these methods often struggle to take the implications of actions on the ability to fine future parking violations into account. This paper proposes SATOP, a novel spatial-aware deep reinforcement learning approach for TOP. Our novel state encoder creates a representation of each action, leveraging the spatial relationships between parking spots, the agent, and the action. Furthermore, we propose a novel message-passing module for learning future inter-action correlations in the given environment. Thus, the agent can estimate the potential to fine further parking violations after executing an action. We evaluate our method using an environment based on real-world data from Melbourne. Our results show that SATOP consistently outperforms state-of-the-art TOP agents and is able to fine up to 22% more parking offenses.
引用
收藏
页码:869 / 877
页数:9
相关论文
共 22 条
[1]  
Ba JL, 2016, arXiv
[2]  
Bello I., 2017, arXiv, DOI DOI 10.48550/ARXIV.1611.09940
[3]   Solving Multi-Agent Routing Problems Using Deep Attention Mechanisms [J].
Bono, Guillaume ;
Dibangoye, Jilles S. ;
Simonin, Olivier ;
Matignon, Laetitia ;
Pereyron, Florian .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (12) :7804-7813
[4]  
Clevert DA, 2016, Arxiv, DOI arXiv:1511.07289
[5]   Stochastic vehicle routing [J].
Gendreau, M ;
Laporte, G ;
Seguin, R .
EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 1996, 88 (01) :3-12
[6]  
Haarnoja T, 2018, PR MACH LEARN RES, V80
[7]   Identity Mappings in Deep Residual Networks [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :630-645
[8]   Heterogeneous Pointer Network for Travelling Officer Problem [J].
He, Rongguang ;
Xiao, Xiao ;
Kang, Yufan ;
Zhao, Hongyu ;
Shao, Wei .
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
[9]   Optimizing Large-Scale Fleet Management on a Road Network using Multi-Agent Deep Reinforcement Learning with Graph Neural Network [J].
Kim, Juhyeon ;
Kim, Kihyun .
2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, :990-995
[10]  
Kool W, 2019, Arxiv, DOI [arXiv:1803.08475, 10.48550/arXiv.1803.08475]