SMAC-tuned Deep Q-learning for Ramp Metering

被引:0
|
作者
ElSamadisy, Omar [1 ,3 ]
Abdulhai, Yazeed [1 ]
Xue, Haoyuan [2 ]
Smirnov, Ilia [1 ]
Khalil, Elias B. [2 ]
Abdulhai, Baher [1 ]
机构
[1] Univ Toronto, Dept Civil Engn, Toronto, ON, Canada
[2] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON, Canada
[3] Arab Acad Sci Technol & Maritime Transport, Coll Engn & Technol, Dept Elect Commun Engn, Alexandria, Egypt
关键词
Ramp metering; Reinforcement learning; Hyperparameter tuning;
D O I
10.1109/SM57895.2023.10112246
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The demand for transportation increases as the population of a city grows, and significant expansion is not conceivable because of spatial, financial, and environmental limitations. As a result, improving infrastructure efficiency is becoming increasingly critical. Ramp metering with deep reinforcement learning (RL) is a method to tackle this problem. However, fine-tuning RL hyperparameters for RM is yet to be explored in the literature, potentially leaving performance improvements on the table. In this paper, the Sequential Model-based Algorithm Configuration (SMAC) method is used to finetune the values of two essential hyperparameters for the deep reinforcement learning ramp metering model, the discount factor and the decay of the explore/exploit ratio. Around 350 experiments with different configurations were run with PySMAC (a python interface to the hyperparameter optimization tool SMAC) and compared to Random search as a baseline. It is found that the best reward discount factor reflects that the RL agent should focus on immediate rewards and not pay much attention to future rewards. On the other hand, the selected value for the exploration ratio decay rate shows that the RL agent should prefer to decrease the exploration rate early. Both random search and SMAC show the same performance improvement of 19
引用
收藏
页码:65 / 72
页数:8
相关论文
共 50 条
  • [31] A Deep Reinforcement Learning Approach for Ramp Metering Based on Traffic Video Data
    Liu, Bing
    Tang, Yu
    Ji, Yuxiong
    Shen, Yu
    Du, Yuchuan
    JOURNAL OF ADVANCED TRANSPORTATION, 2021, 2021
  • [32] Q-LEARNING
    WATKINS, CJCH
    DAYAN, P
    MACHINE LEARNING, 1992, 8 (3-4) : 279 - 292
  • [33] A Deep Q-Learning Dynamic Spectrum Sharing Experiment
    Shea, John M.
    Wong, Tan F.
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [34] Hyperparameter Optimization for Tracking with Continuous Deep Q-Learning
    Dong, Xingping
    Shen, Jianbing
    Wang, Wenguan
    Liu, Yu
    Shao, Ling
    Porikli, Fatih
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 518 - 527
  • [35] Deep Spatial Q-Learning for Infectious Disease Control
    Zhishuai Liu
    Jesse Clifton
    Eric B. Laber
    John Drake
    Ethan X. Fang
    Journal of Agricultural, Biological and Environmental Statistics, 2023, 28 : 749 - 773
  • [36] A Deep Q-Learning Approach for GPU Task Scheduling
    Luley, Ryan S.
    Qiu, Qinru
    2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [37] Stabilizing deep Q-learning with Q-graph-based bounds
    Hoppe, Sabrina
    Giftthaler, Markus
    Krug, Robert
    Toussaint, Marc
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2023, 42 (09): : 633 - 654
  • [38] Autonomous Warehouse Robot using Deep Q-Learning
    Peyas, Ismot Sadik
    Hasan, Zahid
    Tushar, Md Rafat Rahman
    Al Musabbir
    Azni, Raisa Mehjabin
    Siddique, Shahnewaz
    2021 IEEE REGION 10 CONFERENCE (TENCON 2021), 2021, : 857 - 862
  • [39] Deep Q-Learning for Dry Stacking Irregular Objects
    Liu, Yifang
    Shamsi, Seyed Mahdi
    Fang, Le
    Chen, Changyou
    Napp, Nils
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 1569 - 1576
  • [40] Training a dynamically configurable classifier with deep Q-learning
    Malashin, R. O.
    Boiko, A. A.
    JOURNAL OF OPTICAL TECHNOLOGY, 2022, 89 (08) : 437 - 447