SMAC-tuned Deep Q-learning for Ramp Metering

被引:0
|
作者
ElSamadisy, Omar [1 ,3 ]
Abdulhai, Yazeed [1 ]
Xue, Haoyuan [2 ]
Smirnov, Ilia [1 ]
Khalil, Elias B. [2 ]
Abdulhai, Baher [1 ]
机构
[1] Univ Toronto, Dept Civil Engn, Toronto, ON, Canada
[2] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON, Canada
[3] Arab Acad Sci Technol & Maritime Transport, Coll Engn & Technol, Dept Elect Commun Engn, Alexandria, Egypt
关键词
Ramp metering; Reinforcement learning; Hyperparameter tuning;
D O I
10.1109/SM57895.2023.10112246
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The demand for transportation increases as the population of a city grows, and significant expansion is not conceivable because of spatial, financial, and environmental limitations. As a result, improving infrastructure efficiency is becoming increasingly critical. Ramp metering with deep reinforcement learning (RL) is a method to tackle this problem. However, fine-tuning RL hyperparameters for RM is yet to be explored in the literature, potentially leaving performance improvements on the table. In this paper, the Sequential Model-based Algorithm Configuration (SMAC) method is used to finetune the values of two essential hyperparameters for the deep reinforcement learning ramp metering model, the discount factor and the decay of the explore/exploit ratio. Around 350 experiments with different configurations were run with PySMAC (a python interface to the hyperparameter optimization tool SMAC) and compared to Random search as a baseline. It is found that the best reward discount factor reflects that the RL agent should focus on immediate rewards and not pay much attention to future rewards. On the other hand, the selected value for the exploration ratio decay rate shows that the RL agent should prefer to decrease the exploration rate early. Both random search and SMAC show the same performance improvement of 19
引用
收藏
页码:65 / 72
页数:8
相关论文
共 50 条
  • [41] Deep Q-Learning in Robotics: Improvement of Accuracy and Repeatability
    Sumanas, Marius
    Petronis, Algirdas
    Bucinskas, Vytautas
    Dzedzickis, Andrius
    Virzonis, Darius
    Morkvenaite-Vilkonciene, Inga
    SENSORS, 2022, 22 (10)
  • [42] Deep Spatial Q-Learning for Infectious Disease Control
    Liu, Zhishuai
    Clifton, Jesse
    Laber, Eric B.
    Drake, John
    Fang, Ethan X.
    JOURNAL OF AGRICULTURAL BIOLOGICAL AND ENVIRONMENTAL STATISTICS, 2023, 28 (04) : 749 - 773
  • [43] Deep spectral Q-learning with application to mobile health
    Gao, Yuhe
    Shi, Chengchun
    Song, Rui
    STAT, 2023, 12 (01):
  • [44] Split Deep Q-Learning for Robust Object Singulation
    Sarantopoulos, Iason
    Kiatos, Marios
    Doulgeri, Zoe
    Malassiotis, Sotiris
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6225 - 6231
  • [45] Mixed-Policy Asynchronous Deep Q-Learning
    Simoes, David
    Lau, Nuno
    Reis, Luis Paulo
    ROBOT 2017: THIRD IBERIAN ROBOTICS CONFERENCE, VOL 2, 2018, 694 : 129 - 140
  • [46] Backward Q-learning: The combination of Sarsa algorithm and Q-learning
    Wang, Yin-Hao
    Li, Tzuu-Hseng S.
    Lin, Chih-Jui
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) : 2184 - 2193
  • [47] Learning Motion Policy for Mobile Robots using Deep Q-Learning
    Kwak, Nosan
    Yoon, Sukjune
    Roh, Kyungshik
    PROCEEDINGS 2017 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI), 2017, : 805 - 810
  • [48] Learning Robot Grasping from a Random Pile with Deep Q-Learning
    Chen, Bin
    Su, Jianhua
    Wang, Lili
    Gu, Qipeng
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2021, PT II, 2021, 13014 : 142 - 152
  • [49] Enhanced Machine Learning Algorithms: Deep Learning, Reinforcement Learning, ana Q-Learning
    Park, Ji Su
    Park, Jong Hyuk
    JOURNAL OF INFORMATION PROCESSING SYSTEMS, 2020, 16 (05): : 1001 - 1007
  • [50] Deep Q-Learning With Q-Matrix Transfer Learning for Novel Fire Evacuation Environment
    Sharma, Jivitesh
    Andersen, Per-Arne
    Granmo, Ole-Christoffer
    Goodwin, Morten
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (12): : 7363 - 7381