Deep Q-Learning-Based Dynamic Management of a Robotic Cluster

被引:4
|
作者
Gautier, Paul [1 ]
Laurent, Johann [1 ]
Diguet, Jean-Philippe [2 ]
机构
[1] Univ Bretagne Sud, Lab STICC, UMR6285 CNRS, F-56100 Lorient, France
[2] IRL2010 CNRS, CROSSING, Adelaide, SA 5000, Australia
关键词
Task analysis; Robots; Drones; Resource management; Computational modeling; Robot kinematics; Servers; MRS; task distribution; robotic cluster; multi-agent systems; reinforcement learning; deep Q-learning; ALLOCATION; SYSTEMS;
D O I
10.1109/TASE.2022.3205651
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The ever-increasing demands for autonomy and precision have led to the development of heavily computational multi-robot system (MRS). However, numerous missions exclude the use of robotic cloud. Another solution is to use the robotic cluster to locally distribute the computational load. This complex distribution requires adaptability to come up with a dynamic and uncertain environment. Classical approaches are too limited to solve this problem, but recent advances in reinforcement learning and deep learning offer new opportunities. In this paper we propose a new Deep Q-Network (DQN) based approaches where the MRS learns to distribute tasks directly from experience. Since the problem complexity leads to a curse of dimensionality, we use two specific methods, a new branching architecture, called Branching Dueling Q-Network (BDQ), and our own optimized multi-agent solution and we compare them with classical Market-based approaches as well as with non-distributed and purely local solutions. Our study shows the relevancy of learning-based methods for task mapping and also highlight the BDQ architecture capacity to solve high dimensional state space problems. Note to Practitioners-A lot of applications in industry like area exploration and monitoring can be efficiently delegated to a group of small-size robots or autonomous vehicles with advantages like reliability and cost in respect of single-robot solutions. But autonomy requires high and increasing compute-intensive tasks such as computer-vision. On the other hand small robots have energy constraints, limited embedded computing capacities and usually restricted and/or unreliable communications that limit the use of cloud resources. An alternative solution to cope with this problem consists in sharing the computing resources of the group of robots. Previous work was a proof of concept limited to the parallelisation of a single specific task. In this paper we formalize a general method that allows the group of robots to learn on the field how to efficiently distribute tasks in order to optimize the execution time of a mission under energy constraint. We demonstrate the relevancy of our solution over market-based and non-distributed approaches by means of intensive simulations. This successful study is a necessary first step towards distribution and parallelisation of computation tasks over a robotic cluster. The next steps, not tested yet, will address hardware in the loop simulation and finally a real-life mission with a group of robots.
引用
收藏
页码:2503 / 2515
页数:13
相关论文
共 50 条
  • [41] Reward Function Learning for Q-learning-Based Geographic Routing Protocol
    Jin, Weiqi
    Gu, Rentao
    Ji, Yuefeng
    IEEE COMMUNICATIONS LETTERS, 2019, 23 (07) : 1236 - 1239
  • [42] QRAVDR: A deep Q-learning-based RSU-Assisted Video Data Routing algorithm for VANETs
    Ma, Huahong
    Li, Shuangjin
    Wu, Honghai
    Xing, Ling
    Zhang, Xiaohui
    AD HOC NETWORKS, 2025, 171
  • [43] Deep Q-Learning-Based Transmission Power Control of a High Altitude Platform Station with Spectrum Sharing
    Jo, Seongjun
    Yang, Wooyeol
    Choi, Haing Kun
    Noh, Eonsu
    Jo, Han-Shin
    Park, Jaedon
    SENSORS, 2022, 22 (04)
  • [44] Deep Q-learning-based resource allocation for solar-powered users in cognitive radio networks
    Giang, Hoang Thi Huong
    Thanh, Pham Duy
    Koo, Insoo
    ICT EXPRESS, 2021, 7 (01): : 49 - 59
  • [45] Double Deep Q-Learning-Based Distributed Operation of Battery Energy Storage System Considering Uncertainties
    Bui, Van-Hai
    Hussain, Akhtar
    Kim, Hak-Man
    IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (01) : 457 - 469
  • [46] Q-Learning-Based Pesticide Contamination Prediction in Vegetables and Fruits
    Sellamuthu, Kandasamy
    Kaliappan, Vishnu Kumar
    COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2023, 45 (01): : 715 - 736
  • [47] A Multiagent Q-Learning-Based Optimal Allocation Approach for Urban Water Resource Management System
    Ni, Jianjun
    Liu, Minghua
    Ren, Li
    Yang, Simon X.
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2014, 11 (01) : 204 - 214
  • [48] Multiagent Q-Learning-Based Mobility Management for Multi-Connectivity in mmWAVE Cellular Systems
    Ryu, Si A.
    Kim, Duk Kyung
    SENSORS, 2023, 23 (17)
  • [49] Q-learning-based feedback linearization method for unknown dynamics
    Sun, Yipu
    Chen, Xin
    He, Wenpeng
    Wang, Luo
    Fukushima, Edwardo F.
    She, Jinhua
    2023 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS, ICM, 2023,
  • [50] Deep Q-Learning-Based Molecular Graph Generation for Chemical Structure Prediction From Infrared Spectra
    Ellis J.D.
    Iqbal R.
    Yoshimatsu K.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (02): : 634 - 646