Deep Q-Learning-Based Dynamic Management of a Robotic Cluster

被引:4
|
作者
Gautier, Paul [1 ]
Laurent, Johann [1 ]
Diguet, Jean-Philippe [2 ]
机构
[1] Univ Bretagne Sud, Lab STICC, UMR6285 CNRS, F-56100 Lorient, France
[2] IRL2010 CNRS, CROSSING, Adelaide, SA 5000, Australia
关键词
Task analysis; Robots; Drones; Resource management; Computational modeling; Robot kinematics; Servers; MRS; task distribution; robotic cluster; multi-agent systems; reinforcement learning; deep Q-learning; ALLOCATION; SYSTEMS;
D O I
10.1109/TASE.2022.3205651
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The ever-increasing demands for autonomy and precision have led to the development of heavily computational multi-robot system (MRS). However, numerous missions exclude the use of robotic cloud. Another solution is to use the robotic cluster to locally distribute the computational load. This complex distribution requires adaptability to come up with a dynamic and uncertain environment. Classical approaches are too limited to solve this problem, but recent advances in reinforcement learning and deep learning offer new opportunities. In this paper we propose a new Deep Q-Network (DQN) based approaches where the MRS learns to distribute tasks directly from experience. Since the problem complexity leads to a curse of dimensionality, we use two specific methods, a new branching architecture, called Branching Dueling Q-Network (BDQ), and our own optimized multi-agent solution and we compare them with classical Market-based approaches as well as with non-distributed and purely local solutions. Our study shows the relevancy of learning-based methods for task mapping and also highlight the BDQ architecture capacity to solve high dimensional state space problems. Note to Practitioners-A lot of applications in industry like area exploration and monitoring can be efficiently delegated to a group of small-size robots or autonomous vehicles with advantages like reliability and cost in respect of single-robot solutions. But autonomy requires high and increasing compute-intensive tasks such as computer-vision. On the other hand small robots have energy constraints, limited embedded computing capacities and usually restricted and/or unreliable communications that limit the use of cloud resources. An alternative solution to cope with this problem consists in sharing the computing resources of the group of robots. Previous work was a proof of concept limited to the parallelisation of a single specific task. In this paper we formalize a general method that allows the group of robots to learn on the field how to efficiently distribute tasks in order to optimize the execution time of a mission under energy constraint. We demonstrate the relevancy of our solution over market-based and non-distributed approaches by means of intensive simulations. This successful study is a necessary first step towards distribution and parallelisation of computation tasks over a robotic cluster. The next steps, not tested yet, will address hardware in the loop simulation and finally a real-life mission with a group of robots.
引用
收藏
页码:2503 / 2515
页数:13
相关论文
共 50 条
  • [1] Deep Q-Learning-Based Dynamic Network Slicing and Task Offloading in Edge Network
    Chiang, Yao
    Hsu, Chih-Ho
    Chen, Guan-Hao
    Wei, Hung-Yu
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2023, 20 (01): : 369 - 384
  • [2] Deep Q-Learning-Based Resource Management in IRS-Assisted VLC Systems
    Hammadi, Ahmed Al
    Bariah, Lina
    Muhaidat, Sami
    Al-Qutayri, Mahmoud
    Sofotasios, Paschalis C.
    Debbah, Merouane
    IEEE Transactions on Machine Learning in Communications and Networking, 2024, 2 : 34 - 48
  • [3] QRED: A Q-Learning-based Active Queue Management Scheme
    Su, Yuhan
    Huang, Lianfen
    Feng, Chenwei
    JOURNAL OF INTERNET TECHNOLOGY, 2018, 19 (04): : 1169 - 1178
  • [4] A Q-Learning-Based Approach for Deploying Dynamic Service Function Chains
    Sun, Jian
    Huang, Guanhua
    Sun, Gang
    Yu, Hongfang
    Sangaiah, Arun Kumar
    Chang, Victor
    SYMMETRY-BASEL, 2018, 10 (11):
  • [5] Deep Federated Q-Learning-Based Network Slicing for Industrial IoT
    Messaoud, Seifeddine
    Bradai, Abbas
    Ben Ahmed, Olfa
    Pham Tran Anh Quang
    Atri, Mohamed
    Hossain, M. Shamim
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (08) : 5572 - 5582
  • [6] Dynamic Q-Learning-Based Optimized Load Balancing Technique in Cloud
    Muthusamy, Arvindhan
    Dhanaraj, Rajesh Kumar
    Mobile Information Systems, 2023, 2023
  • [7] Q-learning-based algorithms for dynamic transmission control in IoT equipment
    Malekijou, Hanieh
    Hakami, Vesal
    Javan, Nastooh Taheri
    Malekijoo, Amirhossein
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (01): : 75 - 108
  • [8] Q-learning-based navigation for mobile robots in continuous and dynamic environments
    Maoudj, Abderraouf
    Christensen, Anders Lyhne
    2021 IEEE 17TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2021, : 1338 - 1345
  • [9] Double-deep Q-learning-based handover management in mmWave heterogeneous networks with dual connectivity
    Wang, Hao
    Li, Bo
    TRANSACTIONS ON EMERGING TELECOMMUNICATIONS TECHNOLOGIES, 2024, 35 (01)
  • [10] Deep Q-Learning-Based Optimization of Path Planning and Control in Robotic Arms for High-Precision Computational Efficiency
    Li, Yuan
    Min, Byung-Won
    Liu, Haozhi
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2025, 16 (01) : 1199 - 1207