Deep Reinforcement Learning Enables Joint Trajectory and Communication in Internet of Robotic Things

被引:0
作者
Luo, Ruyu [1 ]
Tian, Hui [1 ]
Ni, Wanli [2 ]
Cheng, Julian [3 ]
Chen, Kwang-Cheng [4 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[2] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[3] Univ British Columbia, Sch Engn, Kelowna, BC V1V 1V7, Canada
[4] Univ S Florida, Dept Elect Engn, Tampa, FL 33620 USA
基金
中国国家自然科学基金;
关键词
Ultra reliable low latency communication; Trajectory; Resource management; Robots; NOMA; Wireless communication; Decoding; Deep reinforcement learning; Internet of Robotic Things; trajectory design; ultra-reliable low-latency communications; RESOURCE-ALLOCATION; URLLC; NOMA; OPTIMIZATION; CAPACITY;
D O I
10.1109/TWC.2024.3462450
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Internet of Robotic Things (IoRT) emphasizes the integrated robotic, artificial intelligence computing, and communication technologies, enabling more sophisticated operations and decision-making. As a crucial element of IoRT, mission-critical applications, such as industrial manufacturing and emergency services, impose stringent requirements on ultra-reliable and low-latency communication (URLLC). The paper focuses on addressing URLLC challenges in the context of IoRT, particularly when autonomous mobile robots (AMRs) coexist with static sensors. We prioritize safe and efficient AMRs' travel through trajectory design and communication resource allocation in IoRT systems without the need of any prior knowledge. To enhance network connectivity and exploit diversity gains, we introduce the flexible decoding and free clustering as the next-generation multiple access technologies in spectrum-limited downlink IoRT system. Then, aiming at minimizing the decoding error probability and travel time, we formulate a long-term multi-objective optimization problem by jointly designing AMRs' trajectory and communication resource. To accommodate the inherent dynamics and unpredictability in the IoRT system, we introduce a multi-agent actor-critic deep reinforcement learning (DRL) framework, offering four distinct implementations, each accompanied by comprehensive complexity analyses. Simulation results reveal the following insights: 1) in terms of DRL implementations, off-policy algorithms with deterministic policies outperform their on-policy counterparts, achieving approximately a 67% increase in rewards; 2) In terms of communication schemes, our proposed flexible decoding and free clustering strategies under designed trajectories can effectively reduce decoding errors; and 3) In terms of algorithm optimality, our DRL framework shows superior flexibility and adaptability in communication environments compared to traditional A* search and heuristic methods.
引用
收藏
页码:18154 / 18168
页数:15
相关论文
共 56 条
[11]  
Foerster JN, 2018, AAAI CONF ARTIF INTE, P2974
[12]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[13]   Intelligent Trajectory Design for RIS-NOMA Aided Multi-Robot Communications [J].
Gao, Xinyu ;
Mu, Xidong ;
Yi, Wenqiang ;
Liu, Yuanwei .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2023, 22 (11) :7648-7662
[14]   Trajectory and Passive Beamforming Design for IRS-aided Multi-Robot NOMA Indoor Networks [J].
Gao, Xinyu ;
Liu, Yuanwei ;
Mu, Xidong .
IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
[15]   Toward 6G Networks: Use Cases and Technologies [J].
Giordani, Marco ;
Polese, Michele ;
Mezzavilla, Marco ;
Rangan, Sundeep ;
Zorzi, Michele .
IEEE COMMUNICATIONS MAGAZINE, 2020, 58 (03) :55-61
[16]   A Review of Motion Planning Techniques for Automated Vehicles [J].
Gonzalez, David ;
Perez, Joshue ;
Milanes, Vicente ;
Nashashibi, Fawzi .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2016, 17 (04) :1135-1145
[17]   Deep Reinforcement Learning for Practical Phase-Shift Optimization in RIS-Aided MISO URLLC Systems [J].
Hashemi, Ramin ;
Ali, Samad ;
Mahmood, Nurul Huda ;
Latva-Aho, Matti .
IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (10) :8931-8943
[18]   Voronoi-Based Multi-Robot Autonomous Exploration in Unknown Environments via Deep Reinforcement Learning [J].
Hu, Junyan ;
Niu, Hanlin ;
Carrasco, Joaquin ;
Lennox, Barry ;
Arvin, Farshad .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (12) :14413-14423
[19]  
Jayant AK, 2022, ADV NEUR IN
[20]   Trajectory Generation of a Two-Wheeled Mobile Robot in an Uncertain Environment [J].
Kim, Joonyoung .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2020, 67 (07) :5586-5594