Exploiting Transformer in Sparse Reward Reinforcement Learning for Interpretable Temporal Logic Motion Planning

被引:4
|
作者
Zhang, Hao [1 ]
Wang, Hao [1 ]
Kan, Zhen [1 ]
机构
[1] Univ Sci & Technol China, Dept Automat, Hefei 230026, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Transformers; Robots; Reinforcement learning; Planning; Learning automata; Encoding; Linear temporal logic; motion planning; reinforcement learning;
D O I
10.1109/LRA.2023.3290511
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Automaton based approaches have enabled robots to perform various complex tasks. However, most existing automaton based algorithms highly rely on the manually customized representation of states for the considered task, limiting its applicability in deep reinforcement learning algorithms. To address this issue, by incorporating Transformer into reinforcement learning, we develop a Double-Transformer-guided Temporal Logic framework (T2TL) that exploits the structural feature of Transformer twice, i.e., first encoding the LTL instruction via the Transformer module for efficient understanding of task instructions during the training and then encoding the context variable via the Transformer again for improved task performance. Particularly, the LTL instruction is specified by co-safe LTL. As a semantics-preserving rewriting operation, LTL progression is exploited to decompose the complex task into learnable sub-goals, which not only converts non-Markovian reward decision processes to Markovian ones, but also improves the sampling efficiency by simultaneous learning of multiple sub-tasks. An environment-agnostic LTL pre-training scheme is further incorporated to facilitate the learning of the Transformer module resulting in an improved representation of LTL. The simulation results demonstrate the effectiveness of the T2TL framework.
引用
收藏
页码:4831 / 4838
页数:8
相关论文
共 50 条
  • [21] A review of mobile robot motion planning methods: from classical motion planning workflows to reinforcement learning-based architectures
    Dong, Lu
    He, Zichen
    Song, Chunwei
    Sun, Changyin
    JOURNAL OF SYSTEMS ENGINEERING AND ELECTRONICS, 2023, 34 (02) : 439 - 459
  • [22] HMRL: Hyper-Meta Learning for Sparse Reward Reinforcement Learning Problem
    Hua, Yun
    Wang, Xiangfeng
    Jin, Bo
    Li, Wenhao
    Yan, Junchi
    He, Xiaofeng
    Zha, Hongyuan
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 637 - 645
  • [23] Reinforcement learning techniques applied to the motion planning of a robotic manipulator
    Ribeiro, Francisco M.
    Pinto, Vitor H.
    2022 IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC), 2022, : 173 - 178
  • [24] Energy-Efficient Reinforcement Learning for Motion Planning of AUV
    Wen, Jiayi
    Zhu, Jingwei
    Lin, Yejin
    Zhang, Guichen
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON UNDERWATER SYSTEM TECHNOLOGY: THEORY AND APPLICATIONS, USYS, 2022,
  • [25] Policy-based deep reinforcement learning for sparse reward environment
    Kim M.
    Kim J.-S.
    Transactions of the Korean Institute of Electrical Engineers, 2021, 70 (03): : 506 - 514
  • [26] Sparse reward for reinforcement learning-based continuous integration testing
    Yang, Yang
    Li, Zheng
    Shang, Ying
    Li, Qianyu
    JOURNAL OF SOFTWARE-EVOLUTION AND PROCESS, 2023, 35 (06)
  • [27] Reinforcement Learning for Multi-Agent Systems with Temporal Logic Specifications
    Terashima, Keita
    Kobayashi, Koichi
    Yamashita, Yuh
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2024, E107A (01) : 31 - 37
  • [28] Safe Multiagent Motion Planning Under Uncertainty for Drones Using Filtered Reinforcement Learning
    Safaoui, Sleiman
    Vinod, Abraham P.
    Chakrabarty, Ankush
    Quirynen, Rien
    Yoshikawa, Nobuyuki
    Di Cairano, Stefano
    IEEE TRANSACTIONS ON ROBOTICS, 2024, 40 (2529-2542) : 2529 - 2542
  • [29] Comfort-Oriented Motion Planning for Automated Vehicles Using Deep Reinforcement Learning
    Rajesh, Nishant
    Zheng, Yanggu
    Shyrokau, Barys
    IEEE OPEN JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 4 : 348 - 359
  • [30] Data-Driven Motion Planning: A Survey on Deep Neural Networks, Reinforcement Learning, and Large Language Model Approaches
    de Carvalho, Gabriel Peixoto
    Sawanobori, Tetsuya
    Horii, Takato
    IEEE ACCESS, 2025, 13 : 52195 - 52245