Causal deconfounding deep reinforcement learning for mobile robot motion planning

被引:1
作者
Tang, Wenbing [1 ,2 ]
Wu, Fenghua [2 ]
Lin, Shang-wei [2 ]
Ding, Zuohua [3 ]
Liu, Jing [1 ]
Liu, Yang [2 ]
He, Jifeng [1 ]
机构
[1] East China Normal Univ, Shanghai Key Lab Trustworthy Comp, Shanghai 200062, Peoples R China
[2] Nanyang Technol Univ, Coll Comp & Data Sci, Singapore 639798, Singapore
[3] Zhejiang Sci Tech Univ, Sch Comp Sci & Technol, Hangzhou 310018, Peoples R China
关键词
Backdoor paths; Causal inference; Deep reinforcement learning; Mobile robots; Motion planning; MODEL;
D O I
10.1016/j.knosys.2024.112406
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) has emerged as an efficient approach for motion planning in mobile robot systems. It leverages the offline training process to enhance real-time computation efficiency. In DRLbased methods, the DRL models are trained to compute an action based on the current state of the robot and the surrounding obstacles. However, the trained models may capture spurious correlations through potential confounders, resulting in non-robust state representations, which limits the models' robustness and generalizability. In this paper, we propose a Causal Deconfounding DRL method for Motion Planning, CD-DRL-MP, to address spurious correlations and learn robust and generalizable policies. Specifically, we formalize the temporal causal relationships between states and actions using a structural causal model. We then extract the minimal sufficient state representation set by blocking the backdoor paths in the causal model. Finally, using the representation set, CD-DRL-MP learns the causal effect between states and actions while mitigating the detrimental influence of potential confounders and computes motion commands for mobile robots. Comprehensive experiments show that the proposed method significantly outperforms non-causal DRL methods and existing causal methods, while guaranteeing good robustness and generalizability.
引用
收藏
页数:12
相关论文
共 44 条
[1]   Multiple Mobile Robot Task and Motion Planning: A Survey [J].
Antonyshyn, Luke ;
Silveira, Jefferson ;
Givigi, Sidney ;
Marshall, Joshua .
ACM COMPUTING SURVEYS, 2023, 55 (10)
[2]  
Bica I, 2021, ADV NEUR IN, V34
[3]   Optimal Probabilistic Motion Planning With Potential Infeasible LTL Constraints [J].
Cai, Mingyu ;
Xiao, Shaoping ;
Li, Zhijun ;
Kan, Zhen .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (01) :301-316
[4]  
Chen CG, 2019, IEEE INT CONF ROBOT, P6015, DOI [10.1109/icra.2019.8794134, 10.1109/ICRA.2019.8794134]
[5]   Conditional DQN-Based Motion Planning With Fuzzy Logic for Autonomous Driving [J].
Chen, Long ;
Hu, Xuemin ;
Tang, Bo ;
Cheng, Yu .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (04) :2966-2977
[6]   CDRM: Causal disentangled representation learning for missing data [J].
Chen, Mingjie ;
Wang, Hongcheng ;
Wang, Ruxin ;
Peng, Yuzhong ;
Zhang, Hao .
KNOWLEDGE-BASED SYSTEMS, 2024, 299
[7]  
De Haan P., 2019, ADV NEUR IN, V32
[8]   Causality-based counterfactual explanation for classification models [J].
Duong, Tri Dung ;
Li, Qian ;
Xu, Guandong .
KNOWLEDGE-BASED SYSTEMS, 2024, 300
[9]  
Everett M, 2018, IEEE INT C INT ROBOT, P3052, DOI 10.1109/IROS.2018.8593871
[10]  
Furrer F, 2016, STUD COMPUT INTELL, V625, P595, DOI 10.1007/978-3-319-26054-9_23