Robust Imitation Learning from Noisy Demonstrations

被引:0
作者
Tangkaratt, Voot [1 ]
Charoenphakdee, Nontawat [1 ,2 ]
Sugiyama, Masashi [1 ,2 ]
机构
[1] RIKEN, Tokyo, Japan
[2] Univ Tokyo, Tokyo, Japan
来源
24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS) | 2021年 / 130卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Robust learning from noisy demonstrations is a practical but highly challenging problem in imitation learning. In this paper, we first theoretically show that robust imitation learning can be achieved by optimizing a classification risk with a symmetric loss. Based on this theoretical finding, we then propose a new imitation learning method that optimizes the classification risk by effectively combining pseudo-labeling with co-training. Unlike existing methods, our method does not require additional labels or strict assumptions about noise distributions. Experimental results on continuous-control benchmarks show that our method is more robust compared to state-of-the-art methods.
引用
收藏
页码:298 / +
页数:11
相关论文
共 35 条
[1]  
Angluin D., 1988, MACH LEARN
[2]  
[Anonymous], 2018, C ROB LEARN
[3]  
[Anonymous], 2017, ADV NEURAL INFORM PR
[4]  
[Anonymous], 2014, ADV NEURAL INFORM PR
[5]  
[Anonymous], 2016, ADV NEURAL INFORM PR
[6]  
[Anonymous], 2000, INT C MACH LEARN
[7]  
[Anonymous], 2017, ADV NEURAL INFORM PR
[8]  
[Anonymous], fi=Cfffifi$C>OS Improved training of Wasserstein GANs
[9]  
Blum A, 1998, INT C COMP LEARN THE
[10]  
Brodersen K. H., 2010, INT C PATTERN RECOGN