Learning from Unscripted Deictic Gesture and Language for Human-Robot Interactions

被引:0
作者
Matuszek, Cynthia [1 ]
Bo, Liefeng [1 ]
Zettlemoyer, Luke [1 ]
Fox, Dieter [1 ]
机构
[1] Univ Washington, Comp Sci & Engn, Box 352350, Seattle, WA 98195 USA
来源
PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE | 2014年
关键词
RECOGNITION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As robots become more ubiquitous, it is increasingly important for untrained users to be able to interact with them intuitively. In this work, we investigate how people refer to objects in the world during relatively unstructured communication with robots. We collect a corpus of deictic interactions from users describing objects, which we use to train language and gesture models that allow our robot to determine what objects arc being indicated. We introduce a temporal extension to state-of-the-art hierarchical matching pursuit features to support gesture understanding, and demonstrate that combining multiple communication modalities more effectively capture user intent than relying on a single type of input. Finally, we present initial interactions with a robot that uses the learned models to follow commands.
引用
收藏
页码:2556 / 2563
页数:8
相关论文
共 29 条
[1]  
[Anonymous], P 7 ANN ACM IEEE INT
[2]  
[Anonymous], P 2012 INT C MACH LE
[3]  
[Anonymous], 2012, Experiments in Cultural Language Evolution
[4]  
[Anonymous], MACHINE LEARNING J L
[5]  
[Anonymous], 2011, ADV NEURAL INFORM PR
[6]  
[Anonymous], J CHILD LANGUAGE
[7]   Coordinating with each other in a material world [J].
Clark, HH .
DISCOURSE STUDIES, 2005, 7 (4-5) :507-525
[8]  
Clark HH, 2003, POINTING: WHERE LANGAUAGE, CULTURE, AND COGNITON MEET, P243
[9]  
Farhadi A, 2009, PROC CVPR IEEE, P1778, DOI 10.1109/CVPRW.2009.5206772
[10]  
FitzGerald N., 2013, Empirical Methods on Natural Language Processing (EMNLP)