Visual Exploration and Analysis of Human-Robot Interaction Rules

被引:3
作者
Zhang, Hui [1 ]
Boyles, Michael J. [1 ]
机构
[1] Indiana Univ, Pervas Technol Inst, Bloomington, IN 47405 USA
来源
VISUALIZATION AND DATA ANALYSIS 2013 | 2013年 / 8654卷
关键词
Information Visualization; Visual Data Mining; Joint Attention; Human-Robot Interaction;
D O I
10.1117/12.2002536
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a novel interaction paradigm for the visual exploration, manipulation and analysis of human-robot interaction (HRI) rules; our development is implemented using a visual programming interface and exploits key techniques drawn from both information visualization and visual data mining to facilitate the interaction design and knowledge discovery process. HRI is often concerned with manipulations of multi-modal signals, events, and commands that form various kinds of interaction rules. Depicting, manipulating and sharing such design-level information is a compelling challenge. Furthermore, the closed loop between HRI programming and knowledge discovery from empirical data is a relatively long cycle. This, in turn, makes design-level verification nearly impossible to perform in an earlier phase. In our work, we exploit a drag-and-drop user interface and visual languages to support depicting responsive behaviors from social participants when they interact with their partners. For our principal test case of gaze-contingent HRI interfaces, this permits us to program and debug the robots' responsive behaviors through a graphical data-flow chart editor. We exploit additional program manipulation interfaces to provide still further improvement to our programming experience: by simulating the interaction dynamics between a human and a robot behavior model, we allow the researchers to generate, trace and study the perception-action dynamics with a social interaction simulation to verify and refine their designs. Finally, we extend our visual manipulation environment with a visual data-mining tool that allows the user to investigate interesting phenomena such as joint attention and sequential behavioral patterns from multiple multi-modal data streams. We have created instances of HRI interfaces to evaluate and refine our development paradigm. As far as we are aware, this paper reports the first program manipulation paradigm that integrates visual programming interfaces, information visualization, and visual data mining methods to facilitate designing, comprehending, and evaluating HRI interfaces.
引用
收藏
页数:14
相关论文
共 16 条
[1]   Infant-like social interactions between a robot and a human caregiver. [J].
Breazeal, C ;
Scassellati, B .
ADAPTIVE BEHAVIOR, 2000, 8 (01) :49-74
[2]  
Cassell J., 2007, Genesis Redux : Essays in the History and Philosophy of Artificial Intelligence, P346
[3]  
Diprose JP, 2011, S VIS LANG HUM CEN C, P229, DOI 10.1109/VLHCC.2011.6070406
[4]  
Fricker D., 2011, Development and Learning (ICDL), 2011 IEEE International Conference on, V2, P1, DOI DOI 10.1109/DEVLRN.2011.6037334
[5]  
Guyet Thomas, 2008, 2008 IEEE International Conference on Data Mining Workshops, P218, DOI 10.1109/ICDMW.2008.16
[6]  
Hosoda K., 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566), P2867
[7]  
Kipp M, 2008, LECT NOTES COMPUT SC, V5208, P191
[8]  
Li Jimmy, 2011, 2011 IEEE International Conference on Robotics and Automation, P4846
[9]  
Nakano YI, 2010, IUI 2010, P139
[10]   Tactile programming: A unified manipulation paradigm supporting program comprehension, composition and sharing [J].
Repenning, A ;
Ambach, J .
IEEE SYMPOSIUM ON VISUAL LANGUAGES, PROCEEDINGS, 1996, :102-109