Human-robot dialogue annotation for multi-modal common ground

被引:0
|
作者
Bonial, Claire [1 ]
Lukin, Stephanie M. [1 ]
Abrams, Mitchell [1 ]
Baker, Anthony [1 ]
Donatelli, Lucia [2 ]
Foots, Ashley [1 ]
Hayes, Cory J. [1 ]
Henry, Cassidy [1 ]
Hudson, Taylor [3 ]
Marge, Matthew [4 ]
Pollard, Kimberly A. [1 ]
Artstein, Ron [5 ]
Traum, David [5 ]
Voss, Clare R. [1 ]
机构
[1] DEVCOM Army Res Lab, Adelphi, MD 21005 USA
[2] Vrije Univ, Amsterdam, Netherlands
[3] Oak Ridge Associated Univ, Oak Ridge, TN USA
[4] DARPA, Arlington, VA USA
[5] USC Inst Creat Technol, Playa Vista, CA USA
关键词
Situated dialogue; Semantics; Multi-floor dialogue; Multi-modal dialogue;
D O I
10.1007/s10579-024-09784-2
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In this paper, we describe the development of symbolic representations annotated on human-robot dialogue data to make dimensions of meaning accessible to autonomous systems participating in collaborative, natural language dialogue, and to enable common ground with human partners. A particular challenge for establishing common ground arises in remote dialogue (occurring in disaster relief or search-and-rescue tasks), where a human and robot are engaged in a joint navigation and exploration task of an unfamiliar environment, but where the robot cannot immediately share high quality visual information due to limited communication constraints. Engaging in a dialogue provides an effective way to communicate, while on-demand or lower-quality visual information can be supplemented for establishing common ground. Within this paradigm, we capture propositional semantics and the illocutionary force of a single utterance within the dialogue through our Dialogue-AMR annotation, an augmentation of Abstract Meaning Representation. We then capture patterns in how different utterances within and across speaker floors relate to one another in our development of a multi-floor Dialogue Structure annotation schema. Finally, we begin to annotate and analyze the ways in which the visual modalities provide contextual information to the dialogue for overcoming disparities in the collaborators' understanding of the environment. We conclude by discussing the use-cases, architectures, and systems we have implemented from our annotations that enable physical robots to autonomously engage with humans in bi-directional dialogue and navigation.
引用
收藏
页数:51
相关论文
共 50 条
  • [31] Investigating the Role of Multi-modal Social Cues in Human-Robot Collaboration in Industrial Settings
    Cao, Hoang-Long
    Scholz, Constantin
    De Winter, Joris
    El Makrini, Ilias
    Vanderborght, Bram
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2023, 15 (07) : 1169 - 1179
  • [32] Prevention of accomplishing synchronous multi-modal human-robot cooperation by using visual rhythms
    Yonekura, Kenta
    Kim, Chyon Hae
    Nakadai, Kazuhiro
    Tsujino, Hiroshi
    Yokoi, Kazuhito
    ADVANCED ROBOTICS, 2015, 29 (14) : 901 - 912
  • [33] Context-aware selection of multi-modal conversational fillers in human-robot dialogues
    Galle, Matthias
    Kynev, Ekaterina
    Monet, Nicolas
    Legras, Christophe
    2017 26TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2017, : 317 - 322
  • [34] Investigating the Role of Multi-modal Social Cues in Human-Robot Collaboration in Industrial Settings
    Hoang-Long Cao
    Constantin Scholz
    Joris De Winter
    Ilias El Makrini
    Bram Vanderborght
    International Journal of Social Robotics, 2023, 15 : 1169 - 1179
  • [35] Real-Time Multi-Modal Human-Robot Collaboration Using Gestures and Speech
    Chen, Haodong
    Leu, Ming C.
    Yin, Zhaozheng
    JOURNAL OF MANUFACTURING SCIENCE AND ENGINEERING-TRANSACTIONS OF THE ASME, 2022, 144 (10):
  • [36] Multi-modal referring expressions in human-human task descriptions and their implications for human-robot interaction
    Gross, Stephanie
    Krenn, Brigitte
    Scheutz, Matthias
    INTERACTION STUDIES, 2016, 17 (02) : 180 - 210
  • [37] Design of multi-modal feedback channel of human-robot cognitive interface for teleoperation in manufacturing
    Zheng, Chen
    Wang, Kangning
    Gao, Shiqi
    Yu, Yang
    Wang, Zhanxi
    Tang, Yunlong
    JOURNAL OF INTELLIGENT MANUFACTURING, 2024,
  • [38] Fostering common ground in human-robot interaction
    Kiesler, S
    2005 IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2005, : 729 - 734
  • [39] A Multi-party Multi-modal Dataset for Focus of Visual Attention in Human-human and Human-robot Interaction
    Stefanov, Kalin
    Beskow, Jonas
    LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 4440 - 4444
  • [40] OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
    Stephan, Benedict
    Koehler, Mona
    Mueller, Steffen
    Zhang, Yan
    Gross, Horst-Michael
    Notni, Gunther
    SENSORS, 2023, 23 (18)