Pointing Gestures for Human-Robot Interaction in Service Robotics: A Feasibility Study

被引:0
作者
Pozzi, Luca [1 ]
Gandolla, Marta [2 ]
Roveda, Loris [3 ]
机构
[1] Politecn Milan, WE COBOT Lab Polo Territoriale Lecco, Mech Dept, Lecce, Italy
[2] Politecn Milan, Mech Dept, Milan, Italy
[3] Univ Svizzera Italiana USI, Scuola Univ Profess Svizzera Italiana SUPSI, Ist Dalle Molle Studi Intelligenza Artificiale ID, Lugano, Switzerland
来源
COMPUTERS HELPING PEOPLE WITH SPECIAL NEEDS, ICCHP-AAATE 2022, PT II | 2022年
关键词
Human-Robot Interaction; Pointing; Service robotics; Action detection;
D O I
10.1007/978-3-031-08645-8_54
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Research in service robotics strives at having a positive impact on people's quality of life by the introduction of robotic helpers for everyday activities. From this ambition arises the need of enabling natural communication between robots and ordinary people. For this reason, Human-Robot Interaction (HRI) is an extensively investigated topic, exceeding language-based exchange of information, to include all the relevant facets of communication. Each aspect of communication (e.g. hearing, sight, touch) comes with its own peculiar strengths and limits, thus they are often combined to improve robustness and naturalness. In this contribution, an HRI framework is presented, based on pointing gestures as the preferred interaction strategy. Pointing gestures are selected as they are an innate behavior to direct another attention, and thus could represent a natural way to require a service to a robot. To complement the visual information, the user could be prompted to give voice commands to resolve ambiguities and prevent the execution of unintended actions. The two layers (perceptive and semantic) architecture of the proposed HRI system is described. The perceptive layer is responsible for objects mapping, action detection, and assessment of the indicated direction. Moreover, it has to listen to uses' voice commands. To avoid privacy issues and not burden the computational resources of the robot, the interaction would be triggered by a wake-word detection system. The semantic layer receives the information processed by the perceptive layer and determines which actions are available for the selected object. The decision is based on object's characteristics, contextual information and user vocal feedbacks are exploited to resolve ambiguities. A pilot implementation of the semantic layer is detailed, and qualitative results are shown. The preliminary findings on the validity of the proposed system, as well as on the limitations of a purely vision-based approach, are discussed.
引用
收藏
页码:461 / 468
页数:8
相关论文
共 15 条
  • [11] Ozturkcan S., 2021, J INF TECHNOL TEACH
  • [12] Pages J., 2016, INT WORKSH ROB MOD
  • [13] Shan JJ, 2014, 2014 IEEE WORKSHOP ON ADVANCED ROBOTICS AND ITS SOCIAL IMPACTS (ARSO), P69, DOI 10.1109/ARSO.2014.7020983
  • [14] Pointing Estimation for Human-Robot Interaction Using Hand Pose, Verbal Cues, and Confidence Heuristics
    Showers, Andrew
    Si, Mei
    [J]. SOCIAL COMPUTING AND SOCIAL MEDIA: TECHNOLOGIES AND ANALYTICS, SCSM 2018, PT II, 2018, 10914 : 403 - 412
  • [15] Brave new world: service robots in the frontline
    Wirtz, Jochen
    Patterson, Paul G.
    Kunz, Werner H.
    Gruber, Thorsten
    Lu, Vinh Nhat
    Paluch, Stefanie
    Martins, Antje
    [J]. JOURNAL OF SERVICE MANAGEMENT, 2018, 29 (05) : 907 - 931