Learning Multimodal Representations for Sample-efficient Recognition of Human Actions

被引:0
|
作者
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
de Matos, David Martins [1 ,2 ]
Paiva, Ana [1 ,2 ]
Inamura, Tetsunari [3 ,4 ]
机构
[1] Univ Lisbon, INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
[3] SOKENDAI Grad Univ Adv Studies, Natl Inst Informat, Chiyoda Ku, 2-1-2 Hitotsubashi, Tokyo, Japan
[4] SOKENDAI Grad Univ Adv Studies, Dept Informat, Chiyoda Ku, 2-1-2 Hitotsubashi, Tokyo, Japan
来源
2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) | 2019年
关键词
D O I
10.1109/iros40897.2019.8967635
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans interact in rich and diverse ways with the environment. However, the representation of such behavior by artificial agents is often limited. In this work we present motion concepts, a novel multimodal representation of human actions in a household environment. A motion concept encompasses a probabilistic description of the kinematics of the action along with its contextual background, namely the location and the objects held during the performance. We introduce a novel algorithm which learns and recognizes motion concepts from action demonstrations, named Online Motion Concept Learning (OMCL). The algorithm is evaluated on a virtual-reality household environment with the presence of a human avatar. OMCL outperforms standard motion recognition algorithms on an one-shot recognition task, attesting to its potential for sample-efficient recognition of human actions.
引用
收藏
页码:4288 / 4293
页数:6
相关论文
共 50 条
  • [21] A New Sample-Efficient PAC Reinforcement Learning Algorithm
    Zehfroosh, Ashkan
    Tanner, Herbert G.
    2020 28TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2020, : 788 - 793
  • [22] Learning Sample-Efficient Target Reaching for Mobile Robots
    Khan, Arbaaz
    Kumar, Vijay
    Ribeiro, Alejandro
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 3080 - 3087
  • [23] Conditional Abstraction Trees for Sample-Efficient Reinforcement Learning
    Dadvar, Mehdi
    Nayyar, Rashmeet Kaur
    Srivastava, Siddharth
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 485 - 495
  • [24] Learning future representation with synthetic observations for sample-efficient reinforcement learning
    Xin Liu
    Yaran Chen
    Haoran Li
    Dongbin Zhao
    Science China Information Sciences, 2025, 68 (5)
  • [25] Learning future representation with synthetic observations for sample-efficient reinforcement learning
    Xin LIU
    Yaran CHEN
    Haoran LI
    Dongbin ZHAO
    Science China(Information Sciences), 2025, 68 (05) : 20 - 37
  • [26] Sample-efficient Deep Reinforcement Learning with Imaginary Rollouts for Human-Robot Interaction
    Thabet, Mohammad
    Patacchiola, Massimiliano
    Cangelosi, Angelo
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 5079 - 5085
  • [27] Sample-Efficient Reinforcement Learning for Pose Regulation of a Mobile Robot
    Brescia, Walter
    De Cicco, Luca
    Mascolo, Saverio
    2022 11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (ICCAIS), 2022, : 42 - 47
  • [28] Sample-Efficient Deep Reinforcement Learning with Directed Associative Graph
    Yang, Dujia
    Qin, Xiaowei
    Xu, Xiaodong
    Li, Chensheng
    Wei, Guo
    CHINA COMMUNICATIONS, 2021, 18 (06) : 100 - 113
  • [29] Safe and Sample-Efficient Reinforcement Learning Algorithms for Factored Environments
    Simao, Thiago D.
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 6460 - 6461
  • [30] Sample-Efficient Policy Learning based on Completely Behavior Cloning
    Zou, Qiming
    Wang, Ling
    Li, Yu
    Liu, Jie
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 2543 - 2548