Cross-View Human Intention Recognition for Human-Robot Collaboration

被引:5
作者
Ni, Shouxiang [1 ]
Zhao, Lindong [1 ]
Li, Ang [1 ]
Wu, Dan [2 ]
Zhou, Liang [1 ]
机构
[1] Nanjing Univ Posts & Telecommun, Nanjing, Peoples R China
[2] Army Engn Univ PLA, Nanjing, Peoples R China
基金
中国国家自然科学基金;
关键词
Measurement; Face recognition; Wireless networks; Semantics; Collaboration; Machine learning; Production facilities; Human-robot interaction;
D O I
10.1109/MWC.018.2200514
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Benefiting from the promise of sixth generation (6G) wireless networks, multimodal machine learning based on exploiting complementarity among video, audio, and haptic signals, becomes a key enabler for human intention recognition, which is critical to realize effective human-robot collaboration in Industry 4.0 scenarios. However, as multimodal human intention recognition is limited by expensive equipment and a demanding environment, it is hard to strike an efficient trade-off between inference accuracy and system overhead. Naturally, how to induce more intention semantics from readily available videos emerges as a fundamental issue for human intention recognition. In this article, we use cross-view human intention recognition to solve the above issue and demonstrate the effectiveness of our method with well-designed evaluation metrics. Specifically, we first compensate for the scarcity of intention semantics in the body view by adding a face view. Second, we deploy the cross-view generative model to capture intention semantics induced by the mutual generation of two views. Finally, in the human-robot collaboration experiments, our method gets closer to human performance regarding response time and inference accuracy.
引用
收藏
页码:189 / 195
页数:7
相关论文
共 15 条
  • [1] Multimodal Machine Learning: A Survey and Taxonomy
    Baltrusaitis, Tadas
    Ahuja, Chaitanya
    Morency, Louis-Philippe
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) : 423 - 443
  • [2] Representation Learning: A Review and New Perspectives
    Bengio, Yoshua
    Courville, Aaron
    Vincent, Pascal
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) : 1798 - 1828
  • [3] Negative Information Measurement at AI Edge: A New Perspective for Mental Health Monitoring
    Chen, Min
    Shen, Ke
    Wang, Rui
    Miao, Yiming
    Jiang, Yingying
    Hwang, Kai
    Hao, Yixue
    Tao, Guangming
    Hu, Long
    Liu, Zhongchun
    [J]. ACM TRANSACTIONS ON INTERNET TECHNOLOGY, 2022, 22 (03)
  • [4] The relevance of signal timing in human-robot collaborative manipulation
    Cini, F.
    Banfi, T.
    Ciuti, G.
    Craighero, L.
    Controzzi, M.
    [J]. SCIENCE ROBOTICS, 2021, 6 (58)
  • [5] Toward 6G Networks: Use Cases and Technologies
    Giordani, Marco
    Polese, Michele
    Mezzavilla, Marco
    Rangan, Sundeep
    Zorzi, Michele
    [J]. IEEE COMMUNICATIONS MAGAZINE, 2020, 58 (03) : 55 - 61
  • [6] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [7] Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [8] Survey of Human-Robot Collaboration in Industrial Settings: Awareness, Intelligence, and Compliance
    Kumar, Shitij
    Savur, Celal
    Sahin, Ferat
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (01): : 280 - 297
  • [9] A Vision of 6G Wireless Systems: Applications, Trends, Technologies, and Open Research Problems
    Saad, Walid
    Bennis, Mehdi
    Chen, Mingzhe
    [J]. IEEE NETWORK, 2020, 34 (03): : 134 - 142
  • [10] Simonyan K, 2014, ADV NEUR IN, V27