Effective human-object interaction recognition for edge devices in intelligent space

被引:3
作者
Ozaki, Haruhiro [1 ]
Tran, Dinh Tuan [2 ]
Lee, Joo-Ho [2 ]
机构
[1] Ritsumeikan Univ, Grad Sch Informat Sci & Engn, Shiga, Japan
[2] Ritsumeikan Univ, Coll Informat Sci & Engn, Shiga, Japan
关键词
Human-object interaction; computer vision; machine learning; edge computing; system integration; intelligent space;
D O I
10.1080/18824889.2023.2292353
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To enable machines to understand human-centric images and videos, they need the capability to detect human-object interactions. This capability has been studied using various approaches, but previous research has mainly focused only on recognition accuracy using widely used open datasets. Given the need for advanced machine-learning systems that provide spatial analysis and services, the recognition model should be robust to various changes, have high extensibility, and provide sufficient recognition speed even with minimal computational overhead. Therefore, we propose a novel method that combines the skeletal method with object detection to accurately predict a set of $ \langle $ < human, verb, object $ \rangle $ > triplets in a video frame considering the robustness, extensibility, and lightweight of the model. Training a model with similar perceptual elements to those of humans produces sufficient accuracy for advanced social systems, even with only a small training dataset. The proposed model is trained using only the coordinates of the object and human landmarks, making it robust to various situations and lightweight compared with deep-learning methods. In the experiment, a scenario in which a human is working on a desk is simulated and an algorithm is trained on object-specific interactions. The accuracy of the proposed model was evaluated using various types of datasets.
引用
收藏
页码:1 / 9
页数:9
相关论文
共 25 条
[1]  
Bazarevsky V, 2020, Arxiv, DOI [arXiv:2006.10204, DOI 10.48550/ARXIV.2006.10204]
[2]   Learning to Detect Human-Object Interactions [J].
Chao, Yu-Wei ;
Liu, Yunfan ;
Liu, Xieyang ;
Zeng, Huayi ;
Deng, Jia .
2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, :381-389
[3]   Revisiting Skeleton-based Action Recognition [J].
Duan, Haodong ;
Zhao, Yue ;
Chen, Kai ;
Lin, Dahua ;
Dai, Bo .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :2959-2968
[4]   Understanding Atomic Hand-Object Interaction With Human Intention [J].
Fan, Hehe ;
Zhuo, Tao ;
Yu, Xin ;
Yang, Yi ;
Kankanhalli, Mohan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) :275-285
[5]   Detecting and Recognizing Human-Object Interactions [J].
Gkioxari, Georgia ;
Girshick, Ross ;
Dollar, Piotr ;
He, Kaiming .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8359-8367
[6]   Observing Human-Object Interactions: Using Spatial and Functional Compatibility for Recognition [J].
Gupta, Abhinav ;
Kembhavi, Aniruddha ;
Davis, Larry S. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (10) :1775-1789
[7]  
Irie K., 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566), P193
[8]   Development of ROS2-TMS: new software platform for informationally structured environment [J].
Itsuka, Tomoya ;
Song, Minsoo ;
Kawamura, Akihiro ;
Kurazume, Ryo .
ROBOMECH JOURNAL, 2022, 9 (01)
[9]   HOTR: End-to-End Human-Object Interaction Detection with Transformers [J].
Kim, Bumsoo ;
Lee, Junhyun ;
Kang, Jaewoo ;
Kim, Eun-Sol ;
Kim, Hyunwoo J. .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :74-83
[10]   Intelligent Space - concept and contents [J].
Lee, JH ;
Hashimoto, H .
ADVANCED ROBOTICS, 2002, 16 (03) :265-280