Enhancing Recognition of Human-Object Interaction from Visual Data Using Egocentric Wearable Camera

被引:0
|
作者
Hamid, Danish [1 ]
Ul Haq, Muhammad Ehatisham [1 ]
Yasin, Amanullah [1 ]
Murtaza, Fiza [1 ]
Azam, Muhammad Awais [2 ]
机构
[1] Air Univ, Fac Comp & Artificial Intelligence FCAI, Dept Creat Technol, Islamabad 44000, Pakistan
[2] Whitecliffe, Technol & Innovat Res Grp, Sch Informat Technol, Wellington 6145, New Zealand
关键词
egocentric; hand pose; human-object interaction; machine learning; object recognition; wearable camera;
D O I
10.3390/fi16080269
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection and human action recognition have great significance in many real-world applications. Understanding how a human being interacts with different objects, i.e., human-object interaction, is also crucial in this regard since it enables diverse applications related to security, surveillance, and immersive reality. Thus, this study explored the potential of using a wearable camera for object detection and human-object interaction recognition, which is a key technology for the future Internet and ubiquitous computing. We propose a system that uses an egocentric camera view to recognize objects and human-object interactions by analyzing the wearer's hand pose. Our novel idea leverages the hand joint data of the user, which were extracted from the egocentric camera view, for recognizing different objects and related interactions. Traditional methods for human-object interaction rely on a third-person, i.e., exocentric, camera view by extracting morphological and color/texture-related features, and thus, often fall short when faced with occlusion, camera variations, and background clutter. Moreover, deep learning-based approaches in this regard necessitate substantial data for training, leading to a significant computational overhead. Our proposed approach capitalizes on hand joint data captured from an egocentric perspective, offering a robust solution to the limitations of traditional methods. We propose a machine learning-based innovative technique for feature extraction and description from 3D hand joint data by presenting two distinct approaches: object-dependent and object-independent interaction recognition. The proposed method offered advantages in computational efficiency compared with deep learning methods and was validated using the publicly available HOI4D dataset, where it achieved a best-case average F1-score of 74%. The proposed system paves the way for intuitive human-computer collaboration within the future Internet, enabling applications like seamless object manipulation and natural user interfaces for smart devices, human-robot interactions, virtual reality, and augmented reality.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Decoding the time-course of object recognition in the human brain: From visual features to categorical decisions
    Contini, Erika W.
    Wardle, Susan G.
    Carlson, Thomas A.
    NEUROPSYCHOLOGIA, 2017, 105 : 165 - 176
  • [42] Wearable-based human flow experience recognition enhanced by transfer learning methods using emotion data
    Irshad, Muhammad Tausif
    Li, Frederic
    Nisar, Muhammad Adeel
    Huang, Xinyu
    Buss, Martje
    Kloep, Leonie
    Peifer, Corinna
    Kozusznik, Barbara
    Pollak, Anita
    Pyszka, Adrian
    Flak, Olaf
    Grzegorzek, Marcin
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 166
  • [43] A Hierarchical Multitask Learning Approach for the Recognition of Activities of Daily Living Using Data from Wearable Sensors
    Nisar, Muhammad Adeel
    Shirahama, Kimiaki
    Irshad, Muhammad Tausif
    Huang, Xinyu
    Grzegorzek, Marcin
    SENSORS, 2023, 23 (19)
  • [44] Simultaneous joint and object trajectory templates for human activity recognition from 3-D data
    Ghodsi, Saeed
    Mohammadzade, Hoda
    Korki, Erfan
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2018, 55 : 729 - 741
  • [45] Instance Based Human Physical Activity(HPA) Recognition Using Shimmer2 Wearable Sensor Data sets
    Doreswamy
    Yogesh, K. M.
    2017 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATIONS AND INFORMATICS (ICACCI), 2017, : 995 - 999
  • [46] Visual Object Recognition From Single-Trial EEG Signals Using Machine Learning Wrapper Techniques
    Yavandhasani, Mojtaba
    Ghaderi, Foad
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2022, 69 (07) : 2176 - 2183
  • [47] A highly efficient human activity classification method using mobile data from wearable sensors
    Yuan, Lihuan
    Liu, Qiang
    Lu, Mengchi
    Zhou, Sihang
    Zhu, Chengzhang
    Yin, Jianping
    INTERNATIONAL JOURNAL OF SENSOR NETWORKS, 2017, 25 (02) : 86 - 92
  • [48] Robust Multi-User In-Hand Object Recognition in Human-Robot Collaboration Using a Wearable Force-Myography Device
    Bamani, Eran
    Kahanowich, Nadav D.
    Ben-David, Inbar
    Sintov, Avishai
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (01) : 104 - 111
  • [49] Data processing for object recognition from satellite images using convolutional neural networks: a case study in solar panel recognition with FORMOSAT-2
    Chen, Bo-Wei
    Hsu, Hwai-Jung
    Chang, Yu-Yun
    Liu, Cynthia S. J.
    Huang, Winfred
    JOURNAL OF APPLIED REMOTE SENSING, 2024, 18 (03)
  • [50] Statistical and Machine Learning-Based Recognition of Coughing Events Using Triaxial Accelerometer Sensor Data From Multiple Wearable Points
    Doddabasappla, Kruthi
    Vyas, Rushi
    IEEE SENSORS LETTERS, 2021, 5 (06)