A methodology for semantic action recognition based on pose and human-object interaction in avocado harvesting processes

被引:10
|
作者
Vasconez, J. P. [1 ]
Admoni, H. [2 ]
Auat Cheein, F. [3 ]
机构
[1] Escuela Politec Nacl, Artificial Intelligence & Comp Vis Res Lab, Quito 170517, Ecuador
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[3] Univ Tecn Federico Santa Maria, Dept Elect Engn, Valparaiso, Chile
关键词
Semantic human action recognition; Human-object interaction; Avocado harvesting process; Human?machine collaboration; AGRICULTURE; PRODUCTS;
D O I
10.1016/j.compag.2021.106057
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
The agricultural industry could greatly benefit from an intelligent system capable of supporting field workers to increase production. Such a system would need to monitor human workers, their current actions, their intentions, and possible future actions, which are the focus of this work. Herein, we propose and validate a methodology to recognize human actions during the avocado harvesting process in a Chilean farm based on combined object-pose semantic information using RGB still images. We use Faster R-CNN ?Region Convolutional Neural Network? with Inception V2 convolutional object detection to recognize 17 categories, which include among others, field workers, tools, crops, and vehicles. Then, we use a convolutional-based 2D pose estimation method called OpenPose to detect 18 human skeleton joints. Both the object and the pose features are processed, normalized, and combined into a single feature vector. We test four classifiers ?Support vector machine, Decision trees, KNearest-Neighbour, and Bagged trees? on the combined object-pose feature vectors to evaluate action classification performance. We also test such results using principal component analysis on the four classifiers to reduce dimensionality. Accuracy and inference time are analyzed for all the classifiers using 10 action categories, related to the avocado harvesting process. The results show that it is possible to detect human actions during harvesting, obtaining average accuracy performances (among all action categories) ranging from 57% to 99%, depending on the classifier used. The latter can be used to support an intelligent system, such as robots, interacting with field workers aimed at increasing productivity.
引用
收藏
页数:12
相关论文
共 36 条
  • [21] PHYSIOLOGY AND SENSORIAL BASED QUANTIFICATION OF HUMAN-OBJECT INTERACTION - THE QOSI MATRIX
    Balters, Stephanie
    Jensen, Bisballe Matilde
    Steinert, Martin
    ICED 15, VOL 11: HUMAN BEHAVIOUR IN DESIGN, DESIGN EDUCATION, 2015,
  • [22] Detection of Anomalous Behavior of Manufacturing Workers Using Deep Learning-Based Recognition of Human-Object Interaction
    Rijayanti, Rita
    Hwang, Mintae
    Jin, Kyohong
    APPLIED SCIENCES-BASEL, 2023, 13 (15):
  • [23] Language-guided graph parsing attention network for human-object interaction recognition
    Li, Qiyue
    Xie, Xuemei
    Zhang, Jin
    Shi, Guangming
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 89
  • [24] Human-object interaction detection based on cascade multi-scale transformer
    Limin Xia
    Xiaoyue Ding
    Applied Intelligence, 2024, 54 : 2831 - 2850
  • [25] ACP plus plus : Action Co-Occurrence Priors for Human-Object Interaction Detection
    Kim, Dong-Jin
    Sun, Xiao
    Choi, Jinsoo
    Lin, Stephen
    Kweon, In So
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 9150 - 9163
  • [26] Human-object interaction detection based on cascade multi-scale transformer
    Xia, Limin
    Ding, Xiaoyue
    APPLIED INTELLIGENCE, 2024, 54 (03) : 2831 - 2850
  • [27] LiDARCapV2: 3D human pose estimation with human-object interaction from LiDAR point clouds
    Zhang, Jingyi
    Mao, Qihong
    Shen, Siqi
    Xu, Lan
    Wang, Cheng
    PATTERN RECOGNITION, 2024, 156
  • [28] Enhancing Recognition of Human-Object Interaction from Visual Data Using Egocentric Wearable Camera
    Hamid, Danish
    Ul Haq, Muhammad Ehatisham
    Yasin, Amanullah
    Murtaza, Fiza
    Azam, Muhammad Awais
    FUTURE INTERNET, 2024, 16 (08)
  • [29] Interactivity Recognition Graph Neural Network (IR-GNN) Model for Improving Human-Object Interaction Detection
    Zhang, Jiali
    Yunos, Zuriahati Mohd
    Haron, Habibollah
    ELECTRONICS, 2023, 12 (02)
  • [30] Towards Hard-Positive Query Mining for DETR-Based Human-Object Interaction Detection
    Zhong, Xubin
    Ding, Changxing
    Li, Zijian
    Huang, Shaoli
    COMPUTER VISION - ECCV 2022, PT XXVII, 2022, 13687 : 444 - 460