Collecting public RGB-D datasets for human daily activity recognition

被引:7
|
作者
Wu, Hanbo [1 ]
Ma, Xin [1 ]
Zhang, Zhimeng [1 ]
Wang, Haibo [1 ]
Li, Yibin [1 ]
机构
[1] Shandong Univ, Sch Control Sci & Engn, 17923 Jingshi Rd, Jinan, Shandong, Peoples R China
来源
INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS | 2017年 / 14卷 / 04期
关键词
Human daily activity recognition; public RGB-D data sets merging; large-scale RGB-D activity data set; depth motion maps; depth cuboid similarity feature; curvature space scale; OBJECT RECOGNITION; FUSION; MODEL;
D O I
10.1177/1729881417709079
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Human daily activity recognition has been a hot spot in the field of computer vision for many decades. Despite best efforts, activity recognition in naturally uncontrolled settings remains a challenging problem. Recently, by being able to perceive depth and visual cues simultaneously, RGB-D cameras greatly boost the performance of activity recognition. However, due to some practical difficulties, the publicly available RGB-D data sets are not sufficiently large for benchmarking when considering the diversity of their activities, subjects, and background. This severely affects the applicability of complicated learning-based recognition approaches. To address the issue, this article provides a large-scale RGB-D activity data set by merging five public RGB-D data sets that differ from each other on many aspects such as length of actions, nationality of subjects, or camera angles. This data set comprises 4528 samples depicting 7 action categories (up to 46 subcategories) performed by 74 subjects. To verify the challengeness of the data set, three feature representation methods are evaluated, which are depth motion maps, spatiotemporal depth cuboid similarity feature, and curvature space scale. Results show that the merged large-scale data set is more realistic and challenging and therefore more suitable for benchmarking.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 50 条
  • [21] Deep learning and RGB-D based human action, human-human and human-object interaction recognition: A survey?
    Khaire, Pushpajit
    Kumar, Praveen
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2022, 86
  • [22] A Recognition Method for Overlapped Objects using Multiple RGB-D Sensors
    Morimoto, Masakazu
    Yukitou, Mitsuhiro
    2018 WORLD AUTOMATION CONGRESS (WAC), 2018, : 18 - 21
  • [23] Object recognition and robot grasping technology based on RGB-D data
    Yu, Sheng
    Zhai, Di-Hua
    Wu, Haocun
    Yang, Hongda
    Xia, Yuanqing
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 3869 - 3874
  • [24] Perception Subsystem for Object Recognition and Pose Estimation in RGB-D Images
    Kornuta, Tomasz
    Laszkowski, Michal
    CHALLENGES IN AUTOMATION, ROBOTICS AND MEASUREMENT TECHNIQUES, 2016, 440 : 597 - 607
  • [25] A comparative study of data fusion for RGB-D based visual recognition
    Sanchez-Riera, Jordi
    Hua, Kai-Lung
    Hsiao, Yuan-Sheng
    Lim, Tekoing
    Hidayati, Shintami C.
    Cheng, Wen-Huang
    PATTERN RECOGNITION LETTERS, 2016, 73 : 1 - 6
  • [26] Fast Location and Recognition of Green Apple Based on RGB-D Image
    Sun, Meili
    Xu, Liancheng
    Luo, Rong
    Lu, Yuqi
    Jia, Weikuan
    FRONTIERS IN PLANT SCIENCE, 2022, 13
  • [27] RGB-D OBJECT RECOGNITION WITH MULTIMODAL DEEP CONVOLUTIONAL NEURAL NETWORKS
    Rahman, Mohammad Muntasir
    Tan, Yanhao
    Xue, Jian
    Lu, Ke
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 991 - 996
  • [28] Arbitrary-View Human Action Recognition: A Varying-View RGB-D Action Dataset
    Ji, Yanli
    Yang, Yang
    Shen, Fumin
    Shen, Heng Tao
    Zheng, Wei-Shi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (01) : 289 - 300
  • [29] Infrared and 3D Skeleton Feature Fusion for RGB-D Action Recognition
    De Boissiere, Alban Main
    Noumeir, Rita
    IEEE ACCESS, 2020, 8 (08): : 168297 - 168308
  • [30] Human Visual Scanpath Prediction Based on RGB-D Saliency
    Han, Rui
    Xiao, Shuangjiu
    PROCEEDINGS OF 2018 INTERNATIONAL CONFERENCE ON IMAGE AND GRAPHICS PROCESSING (ICIGP 2018), 2018, : 180 - 184