Vision-based action recognition of construction workers using dense trajectories

被引:124
作者
Yang, Jun [1 ]
Shi, Zhongke [2 ]
Wu, Ziyan [1 ]
机构
[1] Northwestern Polytech Univ, Sch Mech Civil Engn & Architecture, Xian, Peoples R China
[2] Northwestern Polytech Univ, Sch Automat, Xian, Peoples R China
基金
中国国家自然科学基金;
关键词
Worker; Action recognition; Construction; Computer vision; Dense trajectories; RESOURCES; TRACKING; SPACE;
D O I
10.1016/j.aei.2016.04.009
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Wide spread monitoring cameras on construction sites provide large amount of information for construction management. The emerging of computer vision and machine learning technologies enables automated recognition of construction activities from videos. As the executors of construction, the activities of construction workers have strong impact on productivity and progress. Compared to machine work, manual work is more subjective and may differ largely in operation flow and productivity among different individuals. Hence only a handful of work studies on vision based action recognition of construction workers. Lacking of publicly available datasets is one of the main reasons that currently hinder advancement. The paper studies worker actions comprehensively, abstracts 11 common types of actions from 5 kinds of trades and establishes a new real world video dataset with 1176 instances. For action recognition, a cutting-edge video description method, dense trajectories, has been applied. Support vector machines are integrated with a bag-of-features pipeline for action learning and classification. Performances on multiple types of descriptors (Histograms of Oriented Gradients - HOG, Histograms of Optical Flow - HOF, Motion Boundary Histogram - MBH) and their combination have been evaluated. Discussion on different parameter settings and comparison to the state-of-the-art method are provided. Experimental results show that the system with codebook size 500 and MBH descriptor has achieved an average accuracy of 59% for worker action recognition, outperforming the state-of-the-art result by 24%. (C) 2016 Elsevier Ltd. All rights reserved.
引用
收藏
页码:327 / 336
页数:10
相关论文
共 66 条
[1]  
[Anonymous], AUTOMATED 3D VISION
[2]  
[Anonymous], CONSTR RES C ASCE SE
[3]  
[Anonymous], IEEE I CONF COMP VIS
[4]  
[Anonymous], 32 INT S AUT ROB CON
[5]  
[Anonymous], 2008, CMURITR0822
[6]  
[Anonymous], 2014, COMPREHENSIVE METHOD
[7]  
[Anonymous], 2009, INT WORKSH COMP CIV
[8]  
[Anonymous], IR252 2A GUID ACT AN
[9]  
[Anonymous], BEHAV COMPUTER ASSIS
[10]   Server-Customer Interaction Tracker: Computer Vision-Based System to Estimate Dirt-Loading Cycles [J].
Azar, Ehsan Rezazadeh ;
Dickinson, Sven ;
McCabe, Brenda .
JOURNAL OF CONSTRUCTION ENGINEERING AND MANAGEMENT, 2013, 139 (07) :785-794