Three-dimensional structural displacement measurement using monocular vision and deep learning based pose estimation

被引:48
作者
Sun, Chujin [1 ]
Gu, Donglian [2 ]
Lu, Xinzheng [1 ]
机构
[1] Tsinghua Univ, Dept Civil Engn, Key Lab Civil Engn Safety & Durabil China, Educ Minist, Beijing 100084, Peoples R China
[2] Univ Sci & Technol Beijing, Sch Civil & Resource Engn, Res Inst Urbanizat & Urban Safety, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Monocular vision; Displacement measurement; Three-dimensional; Pose estimation; Deep learning; DIGITAL IMAGE CORRELATION; COMPUTER VISION; CIVIL INFRASTRUCTURE; INSPECTION;
D O I
10.1016/j.ymssp.2023.110141
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Computer vision-based displacement measurement methods have received increasing attention for the structural health monitoring of buildings and infrastructures owing to their advantages over traditional contact sensors. Meanwhile, surveillance cameras widely equipped in urban areas can record a large number of images and videos of buildings and infrastructure, which have the potential to support structural analysis in structural health monitoring or engineering in-vestigations. The three-dimensional (3D) displacement of structures is important for structural analysis. It is challenging for the existing vision-based measurement methods to obtain all the 3D displacement components because they require either multi-view camera systems or additional specially designed targets, which makes it difficult to meet the requirements of measurement applications based on urban surveillance cameras. Therefore, this study proposes a 3D structural displacement measurement method using monocular vision and deep learning based pose esti-mation. The method uses virtual rendering to synthesize the training set based on the 3D models of the target objects, then trains the deep learning model DPOD (Dense Pose Object Detector) to estimate the poses of the target object, and finally measures the 3D translation of the structures based on the original and destination poses or the original pose and keypoint matching. The effectiveness of the proposed method was validated through static and dynamic experiments. The results showed that the method can meet the needs of obtaining 3D structural displacement and has good accuracy in identifying the principal frequencies of the dynamic responses. The pro-posed method can support the 3D displacement measurements of buildings and infrastructure based on urban surveillance cameras.
引用
收藏
页数:26
相关论文
共 60 条
[1]  
[Anonymous], 2022, CO COCOMMON OBJECTS
[2]  
[Anonymous], 2008, EUR IT CHAPT C
[3]   SilhoNet: An RGB Method for 6D Object Pose Estimation [J].
Billings, Gideon ;
Johnson-Roberson, Matthew .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04) :3727-3734
[4]  
Brachmann E, 2014, LECT NOTES COMPUT SC, V8690, P536, DOI 10.1007/978-3-319-10605-2_35
[5]   Flexible videogrammetric technique for three-dimensional structural vibration measurement [J].
Chang, C. C. ;
Ji, Y. F. .
JOURNAL OF ENGINEERING MECHANICS, 2007, 133 (06) :656-664
[6]   Three-Dimensional Structural Translation and Rotation Measurement Using Monocular Videogrammetry [J].
Chang, C. C. ;
Xiao, X. H. .
JOURNAL OF ENGINEERING MECHANICS, 2010, 136 (07) :840-848
[7]   Video Camera-Based Vibration Measurement for Civil Infrastructure Applications [J].
Chen, Justin G. ;
Davis, Abe ;
Wadhwa, Neal ;
Durand, Fredo ;
Freeman, William T. ;
Buyukozturk, Oral .
JOURNAL OF INFRASTRUCTURE SYSTEMS, 2017, 23 (03)
[8]   A review of computer vision-based structural health monitoring at local and global levels [J].
Dong, Chuan-Zhi ;
Catbas, F. Necati .
STRUCTURAL HEALTH MONITORING-AN INTERNATIONAL JOURNAL, 2021, 20 (02) :692-743
[9]   Computer vision for SHM of civil infrastructure: From dynamic response measurement to damage detection - A review [J].
Feng, Dongming ;
Feng, Maria Q. .
ENGINEERING STRUCTURES, 2018, 156 :105-117
[10]   Experimental validation of cost-effective vision-based structural health monitoring [J].
Feng, Dongming ;
Feng, Maria Q. .
MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2017, 88 :199-211