CADDY Underwater Stereo-Vision Dataset for Human-Robot Interaction (HRI) in the Context of Diver Activities

被引:41
作者
Chavez, Arturo Gomez [1 ]
Ranieri, Andrea [2 ]
Chiarella, Davide [3 ]
Zereik, Enrica [2 ]
Babic, Anja [4 ]
Birk, Andreas [1 ]
机构
[1] Jacobs Univ Bremen, Comp Sci & Elect Engn, Robot Grp, Campus Ring 1, D-28759 Bremen, Germany
[2] CNR, Inst Marine Engn, Via E De Marini 6, I-16149 Genoa, Italy
[3] CNR, Inst Computat Linguist, Via E De Marini 6, I-16149 Genoa, Italy
[4] Univ Zagreb, Fac Elect Engn & Comp, Unska Ul 3, Zagreb 10000, Croatia
关键词
dataset; underwater imaging; image processing; marine robotics; field robotics; human-robot interaction; stereo vision; object classification; human pose estimation; REFERENCE QUALITY ASSESSMENT; MODEL;
D O I
10.3390/jmse7010016
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
In this article, we present a novel underwater dataset collected from several field trials within the EU FP7 project Cognitive autonomous diving buddy (CADDY), where an Autonomous Underwater Vehicle (AUV) was used to interact with divers and monitor their activities. To our knowledge, this is one of the first efforts to collect a large public dataset in underwater environments with the purpose of studying and boosting object classification, segmentation and human pose estimation tasks. The first part of the dataset contains stereo camera recordings (approximate to 10 K) of divers performing hand gestures to communicate with an AUV in different environmental conditions. The gestures can be used to test the robustness of visual detection and classification algorithms in underwater conditions, e.g., under color attenuation and light backscatter. The second part includes stereo footage (approximate to 12.7 K) of divers free-swimming in front of the AUV, along with synchronized measurements from Inertial Measurement Units (IMU) located throughout the diver's suit (DiverNet), which serve as ground-truth for human pose and tracking methods. In both cases, these rectified images allow the investigation of 3D representation and reasoning pipelines from low-texture targets commonly present in underwater scenarios. This work describes the recording platform, sensor calibration procedure plus the data format and the software utilities provided to use the dataset.
引用
收藏
页数:14
相关论文
共 29 条
[1]  
[Anonymous], P 2018 IEEE INT C RO
[2]  
Bonin F., 2011, Journal of Maritime Research, V8, P65
[3]  
Chavez A.G., 2017, P OCEANS 2017 AB UK
[4]  
Cheng D., 2015, P 2015 IEEE INT C CO
[5]   Illuminant estimation for color constancy: why spatial-domain methods work and the role of the color distribution [J].
Cheng, Dongliang ;
Prasad, Dilip K. ;
Brown, Michael S. .
JOURNAL OF THE OPTICAL SOCIETY OF AMERICA A-OPTICS IMAGE SCIENCE AND VISION, 2014, 31 (05) :1049-1058
[6]  
Chiarella D., 2015, P OCEANS 2015 GEN IT
[7]   A Novel Gesture-Based Language for Underwater Human-Robot Interaction [J].
Chiarella, Davide ;
Bibuli, Marco ;
Bruzzone, Gabriele ;
Caccia, Massimo ;
Ranieri, Andrea ;
Zereik, Enrica ;
Marconi, Lucia ;
Cutugno, Paola .
JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2018, 6 (03)
[8]  
Ciurea F., 2003, P 11 COL IM C SCOTTS
[9]   No-Reference Quality Assessment of Contrast-Distorted Images Based on Natural Scene Statistics [J].
Fang, Yuming ;
Ma, Kede ;
Wang, Zhou ;
Lin, Weisi ;
Fang, Zhijun ;
Zhai, Guangtao .
IEEE SIGNAL PROCESSING LETTERS, 2015, 22 (07) :838-842
[10]  
Garcia R., 2011, P OCEANS 2011 IEEE S