VAST: Visual and Spectral Terrain Classification in Unstructured Multi-Class Environments

被引:4
作者
Hanson, Nathaniel [1 ]
Shaham, Michael [1 ]
Erdogmus, Deniz [1 ]
Padir, Taskin [1 ]
机构
[1] Northeastern Univ, Inst Experiential Robot, Boston, MA 02115 USA
来源
2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) | 2022年
关键词
Sensor Fusion; Datasets for Robot Vision; Field Robots; Spectroscopy in Robotics; SPECTROSCOPY;
D O I
10.1109/IROS47612.2022.9982078
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Terrain classification is a challenging task for robots operating in unstructured environments. Existing classification methods make simplifying assumptions, such as a reduced number of classes, clearly segmentable roads, or good lighting conditions, and focus primarily on one sensor type. These assumptions do not translate well to off-road vehicles, which operate in varying terrain conditions. To provide mobile robots with the capability to identify the terrain being traversed and avoid undesirable surface types, we propose a multimodal sensor suite capable of classifying different terrains. We capture high resolution macro images of surface texture, spectral reflectance curves, and localization data from a 9 degrees of freedom (DOF) inertial measurement unit (IMU) on 11 different terrains at different times of day. Using this dataset, we train individual neural networks on each of the modalities, and then combine their outputs in a fusion network. The fused network achieved an accuracy of 99.98% percent on the test set, exceeding the results of the best individual network component by 0.98%. We conclude that a combination of visual, spectral, and IMU data provides meaningful improvement over state of the art in terrain classification approaches. The data created for this research is available at https://github.com/RIVeR-Lab/vast_data.
引用
收藏
页码:3956 / 3963
页数:8
相关论文
共 40 条
  • [1] On seeing stuff: The perception of materials by humans and machines
    Adelson, EH
    [J]. HUMAN VISION AND ELECTRONIC IMAGING VI, 2001, 4299 : 1 - 12
  • [2] [Anonymous], P IEEE INT C ROB AUT
  • [3] [Anonymous], P NASA SCI TECHN C
  • [4] [Anonymous], THESIS
  • [5] Living in A Material World: How Visual Cues to Material Properties Affect the Way That We Lift Objects and Perceive Their Weight
    Buckingham, Gavin
    Cant, Jonathan S.
    Goodale, Melvyn A.
    [J]. JOURNAL OF NEUROPHYSIOLOGY, 2009, 102 (06) : 3111 - 3118
  • [6] Calderon F., 2007, THESIS
  • [7] A CNN Based Vision-Proprioception Fusion Method for Robust UGV Terrain Classification
    Chen, Yu
    Rastogi, Chirag
    Norris, William R.
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) : 7965 - 7972
  • [8] Integration of simultaneous tactile sensing and visible and near-infrared reflectance spectroscopy in a robot gripper for mango quality assessment
    Cortes, Victoria
    Blanes, Carlos
    Blasco, Jose
    Ortiz, Coral
    Aleixos, Nuria
    Mellado, Martin
    Cubero, Sergio
    Talens, Pau
    [J]. BIOSYSTEMS ENGINEERING, 2017, 162 : 112 - 123
  • [9] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [10] Ebadi K., SEMANTIC MAPPING UNS