Multimodal learning and inference from visual and remotely sensed data

被引:24
|
作者
Rao, Dushyant [1 ]
De Deuge, Mark [1 ]
Nourani-Vatani, Navid [2 ]
Williams, Stefan B. [1 ]
Pizarro, Oscar [1 ]
机构
[1] Univ Sydney, Australian Ctr Field Robot, Rose St Bldg J04, Sydney, NSW 2006, Australia
[2] Siemens AG, Berlin, Germany
基金
澳大利亚研究理事会;
关键词
Multimodal learning; deep learning; marine robotics; classification; semantic mapping; autonomous exploration; CLASSIFICATION;
D O I
10.1177/0278364916679892
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Autonomous vehicles are often tasked to explore unseen environments, aiming to acquire and understand large amounts of visual image data and other sensory information. In such scenarios, remote sensing data may be available a priori, and can help to build a semantic model of the environment and plan future autonomous missions. In this paper, we introduce two multimodal learning algorithms to model the relationship between visual images taken by an autonomous underwater vehicle during a survey and remotely sensed acoustic bathymetry (ocean depth) data that is available prior to the survey. We present a multi-layer architecture to capture the joint distribution between the bathymetry and visual modalities. We then propose an extension based on gated feature learning models, which allows the model to cluster the input data in an unsupervised fashion and predict visual image features using just the ocean depth information. Our experiments demonstrate that multimodal learning improves semantic classification accuracy regardless of which modalities are available at classification time, allows for unsupervised clustering of either or both modalities, and can facilitate mission planning by enabling class-based or image-based queries.
引用
收藏
页码:24 / 43
页数:20
相关论文
共 50 条
  • [31] Feature clustering and ranking for selecting stable features from high dimensional remotely sensed data
    Harris, Dugal
    Van Niekerk, Adriaan
    INTERNATIONAL JOURNAL OF REMOTE SENSING, 2018, 39 (23) : 8934 - 8949
  • [32] Fuzzy modelling of vegetation from remotely sensed imagery
    Foody, GM
    ECOLOGICAL MODELLING, 1996, 85 (01) : 3 - 12
  • [33] Mapping Population Distribution from High Resolution Remotely Sensed Imagery in a Data Poor Setting
    Mossoux, Sophie
    Kervyn, Matthieu
    Soule, Hamid
    Canters, Frank
    REMOTE SENSING, 2018, 10 (09)
  • [34] Distribution and Variation of Forests in China from 2001 to 2011: A Study Based on Remotely Sensed Data
    Zhao, Xiang
    Xu, Peipei
    Zhou, Tao
    Li, Qing
    Wu, Donghai
    FORESTS, 2013, 4 (03) : 632 - 649
  • [35] Digital Soil-Class Mapping from Proximal and Remotely Sensed Data at the Field Level
    Triantafilis, John
    Kerridge, Belinda
    Buchanan, Sam M.
    AGRONOMY JOURNAL, 2009, 101 (04) : 841 - 853
  • [36] Active-Metric Learning for Classification of Remotely Sensed Hyperspectral Images
    Pasolli, Edoardo
    Yang, Hsiuhan Lexie
    Crawford, Melba M.
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2016, 54 (04): : 1925 - 1939
  • [37] Soil Type Identification Using Remotely Sensed Data for Agricultural Purpose
    Jiji G.W.
    Nadar P.
    Jiji, G. Wiselin (jijivevin@yahoo.co.in), 1600, Springer (97): : 405 - 411
  • [38] Scope to map soil management units at the district level from remotely sensed -ray spectrometry and proximal sensed EM induction data
    Jing, Y. Z.
    Huang, J.
    Banks, R.
    Triantafilis, J.
    SOIL USE AND MANAGEMENT, 2017, 33 (04) : 538 - 552
  • [39] Integration of Remotely Sensed Soil Sealing Data in Landslide Susceptibility Mapping
    Luti, Tania
    Segoni, Samuele
    Catani, Filippo
    Munafo, Michele
    Casagli, Nicola
    REMOTE SENSING, 2020, 12 (09)
  • [40] Classification Based Approach for Spectral Signature of Remotely Sensed Temporal Data
    Farouk, H.
    Latif, B. M. Abdel
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2013, 13 (11): : 8 - 17