Multi-Modal Object Tracking and Image Fusion With Unsupervised Deep Learning

被引:11
作者
LaHaye, Nicholas [1 ,2 ]
Ott, Jordan [1 ]
Garay, Michael J. [3 ]
El-Askary, Hesham Mohamed [4 ,5 ,6 ]
Linstead, Erik [5 ,7 ]
机构
[1] Chapman Univ, Computat & Data Sci Dept, Orange, CA 92866 USA
[2] CALTECH, Jet Prop Lab, Proc Algorithms & Calibrat Engn Grp, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[3] CALTECH, Jet Prop Lab, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[4] Chapman Univ, Ctr Excellence Earth Syst Modeling & Observat, Orange, CA 92866 USA
[5] Chapman Univ, Schmid Coll Sci & Technol, Orange, CA 92866 USA
[6] Alexandria Univ, Fac Sci, Dept Environm Sci, Alexandria 21522, Egypt
[7] Chapman Univ, Machine Learning & Assist Technol Lab, Orange, CA 92866 USA
关键词
Bigdata applications; clustering; computer vision; deep belief networks (DBNs); deep learning; CLASSIFICATION; MISR;
D O I
10.1109/JSTARS.2019.2920234
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The number of different modalities for remote sensors continues to grow, bringing with it an increase in the volume and complexity of the data being collected. Although these datasets individually provide valuable information, in aggregate they provide additional opportunities to discover meaningful patterns on a large scale. However, the ability to combine and analyze disparate datasets is challenged by the potentially vast parameter space that results from aggregation. Each dataset in itself requires instrument-specific and dataset-specific knowledge. If the intention is to use multiple, diverse datasets, one needs an understanding of how to translate and combine these parameters in an efficient and effective manner. While there are established techniques for combining datasets from specific domains or platforms, there is no generic, automated method that can address the problem in general. Here, we discuss the application of deep learning to track objects across different image-like data-modalities, given data in a similar spatio-temporal range, and automatically co-register these images. Using deep belief networks combined with unsupervised learning methods, we are able to recognize and separate different objects within image-like data in a structured manner, thus making progress toward the ultimate goal of a generic tracking and fusion pipeline requiring minimal human intervention.
引用
收藏
页码:3056 / 3066
页数:11
相关论文
共 50 条
  • [41] Deep Multi-Modal Metric Learning with Multi-Scale Correlation for Image-Text Retrieval
    Hua, Yan
    Yang, Yingyun
    Du, Jianhe
    ELECTRONICS, 2020, 9 (03)
  • [42] Exploiting Multi-Modal Fusion for Urban Autonomous Driving Using Latent Deep Reinforcement Learning
    Khalil, Yasser H.
    Mouftah, Hussein T.
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (03) : 2921 - 2935
  • [43] Multi-modal fusion deep learning model for excavated soil heterogeneous data with efficient classification
    Guo, Qi-Meng
    Zhan, Liang-Tong
    Yin, Zhen-Yu
    Feng, Hang
    Yang, Guang-Qian
    Chen, Yun-Min
    COMPUTERS AND GEOTECHNICS, 2024, 175
  • [44] A Deep-Learning-Based Multi-Modal Sensor Fusion Approach for Detection of Equipment Faults
    Kullu, Omer
    Cinar, Eyup
    MACHINES, 2022, 10 (11)
  • [45] An Unsupervised Domain Adaptation Method for Multi-Modal Remote Sensing Image Classification
    Liu, Wei
    Qin, Rongjun
    Su, Fulin
    Hu, Kun
    2018 26TH INTERNATIONAL CONFERENCE ON GEOINFORMATICS (GEOINFORMATICS 2018), 2018,
  • [46] Deep learning-based multi-modal computing with feature disentanglement for MRI image synthesis
    Fei, Yuchen
    Zhan, Bo
    Hong, Mei
    Wu, Xi
    Zhou, Jiliu
    Wang, Yan
    MEDICAL PHYSICS, 2021, 48 (07) : 3778 - 3789
  • [47] Multi-modal Fundus Image Registration with Deep Feature Matching and Image Scaling
    Kim, Ju-Chan
    Le, Duc-Tai
    Song, Su Jeong
    Son, Chang-Hwan
    Choo, Hyunseung
    PROCEEDINGS OF THE 2022 16TH INTERNATIONAL CONFERENCE ON UBIQUITOUS INFORMATION MANAGEMENT AND COMMUNICATION (IMCOM 2022), 2022,
  • [48] CovidSafe: A Deep Learning Framework for Covid Detection Using Multi-modal Approach
    Srikanth, Panigrahi
    Behera, Chandan Kumar
    Routhu, Srinivasa Rao
    NEW GENERATION COMPUTING, 2025, 43 (01)
  • [49] Multi-modal deep learning framework for damage detection in social media posts
    Zhang, Jiale
    Liao, Manyu
    Wang, Yanping
    Huang, Yifan
    Chen, Fuyu
    Makiko, Chiba
    PEERJ COMPUTER SCIENCE, 2024, 10
  • [50] Multi-modal deep learning model for auxiliary diagnosis of Alzheimer's disease
    Zhang, Fan
    Li, Zhenzhen
    Zhang, Boyan
    Dua, Haishun
    Wang, Binjie
    Zhang, Xinhong
    NEUROCOMPUTING, 2019, 361 : 185 - 195