Multi-Modal Object Tracking and Image Fusion With Unsupervised Deep Learning

被引:11
|
作者
LaHaye, Nicholas [1 ,2 ]
Ott, Jordan [1 ]
Garay, Michael J. [3 ]
El-Askary, Hesham Mohamed [4 ,5 ,6 ]
Linstead, Erik [5 ,7 ]
机构
[1] Chapman Univ, Computat & Data Sci Dept, Orange, CA 92866 USA
[2] CALTECH, Jet Prop Lab, Proc Algorithms & Calibrat Engn Grp, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[3] CALTECH, Jet Prop Lab, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[4] Chapman Univ, Ctr Excellence Earth Syst Modeling & Observat, Orange, CA 92866 USA
[5] Chapman Univ, Schmid Coll Sci & Technol, Orange, CA 92866 USA
[6] Alexandria Univ, Fac Sci, Dept Environm Sci, Alexandria 21522, Egypt
[7] Chapman Univ, Machine Learning & Assist Technol Lab, Orange, CA 92866 USA
关键词
Bigdata applications; clustering; computer vision; deep belief networks (DBNs); deep learning; CLASSIFICATION; MISR;
D O I
10.1109/JSTARS.2019.2920234
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The number of different modalities for remote sensors continues to grow, bringing with it an increase in the volume and complexity of the data being collected. Although these datasets individually provide valuable information, in aggregate they provide additional opportunities to discover meaningful patterns on a large scale. However, the ability to combine and analyze disparate datasets is challenged by the potentially vast parameter space that results from aggregation. Each dataset in itself requires instrument-specific and dataset-specific knowledge. If the intention is to use multiple, diverse datasets, one needs an understanding of how to translate and combine these parameters in an efficient and effective manner. While there are established techniques for combining datasets from specific domains or platforms, there is no generic, automated method that can address the problem in general. Here, we discuss the application of deep learning to track objects across different image-like data-modalities, given data in a similar spatio-temporal range, and automatically co-register these images. Using deep belief networks combined with unsupervised learning methods, we are able to recognize and separate different objects within image-like data in a structured manner, thus making progress toward the ultimate goal of a generic tracking and fusion pipeline requiring minimal human intervention.
引用
收藏
页码:3056 / 3066
页数:11
相关论文
共 50 条
  • [1] A Quantitative Validation of Multi-Modal Image Fusion and Segmentation for Object Detection and Tracking
    LaHaye, Nicholas
    Garay, Michael J.
    Bue, Brian D.
    El-Askary, Hesham
    Linstead, Erik
    REMOTE SENSING, 2021, 13 (12)
  • [2] An Unsupervised Deep Learning Method for Diffeomorphic Mono-and Multi-modal Image Registration
    Theljani, Anis
    Chen, Ke
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2019, 2020, 1065 : 317 - 326
  • [3] Deep learning supported breast cancer classification with multi-modal image fusion
    Hamdy, Eman
    Zaghloul, Mohamed Saad
    Badawy, Osama
    2021 22ND INTERNATIONAL ARAB CONFERENCE ON INFORMATION TECHNOLOGY (ACIT), 2021, : 319 - 325
  • [4] CIRF: Coupled Image Reconstruction and Fusion Strategy for Deep Learning Based Multi-Modal Image Fusion
    Zheng, Junze
    Xiao, Junyan
    Wang, Yaowei
    Zhang, Xuming
    SENSORS, 2024, 24 (11)
  • [5] Deep unsupervised multi-modal fusion network for detecting driver distraction
    Zhang, Yuxin
    Chen, Yiqiang
    Gao, Chenlong
    NEUROCOMPUTING, 2021, 421 : 26 - 38
  • [6] A comparative review on multi-modal sensors fusion based on deep learning
    Tang, Qin
    Liang, Jing
    Zhu, Fangqi
    SIGNAL PROCESSING, 2023, 213
  • [7] Deep learning and multi-modal fusion for real-time multi-object tracking: Algorithms, challenges, datasets, and comparative study
    Wang, Xuan
    Sun, Zhaojie
    Chehri, Abdellah
    Jeon, Gwanggil
    Song, Yongchao
    INFORMATION FUSION, 2024, 105
  • [8] Exploring Fusion Strategies in Deep Learning Models for Multi-Modal Classification
    Zhang, Duoyi
    Nayak, Richi
    Bashar, Md Abul
    DATA MINING, AUSDM 2021, 2021, 1504 : 102 - 117
  • [9] Classifying Excavator Operations with Fusion Network of Multi-modal Deep Learning Models
    Kim, Jin-Young
    Cho, Sung-Bae
    14TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING MODELS IN INDUSTRIAL AND ENVIRONMENTAL APPLICATIONS (SOCO 2019), 2020, 950 : 25 - 34
  • [10] Efficient Multi-Modal Image Fusion Deep Learning Techniques for Classifying Neurodegenerative Disease Type
    Joy, Johnsymol
    Selvan, Mercy Paul
    TRAITEMENT DU SIGNAL, 2025, 42 (01) : 267 - 275