Multi-Modal Object Tracking and Image Fusion With Unsupervised Deep Learning

被引:11
|
作者
LaHaye, Nicholas [1 ,2 ]
Ott, Jordan [1 ]
Garay, Michael J. [3 ]
El-Askary, Hesham Mohamed [4 ,5 ,6 ]
Linstead, Erik [5 ,7 ]
机构
[1] Chapman Univ, Computat & Data Sci Dept, Orange, CA 92866 USA
[2] CALTECH, Jet Prop Lab, Proc Algorithms & Calibrat Engn Grp, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[3] CALTECH, Jet Prop Lab, 4800 Oak Grove Dr, Pasadena, CA 91109 USA
[4] Chapman Univ, Ctr Excellence Earth Syst Modeling & Observat, Orange, CA 92866 USA
[5] Chapman Univ, Schmid Coll Sci & Technol, Orange, CA 92866 USA
[6] Alexandria Univ, Fac Sci, Dept Environm Sci, Alexandria 21522, Egypt
[7] Chapman Univ, Machine Learning & Assist Technol Lab, Orange, CA 92866 USA
关键词
Bigdata applications; clustering; computer vision; deep belief networks (DBNs); deep learning; CLASSIFICATION; MISR;
D O I
10.1109/JSTARS.2019.2920234
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The number of different modalities for remote sensors continues to grow, bringing with it an increase in the volume and complexity of the data being collected. Although these datasets individually provide valuable information, in aggregate they provide additional opportunities to discover meaningful patterns on a large scale. However, the ability to combine and analyze disparate datasets is challenged by the potentially vast parameter space that results from aggregation. Each dataset in itself requires instrument-specific and dataset-specific knowledge. If the intention is to use multiple, diverse datasets, one needs an understanding of how to translate and combine these parameters in an efficient and effective manner. While there are established techniques for combining datasets from specific domains or platforms, there is no generic, automated method that can address the problem in general. Here, we discuss the application of deep learning to track objects across different image-like data-modalities, given data in a similar spatio-temporal range, and automatically co-register these images. Using deep belief networks combined with unsupervised learning methods, we are able to recognize and separate different objects within image-like data in a structured manner, thus making progress toward the ultimate goal of a generic tracking and fusion pipeline requiring minimal human intervention.
引用
收藏
页码:3056 / 3066
页数:11
相关论文
共 50 条
  • [21] Deep Multi-modal Object Detection for Autonomous Driving
    Ennajar, Amal
    Khouja, Nadia
    Boutteau, Remi
    Tlili, Fethi
    2021 18TH INTERNATIONAL MULTI-CONFERENCE ON SYSTEMS, SIGNALS & DEVICES (SSD), 2021, : 7 - 11
  • [22] A Comprehensive Survey on Deep Learning Multi-Modal Fusion: Methods, Technologies and Applications
    Jiao, Tianzhe
    Guo, Chaopeng
    Feng, Xiaoyue
    Chen, Yuming
    Song, Jie
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 80 (01): : 1 - 35
  • [23] Deep-Learning-Based Multi-Modal Fusion for Fast MR Reconstruction
    Xiang, Lei
    Chen, Yong
    Chang, Weitang
    Zhan, Yiqiang
    Lin, Weili
    Wang, Qian
    Shen, Dinggang
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2019, 66 (07) : 2105 - 2114
  • [24] A Deep Learning-Enhanced Multi-Modal Sensing Platform for Robust Human Object Detection and Tracking in Challenging Environments
    Cheng, Peng
    Xiong, Zinan
    Bao, Yajie
    Zhuang, Ping
    Zhang, Yunqi
    Blasch, Erik
    Chen, Genshe
    ELECTRONICS, 2023, 12 (16)
  • [25] Multi-Modal Hand-Object Pose Estimation With Adaptive Fusion and Interaction Learning
    Hoang, Dinh-Cuong
    Tan, Phan Xuan
    Nguyen, Anh-Nhat
    Vu, Duy-Quang
    Vu, Van-Duc
    Nguyen, Thu-Uyen
    Hoang, Ngoc-Anh
    Phan, Khanh-Toan
    Tran, Duc-Thanh
    Nguyen, Van-Thiep
    Duong, Quang-Tri
    Ho, Ngoc-Trung
    Tran, Cong-Trinh
    Duong, Van-Hiep
    Ngo, Phuc-Quan
    IEEE ACCESS, 2024, 12 : 54339 - 54351
  • [26] Comparison of Multi-Modal Large Language Models with Deep Learning Models for Medical Image Classification
    Than, Joel Chia Ming
    Vong, Wan Tze
    Yong, Kelvin Sheng Chek
    2024 IEEE 8TH INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING APPLICATIONS, ICSIPA, 2024,
  • [27] Deep Multi-modal Learning with Cascade Consensus
    Yang, Yang
    Wu, Yi-Feng
    Zhan, De-Chuan
    Jiang, Yuan
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II, 2018, 11013 : 64 - 72
  • [28] InstaIndoor and multi-modal deep learning for indoor scene recognition
    Glavan, Andreea
    Talavera, Estefania
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (09) : 6861 - 6877
  • [29] Correspondence Learning for Deep Multi-Modal Recognition and Fraud Detection
    Park, Jongchan
    Kim, Min-Hyun
    Choi, Dong-Geol
    ELECTRONICS, 2021, 10 (07)
  • [30] Multi-Modal Deep Learning for Assessing Surgeon Technical Skill
    Kasa, Kevin
    Burns, David
    Goldenberg, Mitchell G.
    Selim, Omar
    Whyne, Cari
    Hardisty, Michael
    SENSORS, 2022, 22 (19)