Multi-Adapter RGBT Tracking

被引:155
作者
Li, Chenglong [1 ,3 ]
Lu, Andong [1 ]
Zheng, Aihua [1 ]
Tu, Zhengzheng [1 ]
Tang, Jin [1 ,2 ]
机构
[1] Anhui Univ, Sch Comp Sci & Technol, Minist Educ, Key Lab Intelligent Comp & Signal Proc, Hefei 230601, Peoples R China
[2] Key Lab Ind Image Proc & Anal Anhui Prov, Hefei 230601, Peoples R China
[3] Anhui Univ, Inst Phys Sci & Informat Technol, Hefei 230601, Peoples R China
来源
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW) | 2019年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCVW.2019.00279
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The task of RGBT tracking aims to take the complementary advantages from visible spectrum and thermal infrared data to achieve robust visual tracking, and receives more and more attention in recent years. Existing works focus on modality-specific information integration by introducing modality weights to achieve adaptive fusion or learning robust feature representations of different modalities. Although these methods could effectively deploy the modality-specific properties, they ignore the potential values of modality-shared cues as well as instance-aware information, which are crucial for effective fusion of different modalities in RGBT tracking. In this paper, we propose a novel Multi-Adapter convolutional Network (MANet) to jointly perform modality-shared, modality-specific and instance-aware feature learning in an end-to-end trained deep framework for RGBT tracking. We design three kinds of adapters within our network. In a specific, the generality adapter is to extract shared object representations, the modality adapter aims at encoding modality-specific information to deploy their complementary advantages, and the instance adapter is to model the appearance properties and temporal variations of a certain object. Moreover, to reduce computational complexity for real-time demand of visual tracking, we design a parallel structure of generic adapter and modality adapter. Extensive experiments on two RGBT tracking benchmark datasets demonstrate the outstanding performance of the proposed tracker against other state-of-the-art RGB and RGBT tracking algorithms.
引用
收藏
页码:2262 / 2270
页数:9
相关论文
共 35 条
[1]  
[Anonymous], 2018, ARXIV181109855
[2]  
Bertinetto L., 2016, PROCEEDINGS OF IEEE
[3]  
Chen B., 2018, P IEEE C EUR C COMP
[4]  
Danelljan M., 2016, P IEEE EUR C COMP VI
[5]   ECO: Efficient Convolution Operators for Tracking [J].
Danelljan, Martin ;
Bhat, Goutam ;
Khan, Fahad Shahbaz ;
Felsberg, Michael .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6931-6939
[6]   Convolutional Features for Correlation Filter Based Visual Tracking [J].
Danelljan, Martin ;
Hager, Gustav ;
Khan, Fahad Shahbaz ;
Felsberg, Michael .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOP (ICCVW), 2015, :621-629
[7]   Thermal cameras and applications: a survey [J].
Gade, Rikke ;
Moeslund, Thomas B. .
MACHINE VISION AND APPLICATIONS, 2014, 25 (01) :245-262
[8]  
Han B., 2017, PROCEEDINGS OF IEEE
[9]  
Hare S., 2011, P IEEE INT C COMP VI
[10]  
He K., 2016, CVPR, DOI [10.1109/CVPR.2016.90, DOI 10.1109/CVPR.2016.90]