Learning Diverse Models for End-to-End Ensemble Tracking

被引:5
作者
Wang, Ning [1 ]
Zhou, Wengang [1 ]
Li, Houqiang [1 ]
机构
[1] Univ Sci & Technol China, Dept Elect Engn & Informat Sci, CAS Key Lab Technol Geospatial Informat Proc & Ap, Hefei 230052, Peoples R China
关键词
Target tracking; Head; Predictive models; Visualization; Training; Computational modeling; Adaptation models; Visual tracking; ensemble learning; end-to-end tracker; diverse models; OBJECT TRACKING;
D O I
10.1109/TIP.2021.3051471
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In visual tracking, how to effectively model the target appearance using limited prior information remains an open problem. In this paper, we leverage an ensemble of diverse models to learn manifold representations for robust object tracking. The proposed ensemble framework includes a shared backbone network for efficient feature extraction and multiple head networks for independent predictions. Trained by the shared data within an identical structure, the mutually correlated head models heavily hinder the potential of ensemble learning. To shrink the representational overlaps among multiple models while encouraging the diversity of individual predictions, we propose the model diversity and response diversity regularization terms during training. By fusing these distinctive prediction results via a fusion module, the tracking variance caused by the distractor objects can be largely restrained. Our whole framework is end-to-end trained in a data-driven manner, avoiding the heuristic designs of multiple base models and fusion strategies. The proposed method achieves state-of-the-art results on seven challenging benchmarks while operating in real-time.
引用
收藏
页码:2220 / 2231
页数:12
相关论文
共 67 条
[1]  
[Anonymous], 2015, ACS SYM SER
[2]  
Bailer C, 2014, LECT NOTES COMPUT SC, V8695, P170, DOI 10.1007/978-3-319-10584-0_12
[3]  
Bansal Nitin, 2018, ADV NEUR IN
[4]   Staple: Complementary Learners for Real-Time Tracking [J].
Bertinetto, Luca ;
Valmadre, Jack ;
Golodetz, Stuart ;
Miksik, Ondrej ;
Torr, Philip H. S. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1401-1409
[5]   Fully-Convolutional Siamese Networks for Object Tracking [J].
Bertinetto, Luca ;
Valmadre, Jack ;
Henriques, Joao F. ;
Vedaldi, Andrea ;
Torr, Philip H. S. .
COMPUTER VISION - ECCV 2016 WORKSHOPS, PT II, 2016, 9914 :850-865
[6]   Learning Discriminative Model Prediction for Tracking [J].
Bhat, Goutam ;
Danelljan, Martin ;
Van Gool, Luc ;
Timofte, Radu .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6181-6190
[7]   Unveiling the Power of Deep Tracking [J].
Bhat, Goutam ;
Johnander, Joakim ;
Danelljan, Martin ;
Khan, Fahad Shahbaz ;
Felsberg, Michael .
COMPUTER VISION - ECCV 2018, PT II, 2018, 11206 :493-509
[8]  
Bolme DS, 2010, PROC CVPR IEEE, P2544, DOI 10.1109/CVPR.2010.5539960
[9]   Attentional Correlation Filter Network for Adaptive Visual Tracking [J].
Choi, Jongwon ;
Chang, Hyung Jin ;
Yun, Sangdoo ;
Fischer, Tobias ;
Demiris, Yiannis ;
Choi, Jin Young .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4828-4837
[10]   Visual Tracking Using Attention-Modulated Disintegration and Integration [J].
Choi, Jongwon ;
Chang, Hyung Jin ;
Jeong, Jiyeoup ;
Demiris, Yiannis ;
Choi, Jin Young .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :4321-4330