DensSiam: End-to-End Densely-Siamese Network with Self-Attention Model for Object Tracking

被引:44
作者
Abdelpakey, Mohamed H. [1 ]
Shehata, Mohamed S. [1 ]
Mohamed, Mostafa M. [2 ,3 ]
机构
[1] Mem Univ Newfoundland, Fac Engn & Appl Sci, St John, NF A1B 3X5, Canada
[2] Univ Calgary, Elect & Comp Engn Dept, Calgary, AB, Canada
[3] Helwan Univ, Biomed Engn Dept, Helwan, Egypt
来源
ADVANCES IN VISUAL COMPUTING, ISVC 2018 | 2018年 / 11241卷
关键词
Object tracking; Siamese-network; Densely-Siamese; Self-attention;
D O I
10.1007/978-3-030-03801-4_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional Siamese neural networks have been recently used to track objects using deep features. Siamese architecture can achieve real time speed, however it is still difficult to find a Siamese architecture that maintains the generalization capability, high accuracy and speed while decreasing the number of shared parameters especially when it is very deep. Furthermore, a conventional Siamese architecture usually processes one local neighborhood at a time, which makes the appearance model local and non-robust to appearance changes. To overcome these two problems, this paper proposes DensSiam, a novel convolutional Siamese architecture, which uses the concept of dense layers and connects each dense layer to all layers in a feed-forward fashion with a similarity-learning function. DensSiam also includes a Self-Attention mechanism to force the network to pay more attention to the non-local features during offline training. Extensive experiments are performed on four tracking benchmarks: OTB2013 and OTB2015 for validation set; and VOT2015, VOT2016 and VOT2017 for testing set. The obtained results show that DensSiam achieves superior results on these benchmarks compared to other current state-of-the-art methods.
引用
收藏
页码:463 / 473
页数:11
相关论文
共 36 条
[1]  
Abadi M., 2015, PREPRINT
[2]  
[Anonymous], PROC ICLR 2015
[3]  
[Anonymous], 2017, arXiv preprint arXiv:1704.04057
[4]  
[Anonymous], 2018, IDEAS HIST MOD CHINA, DOI DOI 10.1163/9789004385580_002
[5]  
[Anonymous], 2016, CVPR CVPR
[6]  
[Anonymous], 2017, ARXIV170406904
[7]   Staple: Complementary Learners for Real-Time Tracking [J].
Bertinetto, Luca ;
Valmadre, Jack ;
Golodetz, Stuart ;
Miksik, Ondrej ;
Torr, Philip H. S. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1401-1409
[8]   Fully-Convolutional Siamese Networks for Object Tracking [J].
Bertinetto, Luca ;
Valmadre, Jack ;
Henriques, Joao F. ;
Vedaldi, Andrea ;
Torr, Philip H. S. .
COMPUTER VISION - ECCV 2016 WORKSHOPS, PT II, 2016, 9914 :850-865
[9]   DeepDriving: Learning Affordance for Direct Perception in Autonomous Driving [J].
Chen, Chenyi ;
Seff, Ari ;
Kornhauser, Alain ;
Xiao, Jianxiong .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2722-2730
[10]   Online Multi-Object Tracking Using CNN-based Single Object Tracker with Spatial-Temporal Attention Mechanism [J].
Chu, Qi ;
Ouyang, Wanli ;
Li, Hongsheng ;
Wang, Xiaogang ;
Liu, Bin ;
Yu, Nenghai .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :4846-4855