Learning Localization-Aware Target Confidence for Siamese Visual Tracking

被引:20
作者
Nie, Jiahao [1 ]
He, Zhiwei [1 ]
Yang, Yuxiang [2 ]
Gao, Mingyu [1 ]
Dong, Zhekang [3 ]
机构
[1] Hangzhou Dianzi Univ, Sch Elect Informat, Hangzhou 310018, Peoples R China
[2] Univ Sci & Technol China, Sch Control Sci & Engn, Hefei 230052, Peoples R China
[3] Zhejiang Univ, Sch Elect Engn, Hangzhou 310058, Peoples R China
基金
中国国家自然科学基金;
关键词
Target tracking; Task analysis; Feature extraction; Training; Location awareness; Visualization; Smoothing methods; Localization-aware components; Siamese tracking paradigm; task misalignment; OBJECT TRACKING;
D O I
10.1109/TMM.2022.3206668
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Siamese tracking paradigm has achieved great success, providing effective appearance discrimination and size estimation by classification and regression. While such a paradigm typically optimizes the classification and regression independently, leading to task misalignment (accurate prediction boxes have no high target confidence scores). In this paper, to alleviate this misalignment, we propose a novel tracking paradigm, called SiamLA. Within this paradigm, a series of simple, yet effective localization-aware components are introduced to generate localization-aware target confidence scores. Specifically, with the proposed localization-aware dynamic label (LADL) loss and localization-aware label smoothing (LALS) strategy, collaborative optimization between the classification and regression is achieved, enabling classification scores to be aware of location state, not just appearance similarity. Besides, we propose a separate localization-aware quality prediction (LAQP) branch to produce location quality scores to further modify the classification scores. To guide a more reliable modification, a novel localization-aware feature aggregation (LAFA) module is designed and embedded into this branch. Consequently, the resulting target confidence scores are more discriminative for the location state, allowing accurate prediction boxes tend to be predicted as high scores. Extensive experiments are conducted on six challenging benchmarks, including GOT10 k, TrackingNet, LaSOT, TNL2K, OTB100 and VOT2018. Our SiamLA achieves competitive performance in terms of both accuracy and efficiency. Furthermore, a stability analysis reveals that our tracking paradigm is relatively stable, implying that the paradigm is potential for real-world applications.
引用
收藏
页码:6194 / 6206
页数:13
相关论文
共 50 条
  • [1] Learning to Rank Proposals for Siamese Visual Tracking
    Tang, Feng
    Ling, Qiang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 8785 - 8796
  • [2] Deep Siamese Cross-Residual Learning for Robust Visual Tracking
    Wu, Fan
    Xu, Tingfa
    Guo, Jie
    Huang, Bo
    Xu, Chang
    Wang, Jihui
    Li, Xiangmin
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (20): : 15216 - 15227
  • [3] Learning Deep Lucas-Kanade Siamese Network for Visual Tracking
    Yao, Siyuan
    Han, Xiaoguang
    Zhang, Hua
    Wang, Xiao
    Cao, Xiaochun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 4814 - 4827
  • [4] Siamese Visual Object Tracking: A Survey
    Ondrasovic, Milan
    Tarabek, Peter
    IEEE ACCESS, 2021, 9 : 110149 - 110172
  • [5] SiamBAN: Target-Aware Tracking With Siamese Box Adaptive Network
    Chen, Zedu
    Zhong, Bineng
    Li, Guorong
    Zhang, Shengping
    Ji, Rongrong
    Tang, Zhenjun
    Li, Xianxian
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) : 5158 - 5173
  • [6] Antidecay LSTM for Siamese Tracking With Adversarial Learning
    Zhao, Fei
    Zhang, Ting
    Wu, Yi
    Tang, Ming
    Wang, Jinqiao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4475 - 4489
  • [7] SiamCAN: Real-Time Visual Tracking Based on Siamese Center-Aware Network
    Zhou, Wenzhang
    Wen, Longyin
    Zhang, Libo
    Du, Dawei
    Luo, Tiejian
    Wu, Yanjun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 3597 - 3609
  • [8] Manipulating Template Pixels for Model Adaptation of Siamese Visual Tracking
    Li, Zhenbang
    Li, Bing
    Gao, Jin
    Li, Liang
    Hu, Weiming
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 1690 - 1694
  • [9] SiamCorners: Siamese Corner Networks for Visual Tracking
    Yang, Kai
    He, Zhenyu
    Pei, Wenjie
    Zhou, Zikun
    Li, Xin
    Yuan, Di
    Zhang, Haijun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 1956 - 1967
  • [10] Target-Aware State Estimation for Visual Tracking
    Zhou, Zikun
    Li, Xin
    Fan, Nana
    Wang, Hongpeng
    He, Zhenyu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 2908 - 2920