UCT: Learning Unified Convolutional Networks for Real-time Visual Tracking

被引:68
|
作者
Zhu, Zheng [1 ,2 ]
Huang, Guan [3 ]
Zou, Wei [1 ,2 ]
Du, Dalong [3 ]
Huang, Chang [3 ]
机构
[1] Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Horizon Robot Inc, Beijing, Peoples R China
基金
国家高技术研究发展计划(863计划); 中国国家自然科学基金;
关键词
OBJECT TRACKING;
D O I
10.1109/ICCVW.2017.231
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNN) based tracking approaches have shown favorable performance in recent benchmarks. Nonetheless, the chosen CNN features are always pre-trained in different task and individual components in tracking systems are learned separately, thus the achieved tracking performance may be suboptimal. Besides, most of these trackers are not designed towards realtime applications because of their time-consuming feature extraction and complex optimization details. In this paper, we propose an end-to-end framework to learn the convolutional features and perform the tracking process simultaneously, namely, a unified convolutional tracker (UCT). Specifically, The UCT treats feature extractor and tracking process (ridge regression) both as convolution operation and trains them jointly, enabling learned CNN features are tightly coupled to tracking process. In online tracking, an efficient updating method is proposed by introducing peak-versus-noise ratio (PNR) criterion, and scale changes are handled efficiently by incorporating a scale branch into network. The proposed approach results in superior tracking performance, while maintaining real-time speed. The standard UCT and UCT-Lite can track generic objects at 41 FPS and 154 FPS without further optimization, respectively. Experiments are performed on four challenging benchmark tracking datasets: OTB2013, OTB2015, VOT2014 and VOT2015, and our method achieves state-ofthe-art results on these benchmarks compared with other real-time trackers.
引用
收藏
页码:1973 / 1982
页数:10
相关论文
共 50 条
  • [1] REGION-BASED FULLY CONVOLUTIONAL SIAMESE NETWORKS FOR ROBUST REAL-TIME VISUAL TRACKING
    Yang, Longchao
    Jiang, Peilin
    Wang, Fei
    Wang, Xuan
    2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 2567 - 2571
  • [3] Robust and Real-Time Visual Tracking with Triplet Convolutional Neural Network
    Kim, Jung Uk
    Kim, Hak Gu
    Ro, Yong Man
    PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, : 280 - 286
  • [4] Learning complementary Siamese networks for real-time high-performance visual tracking
    Tan, Ke
    Xu, Ting-Bing
    Wei, Zhenzhong
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 80
  • [5] Deeper and Wider Siamese Networks for Real-Time Visual Tracking
    Zhang, Zhipeng
    Peng, Houwen
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4586 - 4595
  • [6] Leveraging convolutional neural networks for real-time student attendance tracking
    Yadav, Rajesh
    Gupta, Swati
    Malik, Meenakshi
    Yadav, Poonam
    JOURNAL OF INFORMATION & OPTIMIZATION SCIENCES, 2025, 46 (01): : 43 - 52
  • [7] IMPROVED REAL-TIME VISUAL TRACKING VIA ADVERSARIAL LEARNING
    Zhong, Haoxiang
    Yan, Xiyu
    Jiang, Yong
    Xia, Shu-Tao
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 1853 - 1857
  • [8] Learning Feature Channel Weighting for Real-Time Visual Tracking
    Li, Zhetao
    Zhang, Jie
    Li, Yanchun
    Zhu, Jiang
    Long, Saiqin
    Xue, Dengfeng
    Fan, Longfei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 2190 - 2200
  • [9] Kernelized temporal locality learning for real-time visual tracking
    Liu, Fanghui
    Zhou, Tao
    Fu, Keren
    Yang, Jie
    PATTERN RECOGNITION LETTERS, 2017, 90 : 72 - 79
  • [10] LEARNING ADAPTIVE SELECTION NETWORK FOR REAL-TIME VISUAL TRACKING
    Xiong, Jiangfeng
    Xu, Xiangmin
    Cai, Bolun
    Xing, Xiaofen
    Guo, Kailing
    2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2018,