Adaptive learning-enhanced lightweight network for real-time vehicle density estimation

被引:1
作者
Qin, Ling-Xiao [1 ]
Sun, Hong-Mei [1 ]
Duan, Xiao-Meng [1 ]
Che, Cheng-Yue [1 ]
Jia, Rui-Sheng [1 ]
机构
[1] Shandong Univ Sci & Technol, Coll Comp Sci & Engn, Qingdao 266590, Peoples R China
关键词
Vehicle density estimation; Lightweight networks; Adaptive integration block; Edge devices; CROWD; NET;
D O I
10.1007/s00371-024-03572-3
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In order to maintain competitive density estimation performance, most of the existing works design cumbersome network structures to extract and refine vehicle features, resulting in huge computational resource consumption and storage burden during the inference process, which severely limits their deployment scope and makes it difficult to be applied in practical scenarios. To solve the above problems, we propose a lightweight network for real-time vehicle density estimation (LSENet). Specifically, the network consists of three parts: a pre-trained heavy teacher network, an adaptive integration block and a lightweight student network. First, a teacher network based on a deep single-column transformer is designed as a means to provide effective global dependency and vehicle distribution knowledge for the student network to learn. Second, to address the intermediate layer mismatch and dimensionality inconsistency between the teacher network and the student network, an adaptive integration block is designed to efficiently guide the student network learning by dynamically assigning the self-attention heads that has the most influence on the network decision as a source of distilled knowledge. Finally, to complement the fine-grained features, CNN blocks are designed in parallel with the student network transformer backbone as a way to improve the network's ability to capture vehicle details. Extensive experiments on two vehicle benchmark datasets, TRANCOS and VisDrone2019, show that LSENet achieves an optimal trade-off between density estimation accuracy and operational speed compared to other state-of-the-art methods and is therefore suitable for deployment on computationally resource-poor edge devices. Our codes will be available at https://github.com/goudaner1/LSENet.
引用
收藏
页码:2857 / 2873
页数:17
相关论文
共 66 条
  • [1] Cao W., 2022, Adv. Neural Inf. Process. Syst., V35, P15394
  • [2] Scale Aggregation Network for Accurate and Efficient Crowd Counting
    Cao, Xinkun
    Wang, Zhipeng
    Zhao, Yanyun
    Su, Fei
    [J]. COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 : 757 - 773
  • [3] Flounder-Net: An efficient CNN for crowd counting by aerial photography
    Chen, Jingyu
    Xiu, Shengjie
    Chen, Xiang
    Guo, Hao
    Xie, Xiaohua
    [J]. NEUROCOMPUTING, 2021, 420 : 82 - 89
  • [4] Vehicle Detection in Satellite Images by Hybrid Deep Convolutional Neural Networks
    Chen, Xueyun
    Xiang, Shiming
    Liu, Cheng-Lin
    Pan, Chun-Hong
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2014, 11 (10) : 1797 - 1801
  • [5] Dosovitskiy A., 2021, arXiv
  • [6] Redesigning Multi-Scale Neural Network for Crowd Counting
    Du, Zhipeng
    Shi, Miaojing
    Deng, Jiankang
    Zafeiriou, Stefanos
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 3664 - 3678
  • [7] Fan QF, 2016, IEEE INT VEH SYM, P124, DOI 10.1109/IVS.2016.7535375
  • [8] PCC Net: Perspective Crowd Counting via Spatial Convolutional Network
    Gao, Junyu
    Wang, Qi
    Li, Xuelong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (10) : 3486 - 3498
  • [9] Lopez JG, 2019, INT SYMP IMAGE SIG, P209, DOI [10.1109/ISPA.2019.8868508, 10.1109/ispa.2019.8868508]
  • [10] Extremely Overlapping Vehicle Counting
    Guerrero-Gomez-Olmedo, Ricardo
    Torre-Jimenez, Beatriz
    Lopez-Sastre, Roberto
    Maldonado-Bascon, Saturnino
    Onoro-Rubio, Daniel
    [J]. PATTERN RECOGNITION AND IMAGE ANALYSIS (IBPRIA 2015), 2015, 9117 : 423 - 431