Contraction of Dynamically Masked Deep Neural Networks for Efficient Video Processing

被引:2
|
作者
Rueckauer, Bodo [1 ,2 ,3 ]
Liu, Shih-Chii [1 ,2 ]
机构
[1] Univ Zurich, Inst Neuroinformat, CH-8057 Zurich, Switzerland
[2] Swiss Fed Inst Technol, CH-8057 Zurich, Switzerland
[3] Radboud Univ Nijmegen, Donders Inst Brain Cognit & Behav, NL-6525 XZ Nijmegen, Netherlands
关键词
Deep neural networks; network compression; Taylor approximation; masking;
D O I
10.1109/TCSVT.2021.3066241
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Sequential data such as video are characterized by spatio-temporal redundancies. As of yet, few deep learning algorithms exploit them to decrease the often massive cost during inference. This work leverages correlations in video data to reduce the size and run-time cost of deep neural networks. Drawing upon the simplicity of the typically used ReLU activation function, we replace this function by dynamically updating masks. The resulting network is a simple chain of matrix multiplications and bias additions, which can be contracted into a single weight matrix and bias vector. Inference then reduces to an affine transformation of the input sample with these contracted parameters. We show that the method is akin to approximating the neural network with a first-order Taylor expansion around a dynamically updating reference point. For triggering these updates, one static and three data-driven mechanisms are analyzed. We evaluate the proposed algorithm on a range of tasks, including pose estimation on surveillance data, road detection on KITTI driving scenes, object detection on ImageNet videos, as well as denoising MNIST digits, and obtain compression rates up to 3.6x.
引用
收藏
页码:621 / 633
页数:13
相关论文
共 50 条
  • [1] Dynamically Reconfigurable Deep Learning for Efficient Video Processing in Smart IoT Systems
    Eldash, Omar
    Frost, Adam
    Khali, Kasem
    Kumas, Ashok
    Bayoumi, Magdy
    2020 IEEE 6TH WORLD FORUM ON INTERNET OF THINGS (WF-IOT), 2020,
  • [2] Efficient Processing of Deep Neural Networks: A Tutorial and Survey
    Sze, Vivienne
    Chen, Yu-Hsin
    Yang, Tien-Ju
    Emer, Joel S.
    PROCEEDINGS OF THE IEEE, 2017, 105 (12) : 2295 - 2329
  • [3] Hardware Efficient Convolution Processing Unit for Deep Neural Networks
    Hazarika, Anakhi
    Poddar, Soumyajit
    Rahaman, Hafizur
    2019 2ND INTERNATIONAL SYMPOSIUM ON DEVICES, CIRCUITS AND SYSTEMS (ISDCS 2019), 2019,
  • [4] Linear Approximation of Deep Neural Networks for Efficient Inference on Video Data
    Rueckauer, Bodo
    Liu, Shih-Chii
    2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [5] Toward Efficient and Adaptive Design of Video Detection System with Deep Neural Networks
    Mao, Jiachen
    Yang, Qing
    Li, Ang
    Nixon, Kent W.
    Li, Hai
    Chen, Yiran
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2022, 21 (03)
  • [6] Design Considerations for Efficient Deep Neural Networks on Processing-in-Memory Accelerators
    Yang, Tien-Ju
    Sze, Vivienne
    2019 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2019,
  • [7] Dynamically evolving deep neural networks with continuous online learning
    Zhong, Yuan
    Zhou, Jing
    Li, Ping
    Gong, Jie
    INFORMATION SCIENCES, 2023, 646
  • [8] DeepRecon: Dynamically Reconfigurable Architecture for Accelerating Deep Neural Networks
    Rzayev, Tayyar
    Moradi, Saber
    Albonesi, David H.
    Manohar, Rajit
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 116 - 124
  • [9] Architecture of neural processing unit for deep neural networks
    Lee, Kyuho J.
    HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, 2021, 122 : 217 - 245
  • [10] The neural processing of masked speech
    Scott, Sophie K.
    McGettigan, Carolyn
    HEARING RESEARCH, 2013, 303 : 58 - 66