On-Device Deep Multi-Task Inference via Multi-Task Zipping

被引:5
|
作者
He, Xiaoxi [1 ]
Wang, Xu [2 ,3 ]
Zhou, Zimu [4 ]
Wu, Jiahang [2 ,3 ]
Yang, Zheng [2 ,3 ]
Thiele, Lothar [1 ]
机构
[1] Swiss Fed Inst Technol, Comp Engn & Networks Lab, CH-8052 Zurich, Switzerland
[2] Tsinghua Univ, TNList, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Sch Software, Beijing 100084, Peoples R China
[4] Singapore Management Univ, Sch Comp & Informat Syst, Singapore 178902, Singapore
关键词
Task analysis; Deep learning; Neurons; Training; Biological neural networks; Redundancy; Mobile computing; Deep neural networks; model compression; multi-task learning;
D O I
10.1109/TMC.2021.3124306
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks locally on-device. Yet the complexity of these deep models needs to be trimmed down both within model and cross-model to fit in mobile storage and memory. Previous studies squeeze the redundancy within a single model. In this work, we aim to reduce the redundancy across multiple models. We propose Multi-Task Zipping (MTZ), a framework to automatically merge correlated, pre-trained deep neural networks for cross-model compression. Central in MTZ is a layer-wise neuron sharing and incoming weight updating scheme that induces a minimal change in the error function. MTZ inherits information from each model and demands light retraining to re-boost the accuracy of individual tasks. MTZ supports typical network layers (fully-connected, convolutional and residual) and applies to inference tasks with different input domains. Evaluations show that MTZ can fully merge the hidden layers of two VGG-16 networks with a 3.18% increase in the test error averaged on ImageNet for object classification and CelebA for facial attribute classification, or share 39.61% parameters between the two networks with < 0.5% increase in the test errors. The number of iterations to retrain the combined network is at least 17.8x lower than that of training a single VGG-16 network. Moreover, MTZ can effectively merge nine residual networks for diverse inference tasks and models for different input domains. And with the model merged by MTZ, the latency to switch between these tasks on memory-constrained devices is reduced by 8.71x.
引用
收藏
页码:2878 / 2891
页数:14
相关论文
共 50 条
  • [1] Multi-Task Adapters for On-Device Audio Inference
    Tagliasacchi, Marco
    Quitry, Felix de Chaumont
    Roblek, Dominik
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 630 - 634
  • [2] Multi-Task Zipping via Layer-wise Neuron Sharing
    He, Xiaoxi
    Zhou, Zimu
    Thiele, Lothar
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [4] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [5] LiteMuL: A Lightweight On-Device Sequence Tagger using Multi-task Learning
    Kumari, Sonal
    Agarwal, Vibhav
    Challa, Bharath
    Chalamalasetti, Kranti
    Ghosh, Sourav
    Harshavardhana, Harshavardhana
    Raja, Barath Raj Kandur
    2021 IEEE 15TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC 2021), 2021, : 1 - 8
  • [6] Pareto Multi-task Deep Learning
    Riccio, Salvatore D.
    Dyankov, Deyan
    Jansen, Giorgio
    Di Fatta, Giuseppe
    Nicosia, Giuseppe
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 132 - 141
  • [7] Multi-Receiver Task-Oriented Communications via Multi-Task Deep Learning
    Sagduyu, Yalin E.
    Erpek, Tugba
    Yener, Aylin
    Ulukus, Sennur
    2023 IEEE FUTURE NETWORKS WORLD FORUM, FNWF, 2024,
  • [8] Multi-Task Temporal Shift Attention Networks for On-Device Contactless Vitals Measurement
    Liu, Xin
    Fromm, Josh
    Patel, Shwetak
    McDuff, Daniel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [9] Information Cascades Modeling via Deep Multi-Task Learning
    Chen, Xueqin
    Zhang, Kunpeng
    Zhou, Fan
    Trajcevski, Goce
    Zhong, Ting
    Zhang, Fengli
    PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 885 - 888
  • [10] Multi-task averaging
    Department of Electrical Engineering, University of Washington, Seattle, WA 98103, United States
    Adv. neural inf. proces. syst., 1600, (1169-1177):