On-Device Deep Multi-Task Inference via Multi-Task Zipping

被引:5
|
作者
He, Xiaoxi [1 ]
Wang, Xu [2 ,3 ]
Zhou, Zimu [4 ]
Wu, Jiahang [2 ,3 ]
Yang, Zheng [2 ,3 ]
Thiele, Lothar [1 ]
机构
[1] Swiss Fed Inst Technol, Comp Engn & Networks Lab, CH-8052 Zurich, Switzerland
[2] Tsinghua Univ, TNList, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Sch Software, Beijing 100084, Peoples R China
[4] Singapore Management Univ, Sch Comp & Informat Syst, Singapore 178902, Singapore
关键词
Task analysis; Deep learning; Neurons; Training; Biological neural networks; Redundancy; Mobile computing; Deep neural networks; model compression; multi-task learning;
D O I
10.1109/TMC.2021.3124306
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks locally on-device. Yet the complexity of these deep models needs to be trimmed down both within model and cross-model to fit in mobile storage and memory. Previous studies squeeze the redundancy within a single model. In this work, we aim to reduce the redundancy across multiple models. We propose Multi-Task Zipping (MTZ), a framework to automatically merge correlated, pre-trained deep neural networks for cross-model compression. Central in MTZ is a layer-wise neuron sharing and incoming weight updating scheme that induces a minimal change in the error function. MTZ inherits information from each model and demands light retraining to re-boost the accuracy of individual tasks. MTZ supports typical network layers (fully-connected, convolutional and residual) and applies to inference tasks with different input domains. Evaluations show that MTZ can fully merge the hidden layers of two VGG-16 networks with a 3.18% increase in the test error averaged on ImageNet for object classification and CelebA for facial attribute classification, or share 39.61% parameters between the two networks with < 0.5% increase in the test errors. The number of iterations to retrain the combined network is at least 17.8x lower than that of training a single VGG-16 network. Moreover, MTZ can effectively merge nine residual networks for diverse inference tasks and models for different input domains. And with the model merged by MTZ, the latency to switch between these tasks on memory-constrained devices is reduced by 8.71x.
引用
收藏
页码:2878 / 2891
页数:14
相关论文
共 50 条
  • [41] Multi-task Transfer with Practice
    Pattnaik, Upasana
    Lee, Minwoo
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [42] Boosted multi-task learning
    Olivier Chapelle
    Pannagadatta Shivaswamy
    Srinivas Vadrevu
    Kilian Weinberger
    Ya Zhang
    Belle Tseng
    Machine Learning, 2011, 85 : 149 - 173
  • [43] An overview of multi-task learning
    Zhang, Yu
    Yang, Qiang
    NATIONAL SCIENCE REVIEW, 2018, 5 (01) : 30 - 43
  • [44] Multi-task Network Embedding
    Xu, Linchuan
    Wei, Xiaokai
    Cao, Jiannong
    Yu, Philip S.
    2017 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2017, : 571 - 580
  • [45] Task Variance Regularized Multi-Task Learning
    Mao, Yuren
    Wang, Zekai
    Liu, Weiwei
    Lin, Xuemin
    Hu, Wenbin
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (08) : 8615 - 8629
  • [46] Task Switching Network for Multi-task Learning
    Sun, Guolei
    Probst, Thomas
    Paudel, Danda Pani
    Popovic, Nikola
    Kanakis, Menelaos
    Patel, Jagruti
    Dai, Dengxin
    Van Gool, Luc
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8271 - 8280
  • [47] Multi-task subspace clustering
    Zhong, Guo
    Pun, Chi -Man
    INFORMATION SCIENCES, 2024, 661
  • [48] Multi-task Survival Analysis
    Wang, Lu
    Li, Yan
    Zhou, Jiayu
    Zhu, Dongxiao
    Ye, Jieping
    2017 17TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2017, : 485 - 494
  • [49] Multi-task network embedding
    Linchuan Xu
    Xiaokai Wei
    Jiannong Cao
    Philip S. Yu
    International Journal of Data Science and Analytics, 2019, 8 : 183 - 198
  • [50] On Partial Multi-Task Learning
    He, Yi
    Wu, Baijun
    Wu, Di
    Wu, Xindong
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1174 - 1181