On-Device Deep Multi-Task Inference via Multi-Task Zipping

被引:5
|
作者
He, Xiaoxi [1 ]
Wang, Xu [2 ,3 ]
Zhou, Zimu [4 ]
Wu, Jiahang [2 ,3 ]
Yang, Zheng [2 ,3 ]
Thiele, Lothar [1 ]
机构
[1] Swiss Fed Inst Technol, Comp Engn & Networks Lab, CH-8052 Zurich, Switzerland
[2] Tsinghua Univ, TNList, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Sch Software, Beijing 100084, Peoples R China
[4] Singapore Management Univ, Sch Comp & Informat Syst, Singapore 178902, Singapore
关键词
Task analysis; Deep learning; Neurons; Training; Biological neural networks; Redundancy; Mobile computing; Deep neural networks; model compression; multi-task learning;
D O I
10.1109/TMC.2021.3124306
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks locally on-device. Yet the complexity of these deep models needs to be trimmed down both within model and cross-model to fit in mobile storage and memory. Previous studies squeeze the redundancy within a single model. In this work, we aim to reduce the redundancy across multiple models. We propose Multi-Task Zipping (MTZ), a framework to automatically merge correlated, pre-trained deep neural networks for cross-model compression. Central in MTZ is a layer-wise neuron sharing and incoming weight updating scheme that induces a minimal change in the error function. MTZ inherits information from each model and demands light retraining to re-boost the accuracy of individual tasks. MTZ supports typical network layers (fully-connected, convolutional and residual) and applies to inference tasks with different input domains. Evaluations show that MTZ can fully merge the hidden layers of two VGG-16 networks with a 3.18% increase in the test error averaged on ImageNet for object classification and CelebA for facial attribute classification, or share 39.61% parameters between the two networks with < 0.5% increase in the test errors. The number of iterations to retrain the combined network is at least 17.8x lower than that of training a single VGG-16 network. Moreover, MTZ can effectively merge nine residual networks for diverse inference tasks and models for different input domains. And with the model merged by MTZ, the latency to switch between these tasks on memory-constrained devices is reduced by 8.71x.
引用
收藏
页码:2878 / 2891
页数:14
相关论文
共 50 条
  • [31] Multi-task Crowdsourcing via an Optimization Framework
    Zhou, Yao
    Ying, Lei
    He, Jingrui
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2019, 13 (03)
  • [32] Multi-task clustering via domain adaptation
    Zhang, Zhihao
    Zhou, Jie
    PATTERN RECOGNITION, 2012, 45 (01) : 465 - 473
  • [33] Identification of Distorted RF Components via Deep Multi-Task Learning
    Aygul, Mehmet Ali
    Memisoglu, Ebubekir
    Cirpan, Hakan Ali
    Arslan, Huseyin
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [34] Joint face alignment and segmentation via deep multi-task learning
    Zhao, Yucheng
    Tang, Fan
    Dong, Weiming
    Huang, Feiyue
    Zhang, Xiaopeng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (10) : 13131 - 13148
  • [35] Joint face alignment and segmentation via deep multi-task learning
    Yucheng Zhao
    Fan Tang
    Weiming Dong
    Feiyue Huang
    Xiaopeng Zhang
    Multimedia Tools and Applications, 2019, 78 : 13131 - 13148
  • [36] Multi-step Forecasting via Multi-task Learning
    Jawed, Shayan
    Rashed, Ahmed
    Schmidt-Thieme, Lars
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 790 - 799
  • [37] Multi-Asset Market Making via Multi-Task Deep Reinforcement Learning
    Haider, Abbas
    Hawe, Glenn, I
    Wang, Hui
    Scotney, Bryan
    MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE (LOD 2021), PT II, 2022, 13164 : 353 - 364
  • [38] Interest Inference via Structure-Constrained Multi-Source Multi-Task Learning
    Song, Xuemeng
    Nie, Liqiang
    Zhang, Luming
    Liu, Maofu
    Chua, Tat-Seng
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 2371 - 2377
  • [39] Multi-task Deep Reinforcement Learning for Scalable Parallel Task Scheduling
    Zhang, Lingxin
    Qi, Qi
    Wang, Jingyu
    Sun, Haifeng
    Liao, Jianxin
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2992 - 3001
  • [40] Quantifying Task Priority for Multi-Task Optimization
    Jeong, Wooseong
    Yoon, Kuk-Jin
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 363 - 372