Multi-Task Learning with Knowledge Distillation for Dense Prediction

被引:2
|
作者
Xu, Yangyang [1 ,2 ]
Yang, Yibo [4 ]
Zhang, Lefei [1 ,2 ,3 ]
机构
[1] Wuhan Univ, Inst Artificial Intelligence, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[3] Hubei Luojia Lab, Wuhan, Peoples R China
[4] King Abdullah Univ Sci & Technol, Jeddah, Saudi Arabia
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01970
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While multi-task learning (MTL) has become an attractive topic, its training usually poses more difficulties than the single-task case. How to successfully apply knowledge distillation into MTL to improve training efficiency and model performance is still a challenging problem. In this paper, we introduce a new knowledge distillation procedure with an alternative match for MTL of dense prediction based on two simple design principles. First, for memory and training efficiency, we use a single strong multitask model as a teacher during training instead of multiple teachers, as widely adopted in existing studies. Second, we employ a less sensitive Cauchy-Schwarz (CS) divergence instead of the Kullback-Leibler (KL) divergence and propose a CS distillation loss accordingly. With the less sensitive divergence, our knowledge distillation with an alternative match is applied for capturing inter-task and intratask information between the teacher model and the student model of each task, thereby learning more "dark knowledge" for effective distillation. We conducted extensive experiments on dense prediction datasets, including NYUD-v2 and PASCAL-Context, for multiple vision tasks, such as semantic segmentation, human parts segmentation, depth estimation, surface normal estimation, and boundary detection. The results show that our proposed method decidedly improves model performance and the practical inference efficiency.
引用
收藏
页码:21493 / 21502
页数:10
相关论文
共 50 条
  • [31] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [32] Knowledge triple mining via multi-task learning
    Zhang, Zhao
    Zhuang, Fuzhen
    Li, Xuebing
    Niu, Zheng-Yu
    He, Jia
    He, Qing
    Xiong, Hui
    INFORMATION SYSTEMS, 2019, 80 : 64 - 75
  • [33] Going Beyond Multi-Task Dense Prediction with Synergy Embedding Models
    Huang, Huimin
    Huang, Yawen
    Lin, Lanfen
    Tong, Ruofeng
    Chen, Yen-Wei
    Zheng, Hao
    Li, Yuexiang
    Zheng, Yefeng
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 28181 - 28190
  • [34] Structured Multi-task Learning for Molecular Property Prediction
    Liu, Shengchao
    Qu, Meng
    Zhang, Zuobai
    Cai, Huiyu
    Tang, Jian
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [35] Water Quality Prediction Based on Multi-Task Learning
    Wu, Huan
    Cheng, Shuiping
    Xin, Kunlun
    Ma, Nian
    Chen, Jie
    Tao, Liang
    Gao, Min
    INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH, 2022, 19 (15)
  • [36] Enhancement of acute toxicity prediction by multi-task learning
    Sosnin, Sergey
    Karlov, Dmitry
    Tetko, Igor
    Fedorov, Maxim
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2019, 257
  • [37] Situation Aware Multi-Task Learning for Traffic Prediction
    Deng, Dingxiong
    Shahabi, Cyrus
    Demiryurek, Ugur
    Zhu, Linhong
    2017 17TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2017, : 81 - 90
  • [38] Deep Multi-task Learning for Air Quality Prediction
    Wang, Bin
    Yan, Zheng
    Lu, Jie
    Zhang, Guangquan
    Li, Tianrui
    NEURAL INFORMATION PROCESSING (ICONIP 2018), PT V, 2018, 11305 : 93 - 103
  • [39] Multi-task Learning for Mortality Prediction in LDCT Images
    Guo, Hengtao
    Kruger, Melanie
    Wang, Ge
    Kalra, Mannudeep K.
    Yan, Pingkun
    MEDICAL IMAGING 2020: COMPUTER-AIDED DIAGNOSIS, 2020, 11314
  • [40] Tomato leaf disease recognition based on multi-task distillation learning
    Liu, Bo
    Wei, Shusen
    Zhang, Fan
    Guo, Nawei
    Fan, Hongyu
    Yao, Wei
    FRONTIERS IN PLANT SCIENCE, 2024, 14