Multi-Task Learning with Knowledge Distillation for Dense Prediction

被引:2
|
作者
Xu, Yangyang [1 ,2 ]
Yang, Yibo [4 ]
Zhang, Lefei [1 ,2 ,3 ]
机构
[1] Wuhan Univ, Inst Artificial Intelligence, Wuhan, Peoples R China
[2] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
[3] Hubei Luojia Lab, Wuhan, Peoples R China
[4] King Abdullah Univ Sci & Technol, Jeddah, Saudi Arabia
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01970
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While multi-task learning (MTL) has become an attractive topic, its training usually poses more difficulties than the single-task case. How to successfully apply knowledge distillation into MTL to improve training efficiency and model performance is still a challenging problem. In this paper, we introduce a new knowledge distillation procedure with an alternative match for MTL of dense prediction based on two simple design principles. First, for memory and training efficiency, we use a single strong multitask model as a teacher during training instead of multiple teachers, as widely adopted in existing studies. Second, we employ a less sensitive Cauchy-Schwarz (CS) divergence instead of the Kullback-Leibler (KL) divergence and propose a CS distillation loss accordingly. With the less sensitive divergence, our knowledge distillation with an alternative match is applied for capturing inter-task and intratask information between the teacher model and the student model of each task, thereby learning more "dark knowledge" for effective distillation. We conducted extensive experiments on dense prediction datasets, including NYUD-v2 and PASCAL-Context, for multiple vision tasks, such as semantic segmentation, human parts segmentation, depth estimation, surface normal estimation, and boundary detection. The results show that our proposed method decidedly improves model performance and the practical inference efficiency.
引用
收藏
页码:21493 / 21502
页数:10
相关论文
共 50 条
  • [41] Multi-Task Learning Using Task Dependencies for Face Attributes Prediction
    Fan, Di
    Kim, Hyunwoo
    Kim, Junmo
    Liu, Yunhui
    Huang, Qiang
    APPLIED SCIENCES-BASEL, 2019, 9 (12):
  • [42] Weather-degraded image semantic segmentation with multi-task knowledge distillation
    Li, Zhi
    Wu, Xing
    Wang, Jianjia
    Guo, Yike
    IMAGE AND VISION COMPUTING, 2022, 127
  • [43] Multi-Task Knowledge Distillation with Embedding Constraints for Scholarly Keyphrase Boundary Classification
    Park, Seo Yeon
    Caragea, Cornelia
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 13026 - 13042
  • [44] Structured Knowledge Distillation for Dense Prediction
    Liu, Yifan
    Shu, Changyong
    Wang, Jingdong
    Shen, Chunhua
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) : 7035 - 7049
  • [45] Multi-task safe reinforcement learning for navigating intersections in dense traffic
    Liu, Yuqi
    Gao, Yinfeng
    Zhang, Qichao
    Ding, Dawei
    Zhao, Dongbin
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2023, 360 (17): : 13737 - 13760
  • [46] Multi-source-Load Prediction Based on Multi-task Learning
    Yan, Zhaokang
    Cheng, Sida
    Shen, Jingwen
    Jiang, Hanyuan
    Ma, Gang
    Zou, Wenjin
    PROCEEDINGS OF 2023 INTERNATIONAL CONFERENCE ON WIRELESS POWER TRANSFER, VOL 4, ICWPT 2023, 2024, 1161 : 266 - 273
  • [47] MIND: Multi-Task Incremental Network Distillation
    Bonato, Jacopo
    Pelosin, Francesco
    Sabetta, Luigi
    Nicolosi, Alessandro
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11105 - 11113
  • [48] A Multi-Task Learning Approach for Recommendation based on Knowledge Graph
    Yan, Cairong
    Liu, Shuai
    Zhang, Yanting
    Wang, Zijian
    Wang, Pengwei
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [49] Multi-Task Feature Learning for Knowledge Graph Enhanced Recommendation
    Wang, Hongwei
    Zhang, Fuzheng
    Zhao, Miao
    Li, Wenjie
    Xie, Xing
    Guo, Minyi
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 2000 - 2010
  • [50] Knowledge-Enhanced Multi-task Learning for Course Recommendation
    Ban, Qimin
    Wu, Wen
    Hu, Wenxin
    Lin, Hui
    Zheng, Wei
    He, Liang
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT II, 2022, : 85 - 101