TGGS network: A multi-task learning network for gradient-guided knowledge sharing

被引:0
作者
Huang, Yongjie [1 ]
Han, Xiao [1 ]
Chen, Man [1 ]
Pan, Zhisong [1 ]
机构
[1] Army Engn Univ PLA, Command Control Engn Coll, Nanjing 210007, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-task learning; Gradient; Knowledge sharing; Image classification; Time series prediction;
D O I
10.1016/j.knosys.2024.112254
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-task learning (MTL) has been widely used in various fields, such as time series data prediction and image classification. Most existing deep MTL methods achieve joint learning by sharing common knowledge with task relevance. However, they rely solely on self-exploration to mine common knowledge across the tasks, which is unguided and purposeless. Unlike these methods, in this paper, we propose a novel Task Gradient Guided Sharing (TGGS) MTL network for purposefully guiding the model to share knowledge. Specifically, we define a confidence matrix for each task for selecting effective knowledge from the common knowledge to share. To realize an effective guidance, we align the confidence matrix with a class gradient mapping (CGM) computed based on the task-specific gradients. Consequently, our method effectively employs task-specific gradients to guide the model in sharing useful knowledge. More importantly, the confidence matrix explains the complex correlations among tasks in MTL. As our method exclusively focuses on guided knowledge sharing, the TGGS network can seamlessly integrate with diverse MTL sharing strategies. We conduct extensive experiments on time series datasets and image datasets. The experimental results demonstrate our method significantly improves the performance of these MTL strategies.
引用
收藏
页数:13
相关论文
共 40 条
[1]  
Adebayo J, 2018, ADV NEUR IN, V31
[2]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[3]   A Model-Agnostic Approach to Mitigate Gradient Interference for Multi-Task Learning [J].
Chai, Heyan ;
Yin, Zhe ;
Ding, Ye ;
Liu, Li ;
Fang, Binxing ;
Liao, Qing .
IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (12) :7810-7823
[4]  
de Campos TE, 2009, VISAPP 2009: PROCEEDINGS OF THE FOURTH INTERNATIONAL CONFERENCE ON COMPUTER VISION THEORY AND APPLICATIONS, VOL 2, P273
[5]  
DeVries T, 2018, Arxiv, DOI arXiv:1802.04865
[6]   MSSM: A Multiple-level Sparse Sharing Model for Efficient Multi-Task Learning [J].
Ding, Ke ;
Dong, Xin ;
He, Yong ;
Cheng, Lei ;
Fu, Chilin ;
Huan, Zhaoxin ;
Li, Hai ;
Yan, Tan ;
Zhang, Liang ;
Zhang, Xiaolu ;
Mo, Linjian .
SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, :2237-2241
[7]   Learning twofold heterogeneous multi-task by sharing similar convolution kernel pairs [J].
Feng, Quan ;
Yao, JiaYu ;
Zhong, Yingyu ;
Li, Ping ;
Pan, Zhisong .
KNOWLEDGE-BASED SYSTEMS, 2022, 252
[8]   Predicting LQ45 financial sector indices using RNN-LSTM [J].
Hansun, Seng ;
Young, Julio Christian .
JOURNAL OF BIG DATA, 2021, 8 (01)
[9]  
Hashimoto K., P 2017 C EMP METH NA, V2017, P1923
[10]   MultiTask-CenterNet (MCN): Efficient and Diverse Multitask Learning using an Anchor Free Approach [J].
Heuer, Falk ;
Mantowsky, Sven ;
Bukhari, Syed Saqib ;
Schneider, Georg .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, :997-1005