Learning twofold heterogeneous multi-task by sharing similar convolution kernel pairs

被引:7
作者
Feng, Quan [1 ,2 ]
Yao, JiaYu [3 ]
Zhong, Yingyu [4 ]
Li, Ping [5 ,6 ]
Pan, Zhisong [7 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Jiangsu, Peoples R China
[2] Nanjing Univ Aeronaut & Astronaut, MIIT Key Lab Pattern Anal & Machine Intelligence, Nanjing 211106, Jiangsu, Peoples R China
[3] 4th Paradigm Data & Technol Co Ltd, Beijing 100085, Peoples R China
[4] SANY CONSTRUCT TECHNOL CO LTD, Changsha 210023, Hunan, Peoples R China
[5] Nanjing Univ Posts & Telecommun, Coll Comp Sci, Nanjing 210023, Jiangsu, Peoples R China
[6] Jiangsu Key Lab Big Data Secur & Intelligent Proc, Nanjing 210023, Jiangsu, Peoples R China
[7] Army Engn Univ PLA, Coll Command & Control Engn, Nanjing 210023, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Heterogeneous tasks; Multi-task learning; Convolution kernel sharing;
D O I
10.1016/j.knosys.2022.109396
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Heterogeneous multi-task learning (HMTL) is an important topic in multi-task learning (MTL). Most existing HMTL methods usually solve either scenario where all tasks reside in the same input (feature) space yet unnecessarily the consistent output (label) space or scenario where their input (feature) spaces are heterogeneous while the output (label) space is consistent. However, to the best of our knowledge, there is limited study on twofold heterogeneous MTL (THMTL) scenario where the input and the output spaces are both inconsistent or heterogeneous. In order to handle this complicated scenario, in this paper, we design a simple and effective multi-task adaptive learning (MTAL) network to learn multiple tasks in such THMTL setting. Specifically, we explore and utilize the inherent relationship between tasks for knowledge sharing from similar convolution kernels in individual layers of the MTAL network. To realize the sharing, we weightedly aggregate any pair of convolutional kernels with their similarity greater than some threshold rho. Consequently, our model effectively performs cross-task learning while suppresses the intra-redundancy of the entire network. Finally, we conduct end-to-end training. Our experimental results demonstrate the significant competitiveness of our method in comparison with the current state-of-the-art methods. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:13
相关论文
共 56 条
[1]   A Bayesian information theoretic model of learning to learn via multiple task sampling [J].
Baxter, J .
MACHINE LEARNING, 1997, 28 (01) :7-39
[2]   Heterogeneous Image Features Integration via Multi-Modal Semi-Supervised Learning Model [J].
Cai, Xiao ;
Nie, Feiping ;
Cai, Weidong ;
Huang, Heng .
2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, :1737-1744
[3]   A multi-task learning approach for improving travel recommendation with keywords generation [J].
Chen, Lei ;
Cao, Jie ;
Zhu, Guixiang ;
Wang, Youquan ;
Liang, Weichao .
KNOWLEDGE-BASED SYSTEMS, 2021, 233 (233)
[4]  
Chen Z, 2018, PR MACH LEARN RES, V80
[5]  
Deng F., 2011, 2021 13 INT C MACHIN, P450
[6]  
Denil M., 2013, Adv. Neural Inf. Process. Syst.
[7]   MSSM: A Multiple-level Sparse Sharing Model for Efficient Multi-Task Learning [J].
Ding, Ke ;
Dong, Xin ;
He, Yong ;
Cheng, Lei ;
Fu, Chilin ;
Huan, Zhaoxin ;
Li, Hai ;
Yan, Tan ;
Zhang, Liang ;
Zhang, Xiaolu ;
Mo, Linjian .
SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, :2237-2241
[8]  
Dubey A., 2022, MACHINE LEARNING APP
[9]  
Duong L, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, P845
[10]   Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture [J].
Eigen, David ;
Fergus, Rob .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2650-2658