Trainable Weights for Multitask Learning

被引:0
|
作者
Ryu, Chaeeun [1 ]
Lee, Changwoo [2 ,3 ]
Choi, Hyuk Jin [4 ]
Lee, Chang-Hyun [5 ]
Jeon, Byoungjun [6 ]
Chie, Eui Kyu [7 ,8 ]
Kim, Young-Gon [2 ,9 ]
机构
[1] Sungkyunkwan Univ, Dept Comp Educ, Seoul 03063, South Korea
[2] Seoul Natl Univ Hosp, Dept Transdisciplinary Med, Seoul 03080, South Korea
[3] Seoul Natl Univ, Dept Med Device Dev, Coll Med, Seoul 03080, South Korea
[4] Seoul Natl Univ Hosp Healthcare Syst Gangnam Ctr, Dept Ophthalmol, Seoul 06236, South Korea
[5] Seoul Natl Univ Hosp, Dept Neurosurg, Seoul 03080, South Korea
[6] Seoul Natl Univ, Dept Neurosurg, Coll Med, Seoul 03080, South Korea
[7] Seoul Natl Univ Hosp, Dept Radiat Oncol, Seoul 03080, South Korea
[8] Seoul Natl Univ, Dept Radiat Oncol, Coll Med, Seoul 03080, South Korea
[9] Seoul Natl Univ, Dept Med, Coll Med, Seoul 03080, South Korea
关键词
Auxiliary task learning; incremental learning multitask learning; trainable parameters;
D O I
10.1109/ACCESS.2023.3319072
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The research on multi-task learning has been steadily increasing due to its advantages, such as preventing overfitting, averting catastrophic forgetting, solving multiple inseparable tasks, and coping with data shortage. Here, we question whether to incorporate different orderings of feature levels based on distinct characteristics of tasks and their interrelationships in multitask learning. While in many classification tasks, leveraging the features extracted from the last layer is common, we thought that given different characteristics of tasks there might be a need to encompass different representation levels, i.e., different orderings of feature levels. Hence, we utilized the knowledge of different representation levels by features extracted from the various blocks of the main module and applied trainable parameters as weights on the features. This indicates that we optimized the solution to the question by learning to weigh the features in a task-specific manner and solving tasks with a combination of newly weighted features. Our method SimPara presents a modular topology of multitask learning that is efficient in terms of memory and computation, effective, and easily applicable to diverse tasks or models. To show that our approach is task-agnostic and highly applicable, we demonstrate its effectiveness in auxiliary task learning, active learning, and multilabel learning settings. This work underscores that by simply learning weights to better order the features learned by a single backbone, we can incur better task-specific performance of the model.
引用
收藏
页码:105633 / 105641
页数:9
相关论文
共 50 条
  • [1] Beyond weights adaptation: A new neuron model with trainable activation function and its supervised learning
    Wu, YS
    Zhao, MS
    Ding, XQ
    1997 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, 1997, : 1152 - 1157
  • [2] Learning to Multitask
    Zhang, Yu
    Wei, Ying
    Yang, Qiang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] Multitask learning
    Caruana, R
    MACHINE LEARNING, 1997, 28 (01) : 41 - 75
  • [4] Multitask Learning
    Rich Caruana
    Machine Learning, 1997, 28 : 41 - 75
  • [5] Online multitask learning
    Dekel, Ofer
    Long, Philip M.
    Singer, Yoram
    LEARNING THEORY, PROCEEDINGS, 2006, 4005 : 453 - 467
  • [6] Multitask Coactive Learning
    Goetschalckx, Robby
    Fern, Alan
    Tadepalli, Prasad
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3518 - 3524
  • [7] Semisupervised Multitask Learning
    Liu, Qiuhua
    Liao, Xuejun
    Li, Hui
    Stack, Jason R.
    Carin, Lawrence
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (06) : 1074 - 1086
  • [8] TRAINABLE, BAYESIAN SYMMETRIES FOR REINFORCEMENT LEARNING
    Lu, Qingmei
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON ADVANCED COMPUTER THEORY AND ENGINEERING (ICACTE 2009), VOLS 1 AND 2, 2009, : 1079 - 1086
  • [9] On Multiplicative Multitask Feature Learning
    Wang, Xin
    Bi, Jinbo
    Yu, Shipeng
    Sun, Jiangwen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [10] HIERARCHICAL MULTITASK LEARNING WITH CTC
    Sanabria, Ramon
    Metze, Florian
    2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 485 - 490