Trainable Weights for Multitask Learning

被引:0
|
作者
Ryu, Chaeeun [1 ]
Lee, Changwoo [2 ,3 ]
Choi, Hyuk Jin [4 ]
Lee, Chang-Hyun [5 ]
Jeon, Byoungjun [6 ]
Chie, Eui Kyu [7 ,8 ]
Kim, Young-Gon [2 ,9 ]
机构
[1] Sungkyunkwan Univ, Dept Comp Educ, Seoul 03063, South Korea
[2] Seoul Natl Univ Hosp, Dept Transdisciplinary Med, Seoul 03080, South Korea
[3] Seoul Natl Univ, Dept Med Device Dev, Coll Med, Seoul 03080, South Korea
[4] Seoul Natl Univ Hosp Healthcare Syst Gangnam Ctr, Dept Ophthalmol, Seoul 06236, South Korea
[5] Seoul Natl Univ Hosp, Dept Neurosurg, Seoul 03080, South Korea
[6] Seoul Natl Univ, Dept Neurosurg, Coll Med, Seoul 03080, South Korea
[7] Seoul Natl Univ Hosp, Dept Radiat Oncol, Seoul 03080, South Korea
[8] Seoul Natl Univ, Dept Radiat Oncol, Coll Med, Seoul 03080, South Korea
[9] Seoul Natl Univ, Dept Med, Coll Med, Seoul 03080, South Korea
关键词
Auxiliary task learning; incremental learning multitask learning; trainable parameters;
D O I
10.1109/ACCESS.2023.3319072
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The research on multi-task learning has been steadily increasing due to its advantages, such as preventing overfitting, averting catastrophic forgetting, solving multiple inseparable tasks, and coping with data shortage. Here, we question whether to incorporate different orderings of feature levels based on distinct characteristics of tasks and their interrelationships in multitask learning. While in many classification tasks, leveraging the features extracted from the last layer is common, we thought that given different characteristics of tasks there might be a need to encompass different representation levels, i.e., different orderings of feature levels. Hence, we utilized the knowledge of different representation levels by features extracted from the various blocks of the main module and applied trainable parameters as weights on the features. This indicates that we optimized the solution to the question by learning to weigh the features in a task-specific manner and solving tasks with a combination of newly weighted features. Our method SimPara presents a modular topology of multitask learning that is efficient in terms of memory and computation, effective, and easily applicable to diverse tasks or models. To show that our approach is task-agnostic and highly applicable, we demonstrate its effectiveness in auxiliary task learning, active learning, and multilabel learning settings. This work underscores that by simply learning weights to better order the features learned by a single backbone, we can incur better task-specific performance of the model.
引用
收藏
页码:105633 / 105641
页数:9
相关论文
共 50 条
  • [21] Exploring Multitask Learning for Steganalysis
    Makelberge, Julie
    Ker, Andrew D.
    MEDIA WATERMARKING, SECURITY, AND FORENSICS 2013, 2013, 8665
  • [22] Multitask Soft Option Learning
    Igl, Maximilian
    Gambardella, Andrew
    He, Jinke
    Nardelli, Nantas
    Siddharth, N.
    Bohmer, Wendelin
    Whiteson, Shimon
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 969 - 978
  • [23] A dozen tricks with multitask learning
    Caruana, R
    NEURAL NETWORKS: TRICKS OF THE TRADE, 1998, 1524 : 165 - 191
  • [24] Multitask Learning With Enhanced Modules
    Zheng, Zishuo
    Wei, Yadong
    Zhao, Zixu
    Wu, Xindi
    Li, Zhengcheng
    Ren, Pengju
    2018 IEEE 23RD INTERNATIONAL CONFERENCE ON DIGITAL SIGNAL PROCESSING (DSP), 2018,
  • [25] Is Multitask Learning Always Better?
    Mattick, Alexander
    Mayr, Martin
    Maier, Andreas
    Christlein, Vincent
    DOCUMENT ANALYSIS SYSTEMS, DAS 2022, 2022, 13237 : 674 - 687
  • [26] The benefit of multitask representation learning
    Maurer, Andreas
    Pontil, Massimiliano
    Romera-Paredes, Bernardino
    Journal of Machine Learning Research, 2016, 17
  • [27] Learning Task Sampling Policy for Multitask Learning
    Sundararaman, Dhanasekar
    Tsai, Henry
    Lee, Kuang-Huei
    Turc, Iulia
    Carin, Lawrence
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 4410 - 4415
  • [28] A Trainable Spaced Repetition Model for Language Learning
    Settles, Burr
    Meeder, Brendan
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1848 - 1858
  • [29] Trainable Undersampling for Class-Imbalance Learning
    Peng, Minlong
    Zhang, Qi
    Xing, Xiaoyu
    Gui, Tao
    Huang, Xuanjing
    Jiang, Yu-Gang
    Ding, Keyu
    Chen, Zhigang
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 4707 - 4714
  • [30] Trainable Compound Activation Functions for Machine Learning
    Baggenstoss, Paul M.
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 1382 - 1386