Task-Aware Dynamic Model Optimization for Multi-Task Learning

被引:0
作者
Choi, Sujin [1 ]
Jin, Hyundong [2 ]
Kim, Eunwoo [1 ,2 ]
机构
[1] Chung Ang Univ, Dept Artificial Intelligence, Seoul 06974, South Korea
[2] Chung Ang Univ, Sch Comp Sci & Engn, Seoul 06974, South Korea
关键词
Multi-task learning; resource-efficient learning; model optimization;
D O I
10.1109/ACCESS.2023.3339793
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-task learning (MTL) is a field in which a deep neural network simultaneously learns knowledge from multiple tasks. However, achieving resource-efficient MTL remains challenging due to entangled network parameters across tasks and varying task-specific complexity. Existing methods employ network compression techniques while maintaining comparable performance, but they often compress uniformly across all tasks without considering individual complexity. This can lead to suboptimal solutions due to entangled network parameters and memory inefficiency, as the parameters for each task may be insufficient or excessive. To address these challenges, we propose a framework called Dynamic Model Optimization (DMO) that dynamically allocates network parameters to groups based on task-specific complexity. This framework consists of three key steps: measuring task similarity and task difficulty, grouping tasks, and allocating parameters. This process involves the calculation of both weight and loss similarities across tasks and employs sample-wise loss as a measure of task difficulty. Tasks are grouped based on their similarities, and parameters are allocated with dynamic pruning according to task difficulty within their respective groups. We apply the proposed framework to MTL with various classification datasets. Experimental results demonstrate that the proposed approach achieves high performance while taking fewer network parameters than other MTL methods.
引用
收藏
页码:137709 / 137717
页数:9
相关论文
共 37 条
[1]   Deep Elastic Networks with Model Selection for Multi-Task Learning [J].
Ahn, Chanho ;
Kim, Eunwoo ;
Oh, Songhwai .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6528-6537
[2]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[3]  
Chen Z, 2018, PR MACH LEARN RES, V80
[4]  
Cheng Y, 2020, Arxiv, DOI arXiv:1710.09282
[5]   Describing Textures in the Wild [J].
Cimpoi, Mircea ;
Maji, Subhransu ;
Kokkinos, Iasonas ;
Mohamed, Sammy ;
Vedaldi, Andrea .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :3606-3613
[6]  
Deng L., 2012, IEEE Signal Processing Magazine, V29, P141, DOI DOI 10.1109/MSP.2012.2211477
[7]  
Duong L, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, P845
[8]   Representation Similarity Analysis for Efficient Task Taxonomy & Transfer Learning [J].
Dwivedi, Kshitij ;
Roig, Gemma .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :12379-12388
[9]  
Fifty Christopher, 2021, ADV NEUR IN, V34
[10]  
Frankle Jonathan, 2018, arXiv