Task-Aware Dynamic Model Optimization for Multi-Task Learning

被引:0
作者
Choi, Sujin [1 ]
Jin, Hyundong [2 ]
Kim, Eunwoo [1 ,2 ]
机构
[1] Chung Ang Univ, Dept Artificial Intelligence, Seoul 06974, South Korea
[2] Chung Ang Univ, Sch Comp Sci & Engn, Seoul 06974, South Korea
关键词
Multi-task learning; resource-efficient learning; model optimization;
D O I
10.1109/ACCESS.2023.3339793
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-task learning (MTL) is a field in which a deep neural network simultaneously learns knowledge from multiple tasks. However, achieving resource-efficient MTL remains challenging due to entangled network parameters across tasks and varying task-specific complexity. Existing methods employ network compression techniques while maintaining comparable performance, but they often compress uniformly across all tasks without considering individual complexity. This can lead to suboptimal solutions due to entangled network parameters and memory inefficiency, as the parameters for each task may be insufficient or excessive. To address these challenges, we propose a framework called Dynamic Model Optimization (DMO) that dynamically allocates network parameters to groups based on task-specific complexity. This framework consists of three key steps: measuring task similarity and task difficulty, grouping tasks, and allocating parameters. This process involves the calculation of both weight and loss similarities across tasks and employs sample-wise loss as a measure of task difficulty. Tasks are grouped based on their similarities, and parameters are allocated with dynamic pruning according to task difficulty within their respective groups. We apply the proposed framework to MTL with various classification datasets. Experimental results demonstrate that the proposed approach achieves high performance while taking fewer network parameters than other MTL methods.
引用
收藏
页码:137709 / 137717
页数:9
相关论文
共 37 条
[11]   NDDR-CNN: Layerwise Feature Fusing in Multi-Task CNNs by Neural Discriminative Dimensionality Reduction [J].
Gao, Yuan ;
Ma, Jiayi ;
Zhao, Mingbo ;
Liu, Wei ;
Yuille, Alan L. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3200-3209
[12]   Dynamic Task Prioritization for Multitask Learning [J].
Guo, Michelle ;
Haque, Albert ;
Huang, De-An ;
Yeung, Serena ;
Li Fei-Fei .
COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 :282-299
[13]  
Han S, 2015, ADV NEUR IN, V28
[14]   A DATABASE FOR HANDWRITTEN TEXT RECOGNITION RESEARCH [J].
HULL, JJ .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1994, 16 (05) :550-554
[15]   CLIQUES OF A GRAPH - VARIATIONS ON BRON-KERBOSCH ALGORITHM [J].
JOHNSTON, HC .
INTERNATIONAL JOURNAL OF COMPUTER & INFORMATION SCIENCES, 1976, 5 (03) :209-238
[16]   Multi-Task Learning Using Uncertainty to Weigh Losses for Scene Geometry and Semantics [J].
Kendall, Alex ;
Gal, Yarin ;
Cipolla, Roberto .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7482-7491
[17]   NestedNet: Learning Nested Sparse Structures in Deep Neural Networks [J].
Kim, Eunwoo ;
Ahn, Chanho ;
Oh, Songhwai .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8669-8678
[18]   UberNet: Training a Universal Convolutional Neural Network for Low-, Mid-, and High-Level Vision using Diverse Datasets and Limited Memory [J].
Kokkinos, Iasonas .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5454-5463
[19]  
Kornblith S, 2019, PR MACH LEARN RES, V97
[20]  
Krizhevsky A., 2009, LEARNING MULTIPLE LA