Attention-Aware Multi-Task Convolutional Neural Networks

被引:12
作者
Lyu, Kejie [1 ]
Li, Yingming [1 ]
Zhang, Zhongfei [1 ]
机构
[1] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou 310027, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Deep learning; Feature extraction; Training; Estimation; Semantics; Convolutional neural networks; Attention mechanism; convolutional neural network; multi-task learning; representation sharing;
D O I
10.1109/TIP.2019.2944522
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-task deep learning methods learn multiple tasks simultaneously and share representations amongst them, so information from related tasks improves learning within one task. The generalization capabilities of the produced models are substantially enhanced. Typical multi-task deep learning models usually share representations of different tasks in lower layers of the network, and separate representations of different tasks in higher layers. However, different groups of tasks always have different requirements for sharing representations, so the required design criterion does not necessarily guarantee that the obtained network architecture is optimal. In addition, most existing methods ignore the redundancy problem and lack the pre-screening process for representations before they are shared. Here, we propose a model called Attention-aware Multi-task Convolutional Neural Network, which automatically learns appropriate sharing through end-to-end training. The attention mechanism is introduced into our architecture to suppress redundant contents contained in the representations. The shortcut connection is adopted to preserve useful information. We evaluate our model by carrying out experiments on different task groups and different datasets. Our model demonstrates an improvement over existing techniques in many experiments, indicating the effectiveness and the robustness of the model. We also demonstrate the importance of attention mechanism and shortcut connection in our model.
引用
收藏
页码:1867 / 1878
页数:12
相关论文
共 41 条
[1]  
[Anonymous], 2015, ARXIV PREPRINT ARXIV
[2]  
[Anonymous], 2013, 31 INT C MACH LEARN
[3]  
[Anonymous], 2016, ARXIV PREPRINT ARXIV
[4]  
[Anonymous], 2015, IEEE I CONF COMP VIS, DOI DOI 10.1109/ICCV.2015.123
[5]  
[Anonymous], 2014, P 2 INT C LEARNING R
[6]  
[Anonymous], 2017, P IEEE INT C COMPUTE
[7]  
[Anonymous], 2017, ARXIV170605098
[8]   Convolutional Random Walk Networks for Semantic Image Segmentation [J].
Bertasius, Gedas ;
Torresani, Lorenzo ;
Yu, Stella X. ;
Shi, Jianbo .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6137-6145
[9]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[10]  
Chen L, 2015, 2ND INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY FOR EDUCATION (ICTE 2015), P1