Deep Multi-task Augmented Feature Learning via Hierarchical Graph Neural Network

被引:5
作者
Guo, Pengxin [1 ]
Deng, Chang [3 ]
Xu, Linjie [4 ]
Huang, Xiaonan [1 ]
Zhang, Yu [1 ,2 ]
机构
[1] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen, Peoples R China
[2] Peng Cheng Lab, Shenzhen, Peoples R China
[3] Univ Chicago, Comm Computat & Appl Math, Chicago, IL 60637 USA
[4] Queen Mary Univ London, Game AI Grp, London, England
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES | 2021年 / 12975卷
关键词
Multi-task learning; Feature learning; Graph neural network;
D O I
10.1007/978-3-030-86486-6_33
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep multi-task learning attracts much attention in recent years as it achieves good performance in many applications. Feature learning is important to deep multi-task learning for sharing common information among tasks. In this paper, we propose a Hierarchical Graph Neural Network (HGNN) to learn augmented features for deep multi-task learning. The HGNN consists of two-level graph neural networks. In the low level, an intra-task graph neural network is responsible of learning a powerful representation for each data point in a task by aggregating its neighbors. Based on the learned representation, a task embedding can be generated for each task in a similar way to max pooling. In the second level, an inter-task graph neural network updates task embeddings of all the tasks based on the attention mechanism to model task relations. Then the task embedding of one task is used to augment the feature representation of data points in this task. Moreover, for classification tasks, an inter-class graph neural network is introduced to conduct similar operations on a finer granularity, i.e., the class level, to generate class embeddings for each class in all the tasks using class embeddings to augment the feature representation. The proposed feature augmentation strategy can be used in many deep multi-task learning models. Experiments on real-world datasets show the significant performance improvement when using this strategy.
引用
收藏
页码:538 / 553
页数:16
相关论文
共 41 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
Ando RK, 2005, J MACH LEARN RES, V6, P1817
[3]  
[Anonymous], 2009, P 26 ANN INT C MACHI, DOI [DOI 10.1145/1553374.1553458, 10.1145/1553374.1553458, 10.1145/1553374]
[4]  
[Anonymous], 2007, P ADV NEUR INF PROC
[5]  
[Anonymous], 2015, ACS SYM SER
[6]  
[Anonymous], 2010, Advances in neural information processing systems
[7]  
Argyriou A., 2006, Advances in Neural Information Processing Systems, P41
[8]  
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690
[9]  
Bickel S., 2008, P 25 INT C MACH LEAR, P56, DOI DOI 10.1145/1390156.1390164
[10]  
Caputo B., 2014, Lect. Notes Comput. Sci., P192