Domain-Adversarial Graph Neural Networks for Text Classification

被引:24
作者
Wu, Man [1 ]
Pan, Shirui [2 ]
Zhu, Xingquan [1 ]
Zhou, Chuan [3 ,4 ]
Pan, Lei [5 ]
机构
[1] Florida Atlantic Univ, Dept Comp & Elect Engn & Comp Sci, Boca Raton, FL 33431 USA
[2] Monash Univ, Fac Informat Technol, Melbourne, Vic, Australia
[3] Chinese Acad Sci, Acad Math & Syst Sci, Beijing, Peoples R China
[4] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[5] Deakin Univ, Sch Informat Technol, Geelong, Vic 3220, Australia
来源
2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019) | 2019年
基金
美国国家科学基金会;
关键词
Graph neural networks; cross-domain learning; text classification;
D O I
10.1109/ICDM.2019.00075
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text classification, in cross-domain setting, is a challenging task. On the one hand, data from other domains are often useful to improve the learning on the target domain; on the other hand, domain variance and hierarchical structure of documents from words, key phrases, sentences, paragraphs, etc. make it difficult to align domains for effective learning. To date, existing cross-domain text classification methods mainly strive to minimize feature distribution differences between domains, and they typically suffer from three major limitations - (1) difficult to capture semantics in non-consecutive phrases and long-distance word dependency because of treating texts as word sequences, (2) neglect of hierarchical coarse-grained structures of document for feature learning, and (3) narrow focus of the domains at instance levels, without using domains as supervisions to improve text classification. This paper proposes an end-to-end, domain-adversarial graph neural networks (DAGNN), for cross-domain text classification. Our motivation is to model documents as graphs and use a domain-adversarial training principle to lean features from each graph (as well as learning the separation of domains) for effective text classification. At the instance level, DAGNN uses a graph to model each document, so that it can capture non-consecutive and long-distance semantics. At the feature level, DAGNN uses graphs from different domains to jointly train hierarchical graph neural networks in order to learn good features. At the learning level, DAGNN proposes a domain-adversarial principle such that the learned features not only optimally classify documents but also separates domains. Experiments on benchmark datasets demonstrate the effectiveness of our method in cross-domain classification tasks.
引用
收藏
页码:648 / 657
页数:10
相关论文
共 36 条
  • [31] Wu Z., 2019, 190100596 ARXIV
  • [32] Xue Gui-Rong., 2008, Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR '08, P627, DOI DOI 10.1145/1390334.1390441
  • [33] Yan XF, 2002, 2002 IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, P721, DOI 10.1109/ICDM.2002.1184038
  • [34] Yao L., 2018, Graph convolutional networks for text classification
  • [35] Ying Zhitao., 2018, ADV NEURAL INFORM PR, P4800
  • [36] Zhuang F., 2010, P 19 INT C INF KNOWL, P359