Domain-Adversarial Graph Neural Networks for Text Classification

被引:24
作者
Wu, Man [1 ]
Pan, Shirui [2 ]
Zhu, Xingquan [1 ]
Zhou, Chuan [3 ,4 ]
Pan, Lei [5 ]
机构
[1] Florida Atlantic Univ, Dept Comp & Elect Engn & Comp Sci, Boca Raton, FL 33431 USA
[2] Monash Univ, Fac Informat Technol, Melbourne, Vic, Australia
[3] Chinese Acad Sci, Acad Math & Syst Sci, Beijing, Peoples R China
[4] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[5] Deakin Univ, Sch Informat Technol, Geelong, Vic 3220, Australia
来源
2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019) | 2019年
基金
美国国家科学基金会;
关键词
Graph neural networks; cross-domain learning; text classification;
D O I
10.1109/ICDM.2019.00075
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text classification, in cross-domain setting, is a challenging task. On the one hand, data from other domains are often useful to improve the learning on the target domain; on the other hand, domain variance and hierarchical structure of documents from words, key phrases, sentences, paragraphs, etc. make it difficult to align domains for effective learning. To date, existing cross-domain text classification methods mainly strive to minimize feature distribution differences between domains, and they typically suffer from three major limitations - (1) difficult to capture semantics in non-consecutive phrases and long-distance word dependency because of treating texts as word sequences, (2) neglect of hierarchical coarse-grained structures of document for feature learning, and (3) narrow focus of the domains at instance levels, without using domains as supervisions to improve text classification. This paper proposes an end-to-end, domain-adversarial graph neural networks (DAGNN), for cross-domain text classification. Our motivation is to model documents as graphs and use a domain-adversarial training principle to lean features from each graph (as well as learning the separation of domains) for effective text classification. At the instance level, DAGNN uses a graph to model each document, so that it can capture non-consecutive and long-distance semantics. At the feature level, DAGNN uses graphs from different domains to jointly train hierarchical graph neural networks in order to learn good features. At the learning level, DAGNN proposes a domain-adversarial principle such that the learned features not only optimally classify documents but also separates domains. Experiments on benchmark datasets demonstrate the effectiveness of our method in cross-domain classification tasks.
引用
收藏
页码:648 / 657
页数:10
相关论文
共 36 条
  • [1] [Anonymous], 2007, SIGKDD
  • [2] [Anonymous], 2016, 30 AAAI C ART INT
  • [3] [Anonymous], 2012, P ICML
  • [4] [Anonymous], 2014, UNSUPERVISED DOMAIN
  • [5] [Anonymous], 2015, INT C MACHINE LEARNI
  • [6] [Anonymous], 2018, ARXIV180408049
  • [7] Bastings Joost, 2017, ARXIV170404675
  • [8] A neural probabilistic language model
    Bengio, Y
    Ducharme, R
    Vincent, P
    Jauvin, C
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) : 1137 - 1155
  • [9] Chen M., 2011, P ADV NEUR INF PROC, V24, P2456
  • [10] Domain Transfer Multiple Kernel Learning
    Duan, Lixin
    Tsang, Ivor W.
    Xu, Dong
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (03) : 465 - 479