Hierarchical Multi-label Text Classification: An Attention-based Recurrent Network Approach

被引:123
作者
Huang, Wei [1 ]
Chen, Enhong [1 ]
Liu, Qi [1 ]
Chen, Yuying [1 ,2 ]
Huang, Zai [1 ]
Liu, Yang [1 ]
Zhao, Zhou [3 ]
Zhang, Dan [4 ]
Wang, Shijin [4 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei, Peoples R China
[2] Ant Financial Serv Grp, Hangzhou, Peoples R China
[3] Zhejiang Univ, Sch Comp Sci & Technol, Hangzhou, Peoples R China
[4] iFLYTEK Res, Hefei, Peoples R China
来源
PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19) | 2019年
基金
中国国家自然科学基金;
关键词
Hierarchical Multi-label Text Classification; Attention Mechanism; Hierarchical Attention Networks;
D O I
10.1145/3357384.3357885
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Hierarchical multi-label text classification (HMTC) is a fundamental but challenging task of numerous applications (e.g., patent annotation), where documents are assigned to multiple categories stored in a hierarchical structure. Categories at different levels of a document tend to have dependencies. However, the majority of prior studies for the HMTC task employ classifiers to either deal with all categories simultaneously or decompose the original problem into a set of flat multi-label classification subproblems, ignoring the associations between texts and the hierarchical structure and the dependencies among different levels of the hierarchical structure. To that end, in this paper, we propose a novel framework called Hierarchical Attention-based Recurrent Neural Network (HARNN) for classifying documents into the most relevant categories level by level via integrating texts and the hierarchical category structure. Specifically, we first apply a documentation representing layer for obtaining the representation of texts and the hierarchical structure. Then, we develop an hierarchical attention-based recurrent layer to model the dependencies among different levels of the hierarchical structure in a top-down fashion. Here, a hierarchical attention strategy is proposed to capture the associations between texts and the hierarchical structure. Finally, we design a hybrid method which is capable of predicting the categories of each level while classifying all categories in the entire hierarchical structure precisely. Extensive experimental results on two real-world datasets demonstrate the effectiveness and explanatory power of HARNN.
引用
收藏
页码:1051 / 1060
页数:10
相关论文
共 37 条
[1]  
Abadi M., 2015, TENSORFLOW LARGE SCA, DOI DOI 10.48550/ARXIV.1603.04467
[2]  
[Anonymous], 2012, IJCNN 2012, DOI [10.1109/IJCNN.2012.6252736, DOI 10.1109/IJCNN.2012.6252736]
[3]  
[Anonymous], 2013, PREPRINT ARXIV 1308
[4]  
[Anonymous], 2017, 5 INT C LEARNING REP
[5]  
[Anonymous], 2004, P 13 ACM INT C INF K
[6]  
[Anonymous], P 21 INT C MACH LEAR
[7]  
[Anonymous], 2006, P 23 INT C MACHINE L, DOI [10.1145/1143844.1143874, DOI 10.1145/1143844.1143874]
[8]  
[Anonymous], 2018, P 35 INT C MACH LEAR
[9]  
[Anonymous], 33 AAAI C ART INT
[10]   Refined Experts Improving Classification in Large Taxonomies [J].
Bennett, Paul N. ;
Nguyen, Nam .
PROCEEDINGS 32ND ANNUAL INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2009, :11-18