HLC: hierarchically-aware label correlation for hierarchical text classification

被引:2
作者
Kumar, Ashish [1 ]
Toshinwal, Durga [1 ]
机构
[1] Indian Inst Technol Roorkee, Dept Comp Sci & Engn, Roorkee, Uttrakahnd, India
关键词
Hierarchical text classification; Multi-label classification; NLP; Deep neural networks;
D O I
10.1007/s10489-023-05257-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical Text Classification (HTC) leverages the hierarchical structure of labels to enhance text categorization. Existing methods use a combination of text and structure encoders to generate a composite representation. However, these methods may face challenges when encoding the hierarchy and capturing the label correlations that convey information about the relationships and dependencies among the labels. To address these challenges, we introduce the Hierarchy-Aware Label Correlation (HLC) model in this paper. HLC adopts a customized Graphomer as its structure encoder to learn the hierarchy. Graphormer utilizes self-attention to capture global dependencies and explicit structure encoding mechanisms to model relationships among labels. Additionally, HLC is optimized on Cross-Entropy with Anchor Label (CEAL) loss function, specifically designed to learn the hierarchical label correlations. CEAL introduces an anchor label with a fixed score of zero, distinguishing target labels from non-target ones. This distinctive approach encourages HLC to predict higher scores for true target labels and lower scores for non-target labels compared to the anchor label. We conducted experiments on three benchmark datasets and compared them with existing methods. The results suggest that HLC can be an effective method for HTC.
引用
收藏
页码:1602 / 1618
页数:17
相关论文
共 40 条
[1]  
Agrawal Rahul, 2013, P 22 INT C WORLD WID, P13, DOI [10.1145/2488388.2488391, 10.1145/2488388]
[2]  
Aly R, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019:): STUDENT RESEARCH WORKSHOP, P323
[3]  
Banerjee S, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P6295
[4]  
Cesa-Bianchi Nicolo, 2006, Proceedings of the 23rd international conference on Machine learning, P177
[5]  
Chen BL, 2020, AAAI CONF ARTIF INTE, V34, P7496
[6]  
Chen HB, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4370
[7]   On the cost-effectiveness of neural and non-neural approaches and representations for text classification: A comprehensive comparative study [J].
Cunha, Washington ;
Mangaravite, Vitor ;
Gomes, Christian ;
Canuto, Sergio ;
Resende, Elaine ;
Nascimento, Cecilia ;
Viegas, Felipe ;
Franca, Celso ;
Martins, Wellington Santos ;
Almeida, Jussara M. ;
Rosa, Thierson ;
Rocha, Leonardo ;
Goncalves, Marcos Andre .
INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (03)
[8]  
Deng ZF, 2021, 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), P3259
[9]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[10]  
Dror R, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P1383