Modeling Text-Label Alignment for Hierarchical Text Classification

被引:1
作者
Kumar, Ashish [1 ]
Toshniwal, Durga [1 ]
机构
[1] Indian Inst Technol Roorkee, Roorkee, Uttar Pradesh, India
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES-RESEARCH TRACK, PT VI, ECML PKDD 2024 | 2024年 / 14946卷
关键词
Multi-Label Classification; NLP; Representation Learning;
D O I
10.1007/978-3-031-70365-2_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical Text Classification (HTC) aims to categorize text data based on a structured label hierarchy, resulting in predicted labels forming a sub-hierarchy tree. The semantics of the text should align with the semantics of the labels in this sub-hierarchy. With the sub-hierarchy changing for each sample, the dynamic nature of text-label alignment poses challenges for existing methods, which typically process text and labels independently. To overcome this limitation, we propose a Text-Label Alignment (TLA) loss specifically designed to model the alignment between text and labels. We obtain a set of negative labels for a given text and its positive label set. By leveraging contrastive learning, the TLA loss pulls the text closer to its positive label and pushes it away from its negative label in the embedding space. This process aligns text representations with related labels while distancing them from unrelated ones. Building upon this framework, we introduce the Hierarchical Text-Label Alignment (HTLA) model, which leverages BERT as the text encoder and GPTrans as the graph encoder and integrates text-label embeddings to generate hierarchy-aware representations. Experimental results on benchmark datasets and comparison with existing baselines demonstrate the effectiveness of HTLA for HTC.
引用
收藏
页码:163 / 179
页数:17
相关论文
共 27 条
[1]  
Aly R, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019:): STUDENT RESEARCH WORKSHOP, P323
[2]  
Chen BL, 2020, AAAI CONF ARTIF INTE, V34, P7496
[3]  
Chen HB, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4370
[4]  
Chen T, 2020, PR MACH LEARN RES, V119
[5]  
Chen Z, 2023, PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, P3559
[6]   On the cost-effectiveness of neural and non-neural approaches and representations for text classification: A comprehensive comparative study [J].
Cunha, Washington ;
Mangaravite, Vitor ;
Gomes, Christian ;
Canuto, Sergio ;
Resende, Elaine ;
Nascimento, Cecilia ;
Viegas, Felipe ;
Franca, Celso ;
Martins, Wellington Santos ;
Almeida, Jussara M. ;
Rosa, Thierson ;
Rocha, Leonardo ;
Goncalves, Marcos Andre .
INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (03)
[7]  
Deng ZF, 2021, 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), P3259
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]  
Dumais S., 2000, P 23 ANN INT ACM SIG, P256
[10]  
Gopal S, 2013, 19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), P257