A Study on Hierarchical Text Classification as a Seq2seq Task

被引:1
作者
Torba, Fatos [1 ,2 ]
Gravier, Christophe [2 ]
Laclau, Charlotte [3 ]
Kammoun, Abderrhammen [1 ]
Subercaze, Julien [1 ]
机构
[1] AItenders, St Etienne, France
[2] CNRS, Lab Hubert Curien, UMR 5516, St Etienne, France
[3] Inst Polytech Paris, Telecom Paris, Paris, France
来源
ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT III | 2024年 / 14610卷
关键词
Hierarchical text classification; generative model; reproducibility;
D O I
10.1007/978-3-031-56063-7_20
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the progress of generative neural models, Hierarchical Text Classification (HTC) can be cast as a generative task. In this case, given an input text, the model generates the sequence of predicted class labels taken from a label tree of arbitrary width and depth. Treating HTC as a generative task introduces multiple modeling choices. These choices vary from choosing the order for visiting the class tree and therefore defining the order of generating tokens, choosing either to constrain the decoding to labels that respect the previous level predictions, up to choosing the pre-trained Language Model itself. Each HTC model therefore differs from the others from an architectural standpoint, but also from the modeling choices that were made. Prior contributions lack transparent modeling choices and open implementations, hindering the assessment of whether model performance stems from architectural or modeling decisions. For these reasons, we propose with this paper an analysis of the impact of different modeling choices along with common model errors and successes for this task. This analysis is based on an open framework coming along this paper that can facilitate the development of future contributions in the field by providing datasets, metrics, error analysis toolkit and the capability to readily test various modeling choices for one given model.
引用
收藏
页码:287 / 296
页数:10
相关论文
共 21 条
[1]  
Bhambhoria R, 2023, 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, P1782
[2]  
Cao PF, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P3105
[3]  
Chen HB, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4370
[4]  
Chen L., 2022, P 2022 C EMPIRICAL M, P390, DOI 10.18653/v1/2022.emnlp-industry.39
[5]  
Deng ZF, 2021, 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), P3259
[6]  
Giunchiglia E., 2020, NeurIPS
[7]  
Huang Wei, 2022, P 29 INT C COMP LING, P1116
[8]  
Jiang H., 2020, P 2 WORKSH FIN TECHN, P1
[9]  
Jiang Ting, 2022, P 2022 C EMPIRICAL M, P4030
[10]   HDLTex: Hierarchical Deep Learning for Text Classification [J].
Kowsari, Kamran ;
Brown, Donald E. ;
Heidarysafa, Mojtaba ;
Meimandi, Kiana Jafari ;
Gerber, Matthew S. ;
Barnes, Laura E. .
2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, :364-371