共 43 条
A Dual-branch Learning Model with Gradient-balanced Loss for Long-tailed Multi-label Text Classification
被引:2
|作者:
Yao, Yitong
[1
]
Zhang, Jing
[1
]
Zhang, Peng
[1
]
Sun, Yueheng
[1
]
机构:
[1] Tianjin Univ, Tianjin, Peoples R China
关键词:
Multi-label text classification;
long-tailed learning;
dual-branch structure;
re-weighting loss function;
D O I:
10.1145/3597416
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Multi-label text classification has awide range of applications in the realworld. However, the data distribution in the real world is often imbalanced, which leads to serious long-tailed problems. For multi-label classification, due to the vast scale of datasets and existence of label co-occurrence, how to effectively improve the prediction accuracy of tail labels without degrading the overall precision becomes an important challenge. To address this issue, we propose A Dual-Branch Learning Model with Gradient-Balanced Loss (DBGB) based on the paradigm of existing pre-trained multi-label classification SOTA models. Our model consists of two main long-tailed module improvements. First, with the shared text representation, the dual-classifier is leveraged to process two kinds of label distributions; one is the original data distribution and the other is the under-sampling distribution for head labels to strengthen the prediction for tail labels. Second, the proposed gradient-balanced loss can adaptively suppress the negative gradient accumulation problem related to labels, especially tail labels. We perform extensive experiments on three multi-label text classification datasets. The results show that the proposed method achieves competitive performance on overall prediction results compared to the state-of-the-art methods in solving the multi-label classification, with significant improvement on tail-label accuracy.
引用
收藏
页数:24
相关论文