Improving topic disentanglement via contrastive learning

被引:11
作者
Zhou, Xixi [1 ]
Bu, Jiajun [1 ]
Zhou, Sheng [1 ]
Yu, Zhi [1 ]
Zhao, Ji [2 ]
Yan, Xifeng [3 ]
机构
[1] Zhejiang Univ, 38 Zheda Rd, Hangzhou 310027, Peoples R China
[2] Shanghai Pudong Dev Bank Co Ltd, Shanghai Branch, Shanghai 200123, Peoples R China
[3] Univ Calif Santa Barbara, Santa Barbara, CA 93106 USA
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Topic model; Contrastive learning; Disentanglement;
D O I
10.1016/j.ipm.2022.103164
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the emergence and development of deep generative models, such as the variational auto-encoders (VAEs), the research on topic modeling successfully extends to a new area: neural topic modeling, which aims to learn disentangled topics to understand the data better. However, the original VAE framework had been shown to be limited in disentanglement performance, bringing their inherent defects to a neural topic model (NTM). In this paper, we put forward that the optimization objectives of contrastive learning are consistent with two important goals (alignment and uniformity) of well-disentangled topic learning. Also, the optimization objectives of contrastive learning are consistent with two key evaluation measures for topic models, topic coherence and topic diversity. So, we come to the important conclusion that alignment and uniformity of disentangled topic learning can be quantified with topic coherence and topic diversity. Accordingly, we are inspired to propose the Contrastive Disentangled Neural Topic Model (CNTM). By representing both words and topics as low-dimensional vectors in the same embedding space, we apply contrastive learning to neural topic modeling to produce factorized and disentangled topics in an interpretable manner. We compare our proposed CNTM with strong baseline models on widely-used metrics. Our model achieves the best topic coherence scores under the most general evaluation setting (100% proportion topic selected) with 25.0%, 10.9%, 24.6%, and 51.3% improvements above the second-best models' scores reported on four datasets of 20 Newsgroups, Web Snippets, Tag My News, and Reuters, respectively. Our method also gets the second-best topic diversity scores on the dataset of 20Newsgroups and Web Snippets. Our experimental results show that CNTM can effectively leverage the disentanglement ability from contrastive learning to solve the inherent defect of neural topic modeling and obtain better topic quality.
引用
收藏
页数:19
相关论文
共 57 条
[1]  
Aletras N., 2013, P 10 INT C COMPUTATI, P13
[2]  
[Anonymous], 2013, INT C LEARNING REPRE
[3]  
Bahrainian SA, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, P3341
[4]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[5]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[6]  
Bowman Samuel R, 2015, Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning
[7]  
Burgess C., 2017, NIPS 2017 DISENTANGL
[8]  
Burkhardt S, 2019, J MACH LEARN RES, V20
[9]  
Chen RTQ, 2018, 32 C NEURAL INFORM P, V31
[10]  
Chen Ting, 2020, ICML