A Cross-Sentence Latent Variable Model for Semi-Supervised Text Sequence Matching

被引:0
作者
Choi, Jihun [1 ]
Kim, Taeuk [1 ]
Lee, Sang-goo [1 ]
机构
[1] Seoul Natl Univ, Dept Comp Sci & Engn, Seoul, South Korea
来源
57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019) | 2019年
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a latent variable model for predicting the relationship between a pair of text sequences. Unlike previous auto-encoding- based approaches that consider each sequence separately, our proposed framework utilizes both sequences within a single model by generating a sequence that has a given relationship with a source sequence. We further extend the cross-sentence generating framework to facilitate semi-supervised training. We also define novel semantic constraints that lead the decoder network to generate semantically plausible and diverse sequences. We demonstrate the effectiveness of the proposed model from quantitative and qualitative experiments, while achieving state-of-the-art results on semi-supervised natural language inference and paraphrase identification.
引用
收藏
页码:4747 / 4761
页数:15
相关论文
共 61 条
[1]  
Alemi A., 2018, INT C MACHINE LEARNI, P159
[2]  
[Anonymous], INT C LEARN REPR TOU
[3]  
[Anonymous], 2006, Semi-Supervised Learning, DOI DOI 10.7551/MITPRESS/9780262033589.003.0003
[4]  
[Anonymous], IEEE INT C NETW SENS
[5]  
[Anonymous], 2016, P 29 C NEUR INF PROC
[6]  
[Anonymous], ADAPTIVE COMPUTATION
[7]  
[Anonymous], INT C LEARN REPR TOU
[8]  
[Anonymous], 2015, NEURIPS
[9]  
[Anonymous], P 27 INT JOINT C ART
[10]  
[Anonymous], 2008, Technical report