Feature Differentiation and Fusion for Semantic Text Matching

被引:2
作者
Peng, Rui [1 ]
Hong, Yu [1 ]
Jin, Zhiling [1 ]
Yao, Jianmin [1 ]
Zhou, Guodong [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
来源
ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT II | 2023年 / 13981卷
基金
美国国家科学基金会; 国家重点研发计划;
关键词
Semantic Text Matching; Deep neural networks; Natural Language Processing; NETWORK;
D O I
10.1007/978-3-031-28238-6_3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Semantic Text Matching (STM for short) stands for the task of automatically determining the semantic similarity for a pair of texts. It has been widely applied in a variety of downstream tasks, e.g., information retrieval and question answering. The most recent works of STM leverage Pre-trained Language Models (abbr., PLMs) due to their remarkable capacity for representation learning. Accordingly, significant improvements have been achieved. However, our findings show that PLMs fail to capture task-specific features that signal hardly-perceptible changes in semantics. To overcome the issue, we propose a two-channel Feature Differentiation and Fusion network (FDF). It utilizes a PLM-based encoder to extract features separately from the unabridged texts and those abridged by deduplication. On this basis, gated feature fusion and interaction are conducted across the channels to expand text representations with attentive and distinguishable features. Experiments on the benchmarks QQP, MRPC and BQ show that FDF obtains substantial improvements compared to the baselines and outperforms the state-of-the-art STM models.
引用
收藏
页码:32 / 46
页数:15
相关论文
共 24 条
[1]  
Chen J, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P4946
[2]   Enhanced LSTM for Natural Language Inference [J].
Chen, Qian ;
Zhu, Xiaodan ;
Ling, Zhenhua ;
Wei, Si ;
Jiang, Hui ;
Inkpen, Diana .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :1657-1668
[3]   Pre-Training With Whole Word Masking for Chinese BERT [J].
Cui, Yiming ;
Che, Wanxiang ;
Liu, Ting ;
Qin, Bing ;
Yang, Ziqing .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 :3504-3514
[4]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[5]  
Dolan WilliamB., P 3 INT WORKSHOP PAR
[6]  
Gao J., 2019, INT C LEARN REPR, P1
[7]  
Gong Y, 2017, ARXIV170904348
[8]   Text-Attentional Convolutional Neural Network for Scene Text Detection [J].
He, Tong ;
Huang, Weilin ;
Qiao, Yu ;
Yao, Jian .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (06) :2529-2541
[9]  
Iyer S., 2017, First quora dataset release: Question pairs. data
[10]  
Kim S, 2019, AAAI CONF ARTIF INTE, P6586