A Deep Fusion Matching Network Semantic Reasoning Model

被引:118
作者
Zheng, Wenfeng [1 ]
Zhou, Yu [1 ]
Liu, Shan [1 ]
Tian, Jiawei [1 ]
Yang, Bo [1 ]
Yin, Lirong [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Automat, Chengdu 610054, Peoples R China
[2] Louisiana State Univ, Dept Geog & Anthropol, Baton Rouge, LA 70803 USA
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 07期
关键词
sentence representation; semantic reasoning; attention mechanism; long-short memory network; deep fusion matching network; NEURAL-NETWORK;
D O I
10.3390/app12073416
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As the vital technology of natural language understanding, sentence representation reasoning technology mainly focuses on sentence representation methods and reasoning models. Although the performance has been improved, there are still some problems, such as incomplete sentence semantic expression, lack of depth of reasoning model, and lack of interpretability of the reasoning process. Given the reasoning model's lack of reasoning depth and interpretability, a deep fusion matching network is designed in this paper, which mainly includes a coding layer, matching layer, dependency convolution layer, information aggregation layer, and inference prediction layer. Based on a deep matching network, the matching layer is improved. Furthermore, the heuristic matching algorithm replaces the bidirectional long-short memory neural network to simplify the interactive fusion. As a result, it improves the reasoning depth and reduces the complexity of the model; the dependency convolution layer uses the tree-type convolution network to extract the sentence structure information along with the sentence dependency tree structure, which improves the interpretability of the reasoning process. Finally, the performance of the model is verified on several datasets. The results show that the reasoning effect of the model is better than that of the shallow reasoning model, and the accuracy rate on the SNLI test set reaches 89.0%. At the same time, the semantic correlation analysis results show that the dependency convolution layer is beneficial in improving the interpretability of the reasoning process.
引用
收藏
页数:18
相关论文
共 38 条
[1]  
[Anonymous], 2014, Advances in Neural Information Processing Systems
[2]  
[Anonymous], 2015, Learning natural language inference with LSTM
[3]   ESCA: Event-Based Split-CNN Architecture with Data-Level Parallelism on UltraScale plus FPGA [J].
Bhowmik, Pankaj ;
Pantho, Md Jubaer Hossain ;
Mbongue, Joel Mandebi ;
Bobda, Christophe .
2021 IEEE 29TH ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM 2021), 2021, :176-180
[4]  
Bowman S. R., 2015, C EMP METH NAT LANG, DOI DOI 10.18653/V1/D15-1075
[5]  
Chen D., 2014, P 2014 C EMP METH NA, P740, DOI 10.3115/v1/d14-1082
[6]   Enhanced LSTM for Natural Language Inference [J].
Chen, Qian ;
Zhu, Xiaodan ;
Ling, Zhenhua ;
Wei, Si ;
Jiang, Hui ;
Inkpen, Diana .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :1657-1668
[7]  
Cheng Z., 2021, INT J COMPUT INF ENG, V15, P119
[8]  
Chung Junyoung, 2014, ARXIV
[9]  
De Marneffe M. C., 2006, Lrec, P449
[10]  
Duan CQ, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4033