Cross-sentence Pre-trained model for Interactive QA matching

被引:0
作者
Wu, Jinmeng [1 ,2 ]
Hao, Yanbin [3 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan, Peoples R China
[2] Univ Liverpool, Sch Elect Engn Elect & Comp Sci, Brownlow Hill, Liverpool, Merseyside, England
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
来源
PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020) | 2020年
基金
中国国家自然科学基金;
关键词
Question answering; Interactive matching; Pre-trained language model; Context jump Dependencies;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Semantic matching measures the dependencies between query and answer representations, which is an important criterion for evaluating whether the matching is successful. In fact, such matching does not examine each sentence individually, because the context information between sentences should be considered equally important to the syntactic context inside a sentence. Considering the above, we propose a novel QA matching model, built upon a cross-sentence context-aware architecture. Specifically, an interactive attention mechanism with a pre-trained language model is presented to automatically select salient positional answer representations that contribute more significantly to the answer relevance of a given question. In addition to the context information captured at each word position, we incorporate a quantity of context jump dependencies to leverage the attention weight formulation. This can capture the amount of useful information brought by the next word, is computed by modeling the joint probability between two adjacent word states. The proposed method is compared with multiple state-of-the-art methods using the TREC library, WikiQA, and the Yahoo! community question datasets. Experimental results show that the proposed method outperforms satisfactorily the competing ones.
引用
收藏
页码:5417 / 5424
页数:8
相关论文
共 50 条
[31]   ARoBERT: An ASR Robust Pre-Trained Language Model for Spoken Language Understanding [J].
Wang, Chengyu ;
Dai, Suyang ;
Wang, Yipeng ;
Yang, Fei ;
Qiu, Minghui ;
Chen, Kehan ;
Zhou, Wei ;
Huang, Jun .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 :1207-1218
[32]   CAM-BERT: Chinese Aerospace Manufacturing Pre-trained Language Model [J].
Dai, Jinchi ;
Wang, Shengren ;
Wang, Peiyan ;
Li, Ruiting ;
Chen, Jiaxin ;
Li, Xinrong .
2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, :361-365
[33]   FGeneBERT: function-driven pre-trained gene language model for metagenomics [J].
Duan, Chenrui ;
Zang, Zelin ;
Xu, Yongjie ;
He, Hang ;
Li, Siyuan ;
Liu, Zihan ;
Lei, Zhen ;
Zheng, Ju-Sheng ;
Li, Stan Z. .
BRIEFINGS IN BIOINFORMATICS, 2025, 26 (02)
[34]   SSAP: Storylines and Sentiment Aware Pre-Trained Model for Story Ending Generation [J].
Liu, Yongkang ;
Huang, Qingbao ;
Li, Jing ;
Mo, Linzhang ;
Cai, Yi ;
Li, Qing .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 :686-694
[35]   NMT Enhancement based on Knowledge Graph Mining with Pre-trained Language Model [J].
Yang, Hao ;
Qin, Ying ;
Deng, Yao ;
Wang, Minghan .
2020 22ND INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY (ICACT): DIGITAL SECURITY GLOBAL AGENDA FOR SAFE SOCIETY!, 2020, :185-189
[36]   Prompting disentangled embeddings for knowledge graph completion with pre-trained language model [J].
Geng, Yuxia ;
Chen, Jiaoyan ;
Zeng, Yuhang ;
Chen, Zhuo ;
Zhang, Wen ;
Pan, Jeff Z. ;
Wang, Yuxiang ;
Xu, Xiaoliang .
EXPERT SYSTEMS WITH APPLICATIONS, 2025, 268
[37]   Better Few-Shot Text Classification with Pre-trained Language Model [J].
Chen, Zheng ;
Zhang, Yunchen .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 :537-548
[38]   Explainable reasoning over temporal knowledge graphs by pre-trained language model [J].
Li, Qing ;
Wu, Guanzhong .
INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (01)
[39]   A Light Bug Triage Framework for Applying Large Pre-trained Language Model [J].
Lee, Jaehyung ;
Han, Kisun ;
Yu, Hwanjo .
PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,
[40]   Pre-trained Language Model-based Retrieval and Ranking forWeb Search [J].
Zou, Lixin ;
Lu, Weixue ;
Liu, Yiding ;
Cai, Hengyi ;
Chu, Xiaokai ;
Ma, Dehong ;
Shi, Daiting ;
Sun, Yu ;
Cheng, Zhicong ;
Gu, Simiu ;
Wang, Shuaiqiang ;
Yin, Dawei .
ACM TRANSACTIONS ON THE WEB, 2023, 17 (01)