Cross-sentence Pre-trained model for Interactive QA matching

被引:0
作者
Wu, Jinmeng [1 ,2 ]
Hao, Yanbin [3 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Wuhan, Peoples R China
[2] Univ Liverpool, Sch Elect Engn Elect & Comp Sci, Brownlow Hill, Liverpool, Merseyside, England
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
来源
PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020) | 2020年
基金
中国国家自然科学基金;
关键词
Question answering; Interactive matching; Pre-trained language model; Context jump Dependencies;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Semantic matching measures the dependencies between query and answer representations, which is an important criterion for evaluating whether the matching is successful. In fact, such matching does not examine each sentence individually, because the context information between sentences should be considered equally important to the syntactic context inside a sentence. Considering the above, we propose a novel QA matching model, built upon a cross-sentence context-aware architecture. Specifically, an interactive attention mechanism with a pre-trained language model is presented to automatically select salient positional answer representations that contribute more significantly to the answer relevance of a given question. In addition to the context information captured at each word position, we incorporate a quantity of context jump dependencies to leverage the attention weight formulation. This can capture the amount of useful information brought by the next word, is computed by modeling the joint probability between two adjacent word states. The proposed method is compared with multiple state-of-the-art methods using the TREC library, WikiQA, and the Yahoo! community question datasets. Experimental results show that the proposed method outperforms satisfactorily the competing ones.
引用
收藏
页码:5417 / 5424
页数:8
相关论文
共 50 条
  • [21] Robustly Pre-Trained Neural Model for Direct Temporal Relation Extraction
    Guan, Hong
    Li, Jianfu
    Xu, Hua
    Devarakonda, Murthy
    [J]. 2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 501 - 502
  • [22] Biomedical-domain pre-trained language model for extractive summarization
    Du, Yongping
    Li, Qingxiao
    Wang, Lulin
    He, Yanqing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 199 (199)
  • [23] Leveraging Pre-Trained Language Model for Summary Generation on Short Text
    Zhao, Shuai
    You, Fucheng
    Liu, Zeng Yuan
    [J]. IEEE ACCESS, 2020, 8 : 228798 - 228803
  • [24] Using Noise and External Knowledge to Enhance Chinese Pre-trained Model
    Ma, Haoyang
    Li, Zeyu
    Guo, Hongyu
    [J]. 2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 476 - 480
  • [25] EMBERT: A Pre-trained Language Model for Chinese Medical Text Mining
    Cai, Zerui
    Zhang, Taolin
    Wang, Chengyu
    He, Xiaofeng
    [J]. WEB AND BIG DATA, APWEB-WAIM 2021, PT I, 2021, 12858 : 242 - 257
  • [26] An additional approach to pre-trained code model with multilingual natural languages
    Kajiura, Teruno
    Souma, Nao
    Sato, Miyu
    Takahashi, Mai
    Kuramitsu, Kimio
    [J]. 2022 29TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE, APSEC, 2022, : 580 - 581
  • [27] Question-answering Forestry Pre-trained Language Model: ForestBERT
    Tan, Jingwei
    Zhang, Huaiqing
    Liu, Yang
    Yang, Jie
    Zheng, Dongping
    [J]. Linye Kexue/Scientia Silvae Sinicae, 2024, 60 (09): : 99 - 110
  • [28] Classifying Code Comments via Pre-trained Programming Language Model
    Li, Ying
    Wang, Haibo
    Zhang, Huaien
    Tan, Shin Hwei
    [J]. 2023 IEEE/ACM 2ND INTERNATIONAL WORKSHOP ON NATURAL LANGUAGE-BASED SOFTWARE ENGINEERING, NLBSE, 2023, : 24 - 27
  • [29] A Pre-trained Language Model for Medical Question Answering Based on Domain Adaption
    Liu, Lang
    Ren, Junxiang
    Wu, Yuejiao
    Song, Ruilin
    Cheng, Zhen
    Wang, Sibo
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II, 2022, 13552 : 216 - 227
  • [30] ARoBERT: An ASR Robust Pre-Trained Language Model for Spoken Language Understanding
    Wang, Chengyu
    Dai, Suyang
    Wang, Yipeng
    Yang, Fei
    Qiu, Minghui
    Chen, Kehan
    Zhou, Wei
    Huang, Jun
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 1207 - 1218