TBNF:A Transformer-based Noise Filtering Method for Chinese Long-form Text Matching

被引:2
作者
Gan, Ling [1 ]
Hu, Liuhui [2 ]
Tan, Xiaodong [1 ]
Du, Xinrui [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Software Engn, Chongqing, Peoples R China
关键词
Long text matching; Noise filtering; Transformer; PageRank;
D O I
10.1007/s10489-023-04607-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the field of deep matching, a large amount of noisy data in Chinese long texts affects the matching effect. Most long-form text matching models use all text data indiscriminately, which results in a large amount of noisy data, and thus the PageRank algorithm is combined with Transformer to filter noise. For sentence-level noise detection, after calculating the overlap rate of words to evaluate the similarity, a sentence-level relationship graph is constructed and filtered by using the PageRank algorithm; for word-level noise detection, based on the attention score in Transformer, a word graph is established, then the PageRank algorithm is executed on graph, combined with self-attention weights, to select keywords to highlight topic relevance, the noisy words are filtered sequentially at different layers in the module, layer by layer. In addition, during the model training, PolyLoss is applied to replace the traditional binary Cross-Entropy loss function, thus reducing the difficulty of hyperparameter tuning. Finally, a better filtering strategy is proposed and experiments are conducted to verify it on two Chinese long-form text matching datasets. The result shows that the matching model based on the noise filtering strategy of this paper can better filter the noise and capture the matching signal more accurately.
引用
收藏
页码:22313 / 22327
页数:15
相关论文
共 46 条
  • [1] Latent Dirichlet allocation
    Blei, DM
    Ng, AY
    Jordan, MI
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) : 993 - 1022
  • [2] The anatomy of a large-scale hypertextual Web search engine
    Brin, S
    Page, L
    [J]. COMPUTER NETWORKS AND ISDN SYSTEMS, 1998, 30 (1-7): : 107 - 117
  • [3] Child R, 2019, ARXIV
  • [4] Convolutional Neural Networks for Soft-Matching N-Grams in Ad-hoc Search
    Dai, Zhuyun
    Xiong, Chenyan
    Callan, Jamie
    Liu, Zhiyuan
    [J]. WSDM'18: PROCEEDINGS OF THE ELEVENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2018, : 126 - 134
  • [5] Dai ZH, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P2978
  • [6] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [7] Modeling Diverse Relevance Patterns in Ad-hoc Retrieval
    Fan, Yixing
    Guo, Jiafeng
    Lan, Yanyan
    Xu, Jun
    Zhai, Chengxiang
    Cheng, Xueqi
    [J]. ACM/SIGIR PROCEEDINGS 2018, 2018, : 375 - 384
  • [8] A Deep Relevance Matching Model for Ad-hoc Retrieval
    Guo, Jiafeng
    Fan, Yixing
    Ai, Qingyao
    Croft, W. Bruce
    [J]. CIKM'16: PROCEEDINGS OF THE 2016 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2016, : 55 - 64
  • [9] Hu BT, 2014, ADV NEUR IN, V27
  • [10] Huang PS, 2013, PROCEEDINGS OF THE 22ND ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM'13), P2333