Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and Challenges

被引:19
作者
Wang, Jiajia [1 ]
Huang, Jimmy Xiangji [2 ]
Tu, Xinhui [3 ]
Wang, Junmei [4 ]
Huang, Angela Jennifer [5 ]
Laskar, Md Tahmid Rahman [6 ,7 ]
Bhuiyan, Amran [2 ]
机构
[1] Henan Univ Technol, Sch Sci, Zhengzhou 450001, Henan, Peoples R China
[2] York Univ, Informat Retrieval & Knowledge Management Res Lab, 4700 Keele St, Toronto, ON M3J 1P3, Canada
[3] Cent China Normal Univ, Sch Comp Sci, Wuhan 430079, Hubei, Peoples R China
[4] Hangzhou Dianzi Univ, Sch Comp, Hangzhou 310018, Peoples R China
[5] York Univ, Lassonde Sch Engn, Toronto, ON M3J 2S5, Canada
[6] York Univ, 4700 Keele St, Toronto, ON M3J 1P3, Canada
[7] Dialpad Inc, 4700 Keele St, Toronto, ON M3J 1P3, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
BERT; information retrieval; natural language processing; artificial intelligence;
D O I
10.1145/3648471
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Recent years have witnessed a substantial increase in the use of deep learning to solve various natural language processing (NLP) problems. Early deep learning models were constrained by their sequential or unidirectional nature, such that they struggled to capture the contextual relationships across text inputs. The introduction of bidirectional encoder representations from transformers (BERT) leads to a robust encoder for the transformer model that can understand the broader context and deliver state-of-the-art performance across various NLP tasks. This has inspired researchers and practitioners to apply BERT to practical problems, such as information retrieval (IR). A survey that focuses on a comprehensive analysis of prevalent approaches that apply pretrained transformer encoders like BERT to IR can thus be useful for academia and the industry. In light of this, we revisit a variety of BERT-based methods in this survey, cover a wide range of techniques of IR, and group them into six high-level categories: (i) handling long documents, (ii) integrating semantic information, (iii) balancing effectiveness and efficiency, (iv) predicting the weights of terms, (v) query expansion, and (vi) document expansion. We also provide links to resources, including datasets and toolkits, for BERT-based IR systems. Additionally, we highlight the advantages of employing encoder-based BERT models in contrast to recent large language models like ChatGPT, which are decoder-based and demand extensive computational resources. Finally, we summarize the comprehensive outcomes of the survey and suggest directions for future research in the area.
引用
收藏
页数:33
相关论文
共 155 条
[1]   Improving BERT-based Query-by-Document Retrieval with Multi-task Optimization [J].
Abolghasemi, Amin ;
Verberne, Suzan ;
Azzopardi, Leif .
ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 :3-12
[2]  
Alsentzer E, 2019, Arxiv, DOI [arXiv:1904.03323, DOI 10.48550/ARXIV.1904.03323]
[3]  
Anil R, 2023, Arxiv, DOI [arXiv:2305.10403, 10.48550/arXiv.2305.10403, DOI 10.48550/ARXIV.2305.10403]
[4]  
Babashzadeh A., 2023, HLTH INFORM SCI
[5]   Working memory [J].
Baddeley, Alan .
CURRENT BIOLOGY, 2010, 20 (04) :R136-R140
[6]  
Bai Y, 2020, Arxiv, DOI arXiv:2010.00768
[7]  
Brown TB, 2020, ADV NEUR IN, V33
[8]  
Cer D, 2018, Arxiv, DOI [arXiv:1803.11175, 10.48550/ARXIV.1803.11175, DOI 10.48550/ARXIV.1803.11175]
[9]   Enhancing Recurrent Neural Networks with Positional Attention for Question Answering [J].
Chen, Qin ;
Hu, Qinmin ;
Huang, Jimmy Xiangji ;
He, Liang ;
An, Weijie .
SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, :993-996
[10]  
Chowdhery A, 2023, J MACH LEARN RES, V24