Advancing continual lifelong learning in neural information retrieval: Definition, dataset, framework, and empirical evaluation

被引:1
作者
Hou, Jingrui [1 ]
Cosma, Georgina [1 ]
Finke, Axel [2 ]
机构
[1] Loughborough Univ, Sch Sci, Dept Comp Sci, Epinal Way, Loughborough LE11 3TU, Leics, England
[2] Loughborough Univ, Dept Math Sci, Epinal Way, Loughborough LE11 3TU, Leics, England
关键词
Neural information retrieval; Continual learning; Catastrophic forgetting; Topic shift; Data augmentation;
D O I
10.1016/j.ins.2024.121368
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Continual learning refers to the capability of a machine learning model to learn and adapt to new information, without compromising its performance on previously learned tasks. Although several studies have investigated continual learning methods for neural information retrieval (NIR) tasks, a well-defined task definition is still lacking, and it is unclear how typical learning strategies perform in this context. To address this challenge, a systematic task definition of continual NIR is presented, along with a multiple-topic dataset that simulates continuous information retrieval. A comprehensive continual neural information retrieval framework consisting of typical retrieval models and continual learning strategies is then proposed. Empirical evaluations illustrate that the proposed framework can successfully prevent catastrophic forgetting in neural information retrieval and enhance performance on previously learned tasks. The results also indicate that embedding-based retrieval models experience a decline in their continual learning performance as the topic shift distance and dataset volume of new tasks increase. In contrast, pretraining-based models do not show any such correlation. Adopting suitable learning strategies can mitigate the effects of topic shift and data augmentation in continual neural information retrieval.
引用
收藏
页数:17
相关论文
共 46 条
[1]  
Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
[2]   Memory Aware Synapses: Learning What (not) to Forget [J].
Aljundi, Rahaf ;
Babiloni, Francesca ;
Elhoseiny, Mohamed ;
Rohrbach, Marcus ;
Tuytelaars, Tinne .
COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 :144-161
[3]   Expert Gate: Lifelong Learning with a Network of Experts [J].
Aljundi, Rahaf ;
Chakravarty, Punarjay ;
Tuytelaars, Tinne .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7120-7129
[4]  
[Anonymous], 2013, An Introduction to Information Retrieval
[5]  
Biesialska Magdalena, 2020, P 28 INT C COMP LING, P6523, DOI [10.18653/v1/2020.coling- main.574, DOI 10.18653/V1/2020.COLINGMAIN.574, 10.18653/v1/2020.coling-main.574]
[6]   Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence [J].
Chaudhry, Arslan ;
Dokania, Puneet K. ;
Ajanthan, Thalaiyasingam ;
Torr, Philip H. S. .
COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 :556-572
[7]   Continual Learning for Generative Retrieval over Dynamic Corpora [J].
Chen, Jiangui ;
Zhang, Ruqing ;
Guo, Jiafeng ;
de Rijke, Maarten ;
Chen, Wei ;
Fan, Yixing ;
Cheng, Xueqi .
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, :306-315
[8]   A Continual Learning Survey: Defying Forgetting in Classification Tasks [J].
De Lange, Matthias ;
Aljundi, Rahaf ;
Masana, Marc ;
Parisot, Sarah ;
Jia, Xu ;
Leonardis, Ales ;
Slabaugh, Greg ;
Tuytelaars, Tinne .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) :3366-3385
[9]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[10]   Catastrophic forgetting in connectionist networks [J].
French, RM .
TRENDS IN COGNITIVE SCIENCES, 1999, 3 (04) :128-135