Evaluation of Lifelong Learning Systems

被引:0
作者
Prokopalo, Yevhenii [1 ]
Meignier, Sylvain [1 ]
Galibert, Olivier [2 ]
Barrault, Loic [3 ]
Larcher, Anthony [1 ]
机构
[1] Le Mans Univ, LIUM, Le Mans, France
[2] LNE, Paris, France
[3] Univ Sheffield, Sheffield, S Yorkshire, England
来源
PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020) | 2020年
关键词
Evaluation; lifelong learning; human assisted learning;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Current intelligent systems need the expensive support of machine learning experts to sustain their performance level when used on a daily basis. To reduce this cost, i.e. remaining free from any machine learning expert, it is reasonable to implement lifelong (or continuous) learning intelligent systems that will continuously adapt their model when facing changing execution conditions. In this work, the systems are allowed to refer to human domain experts who can provide the system with relevant knowledge about the task. Nowadays, the fast growth of lifelong learning systems development rises the question of their evaluation. In this article we propose a generic evaluation methodology for the specific case of lifelong learning systems. Two steps will be considered. First, the evaluation of human-assisted learning (including active and/or interactive learning) outside the context of lifelong learning. Second, the system evaluation across time, with propositions of how a lifelong learning intelligent system should be evaluated when including human assisted learning or not.
引用
收藏
页码:1833 / 1841
页数:9
相关论文
共 15 条
[11]   BLEU: a method for automatic evaluation of machine translation [J].
Papineni, K ;
Roukos, S ;
Ward, T ;
Zhu, WJ .
40TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2002, :311-318
[12]  
Perez-Dattari Rodrigo, 2018, INT S EXP ROB, P353
[13]  
Prince Simon, 2012, Computer Vision: Models, Learning, and Inference, Vfirst
[14]  
Siddhant A, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P2904
[15]  
Snover M, 2006, A study of translation edit rate with targeted human annotation