On Reproducible Implementations in Unsupervised Concept Drift Detection Algorithms Research

被引:1
作者
Lukats, Daniel [1 ,2 ]
Stahl, Frederic [1 ]
机构
[1] German Res Ctr Artificial Intelligence, Marie Curie Str1, D-26129 Oldenburg, Germany
[2] Carl Von Ossietzky Univ Oldenburg, Ammerlander Heerstr 114-118, D-26129 Oldenburg, Germany
来源
ARTIFICIAL INTELLIGENCE XL, AI 2023 | 2023年 / 14381卷
关键词
Reproducibility; Pseudo code; Data mining;
D O I
10.1007/978-3-031-47994-6_16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In order to create reproducible experimentation and algorithms in machine learning and data mining research, reproducible descriptions of the algorithms are needed. These can be in the form of source code, pseudo code and prose. Efforts in academia commonly focus on accessibility of source code. Based on an internal study reproducing unsupervised concept drift detectors, this work argues that a publication's content is equally important and highlights common issues affecting attempts at implementing unsupervised concept drift detectors. These include major issues prohibiting implementation entirely, as well as minor issues, which demand increased effort from the developer. The paper proposes the use of a checklist as a consistent tool to ensure better quality and reproducible publications of algorithms. The issues highlighted in this work could mark a starting point, although future work is required to ensure representation of more diverse areas of research in artificial intelligence.
引用
收藏
页码:204 / 209
页数:6
相关论文
共 15 条
[1]  
aaai-23.aaai.org, Reproducibility Checklist | AAAI 2023 Conference
[2]   WHAT DOES PSEUDO-CODE DO - A PSYCHOLOGICAL ANALYSIS OF THE USE OF PSEUDO-CODE BY EXPERIENCED PROGRAMMERS [J].
BELLAMY, RKE .
HUMAN-COMPUTER INTERACTION, 1994, 9 (02) :225-246
[3]  
choosealicense.com, No License | Choose a License
[4]  
docs.github.com, Licensing a Repository
[5]  
Gawande A, 2009, The Checklist Manifesto. How to Get Things Right
[6]  
github.com, Tips for Publishing Research Code
[7]   Reproducibility standards for machine learning in the life sciences [J].
Heil, Benjamin J. ;
Hoffman, Michael M. ;
Markowetz, Florian ;
Lee, Su-In ;
Greene, Casey S. ;
Hicks, Stephanie C. .
NATURE METHODS, 2021, 18 (10) :1132-1135
[8]   LITERATE PROGRAMMING [J].
KNUTH, DE .
COMPUTER JOURNAL, 1984, 27 (02) :97-111
[9]   Reproducibility of studies on text mining for citation screening in systematic reviews: Evaluation and checklist [J].
Olorisade, Babatunde Kazeem ;
Brereton, Pearl ;
Andras, Peter .
JOURNAL OF BIOMEDICAL INFORMATICS, 2017, 73 :1-13
[10]  
paperswithcode.com, Papers with Code