Supervised learning in the presence of concept drift: a modelling framework

被引:5
作者
Straat, M. [1 ]
Abadi, F. [2 ]
Kan, Z. [1 ]
Goepfert, C. [3 ]
Hammer, B. [3 ]
Biehl, M. [1 ]
机构
[1] Univ Groningen, Bernoulli Inst Math Comp Sci & Artificial Intelli, Nijenborgh 9, NL-9747 AG Groningen, Netherlands
[2] Aksum Univ, Comp Sci Dept, Inst Engn & Technol, Axum, Tigray, Ethiopia
[3] Bielefeld Univ, Machine Learning Grp, CITEC, D-33594 Bielefeld, Germany
关键词
Classification; Regression; Supervised learning; Drifting concepts; Learning vector quantization; Layered neural networks; STATISTICAL-MECHANICS; ONLINE; DYNAMICS; ALGORITHMS; PHYSICS;
D O I
10.1007/s00521-021-06035-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a modelling framework for the investigation of supervised learning in non-stationary environments. Specifically, we model two example types of learning systems: prototype-based learning vector quantization (LVQ) for classification and shallow, layered neural networks for regression tasks. We investigate so-called student-teacher scenarios in which the systems are trained from a stream of high-dimensional, labeled data. Properties of the target task are considered to be non-stationary due to drift processes while the training is performed. Different types of concept drift are studied, which affect the density of example inputs only, the target rule itself, or both. By applying methods from statistical physics, we develop a modelling framework for the mathematical analysis of the training dynamics in non-stationary environments. Our results show that standard LVQ algorithms are already suitable for the training in non-stationary environments to a certain extent. However, the application of weight decay as an explicit mechanism of forgetting does not improve the performance under the considered drift processes. Furthermore, we investigate gradient-based training of layered neural networks with sigmoidal activation functions and compare with the use of rectified linear units. Our findings show that the sensitivity to concept drift and the effectiveness of weight decay differs significantly between the two types of activation function.
引用
收藏
页码:101 / 118
页数:18
相关论文
共 50 条
[21]   A META-LEARNING METHOD FOR CONCEPT DRIFT [J].
Wang, Runxin ;
Shi, Lei ;
Foghlu, Micheal O. ;
Robson, Eric .
KDIR 2010: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND INFORMATION RETRIEVAL, 2010, :257-262
[22]   Learning Under Concept Drift for Regression-A Systematic Literature Review [J].
Lima, Marilia ;
Neto, Manoel ;
Silva Filho, Telmo ;
Fagundes, Roberta A. de A. .
IEEE ACCESS, 2022, 10 :45410-45429
[23]   Concept Drift Evolution In Machine Learning Approaches: A Systematic Literature Review [J].
Hashmani, Manzoor Ahmed ;
Jameel, Syed Muslim ;
Rehman, Mobashar ;
Inoue, Atsushi .
INTERNATIONAL JOURNAL ON SMART SENSING AND INTELLIGENT SYSTEMS, 2020, 13 (01) :1-16
[24]   A unified kernel sparse representation framework for supervised learning problems [J].
Ye, Junyou ;
Yang, Zhixia ;
Zhu, Yongqi ;
Zhang, Zheng .
NEURAL COMPUTING & APPLICATIONS, 2024, 36 (09) :4907-4930
[25]   Enhancing Semi-Supervised Learning With Concept Drift Detection and Self-Training: A Study on Classifier Diversity and Performance [J].
Perez, Jose L. M. ;
Barros, Roberto S. M. ;
Santos, Silas G. T. C. .
IEEE ACCESS, 2025, 13 :24681-24697
[26]   The Impact of Latency on Online Classification Learning with Concept Drift [J].
Marrs, Gary R. ;
Hickey, Ray J. ;
Black, Michaela M. .
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, 2010, 6291 :459-469
[27]   Adaptive online learning for classification under concept drift [J].
Goel, Kanu ;
Batra, Shalini .
INTERNATIONAL JOURNAL OF COMPUTATIONAL SCIENCE AND ENGINEERING, 2021, 24 (02) :128-135
[28]   Concept drift detection and accelerated convergence of online learning [J].
Guo, Husheng ;
Li, Hai ;
Sun, Ni ;
Ren, Qiaoyan ;
Zhang, Aijuan ;
Wang, Wenjian .
KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 65 (03) :1005-1043
[29]   Asynchronous Federated Learning for Sensor Data with Concept Drift [J].
Chen, Yujing ;
Chai, Zheng ;
Cheng, Yue ;
Rangwala, Huzefa .
2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, :4822-4831
[30]   A theoretical framework for supervised learning from regions [J].
Gnecco, Giorgio ;
Gori, Marco ;
Melacci, Stefano ;
Sanguineti, Marcello .
NEUROCOMPUTING, 2014, 129 :25-32