Survey of Online Learning Algorithms for Streaming Data Classification

被引:0
|
作者
Zhai T.-T. [1 ,2 ]
Gao Y. [2 ]
Zhu J.-W. [1 ]
机构
[1] School of Information Engineering, Yangzhou University, Yangzhou
[2] State Key Laboratory for Novel Software Technology (Nanjing University), Nanjing
来源
Ruan Jian Xue Bao/Journal of Software | 2020年 / 31卷 / 04期
基金
中国国家自然科学基金;
关键词
Concept drifting; Curse of dimensionality; Evolving data stream classification; Online learning; Sparse online learning; Streaming data classification;
D O I
10.13328/j.cnki.jos.005916
中图分类号
学科分类号
摘要
The objective of streaming data classification is to learn incrementally a decision function that maps input variables to a label variable, from continuously arriving streaming data, so as to accurately classify the test data that may arrive anytime. The online learning paradigm, as an incremental machine learning technology, is an effective tool for classification of streaming data. This paper mainly summarizes, from the perspective of online learning, the recent development of algorithms for streaming data classification. Specifically, the basic framework and the performance evaluation methodology of online learning are first introduced. Then, the latest development of online learning algorithms for general streaming data, for alleviating the "curse of dimensionality" problem in high-dimensional streaming data, and for resolving the "concept drifting" problem in evolving streaming data are reviewed respectively. Finally, future challenges and promising research directions for classification of high-dimensional and evolving streaming data are also discussed. © Copyright 2020, Institute of Software, the Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:912 / 931
页数:19
相关论文
共 137 条
  • [11] Daniely A., Gonen A., Shalev-Shwartz S., Strongly adaptive online learning, Proc. of the Int'l Conf. on Machine Learning (ICML 2015), pp. 1405-1411, (2015)
  • [12] Gyorgy A., Szepesvari C., Shifting regret, mirror descent, and matrices, Proc. of the Int'l Conf. on Machine Learning (ICML 2016), pp. 2943-2951, (2016)
  • [13] Shamir O., Szlak L., Online learning with local permutations and delayed feedback, Proc. of the 34th Int'l Conf. on Machine Learning (ICML 2017), pp. 3086-3094, (2017)
  • [14] Quanrud K., Khashabi D., Online learning with adversarial delays, Proc. of the Advances in Neural Information Processing Systems (NIPS 2015), pp. 1270-1278, (2015)
  • [15] Luo H., Agarwal A., Cesa-Bianchi N., Langford J., Efficient second order online learning by sketching, Proc. of the Advances in Neural Information Processing Systems (NIPS 2016), pp. 902-910, (2016)
  • [16] Erven T., Koolen W.M., MetaGrad: Multiple learning rates in online learning, Proc. of the Advances in Neural Information Processing Systems (NIPS 2016), pp. 3666-3674, (2016)
  • [17] Luo H., Wei C.Y., Zheng K., Efficient online portfolio with logarithmic regret, Proc. of the Advances in Neural Information Processing Systems (NIPS 2018), pp. 8245-8255, (2018)
  • [18] Gillen S., Jung C., Kearns M.J., Roth A., Online learning with an unknown fairness metric, Proc. of the Advances in Neural Information Processing Systems (NIPS 2018), pp. 2605-2614, (2018)
  • [19] Lu J., Hoi S.C.H., Wang J., Zhao P., Liu Z., Large scale online kernel learning, Journal of Machine Learning Research, 17, pp. 47:1-47:43, (2016)
  • [20] Shi T., Zhu J., Online Bayesian passive-aggressive learning, Journal of Machine Learning Research, 18, pp. 33:1-33:39, (2017)