Online Learning Algorithms

被引:18
作者
Cesa-Bianchi, Nicolo [1 ,2 ]
Orabona, Francesco [3 ]
机构
[1] Univ Milan, Dept Comp Sci, I-20133 Milan, Italy
[2] Univ Milan, Data Sci Res Ctr, I-20133 Milan, Italy
[3] Boston Univ, Dept Elect & Comp Engn, Boston, MA 02215 USA
来源
ANNUAL REVIEW OF STATISTICS AND ITS APPLICATION, VOL 8, 2021 | 2021年 / 8卷
基金
美国国家科学基金会;
关键词
regret minimization; convex optimization; regression; classification; SUBGRADIENT METHODS; PERCEPTRON; GRADIENT; DESCENT; MODEL;
D O I
10.1146/annurev-statistics-040620-035329
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Online learning is a framework for the design and analysis of algorithms that build predictive models by processing data one at the time. Besides being computationally efficient, online algorithms enjoy theoretical performance guarantees that do not rely on statistical assumptions on the data source. In this review, we describe some of the most important algorithmic ideas behind online learning and explain the main mathematical tools for their analysis. Our reference framework is online convex optimization, a sequential version of convex optimization within which most online algorithms are formulated. More specifically, we provide an in-depth description of online mirror descent and follow the regularized leader, two of the most fundamental algorithms in online learning. As the tuning of parameters is a typically difficult task in sequential data analysis, in the last part of the review we focus on coin-betting, an information-theoretic approach to the design of parameter-free online algorithms with good theoretical guarantees.
引用
收藏
页码:165 / 190
页数:26
相关论文
共 78 条
[1]  
Abernethy J. D., 2008, COLT, P263
[2]   Convex Optimization: Algorithms and Complexity [J].
不详 .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2015, 8 (3-4) :232-+
[3]  
[Anonymous], 2003, PROC INT C MACHINE L
[4]  
[Anonymous], 2013, PROC C LEARN THEORY
[5]  
[Anonymous], 1951, P 2 BERKELEY S MATH
[6]  
Arora R., 2012, P 29 INT C MACH LEAR
[7]  
Arora S, 2012, Theory of Computing, V8, P121, DOI 10.4086/toc.2012.v008a006
[8]   Relative loss bounds for on-line density estimation with the exponential family of distributions [J].
Azoury, KS ;
Warmuth, MK .
MACHINE LEARNING, 2001, 43 (03) :211-246
[9]   Mirror descent and nonlinear projected subgradient methods for convex optimization [J].
Beck, A ;
Teboulle, M .
OPERATIONS RESEARCH LETTERS, 2003, 31 (03) :167-175
[10]  
Berger James O, 2013, Statistical decision theory and Bayesian analysis