Randomized Block Cubic Newton Method

被引:0
|
作者
Doikov, Nikita [1 ]
Richtarik, Peter [2 ,3 ,4 ]
机构
[1] Natl Res Univ Higher Sch Econ, Samsung HSE Lab, Moscow, Russia
[2] King Abdullah Univ Sci & Technol, Thuwal, Saudi Arabia
[3] Univ Edinburgh, Edinburgh, Midlothian, Scotland
[4] Moscow Inst Phys & Technol, Dolgoprudnyi, Russia
来源
INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80 | 2018年 / 80卷
基金
俄罗斯科学基金会;
关键词
REGULARIZATION; DESCENT;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of minimizing the sum of three convex functions-a differentiable, twice-differentiable and a non-smooth term-in a high dimensional setting. To this effect we propose and analyze a randomized block cubic Newton (RBCN) method, which in each iteration builds a model of the objective function formed as the sum of the natural models of its three components: a linear model with a quadratic regularizer for the differentiable term, a quadratic model with a cubic regularizer for the twice differentiable term, and perfect (proximal) model for the nonsmooth term. Our method in each iteration minimizes the model over a random subset of blocks of the search variable. RBCN is the first algorithm with these properties, generalizing several existing methods, matching the best known bounds in all special cases. We establish O(1/epsilon), O(1/root epsilon) and O(log(1/epsilon)) rates under different assumptions on the component functions. Lastly, we show numerically that our method outperforms the state of the art on a variety of machine learning problems, including cubically regularized least-squares, logistic regression with constraints, and Poisson regression.
引用
收藏
页数:9
相关论文
共 50 条