Distributed Evolution Strategies for Black-Box Stochastic Optimization

被引:7
作者
He, Xiaoyu [1 ,2 ]
Zheng, Zibin [1 ]
Chen, Chuan [1 ]
Zhou, Yuren [1 ]
Luo, Chuan [3 ]
Lin, Qingwei [4 ]
机构
[1] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[3] Beihang Univ, Sch Software, Beijing 100191, Peoples R China
[4] Microsoft Res, Beijing 100080, Peoples R China
基金
中国国家自然科学基金;
关键词
Smoothing methods; Stochastic processes; Convergence; Optimization methods; Machine learning; Linear programming; Distributed databases; Evolution strategies; distributed optimization; black-box optimization; stochastic optimization; zeroth-order methods; ALGORITHMS; CONVERGENCE;
D O I
10.1109/TPDS.2022.3168873
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
This work concerns the evolutionary approaches to distributed stochastic black-box optimization, in which each worker can individually solve an approximation of the problem with nature-inspired algorithms. We propose a distributed evolution strategy (DES) algorithm grounded on a proper modification to evolution strategies, a family of classic evolutionary algorithms, as well as a careful combination with existing distributed frameworks. On smooth and nonconvex landscapes, DES has a convergence rate competitive to existing zeroth-order methods, and can exploit the sparsity, if applicable, to match the rate of first-order methods. The DES method uses a Gaussian probability model to guide the search and avoids the numerical issue resulted from finite-difference techniques in existing zeroth-order methods. The DES method is also fully adaptive to the problem landscape, as its convergence is guaranteed with any parameter setting. We further propose two alternative sampling schemes which significantly improve the sampling efficiency while leading to similar performance. Simulation studies on several machine learning problems suggest that the proposed methods show much promise in reducing the convergence time and improving the robustness to parameter settings.
引用
收藏
页码:3718 / 3731
页数:14
相关论文
共 55 条
[1]   Information-Theoretic Lower Bounds on the Oracle Complexity of Stochastic Convex Optimization [J].
Agarwal, Alekh ;
Bartlett, Peter L. ;
Ravikumar, Pradeep ;
Wainwright, Martin J. .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2012, 58 (05) :3235-3249
[2]   Drift Theory in Continuous Search Spaces: Expected Hitting Time of the (1+1)-ES with 1/5 Success Rule [J].
Akimoto, Youhei ;
Auger, Anne ;
Glasmachers, Tobias .
GECCO'18: PROCEEDINGS OF THE 2018 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2018, :801-808
[3]   Evolution strategies – A comprehensive introduction [J].
Hans-Georg Beyer ;
Hans-Paul Schwefel .
Natural Computing, 2002, 1 (1) :3-52
[4]  
Astete-Morales S., 2015, P 13 ACM C FDN GEN A, P76
[5]   LINEAR CONVERGENCE OF COMPARISON-BASED STEP-SIZE ADAPTIVE RANDOMIZED SEARCH VIA STABILITY OF MARKOV CHAINS [J].
Auger, Anne ;
Hansen, Nikolaus .
SIAM JOURNAL ON OPTIMIZATION, 2016, 26 (03) :1589-1624
[6]  
Baydin AG, 2018, J MACH LEARN RES, V18
[7]  
Bernstein J, 2018, PR MACH LEARN RES, V80
[8]   Toward a Steady-State Analysis of an Evolution Strategy on a Robust Optimization Problem With Noise-Induced Multimodality [J].
Beyer, Hans-Georg ;
Sendhoff, Bernhard .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2017, 21 (04) :629-643
[9]   Optimization Methods for Large-Scale Machine Learning [J].
Bottou, Leon ;
Curtis, Frank E. ;
Nocedal, Jorge .
SIAM REVIEW, 2018, 60 (02) :223-311
[10]   Adaptive Stochastic Optimization: A Framework for Analyzing Stochastic Optimization Algorithms [J].
Curtis, Frank E. ;
Scheinberg, Katya .
IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (05) :32-42