Distributed Evolution Strategies for Black-Box Stochastic Optimization

被引:7
作者
He, Xiaoyu [1 ,2 ]
Zheng, Zibin [1 ]
Chen, Chuan [1 ]
Zhou, Yuren [1 ]
Luo, Chuan [3 ]
Lin, Qingwei [4 ]
机构
[1] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[3] Beihang Univ, Sch Software, Beijing 100191, Peoples R China
[4] Microsoft Res, Beijing 100080, Peoples R China
基金
中国国家自然科学基金;
关键词
Smoothing methods; Stochastic processes; Convergence; Optimization methods; Machine learning; Linear programming; Distributed databases; Evolution strategies; distributed optimization; black-box optimization; stochastic optimization; zeroth-order methods; ALGORITHMS; CONVERGENCE;
D O I
10.1109/TPDS.2022.3168873
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
This work concerns the evolutionary approaches to distributed stochastic black-box optimization, in which each worker can individually solve an approximation of the problem with nature-inspired algorithms. We propose a distributed evolution strategy (DES) algorithm grounded on a proper modification to evolution strategies, a family of classic evolutionary algorithms, as well as a careful combination with existing distributed frameworks. On smooth and nonconvex landscapes, DES has a convergence rate competitive to existing zeroth-order methods, and can exploit the sparsity, if applicable, to match the rate of first-order methods. The DES method uses a Gaussian probability model to guide the search and avoids the numerical issue resulted from finite-difference techniques in existing zeroth-order methods. The DES method is also fully adaptive to the problem landscape, as its convergence is guaranteed with any parameter setting. We further propose two alternative sampling schemes which significantly improve the sampling efficiency while leading to similar performance. Simulation studies on several machine learning problems suggest that the proposed methods show much promise in reducing the convergence time and improving the robustness to parameter settings.
引用
收藏
页码:3718 / 3731
页数:14
相关论文
共 55 条
[31]  
Lian XR, 2015, ADV NEUR IN, V28
[32]  
Liu S., 2019, PROC 7 INT C LEARN R
[33]   Sensor Selection for Estimation with Correlated Measurement Noise [J].
Liu, Sijia ;
Chepuri, Sundeep Prabhakar ;
Fardad, Makan ;
Masazade, Engin ;
Leus, Geert ;
Varshney, Pramod K. .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (13) :3509-3522
[34]  
McMahan HB, 2017, PR MACH LEARN RES, V54, P1273
[35]   Stochastic Quasi-Newton Methods [J].
Mokhtari, Aryan ;
Ribeiro, Alejandro .
PROCEEDINGS OF THE IEEE, 2020, 108 (11) :1906-1922
[36]   Random Gradient-Free Minimization of Convex Functions [J].
Nesterov, Yurii ;
Spokoiny, Vladimir .
FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2017, 17 (02) :527-566
[37]  
Nocedal J., 2006, NUMERICAL OPT, V2
[38]   A Survey of Stochastic Simulation and Optimization Methods in Signal Processing [J].
Pereyra, Marcelo ;
Schniter, Philip ;
Chouzenoux, Emilie ;
Pesquet, Jean-Christophe ;
Tourneret, Jean-Yves ;
Hero, Alfred O., III ;
McLaughlin, Steve .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2016, 10 (02) :224-241
[39]  
Qian C, 2018, EVOL COMPUT, V26, P237, DOI [10.1162/evco_a_00201, 10.1162/EVCO_a_00201]
[40]   Noisy evolutionary optimization algorithms - A comprehensive survey [J].
Rakshit, Pratyusha ;
Konar, Amit ;
Das, Swagatam .
SWARM AND EVOLUTIONARY COMPUTATION, 2017, 33 :18-45