Stochastic Approximation for Expensive One-Bit Feedback Systems

被引:0
作者
Zhang, Xiaoqin [1 ]
Ma, Huimin [1 ]
Wen, Jinghuan [1 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
关键词
stochastic approximation; parameter optimization; one-bit feedback system; regression; Maximum Likelihood Estimation (MLE); PARTICLE SWARM; OPTIMIZATION; CONVERGENCE;
D O I
10.23919/TST.2017.7914203
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One-bit feedback systems generate binary data as their output and the system performance is usually measured by the success rate with a fixed parameter combination. Traditional methods need many executions for parameter optimization. Hence, it is impractical to utilize these methods in Expensive One-Bit Feedback Systems (EOBFSs), where a single system execution is costly in terms of time or money. In this paper, we propose a novel algorithm, named Iterative Regression and Optimization (IRO), for parameter optimization and its corresponding scheme based on the Maximum Likelihood Estimation (MLE) method and Particle Swarm Optimization (PSO) method, named MLEPSO-IRO, for parameter optimization in EOBFSs. The IRO algorithm is an iterative algorithm, with each iteration comprising two parts: regression and optimization. Considering the structure of IRO and the Bernoulli distribution property of the output of EOBFSs, MLE and a modified PSO are selected to implement the regression and optimization sections, respectively, in MLEPSO-IRO. We also provide a theoretical analysis for the convergence of MLEPSO-IRO and provide numerical experiments on hypothesized EOBFSs and one real EOBFS in comparison to traditional methods. The results indicate that MLEPSO-IRO can provide a much better result with only a small amount of system executions.
引用
收藏
页码:317 / 327
页数:11
相关论文
共 15 条
[1]  
Bishop C.M., 2006, PATTERN RECOGN, V4, P738, DOI DOI 10.1117/1.2819119
[2]   The particle swarm - Explosion, stability, and convergence in a multidimensional complex space [J].
Clerc, M ;
Kennedy, J .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (01) :58-73
[3]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[4]   Mini-batch stochastic approximation methods for nonconvex stochastic composite optimization [J].
Ghadimi, Saeed ;
Lan, Guanghui ;
Zhang, Hongchao .
MATHEMATICAL PROGRAMMING, 2016, 155 (1-2) :267-305
[5]  
Hou Yu, 2012, International Conference on Automatic Control and Artificial Intelligence (ACAI 2012), P496
[6]  
Kennedy J., 1995, 1995 IEEE International Conference on Neural Networks Proceedings (Cat. No.95CH35828), P1942, DOI 10.1109/ICNN.1995.488968
[7]   STOCHASTIC ESTIMATION OF THE MAXIMUM OF A REGRESSION FUNCTION [J].
KIEFER, J ;
WOLFOWITZ, J .
ANNALS OF MATHEMATICAL STATISTICS, 1952, 23 (03) :462-466
[8]  
Kolda TG, 2003, SIAM REV, V45, P385, DOI [10.1137/S003614450242889, 10.1137/S0036144502428893]
[9]  
Maryak JL, 2001, P AMER CONTR CONF, P756, DOI 10.1109/ACC.2001.945806
[10]   A STOCHASTIC APPROXIMATION METHOD [J].
ROBBINS, H ;
MONRO, S .
ANNALS OF MATHEMATICAL STATISTICS, 1951, 22 (03) :400-407