Active Learning and Bayesian Optimization: A Unified Perspective to Learn with a Goal

被引:16
作者
Di Fiore, Francesco [1 ]
Nardelli, Michela [1 ]
Mainini, Laura [1 ,2 ,3 ]
机构
[1] Politecn Torino, Dept Mech & Aerosp Engn, Turin, Italy
[2] Imperial Coll London, Dept Aeronaut, London, England
[3] Brahmal Vasudevan Inst Sustainable Aviat, London, England
关键词
MULTI-FIDELITY OPTIMIZATION; STOCHASTIC COLLOCATION; GLOBAL OPTIMIZATION; QUERY; INFERENCE; REGRESSION; SUPPORT; SEARCH; DESIGN; OUTPUT;
D O I
10.1007/s11831-024-10064-z
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Science and Engineering applications are typically associated with expensive optimization problem to identify optimal design solutions and states of the system of interest. Bayesian optimization and active learning compute surrogate models through efficient adaptive sampling schemes to assist and accelerate this search task toward a given optimization goal. Both those methodologies are driven by specific infill/learning criteria which quantify the utility with respect to the set goal of evaluating the objective function for unknown combinations of optimization variables. While the two fields have seen an exponential growth in popularity in the past decades, their dualism and synergy have received relatively little attention to date. This paper discusses and formalizes the synergy between Bayesian optimization and active learning as symbiotic adaptive sampling methodologies driven by common principles. In particular, we demonstrate this unified perspective through the formalization of the analogy between the Bayesian infill criteria and active learning criteria as driving principles of both the goal-driven procedures. To support our original perspective, we propose a general classification of adaptive sampling techniques to highlight similarities and differences between the vast families of adaptive sampling, active learning, and Bayesian optimization. Accordingly, the synergy is demonstrated mapping the Bayesian infill criteria with the active learning criteria, and is formalized for searches informed by both a single information source and multiple levels of fidelity. In addition, we provide guidelines to apply those learning criteria investigating the performance of different Bayesian schemes for a variety of benchmark problems to highlight benefits and limitations over mathematical properties that characterize real-world applications.
引用
收藏
页码:2985 / 3013
页数:29
相关论文
共 158 条
[11]   Goal-oriented, model-constrained optimization for reduction of large-scale systems [J].
Bui-Thanh, T. ;
Willcox, K. ;
Ghattas, O. ;
Waanders, B. van Bloemen .
JOURNAL OF COMPUTATIONAL PHYSICS, 2007, 224 (02) :880-896
[12]  
Bujrbidge R, 2007, LECT NOTES COMPUT SC, V4881, P209
[13]   Batch Mode Active Learning for Regression With Expected Model Change [J].
Cai, Wenbin ;
Zhang, Muhan ;
Zhang, Ya .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (07) :1668-1681
[14]   Maximizing Expected Model Change for Active Learning in Regression [J].
Cai, Wenbin ;
Zhang, Ya ;
Zhou, Jun .
2013 IEEE 13TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2013, :51-60
[15]  
Catlett J., 1994, MACHINE LEARNING P 1, P148, DOI DOI 10.1016/B978-1-55860-335-6.50026-X
[16]   Batch Mode Active Sampling Based on Marginal Probability Distribution Matching [J].
Chattopadhyay, Rita ;
Wang, Zheng ;
Fan, Wei ;
Davidson, Ian ;
Panchanathan, Sethuraman ;
Ye, Jieping .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2013, 7 (03)
[17]  
Chaudhuri TR, 1995, 1995 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS PROCEEDINGS, VOLS 1-6, P1338, DOI 10.1109/ICNN.1995.487351
[18]   SEQUENTIAL DESIGN OF EXPERIMENTS [J].
CHERNOFF, H .
ANNALS OF MATHEMATICAL STATISTICS, 1959, 30 (03) :755-770
[19]  
Cohn D, 1993, ADV NEURAL INF PROCE, V6
[20]  
Dasgupta S., 2008, P 25 INT C MACH LEAR, P208