Bosh: Bayesian optimization by sampling hierarchically

HB Moss, DS Leslie, P Rayson - arXiv preprint arXiv:2007.00939, 2020 - arxiv.org
arXiv preprint arXiv:2007.00939, 2020arxiv.org
Deployments of Bayesian Optimization (BO) for functions with stochastic evaluations, such
as parameter tuning via cross validation and simulation optimization, typically optimize an
average of a fixed set of noisy realizations of the objective function. However, disregarding
the true objective function in this manner finds a high-precision optimum of the wrong
function. To solve this problem, we propose Bayesian Optimization by Sampling
Hierarchically (BOSH), a novel BO routine pairing a hierarchical Gaussian process with an …
Deployments of Bayesian Optimization (BO) for functions with stochastic evaluations, such as parameter tuning via cross validation and simulation optimization, typically optimize an average of a fixed set of noisy realizations of the objective function. However, disregarding the true objective function in this manner finds a high-precision optimum of the wrong function. To solve this problem, we propose Bayesian Optimization by Sampling Hierarchically (BOSH), a novel BO routine pairing a hierarchical Gaussian process with an information-theoretic framework to generate a growing pool of realizations as the optimization progresses. We demonstrate that BOSH provides more efficient and higher-precision optimization than standard BO across synthetic benchmarks, simulation optimization, reinforcement learning and hyper-parameter tuning tasks.
arxiv.org