Abstract

Many real-world functions are defined over both categorical and category-specific continuous variables and thus cannot be optimized by traditional Bayesian optimization (BO) methods. To optimize such functions, we propose a new method that formulates the problem as a multi-armed bandit problem, wherein each category corresponds to an arm with its reward distribution centered around the optimum of the objective function in continuous variables. Our goal is to identify the best arm and the maximizer of the corresponding continuous function simultaneously. Our algorithm uses a Thompson sampling scheme that helps connecting both multi-arm bandit and BO in a unified framework. We extend our method to batch BO to allow parallel optimization when multiple resources are available. We theoretically analyze our method for convergence and prove sub-linear regret bounds. We perform a variety of experiments: optimization of several benchmark functions, hyper-parameter tuning of a neural network, and automatic selection of the best machine learning model along with its optimal hyper-parameters (a.k.a automated machine learning). Comparisons with other methods demonstrate the effectiveness of our proposed method.

Highlights

  • Bayesian optimization (BO) (Shahriari et al 2016) provides a powerful and efficient framework for global optimization of expensive black-box functions

  • We develop a BO algorithm that can handle both categorical and continuous variables even if each category involves a different set of continuous variables

  • We have introduced a novel BO method to globally optimize expensive black-box functions involving both categorical and continuous variables

Read more

Summary

Introduction

Bayesian optimization (BO) (Shahriari et al 2016) provides a powerful and efficient framework for global optimization of expensive black-box functions. After converting the categorical variables to one-hot encoding, this approach treats extra variables as continuous in [0, 1] and uses a typical BO algorithm to optimize them This type of encoding imposes equal measure of covariance between all category pairs, totally ignoring the fact that they may have different or no correlations at all. The recommendations can get repeated as they are generated via rounding-off at the end To address the latter problem, (Garrido-Merchan and Hernandez-Lobato 2018) assumed that the objective function does not change its values except at the designated points of 0 and 1. This is achieved by using a kernel function that computes covariances after the input is rounded off. This makes the resulting acquisition function step-wise, which is hard to optimize

Objectives
Methods
Findings
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.