5 Exploring the search space with bandit-style policies
This chapter covers
- The multi-armed bandit problem and how it’s related to BayesOpt
- The Upper Confidence Bound policy in BayesOpt
- The Thompson sampling policy in BayesOpt
Which slot machine should you play at a casino to maximize your winnings? How can you develop a strategy to intelligently try out multiple slot machines and narrow down the most profitable machine? What does this problem have to do with BayesOpt? These are the questions this chapter will help us answer.
Chapter 4 was our introduction to BayesOpt policies, which decide how the search space should be explored and inspected. The exploration strategy of a BayesOpt policy should guide us toward the optimum of the objective ...
Get Bayesian Optimization in Action now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.