Scalable Global Optimization via Local Bayesian Optimization David - - PowerPoint PPT Presentation

scalable global optimization via local bayesian
SMART_READER_LITE
LIVE PREVIEW

Scalable Global Optimization via Local Bayesian Optimization David - - PowerPoint PPT Presentation

Scalable Global Optimization via Local Bayesian Optimization David Eriksson Uber AI eriksson@uber.com Matthias Poloczek Michael Pearce Jake Gardner Ryan Turner Global Optimization Find such that


slide-1
SLIDE 1

Scalable Global Optimization via Local Bayesian Optimization

Jake Gardner Michael Pearce Matthias Poloczek David Eriksson Uber AI eriksson@uber.com Ryan Turner

slide-2
SLIDE 2

Global Optimization

Find 𝑦∗ ∈ Ω such that 𝑔 𝑦∗ ≤ 𝑔 𝑦 , ∀𝑦 ∈ Ω

  • 𝑔 is a continuous, computationally expensive, black-box function
  • Ω ⊂ ℝ+ is a hyper-rectangle

Design of aerodynamic structures Planning and control

slide-3
SLIDE 3

Bayesian Optimization (BO)

Common restrictions:

  • A few hundred

evaluations

  • Less than 10 tunable

parameters

True function Sample Observed points Next point

slide-4
SLIDE 4

Bayesian Optimization (BO)

Common restrictions:

  • A few hundred

evaluations

  • Less than 10 tunable

parameters

True function Sample Observed points Next point

slide-5
SLIDE 5

Bayesian Optimization (BO)

Common restrictions:

  • A few hundred

evaluations

  • Less than 10 tunable

parameters

True function Sample Observed points Next point

slide-6
SLIDE 6

High-dimensional BO is challenging

Challenges:

  • 1. The search space grows exponentially with dimensionality
  • 2. A global GP model may not fit the data everywhere
  • 3. Large areas of uncertainty leads to over-exploration

Previous work makes strong assumptions:

  • Additive structure
  • Low-dimensional structure
slide-7
SLIDE 7

Trust-region methods

Main idea:

  • Optimize a (simple)

model in a local region

  • Expand/shrink this region

based on progress

  • Only requires a locally

accurate model

Linear (e.g. COBYLA) Quadratic (e.g. BOBYQA) GP (TuRBO, this paper)

slide-8
SLIDE 8

Trust-region BO (TuRBO)

  • 1. Avoids over-exploration by using a trust-region framework
  • 2. Balances exploration/exploitation by using BO inside the trust-region
  • 3. Uses Thompson sampling to scale to large batch sizes

Trust Region Update

GP Model True Function

slide-9
SLIDE 9

Experimental results

Robot pushing: 10,000 evaluations, batch size 50 Rover trajectory planning: 20,000 evaluations, batch size 100

slide-10
SLIDE 10

Experimental results

200D Ackley function: 10,000 evaluations, batch size 100

2000 4000 6000 8000 10000 Number of evaluations 4 6 8 10 12 14 16 Value

TuRBO-1 Thompson BOCK Bohamiann HeSBO CMA-ES BOBYQA Nelder-Mead BFGS Random

slide-11
SLIDE 11

Summary

TuRBO:

  • Achieves excellent results for high-dimensional problems
  • Combines BO with trust-regions to avoid over-exploration
  • Makes no assumptions about low-dimensional structure

Paper: https://arxiv.org/abs/1910.01739 Code: https://github.com/uber-research/TuRBO Poster #9