time bounded sequential parameter optimization
play

Time-Bounded Sequential Parameter Optimization Frank Hutter, Holger - PowerPoint PPT Presentation

Time-Bounded Sequential Parameter Optimization Frank Hutter, Holger H. Hoos, Kevin Leyton-Brown, Kevin P. Murphy Department of Computer Science University of British Columbia Canada { hutter, hoos, kevinlb, murphyk } @cs.ubc.ca Automated


  1. Time-Bounded Sequential Parameter Optimization Frank Hutter, Holger H. Hoos, Kevin Leyton-Brown, Kevin P. Murphy Department of Computer Science University of British Columbia Canada { hutter, hoos, kevinlb, murphyk } @cs.ubc.ca

  2. Automated Parameter Optimization Most algorithms have parameters ◮ Decisions that are left open during algorithm design ◮ Instantiate to optimize empirical performance 2

  3. Automated Parameter Optimization Most algorithms have parameters ◮ Decisions that are left open during algorithm design ◮ Instantiate to optimize empirical performance ◮ E.g. local search – neighbourhoods, restarts, types of perturbations, tabu length (or range for it), etc 2

  4. Automated Parameter Optimization Most algorithms have parameters ◮ Decisions that are left open during algorithm design ◮ Instantiate to optimize empirical performance ◮ E.g. local search – neighbourhoods, restarts, types of perturbations, tabu length (or range for it), etc ◮ E.g., tree search – Branching heuristics, no-good learning, restarts, pre-processing, etc 2

  5. Automated Parameter Optimization Most algorithms have parameters ◮ Decisions that are left open during algorithm design ◮ Instantiate to optimize empirical performance ◮ E.g. local search – neighbourhoods, restarts, types of perturbations, tabu length (or range for it), etc ◮ E.g., tree search – Branching heuristics, no-good learning, restarts, pre-processing, etc Automatically find good instantiation of parameters ◮ Eliminate most tedious part of algorithm design and end use ◮ Save development time & improve performance 2

  6. Parameter Optimization Methods ◮ Lots of work on numerical parameters, e.g. – CALIBRA [Adenso-Diaz & Laguna, ’06] – Population-based, e.g. CMA-ES [Hansen et al, ’95-present] 3

  7. Parameter Optimization Methods ◮ Lots of work on numerical parameters, e.g. – CALIBRA [Adenso-Diaz & Laguna, ’06] – Population-based, e.g. CMA-ES [Hansen et al, ’95-present] ◮ Categorical parameters – Racing algorithms, F-Race [Birattari et al., ’02-present] – Iterated Local Search, ParamILS [Hutter et al., AAAI ’07 & JAIR’09] 3

  8. Parameter Optimization Methods ◮ Lots of work on numerical parameters, e.g. – CALIBRA [Adenso-Diaz & Laguna, ’06] – Population-based, e.g. CMA-ES [Hansen et al, ’95-present] ◮ Categorical parameters – Racing algorithms, F-Race [Birattari et al., ’02-present] – Iterated Local Search, ParamILS [Hutter et al., AAAI ’07 & JAIR’09] ◮ Success of parameter optimization – Many parameters ( e.g. , CPLEX with 63 parameters) – Large speedups (sometimes orders of magnitude!) – For many problems: SAT, MIP, time-tabling, protein folding, ... 3

  9. Limitations of Model-Free Parameter Optimization Model-free methods only return the best parameter setting ◮ Often that is all you need – E.g.: end user can customize algorithm 4

  10. Limitations of Model-Free Parameter Optimization Model-free methods only return the best parameter setting ◮ Often that is all you need – E.g.: end user can customize algorithm ◮ But sometimes we would like to know more – How important is each of the parameters? – Which parameters interact? – For which types of instances is a parameter setting good? � Inform algorithm designer 4

  11. Limitations of Model-Free Parameter Optimization Model-free methods only return the best parameter setting ◮ Often that is all you need – E.g.: end user can customize algorithm ◮ But sometimes we would like to know more – How important is each of the parameters? – Which parameters interact? – For which types of instances is a parameter setting good? � Inform algorithm designer Response surface models can help ◮ Predictive models of algorithm performance with given parameter settings 4

  12. Sequential Parameter Optimization (SPO) ◮ Original SPO [Bartz-Beielstein et al., ’05-present] ◮ SPO toolbox ◮ Set of interactive tools for parameter optimization 5

  13. Sequential Parameter Optimization (SPO) ◮ Original SPO [Bartz-Beielstein et al., ’05-present] ◮ SPO toolbox ◮ Set of interactive tools for parameter optimization ◮ Studied SPO components [Hutter et al, GECCO-09] ◮ Want completely automated tool � More robust version: SPO + 5

  14. Sequential Parameter Optimization (SPO) ◮ Original SPO [Bartz-Beielstein et al., ’05-present] ◮ SPO toolbox ◮ Set of interactive tools for parameter optimization ◮ Studied SPO components [Hutter et al, GECCO-09] ◮ Want completely automated tool � More robust version: SPO + ◮ This work: TB-SPO, reduce computational overheads 5

  15. Sequential Parameter Optimization (SPO) ◮ Original SPO [Bartz-Beielstein et al., ’05-present] ◮ SPO toolbox ◮ Set of interactive tools for parameter optimization ◮ Studied SPO components [Hutter et al, GECCO-09] ◮ Want completely automated tool � More robust version: SPO + ◮ This work: TB-SPO, reduce computational overheads ◮ Ongoing work: extend TB-SPO to handle – Categorical parameters – Multiple benchmark instances 5

  16. Sequential Parameter Optimization (SPO) ◮ Original SPO [Bartz-Beielstein et al., ’05-present] ◮ SPO toolbox ◮ Set of interactive tools for parameter optimization ◮ Studied SPO components [Hutter et al, GECCO-09] ◮ Want completely automated tool � More robust version: SPO + ◮ This work: TB-SPO, reduce computational overheads ◮ Ongoing work: extend TB-SPO to handle – Categorical parameters – Multiple benchmark instances – Very promising results for both 5

  17. Outline 1. Sequential Model-Based Optimization 2. Reducing the Computational Overhead Due To Models 3. Conclusions 6

  18. Outline 1. Sequential Model-Based Optimization 2. Reducing the Computational Overhead Due To Models 3. Conclusions 7

  19. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 30 . . 25 True function . . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x 8

  20. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 0. Run algorithm with initial parameter settings 30 . 25 True function Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x 8

  21. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 0. Run algorithm with initial parameter settings 30 . . 25 . Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x 8

  22. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 0. Run algorithm with initial parameter settings 1. Fit a model to the data 30 DACE mean prediction DACE mean +/− 2*stddev 25 . Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x 8

  23. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 0. Run algorithm with initial parameter settings 1. Fit a model to the data 2. Use model to pick promising parameter setting 30 DACE mean prediction DACE mean +/− 2*stddev 25 . Function evaluations EI (scaled) 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x 8

  24. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 0. Run algorithm with initial parameter settings 1. Fit a model to the data 2. Use model to pick promising parameter setting 3. Perform an algorithm run with that parameter setting 30 DACE mean prediction DACE mean +/− 2*stddev 25 True function Function evaluations EI (scaled) 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x 8

  25. Sequential Model-Based Optimization (SMBO) Blackbox function optimization; function = algo. performance 0. Run algorithm with initial parameter settings 1. Fit a model to the data 2. Use model to pick promising parameter setting 3. Perform an algorithm run with that parameter setting ◮ Repeat 1-3 until time is up 30 30 DACE mean prediction DACE mean prediction DACE mean +/− 2*stddev DACE mean +/− 2*stddev 25 True function 25 True function Function evaluations Function evaluations EI (scaled) EI (scaled) 20 20 response y response y 15 15 10 10 5 5 0 0 −5 −5 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 parameter x parameter x First step Second step 8

  26. Computational Overhead due to Models: Example Example times 0. Run algorithm with initial parameter settings 1. Fit a model to the data 2. Use model to pick promising parameter setting 3. Perform an algorithm run with that parameter setting ◮ Repeat 1-3 until time is up 30 30 DACE mean prediction DACE mean prediction DACE mean +/− 2*stddev DACE mean +/− 2*stddev 25 True function 25 True function Function evaluations Function evaluations EI (scaled) EI (scaled) 20 20 response y response y 15 15 10 10 5 5 0 0 −5 −5 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 parameter x parameter x First step Second step 9

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend