an experimental investigation of model based parameter
play

An Experimental Investigation of Model-Based Parameter Optimization: - PowerPoint PPT Presentation

An Experimental Investigation of Model-Based Parameter Optimization: SPO and Beyond Frank Hutter, Holger H. Hoos, Kevin Leyton-Brown, Kevin P. Murphy Department of Computer Science University of British Columbia Canada { hutter, hoos,


  1. An Experimental Investigation of Model-Based Parameter Optimization: SPO and Beyond Frank Hutter, Holger H. Hoos, Kevin Leyton-Brown, Kevin P. Murphy Department of Computer Science University of British Columbia Canada { hutter, hoos, kevinlb, murphyk } @cs.ubc.ca

  2. Motivation for Parameter Optimization Genetic Algorithms & Evolutionary Strategies are + Very flexible frameworks 2

  3. Motivation for Parameter Optimization Genetic Algorithms & Evolutionary Strategies are + Very flexible frameworks – Tedious to configure for a new domain ◮ Population size ◮ Mating scheme ◮ Mutation rate ◮ Search operators ◮ Hybridizations, ... 2

  4. Motivation for Parameter Optimization Genetic Algorithms & Evolutionary Strategies are + Very flexible frameworks – Tedious to configure for a new domain ◮ Population size ◮ Mating scheme ◮ Mutation rate ◮ Search operators ◮ Hybridizations, ... Automated parameter optimization can help ◮ High-dimensional optimization problem ◮ Automate � saves time & improves results 2

  5. Parameter Optimization Methods ◮ Numerical parameters – See Blackbox optimization workshop (this GECCO) – Algorithm parameters: CALIBRA [Adenso-Diaz & Laguna, ’06] 3

  6. Parameter Optimization Methods ◮ Numerical parameters – See Blackbox optimization workshop (this GECCO) – Algorithm parameters: CALIBRA [Adenso-Diaz & Laguna, ’06] ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] 3

  7. Parameter Optimization Methods ◮ Numerical parameters – See Blackbox optimization workshop (this GECCO) – Algorithm parameters: CALIBRA [Adenso-Diaz & Laguna, ’06] ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] ◮ Many categorical parameters – Genetic algorithms [Terashima-Mar´ ın, Ross & Valenzuela-R´ endon, ’99] 3

  8. Parameter Optimization Methods ◮ Numerical parameters – See Blackbox optimization workshop (this GECCO) – Algorithm parameters: CALIBRA [Adenso-Diaz & Laguna, ’06] ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] ◮ Many categorical parameters – Genetic algorithms [Terashima-Mar´ ın, Ross & Valenzuela-R´ endon, ’99] – Iterated Local Search [Hutter, Hoos, Leyton-Brown & St¨ utzle, ’07-’09] � Dozens of parameters ( e.g. , CPLEX with 63 parameters) � For many problems: SAT, MIP, time-tabling, protein folding, MPE, ... 3

  9. Parameter Optimization Methods Model-free Parameter Optimization ◮ Numerical parameters: see BBOB workshop (this GECCO) ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] ◮ Many categorical parameters [ e.g. , Terashima-Mar´ ın, Ross & Valenzuela-R´ endon, ’99, Hutter, Hoos, Leyton-Brown & St¨ utzle, ’07-’09] 4

  10. Parameter Optimization Methods Model-free Parameter Optimization ◮ Numerical parameters: see BBOB workshop (this GECCO) ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] ◮ Many categorical parameters [ e.g. , Terashima-Mar´ ın, Ross & Valenzuela-R´ endon, ’99, Hutter, Hoos, Leyton-Brown & St¨ utzle, ’07-’09] Model-based Parameter Optimization 4

  11. Parameter Optimization Methods Model-free Parameter Optimization ◮ Numerical parameters: see BBOB workshop (this GECCO) ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] ◮ Many categorical parameters [ e.g. , Terashima-Mar´ ın, Ross & Valenzuela-R´ endon, ’99, Hutter, Hoos, Leyton-Brown & St¨ utzle, ’07-’09] Model-based Parameter Optimization ◮ Methods – Fractional factorial designs [ e.g. , Ridge & Kudenko, ’07] – Sequential Parameter Optimization (SPO) [Bartz-Beielstein, Preuss, Lasarczyk, ’05-’09] 4

  12. Parameter Optimization Methods Model-free Parameter Optimization ◮ Numerical parameters: see BBOB workshop (this GECCO) ◮ Few categorical parameters: racing algorithms [Birattari, St¨ utzle, Paquete & Varrentrapp, ’02] ◮ Many categorical parameters [ e.g. , Terashima-Mar´ ın, Ross & Valenzuela-R´ endon, ’99, Hutter, Hoos, Leyton-Brown & St¨ utzle, ’07-’09] Model-based Parameter Optimization ◮ Methods – Fractional factorial designs [ e.g. , Ridge & Kudenko, ’07] – Sequential Parameter Optimization (SPO) [Bartz-Beielstein, Preuss, Lasarczyk, ’05-’09] ◮ Can use model for more than optimization – Importance of each parameter – Interaction between parameters 4

  13. Outline 1. Sequential Model-Based Optimization (SMBO): Introduction 2. Comparing Two SMBO Methods: SPO vs SKO 3. Components of SPO: Model Quality 4. Components of SPO: Sequential Experimental Design 5. Conclusions and Future Work 5

  14. Outline 1. Sequential Model-Based Optimization (SMBO): Introduction 2. Comparing Two SMBO Methods: SPO vs SKO 3. Components of SPO: Model Quality 4. Components of SPO: Sequential Experimental Design 5. Conclusions and Future Work 6

  15. SMBO: Introduction 30 . . 25 True function . . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x First step of SMBO 7

  16. SMBO: Introduction 1. Get response values at initial design points 30 . 25 True function Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x First step of SMBO 7

  17. SMBO: Introduction 1. Get response values at initial design points 30 . . 25 . Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x First step of SMBO 7

  18. SMBO: Introduction 1. Get response values at initial design points 2. Fit a model to the data 30 DACE mean prediction DACE mean +/− 2*stddev 25 . Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x First step of SMBO 7

  19. SMBO: Introduction 1. Get response values at initial design points 2. Fit a model to the data 3. Use model to pick most promising next design point (based on expected improvement criterion) 30 DACE mean prediction DACE mean +/− 2*stddev 25 . Function evaluations EI (scaled) 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x First step of SMBO 7

  20. SMBO: Introduction 1. Get response values at initial design points 2. Fit a model to the data 3. Use model to pick most promising next design point (based on expected improvement criterion) 30 DACE mean prediction DACE mean +/− 2*stddev 25 True function Function evaluations EI (scaled) 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x First step of SMBO 7

  21. SMBO: Introduction 1. Get response values at initial design points 2. Fit a model to the data 3. Use model to pick most promising next design point (based on expected improvement criterion) 4. Repeat 2. and 3. until time is up 30 30 DACE mean prediction DACE mean prediction DACE mean +/− 2*stddev DACE mean +/− 2*stddev 25 True function 25 True function Function evaluations Function evaluations EI (scaled) EI (scaled) 20 20 response y response y 15 15 10 10 5 5 0 0 −5 −5 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 parameter x parameter x First step of SMBO Second step of SMBO 7

  22. Outline 1. Sequential Model-Based Optimization (SMBO): Introduction 2. Comparing Two SMBO Methods: SPO vs SKO 3. Components of SPO: Model Quality 4. Components of SPO: Sequential Experimental Design 5. Conclusions and Future Work 8

  23. Dealing with Noise: SKO vs SPO ◮ Method I (used in SKO) [Huang, Allen, Notz & Zeng, ’06.] – Fit standard GP assuming Gaussian observation noise – Can only fit the mean of the responses 30 GP mean prediction GP mean +/− 2*stddev 25 True function Function evaluations . 20 response y 15 10 5 0 −5 0 0.2 0.4 0.6 0.8 1 parameter x Method I: noisy fit of original response 9

  24. Dealing with Noise: SKO vs SPO ◮ Method I (used in SKO) [Huang, Allen, Notz & Zeng, ’06.] – Fit standard GP assuming Gaussian observation noise – Can only fit the mean of the responses ◮ Method II (used in SPO) [Bartz-Beielstein, Preuss, Lasarczyk, ’05-’09] – Compute statistic of empirical distribution of responses at each design point – Fit noise-free GP to that 30 30 GP mean prediction DACE mean prediction GP mean +/− 2*stddev DACE mean +/− 2*stddev 25 True function 25 True function Function evaluations Function evaluations . . 20 20 response y response y 15 15 10 10 5 5 0 0 −5 −5 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 parameter x parameter x Method I: noisy fit of original response Method II: noise-free fit of cost statistic 9

  25. Experiment: SPO vs SKO for Tuning CMA-ES ◮ CMA-ES [Hansen et al., ’95-’09] – Evolutionary strategy for global optimization – State-of-the-art (see BBOB workshop this GECCO) – Parameters: population size, number of parents, learning rate, damping parameter 10

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend