automatic algorithm configuration
play

Automatic Algorithm Configuration Design choices and parameters - PDF document

HEURISTIC OPTIMIZATION Automatic Algorithm Configuration Design choices and parameters everywhere Todays high-performance optimizers involve a large number of design choices and parameter settings I exact solvers I design choices include


  1. HEURISTIC OPTIMIZATION Automatic Algorithm Configuration Design choices and parameters everywhere Todays high-performance optimizers involve a large number of design choices and parameter settings I exact solvers I design choices include alternative models, pre-processing, variable selection, value selection, branching rules . . . I many design choices have associated numerical parameters I example: CPLEX 10.1.1 has 159 user-specifiable parameters, about 80 influence the solver’s search mechanism I approximate algorithms I design choices include solution representation, operators, neighborhood, pre-processing, strategies, . . . I many design choices have associated numerical parameters I example: multi-objective ACO algorithms with 22 parameters (plus several still hidden ones): see later Heuristic Optimization 2011 2

  2. Example: Ant Colony Optimization Heuristic Optimization 2011 3 ��������� ����������� �������������� ��������� ������������� ������������� ����������� ����������������� ���������������� Heuristic Optimization 2011 4

  3. Probabilistic solution construction g j ? ! ij " ij , i k Heuristic Optimization 2011 5 ACO design choices and numerical parameters I solution construction I choice of pheromone model I choice of heuristic information I choice of constructive procedure I numerical parameters I α , β influence the weight of pheromone and heuristic information, respectively I q 0 determines greediness of construction procedure I m , the number of ants I pheromone update I which ants deposit pheromone and how much? I numerical parameters I ρ : evaporation rate I τ 0 : initial pheromone level I local search I . . . many more . . . Heuristic Optimization 2011 6

  4. Designing optimization algorithms Challenges I many alternative design choices I nonlinear interactions among algorithm components and/or parameters I performance assessment is di ffi cult Traditional design approach I trial–and–error design guided by expertise/intuition prone to over-generalizations, implicit independence assumptions, limited exploration of design alternatives Can we make this approach more principled and automatic? Heuristic Optimization 2011 7 Towards automatic algorithm configuration Automated algorithm configuration I apply powerful search techniques to design algorithms I use computation power to explore algorithm design spaces I free human creativity for higher level tasks Heuristic Optimization 2011 8

  5. Why automatic algorithm configuration? I improvement over manual, ad-hoc methods for tuning I reduction of development time and human intervention I increase number of considerable degrees of freedom I empirical studies, comparisons of algorithms I support for end users of algorithms Heuristic Optimization 2011 9 Configuration is a stochastic optimization/learning problem Random influences I stochasticity of the parameterized algorithm I stochasticity due to the “sampling” of the instance to be tackled Learning aspects I algorithm configuration should solve unseen instances Configuration problem is a stochastic mixed discrete–continuous optimization problem with machine learning aspects Heuristic Optimization 2011 10

  6. Main configuration approaches O ffl ine configuration I configure algorithm before deploying it I configuration done on training instances Online configuration I adapt parameter setting while solving an instance I typically limited to a set of known crucial algorithm parameters Heuristic Optimization 2011 11 O ffl ine configuration Remark: Configuration scenario requires the definition of performance measure to be optimized I maximize solution quality (within given computation time) I minimize computation time (to reach optimal solution) Heuristic Optimization 2011 12

  7. Towards a shift of paradigm in algorithm design ��������� ������������� ������������� ��������������� ����������������� Heuristic Optimization 2011 13 Towards a shift of paradigm in algorithm design ��������� ������������� ������������� ��������������� ����������������� Heuristic Optimization 2011 14

  8. Towards a shift of paradigm in algorithm design ��������� ������������� ������������� ��������������� ����������������� ��������� ������� ���������� Heuristic Optimization 2011 15 Approaches to configuration and tuning I numerical optimization techniques I e.g. CMA-ES [Hansen & Ostermeier, 2001], MADS [Audet & Orban, 2006] I heuristic search methods I e.g. ParamILS [Hutter, Hoos, Leyton-Brown, St¨ utzle, 2009], genetic programming [Fukunaga, 2002], gender-based GA [Sellman et al, 2010], . . . I experimental design, ANOVA I e.g. CALIBRA [Adenso-Diaz & Laguna, 2006], [Ridge, Kudenko, 2007, Ruiz Maroto, 2006, Coy et al., 2000] I response surface methods (model-based optimization) I e.g. SPO [Bartz-Beielstein, 2006], SMAC [Hutter, Hoos, Leyton-Brown, 2011] I sequential statistical testing, F-race, iterated F-race I e.g. [Birratari, St¨ utzle, Paquete, Varrentrap, 2002;Balaprakash, Birattari, St¨ utzle, 2007] Heuristic Optimization 2011 16

  9. Example of application scenario I Mario collects phone orders for 30 minutes I scheduling deliveries is an optimization problem I a di ff erent instance arises every 30 minutes I limited amount of time for scheduling, say one minute I good news: Mario has an SLS algorithm! I . . . but the SLS algorithm must be tuned I You have a limited amount of time for configuring it, say one week Criterion: Good configurations find good solutions for future instances! Heuristic Optimization 2011 17 Brute-force approach to configuration 1. sample a set of configurations Θ 0 ⊆ Θ 2. estimate C ( θ ) for each θ ∈ Θ 0 3. return the configuration with the lowest estimate Disadvantages of brute-force configuration 1. one needs to determine a priori how many runs on each candidate configuration 2. poor performing candidate configurations are evaluated with same computational e ff ort as good ones Heuristic Optimization 2011 18

  10. Remark : Estimation of expected cost Given I n runs for estimating expected cost of configuration θ I a large number of instances Question I how many runs on how many instances to minimize variance of estimate Answer I one run on each of n instances (Birattari, 2004) Heuristic Optimization 2011 19 The racing approach Θ I start with a set of initial candidates I consider a stream of instances I sequentially evaluate candidates I discard inferior candidates as su ffi cient evidence is gathered against them I . . . repeat until a winner is selected or until computation time expires i Heuristic Optimization 2011 20

  11. The F-Race algorithm Statistical testing 1. family-wise tests for di ff erences among configurations I Friedman two-way analysis of variance by ranks 2. if Friedman rejects H 0 , perform pairwise comparisons to best configuration I apply Friedman post-hoc test Predecessors I racing algorithms in model-selection Maron & Moore (1994) Heuristic Optimization 2011 21 Sampling configurations F-race is a method for the selection of the best configuration and independent of the way the set of configurations is sampled Sampling configurations and F-race I full factorial design I random sampling design I iterative refinement of a sampling model (iterative F-race) Heuristic Optimization 2011 22

  12. Full factorial design I full factorial design (FFD) was used in the first applications of F-race to make comparisons to other ways of doing races I levels for FFD can be determined manually, randomly, etc. I FFD has significant disadvantages I expertise for selecting the levels of each parameter I exponential growth with number d of parameters: l d Heuristic Optimization 2011 23 Random sampling design I Define a probability measure P X on the space X of parameter values I Sample the configurations I Apply F-Race to select the best I Performance attributed to the number of samples I advantages I arbitrary number of candidate configurations is sampled I no a priori definition of levels necessary I covers uniformly the parameter space Heuristic Optimization 2011 24

  13. Iterative re-finement I modify the probability measure: I using previously seen promising configurations to favor the search towards promising regions I sample from the newly defined distribution I apply F-race I iterate through this process Heuristic Optimization 2011 25 I sample configurations from initial distribution While not terminate() 1. apply F-Race 2. modify the distribution 3. sample configurations with selection probability Heuristic Optimization 2011 26

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend