model based evolutionary algorithms part 1 estimation of
play

Model-Based Evolutionary Algorithms Part 1: Estimation of - PowerPoint PPT Presentation

Model-Based Evolutionary Algorithms Part 1: Estimation of Distribution Algorithms Dirk Thierens Universiteit Utrecht The Netherlands 1/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 1 / 43 What ? Evolutionary


  1. Model-Based Evolutionary Algorithms Part 1: Estimation of Distribution Algorithms Dirk Thierens Universiteit Utrecht The Netherlands 1/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 1 / 43

  2. What ? Evolutionary Algorithms Population-based, stochastic search algorithms Exploitation: selection Exploration: mutation & crossover Model-Based Evolutionary Algorithms Population-based, stochastic search algorithms Exploitation: selection Exploration: Learn a model from selected solutions 1 Generate new solutions from the model (& population) 2 2/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 2 / 43

  3. What ? Probabilistic Model-Based Evolutionary Algorithms (MBEA) a.k.a. Estimation of Distribution Algorithms (EDAs) a.k.a. Probabilistic Model-Building Genetic Algorithms a.k.a. Iterated Density Estimation Evolutionary Algorithms MBEA = Evolutionary Computing + Machine Learning Note: model not necessarily probabilistic 3/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 3 / 43

  4. Why ? Goal: Black Box Optimization Little known about the structure of the problem Clean separation optimizer from problem definition Easy and generally applicable Approach * Classical EAs: need suitable representation & variation operators * Model-Based EAs: learn structure from good solutions 4/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 4 / 43

  5. Discrete Representation Typically binary representation Higher order cardinality: similar approach 5/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 5 / 43

  6. Probabilistic Model-Building Genetic Algorithm Type of Models Univariate: no statistical interaction between variables considered. Bivariate: pairwise dependencies learned. Multivariate: higher-order interactions modeled. 6/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 6 / 43

  7. Univariate PMBGA Model * Model: probability vector [ p 1 , . . . , p ℓ ] ( ℓ : string length) * p i : probability of value 1 at string position i * p ( X ) = � ℓ i = 1 p ( x i ) ( p ( x i ) : univariate marginal distribution) Learn model: count proportions of 1 in selected population Sample model: generate new solutions with specified probabilities 7/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 7 / 43

  8. Univariate PMBGA Different Variants PBIL (Baluja; 1995) ◮ Prob. vector incrementally updated over successive generations UMDA (M¨ uhlenbein, Paass; 1996) ◮ No incremental updates: example above Compact GA (Harik, Lobo, Goldberg; 1998) ◮ Models steady-state GA with tournament selection DEUM (Shakya, McCall, Brown; 2004) ◮ Uses Markov Random Field modeling 8/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 8 / 43

  9. A hard problem for the univariate model Data Marginal Product model 000000 ˆ ˆ P ( X 0 X 1 X 2 ) P ( X 3 X 4 X 5 ) 111111 000 0.3 0.3 010101 001 0.0 0.0 101010 010 0.2 0.2 000010 011 0.0 0.0 111000 100 0.0 0.0 010111 101 0.1 0.1 111000 110 0.0 0.0 000111 111 0.4 0.4 111111 Univariate model ˆ P ( X 0 ) ˆ P ( X 1 ) ˆ P ( X 2 ) ˆ P ( X 3 ) ˆ P ( X 4 ) ˆ P ( X 5 ) 0 0.5 0.4 0.5 0.5 0.4 0.5 1 0.5 0.6 0.5 0.5 0.6 0.5 What is the probability of generating 111111? Univariate model: 0 . 5 · 0 . 6 · 0 . 5 · 0 . 5 · 0 . 6 · 0 . 5 = 0 . 0225 MP model: 0 . 4 · 0 . 4 = 0 . 16 (7 times larger!) 9/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 9 / 43

  10. Learning problem structure on the fly Without a “good” decomposition of the problem, important partial solutions (building blocks) are likely to get disrupted in variation. Disruption leads to inefficiency. Can we automatically configure the model structure favorably? Selection increases proportion of good building blocks and thus “correlations” between variables of these building blocks. So, learn which variables are “correlated”. See the population (or selection) as a data set. Apply statistics / probability theory / probabilistic modeling. 10/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 10 / 43

  11. Bivariate PMBGA Model Need more than just probabilities of bit values Model pairwise interactions: conditional probabilities MIMIC (de Bonet, Isbell, Viola; 1996) ◮ Dependency Chain COMIT (Baluja, Davies; 1997) ◮ Dependency Tree BMDA (Pelikan , M¨ uhlenbein; 1998) ◮ Independent trees (forest) 11/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 11 / 43

  12. Entropy Random variable X with probability distribution function p ( X ) Entropy H ( X ) is a measure of uncertainty about a random variable X: � H ( X ) = − p ( x ) log p ( x ) x ∈ X Conditional entropy H ( Y | X ) is a measure of uncertainty remaining about Y after X is known (what X does not say about Y): p ( x , y ) log p ( x ) � � H ( Y | X ) = p ( x , y ) x ∈ X y ∈ Y 12/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 12 / 43

  13. Mutual information The mutual information I ( X , Y ) of two random variables is a measure of the variables’ mutual dependence. Mutual information is more general than the correlation coefficient (= linear relation between real-valued variables) Mutual information determines how similar the joint distribution p ( X , Y ) is to the products of factored marginal distribution p ( X ) p ( Y ) : p ( x , y ) log p ( x , y ) � � I ( X , Y ) = p ( x ) p ( y ) x ∈ X y ∈ Y 13/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 13 / 43

  14. Mutual information and entropy Mutual information in relation to entropy: I ( X , Y ) = H ( Y ) − H ( Y | X ) = H ( X ) − H ( X | Y ) = H ( X ) + H ( Y ) − H ( X , Y ) Mutual information can thus be seen as the amount of uncertainty in Y, minus the amount of uncertainty in Y which remains after X is known, which is equivalent to the amount of uncertainty in Y which is removed by knowing X Mutual information is the amount of information (that is, reduction in uncertainty) that knowing either variable provides about the other 14/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 14 / 43

  15. Bivariate PMBGA MIMIC Model: chain of pairwise dependencies. p ( X ) = � ℓ − 1 i = 1 p ( x i + 1 | x i ) p ( x 1 ) . MIMIC greedily searches for the optimal permutation of variables that minimizes Kullack-Leibler divergence. 15/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 15 / 43

  16. Bivariate PMBGA MIMIC Joint probability distribution over a set of random variables, X = X i is: p ( X ) = p ( X 1 | X 2 ... X n ) p ( X 2 | X 3 ... X n ) ... p ( X n − 1 | X n ) p ( X n ) Given only pairwise conditional probabilities, p ( X i | X j ) and unconditional probabilities, p ( X i ) , we want to approximate the true joint distribution as close as possible Given a permutation of numbers between 1 and n : π = i 1 i 2 ... i n define a class of probability distributions p π ( X ) : p π ( X ) = p ( X i 1 | X i 2 ) p ( X i 2 | X i 3 ) ... p ( X i n − 1 | X i n ) p ( X i n ) 16/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 16 / 43

  17. Bivariate PMBGA MIMIC Goal is to find a permutation π that maximizes the agreement between p π ( X ) and the true joint distribution p ( X ) Agreement between distributions can be measured by the Kullback-Leibler divergence: p ( x ) log p ( x ) � D ( p ( X ) || p π ( X )) = p π ( x ) x ∈ X = − H ( p ) + H ( X i 1 | X i 2 ) + ... + H ( X i n − 1 | X i n ) + H ( X i n ) The optimal permutation π minimizes the sum of the conditional entropies: H ( X i 1 | X i 2 ) + ... + H ( X i n − 1 | X i n ) + H ( X i n ) 17/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 17 / 43

  18. Bivariate PMBGA MIMIC: algorithm i n = argmin j H ( X j ) 1 i k = argmin t H ( X t | X i k + 1 ) , where 2 t � = i k + 1 ... i n and k = n − 1 , n − 2 , ..., 2 , 1 Generating samples from the distribution: Choose a value for X i n based on the probability p ( X i n ) 1 for k = n − 1 , n − 2 , ..., 2 , 1, choose an element X i k based on the 2 conditional probability p ( X i k | X i k + 1 ) Both algorithms run in O ( n 2 ) 18/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 18 / 43

  19. Bivariate PMBGA COMIT Optimal dependency tree instead of linear chain. Compute fully connected weighted graph between problem variables. Weights are the mutual information I ( X , Y ) between the variables. p ( x , y ) I ( X , Y ) = � � x ∈ X p ( x , y ) log p ( x ) p ( y ) . y ∈ Y COMIT computes the maximum spanning tree of the weighted graph. 19/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 19 / 43

  20. Bivariate PMBGA COMIT The approximating probability model is restricted to factorizations in which the conditional probability distribution for any random variable depends on the value of at most one other random variable: n � p ( X ) = p ( X i | X parent ( i ) ) i = 1 p ( X ) is the class of distributions with a tree as graphical model 20/ ?? Dirk Thierens (Universiteit Utrecht) Model-Based Evolutionary Algorithms 20 / 43

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend