l ecture 16 s warm i ntelligence 2 p article s warm o
play

L ECTURE 16: S WARM I NTELLIGENCE 2 / P ARTICLE S WARM O PTIMIZATION - PowerPoint PPT Presentation

15-382 C OLLECTIVE I NTELLIGENCE - S18 L ECTURE 16: S WARM I NTELLIGENCE 2 / P ARTICLE S WARM O PTIMIZATION 2 I NSTRUCTOR : G IANNI A. D I C ARO BACKGROUND: REYNOLDS BOIDS Reynolds created a model of coordinated animal motion in which the


  1. 15-382 C OLLECTIVE I NTELLIGENCE - S18 L ECTURE 16: S WARM I NTELLIGENCE 2 / P ARTICLE S WARM O PTIMIZATION 2 I NSTRUCTOR : G IANNI A. D I C ARO

  2. BACKGROUND: REYNOLDS’ BOIDS Reynolds created a model of coordinated animal motion in which the agents (boids) obeyed three simple local rules: Separation : steer to Alignment : steer towards Cohesion: steer to move avoid crowding local the average heading of toward the average position flockmates local flockmates of local flockmates Field of view Reynolds, C.W.: Flocks, herds and schools: a distributed https://www.youtube.com/watch?v=QbUPfMXXQIY behavioral model. Computer Graphics, 21(4), p.25-34, 1987 Play with Boids, PSO, CA, and … with NetLogo https://ccl.northwestern.edu/netlogo/ 2

  3. BOIDS + ROOSTING BEHAVIOR Kennedy and Eberhart included a roost, or, more generally, an attraction point (e.g, a prey ) in a simplified Boids-like simulation, such that each agent: • is attracted to the location of the roost, • remembers where it was closer to the roost, • shares information with its neighbors about its closest location to the roost Eventually, (almost) all agents land on the roost What if: • roost = ( unknown) extremum of a function • distance to the roost = quality of current agent position on the optimization landscape 3

  4. PARTICLE SWARM OPTIMIZATION (PSO) • PSO consists of a swarm of bird-like particles ➔ Multi-agent system • At each discrete time step, each particle is found at a position in the search space: it encodes a solution point 𝒚 ∈ 𝛹 𝑜 ⊆ ℝ 𝑜 for an optimization problem with objective function f( 𝒚 ) : 𝛹 𝑜 ⟼ ℝ m (black-box or white-box problem) • The fitness of each particle represents the quality of its position on the optimization landscape ( note: this can be virtually anything , think about robots looking for energy) • Particles move over the search space with a certain velocity • Each particle has: I nternal state + ( Neighborhood ⟷ Network of social connections) • At each time step, the velocity (both direction and speed ) of each particle is influenced + random, by: • pbest: its own best position found so far • lbest: the best solution that was found so far by the teammates in its social neighbor, and/or • gbest: the global best solution so far • “ Eventually” the swarm will converge to optimal positions { ~ x pbest , N ( p ) } x, ~ v, ~ 4

  5. NEIGHBORHOODS Geographical Social Global 5

  6. VECTOR COMBINATION OF MULTIPLE BIASES ~ vt r 2 · ( ~ xt ) xlbest − ~ r 1 · ( ~ xt ) xpbest − ~ xt +1 ~ !~ vt ~ xlbest xlbest − ~ ~ xt xt ~ ~ xpbest − ~ xt ~ xpbest 6

  7. PARTICLE SWARM OPTIMIZATION (PSO) element-wise multiplication operator r 1 = U (0 , � 1 ) r 2 = U (0 , � 2 ) ~ ~ ɸ are acceleration coefficients determining scale of forces in the direction of individual and social biases 7

  8. VECTOR COMBINATION OF MULTIPLE BIASES • Makes the particle move in the same direction and with the same velocity 1. Inertia • Improves the individual • Makes the particle return to a previous 2. Personal Influence position, better than the current Conservative • 3. Social Influence • Makes the particle follow the best neighbors direction Search for new Exploits what solutions good so far 8

  9. PSO AT WORK (MAX OPTIMIZATION PROBLEM) Example slides from Pinto et al. 9

  10. PSO AT WORK 10

  11. PSO AT WORK 11

  12. PSO AT WORK 12

  13. PSO AT WORK 13

  14. PSO AT WORK 14

  15. PSO AT WORK 15

  16. PSO AT WORK 16

  17. PSO AT WORK 17

  18. PSO AT WORK 18

  19. GOOD AND BAD POINTS OF BASIC PSO • Advantages • Quite insensitive to scaling of design variables • Simple implementation • Easily parallelized for concurrent processing • Derivative free / Black-box optimization • Very few algorithm parameters • Very efficient global search algorithm • Disadvantages • Tendency to a fast and premature convergence in mid optimum points • Slow convergence in refined search stage (weak local search ability) 19

  20. GOOD NEIGHBORHOOD TOPOLOGY? 20

  21. GOOD NEIGHBORHOOD TOPOLOGY? • Also considered were: • Clustering topologies (islands) • Dynamic topologies • … • No clear way of saying which topology is the best • Exploration / exploitation dilemma • Some neighborhood topologies are better for local search others for global search • lbest neighborhood topologies seems better for global search, • gbest topologies seem better for local search 21

  22. ACCELERATION COEFFICIENTS • The boxes show the distribution of the random vectors of the attracting forces of the local best and global best • The acceleration coefficients determine the scale distribution of the random individual / cognitive component vector 𝜚 1 and the social component vector 𝜚 2 22

  23. ACCELERATION COEFFICIENTS • ɸ 1 >0, ɸ 2 =0 particles are independent hill-climbers • ɸ 1 =0, ɸ 2 >0 swarm is one stochastic hill-climber • ɸ 1 = ɸ 2 >0 particles are attracted to the average of p i and g i • ɸ 2 > ɸ 1 more beneficial for unimodal problems • ɸ 1 > ɸ 2 more beneficial for multimodal problems • low ɸ 1 , ɸ 2 smooth particle trajectories • high ɸ 1 , ɸ 2 more acceleration, abrupt movements • Adaptive acceleration coefficients have also been proposed, for example to have ɸ 1 , ɸ 2 decreased over time (e.g., Simulated Annealing) 23

  24. ORIGINAL PSO: ISSUES • The acceleration coefficients should be set sufficiently high to cover large search spaces • High acceleration coefficients result in less stable systems in which the velocity has a tendency to explode! • To fix this, the velocity v is usually kept within the range [- v max , v max ] • However, limiting the velocity does not necessarily prevent particles from leaving the search space, nor does it help to guarantee convergence :( 24

  25. INERTIA COEFFICIENT • The i nertia weight ω was introduced to control the velocity explosion r 1 � ~ r 2 � ~ � � ~ ~ � + ~ � � nd ��� d ��� + ~ � soc ��� • If ω , ɸ 1 , ɸ 2 are set “correctly”, this update rule allows for convergence without the use of v max • The inertia weight can be used to control the balance between exploration and exploitation: • ω ≥ 1: velocities increase over time, swarm diverges • 0 < ω < 1: particles decelerate, convergence depends on ɸ 1 , ɸ 2 25

  26. CONSTRICTION COEFFICIENT • Take away some ‘guesswork’ for setting ω , ɸ 1 , ɸ 2 • The constriction coefficient is an “elegant” method for preventing explosion, ensuring convergence and eliminating the parameter v max • The constriction coefficient was introduced as: 26

  27. FULLY INFORMED PSO (FIPS) • Best → Average: Each particle is affected by all of its K neighbors by taking the average (personal best can or cannot be included) • The velocity update in FIPS is: • FIPS outperforms the canonical PSO’s on most test-problems • The performance of FIPS is generally more dependent on the neighborhood topology (global best neighborhood topology is recommended) 27

  28. TYPICAL BENCHMARK FUNCTIONS 28

  29. PERFORMANCE VARIANCE • Minimization problems • Best solution found over the iterations • Improvement in performance differs according to the different strategic choices • No a single winner • Early stagnation of performance can exist 29

  30. BINARY / DISCRETE PSO • A simple modification to the continuous one • Velocity remains continuous using the original update rule • Positions are updated using the velocity as a probability threshold to determine whether the j-th component of the i-th particle is a zero or a one 30

  31. ANALYSIS, GUARANTEES • Hard because: 
 - Stochastic search algorithm 
 - Complex group dynamics 
 - Performance depends on the search landscape • Theoretical analysis has been done with simplified PSOs on simplified problems • Graphical examinations of the trajectories of individual particles and their responses to variations in key parameters • Empirical performance distributions 31

  32. SUMMARY PSO • Inspired by social and roosting behaviors in bird flocking • Easy to implement, easy to get good results with “wise” parameter tuning (but just a few parameters) • Computationally light • Exploitation-Exploration dilemma • A number of variants • A few theoretical properties (hard to derive for general cases) • Mostly applied to continuous function optimization, but also to combinatorial optimization, and robotics / distributed systems • References: • Swarm Intelligence, J. Kennedy, R. Eberhart, Y. Shi, Morgan Kaufmann, 2001 • Computational Intelligence, A. Engelbrecht, Wiley, 2007 32

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend