Latent Class Models for Algorithm Portfolio Methods Bryan - - PowerPoint PPT Presentation

latent class models for algorithm portfolio methods
SMART_READER_LITE
LIVE PREVIEW

Latent Class Models for Algorithm Portfolio Methods Bryan - - PowerPoint PPT Presentation

Latent Class Models for Algorithm Portfolio Methods Bryan Silverthorn and Risto Miikkulainen Department of Computer Science The University of Texas at Austin 13 July 2010 Our setting: hard computational problems Many important computational


slide-1
SLIDE 1

Latent Class Models for Algorithm Portfolio Methods

Bryan Silverthorn and Risto Miikkulainen

Department of Computer Science The University of Texas at Austin

13 July 2010

slide-2
SLIDE 2

Our setting: hard computational problems

Many important computational questions, such as satisfiability (SAT), are intractable in the worst case.

Definition (SAT)

Does a truth assignment exist that makes a given Boolean expression true? But heuristics often work: enormous problem instances can be solved. SAT solvers are now in routine use for applications such as hardware verification... with up to a million variables. (Kautz and Selman, 2007)

2 / 17

slide-3
SLIDE 3

Which solver do you choose? (2009 SAT competition)

march hi hybridGM3 gnovelty+2 MiniSat-2.1 clasp ag2wsat09++ SApperloT march KS SATzilla09 R precosat TNM glucose iPAWS adaptg2wsat+ MXC gnovelty+ gnovelty+-T SATzilla09 C SATzilla07 C SATzilla09 I kw Rsat minisat CircUs LySAT-i IUT BMB SAT LySAT-c MiniSAT-09z VARSAT-I ManySAT 1.1 minisat cumr rsat picosat SATzilla07 R 20 40 60 80 100 120

Instances where solver was best

3 / 17

slide-4
SLIDE 4

Algorithm portfolios: what and why

Definition

An algorithm portfolio is a pool of algorithms (“solvers”) and a method for scheduling their execution. Portfolios can reduce effort by choosing solvers automatically, and improve performance by allocating resources more effectively. Existing portfolios, such as SATzilla (Xu et al. 2008), often use classifiers trained on feature information to predict solver performance.

4 / 17

slide-5
SLIDE 5

How should we predict solver performance?

Research Questions

What predictions can we make with minimal information? What assumptions are needed to make useful predictions? Do they hold sufficiently well in practice? To explore these questions, we will build unifying generative models of solver behavior and evaluate them in the SAT domain.

5 / 17

slide-6
SLIDE 6

Assumptions that make modeling easier

Outcomes of runs are discrete, few, and fixed. Utilities of outcomes are known. Durations of runs are discrete, few, and fixed. Learning is offline, but action is online. Tasks are drawn IID from some distribution. Information is obtained from outcomes alone.

6 / 17

slide-7
SLIDE 7

Architecture of a model-based portfolio

Training Outcomes Training Tasks Solvers

7 / 17

slide-8
SLIDE 8

Architecture of a model-based portfolio

Training Outcomes Training Tasks Model Solvers

7 / 17

slide-9
SLIDE 9

Architecture of a model-based portfolio

Training Outcomes Training Tasks Model Predicted Outcomes Solvers

7 / 17

slide-10
SLIDE 10

Architecture of a model-based portfolio

Training Outcomes Training Tasks Model Predicted Outcomes Action Selection Solvers

7 / 17

slide-11
SLIDE 11

Architecture of a model-based portfolio

Training Outcomes Training Tasks Model Predicted Outcomes Action Selection Test Outcomes Test Tasks Solvers

7 / 17

slide-12
SLIDE 12

Basic structure in solver behavior

Inter-algorithm correlations: solvers can be (dis)similar.

Example

“If solver X yielded outcome A on this task, solver Y likely will as well.” Inter-task correlations: tasks can be (dis)similar.

Example

“If solver X yielded outcome A on task 1, it likely will on task 2 as well.” Inter-duration correlations: runs can have (dis)similar outcomes.

Example

“If solver X did not quickly yield outcome A on this task, it never will.”

8 / 17

slide-13
SLIDE 13

Conditional independence in solver behavior

The outcome of a solver run is a function of only three inputs: the task on which it is executed, the duration of the run, and the seed of any internal pseudorandom sequence. This strong local independence suggests a possible model: take actions to be solver-duration pairs, and assume that tasks cluster into classes. Classes then capture the basic three aspects of solver behavior.

9 / 17

slide-14
SLIDE 14

Multinomial latent class model of search

ξ

Model Key

ξ class prior

10 / 17

slide-15
SLIDE 15

Multinomial latent class model of search

ξ β

Model

β ∼ Dir(ξ)

Key

ξ class prior β class distr.

10 / 17

slide-16
SLIDE 16

Multinomial latent class model of search

T ξ β k

Model

β ∼ Dir(ξ) kt ∼ Mult(β) t ∈ 1 . . . T

Key

ξ class prior β class distr. k class T tasks

10 / 17

slide-17
SLIDE 17

Multinomial latent class model of search

S T ξ β k α

Model

β ∼ Dir(ξ) kt ∼ Mult(β) t ∈ 1 . . . T

Key

ξ class prior α

  • utcome prior

β class distr. k class T tasks S actions

10 / 17

slide-18
SLIDE 18

Multinomial latent class model of search

S T K ξ β k θ α

Model

β ∼ Dir(ξ) kt ∼ Mult(β) t ∈ 1 . . . T θs,k ∼ Dir(αs) s ∈ 1 . . . S k ∈ 1 . . . K

Key

ξ class prior α

  • utcome prior

β class distr. θ

  • utcome distr.

k class T tasks S actions K classes

10 / 17

slide-19
SLIDE 19

Multinomial latent class model of search

S T Rs,t K ξ β k

  • θ

α

Model

β ∼ Dir(ξ) kt ∼ Mult(β) t ∈ 1 . . . T θs,k ∼ Dir(αs) s ∈ 1 . . . S k ∈ 1 . . . K

  • t,s,r ∼ Mult(θs,kt)

t ∈ 1 . . . T s ∈ 1 . . . S r ∈ 1 . . . Rs,t

Key

ξ class prior α

  • utcome prior

β class distr. θ

  • utcome distr.

k class

  • utcome

T tasks S actions K classes R runs

10 / 17

slide-20
SLIDE 20

Burstiness: another important aspect of solver behavior

Definition

Burstiness is the tendency of some random events to recur. Solver outcomes recur—for some solvers more than others.

Example

“If solver X yields outcome A on this task, it will again; not so for Y.” Deterministic solvers are entirely bursty. Randomized solvers are less so. Burstiness also appears in text data. The Dirichlet compound multinomial (DCM) distribution has modeled it well in that domain. (Madsen et al., 2005)

11 / 17

slide-21
SLIDE 21

Multinomial latent class model of search

S T Rs,t K ξ β k

  • θ

α

Model

β ∼ Dir(ξ) kt ∼ Mult(β) t ∈ 1 . . . T θs,k ∼ Dir(αs) s ∈ 1 . . . S k ∈ 1 . . . K

  • t,s,r ∼ Mult(θs,kt)

t ∈ 1 . . . T s ∈ 1 . . . S r ∈ 1 . . . Rs,t

Key

ξ class prior α

  • utcome prior

β class distr. θ

  • utcome distr.

k class

  • utcome

T tasks S actions K classes R runs

12 / 17

slide-22
SLIDE 22

DCM (bursty) latent class model of search

S T Rs,t K ξ β k

  • θ

α

Model

β ∼ Dir(ξ) kt ∼ Mult(β) t ∈ 1 . . . T θt,s ∼ Dir(αs,kt) s ∈ 1 . . . S t ∈ 1 . . . T

  • t,s,r ∼ Mult(θt,s)

t ∈ 1 . . . T s ∈ 1 . . . S r ∈ 1 . . . Rs,t

Key

ξ class prior α

  • utcome root

β class distr. θ

  • utcome distr.

k class

  • utcome

T tasks S actions K classes R runs

13 / 17

slide-23
SLIDE 23

Greedy, discounted selection

One efficient approach is to choose the next action according to immediate expected utility without regard to later actions. This approach gives us a hard policy that chooses the expected-best action, and a soft policy that draws actions proportional to expected utility. Actions are solver-duration pairs: they have wildly different costs. An obvious response is to reduce an action’s expected utility by its cost, discounting by γc for a c-second run and factor γ.

14 / 17

slide-24
SLIDE 24

Experimental procedure

In our experiments, we use every individual solver from the latest SAT competition, and every problem instance from its three benchmark collections; in repeated trials, we run the solvers on a randomly-drawn training set, fit a model to that training data, and then run a portfolio using that model on the remaining test set.

Empirical Questions

For each combination of model and action selection policy, how does its performance compare to its subsolvers? how does its performance compare to that of other portfolios?

15 / 17

slide-25
SLIDE 25

Portfolio performance (on the random collection)

5 10 15 20 25 30 35 40 45 50 55 60 65 Number of latent classes (K) 340 360 380 400 420 440 460 SAT instances solved

DCM Multinomial Best Single SATzilla

16 / 17

slide-26
SLIDE 26

Recapitulation

These results suggest that models can capture useful patterns given little information, and these latent class models can be applied to a portfolio in practice. Research in progress aims to extend these models to capture dynamic information, and improve action planning to better exploit their predictions.

20 40 60 80 100 120 200 400 600 800 1000 1200

17 / 17

slide-27
SLIDE 27

Thanks!—Questions?

18 / 17

slide-28
SLIDE 28

References I

Matteo Gagliolo and Jurgen Schmidhuber. Learning Restart Strategies. IJCAI 2007. Carla Gomes and Bart Selman. Algorithm Portfolio Design: Theory vs. Practice. UAI 1997. Eric Horvitz, Ruan Yongshao, Carla Gomes, Henry Kautz, Bart Selman, and David Chickering. A Bayesian Approach to Tackling Hard Computational Problems. UAI 2001. Bernardo Huberman, Rajan Lukose, and Tad Hogg. An Economics Approach to Hard Computational Problems. Science, 275(5296), 1997.

19 / 17

slide-29
SLIDE 29

References II

Frank Hutter, Domagoj Babi´ c, Holger H. Hoos, and Alan J. Hu. Boosting Verification by Automatic Tuning of Decision Procedures. FMCAD 2007. Frank Hutter, Holger H. Hoos, Kevin Leyton-Brown, and Thomas St¨ utzle. ParamILS: An Automatic Algorithm Configuration Framework. Journal of Artificial Intelligence Research, 2009. Henry Kautz and Bart Selman. The state of SAT. Discrete Applied Mathematics, 12, 2007. Ashiqur KhudaBukhsh, Lin Xu, Holger H. Hoos, Kevin Leyton-Brown. SATenstein: Automatically Building Local Search SAT Solvers From Components. IJCAI 2009.

20 / 17

slide-30
SLIDE 30

References III

Rasmus Madsen, David Kauchak, and Charles Elkan. Modeling Word Burstiness Using the Dirichlet Distribution. ICML 2005. David Mimno and Andrew McCallum. Topic Models Conditioned on Arbitrary Features with Dirichlet-multinomial Regression. UAI 2008. Mladen Nikoli´ c, Filip Mari´ c, and Predrag Janiˇ ci´ c. Instance-Based Selection of Policies for SAT Solvers. SAT 2009. Eugene Nudelman, Kevin Leyton-Brown, Holger H. Hoos, Alex Devkar, and Yoav Shoham. Understanding Random SAT: Beyond the Clauses-to-Variables Ratio. CP 2004.

21 / 17

slide-31
SLIDE 31

References IV

  • J. R. Rice.

The Algorithm Selection Problem. Advances in Computers, 15, 1976.

22 / 17