Adaptive Submodularity: A New Approach to Active Learning and - - PowerPoint PPT Presentation

adaptive submodularity a new approach to active learning
SMART_READER_LITE
LIVE PREVIEW

Adaptive Submodularity: A New Approach to Active Learning and - - PowerPoint PPT Presentation

Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin and Andreas Krause (FYI, a powerpoint version of these slides is available on Daniel Golovins website.) 1 Max K-Cover (Oil Spill Edition) 2


slide-1
SLIDE 1

1

Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization

Daniel Golovin and Andreas Krause

(FYI, a powerpoint version of these slides is available on Daniel Golovin‟s website.)

slide-2
SLIDE 2

2

Max K-Cover (Oil Spill Edition)

slide-3
SLIDE 3

3

Submodularity

Time Time Discrete diminishing returns property for set functions. ``Playing an action at an earlier stage

  • nly increases its marginal benefit''
slide-4
SLIDE 4

4

The Greedy Algorithm

Theorem [Nemhauser et al „78]

slide-5
SLIDE 5

5

Stochastic Max K-Cover

Asadpour & Saberi (`08): (1-1/e)-approx if sensors (independently) either work perfectly or fail completely. Bayesian: Known failure distribution. Adaptive: Deploy a sensor and see what you get. Repeat K times.

0.5 0.2 0.3 At 1st location

slide-6
SLIDE 6

6

Adaptive Submodularity

Time Playing an action at an earlier stage

  • nly increases its marginal benefit

expected (taken over its outcome)

Gain more Gain less

(i.e., at an ancestor)

Select Item Stochastic Outcome

slide-7
SLIDE 7

7

Adaptive Monotonicity

slide-8
SLIDE 8

8

What is it good for?

Allows us to generalize various results to the adaptive realm, including:

  • (ln(n)+1)-approximation for Set Cover
  • (1-1/e)-approximation for Max K-Cover, submodular

maximization subject to a cardinality constraint

slide-9
SLIDE 9

9

Recall the Greedy Algorithm

Theorem [Nemhauser et al „78]

slide-10
SLIDE 10

10

The Adaptive-Greedy Algorithm

Theorem

slide-11
SLIDE 11

11

[Adapt-monotonicity]

  • (

)

  • [Adapt-submodularity]
slide-12
SLIDE 12

12

The world-state dictates which path in the tree we‟ll take.

  • 1. For each node at layer i+1,
  • 2. Sample path to layer j,
  • 3. Play the resulting layer j action at layer i+1.

How to play layer j at layer i+1

By adapt. submod., playing a layer earlier

  • nly increases it‟s

marginal benefit

slide-13
SLIDE 13

13

[Adapt-monotonicity]

  • (

)

  • (

)

  • [Def. of adapt-greedy]

( )

  • [Adapt-submodularity]
slide-14
SLIDE 14

14

slide-15
SLIDE 15

15

2 1 3

Stochastic Max Cover is Adapt-Submod

1 3

Gain more Gain less adapt-greedy is a (1-1/e) ≈ 63% approximation to the adaptive optimal solution. Random sets distributed independently.

slide-16
SLIDE 16

16

Stochastic Min Cost Cover

 Adaptively get a threshold amount of value.  Minimize expected number of actions.  If objective is adapt-submod and

monotone, we get a logarithmic approximation.

[Goemans & Vondrak, LATIN „06] [Liu et al., SIGMOD „08] [Feige, JACM „98] [Guillory & Bilmes, ICML „10]

c.f., Interactive Submodular Set Cover

slide-17
SLIDE 17

17

Optimal Decision Trees

x1 x2 x3 1 1 = = =

Garey & Graham, 1974; Loveland, 1985; Arkin et al., 1993; Kosaraju et al., 1999; Dasgupta, 2004; Guillory & Bilmes, 2009; Nowak, 2009; Gupta et al., 2010

“Diagnose the patient as cheaply as possible (w.r.t. expected cost)” 1 1

slide-18
SLIDE 18

18

Objective = probability mass of hypotheses you have ruled out. It‟s Adaptive Submodular.

Outcome = 1 Outcome = 0 Test x Test w Test v

slide-19
SLIDE 19

19

Conclusions

 New structural property useful for design & analysis of

adaptive algorithms

 Powerful enough to recover and generalize many known

results in a unified manner. (We can also handle costs)

 Tight analyses and optimal approximation factors in

many cases. 2 1 3

x1 x2 x3 1 1 1

0.5 0.3 0.5 0.4 0.2 0.2 0.5