Potential Heuristics in Satisficing Planning Alexander Rovner - - PowerPoint PPT Presentation

potential heuristics in satisficing planning
SMART_READER_LITE
LIVE PREVIEW

Potential Heuristics in Satisficing Planning Alexander Rovner - - PowerPoint PPT Presentation

Definitions DDA Complexity Approximation Algorithms Results Conclusion Potential Heuristics in Satisficing Planning Alexander Rovner University of Basel February 12, 2020 Definitions DDA Complexity Approximation Algorithms Results


slide-1
SLIDE 1

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics in Satisficing Planning

Alexander Rovner

University of Basel

February 12, 2020

slide-2
SLIDE 2

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Classical Planning

SAS+ Planning Task Π = V , I, γ, O:

slide-3
SLIDE 3

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Classical Planning

SAS+ Planning Task Π = V , I, γ, O: state variables V = {player-pos, box-pos}

slide-4
SLIDE 4

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Classical Planning

SAS+ Planning Task Π = V , I, γ, O: state variables V = {player-pos, box-pos} initial state I goal state s⋆ ⊇ γ

slide-5
SLIDE 5

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Classical Planning

SAS+ Planning Task Π = V , I, γ, O: state variables V = {player-pos, box-pos} initial state I goal state s⋆ ⊇ γ set of operators O, where each o ∈ O has a precondition, effect, and a cost

slide-6
SLIDE 6

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Classical Planning

SAS+ Planning Task Π = V , I, γ, O: state variables V = {player-pos, box-pos} initial state I goal state s⋆ ⊇ γ set of operators O, where each o ∈ O has a precondition, effect, and a cost Goal: find a sequence of actions that transforms I into a goal state

slide-7
SLIDE 7

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics

Task induces a graph called transition system/state space. Use search algorithm (e.g. A*, GBFS) to find a path from the initial state to some goal state. Search algorithms are guided towards the goal by heuristic functions.

slide-8
SLIDE 8

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics

Task induces a graph called transition system/state space. Use search algorithm (e.g. A*, GBFS) to find a path from the initial state to some goal state. Search algorithms are guided towards the goal by heuristic functions. In this thesis: potential heuristics.

slide-9
SLIDE 9

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics

Definition: Potential Heuristics Linear combination of features F ∈ F that are present in the given state s: hpot(s) :=

  • F∈F

w(F)[F ⊆ s] where w(F) is the weight of feature F and F is a set of facts.

slide-10
SLIDE 10

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics

Definition: Potential Heuristics Linear combination of features F ∈ F that are present in the given state s: hpot(s) :=

  • F∈F

w(F)[F ⊆ s] where w(F) is the weight of feature F and F is a set of facts. Central Question: how to select weights w(F) for each F ∈ F?

slide-11
SLIDE 11

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics

Definition: Potential Heuristics Linear combination of features F ∈ F that are present in the given state s: hpot(s) :=

  • F∈F

w(F)[F ⊆ s] where w(F) is the weight of feature F and F is a set of facts. Central Question: how to select weights w(F) for each F ∈ F? In Optimal Planning: choose w(F) such that hpot is admissible

slide-12
SLIDE 12

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Potential Heuristics

Definition: Potential Heuristics Linear combination of features F ∈ F that are present in the given state s: hpot(s) :=

  • F∈F

w(F)[F ⊆ s] where w(F) is the weight of feature F and F is a set of facts. Central Question: how to select weights w(F) for each F ∈ F? In Optimal Planning: choose w(F) such that hpot is admissible In Satisficing Planning: we focus on heuristics that are descending and dead-end avoiding (DDA)

slide-13
SLIDE 13

Definitions DDA Complexity Approximation Algorithms Results Conclusion

DDA Heuristics

s0 start s1 s2 s3 s4 s5 s6

slide-14
SLIDE 14

Definitions DDA Complexity Approximation Algorithms Results Conclusion

DDA Heuristics

s0 start s1 s2 s3 s4 s5 s6 States that are reachable and solvable are called alive.

slide-15
SLIDE 15

Definitions DDA Complexity Approximation Algorithms Results Conclusion

DDA Heuristics

5 start 6 3 6 4 A heuristic is descending if every alive non-goal state has an improving successor.

slide-16
SLIDE 16

Definitions DDA Complexity Approximation Algorithms Results Conclusion

DDA Heuristics

5 start 6 3 6 4 3 A heuristic is dead-end avoiding if only alive successors are improving.

slide-17
SLIDE 17

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Complexity of Computing DDA Heuristics

Central Question: How hard is it to come up with a DDA heuristic?

slide-18
SLIDE 18

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Complexity of Computing DDA Heuristics

Central Question: How hard is it to come up with a DDA heuristic? Definition: IsDDA decision problem Given: heuristic h and task Π Question: is h DDA in task Π?

slide-19
SLIDE 19

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Complexity of Computing DDA Heuristics

Central Question: How hard is it to come up with a DDA heuristic? Definition: IsDDA decision problem Given: heuristic h and task Π Question: is h DDA in task Π? Claim IsDDA is a PSPACE-complete problem.

slide-20
SLIDE 20

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Complexity of Computing DDA Heuristics

Central Question: How hard is it to come up with a DDA heuristic? Definition: IsDDA decision problem Given: heuristic h and task Π Question: is h DDA in task Π? Claim IsDDA is a PSPACE-complete problem. Proof idea: show that NotDDA (complement of IsDDA) is PSPACE-complete and use the fact that PSPACE=coPSPACE.

slide-21
SLIDE 21

Definitions DDA Complexity Approximation Algorithms Results Conclusion

PSPACE-hardness of NotDDA

Key Observations

1 If task Π is unsolvable then it has no alive states. 2 In tasks without alive states, any heuristic is DDA.

Proof: NotDDA is PSPACE-hard Reduction from PlanEx: given task Π...

slide-22
SLIDE 22

Definitions DDA Complexity Approximation Algorithms Results Conclusion

PSPACE-hardness of NotDDA

Key Observations

1 If task Π is unsolvable then it has no alive states. 2 In tasks without alive states, any heuristic is DDA.

Proof: NotDDA is PSPACE-hard Reduction from PlanEx: given task Π... construct a heuristic that is never DDA (e.g. ˆ h(s) = 0 ∀s)

slide-23
SLIDE 23

Definitions DDA Complexity Approximation Algorithms Results Conclusion

PSPACE-hardness of NotDDA

Key Observations

1 If task Π is unsolvable then it has no alive states. 2 In tasks without alive states, any heuristic is DDA.

Proof: NotDDA is PSPACE-hard Reduction from PlanEx: given task Π... construct a heuristic that is never DDA (e.g. ˆ h(s) = 0 ∀s) Π ∈ PlanEx iff Π, ˆ h ∈ NotDDA.

slide-24
SLIDE 24

Definitions DDA Complexity Approximation Algorithms Results Conclusion

PSPACE-hardness of NotDDA

Key Observations

1 If task Π is unsolvable then it has no alive states. 2 In tasks without alive states, any heuristic is DDA.

Proof: NotDDA is PSPACE-hard Reduction from PlanEx: given task Π... construct a heuristic that is never DDA (e.g. ˆ h(s) = 0 ∀s) Π ∈ PlanEx iff Π, ˆ h ∈ NotDDA. Π ∈ PlanEx iff Π, ˆ h ∈ NotDDA.

slide-25
SLIDE 25

Definitions DDA Complexity Approximation Algorithms Results Conclusion

PSPACE-membership of NotDDA

PSPACE algorithm sketch For each state s of the planning task:

1 if s is not alive ⇒ continue 2 for all successors s′ of s: 1

if s′ is not alive and h(s′) < h(s) ⇒ accept

3 if there exists no s′ with h(s′) < h(s) ⇒ accept

  • therwise fail
slide-26
SLIDE 26

Definitions DDA Complexity Approximation Algorithms Results Conclusion

PSPACE-membership of NotDDA

PSPACE algorithm sketch For each state s of the planning task:

1 if s is not alive ⇒ continue 2 for all successors s′ of s: 1

if s′ is not alive and h(s′) < h(s) ⇒ accept

3 if there exists no s′ with h(s′) < h(s) ⇒ accept

  • therwise fail

DDA computation is as hard as planning itself! ⇒ Need approximation algorithms.

slide-27
SLIDE 27

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Naive Approach

Naive Approach: compute weights by solving a MIP model.

slide-28
SLIDE 28

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Naive Approach

Naive Approach: compute weights by solving a MIP model. min (1) s.t.

  • s′∈succ(s)

h(s′) + 1 ≤ h(s) for s ∈ SA (2) h(s′) ≥ h(s) for s, s′ ∈ TD (3) SA: set of all alive states TD: set of all transitions from an alive state to an unsolvable one

slide-29
SLIDE 29

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Naive Approach

Naive Approach: compute weights by solving a MIP model. min (1) s.t.

  • s′∈succ(s)

h(s′) + 1 ≤ h(s) for s ∈ SA (2) h(s′) ≥ h(s) for s, s′ ∈ TD (3) SA: set of all alive states TD: set of all transitions from an alive state to an unsolvable one Problem: Solver usually fails to find an initial solution. ⇒ Add slack variables to the model.

slide-30
SLIDE 30

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Naive Approach

MIP model with slack variables: min

  • s∈SA

αs +

  • s,s′∈TD

β(s,s′) (4) s.t.

  • s′∈succ(s)

h(s′) + 1 − αs ≤ h(s) for s ∈ SA (5) h(s′)+βs,s′ ≥ h(s) for s, s′ ∈ TD (6) αs ≥ 0 for s ∈ SA (7) βs,s′ ≥ 0 for s, s′ ∈ TD (8)

slide-31
SLIDE 31

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Naive Approach

MIP model with slack variables: min

  • s∈SA

αs +

  • s,s′∈TD

β(s,s′) (4) s.t.

  • s′∈succ(s)

h(s′) + 1 − αs ≤ h(s) for s ∈ SA (5) h(s′)+βs,s′ ≥ h(s) for s, s′ ∈ TD (6) αs ≥ 0 for s ∈ SA (7) βs,s′ ≥ 0 for s, s′ ∈ TD (8) Simple first solution: assign large values to all α and β Can stop MIP solver early and work with an approximation. Problem: this does not scale!

slide-32
SLIDE 32

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Forward-Sampling

Simple Alternative: construct the same MIP over a random subset of all states. Main Question: how to generate the subset? ⇒ perform a random walk starting in the initial state The sample will only contain reachable states ⇒ can only assume that they are also solvable

slide-33
SLIDE 33

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Backward-Sampling

Can also generate the sample by walking backwards from some goal This also gives us the goal-distance of each state Idea: sample a pair of states where one is closer to the goal than the other

slide-34
SLIDE 34

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Backward-Sampling

Can also generate the sample by walking backwards from some goal This also gives us the goal-distance of each state Idea: sample a pair of states where one is closer to the goal than the other ⇒ can formulate an LP instead of a MIP

slide-35
SLIDE 35

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Backward-Sampling

Can also generate the sample by walking backwards from some goal This also gives us the goal-distance of each state Idea: sample a pair of states where one is closer to the goal than the other ⇒ can formulate an LP instead of a MIP min

  • (s,s′)∈Ssample

α(s,s′) (9) s.t. h(s) − h(s′) + α(s,s′) ≥ 1 (10) α(s,s′) ≥ 0 for (s, s′) ∈ Ssample (11)

slide-36
SLIDE 36

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Abstract DDA Potential Heuristics

Naive algorithm does not scale due to the large state space

slide-37
SLIDE 37

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Abstract DDA Potential Heuristics

Naive algorithm does not scale due to the large state space Idea: use abstractions to obtain a smaller state space

slide-38
SLIDE 38

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Abstract DDA Potential Heuristics

Naive algorithm does not scale due to the large state space Idea: use abstractions to obtain a smaller state space Abstract DDA Potential Heuristics:

1

use pattern selection algorithm to select an abstraction P

2

create corresponding abstract task ΠP

3

use exact algorithm to compute DDA heuristic hDDA

P

for ΠP

4

use hDDA

P

for searching the original state space

slide-39
SLIDE 39

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Abstract DDA Potential Heuristics

Naive algorithm does not scale due to the large state space Idea: use abstractions to obtain a smaller state space Abstract DDA Potential Heuristics:

1

use pattern selection algorithm to select an abstraction P

2

create corresponding abstract task ΠP

3

use exact algorithm to compute DDA heuristic hDDA

P

for ΠP

4

use hDDA

P

for searching the original state space

we can combine multiple such heuristics by summation

slide-40
SLIDE 40

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Experimental Setup

Setup: 1816 planning tasks 8 GB memory limit 30 min time limit systematically generate all features up to dimension 2

slide-41
SLIDE 41

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Naive Approach

157 out of 1816 tasks solved

slide-42
SLIDE 42

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Naive Approach

157 out of 1816 tasks solved Scalability issues:

too many constraints too many features MIP hardness

slide-43
SLIDE 43

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Forward-Sampling

Scalability issues: too many constraints ⇒ formulate MIP over a sample (sz ∈ {125, 250, 500, 1000}) too many features ⇒ use all features vs. use only 1000 randomly selected ones MIP hardness ⇒ unaddressed

slide-44
SLIDE 44

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Forward-Sampling

Scalability issues: too many constraints ⇒ formulate MIP over a sample (sz ∈ {125, 250, 500, 1000}) too many features ⇒ use all features vs. use only 1000 randomly selected ones MIP hardness ⇒ unaddressed all features 1000 features sz = 125 442 521 sz = 250 431 512 sz = 500 409 493 sz = 1000 381 490

slide-45
SLIDE 45

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Backward-Sampling

Scalability issues: too many constraints ⇒ formulate LP over a sample (sz ∈ {125, 250, 500, 1000}) too many features ⇒ use all features vs. use only 1000 randomly selected ones MIP hardness ⇒ use an LP model

slide-46
SLIDE 46

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Backward-Sampling

Scalability issues: too many constraints ⇒ formulate LP over a sample (sz ∈ {125, 250, 500, 1000}) too many features ⇒ use all features vs. use only 1000 randomly selected ones MIP hardness ⇒ use an LP model all features 1000 features sz = 125 469 538 sz = 250 477 560 sz = 500 479 575 sz = 1000 487 575

slide-47
SLIDE 47

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Single Abstract DDA Heuristic

Scalability issues: too many constraints ⇒ formulate MIP for an abstraction (sz ∈ {256, 512, 1024, 2048}) too many features ⇒ resolved due to abstraction MIP hardness ⇒ unaddressed

slide-48
SLIDE 48

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Single Abstract DDA Heuristic

Scalability issues: too many constraints ⇒ formulate MIP for an abstraction (sz ∈ {256, 512, 1024, 2048}) too many features ⇒ resolved due to abstraction MIP hardness ⇒ unaddressed single abs-DDA single PDB sz = 256 581 732 sz = 512 561 747 sz = 1024 513 758 sz = 2048 455 768

slide-49
SLIDE 49

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Multiple Abstract DDA Heuristics

Scalability issues: too many constraints ⇒ formulate MIP for an abstraction (sz ∈ {128, 256, 512, 1024}) and atomic abstractions too many features ⇒ resolved due to abstraction MIP hardness ⇒ unaddressed

slide-50
SLIDE 50

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage: Multiple Abstract DDA Heuristics

Scalability issues: too many constraints ⇒ formulate MIP for an abstraction (sz ∈ {128, 256, 512, 1024}) and atomic abstractions too many features ⇒ resolved due to abstraction MIP hardness ⇒ unaddressed multiple abs-DDA multiple PDB atomic 1028 1107 sz = 128 1005 1121 sz = 256 1005 1130 sz = 512 1005 1128 sz = 1024 999 1130

slide-51
SLIDE 51

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Coverage

bw-sampling multiple abs-DDA multiple PDBs logistics98 3 8 35 visitall14 20

  • penstacks08

8 30 6 parcprinter11 12 tpp 8 29 9 snake18 18 5 7

slide-52
SLIDE 52

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Heuristic Quality

10−1 101 103 105 107 109 10−1 102 105 108 109 109 single PDB (256) single abs-DDA (256) expansions

slide-53
SLIDE 53

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Heuristic Quality

10−1 101 103 105 107 109 10−1 102 105 108 109 109 multiple PDB (128) multiple abs-DDA (128) expansions

slide-54
SLIDE 54

Definitions DDA Complexity Approximation Algorithms Results Conclusion

Conclusion

DDA heuristics are PSPACE-hard to compute approximation algorithms are necessary ⇒ most promising approach: abs-DDA potential heuristics

  • utscaled by PDBs (PDB computation is more efficient)

Heuristic quality is comparable to PDBs