Partial Operator Induction with Beta Distribution Nil Geisweiller - - PowerPoint PPT Presentation

partial operator induction with beta distribution
SMART_READER_LITE
LIVE PREVIEW

Partial Operator Induction with Beta Distribution Nil Geisweiller - - PowerPoint PPT Presentation

Partial Operator Induction with Beta Distribution Nil Geisweiller AGI-18 Prague 1 Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning 2


slide-1
SLIDE 1

Partial Operator Induction with Beta Distribution

Nil Geisweiller AGI-18 Prague

1

slide-2
SLIDE 2

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

2

slide-3
SLIDE 3

Outline

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

3

slide-4
SLIDE 4

Outline

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

4

slide-5
SLIDE 5

Problem: Models from different contexts

How to combine models obtained from different contexts? Large Contexts → Underfit Small Contexts → Overfit

5

slide-6
SLIDE 6

Problem: Preserve Uncertainty

6

slide-7
SLIDE 7

Problem: Preserve Uncertainty

Exploration vs Exploitation (Thompson Sampling)

7

slide-8
SLIDE 8

Problem: ImplicationLink

ImplicationLink <TV> R S

≡ Second Order P(S|R) Beta Distribution in disguise

8

slide-9
SLIDE 9

Solution

Bayesian Model Averaging / Solomonoff Operator Induction, modified to:

  • 1. Support partial models
  • 2. Produce a probability distribution estimate, rather than

probability estimate.

  • 3. Specialize for Beta distributions

9

slide-10
SLIDE 10

Outline

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

10

slide-11
SLIDE 11

Outline

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

11

slide-12
SLIDE 12

Solomonoff Operator Induction

Bayesian Model Averaging + Universal Distribution Probability Estimate: ˆ P(An+1|Qn+1) =

  • j

aj

n+1

  • i=1

Oj(Ai|Qi) where:

  • Qi = ith question
  • Ai = ith answer
  • Oj = jth operator
  • aj

0 = prior of jth operator 12

slide-13
SLIDE 13

Beta Distribution Operator

Specialization of Solomonoff Operator Induction OpenCog implication link

ImplicationLink <TV> R S

≡ Class of parameterized operators Oj

p(Ai|Qi) = if Rj(Qi) then

   p, if Ai = An+1 1 − p,

  • therwise

13

slide-14
SLIDE 14

Beta Distribution

Probability Density Function: pdfα,β(x) = xα−1(1 − x)β−1 B(α, β) Beta Function: Bx(α, β) = x pα−1(1−p)β−1dp B(α, β) = B1(α, β) Conjugate Prior: pdfm+α,n−m+β(x) ∝ xm(1−x)n−mpdfα,β(x)

14

slide-15
SLIDE 15

Artificial Completion

Oj

p

(Ai|Qi) = if Rj(Qi) then    p, if Ai = An+1 1 − p,

  • therwise

15

slide-16
SLIDE 16

Artificial Completion

Oj

p,C(Ai|Qi) =

if Rj(Qi) then    p, if Ai = An+1 1 − p,

  • therwise

else C(Ai|Qi)

15

slide-17
SLIDE 17

Second Order Solomonoff Operator Induction

Probability Estimate: ˆ P(An+1|Qn+1) =

  • j

aj

n+1

  • i=1

Oj(Ai|Qi) Probability Distribution Estimate: ˆ cdf (An+1|Qn+1)(x) =

  • Oj(An+1|Qn+1)≤x

aj

n

  • i=1

Oj(Ai|Qi)

16

slide-18
SLIDE 18

Combing Solomonoff Operator Induction and Beta Distributions

ˆ cdf (An+1|Qn+1)(x) ∝

  • j

aj

0r jBx(mj+α, nj−mj+β)B(mj+α, nj−mj+β)

where

  • nj = number of observations explained by jth model
  • mj = number of true observations explained by jth model
  • r j = likelihood of the unexplained data

r j =???

17

slide-19
SLIDE 19

Combing Solomonoff Operator Induction and Beta Distributions

ˆ cdf (An+1|Qn+1)(x) ∝

  • j

aj

0r jBx(mj+α, nj−mj+β)B(mj+α, nj−mj+β)

where

  • nj = number of observations explained by jth model
  • mj = number of true observations explained by jth model
  • r j = likelihood of the unexplained data

r j =??? ≈ 2−v(1−c)

  • v = n − nj = number of unexplained observations
  • c = compressability parameter
  • c = 1 → explains remaining data
  • c = 0 → can’t explain remaining data

17

slide-20
SLIDE 20

Outline

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

18

slide-21
SLIDE 21

Outline

Problem: Combining Models from Different Contexts Theory: Solomonoff Operator Induction and Beta Distribution Practice: Inference Control Meta-Learning

19

slide-22
SLIDE 22

Inference Control Meta-learning

Learn how to reason efficiently

20

slide-23
SLIDE 23

Inference Control Meta-learning

Learn how to reason efficiently Methodology:

  • 1. Solve sequence of problems (via reasoning)

20

slide-24
SLIDE 24

Inference Control Meta-learning

Learn how to reason efficiently Methodology:

  • 1. Solve sequence of problems (via reasoning)
  • 2. Store inference traces

20

slide-25
SLIDE 25

Inference Control Meta-learning

Learn how to reason efficiently Methodology:

  • 1. Solve sequence of problems (via reasoning)
  • 2. Store inference traces
  • 3. Mine traces to discover patterns

20

slide-26
SLIDE 26

Inference Control Meta-learning

Learn how to reason efficiently Methodology:

  • 1. Solve sequence of problems (via reasoning)
  • 2. Store inference traces
  • 3. Mine traces to discover patterns
  • 4. Build control rules

Implication <TV> And <inference-pattern> <rule> <good-inference>

20

slide-27
SLIDE 27

Inference Control Meta-learning

Learn how to reason efficiently Methodology:

  • 1. Solve sequence of problems (via reasoning)
  • 2. Store inference traces
  • 3. Mine traces to discover patterns
  • 4. Build control rules

Implication <TV> And <inference-pattern> <rule> <good-inference>

  • 5. Combine control rules to guide future reasoning

20

slide-28
SLIDE 28

Combine Control Rules

Implication <TV1> And <inference-pattern-1> deduction-rule <good-inference>

c = 1 Implication <TV2> And <inference-pattern-2> deduction-rule <good-inference>

=

21

slide-29
SLIDE 29

Combine Control Rules

Implication <TV1> And <inference-pattern-1> deduction-rule <good-inference>

c = 0.5 Implication <TV2> And <inference-pattern-2> deduction-rule <good-inference>

=

22

slide-30
SLIDE 30

Combine Control Rules

Implication <TV1> And <inference-pattern-1> deduction-rule <good-inference>

c = 0.1 Implication <TV2> And <inference-pattern-2> deduction-rule <good-inference>

=

23

slide-31
SLIDE 31

Conclusion

Contribution:

  • Second Order Solomonoff Operator Induction
  • Specialized for Beta Distribution
  • Attempt to Deal with Partial Models

24

slide-32
SLIDE 32

Conclusion

Contribution:

  • Second Order Solomonoff Operator Induction
  • Specialized for Beta Distribution
  • Attempt to Deal with Partial Models

Future Work:

  • Improve Likelihood of Unexplained Data
  • More Experiments (Inference Control Meta-learning)

24

slide-33
SLIDE 33

Conclusion

Contribution:

  • Second Order Solomonoff Operator Induction
  • Specialized for Beta Distribution
  • Attempt to Deal with Partial Models

Future Work:

  • Improve Likelihood of Unexplained Data
  • More Experiments (Inference Control Meta-learning)

Thank you!

24