Making Solomonoff Induction Effective You Can Learn What You Can - - PowerPoint PPT Presentation

making solomonoff induction effective
SMART_READER_LITE
LIVE PREVIEW

Making Solomonoff Induction Effective You Can Learn What You Can - - PowerPoint PPT Presentation

Making Solomonoff Induction Effective You Can Learn What You Can Bound J org Zimmermann and Armin B. Cremers Institute of Computer Science University of Bonn, Germany 1 J org Zimmermann and Armin B. Cremers: Making Solomonoff Induction


slide-1
SLIDE 1

Making Solomonoff Induction Effective

You Can Learn What You Can Bound J¨

  • rg Zimmermann and Armin B. Cremers

Institute of Computer Science University of Bonn, Germany

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

1

slide-2
SLIDE 2

The General Prediction Problem

Given a finite sequence of bits, e.g.:

0010010000111111011010101000100010000101

Question: What is the next bit?

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

2

slide-3
SLIDE 3

Asynchronous Learning Framework (ALF)

A learning system observing and predicting an environment:

work tape work tape ... ... Learning System p Environment q p1 p2 p3 p4 ...

  • 1
  • 2
  • 3
  • 4

...

  • bservations

predictions

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

3

slide-4
SLIDE 4

Solomonoff Induction

  • Bayesian learning in program space.
  • Prior ∼ 2−|p|,

|p| = length of program p in bits.

  • But posterior distribution on program space is not computable!

(the programs stopping to produce output cause trouble).

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

4

slide-5
SLIDE 5

Key Points of our Approach

  • Learning driven by a combined search in program and proof space.
  • Reduction of learnability to provability and set existence axioms.

Axiom systems of reverse mathematics and large cardinal axioms can be used to show that proof-theoretic strength translates into learning strength.

  • Introduction of a new learning framework, the Synchronous Learning

Framework (SLF), which couples the time scales of the learning system and the environment.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

5

slide-6
SLIDE 6

Probabilistic Learning Systems Λ : {0, 1}∗ × {0, 1} → [0, 1]Q

with Λ(x, 0) + Λ(x, 1) = 1 for all x ∈ {0, 1}∗. Λ is an effective probabilistic learning system if Λ is a total recursive function.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

6

slide-7
SLIDE 7

Learnability

Learning as learning in the limit: Eventually the learning system will become near certain about the true continuation of the observed bit sequence.

Definition: An infinite bit sequence s is learnable in the limit by the probabilistic learning system Λ, if for all ǫ > 0 there is an n0 so that for all n ≥ n0 and all k ≥ 1: Λ(k)(s1:n, sn+1:n+k) > 1 − ǫ. Λ(k): extending prediction horizon to k bits by feeding Λ with its own predictions.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

7

slide-8
SLIDE 8

Σ-driven Learning Systems

  • Turning an Axiom System Σ into a Learning System Λ:

Σ − → Λ(Σ)

  • A Σ-driven learning system is a learning system using the

background theory Σ in order to derive totality proofs for recursive functions.

  • These provably recursive functions are used to build a guard

function, which schedules the learning process and guarantees its effectiveness.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

8

slide-9
SLIDE 9

Generator Time Function

The generator time function of a program p is defined as:

Gp : N → N ∪ {∞}

Gp(n) = #transitions executed by p to generate the first n bits.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

9

slide-10
SLIDE 10

Observation Equivalence

sp = the bit sequence generated by program p. Then the observation class [s] of a bit sequence s is defined as:

p ∈ [s] iff s = sp.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

10

slide-11
SLIDE 11

Generator-Predictor Theorem The infinite bit sequence s is learnable by Λ(Σ), if: ∃ p ∈ [s], f recursive function : Σ ⊢ φtot(f) and f ≥d Gp.

φtot(f) = f is a total recursive function. f ≥d g = f dominates g (i.e., ∃n0 ∀n ≥ n0 : f(n) ≥ g(n)).

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

11

slide-12
SLIDE 12

The hard case: infinite number of switches

suspend reactivate

Gp guard function

#observed bits #transitions

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

12

slide-13
SLIDE 13

Σ-driven probabilistic learning system

Idea: retroactive change of prior: ∼ 2−(|p|+switch(p,n)) (Solomonoff prior ∼ 2−|p|) = ⇒ Dynamic Bayesian Inference, i.e., construction of model space and prior probabilities is interleaved with the inference process.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

13

slide-14
SLIDE 14

Conclusions 1

  • The generator-predictor theorem establishes a natural perspective on

the effective core of Solomonoff induction.

  • This shifts the questions related to learnability to questions related

to provability, and therefore into the realm of the foundations of mathematics.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

14

slide-15
SLIDE 15

Synchronous Learning Framework (SLF)

Observation: in real world learning situations, the generator and the learner are not suspended while the other one is busy. s is synchronous :⇐ ⇒ lim sup

n→∞ Gp(n) n

< ∞ for at least one p ∈ [s]. = ⇒ the time scales of the learning system and the environment are coupled.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

15

slide-16
SLIDE 16

Clockification

1 ... 10 00 00 10 11 00 10 10 ... Internal Clock arbitrary computable sequence synchronous sequence clock signal 00 1 10 11

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

16

slide-17
SLIDE 17

Synchronous Learning Framework

  • Clockification transforms every computable bit sequence into a

synchronous one.

  • All synchronous bit sequences are learnable by Λ(Σ), if Σ ⊢ “n2 is a

total recursive function”.

  • Thus in the SLF all effectively generated bit sequences can be

effectively learned.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

17

slide-18
SLIDE 18

Final Conclusion

If the learning system is enhanced by an internal clock:

Effective universal induction is possible!

Hence future research can focus on efficient universal induction.

  • rg Zimmermann and Armin B. Cremers: Making Solomonoff Induction Effective

18