On the exact learnability of graph parameters The case of partition - - PowerPoint PPT Presentation

on the exact learnability of graph parameters
SMART_READER_LITE
LIVE PREVIEW

On the exact learnability of graph parameters The case of partition - - PowerPoint PPT Presentation

On the exact learnability of graph parameters The case of partition functions Nadia Labai TU Wien Joint work with Johann Makowsky Exact learning YES polynomial time. is exactly learnable if the learner finds a correct hypothesis in A


slide-1
SLIDE 1

On the exact learnability of graph parameters

The case of partition functions

Nadia Labai TU Wien Joint work with Johann Makowsky

slide-2
SLIDE 2

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE x

f x

EQUIVALENT h

YES h x f x

  • Value queries - learner sends input x, teacher sends back f x
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-3
SLIDE 3

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE(x)

f x

EQUIVALENT h

YES h x f x

  • Value queries - learner sends input x, teacher sends back f(x)
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-4
SLIDE 4

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE(x)

f(x)

EQUIVALENT h

YES h x f x

  • Value queries - learner sends input x, teacher sends back f(x)
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-5
SLIDE 5

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE x

f x

EQUIVALENT(h)

YES h x f x

  • Value queries - learner sends input x, teacher sends back f(x)
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-6
SLIDE 6

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE x

f x

EQUIVALENT(h)

YES h x f x

  • Value queries - learner sends input x, teacher sends back f(x)
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-7
SLIDE 7

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE x

f x

EQUIVALENT(h)

YES h(x) ̸= f(x)

  • Value queries - learner sends input x, teacher sends back f(x)
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-8
SLIDE 8

Exact learning

Proposed by Angluin in 1987. The scenario includes: a target function f ∈ C, a learner maintaining a hypothesis h and a powerful teacher. Teacher Learner f ∈ C h ∈ C

VALUE x

f x

EQUIVALENT h

YES h x f x

  • Value queries - learner sends input x, teacher sends back f(x)
  • Equivalence queries - learner sends a hypothesis, teacher sends:
  • YES if the hypothesis is correct
  • A counterexample if it is incorrect

C is exactly learnable if the learner finds a correct hypothesis in polynomial time.

2/18

slide-9
SLIDE 9

Existing exact learning algorithms

Exact learning algorithms were developed for word and tree functions representable as automata

  • usually rely on an algebraic characterization of these functions via

Hankel matrices The Hankel matrix

f of a word function f

  • Infinite matrix with rows and

columns indexed by words u1 u2

  • ver

:

f

  • The entry u v is f uv :

f u v

f uv

u1 uj u1 . . . ui . . . f uiuj 3/18

slide-10
SLIDE 10

Existing exact learning algorithms

Exact learning algorithms were developed for word and tree functions representable as automata

  • usually rely on an algebraic characterization of these functions via

Hankel matrices The Hankel matrix Hf of a word function f : Σ⋆ → R

  • Infinite matrix with rows and

columns indexed by words u1 u2

  • ver

:

f

  • The entry u v is f uv :

f u v

f uv

u1 . . . uj . . . u1 . . . ui . . . f(uiuj) 3/18

slide-11
SLIDE 11

Existing exact learning algorithms

Exact learning algorithms were developed for word and tree functions representable as automata

  • usually rely on an algebraic characterization of these functions via

Hankel matrices The Hankel matrix Hf of a word function f : Σ⋆ → R

  • Infinite matrix with rows and

columns indexed by words u1, u2, . . . over Σ: Hf ∈ RΣ⋆×Σ⋆

  • The entry u v is f uv :

f u v

f uv

u1 . . . uj . . . u1 . . . ui . . . f(uiuj) 3/18

slide-12
SLIDE 12

Existing exact learning algorithms

Exact learning algorithms were developed for word and tree functions representable as automata

  • usually rely on an algebraic characterization of these functions via

Hankel matrices The Hankel matrix Hf of a word function f : Σ⋆ → R

  • Infinite matrix with rows and

columns indexed by words u1, u2, . . . over Σ: Hf ∈ RΣ⋆×Σ⋆

  • The entry (u, v) is f(uv):

Hf(u, v) = f(uv)

u1 . . . uj . . . u1 . . . ui . . . f(uiuj) 3/18

slide-13
SLIDE 13

Learning word and tree automata

Typical characterization theorem: A function is representable as an automaton iff its Hankel matrix has finite rank.

  • 1. The proofs usually provide a direct translation from Hankel matrix

to automaton

  • 2. Algorithms iteratively build a submatrix of the Hankel matrix using

query answers

  • 3. Eventually the submatrix is large enough to provide a correct

automaton

4/18

slide-14
SLIDE 14

Learning word and tree automata

Typical characterization theorem: A function is representable as an automaton iff its Hankel matrix has finite rank.

  • 1. The proofs usually provide a direct translation from Hankel matrix

to automaton

  • 2. Algorithms iteratively build a submatrix of the Hankel matrix using

query answers

  • 3. Eventually the submatrix is large enough to provide a correct

automaton

4/18

slide-15
SLIDE 15

Learning word and tree automata

Typical characterization theorem: A function is representable as an automaton iff its Hankel matrix has finite rank.

  • 1. The proofs usually provide a direct translation from Hankel matrix

to automaton

  • 2. Algorithms iteratively build a submatrix of the Hankel matrix using

query answers

  • 3. Eventually the submatrix is large enough to provide a correct

automaton

4/18

slide-16
SLIDE 16

Learning word and tree automata

Typical characterization theorem: A function is representable as an automaton iff its Hankel matrix has finite rank.

  • 1. The proofs usually provide a direct translation from Hankel matrix

to automaton

  • 2. Algorithms iteratively build a submatrix of the Hankel matrix using

query answers

  • 3. Eventually the submatrix is large enough to provide a correct

automaton

4/18

slide-17
SLIDE 17

Similar theorem for MSOL-definable graph parameters

Definition of MSOL-definable graph parameters is not in this talk. Examples include:

  • various counting functions for graphs
  • functions recognized by weighted word and tree automata

Finite Rank Theorem (Godlin, Kotek, Makowsky): If a real-valued graph parameter is MSOL-definable, its connection matrix has finite rank.

5/18

slide-18
SLIDE 18

Similar theorem for MSOL-definable graph parameters

Definition of MSOL-definable graph parameters is not in this talk. Examples include:

  • various counting functions for graphs
  • functions recognized by weighted word and tree automata

Finite Rank Theorem (Godlin, Kotek, Makowsky): If a real-valued graph parameter is MSOL-definable, its connection matrix has finite rank.

5/18

slide-19
SLIDE 19

Connection matrices

The k-connection two k-labeled graphs - take their disjoint union and identify similarly labeled vertices. Example:

1 2 3 1 2 3 1 2 3 Two 3-labeled graphs: Their 3-connection: G1 G2 G1G2

The k-connection matrix C f k of a graph parameter f:

  • Infinite matrix with rows and columns

indexed by k-labeled graphs

  • The entry Gi Gj is f GiGj :

C f k Gi Gj f GiGj

G1 Gj G1 . . . Gi . . . f GiGj

6/18

slide-20
SLIDE 20

Connection matrices

The k-connection two k-labeled graphs - take their disjoint union and identify similarly labeled vertices. Example:

1 2 3 1 2 3 1 2 3 Two 3-labeled graphs: Their 3-connection: G1 G2 G1G2

The k-connection matrix C(f, k) of a graph parameter f:

  • Infinite matrix with rows and columns

indexed by k-labeled graphs

  • The entry (Gi, Gj) is f(GiGj):

C(f, k)Gi,Gj = f(GiGj)

G1 . . . Gj . . . G1 . . . Gi . . . f(GiGj)

6/18

slide-21
SLIDE 21

Can we learn MSOL-definable graph parameters?

Finite Rank Theorem (Godlin, Kotek, Makowsky): If a real-valued graph parameter is MSOL-definable, its connection matrix has finite rank. Two obvious differences between this theorem and typical theorems:

  • 1. This is not a characterization theorem
  • 2. The proof does not provide a translation from the matrix to the

parameter Can we do something anyway?

7/18

slide-22
SLIDE 22

Can we learn MSOL-definable graph parameters?

Finite Rank Theorem (Godlin, Kotek, Makowsky): If a real-valued graph parameter is MSOL-definable, its connection matrix has finite rank. Two obvious differences between this theorem and typical theorems:

  • 1. This is not a characterization theorem
  • 2. The proof does not provide a translation from the matrix to the

parameter Can we do something anyway?

7/18

slide-23
SLIDE 23

Can we learn MSOL-definable graph parameters?

Finite Rank Theorem (Godlin, Kotek, Makowsky): If a real-valued graph parameter is MSOL-definable, its connection matrix has finite rank. Two obvious differences between this theorem and typical theorems:

  • 1. This is not a characterization theorem
  • 2. The proof does not provide a translation from the matrix to the

parameter Can we do something anyway?

7/18

slide-24
SLIDE 24

Can we learn MSOL-definable graph parameters?

Finite Rank Theorem (Godlin, Kotek, Makowsky): If a real-valued graph parameter is MSOL-definable, its connection matrix has finite rank. Two obvious differences between this theorem and typical theorems:

  • 1. This is not a characterization theorem
  • 2. The proof does not provide a translation from the matrix to the

parameter Can we do something anyway?

7/18

slide-25
SLIDE 25

Today:

An exact learning algorithm for partition functions

slide-26
SLIDE 26

Partition functions

A partition function is defined by an R-weighted graph H(α, β)

  • α is the vertex-weights function
  • β is the edge-weights function

The partition function hom H counts weighted homomorphisms. For a graph G and a homomorphism t G H,

  • multiply the vertex weights:

v V G

t v

  • multiply the edge weights:

u v V G 2

t u t v And take the sum over all homomorphisms: hom G H

t G H v V G

t v

u v V G 2

t u t v

8/18

slide-27
SLIDE 27

Partition functions

A partition function is defined by an R-weighted graph H(α, β)

  • α is the vertex-weights function
  • β is the edge-weights function

The partition function hom(−, H(α, β)) counts weighted homomorphisms. For a graph G and a homomorphism t : G → H,

  • multiply the vertex weights: ∏

v∈V(G) α(t(v))

  • multiply the edge weights: ∏

(u,v)∈V(G)2 β(t(u), t(v))

And take the sum over all homomorphisms: hom(G, H(α, β)) = ∑

t:G→H

v∈V(G)

α(t(v)) ∏

(u,v)∈V(G)2

β(t(u), t(v))

8/18

slide-28
SLIDE 28

Partition functions

A partition function is defined by an R-weighted graph H(α, β)

  • α is the vertex-weights function
  • β is the edge-weights function

The partition function hom(−, H(α, β)) counts weighted homomorphisms. For a graph G and a homomorphism t : G → H,

  • multiply the vertex weights: ∏

v∈V(G) α(t(v))

  • multiply the edge weights: ∏

(u,v)∈V(G)2 β(t(u), t(v))

And take the sum over all homomorphisms: hom(G, H(α, β)) = ∑

t:G→H

v∈V(G)

α(t(v)) ∏

(u,v)∈V(G)2

β(t(u), t(v))

8/18

slide-29
SLIDE 29

Partition functions

A partition function is defined by an R-weighted graph H(α, β)

  • α is the vertex-weights function
  • β is the edge-weights function

The partition function hom(−, H(α, β)) counts weighted homomorphisms. For a graph G and a homomorphism t : G → H,

  • multiply the vertex weights: ∏

v∈V(G) α(t(v))

  • multiply the edge weights: ∏

(u,v)∈V(G)2 β(t(u), t(v))

And take the sum over all homomorphisms: hom(G, H(α, β)) = ∑

t:G→H

v∈V(G)

α(t(v)) ∏

(u,v)∈V(G)2

β(t(u), t(v))

8/18

slide-30
SLIDE 30

Partition functions

A partition function is defined by an R-weighted graph H(α, β)

  • α is the vertex-weights function
  • β is the edge-weights function

The partition function hom(−, H(α, β)) counts weighted homomorphisms. For a graph G and a homomorphism t : G → H,

  • multiply the vertex weights: ∏

v∈V(G) α(t(v))

  • multiply the edge weights: ∏

(u,v)∈V(G)2 β(t(u), t(v))

And take the sum over all homomorphisms: hom(G, H(α, β)) = ∑

t:G→H

v∈V(G)

α(t(v)) ∏

(u,v)∈V(G)2

β(t(u), t(v))

8/18

slide-31
SLIDE 31

Functions representable as partition functions1

Number of independent sets, whether a graph is Eulerian, number of k-colorings, number of covering edge sets, and other uses in statistical mechanics and approximations of graph properties. Any graph parameter representable as a partition function is MSOL-definable.

1Examples taken from Lovász’s book, Large Networks and Graph Limits

9/18

slide-32
SLIDE 32

Functions representable as partition functions1

Number of independent sets, whether a graph is Eulerian, number of k-colorings, number of covering edge sets,

0.5 0.5 −1 1 1

and other uses in statistical mechanics and approximations of graph properties. Any graph parameter representable as a partition function is MSOL-definable.

1Examples taken from Lovász’s book, Large Networks and Graph Limits

9/18

slide-33
SLIDE 33

Functions representable as partition functions1

Number of independent sets, whether a graph is Eulerian, number of k-colorings, number of covering edge sets, and other uses in statistical mechanics and approximations of graph properties. Any graph parameter representable as a partition function is MSOL-definable.

1Examples taken from Lovász’s book, Large Networks and Graph Limits

9/18

slide-34
SLIDE 34

Functions representable as partition functions1

Number of independent sets, whether a graph is Eulerian, number of k-colorings, number of covering edge sets,

1 −1 1 2 1

and other uses in statistical mechanics and approximations of graph properties. Any graph parameter representable as a partition function is MSOL-definable.

1Examples taken from Lovász’s book, Large Networks and Graph Limits

9/18

slide-35
SLIDE 35

Functions representable as partition functions1

Number of independent sets, whether a graph is Eulerian, number of k-colorings, number of covering edge sets, and other uses in statistical mechanics and approximations of graph properties. Any graph parameter representable as a partition function is MSOL-definable.

1Examples taken from Lovász’s book, Large Networks and Graph Limits

9/18

slide-36
SLIDE 36

Functions representable as partition functions1

Number of independent sets, whether a graph is Eulerian, number of k-colorings, number of covering edge sets, and other uses in statistical mechanics and approximations of graph properties. Any graph parameter representable as a partition function is MSOL-definable.

1Examples taken from Lovász’s book, Large Networks and Graph Limits

9/18

slide-37
SLIDE 37

The setting

  • The learner’s target is some R-weighted graph H(α, β) which

defines the partition function

  • Learner sends VALUE(G) queries and teacher sends back

hom(G, H(α, β))

  • Counterexamples to EQUIVALENT queries are graphs

? ? ? ? ?

hypothesis

α5 α4 α6 α1 α2 α3 β1,2 β2,3 β3,6 β4,6 β1,5 β1,3 β1,6

target

The algorithm relies on results from the theory of graph algebras.

10/18

slide-38
SLIDE 38

The setting

  • The learner’s target is some R-weighted graph H(α, β) which

defines the partition function

  • Learner sends VALUE(G) queries and teacher sends back

hom(G, H(α, β))

  • Counterexamples to EQUIVALENT queries are graphs

? ? ? ? ?

hypothesis

α5 α4 α6 α1 α2 α3 β1,2 β2,3 β3,6 β4,6 β1,5 β1,3 β1,6

target

The algorithm relies on results from the theory of graph algebras.

10/18

slide-39
SLIDE 39

The theoretical backbone

A body of work on the algebraic properties of connection matrices of partition functions, sets up the following result: Theorem [Freedman, Lovász, Schrijver] The k-connection matrix of a rigid partition function2 has rank nk, where n is the size of the weighted graph representing it. Our algorithm utilizes:

  • The existence of a special basis of the space generated by the rows
  • f connection matrices
  • The relationship between this basis and the weighted graph

defining the partition function

2More on this later.

11/18

slide-40
SLIDE 40

The theoretical backbone

A body of work on the algebraic properties of connection matrices of partition functions, sets up the following result: Theorem [Freedman, Lovász, Schrijver] The k-connection matrix of a rigid partition function2 has rank nk, where n is the size of the weighted graph representing it. Our algorithm utilizes:

  • The existence of a special basis of the space generated by the rows
  • f connection matrices
  • The relationship between this basis and the weighted graph

defining the partition function

2More on this later.

11/18

slide-41
SLIDE 41

An overview of the algorithm

  • The algorithm keeps a submatrix of the 1-connection matrix
  • If a counterexample is given, the submatrix is expanded with a new

row and a new column

  • In each iteration, the algorithm:
  • Finds an idempotent basis for the space spanned by the

submatrix

  • Generates a hypothesis from the found basis

12/18

slide-42
SLIDE 42

An overview of the algorithm

  • The algorithm keeps a submatrix of the 1-connection matrix
  • If a counterexample is given, the submatrix is expanded with a new

row and a new column

  • In each iteration, the algorithm:
  • Finds an idempotent basis for the space spanned by the

submatrix

  • Generates a hypothesis from the found basis

12/18

slide-43
SLIDE 43

An overview of the algorithm

  • The algorithm keeps a submatrix of the 1-connection matrix
  • If a counterexample is given, the submatrix is expanded with a new

row and a new column

  • In each iteration, the algorithm:
  • Finds an idempotent basis for the space spanned by the

submatrix

  • Generates a hypothesis from the found basis

12/18

slide-44
SLIDE 44

An overview of the algorithm

  • The algorithm keeps a submatrix of the 1-connection matrix
  • If a counterexample is given, the submatrix is expanded with a new

row and a new column

  • In each iteration, the algorithm:
  • Finds an idempotent basis for the space spanned by the

submatrix

  • Generates a hypothesis from the found basis

12/18

slide-45
SLIDE 45

An overview of the algorithm

  • The algorithm keeps a submatrix of the 1-connection matrix
  • If a counterexample is given, the submatrix is expanded with a new

row and a new column

  • In each iteration, the algorithm:
  • Finds an idempotent basis for the space spanned by the

submatrix

  • Generates a hypothesis from the found basis

12/18

slide-46
SLIDE 46

An overview of the algorithm

  • The algorithm keeps a submatrix of the 1-connection matrix
  • If a counterexample is given, the submatrix is expanded with a new

row and a new column

  • In each iteration, the algorithm:
  • Finds an idempotent basis for the space spanned by the

submatrix

  • Generates a hypothesis from the found basis

12/18

slide-47
SLIDE 47

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT(h)

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M = f(B1B1) . . . f(B1Bm) . . . . . . f(BmB1) . . . f(BmBm)

Bm

1

x Bm

1

x

VALUE queries

p1 pm

1

p1 pm

1

13/18

slide-48
SLIDE 48

Teacher Learner f ∈ C h ∈ C h(x) ̸= f(x)

EQUIVALENT(h)

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M = f(B1B1) . . . f(B1Bm) . . . . . . f(BmB1) . . . f(BmBm)

Bm

1

x Bm

1

x

VALUE queries

p1 pm

1

p1 pm

1

13/18

slide-49
SLIDE 49

Teacher Learner f ∈ C h ∈ C h(x) ̸= f(x)

EQUIVALENT(h)

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M = f(B1B1) . . . f(B1Bm) . . . . . . f(BmB1) . . . f(BmBm)

Bm+1 = x Bm

1

x

VALUE queries

p1 pm

1

p1 pm

1

13/18

slide-50
SLIDE 50

Teacher Learner f ∈ C h ∈ C h(x) ̸= f(x)

EQUIVALENT(h)

augment M with: B1Bm+1, . . . , Bm+1Bm+1, . . . , Bm+1B1

M = f(B1B1) . . . f(B1Bm) . . . . . . f(BmB1) . . . f(BmBm)

Bm

1

x Bm+1 = x

VALUE queries

p1 pm

1

p1 pm

1

13/18

slide-51
SLIDE 51

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm+1, . . . , Bm+1Bm+1, . . . , Bm+1B1

M = f(B1B1) . . . f(B1Bm) . . . . . . f(BmB1) . . . f(BmBm)

Bm

1

x Bm+1 = x

VALUE queries

p1 pm

1

p1 pm

1

13/18

slide-52
SLIDE 52

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

p1 pm

1

p1 pm

1

13/18

slide-53
SLIDE 53

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

find basis p1 pm

1

p1 pm

1

13/18

slide-54
SLIDE 54

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

find basis p1 pm

1

p1 pm

1

VALUE queries

13/18

slide-55
SLIDE 55

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

find basis p1, . . . , pm+1 p1 pm

1

13/18

slide-56
SLIDE 56

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

find basis p1 pm

1

p1, . . . , pm+1 generate hypothesis

13/18

slide-57
SLIDE 57

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

find basis p1 pm

1

p1, . . . , pm+1 generate hypothesis

VALUE queries

13/18

slide-58
SLIDE 58

Teacher Learner f ∈ C h ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

find basis p1 pm

1

p1, . . . , pm+1 generate hypothesis A weighted graph h′

13/18

slide-59
SLIDE 59

Teacher Learner f ∈ C h′ ∈ C h x f x

EQUIVALENT h

augment M with: B1Bm

1

Bm

1Bm 1

Bm

1B1

M =

Bm

1

x Bm

1

x

VALUE queries

f(B1B1) . . . f(B1Bm) f(B1Bm+1) . . . . . . . . . f(BmB1) . . . f(BmBm) f(BmBm+1) f(Bm+1B1) . . . f(Bm+1Bm) f(Bm+1Bm+1)

p1 pm

1

p1 pm

1

13/18

slide-60
SLIDE 60

Complexity

We are over the Blum-Shub-Smale model of computation over the reals

  • Real numbers are stored in single memory cells
  • Arithmetic operations are performed in a single step

For a target with n vertices, the rank of the 1-connection matrix is n. The algorithm has:

  • O n iterations
  • O n2 VALUE queries
  • O n EQUIVALENT queries
  • Solves systems of linear equation – done in POLYTIME n .

14/18

slide-61
SLIDE 61

Complexity

We are over the Blum-Shub-Smale model of computation over the reals

  • Real numbers are stored in single memory cells
  • Arithmetic operations are performed in a single step

For a target with n vertices, the rank of the 1-connection matrix is n. The algorithm has:

  • O(n) iterations
  • O(n2) VALUE queries
  • O(n) EQUIVALENT queries
  • Solves systems of linear equation – done in POLYTIME(n).

14/18

slide-62
SLIDE 62

Limitation to rigid graphs

Our algorithm learns rigid partition functions - the weighted graph has no proper automorphisms.

  • Almost all graphs are rigid: the probability that a uniformly chosen

graph with n vertices is rigid tends to 1 as n → ∞ Recall the algorithm keeps a submatrix of the 1-connection matrix.

  • Lifting the rigidity restriction would require a k-connection

submatrix for unknown k

  • If we can find the correct k quickly enough - the same algorithm

should apply

15/18

slide-63
SLIDE 63

Limitation to rigid graphs

Our algorithm learns rigid partition functions - the weighted graph has no proper automorphisms.

  • Almost all graphs are rigid: the probability that a uniformly chosen

graph with n vertices is rigid tends to 1 as n → ∞ Recall the algorithm keeps a submatrix of the 1-connection matrix.

  • Lifting the rigidity restriction would require a k-connection

submatrix for unknown k

  • If we can find the correct k quickly enough - the same algorithm

should apply

15/18

slide-64
SLIDE 64

Limitation to rigid graphs

Our algorithm learns rigid partition functions - the weighted graph has no proper automorphisms.

  • Almost all graphs are rigid: the probability that a uniformly chosen

graph with n vertices is rigid tends to 1 as n → ∞ Recall the algorithm keeps a submatrix of the 1-connection matrix.

  • Lifting the rigidity restriction would require a k-connection

submatrix for unknown k

  • If we can find the correct k quickly enough - the same algorithm

should apply

15/18

slide-65
SLIDE 65

Limitation to rigid graphs

Our algorithm learns rigid partition functions - the weighted graph has no proper automorphisms.

  • Almost all graphs are rigid: the probability that a uniformly chosen

graph with n vertices is rigid tends to 1 as n → ∞ Recall the algorithm keeps a submatrix of the 1-connection matrix.

  • Lifting the rigidity restriction would require a k-connection

submatrix for unknown k

  • If we can find the correct k quickly enough - the same algorithm

should apply

15/18

slide-66
SLIDE 66

How powerful does the Teacher need to be?

The teacher in exact learning is assumed to be all-powerful. In reality...

  • Answering VALUE queries is generally #P-hard
  • POLYTIME on graphs of bounded tree-width

Can we guarantee VALUE queries will only be on such graphs? When?

  • Answering EQUIVALENT queries:
  • Counterexamples are guaranteed to be small!
  • f size

2 1 n2 n6

  • Whether the answer is YES reduces to the (weighted) graph

isomorphism problem

16/18

slide-67
SLIDE 67

How powerful does the Teacher need to be?

The teacher in exact learning is assumed to be all-powerful. In reality...

  • Answering VALUE queries is generally #P-hard
  • POLYTIME on graphs of bounded tree-width

Can we guarantee VALUE queries will only be on such graphs? When?

  • Answering EQUIVALENT queries:
  • Counterexamples are guaranteed to be small!
  • f size

2 1 n2 n6

  • Whether the answer is YES reduces to the (weighted) graph

isomorphism problem

16/18

slide-68
SLIDE 68

How powerful does the Teacher need to be?

The teacher in exact learning is assumed to be all-powerful. In reality...

  • Answering VALUE queries is generally #P-hard
  • POLYTIME on graphs of bounded tree-width

Can we guarantee VALUE queries will only be on such graphs? When?

  • Answering EQUIVALENT queries:
  • Counterexamples are guaranteed to be small!
  • f size

2 1 n2 n6

  • Whether the answer is YES reduces to the (weighted) graph

isomorphism problem

16/18

slide-69
SLIDE 69

How powerful does the Teacher need to be?

The teacher in exact learning is assumed to be all-powerful. In reality...

  • Answering VALUE queries is generally #P-hard
  • POLYTIME on graphs of bounded tree-width

Can we guarantee VALUE queries will only be on such graphs? When?

  • Answering EQUIVALENT queries:
  • Counterexamples are guaranteed to be small!
  • f size

2 1 n2 n6

  • Whether the answer is YES reduces to the (weighted) graph

isomorphism problem

16/18

slide-70
SLIDE 70

How powerful does the Teacher need to be?

The teacher in exact learning is assumed to be all-powerful. In reality...

  • Answering VALUE queries is generally #P-hard
  • POLYTIME on graphs of bounded tree-width

Can we guarantee VALUE queries will only be on such graphs? When?

  • Answering EQUIVALENT queries:
  • Counterexamples are guaranteed to be small!
  • f size ≤ 2(1 + n2)n6
  • Whether the answer is YES reduces to the (weighted) graph

isomorphism problem

16/18

slide-71
SLIDE 71

Summary and further work

  • We suggested the study of exact learnability of graph parameters
  • Presented an exact learning algorithm for rigid partition functions

as proof-of-concept Future work:

  • Lift the rigidity restriction
  • Investigate requirements from the teacher
  • Learning algorithms for other classes of graph parameters

17/18

slide-72
SLIDE 72

Summary and further work

  • We suggested the study of exact learnability of graph parameters
  • Presented an exact learning algorithm for rigid partition functions

as proof-of-concept Future work:

  • Lift the rigidity restriction
  • Investigate requirements from the teacher
  • Learning algorithms for other classes of graph parameters

17/18

slide-73
SLIDE 73

Summary and further work

  • We suggested the study of exact learnability of graph parameters
  • Presented an exact learning algorithm for rigid partition functions

as proof-of-concept Future work:

  • Lift the rigidity restriction
  • Investigate requirements from the teacher
  • Learning algorithms for other classes of graph parameters

17/18

slide-74
SLIDE 74

References

  • D. Angluin.

Queries and concept learning. Machine Learning, 2(4):319–342, 1987.

  • L. Lovász.

Large Networks and Graph Limits, volume 60 of Colloquium Publications. AMS, 2012. 18/18