Tree Transducers and Tree Adjoining Grammars Historical and Current - - PDF document

tree transducers and tree adjoining grammars historical
SMART_READER_LITE
LIVE PREVIEW

Tree Transducers and Tree Adjoining Grammars Historical and Current - - PDF document

Tree Transducers and Tree Adjoining Grammars Historical and Current Perspectives William C. Rounds University of Michigan, Ann Arbor 1 Outline Some history genesis of tree transducers and tree grammars A little bit on the genesis of


slide-1
SLIDE 1

Tree Transducers and Tree Adjoining Grammars Historical and Current Perspectives William C. Rounds University of Michigan, Ann Arbor

1

slide-2
SLIDE 2

Outline

  • Some history – genesis of tree transducers and tree grammars
  • A little bit on the genesis of feature logic
  • A preliminary attempt to unify transducers, TAGs, and feature

logic

  • A few questions about ongoing work

2

slide-3
SLIDE 3

The 60’s A new religion is born...

3

slide-4
SLIDE 4

The Master

4

slide-5
SLIDE 5

I become a disciple

5

slide-6
SLIDE 6

The Peters-Ritchie result Transformational grammars are Turing-powerful End times for TG?

6

slide-7
SLIDE 7

TG survives! It begins series of mutations into GB, PP, MIN, REC

7

slide-8
SLIDE 8

A failed Math PhD?

8

slide-9
SLIDE 9

Go into computer science, but call it math

9

slide-10
SLIDE 10

Problem

  • Go beyond context-free (reduplication phenomena)
  • Have recursion
  • Avoid Turing-powerful
  • Have a vague resemblance to transformational grammar

10

slide-11
SLIDE 11

Tree automata – salvation!

11

slide-12
SLIDE 12

Thatcher, Wright, Brainerd, Doner, Rabin

  • Tree automaton - recognizable = context-free.
  • Top-down infinite tree automata emptiness is decidable
  • Idea - use top-down to define tree transductions
  • Reinforced by being able to model syntax-directed translation (not

for NL!)

12

slide-13
SLIDE 13

Top-down tree transduction

13

slide-14
SLIDE 14

Further developments

  • Tree transducers which could delay (create their own input), now

called macro tree transducers

  • One-state macro tree transducer = CFG on trees
  • Santa Cruz 1970 – Thatcher, Joshi, Peters, Petrick, Partee, Bach:

Tree Mappings in Linguistics

  • birth of TAGs, aka linear CFGs on trees
  • Natural generalization to graph grammars, links to term rewriting
  • Burgeoning industry in Europe, led by Engelfriet

14

slide-15
SLIDE 15

What was I doing? (1976-1983)

  • Some results on complexity
  • Modelling semantics of concurrency
  • Ignoring tree transducers
  • Learning about bisimulations and modal logic

15

slide-16
SLIDE 16

Learn from your graduate students

  • Bob Kasper (1983-4): What is a disjunctive feature structure?
  • How should these be unified?
  • Write down desired laws for distributing unification over disjunc-

tion

  • With the background of modal logics for concurrency, realize that

feature structures are models for feature logic.

  • PATR-2 actually invents feature logic; we extend to modal version.
  • Make big mistakes proving a completeness theorem.
  • Drew Moshier (1986-7), Larry Moss, Bob Carpenter fix things

16

slide-17
SLIDE 17

Skip to the near present

  • Probabilities, statistics, and corpora
  • Resurgence of (weighted) finite-state transducers on strings as uni-

fying model for speech recognition and generation algorithms (Mohri, Pereira, Riley)

  • Kevin Knight and students propose probabilistic tree transducers

as schemas for MT algorithms

  • Multiplicity of tree transducer models (e.g., semilinear non-deleting

deterministic, with inherited attributes and right-regular looka- head)

  • Can we take any of these off the shelf and actually use them?

17

slide-18
SLIDE 18

Two directions

  • Use linguistic evidence to select relevant class of models (this work-

shop)

  • Use various mathematical means to understand commonalities and

differences among models

  • Shieber: synchronous TAGS and tree transducers
  • Rogers: TAGS as 3D tree automata
  • Take a break from inventing the next variation

18

slide-19
SLIDE 19

Model-theoretic syntax

  • Long tradition of regarding generation as proof, even parsing as

proof

  • In last ten years: what is the model theory for these proof systems?
  • Best-known example: Montague grammar (focus on interpreta-

tion)

  • Now: type-logical syntax (Morrill, Moortgat) and type-logical se-

mantics (Carpenter)

  • Feature logic is another description language for syntax.
  • Attempts to view grammatical derivations as proofs, usually in

logic programs with feature logic as a constraint language.

  • HPSG: fully developed linguistic theory grounded in feature de-

scriptions and unification; grammars as logical constraints on fea- ture structures.

19

slide-20
SLIDE 20

Clean proof theory and accompanying model theory for feature logic ?

  • Incomplete and ongoing work
  • Goals: self-contained proof theory (do not glue onto grammar)
  • Logic should model common grammatical formalisms, to under-

stand them better

  • Some previous work: Keller (extending feature logic to model

TAGs); Vijay-Shanker and Joshi (FTAGs)

20

slide-21
SLIDE 21

Three-dimensional trees (Rogers)

  • 1 2

1, 0 1, 1 1, 0 1, 1 1, 2, 2, 0

21

slide-22
SLIDE 22

Adjunction as a 3D tree

Initial tree Auxiliary trees π

22

slide-23
SLIDE 23

Adjunction as a 3D FS

a b

c a b

a b

3 3

a b

≡ ≡ π b ≡ 3πb a ≡ 3πa Initial tree Auxiliary tree Substitution

23

slide-24
SLIDE 24

Example of Adjunction Rule

n[c:⊥] → {3:n[l:adj[c:pretty], r:n[c:⊥]] (c . = 3rc), 3:n[c:⊥] (c . = 3c)}

np n boy the d adj n pretty l r n c c adj pretty l r n c n boy c ≡ 3 Choice

24

slide-25
SLIDE 25

Rules as logical constraints

np n boy the d adj pretty l r n c n boy c ≡ np n boy the d

n[c:⊥] → {3:n[l:adj[c:pretty], r:n[c:⊥]] (c . = 3rc), 3:n[c:⊥] (c . = 3c)}

Not a model Not a model, either (still unexpanded ns) 3 25

slide-26
SLIDE 26

Quick look at tree transductions

c

b c a b

l l r r

q

in q[in :c[l:⊥, r:⊥]] → 3:c[l:p, r:q] [inr . = 3lin] [inl . = 3rin]

26

slide-27
SLIDE 27

Quick look at tree transductions

l c

b c a b

l l r r

q

c

b c a b

l r r l r

q

p 3 in in in in

q c q[in :c[l:⊥, r:⊥]] → 3:c[l:p, r:q] [inr . = 3lin] [inl . = 3rin] q[in :c[l:⊥, r:⊥]] → 3:c[l:p, r:q] [inr . = 3rin] [inl . = 3lin] 27

slide-28
SLIDE 28

Quick look at tree transductions

l c

b c a b

l l r r

q

c

b c a b

l r r l r q p 3 in in in in 3 q

c b c a b

l r r l r

c c

l r p l in in in in in 3 a 3

p[in :b] → 3:a

p q

q c

3 3

a

b

q q[in :c[l:⊥, r:⊥]] → 3:c[l:p, r:q] [inr . = 3rin] [inl . = 3lin] q[in :c[l:⊥, r:⊥]] → 3:c[l:p, r:q] [inr . = 3lin] [inl . = 3rin] 28

slide-29
SLIDE 29

Theory behind this

  • Disjunctive feature logic programming.
  • A program is set of rules of the formf → L, where f is a feature

structure, and L is a clause, a finite set of feature structures.

  • These rules can be used in proofs, to create a theory, in general

an infinite set of clauses.

  • A feature structure m satisfies the clause L if some element of L

subsumes it.

  • m is a model of the program if for any rule f → L, if f subsumes

m, then m satisfies L.

  • Theorem: the minimal models of the program are the minimal

structures satisfying all clauses of the theory.

  • This way we can get infinite FS as models.
  • There is a sound and complete resolution proof system to go with

all of this.

29

slide-30
SLIDE 30

The resolution rules

  • Logical resolution:

K L f ∈ K g ∈ L f ⊔ g | = M M ∪ (K \ {f}) ∪ (L \ {g}) where K, L, M are clauses.

  • Clause introduction (nonlogical resolution):

M g ∈ M f → L ∈ P f ⊑ g L ∪ (M \ {g})

30

slide-31
SLIDE 31

Questions and further work

  • Can you compile FL specifications into a parser?
  • What about other formalisms, like synchronous TAGS?
  • Can you work probabilities, or more generally, weights, into a fully

declarative formalism?

31

slide-32
SLIDE 32

Thanks!

32