tree transducers and tree adjoining grammars historical
play

Tree Transducers and Tree Adjoining Grammars Historical and Current - PDF document

Tree Transducers and Tree Adjoining Grammars Historical and Current Perspectives William C. Rounds University of Michigan, Ann Arbor 1 Outline Some history genesis of tree transducers and tree grammars A little bit on the genesis of


  1. Tree Transducers and Tree Adjoining Grammars Historical and Current Perspectives William C. Rounds University of Michigan, Ann Arbor 1

  2. Outline • Some history – genesis of tree transducers and tree grammars • A little bit on the genesis of feature logic • A preliminary attempt to unify transducers, TAGs, and feature logic • A few questions about ongoing work 2

  3. The 60’s A new religion is born... 3

  4. The Master 4

  5. I become a disciple 5

  6. The Peters-Ritchie result Transformational grammars are Turing-powerful End times for TG? 6

  7. TG survives! It begins series of mutations into GB, PP, MIN, REC 7

  8. A failed Math PhD? 8

  9. Go into computer science, but call it math 9

  10. Problem • Go beyond context-free (reduplication phenomena) • Have recursion • Avoid Turing-powerful • Have a vague resemblance to transformational grammar 10

  11. Tree automata – salvation! 11

  12. Thatcher, Wright, Brainerd, Doner, Rabin • Tree automaton - recognizable = context-free. • Top-down infinite tree automata emptiness is decidable • Idea - use top-down to define tree transductions • Reinforced by being able to model syntax-directed translation (not for NL!) 12

  13. Top-down tree transduction 13

  14. Further developments • Tree transducers which could delay (create their own input), now called macro tree transducers • One-state macro tree transducer = CFG on trees • Santa Cruz 1970 – Thatcher, Joshi, Peters, Petrick, Partee, Bach: Tree Mappings in Linguistics • birth of TAGs, aka linear CFGs on trees • Natural generalization to graph grammars, links to term rewriting • Burgeoning industry in Europe, led by Engelfriet 14

  15. What was I doing? (1976-1983) • Some results on complexity • Modelling semantics of concurrency • Ignoring tree transducers • Learning about bisimulations and modal logic 15

  16. Learn from your graduate students • Bob Kasper (1983-4): What is a disjunctive feature structure? • How should these be unified? • Write down desired laws for distributing unification over disjunc- tion • With the background of modal logics for concurrency, realize that feature structures are models for feature logic. • PATR-2 actually invents feature logic; we extend to modal version. • Make big mistakes proving a completeness theorem. • Drew Moshier (1986-7), Larry Moss, Bob Carpenter fix things 16

  17. Skip to the near present • Probabilities, statistics, and corpora • Resurgence of (weighted) finite-state transducers on strings as uni- fying model for speech recognition and generation algorithms (Mohri, Pereira, Riley) • Kevin Knight and students propose probabilistic tree transducers as schemas for MT algorithms • Multiplicity of tree transducer models (e.g., semilinear non-deleting deterministic, with inherited attributes and right-regular looka- head) • Can we take any of these off the shelf and actually use them? 17

  18. Two directions • Use linguistic evidence to select relevant class of models (this work- shop) • Use various mathematical means to understand commonalities and differences among models • Shieber: synchronous TAGS and tree transducers • Rogers: TAGS as 3D tree automata • Take a break from inventing the next variation 18

  19. Model-theoretic syntax • Long tradition of regarding generation as proof, even parsing as proof • In last ten years: what is the model theory for these proof systems? • Best-known example: Montague grammar (focus on interpreta- tion) • Now: type-logical syntax (Morrill, Moortgat) and type-logical se- mantics (Carpenter) • Feature logic is another description language for syntax. • Attempts to view grammatical derivations as proofs, usually in logic programs with feature logic as a constraint language. • HPSG: fully developed linguistic theory grounded in feature de- scriptions and unification; grammars as logical constraints on fea- ture structures. 19

  20. Clean proof theory and accompanying model theory for feature logic ? • Incomplete and ongoing work • Goals: self-contained proof theory (do not glue onto grammar) • Logic should model common grammatical formalisms, to under- stand them better • Some previous work: Keller (extending feature logic to model TAGs); Vijay-Shanker and Joshi (FTAGs) 20

  21. Three-dimensional trees (Rogers) � � � � �� 1 �� �� 2 �� �� 2 � , � � �� 0 �� �� 1 � , � � �� 2 � , � 0 �� �� 1 , 1 �� �� 1 � , � 1 �� �� 1 , 0 �� �� 1 � , � 0 �� 21

  22. Adjunction as a 3D tree Initial tree π Auxiliary trees 22

  23. Adjunction as a 3D FS 3 Initial tree c b a a ≡ 3 π a b ≡ 3 πb Auxiliary tree 3 b a b a π ≡ b a ≡ Substitution 23

  24. Example of Adjunction Rule n l r adj n np c pretty n d 3 c n boy the l r adj n ≡ c c boy pretty n [ c : ⊥ ] → { 3: n [ l : adj [ c :pretty] , r : n [ c : ⊥ ]] � ( c . = 3 rc ) , 3: n [ c : ⊥ ] � ( c . = 3 c ) } Choice 24

  25. Rules as logical constraints np n d Not a model boy the np Not a model, either (still unexpanded n s) n d 3 n boy the l r adj n ≡ c c pretty boy n [ c : ⊥ ] → { 3: n [ l : adj [ c :pretty] , r : n [ c : ⊥ ]] � ( c . = 3 rc ) , 3: n [ c : ⊥ ] � ( c . = 3 c ) } 25

  26. Quick look at tree transductions c in q l r c b q [ in : c [ l : ⊥ , r : ⊥ ]] → 3: c [ l : p, r : q ] � [ in r . = 3 l in ] � [ in l . = 3 r in ] l r b a 26

  27. Quick look at tree transductions c in q l r c b q [ in : c [ l : ⊥ , r : ⊥ ]] → 3: c [ l : p, r : q ] � [ in r . = 3 l in ] � [ in l . = 3 r in ] l r b a c in q 3 c l l r r in c b in q p q [ in : c [ l : ⊥ , r : ⊥ ]] → 3: c [ l : p, r : q ] � [ in r . = 3 r in ] � [ in l . = 3 l in ] l r b a 27

  28. Quick look at tree transductions c in q l r c b q [ in : c [ l : ⊥ , r : ⊥ ]] → 3: c [ l : p, r : q ] � [ in r . = 3 l in ] � [ in l . = 3 r in ] l r b a in c q 3 c l r l in r c b in q p q [ in : c [ l : ⊥ , r : ⊥ ]] → 3: c [ l : p, r : q ] � [ in r . = 3 r in ] � [ in l . = 3 l in ] l r q in c 3 b a c p [ in : b ] → 3: a l l in r r 3 q p c b in 3 c in a r r l l q p in 3 b 3 a b a 28

  29. Theory behind this • Disjunctive feature logic programming. • A program is set of rules of the form f → L , where f is a feature structure, and L is a clause , a finite set of feature structures. • These rules can be used in proofs, to create a theory , in general an infinite set of clauses. • A feature structure m satisfies the clause L if some element of L subsumes it. • m is a model of the program if for any rule f → L , if f subsumes m , then m satisfies L . • Theorem : the minimal models of the program are the minimal structures satisfying all clauses of the theory. • This way we can get infinite FS as models. • There is a sound and complete resolution proof system to go with all of this. 29

  30. The resolution rules • Logical resolution: K L f ∈ K g ∈ L f ⊔ g | = M M ∪ ( K \ { f } ) ∪ ( L \ { g } ) where K, L, M are clauses. • Clause introduction (nonlogical resolution): M g ∈ M f → L ∈ P f ⊑ g L ∪ ( M \ { g } ) 30

  31. Questions and further work • Can you compile FL specifications into a parser? • What about other formalisms, like synchronous TAGS? • Can you work probabilities, or more generally, weights, into a fully declarative formalism? 31

  32. Thanks! 32

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend