tree adjoining grammar parsing and vector representations
play

Tree-Adjoining Grammar Parsing and Vector Representations of - PowerPoint PPT Presentation

Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Tree-Adjoining Grammar Parsing and Vector Representations of Supertags Jungo Kasai Yale University December 14, 2017


  1. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Tree-Adjoining Grammar Parsing and Vector Representations of Supertags Jungo Kasai Yale University December 14, 2017

  2. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Outline Background and Motivations 1 Supertagging Models 2 Parsing Models 3 Vector Representations of Supertags 4 Ongoing TAG Parsing Work 5 Applications of TAG 6 Future Work 7

  3. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Outline Background and Motivations 1 Supertagging Models 2 Parsing Models 3 Vector Representations of Supertags 4 Ongoing TAG Parsing Work 5 Applications of TAG 6 Future Work 7

  4. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Syntactic Parsing S NP VP John AdvP VP really V NP likes Mary Why do we need parsing? Does John love Mary? Does Mary love John? Understanding of a sentence depends on the structure

  5. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Context Free Grammars S S → NP VP VP → AdvP VP NP VP AdvP → really VP → V NP John AdvP VP NP → Mary NP → they really V NP NP → John V → like likes Mary V → likes These production rules generate sentences

  6. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Context Free Grammars S S → NP VP VP → AdvP VP NP VP AdvP → really VP → V NP John AdvP VP NP → Mary NP → they really V NP NP → John V → like likes Mary V → likes Fundamental problem : constraints are distributed over separate rules How do we choose V → like or V → likes ?

  7. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Tree-Adjoining Grammar Tree-Adjoining Grammar (TAG) localizes grammatical constraints Finite set of lexicalized elementary trees Finite set of operations (Substitution and Adjunction) are used to combine elementary trees S VP S VP * NP 0 ↓ VP AdvP NP 0 ↓ VP NP Ad ♦ V ♦ V ♦ NP 1 ↓ N ♦ sleep really likes John

  8. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Tree-Adjoining Grammar Substitution

  9. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Tree-Adjoining Grammar Adjunction

  10. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Tree-Adjoining Grammar Adjunction allows for unbounded recursion while still enforcing agreement. John smartly occasionally really only likes Mary...

  11. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Derivation Tree Derivation tree records the operations. Forms a dependency tree (each token has exactly one parent) likes Subst 0 AdjSubst 1 John really Mary ROOT Subst 0 ADJ Subst 1 ROOT John really likes Mary

  12. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Two Steps in TAG Parsing Now the reverse process. Supertagging Assign elementary trees (supertags) to each token. Similar to POS tagging. Parsing Predict operations on the elementary trees. NP * NP * S S NP 0 ↓ S NP 0 ↓ VP NP VP V ♦ -NONE- V ♦ left left

  13. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Outline Background and Motivations 1 Supertagging Models 2 Parsing Models 3 Vector Representations of Supertags 4 Ongoing TAG Parsing Work 5 Applications of TAG 6 Future Work 7

  14. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Supertagging is a bottleneck Supertagger Parser Stag Acc UAS LAS Gold Chart (MICA) 100.00 97.60 97.30 Maxent (MICA) Chart (MICA) 88.52 87.60 85.80 Supertagging is almost parsing There are about 5,000 supertags in the grammar About half of them occur only once in the training data (PTB WSJ Sections 1-22).

  15. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W BiLSTM Supertagging Figure: BiLSTM Supertagger Architecture.

  16. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Supertagging is still a bottleneck Supertagger Parser Stag Acc UAS LAS Maxent (MICA) Chart (MICA) 88.52 BiLSTM Chart (MICA) 89.32

  17. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Supertagging is still a bottleneck Supertagger Parser Stag Acc UAS LAS Gold Chart (MICA) 100.00 97.60 97.30 Maxent (MICA) Chart (MICA) 88.52 87.60 85.80 BiLSTM Chart (MICA) 89.32 90.05 88.32 We can compensate for supertagging errors by exploiting structural similarities across elementary trees. Similarities across supertags are not utilized by the chart parser. We use two alternative families of parsing algorithms

  18. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Outline Background and Motivations 1 Supertagging Models 2 Parsing Models 3 Vector Representations of Supertags 4 Ongoing TAG Parsing Work 5 Applications of TAG 6 Future Work 7

  19. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Parsing Models Prior Work: Unlexicalized Chart-Parser (MICA) [Bangalore et al., 2009] Unlexicalized Transition-based Parser [Kasai et al., 2017, Friedman et al., 2017] Graph-based Parser (work in progress)

  20. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing Arc-Eager System (MALT) [Nivre et al., 2006] ROOT Subst 0 ADJ Subst1 ROOT John really likes Mary

  21. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  22. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  23. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  24. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  25. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  26. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  27. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based Parsing

  28. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Transition-based TAG Parsing How do we learn? Represent the configuration by the top k elements from stack and buffer: { s i , b i } k i = 1 [Chen and Manning, 2014]. Represent s i ( b i ) by the TAG elementary tree and the derived substitution operations performed into s i . Encode the TAG elementary trees and the substitution operations with dense vectors.

  29. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W NN Transition-based Parsing Model Figure: Transition-based Parser Neural Network Architecture.

  30. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Example John really likes Mary Stack Buffer Relations Action ROOT likes Mary { ( ROOT , likes , ROOT ) , ( likes , John , 0 ) · · · } RIGHT:1

  31. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Parsing Results Gold Stags Predicted Stags Parsing Model UAS LAS UAS LAS MICA Chart 97.60 97.30 90.05 88.32 Transition-based 97.67 97.45 90.23 88.77 Table: Results on Section 00. Beam size 16. Predicted supertags are from our BiLSTM supertagger.

  32. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Outline Background and Motivations 1 Supertagging Models 2 Parsing Models 3 Vector Representations of Supertags 4 Ongoing TAG Parsing Work 5 Applications of TAG 6 Future Work 7

  33. Background and Motivations Supertagging Models Parsing Models Vector Representations of Supertags Ongoing TAG Parsing W Embeddings for Elementary Trees

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend