tecto to amr and translation
play

Tecto to AMR and translation Ond rej Bojar, Silvie Cinkov a, Ond - PowerPoint PPT Presentation

Tecto to AMR and translation Ond rej Bojar, Silvie Cinkov a, Ond rej Du sek, Tim OGorman, Martin Popel, Roman Sudarikov, Zde nka Ure sov a August 1, 2014 1 / 24 Introduction 2 / 24 Motivation We are


  1. Tecto to AMR and translation Ondˇ rej Bojar, Silvie Cinkov´ a, Ondˇ rej Duˇ sek, Tim O’Gorman, Martin Popel, Roman Sudarikov, Zdeˇ nka Ureˇ sov´ a August 1, 2014 1 / 24

  2. Introduction 2 / 24

  3. Motivation ◮ We are investigating the value of parallel Abstract Meaning Representations (AMRs) ◮ Question 1: How similar are AMRs made in different languages? How do you compare them? ◮ Question 2: How could we get a large corpus of parallel AMRs? 3 / 24

  4. AMRICA ◮ (AMR Inspector with Cross-language Alignment) ◮ Usual evaluation and alignment methods break across languages. ◮ Extension to Smatch (Cai & Knight 2012). 4 / 24

  5. Smatch Classic 5 / 24

  6. Smatch Classic 6 / 24

  7. Smatch Classic 7 / 24

  8. AMRICA 8 / 24

  9. AMRICA 9 / 24

  10. T-layer to AMR conversion ◮ PCEDT: Large parallel corpus (WSJ) annotated with t-layer for English and Czech ◮ T-layer to AMR conversion would provide a large static parallel AMR corpus. ◮ Could be used dynamically to turn a ”t-layer” parser into an AMR parser. 10 / 24

  11. Why this might work ◮ AMR and t-layer are very similar: ◮ Both abstract away from syntax. ◮ Both make all semantic links in a sentence in a graph format. ◮ Both do coreference ◮ Various minor structural differences. ◮ AMR is more abstract, makes more inference. 11 / 24

  12. “Peter is eager to please” be.ENUNC eager-41 PAT arg1 eager arg0 please-01 PAT arg0 ACT please person PAT ACT name #Gen #Cor name coreference op1 Peter Peter 12 / 24

  13. Merging of Coreferent Nodes eager-41 be.ENUNC arg1 PAT arg0 please-01 eager arg0 person ACT PAT name please name ACT PAT op1 Peter #Gen Peter 13 / 24

  14. Elimination of semantically light words eager-41 eager arg1 arg0 please-01 PAT arg0 ACT please person name ACT name op1 Peter Peter 14 / 24

  15. Semantic Roles and Senses eager-41 eager-41 arg1 arg0 please-01 arg1 arg0 arg0 please-01 person name arg0 name op1 Peter Peter 15 / 24

  16. Add Named Entities eager-41 eager-41 arg1 arg1 arg0 please-01 arg0 please-01 arg0 arg0 person person name name name name op1 op1 Peter Peter 16 / 24

  17. Conversion Procedures ◮ Converted t-trees to AMR format ◮ Added named entities using NER systems (Stanford and NameTag) ◮ Tried two strategies for doing more complex changes to the graphs: ◮ PML-TQ ◮ Tsurgeon ◮ List-based verbalization and semantic role mapping 17 / 24

  18. PML-TQ rules ◮ Based on AMR guidelines (generalized) ◮ For copula, attributes, non-core roles . . . A PML-TQ rule LHS (PML-TQ Query) conditions on a t-subtree t-node Ib_DEL RHS (AMR Subtree) t_lemma{in{{}be},{}become},{}remain}} r conditions on surface t-node Ib2 t-node Ir ARG0 ARG1 functor{={}ACT} functor{={}PAT} formeme{~{}n:.*} formeme{={}adj:compl} a-node tag{={}IN} b2 w t-node Iw functor{={}PAT} Guidelines example: The boy is responsible for the work. 18 / 24

  19. PML-TQ rules Rule application t-tree zone=en n/nervous Conversion result ARG0 ARG1 be.enunc p2/person p/presentation PREDv:fin was name Ondrej nervous n2/name ACTn:subj PATadj:compl Ondrej nervous op1 Matching t-tree presentation "Ondrej" PATn:about+X about the presentation Matching sentence: Ondrej was nervous about the presentation. 19 / 24

  20. Tsurgeon tree transformation rules ◮ We converted to constituency trees so as to use a tree tranformation tool, Tsurgeon (Levy and Andrews 2006) to quickly implement hand-written rules. 20 / 24

  21. Tsurgeon tree transformation rules ◮ Many of the structural differences are just notational differences: eat eat CONJ arg1 and PAT PAT and op1 op2 banana apple banana apple 21 / 24

  22. List-based Methods ◮ Verbalizations are based on dictionary look-ups: ◮ beekeeper → person :ARG0-of keep-01 :ARG1 bee ◮ As are complex predications: give ACT CPHR bless John blessing APP ACT PAT Mary John Mary 22 / 24

  23. Using Existing Resources Other WSJ annotation Propbank Lexicon Lexical Lists Vallex Map t-layer roles to AMR roles X X X Verbalize nouns/adjectives X X Introduce inferrable predicates X Named Entities X X 23 / 24

  24. Results of EN t-to-AMR Conv Semantic Role Mapping Smatch w/o senses Verbalization Lists Named Entities Smatch Baseline (direct conversion) 20 28 Baseline (direct conversion) X 33 41 Baseline (direct conversion) X X 37 45 Baseline (direct conversion) X X X 40 48 PML-TQ (guidelines-based) X X 35 43 PML-TQ (guidelines-based) X X X 38 47 Tsurgeon (rule-based) X X X 44 52 JAMR 44 45 24 / 24

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend