Tecto to AMR and translation Ond rej Bojar, Silvie Cinkov a, Ond - - PowerPoint PPT Presentation

tecto to amr and translation
SMART_READER_LITE
LIVE PREVIEW

Tecto to AMR and translation Ond rej Bojar, Silvie Cinkov a, Ond - - PowerPoint PPT Presentation

Tecto to AMR and translation Ond rej Bojar, Silvie Cinkov a, Ond rej Du sek, Tim OGorman, Martin Popel, Roman Sudarikov, Zde nka Ure sov a August 1, 2014 1 / 24 Introduction 2 / 24 Motivation We are


slide-1
SLIDE 1

Tecto to AMR and translation

Ondˇ rej Bojar, Silvie Cinkov´ a, Ondˇ rej Duˇ sek, Tim O’Gorman, Martin Popel, Roman Sudarikov, Zdeˇ nka Ureˇ sov´ a August 1, 2014

1 / 24

slide-2
SLIDE 2

Introduction

2 / 24

slide-3
SLIDE 3

Motivation

◮ We are investigating the value of parallel

Abstract Meaning Representations (AMRs)

◮ Question 1: How similar are AMRs

made in different languages? How do you compare them?

◮ Question 2: How could we get a large

corpus of parallel AMRs?

3 / 24

slide-4
SLIDE 4

AMRICA

◮ (AMR Inspector with Cross-language

Alignment)

◮ Usual evaluation and alignment methods

break across languages.

◮ Extension to Smatch (Cai & Knight

2012).

4 / 24

slide-5
SLIDE 5

Smatch Classic

5 / 24

slide-6
SLIDE 6

Smatch Classic

6 / 24

slide-7
SLIDE 7

Smatch Classic

7 / 24

slide-8
SLIDE 8

AMRICA

8 / 24

slide-9
SLIDE 9

AMRICA

9 / 24

slide-10
SLIDE 10

T-layer to AMR conversion

◮ PCEDT: Large parallel corpus (WSJ)

annotated with t-layer for English and Czech

◮ T-layer to AMR conversion would

provide a large static parallel AMR corpus.

◮ Could be used dynamically to turn a

”t-layer” parser into an AMR parser.

10 / 24

slide-11
SLIDE 11

Why this might work

◮ AMR and t-layer are very similar:

◮ Both abstract away from syntax. ◮ Both make all semantic links in a

sentence in a graph format.

◮ Both do coreference

◮ Various minor structural differences. ◮ AMR is more abstract, makes more

inference.

11 / 24

slide-12
SLIDE 12

“Peter is eager to please”

be.ENUNC Peter ACT eager PAT please PAT #Gen PAT #Cor ACT coreference name Peter

  • p1

person name eager-41 arg0 please-01 arg1 arg0

12 / 24

slide-13
SLIDE 13

Merging of Coreferent Nodes

be.ENUNC Peter ACT eager PAT please PAT ACT #Gen PAT

name Peter

  • p1

person name eager-41 arg0 please-01 arg1 arg0

13 / 24

slide-14
SLIDE 14

Elimination of semantically light words

eager Peter ACT please PAT ACT

name Peter

  • p1

person name eager-41 arg0 please-01 arg1 arg0

14 / 24

slide-15
SLIDE 15

Semantic Roles and Senses

eager-41 Peter arg0 please-01 arg1 arg0

name Peter

  • p1

person name eager-41 arg0 please-01 arg1 arg0

15 / 24

slide-16
SLIDE 16

Add Named Entities

name Peter

  • p1

person name eager-41 arg0 please-01 arg1 arg0 name Peter

  • p1

person name eager-41 arg0 please-01 arg1 arg0

16 / 24

slide-17
SLIDE 17

Conversion Procedures

◮ Converted t-trees to AMR format ◮ Added named entities using NER

systems (Stanford and NameTag)

◮ Tried two strategies for doing more

complex changes to the graphs:

◮ PML-TQ ◮ Tsurgeon

◮ List-based verbalization and semantic

role mapping

17 / 24

slide-18
SLIDE 18

PML-TQ rules

◮ Based on AMR guidelines (generalized) ◮ For copula, attributes, non-core roles . . .

t-node Ib2 functor{={}ACT} formeme{~{}n:.*} t-node Ib_DEL t_lemma{in{{}be},{}become},{}remain}} a-node tag{={}IN} t-node Iw functor{={}PAT} t-node Ir functor{={}PAT} formeme{={}adj:compl} conditions on surface conditions on a t-subtree

LHS (PML-TQ Query) RHS (AMR Subtree)

b2 r ARG0 w ARG1

Guidelines example: The boy is responsible for the work.

A PML-TQ rule

18 / 24

slide-19
SLIDE 19

PML-TQ rules

t-tree zone=en Ondrej ACTn:subj Ondrej be.enunc PREDv:fin was nervous PATadj:compl nervous presentation PATn:about+X about the presentation

Matching t-tree

n2/name "Ondrej"

  • p1

p2/person name n/nervous ARG0 p/presentation ARG1

Conversion result

Matching sentence: Ondrej was nervous about the presentation.

Rule application

19 / 24

slide-20
SLIDE 20

Tsurgeon tree transformation rules

◮ We converted to constituency trees so

as to use a tree tranformation tool, Tsurgeon (Levy and Andrews 2006) to quickly implement hand-written rules.

20 / 24

slide-21
SLIDE 21

Tsurgeon tree transformation rules

◮ Many of the structural differences are

just notational differences: eat CONJ PAT apple PAT banana and eat arg1

  • p2

apple

  • p1

banana and

21 / 24

slide-22
SLIDE 22

List-based Methods

◮ Verbalizations are based on dictionary

look-ups:

◮ beekeeper → person :ARG0-of keep-01

:ARG1 bee

◮ As are complex predications:

give CPHR APP Mary blessing ACT John bless PAT Mary ACT John

22 / 24

slide-23
SLIDE 23

Using Existing Resources

Vallex Propbank Lexicon Other WSJ annotation Lexical Lists Map t-layer roles to AMR roles X X X Verbalize nouns/adjectives X X Introduce inferrable predicates X Named Entities X X

23 / 24

slide-24
SLIDE 24

Results of EN t-to-AMR Conv

Semantic Role Mapping Named Entities Verbalization Lists Smatch Smatch w/o senses Baseline (direct conversion) 20 28 Baseline (direct conversion) X 33 41 Baseline (direct conversion) X X 37 45 Baseline (direct conversion) X X X 40 48 PML-TQ (guidelines-based) X X 35 43 PML-TQ (guidelines-based) X X X 38 47 Tsurgeon (rule-based) X X X 44 52 JAMR 44 45

24 / 24