Dependency Grammars and Parser LING 571 Deep Processing for NLP - - PowerPoint PPT Presentation

dependency grammars and parser
SMART_READER_LITE
LIVE PREVIEW

Dependency Grammars and Parser LING 571 Deep Processing for NLP - - PowerPoint PPT Presentation

Dependency Grammars and Parser LING 571 Deep Processing for NLP October 16, 2019 Shane Steinert-Threlkeld 1 Ambiguity of the Week 2 Roadmap Dependency Grammars Definition Motivation: Limitations of Context-Free Grammars


slide-1
SLIDE 1

Dependency Grammars and Parser

LING 571 — Deep Processing for NLP October 16, 2019 Shane Steinert-Threlkeld

1

slide-2
SLIDE 2

Ambiguity of the Week

2

slide-3
SLIDE 3

Roadmap

  • Dependency Grammars
  • Definition
  • Motivation:
  • Limitations of Context-Free Grammars
  • Dependency Parsing
  • By conversion to CFG
  • By Graph-based models
  • By transition-based parsing
  • HW4 + Mid-term evaluation

3

slide-4
SLIDE 4

Dependency Grammar

  • [P]CFGs:
  • Phrase-Structure Grammars
  • Focus on modeling constituent structure
  • Dependency grammars:
  • Syntactic structure described in terms of
  • Words
  • Syntactic/semantic relations between words

4

slide-5
SLIDE 5

Dependency Parse

  • A Dependency parse is a tree,* where:
  • Nodes correspond to words in string
  • Edges between nodes represent dependency relations
  • Relations may or may not be labeled (aka typed)
  • *: in very special cases, can argue for cycles

5

slide-6
SLIDE 6

Dependency Parse Example:


They hid the letter on the shelf

6 Argument Dependencies Abbreviation Description nsubj nominal subject csubj clausal subject dobj direct object iobj indirect object pobj

  • bject of preposition

Modifier Dependencies Abbreviation Description tmod temporal modifier appos appositional modifier det determiner prep prepositional modifier

They hid

nsubj

letter

dobj

the

det

shelf

  • n

the

det

slide-7
SLIDE 7

Dependency Parse Example:


They hid the letter on the shelf

7 Argument Dependencies Abbreviation Description nsubj nominal subject csubj clausal subject dobj direct object iobj indirect object pobj

  • bject of preposition

Modifier Dependencies Abbreviation Description tmod temporal modifier appos appositional modifier det determiner prep prepositional modifier

They hid

nsubj

letter

dobj

the

det

shelf

  • n

the

det

slide-8
SLIDE 8

Dependency Parse Example:


They hid the letter on the shelf

8 Argument Dependencies Abbreviation Description nsubj nominal subject csubj clausal subject dobj direct object iobj indirect object pobj

  • bject of preposition

Modifier Dependencies Abbreviation Description tmod temporal modifier appos appositional modifier det determiner prep prepositional modifier

They hid

nsubj

letter

dobj

the

det

shelf

  • n

the

det

slide-9
SLIDE 9

Dependency Parse Example:


They hid the letter on the shelf

9 Argument Dependencies Abbreviation Description nsubj nominal subject csubj clausal subject dobj direct object iobj indirect object pobj

  • bject of preposition

Modifier Dependencies Abbreviation Description tmod temporal modifier appos appositional modifier det determiner prep prepositional modifier

They hid

nsubj

letter

dobj

the

det

shelf

  • n

the

det

slide-10
SLIDE 10

Alternative Representation

10

slide-11
SLIDE 11

Why Dependency Grammar?

  • More natural representation for many tasks
  • Clear encapsulation of predicate-argument structure
  • Phrase structure may obscure, e.g. wh-movement
  • Good match for question-answering, relation extraction
  • Who did what to whom?
  • = (Subject) did (theme) to (patient)
  • Helps with parallel relations between roles in questions, and roles in answers

11

slide-12
SLIDE 12
  • Easier handling of flexible or free word order
  • How does CFG handle variation in word order?

Why Dependency Grammar?

12

S PP Prep On NP N Tuesday NP Pron I VP Verb called-in Adv sick S NP Pron I VP Verb called-in Adv sick PP Prep

  • n

NP N Tuesday

S → PP NP VP S → NP VP PP

slide-13
SLIDE 13
  • English has relatively fixed word order
  • Big problem for languages with freer word order

Why Dependency Grammar?

13

S PP Prep On NP N Tuesday NP Pron I VP Verb called-in Adv sick S NP Pron I VP Verb called-in Adv sick PP Prep

  • n

NP N Tuesday

S → PP NP VP S → NP VP PP

slide-14
SLIDE 14
  • How do dependency structures represent the difference?
  • Same structure
  • Relationships are between words, order insensitive

= temporal modifier

Why Dependency Grammar?

14

called-in I sick

  • n

Tuesday

I called in sick on Tuesday

slide-15
SLIDE 15
  • How do dependency structures represent the difference?
  • Same structure
  • Relationships are between words, order insensitive

= temporal modifier

Why Dependency Grammar?

15

call-in did I sick when

when did I call in sick?

slide-16
SLIDE 16

Natural Efficiencies

  • Phrase Structures:
  • Must derive full trees of many non-terminals
  • Dependency Structures:
  • For each word, identify
  • Syntactic head, h
  • Dependency label, d
  • Inherently lexicalized
  • Strong constraints hold between pairs of words

16

slide-17
SLIDE 17

Visualization

  • Web demos:
  • displaCy: https://explosion.ai/demos/displacy
  • Stanford CoreNLP: http://corenlp.run/
  • spaCy and stanfordnlp Python packages have good built-in parsers
  • LaTeX: tikz-dependency (https://ctan.org/pkg/tikz-dependency)

17

slide-18
SLIDE 18

Resources

  • Universal Dependencies:
  • Consistent annotation scheme (i.e. same POS, dependency labels)
  • Treebanks for >70 languages
  • Sizes: German, Czech, Japanese, Russian, French, Arabic, …

18

slide-19
SLIDE 19

Summary

  • Dependency grammars balance complexity and expressiveness
  • Sufficiently expressive to capture predicate-argument structure
  • Sufficiently constrained to allow efficient parsing
  • Still not perfect
  • “On Tuesday I called in sick” vs. “I called in sick on Tuesday”
  • These feel pragmatically different (e.g. topically), might want to represent

difference syntactically.

19

slide-20
SLIDE 20

Roadmap

  • Dependency Grammars
  • Definition
  • Motivation:
  • Limitations of Context-Free Grammars
  • Dependency Parsing
  • By conversion from CFG
  • By Graph-based models
  • By transition-based parsing

20

slide-21
SLIDE 21

Conversion: PS → DS

  • Can convert Phrase Structure (PS) to Dependency Structure (DS)
  • …without the dependency labels
  • Algorithm:
  • Identify all head children in PS
  • Make head of each non-head-child depend on head of head-child
  • Use a head percolation table to determine headedness

21

slide-22
SLIDE 22

Conversion: PS → DS

22

slide-23
SLIDE 23

Conversion: PS → DS

23

had news economic impact little

  • n

markets financial

slide-24
SLIDE 24

Conversion: PS → DS

24

had news

slide-25
SLIDE 25

Conversion: PS → DS

25

had news economic

slide-26
SLIDE 26

Conversion: PS → DS

26

had news economic impact

slide-27
SLIDE 27

Conversion: PS → DS

27

had news economic impact little

slide-28
SLIDE 28

Conversion: PS → DS

28

had news economic impact little

  • n
slide-29
SLIDE 29

Conversion: PS → DS

29

had news economic impact little

  • n

markets

slide-30
SLIDE 30

Conversion: PS → DS

30

had news economic impact little

  • n

markets financial

slide-31
SLIDE 31

Head Percolation Table

  • Finding the head of an NP:
  • If the rightmost word is preterminal, return
  • …else search Right→Left for first child which is NN, NNP, NNPS…
  • …else search Left→Right for first child which is NP
  • …else search Right→Left for first child which is $, ADJP, PRN
  • …else search Right→Left for first child which is CD
  • …else search Right→Left for first child which is JJ, JJS, RB or QP
  • …else return rightmost word.

31

From J&M Page 411, via Collins (1999)

slide-32
SLIDE 32

Conversion: DS → PS

  • Can map any projective dependency tree to PS tree
  • Projective:
  • Does not contain “crossing” dependencies w.r.t. word order

32

A hearing is scheduled

  • n

the issue today .

att att sbj punc vc tmp issue att root

slide-33
SLIDE 33

Non-Projective DS

33

= Projection

A hearing is scheduled

  • n

the issue today . A is scheduled

  • n

the today issue . hearing

slide-34
SLIDE 34

Projective DS

34

= Projection

Economic news had little effect

  • n

financial markets . Economic news had little effect

  • n

markets financial .

slide-35
SLIDE 35

More Non-Projective Parses

35

O to nové většinou nemá ani zájem a taky na to většinou nemá peníze

root

He is mostly not even interested in the new things and in most cases, he has no money for it either.

From McDonald et. al, 2005

John saw a dog yesterday which was a Yorkshire Terrier

root

slide-36
SLIDE 36

Conversion: DS → PS

  • For each node w with outgoing arcs…
  • …convert the subtree w and its dependents t1,…,tn to a new subtree:
  • Nonterminal: Xw
  • Child: w
  • Subtrees t1,…,tn in original sentence order

36

slide-37
SLIDE 37

Conversion: DS → PS

37

Economic news had little effect

  • n

financial markets .

sbj att

  • bj

att att pc att punc root

slide-38
SLIDE 38

Conversion: DS → PS

38

Economic news had little effect

  • n

financial markets .

sbj att

  • bj

att att pc att punc root

slide-39
SLIDE 39

Conversion: DS → PS

39

Economic news had little effect

  • n

financial markets .

sbj att

  • bj

att att pc att punc root

slide-40
SLIDE 40

Conversion: DS → PS

40

Economic news had little effect

  • n

financial markets .

sbj att

  • bj

att att pc att punc root

slide-41
SLIDE 41

Conversion: DS → PS

  • What about labeled dependencies?
  • Can attach labels to nonterminals associated with non-heads
  • e.g. Xlittle → Xlittle:nmod
  • Doesn’t create typical PS trees
  • Does create fully lexicalized, labeled, context-free trees
  • Can be parsed with any standard CFG parser

41

slide-42
SLIDE 42

42

The dog barked at the cat .

root

ROOT Xbarked Xdog Xthe the dog barked Xat at Xcat Xthe the cat X. .

Example from J. Moore, 2013

slide-43
SLIDE 43

Roadmap

  • Dependency Grammars
  • Definition
  • Motivation:
  • Limitations of Context-Free Grammars
  • Dependency Parsing
  • By conversion to CFG
  • By Graph-based models
  • By transition-based parsing

43

slide-44
SLIDE 44

Graph-based Dependency Parsing

  • Goal: Find the highest scoring dependency tree T̂ for sentence S
  • If S is unambiguous, T is the correct parse
  • If S is ambiguous, T is the highest scoring parse
  • Where do scores come from?
  • Weights on dependency edges by learning algorithm
  • Learned from dependency treebank
  • Where are the grammar rules?
  • …there aren’t any! All data-driven.

44

slide-45
SLIDE 45

Graph-based Dependency Parsing

  • Map dependency parsing to Maximum Spanning Tree (MST)
  • Build fully connected initial graph:
  • Nodes: words in sentence to parse
  • Edges: directed edges between all words
  • + Edges from ROOT to all words
  • Identify maximum spanning tree
  • Tree s.t. all nodes are connected
  • Select such tree with highest weight

45

slide-46
SLIDE 46

Graph-based Dependency Parsing

  • Arc-factored model:
  • Weights depend on end nodes & link
  • Weight of tree is sum of participating arcs

46

slide-47
SLIDE 47

Initial Graph: (McDonald et al, 2005b)

  • John saw Mary
  • All words connected: ROOT only has outgoing arcs
  • Goal: Remove arcs to create a tree covering all words
  • Resulting tree is parse

47

ROOT John saw Mary 10 9 30 3 11 20 30 9

slide-48
SLIDE 48

Maximum Spanning Tree

  • McDonald et al, 2005 use variant of Chu-Liu-Edmonds algorithm for MST (CLE)
  • Sketch of algorithm:
  • For each node, greedily select incoming arc with max weight
  • If the resulting set of arcs forms a tree, this is the MST.
  • If not, there must be a cycle.
  • “Contract” the cycle: Treat it as a single vertex
  • Recalculate weights into/out of the new vertex
  • Recursively do MST algorithm on resulting graph
  • Running time: naïve: O(n3); Tarjan: O(n2)
  • Applicable to non-projective graphs

48

ROOT John saw Mary 10 9 30 3 11 20 30 9

slide-49
SLIDE 49

Step 1 & 2

  • Find, for each word, the highest scoring incoming edge.
  • Is it a tree?
  • No, there’s a cycle.
  • Collapse the cycle
  • And re-examine the edges again

49

ROOT John saw Mary 10 9 30 3 11 20 30 9 ROOT John saw Mary 10 9 30 3 11 20 30 9 ROOT John saw Mary 10 9 30 3 11 20 30 9 ROOT John saw Mary 10 9 30 3 11 20 30 9 ROOT John saw Mary ?? 9 30 3 ?? 20 30 9

slide-50
SLIDE 50

ROOT John saw Mary 10 9 30 3 11 20 30 9

Calculating Weights for Collapsed Vertex

50

s( Mary, C ) 11 + 20 = 31

ROOT John saw Mary 10 9 30 3 31 20 30 9

slide-51
SLIDE 51

ROOT John saw Mary 40 9 30 3 11 9 ROOT John saw Mary 10 9 30 3 11 20 30 9

Calculating Weights for Collapsed Vertex

51

s( ROOT, C ) 10 + 30 = 40

slide-52
SLIDE 52

Step 3

52

ROOT John saw Mary 40 9 30 3 31 20 30 9

  • With cycle collapsed, recurse on step 1:
  • Keep highest weighted incoming edge for each edge
  • Is it a tree?
  • Yes!
  • …but must recover collapsed portions.

ROOT John saw Mary 40 9 30 3 31 20 30 9 ROOT John saw Mary 10 9 30 3 11 20 30 9

slide-53
SLIDE 53

MST Algorithm

53

slide-54
SLIDE 54

Learning Weights

  • Weights for arc-factored model learned from dependency treebank
  • Weights learned for tuple ( wi, wj, l )
  • McDonald et al, 2005a employed discriminative ML
  • MIRA (Crammer and Singer, 2003)
  • Operates on vector of local features

54

slide-55
SLIDE 55

Features for Learning Weights

  • Simple categorical features for (wi, L, wj) including:
  • Identity of wi (or char 5-gram prefix), POS of wi
  • Identity of wj (or char 5-gram prefix), POS of wj
  • Label of L, direction of L
  • Number of words between wi, wj
  • POS tag of wi-1, POS tag of wi+1
  • POS tag of wj-1, POS tag of wj+1
  • Features conjoined with direction of attachment and distance between

words

55

slide-56
SLIDE 56

Dependency Parsing

  • Dependency Grammars:
  • Compactly represent predicate–argument structure
  • Lexicalized, localized
  • Natural handling of flexible word order
  • Dependency parsing:
  • Conversion to phrase structure trees
  • Graph-based parsing (MST), efficient non-proj O(n2)
  • Next time: Transition-based parsing

56

slide-57
SLIDE 57

Further Reading

  • Ryan McDonald, Koby Crammer, and Fernando Pereira. 2005. Online Large-Margin Training of Dependency
  • Parsers. In Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics, pages 91–98.
  • May. [link]
  • Ryan McDonald, Fernando Pereira, K. Ribarov, and Jan Hajič. 2005b. Non-projective dependency parsing using

spanning tree algorithms. In Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing, pages 523–530. Association for Computational Linguistics. [link]

  • Sandra Kübler, Ryan McDonald, and Joakim Nivre. 2009. Dependency Parsing. Morgan & Claypool. [link]
  • Jason M. Eisner. 1996. Three new probabilistic models for dependency parsing: An exploration. In Proceedings of

the 16th Conference on Computational Linguistics, pages 340–345. Association for Computational Linguistics. [link]

  • Michael Collins. 1999. Head-Driven Statistical Models For Natural Language Parsing. [link]

57

slide-58
SLIDE 58

HW #4

58

slide-59
SLIDE 59

Probabilistic Parsing

  • Goals:
  • Learn about PCFGs
  • Implement PCKY
  • Analyze Parsing Evaluation
  • Assess improvements to PCFG Parsing

59

slide-60
SLIDE 60

Tasks

  • 1. Train a PCFG
  • 1. Estimate rule probabilities from treebank
  • 2. Treebank is already in CNF
  • 3. More ATIS data from Penn Treebank
  • 2. Build CKY Parser
  • 1. Modify (your) existing CKY implementation

60

slide-61
SLIDE 61

Tasks

  • 3. Evaluation
  • 1. Evaluate your parser using standard metric
  • 2. We will provide evalb program and gold standard
  • 4. Improvement
  • 1. Improve your parser in some way:
  • 1. Coverage
  • 2. Accuracy
  • 3. Speed
  • 2. Evaluate new parser

61

slide-62
SLIDE 62

Improvement Possibilities

  • Coverage:
  • Some test sentences won’t parse as is!
  • Lexical gaps (aka out-of-vocabulary [OOV] tokens)
  • …remember to model the probabilities, too
  • Better context modeling
  • e.g. — Parent Annotation
  • Better Efficiency
  • e.g. — Heuristic Filtering, Beam Search
  • No “cheating” improvements:
  • improvement can’t change training by looking at test data

62

slide-63
SLIDE 63

evalb

  • evalb available in 


dropbox/19-20/571/hw4/tools

  • evalb […] <gold-file> <test-file>
  • evalb --help for more info

63

slide-64
SLIDE 64

Mid-term Evaluation!

  • Please take a few minutes to provide feedback on this course
  • Completely anonymous
  • All feedback valuable; will incorporate things that can be changed
  • Final week: summary, but also some current/future directions topics TBD

64

http://bit.ly/571-aut19-feedback