lecture 15 dependency parsing
play

Lecture 15: Dependency Parsing Kai-Wei Chang CS @ University of - PowerPoint PPT Presentation

Lecture 15: Dependency Parsing Kai-Wei Chang CS @ University of Virginia kw@kwchang.net Couse webpage: http://kwchang.net/teaching/NLP16 CS6501: NLP 1 How to represent the structure CS6501: NLP 2 Dependency trees v Dependency grammar


  1. Lecture 15: Dependency Parsing Kai-Wei Chang CS @ University of Virginia kw@kwchang.net Couse webpage: http://kwchang.net/teaching/NLP16 CS6501: NLP 1

  2. How to represent the structure CS6501: NLP 2

  3. Dependency trees v Dependency grammar describe the structure of sentences as a graph (tree) v Nodes represent words v Edges represent dependencies v Idea goes back to 4 th century BC in ancient India CS6501: NLP 3

  4. Phrase structure (constituent parse) trees v Can be modeled by Context-free grammars v We will see how constituent parse and dependency parse are related CS6501: NLP 4

  5. Context-free grammars PP → P NP PP → P DT N PP → in the garden Non-terminal: DT, N, P, NP, PP, … Terminal: the, a, ball, garden CS6501: NLP 5

  6. Generate sentences by CFG CS6501: NLP 6

  7. Parse tree defined by CFG 1 2 3 4 5 6 Rule 2 Rule 4 & 1 7 Rule 6 CS6501: NLP 7

  8. Example: noun phrases CS6501: NLP 8

  9. Example: verb phrase CS6501: NLP 9

  10. Sentences CS6501: NLP 10

  11. Constituent Parse From: Kevin Gimpel CS6501: NLP 11

  12. Constituent Parse Non-terminal S à NP VP NP à DT NN NP à DT PP à IN NP VP à VBD PP VP à NP VBD NP VP à NP VB Terminal CS6501: NLP 12

  13. Nonterminal in Penn Treebank CS6501: NLP 13

  14. Probabilistic Context-free Grammar Non-terminal 1.0 S à NP VP 0.6 NP à DT NN 0.4 NP à NP PP 1.0 PP à IN NP 0.5 VP à VBD PP 0.2 VP à NP VBD NP 0.3 VP à NP VB Terminal CS6501: NLP 14

  15. Probabilistic Context-free Grammar v PCFG achieves ~73% on PTB v State-of-the art ~92% v Lexicalized PCFG (Collins 1997) CS6501: NLP 15

  16. CS6501: NLP 16

  17. How to decide head? v Usually use deterministic head rules (e.g., Collins head rules) v Define heads in CFG v S → NP VP v VP → VBD NP PP v NP → DT JJ NN From Noah Smith CS6501: NLP 17

  18. Lexical Head Annotation CS6501: NLP 18

  19. Constituent parse → Dependency Parse CS6501: NLP 19

  20. Constituent parse → Dependency Parse CS6501: NLP 20

  21. Head rules can be used to extract dependency parse from a CFG CS6501: NLP 21

  22. Arrow types show the name of grammatical relations CS6501: NLP 22

  23. Dependency parsing v Can be more flexible (non-projective) v English are mostly projective v Some free word order languages (e.g., Czech) are non-projective CS6501: NLP 23

  24. How to build a dependency tree? v There are several approaches v Graph Algorithms v Consider all word pairs v Create a Maximum Spanning Tree for a sentence v Transition-base Approaches v Similar to how we parse a program: Shift-Reduce Parser v Many other approaches… CS6501: NLP 24

  25. Sources of information for DP v Lexical affinities v [ issues → the ] v [ issues → I ] v Distances v Words usually depend on nearby words v Valency of heads v # dependents for a head CS6501: NLP 25

  26. Graph-Based Approaches [McDonald et al. 2005] v Consider all word pairs and assign scores v Score of a tree = sum of score of edges v Can be solve as a MST problem v Chu-Liu-Edmonds CS6501: NLP 26

  27. Transition-based parser v MaltParser (Nivre et al. 2008) v Similar to a Shift-Reduce Parser v But “reduce” actions can create dependencies v The parser has: v A stack 𝜏 – starts with a “Root” symbol v A buffer 𝛾 – starts with the input sentence v A set of dependency arcs A– starts off empty v Use a set of actions to parse sentences v Many possible action sets CS6501: NLP 27

  28. Arc-Eager Dependency Parser v Shift: → Joe Joe likes Mary likes Mary ROOT ROOT v Left-Arc: Precondition: 𝑥 % ≠ Root & (𝑥 ( ,𝑥 % ) ∉ 𝐵 → Joe likes Mary likes Mary ROOT ROOT Joe CS6501: NLP 28

  29. Arc-Eager Dependency Parser v Right-Arc Mary → like like Mary ROOT ROOT Joe Joe v Reduce: Precondition: (𝑥 ( , 𝑥 % ) ∈ 𝐵 Mary Mary → like like ROOT ROOT Joe Joe CS6501: NLP 29

  30. Arc-Eager Dependency Parser v Start: v Conduct a sequence of actions v Terminate with 𝜏, 𝛾 = ∅ CS6501: NLP 30

  31. CS6501: NLP 31

  32. CS6501: NLP 32

  33. CS6501: NLP 33

  34. CS6501: NLP 34

  35. CS6501: NLP 35

  36. CS6501: NLP 36

  37. CS6501: NLP 37

  38. CS6501: NLP 38

  39. CS6501: NLP 39

  40. CS6501: NLP 40

  41. CS6501: NLP 41

  42. CS6501: NLP 42

  43. CS6501: NLP 43

  44. CS6501: NLP 44

  45. CS6501: NLP 45

  46. CS6501: NLP 46

  47. CS6501: NLP 47

  48. It’s your turn v Happy children like to play with their friend . v Shift → Left-arc → Shift → Left-arc → Right-arc → Shift → Left-arc → Right-arc → Right-arc → Shift → Left-arc → Right-arc → Reduce*3 → Right-arc → Reduce*3 CS6501: NLP 48

  49. From Chris Manning CS6501: NLP 49

  50. CS6501: NLP 50

  51. Structured Prediction –beyond sequence tagging Assign values to a set of interdependent output variables Task Input Output Part-of-speech They operate Pronoun Verb Noun And Noun Tagging ships and banks. Dependency They operate Root They operate ships and banks . Parsing ships and banks. 51

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend